DA-RAG: Dynamic Attributed Community Search for Retrieval-Augmented Generation
Owing to their unprecedented comprehension capabilities, large language models (LLMs) have become indispensable components of modern web search engines. From a technical perspective, this integration represents retrieval-augmented generation (RAG), which enhances LLMs by grounding them in external knowledge bases. A prevalent technical approach in this context is graph-based RAG (G-RAG). However, current G-RAG methodologies frequently underutilize graph topology, predominantly focusing on low-order structures or pre-computed static communities. This limitation affects their effectiveness in addressing dynamic and complex queries. Thus, we propose DA-RAG, which leverages attributed community search (ACS) to extract relevant subgraphs based on the queried question dynamically. DA-RAG captures high-order graph structures, allowing for the retrieval of self-complementary knowledge. Furthermore, DA-RAG is equipped with a chunk-layer oriented graph index, which facilitates efficient multi-granularity retrieval while significantly reducing both computational and economic costs. We evaluate DA-RAG on multiple datasets, demonstrating that it outperforms existing RAG methods by up to 40% in head-to-head comparisons across four metrics while reducing index construction time and token overhead by up to 37% and 41%, respectively.
💡 Research Summary
The paper introduces DA‑RAG, a novel framework for Retrieval‑Augmented Generation (RAG) that dynamically extracts query‑relevant subgraphs from a knowledge graph by leveraging high‑order graph structure and attribute information. Traditional graph‑based RAG (G‑RAG) methods either rely on low‑order adjacency (e.g., 1‑hop neighbors) or pre‑computed static communities, which limits their ability to handle complex, evolving queries. DA‑RAG reframes subgraph retrieval as an Embedding‑Attributed Community Search (EACS) problem, adapting the Attributed Community Search (ACS) paradigm from graph analytics to the RAG setting.
The system consists of two main phases. In the offline phase, documents are first split into semantic chunks using a large language model (LLM). For each chunk, a local knowledge graph (KG) is extracted, and the chunk is represented by a title, description, and embedding. Three interconnected layers form the index: (1) a Semantic Chunk Layer (L_C) where each node is a chunk; (2) a Knowledge Graph Layer (L_KG) containing entity nodes and relational edges; (3) a Similarity Layer (L_S) that connects the same entity nodes based on embedding similarity via a k‑Nearest‑Neighbor graph. Inter‑layer edges link chunk nodes to their constituent entities, creating a hierarchical, multi‑perspective index without costly clustering.
During online retrieval, a coarse‑to‑fine strategy is employed. The query embedding q is first used to run EACS on the coarse Chunk Layer, yielding an initial community H_C that provides a high‑level context. H_C guides the selection of relevant entities, forming two much smaller working subgraphs in L_KG and L_S (G_work_KG and G_work_S). EACS is then applied again on each working subgraph to obtain fine‑grained communities H_KG and H_S. The core of EACS is a k‑truss based cohesion constraint: the retrieved community must be a k‑truss, guaranteeing a minimum number of triangles per edge and thus strong internal connectivity. An adaptive mechanism determines the appropriate k value based on query complexity and subgraph density. Additionally, a custom relevance score mitigates the “free‑rider” effect, ensuring that only nodes semantically aligned with the query are retained.
Experiments on four public datasets (e.g., WikiQA, Natural Questions) and four evaluation metrics (accuracy, BLEU, R‑Recall, token cost) compare DA‑RAG against several state‑of‑the‑art G‑RAG baselines such as LightRAG, HippoRAG, GraphRAG, and ArchRAG. DA‑RAG consistently outperforms baselines, achieving 30‑40 % higher answer quality and reducing index construction time by 37 % and token overhead by 41 %. Ablation studies confirm that both the multi‑layer index and the dynamic k‑truss community search contribute substantially to these gains.
The authors acknowledge limitations: exact k‑truss computation can be expensive on very large graphs, the quality of semantic chunking depends on the underlying LLM and may degrade on domain‑specific texts, and the current design assumes a static knowledge graph, making frequent updates costly. Future work is proposed on approximate k‑truss algorithms, domain‑adaptive chunking, and incremental index updates.
In summary, DA‑RAG presents a compelling advancement for graph‑based RAG by integrating high‑order structural cohesion with semantic guidance, delivering superior answer quality while markedly improving indexing and retrieval efficiency. This work paves the way for more nuanced, cost‑effective LLM‑augmented search systems.
Comments & Academic Discussion
Loading comments...
Leave a Comment