Search papers, labs, and topics across Lattice.
This paper introduces MDER-DR, a novel retrieval-augmented generation (RAG) framework for knowledge graph-based question answering that addresses the limitations of standard KG indexing approaches. MDER (Map-Disambiguate-Enrich-Reduce) generates context-derived triple descriptions and integrates them with entity-level summaries to improve indexing, while DR (Decompose-Resolve) decomposes user queries into resolvable triples for iterative reasoning. Experiments on standard and domain-specific benchmarks demonstrate that MDER-DR significantly outperforms standard RAG baselines, achieving improvements of up to 66% and maintaining cross-lingual robustness.
Forget brittle KG traversals: MDER-DR's entity-centric summaries and decomposed queries boost multi-hop QA accuracy by up to 66% over standard RAG.
Retrieval-Augmented Generation (RAG) over Knowledge Graphs (KGs) suffers from the fact that indexing approaches may lose important contextual nuance when text is reduced to triples, thereby degrading performance in downstream Question-Answering (QA) tasks, particularly for multi-hop QA, which requires composing answers from multiple entities, facts, or relations. We propose a domain-agnostic, KG-based QA framework that covers both the indexing and retrieval/inference phases. A new indexing approach called Map-Disambiguate-Enrich-Reduce (MDER) generates context-derived triple descriptions and subsequently integrates them with entity-level summaries, thus avoiding the need for explicit traversal of edges in the graph during the QA retrieval phase. Complementing this, we introduce Decompose-Resolve (DR), a retrieval mechanism that decomposes user queries into resolvable triples and grounds them in the KG via iterative reasoning. Together, MDER and DR form an LLM-driven QA pipeline that is robust to sparse, incomplete, and complex relational data. Experiments show that on standard and domain specific benchmarks, MDER-DR achieves substantial improvements over standard RAG baselines (up to 66%), while maintaining cross-lingual robustness. Our code is available at https://github.com/DataSciencePolimi/MDER-DR_RAG.