Retrieval-augmented generation (RAG) has demonstrated its ability to enhance Large Language Models (LLMs) by integrating external knowledge sources. However, multi-hop questions, which require the identification of multiple knowledge targets to form a synthesized answer, raise new challenges for RAG systems. Under the multi-hop settings, existing methods often struggle to fully understand the questions with complex semantic structures and are susceptible to irrelevant noise during the retrieval of multiple information targets. To address these limitations, we propose a novel graph representation learning framework for multi-hop question retrieval. We first introduce a Multi-information Level Knowledge Graph (Multi-L KG) to model various information levels for a more comprehensive understanding of multi-hop questions. Based on this, we design a Query-Specific Graph Neural Network (QSGNN) for representation learning on the Multi-L KG. QSGNN employs intra/inter-level message passing mechanisms, and in each message passing the information aggregation is guided by the query, which not only facilitates multi-granular information aggregation but also significantly reduces the impact of noise. To enhance its ability to learn robust representations, we further propose two synthesized data generation strategies for pre-training the QSGNN. Extensive experimental results demonstrate the effectiveness of our framework in multi-hop scenarios, especially in high-hop questions the improvement can reach 33.8\%. The code is available at: https://github.com/Jerry2398/QSGNN.
翻译:检索增强生成(RAG)通过整合外部知识源,已展现出增强大语言模型(LLM)能力的显著效果。然而,多跳问题——需要识别多个知识目标以形成综合答案——为RAG系统带来了新的挑战。在多跳场景下,现有方法往往难以充分理解具有复杂语义结构的问题,且在检索多个信息目标时易受无关噪声干扰。为应对这些局限,本文提出一种面向多跳问题检索的新型图表示学习框架。我们首先引入多信息层级知识图谱(Multi-L KG),通过建模不同信息层级以更全面地理解多跳问题。在此基础上,我们设计了查询特定图神经网络(QSGNN)以在Multi-L KG上进行表示学习。QSGNN采用层级内/层级间消息传递机制,且每次消息传递中的信息聚合均由查询引导,这不仅促进了多粒度信息聚合,也显著降低了噪声影响。为增强其学习鲁棒表示的能力,我们进一步提出两种合成数据生成策略对QSGNN进行预训练。大量实验结果表明,我们的框架在多跳场景中具有显著有效性,尤其在高层级跳数问题上性能提升可达33.8%。代码已开源:https://github.com/Jerry2398/QSGNN。