MarkTechPost@AI 15小时前
Graph-R1: An Agentic GraphRAG Framework for Structured, Multi-Turn Reasoning with Reinforcement Learning
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Graph-R1是一项由多所知名大学联合推出的创新性Agentic GraphRAG框架,它通过端到端的强化学习,显著提升了大型语言模型(LLMs)在知识密集型应用中的事实准确性。与传统基于文本块的检索方法不同,Graph-R1构建了轻量级的知识超图,能够更丰富地编码实体间的语义关系。其核心亮点在于多轮智能检索过程,允许模型在“思考-检索-再思考-生成”的循环中动态优化知识路径,克服了传统单次检索的局限性。通过端到端强化学习,Graph-R1能够同时优化格式遵循、相关性和答案准确性,从而生成更具结构性、可信赖且高度相关的答案。实验结果表明,Graph-R1在多项问答任务上取得了远超现有基线的性能,尤其在准确性、相关性和逻辑连贯性方面表现突出,为下一代智能、知识驱动的语言模型系统开辟了新方向。

💡 Graph-R1构建了一种轻量级的知识超图表示,通过LLM驱动的n-ary关系提取,编码了比传统实体关系图更丰富、语义更强的多向关系,从而增强了智能体的推理能力,同时保持了成本和计算要求的可控性。该方法在构建效率和成本上优于现有图式方法,仅需5.69秒和2.81美元/千tokens即可生成包含大量节点和边的语义丰富的图谱。

🔄 Graph-R1将检索过程设计为多轮智能交互循环(“思考-检索-再思考-生成”),使智能体能够自适应地查询和精炼知识路径,而非依赖一次性检索。智能体在每一步都能自主决定是继续探索还是终止生成答案,并通过实体检索和直接超边检索的融合,提高了检索最相关知识的概率。

🚀 通过集成端到端的强化学习(GRPO),Graph-R1能够统一优化格式遵循、相关性和答案准确性等多个维度。这种奖励机制指导智能体发展出与知识结构和输出质量紧密结合的通用化推理策略,确保生成的答案不仅在语义上准确,而且在结构上符合要求,从而显著提升了生成质量。

📊 Graph-R1在六个标准问答数据集上的基准测试结果显示,其平均F1得分高达57.82(使用Qwen2.5-7B模型),大幅超越了所有现有基线方法。消融实验证明了超图构建、多轮推理和强化学习优化的必要性。同时,Graph-R1在检索和生成成本上表现出显著优势,响应时间短且成本极低,在准确性、相关性和逻辑连贯性等生成质量维度上也获得了最高评分。

🌐 Graph-R1框架展现出强大的泛化能力,在跨数据集的分布外(O.O.D.)设置下仍能保持稳健的性能,O.O.D./I.I.D.比率常高于85%。这表明其在不同领域和数据集之间具有良好的迁移性,尤其适用于对事实准确性、推理透明度有高要求的医疗、法律及企业知识自动化等领域。

Introduction

Large Language Models (LLMs) have set new benchmarks in natural language processing, but their tendency for hallucination—generating inaccurate outputs—remains a critical issue for knowledge-intensive applications. Retrieval-Augmented Generation (RAG) frameworks attempt to solve this by incorporating external knowledge into language generation. However, traditional RAG approaches rely on chunk-based retrieval, which limits their ability to represent complex semantic relationships. Entity-relation graph-based RAG methods (GraphRAG) address some structural limitations, but still face high construction cost, one-shot retrieval inflexibility, and dependence on long-context reasoning and carefully crafted prompts.

Researchers from Nanyang Technological University, National University of Singapore, Beijing Institute of Computer Technology and Application, and Beijing Anzhen Hospital have introduced Graph-R1, an agentic GraphRAG framework powered by end-to-end reinforcement learning.

Image source: https://arxiv.org/pdf/2507.21892v1

Core Innovations of Graph-R1

1. Lightweight Knowledge Hypergraph Construction

Graph-R1 constructs knowledge as a hypergraph, where each knowledge segment is extracted using LLM-driven n-ary relation extraction. This approach encodes richer and more semantically grounded relationships, boosting agentic reasoning capabilities while maintaining manageable cost and computational requirements.

2. Multi-Turn Agentic Retrieval Process

Graph-R1 models retrieval as a multi-turn interaction loop (“think-retrieve-rethink-generate”), allowing the agent to adaptively query and refine its knowledge path, unlike previous methods that use one-shot retrieval.

3. End-to-End Reinforcement Learning Optimization

Graph-R1 uses Group Relative Policy Optimization (GRPO) for end-to-end RL, integrating rewards for format adherence, relevance, and answer correctness. This unified reward guides agents to develop generalizable reasoning strategies tightly aligned with both the knowledge structure and output quality.

Key Findings

Benchmarking on RAG QA Tasks

Graph-R1 was evaluated across six standard QA datasets (2WikiMultiHopQA, HotpotQA, Musique, Natural Questions, PopQA, TriviaQA).

MethodAvg. F1 (Qwen2.5-7B)
NaiveGeneration13.87
StandardRAG15.89
GraphRAG24.87
HyperGraphRAG29.40
Search-R146.19
R1-Searcher42.29
Graph-R157.82

Ablation Analysis

Component ablation demonstrates that removing hypergraph construction, multi-turn reasoning, or RL optimization dramatically reduces performance, validating the necessity of each module within Graph-R1.

Retrieval and Efficiency

Generation Quality

Graph-R1’s generation quality is evaluated across seven dimensions—comprehensiveness, knowledgeability, correctness, relevance, diversity, logical coherence, factuality—and consistently outperforms all RL-based and graph-based baselines, achieving top scores in correctness (86.9), relevance (95.2), and coherence (88.5).

Generalizability

Cross-validation on out-of-distribution (O.O.D.) settings reveals that Graph-R1 maintains robust performance across datasets, with O.O.D./I.I.D. ratios often above 85%, demonstrating strong domain generalization properties.

Theoretical Guarantees

Graph-R1 is supported by information-theoretic analyses:

Algorithmic Workflow (High-Level)

    Knowledge Hypergraph Extraction: LLM extracts n-ary relations to build entity and hyperedge sets.Multi-turn Agentic Reasoning: The agent alternates between reflective thinking, querying, hypergraph retrieval (entity and hyperedge dual paths), and synthesis.GRPO Optimization: RL policy is updated using sampled trajectories and reward normalization, enforcing structure and answer correctness.

Conclusion

Graph-R1 demonstrates that integrating hypergraph-based knowledge representation, agentic multi-turn reasoning, and end-to-end RL delivers unprecedented gains in factual QA performance, retrieval efficiency, and generation quality, charting the path for next-generation agentic and knowledge-driven LLM systems.


FAQ 1: What is the key innovation of Graph-R1 compared to earlier GraphRAG and RAG systems?

Graph-R1 introduces an agentic framework where retrieval is modeled as a multi-turn interaction rather than a single one-shot process. Its main innovations are:

FAQ 2: How does Graph-R1’s retrieval and generation efficiency compare to previous methods?

Graph-R1 is significantly more efficient and effective in both retrieval and answer generation:

FAQ 3: In which scenarios or domains is the Graph-R1 framework most applicable?

Graph-R1 is ideal for complex knowledge-intensive applications demanding both factual accuracy and reasoning transparency, such as:


Check out the Paper here and GitHub Page. Feel free to check out our GitHub Page for Tutorials, Codes and Notebooks

The post Graph-R1: An Agentic GraphRAG Framework for Structured, Multi-Turn Reasoning with Reinforcement Learning appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Graph-R1 Agentic GraphRAG 强化学习 大型语言模型 知识图谱
相关文章