MarkTechPost@AI 03月04日
HippoRAG 2: Advancing Long-Term Memory and Contextual Retrieval in Large Language Models
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

HippoRAG 2是由俄亥俄州立大学和伊利诺伊大学厄巴纳-香槟分校的研究人员开发的,旨在增强大型语言模型(LLM)的长期记忆和上下文检索能力。它通过改进上下文整合和检索,提升了事实召回、理解能力和关联记忆。该框架基于神经生物学原理,包含人工新皮层(LLM)、海马旁区编码器和开放知识图谱(KG)。离线状态下,LLM从段落中提取三元组,连接同义词,整合概念和上下文信息。在线状态下,查询通过基于嵌入的检索映射到相关三元组,然后使用个性化PageRank(PPR)进行上下文感知选择。HippoRAG 2引入了识别记忆来过滤三元组,并通过将查询链接到三元组来加深上下文,从而增强多跳推理并提高QA任务的检索准确性。实验表明,HippoRAG 2在关联记忆任务中比领先的嵌入模型提高了7%,同时保持了强大的事实和上下文理解能力。

🧠 HippoRAG 2是一种受神经生物学启发的LLM长期记忆框架,它通过改进上下文整合和检索来增强原始的HippoRAG模型。它由人工新皮层(LLM),海马旁区编码器和开放知识图谱(KG)组成。

🔗 HippoRAG 2通过离线LLM提取三元组,连接同义词,并整合概念和上下文信息,从而实现知识的结构化。在线状态下,查询被映射到相关三元组,并利用个性化PageRank(PPR)进行上下文感知的选择。

🔍 HippoRAG 2集成了识别记忆,用于过滤三元组,并通过将查询链接到三元组来深化上下文,从而增强多跳推理并提高QA任务的检索准确性。实验结果表明,这种方法在多跳任务中表现出色。

📊 实验评估表明,HippoRAG 2在各种基准测试中表现出强大的鲁棒性,优于现有的结构增强RAG方法。尤其是在多跳QA任务中,它利用Llama-3.3-70B-Instruct和NV-Embed-v2,展示了增强的检索和响应准确性。

LLMs face challenges in continual learning due to the limitations of parametric knowledge retention, leading to the widespread adoption of RAG as a solution. RAG enables models to access new information without modifying their internal parameters, making it a practical approach for real-time adaptation. However, traditional RAG frameworks rely heavily on vector retrieval, which limits their ability to capture complex relationships and associations in knowledge. Recent advancements have integrated structured data, such as knowledge graphs, to enhance reasoning capabilities, improving sense-making and multi-hop connections. While these methods offer improvements in contextual understanding, they often compromise performance on simpler factual recall tasks, highlighting the need for more refined approaches.

Continual learning strategies for LLMs typically fall into three categories: continual fine-tuning, model editing, and non-parametric retrieval. Fine-tuning periodically updates model parameters with new data but is computationally expensive and prone to catastrophic forgetting. Model editing modifies specific parameters for targeted knowledge updates, but its effects remain localized. In contrast, RAG dynamically retrieves relevant external information at inference time, allowing for efficient knowledge updates without altering the model’s parameters. Advanced RAG frameworks, such as GraphRAG and LightRAG, enhance retrieval by structuring knowledge into graphs, improving the model’s ability to synthesize complex information. HippoRAG 2 refines this approach by leveraging structured retrieval while minimizing errors from LLM-generated noise, balancing sense-making and factual accuracy.

HippoRAG 2, developed by researchers from The Ohio State University and the University of Illinois Urbana-Champaign, enhances RAG by improving factual recall, sense-making, and associative memory. Building upon HippoRAG’s Personalized PageRank algorithm, it integrates passages more effectively and refines online LLM utilization. This approach achieves a 7% improvement in associative memory tasks over leading embedding models while maintaining strong factual and contextual understanding. Extensive evaluations show its robustness across various benchmarks, outperforming existing structure-augmented RAG methods. HippoRAG 2 significantly advances non-parametric continual learning, bringing AI systems closer to human-like long-term memory capabilities.

HippoRAG 2 is a neurobiologically inspired long-term memory framework for LLMs, enhancing the original HippoRAG by improving context integration and retrieval. It comprises an artificial neocortex (LLM), a parahippocampal region encoder, and an open knowledge graph (KG). Offline, an LLM extracts triples from passages, linking synonyms and integrating conceptual and contextual information. Online, queries are mapped to relevant triples using embedding-based retrieval, followed by Personalized PageRank (PPR) for context-aware selection. HippoRAG 2 introduces recognition memory for filtering triples and deeper contextualization by linking queries to triples, enhancing multi-hop reasoning and improving retrieval accuracy for QA tasks.

The experimental setup includes three baseline categories: (1) classical retrievers such as BM25, Contriever, and GTR, (2) large embedding models like GTE-Qwen2-7B-Instruct, GritLM-7B, and NV-Embed-v2, and (3) structure-augmented RAG models, including RAPTOR, GraphRAG, LightRAG, and HippoRAG. The evaluation spans three key challenge areas: simple QA (factual recall), multi-hop QA (associative reasoning), and discourse understanding (sense-making). Metrics include passage recall@5 for retrieval and F1 scores for QA. HippoRAG 2, leveraging Llama-3.3-70B-Instruct and NV-Embed-v2, outperforms prior models, particularly in multi-hop tasks, demonstrating enhanced retrieval and response accuracy with its neuropsychology-inspired approach.

In conclusion, the ablation study evaluates the impact of linking, graph construction, and triple filtering methods, showing that deeper contextualization significantly improves HippoRAG 2’s performance. The query-to-triple approach outperforms others, enhancing Recall@5 by 12.5% over NER-to-node. Adjusting reset probabilities in PPR balances phrase and passage nodes, optimizing retrieval. HippoRAG 2 integrates seamlessly with dense retrievers, consistently outperforming them. Qualitative analysis highlights superior multi-hop reasoning. Overall, HippoRAG 2 enhances retrieval and reasoning by leveraging Personalized PageRank, deeper passage integration, and LLMs, offering advancements in long-term memory modeling. Future work may explore graph-based retrieval for improved episodic memory in conversations.


Check out the Paper and GitHub Page. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 80k+ ML SubReddit.

Recommended Read- LG AI Research Releases NEXUS: An Advanced System Integrating Agent AI System and Data Compliance Standards to Address Legal Concerns in AI Datasets

The post HippoRAG 2: Advancing Long-Term Memory and Contextual Retrieval in Large Language Models appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

HippoRAG 2 长期记忆 上下文检索 大型语言模型 知识图谱
相关文章