MarkTechPost@AI 07月21日 10:05
MIRIX: A Modular Multi-Agent Memory System for Enhanced Long-Term Reasoning and Personalization in LLM-Based Agents
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

MIRIX是一个创新的模块化多代理记忆系统,旨在为大语言模型(LLM)代理提供强大的长期记忆能力。与传统的纯文本记忆系统不同,MIRIX整合了跨模态的结构化记忆,包括视觉信息,并采用协调的多代理架构进行记忆管理。它通过“主动检索”机制,能够自主推断主题、检索相关记忆并注入提示,从而减少对过时模型知识的依赖。MIRIX在多模态和对话基准测试中表现出色,在视觉问答和长对话记忆方面均超越现有技术。该系统还支持可穿戴设备,并设想了“记忆市场”以实现安全的记忆共享和个性化。

✨ MIRIX系统核心在于其模块化多代理记忆架构,包含核心记忆、事件记忆、语义记忆、过程记忆、资源记忆和知识库六个专门组件,由元记忆管理器协调,以实现更全面、结构化的信息存储和管理。这超越了简单的文本存储,能够处理用户偏好、交互事件、知识图谱、工作流程、外部资源引用以及敏感凭证等多种信息类型。

🚀 MIRIX的“主动检索”机制是其关键创新之一。它能够在用户输入时自主推断主题,从所有记忆组件中检索相关条目,并通过多种检索策略(如embedding_match, bm25_match, string_match)优化信息提取,最终将检索到的数据标记并注入系统提示,从而显著增强答案的依据性和上下文相关性。

🖼️ MIRIX在多模态记忆方面表现突出,能够通过屏幕截图捕捉视觉信息,并利用Gemini API进行流式上传和处理,实现低于5秒的低延迟记忆更新。在ScreenshotVQA基准测试中,MIRIX的LLM-as-a-Judge准确率比传统的检索增强生成(RAG)基线高出35%,同时大幅降低了检索存储需求。

💬 在对话记忆方面,MIRIX通过LOCOMO基准测试验证了其在长对话中的表现。MIRIX达到了85.38%的平均准确率,显著优于LangMem和Mem0等开源系统,证明了其在保持对话连贯性和上下文感知能力方面的优势。

💡 MIRIX不仅支持跨平台应用,还为轻量级AI可穿戴设备设计了高效的模块化架构,并提出了“记忆市场”的愿景,一个去中心化的生态系统,允许用户安全地共享、货币化记忆,并实现AI的协作个性化,同时强调用户数据主权和隐私保护。

Recent developments in LLM agents have largely focused on enhancing capabilities in complex task execution. However, a critical dimension remains underexplored: memory—the capacity of agents to persist, recall, and reason over user-specific information across time. Without persistent memory, most LLM-based agents remain stateless, unable to build context beyond a single prompt, limiting their usefulness in real-world settings where consistency and personalization are essential.

To address this, MIRIX AI introduces MIRIX, a modular multi-agent memory system explicitly designed to enable robust long-term memory for LLM-based agents. Unlike flat, purely text-centric systems, MIRIX integrates structured memory types across modalities—including visual input—and is built upon a coordinated multi-agent architecture for memory management.

Core Architecture and Memory Composition

MIRIX features six specialized, compositional memory components, each governed by a corresponding Memory Manager:

Meta Memory Manager orchestrates the activities of these six specialized managers, enabling intelligent message routing, hierarchical storage, and memory-specific retrieval operations. Additional agents—with roles like chat and interface—collaborate within this architecture.

Active Retrieval and Interaction Pipeline

A core innovation of MIRIX is its Active Retrieval mechanism. On user input, the system first autonomously infers a topic, then retrieves relevant memory entries from all six components, and finally tags the retrieved data for contextual injection into the resulting system prompt. This process decreases reliance on outdated parametric model knowledge and provides much stronger answer grounding.

Multiple retrieval strategies—including embedding_matchbm25_match, and string_match—are available, ensuring accurate and context-aware access to memory. The architecture allows for further expansion of retrieval tools as needed.

System Implementation and Application

MIRIX is deployed as a cross-platform assistant application developed with React-Electron (for the UI) and Uvicorn (for the backend API). The assistant monitors screen activity by capturing screenshots every 1.5 seconds; only non-redundant screens are kept, and memory updates are triggered in batches after collecting 20 unique screenshots (approximately once per minute). Uploads to the Gemini API are streaming, enabling efficient visual data processing and sub-5-second latency for updating memory from visual inputs.

Users interact through a chat interface, which dynamically draws on the agent’s memory components to generate context-aware, personalized responses. Semantic and procedural memories are rendered as expandable trees or lists, providing transparency and allowing users to audit and inspect what the agent “remembers” about them.

Evaluation on Multimodal and Conversational Benchmarks

MIRIX is validated on two rigorous tasks:

    ScreenshotVQA: A visual question-answering benchmark requiring persistent, long-term memory over high-resolution screenshots. MIRIX outperforms retrieval-augmented generation (RAG) baselines—specifically SigLIP and Gemini—by 35% in LLM-as-a-Judge accuracy, while reducing retrieval storage needs by 99.9% compared to text-heavy methods.LOCOMO: A textual benchmark assessing long-form conversation memory. MIRIX achieves 85.38% average accuracy, outperforming strong open-source systems such as LangMem and Mem0 by over 8 points, and approaching full-context sequence upper bounds.

The modular design enables high performance across both multimodal and text-only inference domains.

Use Cases: Wearables and the Memory Marketplace

MIRIX is designed for extensibility, with support for lightweight AI wearables—including smart glasses and pins—via its efficient, modular architecture. Hybrid deployment allows both on-device and cloud-based memory handling, while practical applications include real-time meeting summarization, granular location and context recall, and dynamic modeling of user habits.

A visionary feature of MIRIX is the Memory Marketplace: a decentralized ecosystem enabling secure memory sharing, monetization, and collaborative AI personalization between users. The Marketplace is designed with fine-grained privacy controls, end-to-end encryption, and decentralized storage to ensure data sovereignty and user self-ownership.

Conclusion

MIRIX represents a significant step toward endowing LLM-based agents with human-like memory. Its structured, multi-agent compositional architecture enables robust memory abstraction, multimodal support, and real-time, contextually grounded reasoning. With empirical gains across challenging benchmarks and an accessible, cross-platform application interface, MIRIX sets a new standard for memory-augmented AI systems.

FAQs

1. What makes MIRIX different from existing memory systems like Mem0 or Zep?
MIRIX introduces multi-component, compositional memory (beyond text passage storage), multimodal support (including vision), and a multi-agent retrieval architecture for more scalable, accurate, and context-rich long-term memory management.

2. How does MIRIX ensure low-latency memory updates from visual inputs?
By using streaming uploads in combination with Gemini APIs, MIRIX is able to update screenshot-based visual memory with under 5 seconds latency, even during active user sessions.

3. Is MIRIX compatible with closed-source LLMs like GPT-4?
Yes. Since MIRIX operates as an external system (and not as a model plugin or retrainer), it can augment any LLM, regardless of its base architecture or licensing, including GPT-4, Gemini, and other proprietary models.


Check out the Paper, GitHub and Project. All credit for this research goes to the researchers of this project.

Sponsorship Opportunity: Reach the most influential AI developers in US and Europe. 1M+ monthly readers, 500K+ community builders, infinite possibilities. [Explore Sponsorship]

The post MIRIX: A Modular Multi-Agent Memory System for Enhanced Long-Term Reasoning and Personalization in LLM-Based Agents appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

MIRIX 大语言模型 LLM代理 长期记忆 多模态记忆
相关文章