MarkTechPost@AI 2024年10月13日
OPTIMA: Enhancing Efficiency and Effectiveness in LLM-Based Multi-Agent Systems
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

本文介绍了LLM在多任务中的重要性及LLM-based MAS的发展,指出其面临的挑战。清华大学和北京邮电大学的研究人员提出OPTIMA框架,用以提高LLM-based MAS的通信效率和任务效果。该框架在多个任务中表现出色,优于基线方法,具有重要意义。

🧠LLM在多种任务中表现出色,其在多智能体系统中的应用备受关注,但面临通信效率和系统整体性能优化的挑战。

💡清华大学和北京邮电大学的研究人员提出OPTIMA框架,采用迭代生成、排序、选择和训练范式,利用平衡任务性能、令牌效率和通信可读性的奖励函数。

📊OPTIMA在信息交换和辩论多智能体设置中进行评估,在不同任务中均优于基线方法,其变体在信息交换任务中表现尤为突出。

🌟OPTIMA的成功得益于其关键创新,包括迭代训练技术、平衡奖励函数和受蒙特卡罗树搜索启发的数据生成方法,对改进通信效率和任务性能有重要作用。

Large Language Models (LLMs) have gained significant attention for their versatility in various tasks, from natural language processing to complex reasoning. A promising application of these models is the development of autonomous multi-agent systems (MAS), which aim to utilize the collective intelligence of multiple LLM-based agents for collaborative problem-solving. However, LLM-based MAS faces two critical challenges: achieving efficient inter-agent communication to minimize computational costs and optimizing the collective performance of the system as a cohesive unit. Current methods fail to solve these challenges, resulting in overly detailed exchanges that increase token usage, longer inference times, and higher computational costs.

Existing methods discussed in this paper include LLM-based MAS and Iterative Refinement of LLMs. The Role-playing in LLM-based MAS for complex reasoning, collaborative software development, and embodied agent interactions have shown promise. Current research has shown that increasing the number and diversity of agents can lead to performance gains. Moreover, iterative refinement paradigms, such as self-reflection mechanisms and parameter updates for example ReST and STaR, have been developed for individual LLMs. However, iterative refinement is yet to be explored in the LLM-based MAS context. These methods are effective in single-agent scenarios but ineffectively adapted to optimize the collective performance of multi-agent systems.

Researchers from Tsinghua University and Beijing University of Posts and Telecommunications have proposed OPTIMA, a novel framework designed to enhance both communication efficiency and task effectiveness in LLM-based MAS. It employs an iterative generate, rank, select, and train paradigm, utilizing a reward function that balances task performance, token efficiency, and communication readability. OPTIMA uses Monte Carlo Tree Search-inspired techniques for data generation, treating conversation turns as tree nodes to explore diverse interaction paths. The method addresses the fundamental challenges in LLM-based MAS, potentially leading to more scalable, efficient, and effective multi-agent systems.

OPTIMA is evaluated on information exchange (IE) and debate multi-agent settings. The IE setting uses datasets like HotpotQA, CBT, etc, with contexts split between agents to support information exchange. The debate setting uses GSM8K, MATH, ARC-C, and MMLU, with one agent as a solver and another as a critic. OPTIMA is compared against single-agent approaches like Chain-of-Thought and Self-Consistency, and multi-agent baselines such as Multi-Agent Debate and AutoForm. Llama 3 8B serves as the base model, focusing on two-agent scenarios and no external tools, allowing a clear analysis of the key elements of multi-agent communication and collaboration.

OPTIMA consistently outperforms baseline methods in both effectiveness and efficiency across different tasks. Its variants show substantial gains in Information Exchange (IE) tasks, especially in multi-hop reasoning scenarios. The iSFT-DPO variant stands out, delivering the best performance while greatly reducing token usage compared to the top baseline. For instance, it improves the F1 score by 38.3% on 2WMHQA while using only 10% of the tokens required by Multi-Agent Debate. In debate tasks, OPTIMA shows better performance and token efficiency for ARC-C and MMLU, while maintaining comparable performance with higher efficiency for MATH and GSM8k tasks.

In conclusion, researchers introduced OPTIMA, a method to enhance communication efficiency and task effectiveness in LLM-based MAS. It demonstrates consistent superiority over single-agent and multi-agent baselines across various tasks. The framework’s key innovations, including iterative training techniques, a balanced reward function, and an MCTS-inspired approach for data generation, contribute to its success in improving communication efficiency and task performance. OPTIMA’s potential to enhance inference scaling laws and adapt to out-of-distribution tasks highlights the importance of efficient communication in multi-agent and LLM systems. Future studies should investigate OPTIMA’s scalability to larger models and more complex scenarios, opening the door to even more advanced multi-agent systems.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter.. Don’t Forget to join our 50k+ ML SubReddit

[Upcoming Event- Oct 17, 2024] RetrieveX – The GenAI Data Retrieval Conference (Promoted)

The post OPTIMA: Enhancing Efficiency and Effectiveness in LLM-Based Multi-Agent Systems appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

OPTIMA LLM-based MAS 通信效率 任务效果
相关文章