MarkTechPost@AI 2024年09月18日
Collaborative Small Language Models for Finance: Meet The Mixture of Agents MoA Framework from Vanguard IMFS
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

为了解决大型语言模型 (LLM) 在金融领域应用中成本效益与性能之间的平衡问题,Vanguard IMFS 团队提出了一种名为混合代理 (MoA) 的框架。MoA 是一种先进的多代理系统,专为检索增强生成 (RAG) 任务而设计,它利用多个小型、专业化的模型协同工作,以更低的成本和更高的准确性回答复杂问题。

💥 **MoA 框架的优势:** MoA 框架通过多个小型、专业化的模型协同工作,克服了传统集成方法的局限性。它模仿研究团队的结构,每个代理都拥有专门的知识和专业知识,使其能够在各种金融领域中表现出色。

🚀 **MoA 的工作原理:** MoA 系统包含多个专业化的代理,每个代理都充当“初级研究员”,专注于特定领域,例如情感分析、财务指标或数学计算。这些代理可以访问不同的数据源,包括数据库、API 和外部文档,从而能够快速有效地处理高度特定信息。

💰 **MoA 的成本效益:** MoA 的成本效益使其非常适合大规模金融应用。Vanguard 的 IMFS 团队报告称,他们的 MoA 系统每月运营成本不到 8000 美元,同时处理来自研究人员团队的查询。与每月成本在 5000 美元到 8000 美元之间但性能明显较低的单模型系统相比,这是一个可比的数字。

📊 **MoA 的性能提升:** 在性能方面,MoA 系统与传统单模型系统相比,在响应质量和效率方面取得了显著改进。在测试中,MoA 系统能够使用两层代理在不到 60 秒的时间内分析数万份财务文件。与单模型系统相比,这些代理在串行推理中的延迟惩罚仅为 4.07 倍,在并行运行时为 2.24 倍。

Language model research has rapidly advanced, focusing on improving how models understand and process language, particularly in specialized fields like finance. Large Language Models (LLMs) have moved beyond basic classification tasks to become powerful tools capable of retrieving and generating complex knowledge. These models work by accessing large data sets and using advanced algorithms to provide insights and predictions. In finance, where the volume of data is immense and requires precise interpretation, LLMs are crucial for analyzing market trends, predicting outcomes, and providing decision-making support.

One major problem researchers face in the LLM field is balancing cost-effectiveness with performance. LLMs are computationally expensive, and as they process larger data sets, the risk of producing inaccurate or misleading information increases, especially in fields like finance, where incorrect predictions can lead to significant losses. Traditional approaches rely heavily on a single dense transformer model, which, while powerful, often needs help with hallucinations, where the model generates incorrect or irrelevant information. Large financial applications requiring fast, accurate, and cost-efficient models amplify this problem.

Researchers have explored several methods to address these challenges, including ensemble models, which involve multiple LLMs working together to improve output accuracy. Ensemble models have successfully reduced errors and improved generalization, especially when dealing with new information not included in the training data. However, the downside of these systems is their cost and slow processing speed, as running multiple models in parallel or sequence requires significant computational power. The financial sector, which deals with massive amounts of data, often finds these solutions impractical due to the high operational costs and time constraints.

Researchers from the Vanguard IMFS (Investment Management FinTech Strategies)  team introduced a new framework called Mixture of Agents (MoA) to overcome the limitations of traditional ensemble methods. MoA is an advanced multi-agent system designed specifically for Retrieval-Augmented Generation (RAG) tasks. Unlike previous models, MoA utilizes a collection of small, specialized models that work together in a highly coordinated manner to answer complex questions with greater accuracy and lower costs. This collaborative network of agents mirrors the structure of a research team, with each agent having expertise and knowledge base, enabling the system to perform better across various financial domains.

The MoA system comprises multiple specialized agents, each acting as a “junior researcher” with a specific focus, such as sentiment analysis, financial metrics, or mathematical computations. For example, the system includes agents like the “10-K/Q Math Agent,” a fine-tuned GPT-4 model designed for handling accounting and financial figures, and the “10-K/Q Sentiment Agent,” a Llama-2 model trained to analyze sentiment in equity markets. Each agent has access to different data sources, including databases, APIs, and external documents, allowing them to process highly specific information quickly and efficiently. This specialization enables the MoA framework to outperform traditional single-model systems in speed and accuracy while keeping operational costs low.

In terms of performance, the MoA system has shown significant improvements in response quality and efficiency compared to traditional single-model systems. During tests, the MoA system could analyze tens of thousands of financial documents in under 60 seconds using two layers of agents. Compared to a single-model system, these agents operate with a latency penalty of only 4.07x in serial inference or 2.24x when running in parallel. A basic MoA system with two Mistral-7B agents was tested in one experiment alongside single-model systems such as GPT-4 and Claude 3 Opus. The MoA system consistently provided more accurate and comprehensive answers. For example, when asked about revenue growth in Apple’s Q1 2023 earnings report, the MoA agents captured 5 out of 7 key points, compared to 4 from Claude and only two from GPT-4. This demonstrates the system’s ability to surface critical information with higher precision and speed.

The cost-effectiveness of MoA makes it highly suitable for large-scale financial applications. Vanguard’s IMFS team reported that their MoA system operates at a total monthly cost of under $8,000 while processing queries from a team of researchers. This is comparable to single-model systems, which cost between $5,000 and $8,000 per month but provide significantly lower performance. The MoA framework’s modular design allows companies to scale their operations based on budget and need, with the flexibility to add or remove agents as necessary. As the system scales, it becomes increasingly efficient, saving time and computational resources.

In conclusion, the Mixture of Agents framework offers a powerful solution for improving the performance of large language models in finance. The researchers successfully addressed critical issues like scalability, cost, and response accuracy by leveraging a collaborative agent-based system. The MoA framework enhances the speed and quality of information retrieval and offers significant cost savings compared to traditional methods. With its ability to process vast amounts of data in a fraction of the time while maintaining high accuracy, MoA is set to become a standard for enterprise-grade applications in finance and beyond. This system represents a significant advancement in LLM technology, providing a scalable, cost-effective, and highly efficient method for handling complex financial data.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 50k+ ML SubReddit

FREE AI WEBINAR: ‘SAM 2 for Video: How to Fine-tune On Your Data’ (Wed, Sep 25, 4:00 AM – 4:45 AM EST)

The post Collaborative Small Language Models for Finance: Meet The Mixture of Agents MoA Framework from Vanguard IMFS appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

金融科技 大型语言模型 混合代理 MoA Vanguard IMFS
相关文章