MarkTechPost@AI 2024年09月19日
TinyAgent: An End-to-End AI Framework for Training and Deploying Task-Specific Small Language Model Agents
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

TinyAgent是一种用于训练和部署任务特定的小型语言模型代理的创新框架,可在本地设备运行,具有多种优势。

🎯TinyAgent框架旨在填补LLM在本地设备应用的空白。它能通过解释自然语言命令,独立执行复杂任务,如信息检索和设备控制,且不依赖云基础设施。

💻TinyAgent使用LLM Compiler框架对开源模型进行修改,以正确执行函数调用。通过精心策划的高质量数据集进行微调,生成了TinyAgent-1.1B和TinyAgent-7B两种变体,它们在处理特定任务时高度精确。

🔧TinyAgent框架的独特工具检索技术有助于缩短推理时的输入提示,使其能更快速有效地选择正确的工具或功能。同时,它还使用量化方法来缩小模型规模和复杂性,以在本地设备上正常运行。

📱TinyAgent已被部署为MacBook上的本地Siri式系统,可理解用户通过文本或语音输入的命令并执行操作,如启动应用、创建提醒和进行信息搜索。该系统在保护隐私的同时,无需互联网连接。

The significant advancements in Large Language Models (LLMs) have led to the development of agentic systems, which integrate several tools and APIs to fulfill user inquiries through function calls. By interpreting natural language commands, these systems can perform sophisticated tasks independently, such as information retrieval and device control. However, much research hasn’t been done on using these LLMs locally, on laptops or smartphones, or at the edge. The primary limitation is the large size and high processing demands of these models, which usually require cloud-based infrastructure to function properly.

In recent research from UC Berkeley and ICSI, the TinyAgent framework has been introduced as an innovative technique to train and deploy task-specific little language model agents in order to fill this gap. Because of their ability to manage function calls, these agents can operate independently on local devices and are not dependent on cloud-based infrastructure. By concentrating on smaller, more effective models that preserve the key functionalities of bigger LLMs and the ability to carry out user requests by coordinating other tools and APIs, TinyAgent provides a comprehensive solution for advancing sophisticated AI capabilities.

The TinyAgent framework starts with open-source models that need to be modified in order to correctly execute function calls. The LLMCompiler framework has been used to accomplish this, fine-tuning the models to guarantee that they can execute commands consistently. The methodical curation of a high-quality dataset designed especially for function-calling jobs is a crucial step in this approach. Using this specific dataset to refine the models, TinyAgent generates two variants: TinyAgent-1.1B and TinyAgent-7B. Despite being much smaller than their larger equivalents, like GPT-4-Turbo, these models are highly precise at handling particular jobs.

A unique tool retrieval technique is one of the main contributions of the TinyAgent framework, as it helps shorten the input prompt during inference. By doing this, the model is able to choose the right tool or function more quickly and effectively, all without being bogged down by extensive or unnecessary input data. To further improve its inference performance, TinyAgent also uses quantization, a method that shrinks the size and complexity of the model. In order to guarantee that the compact models can function properly on local devices, even with constrained computational resources, these optimizations are essential.

The TinyAgent framework has been deployed as a local Siri-like system for the MacBook in order to showcase the system’s real-world applications. Without requiring cloud access, this system can comprehend orders from users sent by text or voice input and carry out actions like starting apps, creating reminders, and doing information searches. By storing user data locally, this localized deployment not only protects privacy but also does away with the necessity for an internet connection, which is a critical feature in situations where reliable access might not be available.

The TinyAgent framework has demonstrated some amazing results. The function-calling capabilities of much bigger models, such as GPT-4-Turbo, have been demonstrated to be met, and in some cases exceeded, by the TinyAgent models despite their reduced size. This is a good accomplishment because it shows that smaller models may accomplish highly specialized tasks effectively and efficiently when they are trained and optimized using the appropriate methods.

In conclusion, TinyAgent offers a great method for enabling edge devices to harness the potential of LLM-driven agentic systems. While retaining strong performance in real-time applications, TinyAgent provides an effective, privacy-focused substitute for cloud-based AI systems by optimizing smaller models for function calling and employing strategies like tool retrieval and quantization.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. If you like our work, you will love our newsletter..

Don’t Forget to join our 50k+ ML SubReddit

FREE AI WEBINAR: ‘SAM 2 for Video: How to Fine-tune On Your Data’ (Wed, Sep 25, 4:00 AM – 4:45 AM EST)

The post TinyAgent: An End-to-End AI Framework for Training and Deploying Task-Specific Small Language Model Agents appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

TinyAgent 语言模型 本地设备 函数调用
相关文章