MarkTechPost@AI 2024年07月18日
Microsoft Researchers Propose Auto Evol-Instruct: An End-to-End AI Framework that Evolves Instruction Datasets Using Large Language Models without Any Human Effort
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

微软研究人员提出了一种名为 Auto Evol-Instruct 的端到端 AI 框架,利用大型语言模型自动进化指令数据集,无需人工干预。该框架通过优化进化方法,提高了指令数据集的复杂性和多样性,从而提升了大型语言模型在各种任务中的表现。

🤔 Auto Evol-Instruct 是一种自动化的框架,通过大型语言模型 (LLM) 自动设计进化方法,无需人工干预,可以有效地将指令数据集应用于各种任务。

🚀 Auto Evol-Instruct 采用了一个初始进化方法,该方法分析输入指令并自动推导出适合给定数据的进化规则,然后通过优化器 LLM 迭代优化这些规则,以确保进化方法的稳定性和有效性。

📊 Auto Evol-Instruct 在多个基准测试中表现出色,例如 MT-Bench、AlpacaEval、GSM8K 和 HumanEval,其性能超过了 GPT-3.5-Turbo 和 WizardLM-70B 等模型,与 Claude2.0 相当。

📈 Auto Evol-Instruct 通过进化轨迹分析和进化方法优化阶段,迭代优化进化方法,确保进化数据集的复杂性和多样性。

💡 Auto Evol-Instruct 解决了手动方法的局限性,通过自动进化指令数据集,为提高大型语言模型的性能和适应性提供了可扩展且高效的解决方案。

Large language models (LLMs) are central to advancements in artificial intelligence, focusing on enhancing the models’ ability to follow detailed instructions. This area of research encompasses methods to improve the quality and complexity of datasets used for training LLMs, ultimately leading to more sophisticated and versatile AI systems. The importance of these improvements cannot be overstated, as they directly impact the models’ performance across various tasks, from natural language understanding to code generation and mathematical reasoning.

A major challenge in this field is the dependency on high-quality instruction datasets, which are difficult to annotate at scale. Manually designed methods require substantial human expertise and resources, making achieving consistent improvements across different tasks challenging. This limitation hinders the performance and adaptability of LLMs, creating a bottleneck in their development. Researchers have been actively exploring ways to overcome this bottleneck, seeking methods to enhance dataset complexity and diversity without requiring extensive human intervention.

Current methods, such as Evol-Instruct, iteratively refine high-quality data using LLMs to improve dataset complexity and diversity. While these methods are effective, they heavily rely on heuristic efforts and expert-designed evolving rules. This reliance can be expensive and time-consuming, particularly when adapting to new tasks. Evol-Instruct, for instance, has shown superior performance across various benchmarks, including MT-Bench, AlpacaEval, GSM8K, and HumanEval. However, each time it is applied to a new task, the methods for execution evolution need to be redesigned, requiring a high level of expertise and considerable costs.

Researchers from Microsoft introduced Auto Evol-Instruct, an automated framework that eliminates the need for human intervention in the instruction evolution process. This innovative approach leverages LLMs to design evolving methods autonomously, enabling cost-effective adaptation to various tasks by altering the input data. The framework begins with a universal initial evolving method that autonomously analyzes the input instructions and formulates evolution rules. These rules are then iteratively optimized by an optimizer LLM, which identifies and addresses issues in the evolving methods, ensuring minimal evolution failure and enhancing the dataset’s complexity and diversity.

Auto Evol-Instruct operates through a detailed process involving multiple stages. Firstly, it employs an initial evolving method that analyzes the input instruction and brainstorms evolution rules suitable for the given data. This method differs from Evol-Instruct, which requires human experts to specify the rules of evolution. Instead, Auto Evol-Instruct uses an evol LLM to devise a comprehensive plan based on the listed methods autonomously and implements this plan to generate the evolved instruction. The evol LLM then thoroughly reviews the evolved instruction, rectifying any unreasonable parts to ensure the final evolved instruction is complex and stable.

The performance of Auto Evol-Instruct was rigorously evaluated across several benchmarks. Using only 10K evolved ShareGPT data for fine-tuning Mixtral-8x7B, the framework achieved an impressive 8.09 on MT-Bench and 91.4 on AlpacaEval, surpassing GPT-3.5-Turbo and WizardLM-70B, and comparable with Claude2.0. Additionally, with just 7K evolved GSM8K training data, Auto Evol-Instruct achieved 82.49 on GSM8K, outperforming GPT-3.5-Turbo, WizardMath-70B, and MetaMath-70B. In code generation, using 20K evolved Code Alpaca to fine-tune DeepSeek-Coder-Base-33B, the framework achieved 77.4 on HumanEval, surpassing GPT-3.5-Turbo and WizardCoder-34B.

A key aspect of Auto Evol-Instruct is its ability to iteratively optimize the evolving method through Evol Trajectory Analysis and Evolving Method Optimization stages. The optimizer LLM analyzes the potential issues and failures exposed in instruction evolution performed by the evol LLM, generating feedback for subsequent optimization. This feedback is then used to refine the evolving method, ensuring the lowest failure rate for a given instruction dataset. This meticulous optimization and analysis ensures that the evolved datasets are complex and diverse, improving instruction tuning.

In conclusion, Auto Evol-Instruct addresses the limitations of manual methods by automating the evolution of instruction datasets. It provides a scalable, efficient solution that enhances the performance and adaptability of LLMs across various tasks. The research demonstrates that methods optimized by Auto Evol-Instruct significantly surpass those crafted by humans, showcasing its potential to advance the field of AI. The framework’s impressive results across multiple benchmarks highlight its effectiveness in improving instruction following, mathematical reasoning, and code generation capabilities.


Check out the Paper. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter

Join our Telegram Channel and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 46k+ ML SubReddit

The post Microsoft Researchers Propose Auto Evol-Instruct: An End-to-End AI Framework that Evolves Instruction Datasets Using Large Language Models without Any Human Effort appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Auto Evol-Instruct 大型语言模型 指令数据集 人工智能 进化
相关文章