MarkTechPost@AI 01月09日
Microsoft AI Just Fully Open-Sourced Phi-4: A Small Language Model Available on Hugging Face Under the MIT License
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

微软在Hugging Face上以MIT许可证开源了其小型语言模型Phi-4。该模型拥有140亿参数,注重数据质量和效率,采用创新的多代理提示、指令反转和自我修订等合成数据生成方法。Phi-4基于Transformer架构,上下文长度达16k tokens,预训练使用了10万亿tokens的合成和高质量数据。其在STEM任务中表现出色,超越了其前代和更大的模型。开源Phi-4旨在促进AI社区的合作、透明度和广泛应用,为研究人员和开发者提供了宝贵的资源,推动了AI在研究、教育和行业中的发展。

💡Phi-4是一个140亿参数的语言模型,它采用高质量的合成数据进行训练,并通过多代理提示、指令反转和自我修订等创新方法生成,从而增强了其推理和问题解决能力。

🚀该模型基于decoder-only Transformer架构,具有16k tokens的扩展上下文长度,使其能够处理大型输入,并且在MMLU和HumanEval等基准测试中表现出色,尤其在STEM领域。

🛠️Phi-4不仅性能强大,还具有高度的定制化能力,支持使用多样化的合成数据集进行微调,以满足特定领域的需求,并且在Hugging Face上提供了详细的文档和API,方便用户集成到自己的项目中。

🔬微软开源Phi-4旨在促进AI领域的合作、透明度和更广泛的应用,同时为研究人员和开发者提供了一个强大的工具,推动AI在研究、教育和行业中的发展。

Microsoft has open-sourced Phi-4, a compact and efficient small language model, on Hugging Face under the MIT license. This decision highlights a shift towards transparency and collaboration in the AI community, offering developers and researchers new opportunities.

What Is Microsoft Phi-4?

Phi-4 is a 14-billion-parameter language model developed with a focus on data quality and efficiency. Unlike many models relying heavily on organic data sources, Phi-4 incorporates high-quality synthetic data generated through innovative methods such as multi-agent prompting, instruction reversal, and self-revision workflows. These techniques enhance its reasoning and problem-solving capabilities, making it suitable for tasks requiring nuanced understanding.

Phi-4 is built on a decoder-only Transformer architecture with an extended context length of 16k tokens, ensuring versatility for applications involving large inputs. Its pretraining involved approximately 10 trillion tokens, leveraging a mix of synthetic and highly curated organic data to achieve strong performance on benchmarks like MMLU and HumanEval.

Features and Benefits

    Compact and Accessible: Runs effectively on consumer-grade hardware.Reasoning-Enhanced: Outperforms its predecessor and larger models on STEM-focused tasks.Customizable: Supports fine-tuning with diverse synthetic datasets tailored for domain-specific needs.Easy Integration: Available on Hugging Face with detailed documentation and APIs.

Why Open Source?

Open-sourcing Phi-4 fosters collaboration, transparency, and wider adoption. Key motivations include:

Technical Innovations in Phi-4

Phi-4’s development was guided by three pillars:

    Synthetic Data: Generated using multi-agent and self-revision techniques, synthetic data forms the core of Phi-4’s training process, enhancing reasoning capabilities and reducing dependency on organic data.Post-Training Enhancements: Techniques such as rejection sampling and Direct Preference Optimization (DPO) improve output quality and alignment with human preferences.Decontaminated Training Data: Rigorous filtering processes ensured the exclusion of overlapping data with benchmarks, improving generalization.

Phi-4 also leverages Pivotal Token Search (PTS) to identify critical decision-making points in its responses, refining its ability to handle reasoning-heavy tasks efficiently.

Accessing Phi-4

Phi-4 is hosted on Hugging Face under the MIT license. Users can:

Impact on AI

By lowering barriers to advanced AI tools, Phi-4 promotes:

Community and Future

Phi-4’s release has been well-received, with developers sharing fine-tuned adaptations and innovative applications. Its ability to excel in STEM reasoning benchmarks demonstrates its potential to redefine what small language models can achieve. Microsoft’s collaboration with Hugging Face is expected to lead to more open-source initiatives, furthering innovation in AI.

Conclusion

The open-sourcing of Phi-4 reflects Microsoft’s commitment to democratizing AI. By making a powerful language model freely available, the company enables a global community to innovate and collaborate. As Phi-4 continues to find diverse applications, it exemplifies the transformative potential of open-source AI in advancing research, education, and industry.


Check out the Paper and Model on Hugging Face. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter and join our Telegram Channel and LinkedIn Group. Don’t Forget to join our 60k+ ML SubReddit.

FREE UPCOMING AI WEBINAR (JAN 15, 2025): Boost LLM Accuracy with Synthetic Data and Evaluation IntelligenceJoin this webinar to gain actionable insights into boosting LLM model performance and accuracy while safeguarding data privacy.

The post Microsoft AI Just Fully Open-Sourced Phi-4: A Small Language Model Available on Hugging Face Under the MIT License appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Phi-4 开源模型 小型语言模型 合成数据 Hugging Face
相关文章