Microsoft Azure Blog Announcements 9小时前
OpenAI’s open‑source model: gpt‑oss on Azure AI Foundry and Windows AI Foundry 
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

微软发布了其AI平台的新时代,将AI从堆栈中的一层转变为核心。通过Azure AI Foundry和Foundry Local,开发者可以统一构建、微调和部署AI应用和智能代理。此次重磅更新包括OpenAI的GPT-OSS开源模型,允许用户在云端或边缘设备上自由运行、适配和部署。GPT-OSS-120b和GPT-OSS-20b模型提供了强大的推理能力和高效的代理任务执行,并支持在企业级GPU或本地Windows设备上运行,为开发者提供了前所未有的灵活性和控制力,加速AI创新。

🚀 **AI成为核心,平台化赋能开发者:** 微软正构建一个从云到边缘的全栈AI应用和代理工厂,使开发者能够不仅使用AI,更能创造AI。Azure AI Foundry提供统一平台,Foundry Local将开源模型带到边缘,实现跨设备推理,Windows AI Foundry进一步将Foundry Local集成至Windows 11,支持本地AI开发生命周期。

💡 **GPT-OSS开源模型带来前所未有的灵活性:** OpenAI首次发布GPT-2后的开源权重模型GPT-OSS,允许开发者以自己的方式运行、适配和部署,例如在单企业GPU上运行GPT-OSS-120b,或在本地运行GPT-OSS-20b。这些模型不仅快速且强大,还易于微调、提炼和优化,以适应特定领域或边缘部署需求。

⚙️ **开放模型加速AI开发与部署:** 开放模型已成为主流,能够驱动自主代理和领域特定副驾驶。通过Azure AI Foundry,开发者可利用参数高效微调方法(LoRA, QLoRA, PEFT),集成专有数据,并在数小时内发布新检查点。模型还可被提炼、量化、修剪上下文长度或应用稀疏性,以满足严格的内存要求,实现云端或边缘的灵活部署。

🌐 **GPT-OSS模型的多样化应用场景:** GPT-OSS-120b是强大的推理引擎,适用于复杂任务如数学、代码和领域问答,能在单数据中心级GPU上运行。GPT-OSS-20b则擅长代理任务,如代码执行和工具使用,可在Windows设备上高效运行,非常适合构建自主助手或嵌入现实工作流。

🤝 **开放与负责任的AI愿景:** 微软致力于提供一个开放的平台,将AI技术带给客户。通过提供GPT-OSS等开源模型,并结合Foundry内置的安全和治理工具,微软支持开发者在开放和专有模型之间自由选择,确保安全、合规和信任,加速AI创新。

AI is no longer a layer in the stack—it’s becoming the stack. This new era calls for tools that are open, adaptable, and ready to run wherever your ideas live—from cloud to edge, from first experiment to scaled deployment. At Microsoft, we’re building a full-stack AI app and agent factory that empowers every developer not just to use AI, but to create with it.

That’s the vision behind our AI platform spanning cloud to edge. Azure AI Foundry provides a unified platform for building, fine-tuning, and deploying intelligent agents with confidence while Foundry Local brings open-source models to the edge—enabling flexible, on-device inferencing across billions of devices. Windows AI Foundry builds on this foundation, integrating Foundry Local into Windows 11 to support a secure, low-latency local AI development lifecycle deeply aligned with the Windows platform. 

With the launch of OpenAI’s gpt‑oss models—its first open-weight release since GPT‑2—we’re giving developers and enterprises unprecedented ability to run, adapt, and deploy OpenAI models entirely on their own terms. 

For the first time, you can run OpenAI models like gpt‑oss‑120b on a single enterprise GPU—or run gpt‑oss‑20b locally. It’s notable that these aren’t stripped-down replicas—they’re fast, capable, and designed with real-world deployment in mind: reasoning at scale in the cloud, or agentic tasks at the edge. 

And because they’re open-weight, these models are also easy to fine-tune, distill, and optimize. Whether you’re adapting for a domain-specific copilot, compressing for offline inference, or prototyping locally before scaling in production, Azure AI Foundry and Foundry Local give you the tooling to do it all—securely, efficiently, and without compromise. 

Open models, real momentum 

Open models have moved from the margins to the mainstream. Today, they’re powering everything from autonomous agents to domain-specific copilots—and redefining how AI gets built and deployed. And with Azure AI Foundry, we’re giving you the infrastructure to move with that momentum: 

In short, open models aren’t just feature-parity replacements—they’re programmable substrates. And Azure AI Foundry provides training pipelines, weight management, and low-latency serving backplane so you can exploit every one of those levers and push the envelope of AI customization. 

Meet gpt‑oss: Two models, infinite possibilities

Today, gpt‑oss-120b and gpt‑oss-20b are available on Azure AI Foundry. gpt‑oss-20b is also available on Windows AI Foundry and will be coming soon on MacOS via Foundry Local. Whether you’re optimizing for sovereignty, performance, or portability, these models unlock a new level of control. 

Both models will soon be API-compatible with the now ubiquitous responses API. That means you can swap them into existing apps with minimal changes—and maximum flexibility. 

Bringing gpt‑oss to Cloud and Edge 

Azure AI Foundry is more than a model catalog—it’s a platform for AI builders. With more than 11,000 models and growing, it gives developers a unified space to evaluate, fine-tune, and productionize models with enterprise-grade reliability and security. 

Today, with gpt‑oss in the catalog, you can: 

For organizations developing scenarios only possible on client devices, Foundry Local brings prominent open-source models to Windows AI Foundry, pre-optimized for inference on your own hardware, supporting CPUs, GPUs, and NPUs, through a simple CLI, API, and SDK.

Whether you’re working in an offline setting, building in a secure network, or running at the edge—Foundry Local and Windows AI Foundry lets you go fully cloud-optional. With the capability to deploy gpt‑oss-20b on modern high-performance Windows PCs, your data stays where you want it—and the power of frontier-class models comes to you. 

This is hybrid AI in action: the ability to mix and match models, optimize performance and cost, and meet your data where it lives. 

Empowering builders and decision makers 

The availability of gpt‑oss on Azure and Windows unlocks powerful new possibilities for both builders and business leaders. 

For developers, open weights mean full transparency. Inspect the model, customize, fine-tune, and deploy on your own terms. With gpt‑oss, you can build with confidence, understanding exactly how your model works and how to improve it for your use case. 

For decision makers, it’s about control and flexibility. With gpt‑oss, you get competitive performance—with no black boxes, fewer trade-offs, and more options across deployment, compliance, and cost. 

A vision for the future: Open and responsible AI, together 

The release of gpt‑oss and its integration into Azure and Windows is part of a bigger story. We envision a future where AI is ubiquitous—and we are committed to being an open platform to bring these innovative technologies to our customers, across all our data centers and devices. 

By offering gpt‑oss through a variety of entry points, we’re doubling down on our commitment to democratize AI. We recognize that our customers will benefit from a diverse portfolio of models—proprietary and open—and we’re here to support whichever path unlocks value for you. Whether you are working with open-source models or proprietary ones, Foundry’s built-in safety and security tools ensure consistent governance, compliance, and trust—so customers can innovate confidently across all model types. 

Finally, our support of gpt-oss is just the latest in our commitment to open tools and standards. In June we announced that GitHub Copilot Chat extension is now open source on GitHub under the MIT license—the first step to make VS Code an open source AI editor. We seek to accelerate innovation with the open-source community and drive greater value to our market leading developer tools. This is what it looks like when research, product, and platform come together. The very breakthroughs we’ve enabled with our cloud at OpenAI are now open tools that anyone can build on—and Azure is the bridge that brings them to life. 

Next steps and resources for navigating gpt‑oss

*See Managed Compute pricing page here.


1Pricing is accurate as of August 2025.

The post OpenAI’s open‑source model: gpt‑oss on Azure AI Foundry and Windows AI Foundry  appeared first on Microsoft Azure Blog.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

GPT-OSS Azure AI Foundry Windows AI Foundry 开源AI 边缘AI
相关文章