Fortune | FORTUNE 前天 01:11
OpenAI launches its first open model in years so it can stop being on the ‘wrong side of history’—while still keeping its most valuable IP under wraps
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

OpenAI近期发布了其期待已久的开源模型,这是自2020年GPT-2以来首次公开模型权重。CEO Sam Altman表示,此举旨在让更多人接触到AI技术,并促进新的研究和产品开发。然而,尽管模型权重公开,其核心的专有架构、训练数据和方法仍未披露。此次发布的gpt-oss-120b和gpt-oss-20b模型主要面向AI开发者,使其能在不同硬件上运行。此举也被视为对中国开源模型崛起的回应,尽管OpenAI未直接将其新模型与中国领先的开源模型进行基准测试。该模型采用Mixture-of-Experts(MoE)架构,旨在在不泄露核心知识产权的前提下,为社区提供有价值的工具。

🚀 OpenAI推出开源模型,旨在普及AI技术并推动创新,CEO Sam Altman强调此举是为让AI造福全人类,并希望促进基于开放AI生态的建设。

💡 尽管此次发布了模型权重,但OpenAI并未公开其最先进模型的专有架构、训练数据和方法,这使得该模型并非完全意义上的“开放”,而是一种平衡商业秘密与技术共享的策略。

💻 新发布的gpt-oss-120b和gpt-oss-20b模型主要面向AI开发者和构建者,其设计允许在单块Nvidia 80GB芯片或消费级Mac笔记本等设备上运行,便于实际应用开发。

📈 此举被视为OpenAI对中国开源模型(如DeepSeek)崛起的回应,旨在加强在开源AI领域的竞争力,尽管OpenAI选择不直接与中国模型进行基准测试,而是鼓励社区自行评估。

⚙️ 模型采用了Mixture-of-Experts(MoE)架构,仅激活特定查询所需的子网络,避免了激活整个模型。通过使用公开已知的架构组件进行训练,OpenAI成功地在提供实用模型的同时,规避了核心知识产权的泄露风险。

Despite what its name suggests, OpenAI hadn’t released an “open” model—one that includes access to the weights, or the numerical parameters often described as the model’s brains—since GPT-2 in 2020. That changed on Tuesday: The company launched a long-awaited open-weight model, in two sizes, that the company says pushes the frontier of reasoning in open-source AI.

“We’re excited to make this model, the result of billions of dollars of research, available to the world to get AI into the hands of the most people possible,” said OpenAI CEO Sam Altman about the release. “As part of this, we are quite hopeful that this release will enable new kinds of research and the creation of new kinds of products.” He emphasized that he is “excited for the world to be building on an open AI stack created in the United States, based on democratic values, available for free to all and for wide benefit.”

OpenAI CEO Sam Altman had teased the upcoming models back in March, two months after admitting, in the wake of the success of China’s open models from DeepSeek, that the company had been “on the wrong side of history” when it came to opening up its models to developers and builders. But while the weights are now public, experts note that OpenAI’s new models are hardly “open.” By no means is it giving away its crown jewels: The proprietary architecture, routing mechanisms, training data and methods that power its most advanced models—including the long-awaited GPT-5, widely expected to be released sometime this month—remain tightly under wraps.

OpenAI is targeting AI builders and developers

The two new model names – gpt-oss-120b and gpt-oss-20b – may be indecipherable to non-engineers, but that’s because OpenAI is setting its sights on AI builders and developers seeking to rapidly build on real-world use cases on their own systems. The company noted that the larger of the two models can run on a single Nvidia 80GB chip while the smaller one fits on consumer hardware like a Mac laptop. 

Greg Brockman, co-founder and president of OpenAI, acknowledged on a press pre-briefing call that “it’s been a long time” since the company had released an open model. He added that it is “something that we view as complementary to the other products that we release” and along with OpenAI’s proprietary models, “combine to really accelerate our mission of ensuring that API benefits all of humanity.”

OpenAI said the new models perform well on reasoning benchmarks, which have emerged as the key measurements for AI performance, with models from OpenAI, Anthropic, Google and DeepSeek fiercely competing over their abilities to tackle multi-step logic, code generation, and complex problem-solving. Ever since the open source DeepSeek R1 shook the industry in January with its reasoning capabilities at a much lower cost, many other Chinese models have followed suit– including Alibaba’s Qwen and Moonshot AI’s Kimi models. While OpenAI said at a press pre-briefing that the new open-weight models are a proactive effort to provide what users want, it is also clearly a strategic response to ramping up open source competition.  

Notably, OpenAI declined to benchmark its new open-weight models against Chinese open-source systems like DeepSeek or Qwen—despite the fact that those models have recently outperformed U.S. rivals on key reasoning benchmarks. In the press briefing, the company said it confident in its benchmarks against its own models and that it would leave it to others in the AI community to test further and “make up their own minds.”

Avoiding leaking intellectual property

OpenAI’s new open-weight models are built using a Mixture-of-Experts (MoE) architecture, in which the system activates only the “experts,” or sub-networks, it needs for a specific input, rather than using the entire model for every query. Dylan Patel, founder of research firm SemiAnalysis, pointed out in a post on X before the release that OpenAI trained the models only using publicly known components of the architecture—meaning the building blocks it used are already familiar to the open-source community. He emphasized that this was a deliberate choice—that by avoiding any proprietary training techniques or architecture innovations, OpenAI could release a genuinely useful model without actually leaking any intellectual property that powers its proprietary frontier models like GPT 4o.

For example, in a model card accompanying the release, OpenAI confirmed that the models use a Mixture of Experts (MoE) architecture with 12 active experts out of 64, but it does not describe the routing mechanism, which is a crucial and proprietary part of the architecture.

“You want to minimize risk to your business, but you [also] want to be maximally useful to the public,” Aleksa Gordic, a former Google DeepMind researcher, told Fortune, adding that companies like Meta and Mistral, which have also focused on open-weight models, also have not included proprietary information.

“They minimize the IP leak and remove any risk to their core business, while at the same time sharing a useful artifact that will enable the startup ecosystem and developers,” he said. “It’s by definition the best they can do given those two opposing objectives.”

Introducing the 2025 Fortune 500

, the definitive ranking of the biggest companies in America. 

Explore this year's list.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

OpenAI 开源模型 AI技术 GPT Mixture-of-Experts
相关文章