AI News 19小时前
Alibaba’s new Qwen reasoning AI model sets open-source records
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

阿里巴巴的Qwen团队发布了其开源推理AI模型Qwen3-235B-A22B-Thinking-2507的新版本,在逻辑推理、复杂数学、科学问题和高级编码方面表现出色,设定了新的行业标准。该模型拥有2350亿参数,但采用MoE架构,每次仅激活约220亿参数,并具备262,144个tokens的超大内存。开发者可通过Hugging Face轻松部署,并可利用Qwen-Agent框架优化其工具调用能力。为获得最佳效果,建议根据任务复杂性调整输出长度,并在提示词中引导模型进行逐步思考。

🌟 **卓越的推理与编码能力**: Qwen3-235B-A22B-Thinking-2507在逻辑推理、复杂数学、科学问题和高级编码等领域展现出非凡实力,在AIME25基准测试中达到92.3分,在LiveCodeBench v6编码测试中获得74.1分,树立了开源模型的标杆。

🧠 **高效的MoE架构与海量参数**: 模型拥有2350亿总参数,但通过混合专家(MoE)架构,每次仅激活约220亿参数,实现了高效运作。这如同一个庞大的专家团队,仅根据任务需求调用最合适的成员,大幅提升了计算效率。

📚 **超长上下文窗口**: 该模型原生支持262,144个tokens的上下文长度,这意味着它能够处理和理解海量信息,对于需要深度分析和关联大量数据的任务具有显著优势,远超许多现有模型。

🛠️ **易于部署与使用**: Qwen3-235B-A22B-Thinking-2507已在Hugging Face上可用,开发者可使用sglang或vllm等工具便捷部署。其Qwen-Agent框架特别适合发挥模型的工具调用能力,为开发创新应用提供了便利。

💡 **优化使用建议**: 为最大化模型性能,建议根据任务复杂性调整输出长度,如复杂挑战建议提升至81,920 tokens。同时,在提示词中加入“逐步思考”等指令,有助于模型提供更精准、结构化的回答。

The Qwen team from Alibaba have just released a new version of their open-source reasoning AI model with some impressive benchmarks.

Meet Qwen3-235B-A22B-Thinking-2507. Over the past three months, the Qwen team has been hard at work scaling up what they call the “thinking capability” of their AI, aiming to improve both the quality and depth of its reasoning.

The result of their efforts is a model that excels at the really tough stuff: logical reasoning, complex maths, science problems, and advanced coding. In these areas that typically require a human expert, this new Qwen model is now setting the standard for open-source models.

On reasoning benchmarks, Qwen’s latest open-source AI model achieves 92.3 on AIME25 and 74.1 on LiveCodeBench v6 for coding. It also holds its own in more general capability tests, scoring 79.7 on Arena-Hard v2, which measures how well it aligns with human preferences.

At its heart, this is a massive reasoning AI model from the Qwen team with 235 billion parameters in total. However, it uses Mixture-of-Experts (MoE), which means it only activates a fraction of those parameters – about 22 billion – at any one time. Think of it like having a huge team of 128 specialists on call, but only the eight best-suited for a specific task are brought in to actually work on it.

Perhaps one of its most impressive features is its massive memory. Qwen’s open-source reasoning AI model has a native context length of 262,144 tokens; a huge advantage for tasks that involve understanding vast amounts of information.

For the developers and tinkerers out there, the Qwen team has made it easy to get started. The model is available on Hugging Face. You can deploy it using tools like sglang or vllm to create your own API endpoint. The team also points to their Qwen-Agent framework as the best way to make use of the model’s tool-calling skills.

To get the best performance from their open-source AI reasoning model, the Qwen team have shared a few tips. They suggest an output length of around 32,768 tokens for most tasks, but for really complex challenges, you should boost that to 81,920 tokens to give the AI enough room to “think”. They also recommend giving the model specific instructions in your prompt, like asking it to “reason step-by-step” for maths problems, to get the most accurate and well-structured answers.

The release of this new Qwen model provides a powerful yet open-source reasoning AI that can rival some of the best proprietary models out there, especially when it comes to complex, brain-bending tasks. It will be exciting to see what developers ultimately build with it.

(Image by Tung Lam)

See also: AI Action Plan: US leadership must be ‘unchallenged’

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Alibaba’s new Qwen reasoning AI model sets open-source records appeared first on AI News.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Qwen3 AI模型 开源 推理能力 自然语言处理
相关文章