The Verge - Artificial Intelligences 02月07日
Researchers trained an OpenAI rival in half an hour for less than $50
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

研究者在26分钟内创建了低成本AI推理模型s1,它以Qwen2.5为基础,用小数据集和低成本进行训练,采用多种技术,且在某些方面超越OpenAI的模型,该模型的出现可能改变整个行业。

🎯s1模型以Qwen2.5为基础,用16个Nvidia H100 GPUs训练。

💡使用蒸馏法,从Google的AI推理模型中提炼答案来优化。

🔍采用测试时缩放技术,让模型在回答前思考更长时间。

Researchers managed to create a low-cost AI reasoning model rivaling OpenAI’s in just 26 minutes, as outlined in a paper published last week. The model, called s1, was refined using a small dataset of 1,000 questions and for under $50, according to TechCrunch.

To do this, researchers at Stanford and the University of Washington used a method known as distillation — which allows smaller models to draw from the answers produced by larger ones — to refine s1 using answers from Google’s AI reasoning model, Gemini 2.0 Flash Thinking Experimental. Google’s terms of service note that you can’t use Gemini’s API to “develop models that compete with” the company’s AI models. The Verge reached out to Google with a request for comment but didn’t immediately hear back.

The researchers based s1 on Qwen2.5, an open-source model from Alibaba Cloud. They initially started with a pool of 59,000 questions to train the model on, but found that the larger data set didn’t offer “substantial gains” over a whittled-down set of just 1,000. The researchers say they trained the model on just 16 Nvidia H100 GPUs.

The s1 model also uses a technique called test-time scaling, allowing the model to “think” for a longer amount of time before producing an answer. As noted in the paper, researchers forced the model to continue reasoning by adding “Wait” to the model’s response. “This can lead the model to doublecheck its answer, often fixing incorrect reasoning steps,” the paper says.

OpenAI’s o1 reasoning model uses a similar approach, something the buzzy AI startup DeepSeek sought to replicate with the launch of its R1 model that it claims was trained at a fraction of the cost. OpenAI has since accused DeepSeek of distilling information from its models to build a competitor, violating its terms of service. As for s1, the researchers claim that s1 “exceeds o1-preview on competition math questions by up to 27%.”

The rise of smaller and cheaper AI models threatens to upend the entire industry. They could prove that major companies like OpenAI, Microsoft, Meta, and Google don’t need to spend billions of dollars training AI, while building massive data centers filled with thousands of Nvidia GPUs.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

s1模型 AI推理 低成本 技术创新
相关文章