TechCrunch News 05月02日 02:36
Ai2’s new small AI model outperforms similarly-sized models from Google, Meta
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Ai2发布了Olmo 2 1B,一个10亿参数的小型AI模型,在多个基准测试中击败了谷歌、Meta和阿里巴巴的同等规模模型。该模型采用Apache 2.0许可,可在Hugging Face上获取,并提供代码和数据集,允许完全复现。小模型对硬件要求低,更易于开发者和爱好者使用。近期涌现了多款小模型,如微软的Phi 4和Qwen的2.5 Omni 3B,Olmo 2 1B同样可以在普通笔记本电脑甚至移动设备上运行。该模型使用4万亿tokens的数据集进行训练,但在商业部署方面,Ai2建议谨慎,因为它可能产生有害或不准确的内容。

💡Ai2发布Olmo 2 1B,一个10亿参数的AI模型,在性能上超越了Google的Gemma 3 1B、Meta的Llama 3.2 1B和阿里巴巴的Qwen 2.5 1.5B。

💻Olmo 2 1B与大型模型相比,对硬件要求较低,可以在普通笔记本电脑甚至移动设备上运行,这使得它对开发者和爱好者更具吸引力。

📚Olmo 2 1B是在一个包含4万亿tokens的数据集上进行训练的,该数据集来自公开的、AI生成的和手动创建的来源。

⚠️Ai2警告称,Olmo 2 1B可能产生有害或不准确的内容,因此不建议在商业环境中部署。

‘Tis the week for small AI models, it seems.

On Thursday, Ai2, the nonprofit AI research institute, released Olmo 2 1B, a 1-billion-parameter model that Ai2 claims beats similarly-sized models from Google, Meta, and Alibaba on several benchmarks. Parameters, sometimes referred to as weights, are the internal components of a model that guide its behavior.

Olmo 2 1B is available under a permissive Apache 2.0 license on the AI dev platform Hugging Face. Unlike most models, Olmo 2 1B can be replicated from scratch; Ai2 has provided the code and data sets (Olmo-mix-1124, Dolmino-mix-1124) used to develop it.

Small models might not be as capable as their behemoth counterparts, but importantly, they don’t require beefy hardware to run. That makes them much more accessible for developers and hobbyists contending with the limitations of lower-end and consumer machines.

There’s been a raft of small model launches over the past few days, from Microsoft’s Phi 4 reasoning family to Qwen’s 2.5 Omni 3B. Most of these — and Olmo 2 1B — can easily run on a modern laptop or even a mobile device.

Ai2 says that Olmo 2 1B was trained on a data set of 4 trillion tokens from publicly available, AI-generated, and manually created sources. Tokens are the raw bits of data models ingest and generate — 1 million tokens is equivalent to about 750,000 words.

On a benchmark measuring arithmetic reasoning, GSM8K, Olmo 2 1B scores better than Google’s Gemma 3 1B, Meta’s Llama 3.2 1B, and Alibaba’s Qwen 2.5 1.5B. Olmo 2 1B also eclipses the performance of those three models on TruthfulQA, a test for evaluating factual accuracy.

Techcrunch event

Berkeley, CA | June 5

BOOK NOW

Ai2 warns that that Olmo 2 1B carries risks, however. Like all AI models, it can produce “problematic outputs” including harmful and “sensitive” content, the organization says, as well as factually inaccurate statements. For these reasons, Ai2 recommends against deploying Olmo 2 1B in commercial settings.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Olmo 2 1B AI模型 Hugging Face 小模型
相关文章