AI News 2024年11月28日
Ai2 OLMo 2: Raising the bar for open language models
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Ai2发布了新一代开源语言模型OLMo 2,该模型拥有7B和13B参数版本,训练数据量高达5万亿tokens。OLMo 2在性能上与同类开源模型相当甚至超越,并在英语学术基准测试中与Llama 3.1等闭源模型竞争。该模型的改进得益于增强训练稳定性、分阶段训练和基于Tülu 3框架的先进后训练方法等创新。Ai2还发布了模型权重、数据、代码、配方、中间检查点和指令微调模型等资源,并推出OLMES评估框架,旨在促进开源AI的发展。

🤔OLMo 2模型在训练数据和性能方面取得了显著提升,其7B和13B参数版本分别在不同基准测试中表现出色,训练数据量高达5万亿tokens,性能与闭源模型Llama 3.1相当甚至超越。

🚀OLMo 2的改进得益于一系列创新技术,包括增强训练稳定性措施、分阶段训练方法以及基于Tülu 3框架的先进后训练方法,例如将非参数层归一化切换到RMSNorm,并实施旋转位置嵌入。

📚OLMo 2采用了分阶段训练方法,第一阶段使用OLMo-Mix-1124数据集,包含约3.9万亿tokens,第二阶段则结合了高质量网络数据和特定领域内容。

🤝Ai2秉承开放科学的理念,发布了模型权重、数据、代码、配方、中间检查点和指令微调模型等资源,并推出OLMES评估框架,用于评估模型的核心能力,例如知识回忆、常识推理和数学推理。

💡OLMo 2的发布提高了开源AI开发的标准,有望加速该领域的创新步伐,同时保持透明度和可访问性。

Ai2 is releasing OLMo 2, a family of open-source language models that advances the democratisation of AI and narrows the gap between open and proprietary solutions.

The new models, available in 7B and 13B parameter versions, are trained on up to 5 trillion tokens and demonstrate performance levels that match or exceed comparable fully open models whilst remaining competitive with open-weight models such as Llama 3.1 on English academic benchmarks.

“Since the release of the first OLMo in February 2024, we’ve seen rapid growth in the open language model ecosystem, and a narrowing of the performance gap between open and proprietary models,” explained Ai2.

The development team achieved these improvements through several innovations, including enhanced training stability measures, staged training approaches, and state-of-the-art post-training methodologies derived from their Tülu 3 framework. Notable technical improvements include the switch from nonparametric layer norm to RMSNorm and the implementation of rotary positional embedding.

OLMo 2 model training breakthrough

The training process employed a sophisticated two-stage approach. The initial stage utilised the OLMo-Mix-1124 dataset of approximately 3.9 trillion tokens, sourced from DCLM, Dolma, Starcoder, and Proof Pile II. The second stage incorporated a carefully curated mixture of high-quality web data and domain-specific content through the Dolmino-Mix-1124 dataset.

Particularly noteworthy is the OLMo 2-Instruct-13B variant, which is the most capable model in the series. The model demonstrates superior performance compared to Qwen 2.5 14B instruct, Tülu 3 8B, and Llama 3.1 8B instruct models across various benchmarks.

(Credit: Ai2)

Commiting to open science

Reinforcing its commitment to open science, Ai2 has released comprehensive documentation including weights, data, code, recipes, intermediate checkpoints, and instruction-tuned models. This transparency allows for full inspection and reproduction of results by the wider AI community.

The release also introduces an evaluation framework called OLMES (Open Language Modeling Evaluation System), comprising 20 benchmarks designed to assess core capabilities such as knowledge recall, commonsense reasoning, and mathematical reasoning.

OLMo 2 raises the bar in open-source AI development, potentially accelerating the pace of innovation in the field whilst maintaining transparency and accessibility.

(Photo by Rick Barrett)

See also: OpenAI enhances AI safety with new red teaming methods

Want to learn more about AI and big data from industry leaders? Check out AI & Big Data Expo taking place in Amsterdam, California, and London. The comprehensive event is co-located with other leading events including Intelligent Automation Conference, BlockX, Digital Transformation Week, and Cyber Security & Cloud Expo.

Explore other upcoming enterprise technology events and webinars powered by TechForge here.

The post Ai2 OLMo 2: Raising the bar for open language models appeared first on AI News.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

OLMo 2 开源语言模型 人工智能 AI
相关文章