MarkTechPost@AI 04月14日
Small Models, Big Impact: ServiceNow AI Releases Apriel-5B to Outperform Larger LLMs with Fewer Resources
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

ServiceNow AI 发布了 Apriel-5B,一款新型小语言模型,旨在解决大型语言模型在资源和部署方面的挑战。Apriel-5B 拥有 48 亿参数,在多种任务上表现出色,同时对硬件要求较低。它包含 Apriel-5B-Base 和 Apriel-5B-Instruct 两个版本,分别用于预训练和指令调优。该模型在训练效率、推理吞吐量和跨领域通用性方面进行了优化,并在多个基准测试中超越了其他更大的模型。 Apriel-5B 的发布为在资源受限的环境中部署实用 AI 提供了新的可能性。

💡Apriel-5B 是 ServiceNow AI 发布的新型小语言模型,包含 48 亿参数,专注于推理吞吐量、训练效率和跨领域通用性。

📚Apriel-5B 有两个版本:Apriel-5B-Base 用于预训练,Apriel-5B-Instruct 针对聊天、推理和任务完成进行了指令调优。

⚙️该模型采用了多种技术优化,包括:旋转位置嵌入 (RoPE),FlashAttention-2,分组查询注意力 (GQA),以及 BFloat16 训练,从而在不依赖特殊硬件或大规模并行化的前提下,保持响应速度和效率。

✅在多个基准测试中,Apriel-5B-Instruct 在通用任务上平均优于 OLMo-2-7B-Instruct 和 Mistral-Nemo-12B-Instruct,在数学相关任务和 IF Eval 上优于 LLaMA-3.1-8B-Instruct。

💰与 OLMo-2-7B 相比,Apriel-5B 训练所需的计算资源减少了 2.3 倍,强调了其训练效率。

As language models continue to grow in size and complexity, so do the resource requirements needed to train and deploy them. While large-scale models can achieve remarkable performance across a variety of benchmarks, they are often inaccessible to many organizations due to infrastructure limitations and high operational costs. This gap between capability and deployability presents a practical challenge, particularly for enterprises seeking to embed language models into real-time systems or cost-sensitive environments.

In recent years, small language models (SLMs) have emerged as a potential solution, offering reduced memory and compute requirements without entirely compromising on performance. Still, many SLMs struggle to provide consistent results across diverse tasks, and their design often involves trade-offs that limit generalization or usability.

ServiceNow AI Releases Apriel-5B: A Step Toward Practical AI at Scale

To address these concerns, ServiceNow AI has released Apriel-5B, a new family of small language models designed with a focus on inference throughput, training efficiency, and cross-domain versatility. With 4.8 billion parameters, Apriel-5B is small enough to be deployed on modest hardware but still performs competitively on a range of instruction-following and reasoning tasks.

The Apriel family includes two versions:

Both models are released under the MIT license, supporting open experimentation and broader adoption across research and commercial use cases.

Architectural Design and Technical Highlights

Apriel-5B was trained on over 4.5 trillion tokens, a dataset carefully constructed to cover multiple task categories, including natural language understanding, reasoning, and multilingual capabilities. The model uses a dense architecture optimized for inference efficiency, with key technical features such as:

These architectural decisions allow Apriel-5B to maintain responsiveness and speed without relying on specialized hardware or extensive parallelization. The instruction-tuned version was fine-tuned using curated datasets and supervised techniques, enabling it to perform well on a range of instruction-following tasks with minimal prompting.

Evaluation Insights and Benchmark Comparisons

Apriel-5B-Instruct has been evaluated against several widely used open models, including Meta’s LLaMA 3.1–8B, Allen AI’s OLMo-2–7B, and Mistral-Nemo-12B. Despite its smaller size, Apriel shows competitive results across multiple benchmarks:

These outcomes suggest that Apriel-5B hits a productive midpoint between lightweight deployment and task versatility, particularly in domains where real-time performance and limited resources are key considerations.

Conclusion: A Practical Addition to the Model Ecosystem

Apriel-5B represents a thoughtful approach to small model design, one that emphasizes balance rather than scale. By focusing on inference throughput, training efficiency, and core instruction-following performance, ServiceNow AI has created a model family that is easy to deploy, adaptable to varied use cases, and openly available for integration.

Its strong performance on math and reasoning benchmarks, combined with a permissive license and efficient compute profile, makes Apriel-5B a compelling choice for teams building AI capabilities into products, agents, or workflows. In a field increasingly defined by accessibility and real-world applicability, Apriel-5B is a practical step forward.


Check out ServiceNow-AI/Apriel-5B-Base and ServiceNow-AI/Apriel-5B-Instruct. All credit for this research goes to the researchers of this project. Also, feel free to follow us on Twitter and don’t forget to join our 85k+ ML SubReddit.

The post Small Models, Big Impact: ServiceNow AI Releases Apriel-5B to Outperform Larger LLMs with Fewer Resources appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Apriel-5B 小语言模型 ServiceNow AI
相关文章