MarkTechPost@AI 13小时前
NVIDIA AI Releases OpenReasoning-Nemotron: A Suite of Reasoning-Enhanced LLMs Distilled from DeepSeek R1 0528
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

NVIDIA最新发布的OpenReasoning-Nemotron系列大语言模型,包括1.5B、7B、14B和32B参数版本,专注于数学、科学和代码领域的复杂推理任务。这些模型通过从671B DeepSeek R1 0528模型进行蒸馏,实现了高水平的推理能力,同时体积更小、效率更高。Nemotron模型在GSM8K、MATH、ARC、HumanEval等多个推理基准测试中表现优异,超越了同等规模的LLaMA2、Mixtral和DeepSeek-Coder模型。该系列模型采用开放且商业友好的许可,易于通过Hugging Face获取,并与NVIDIA NeMo框架深度集成,为开发者提供了强大且易于部署的AI推理解决方案。

✨ **蒸馏自强大基础模型,精炼推理能力**:OpenReasoning-Nemotron系列模型的核心在于其蒸馏技术,将671B参数的DeepSeek R1模型的高级推理能力转移到更小、更高效的版本中。这种方法优先考虑推理泛化能力,而非单纯的token预测,使得这些紧凑型模型在结构化、高认知要求的任务上表现出色,特别是在数学、科学和编程领域。

🚀 **多尺寸模型适配广泛需求**:该系列提供了1.5B、7B、14B和32B四种不同参数规模的模型,以满足多样化的应用场景。1.5B适用于入门级推理,7B适合中等规模推理和代码助手,14B提供高级推理能力,而32B则接近前沿模型性能,专为逻辑密集型任务设计。所有模型均兼容Transformer架构,支持FP16/INT8量化,并针对NVIDIA GPU和NeMo框架进行了优化。

📊 **推理基准测试表现卓越**:在数学(GSM8K, MATH, MMLU)、科学问答(ARC, OpenBookQA, PubMedQA)和编程(HumanEval, MBPP)等关键推理基准上,OpenReasoning-Nemotron模型均取得了显著的性能提升。与同等规模的LLaMA2、Mixtral和DeepSeek-Coder相比,Nemotron模型在各项指标上均有超越,展现了其在推理方面的高度专业化。

🌐 **开放生态与商业友好**:所有OpenReasoning-Nemotron模型均采用开放且商业许可的模式发布,模型卡、评估脚本和推理权重均可在Hugging Face上获取。这极大地降低了模型的使用门槛,并支持与NVIDIA NeMo框架、TensorRT-LLM、ONNX及Hugging Face Transformers等工具链的无缝集成,便于快速部署到研究和生产环境中。

NVIDIA AI has introduced OpenReasoning-Nemotron, a family of large language models (LLMs) designed to excel in complex reasoning tasks across mathematics, science, and code. This model suite—comprising 1.5B, 7B, 14B, and 32B parameter versions—has been distilled from the 671B DeepSeek R1 0528 model, capturing its high-level reasoning capabilities in significantly smaller and more efficient models.

The release positions NVIDIA as a leading contributor to the open-source LLM ecosystem, delivering models that push state-of-the-art (SOTA) performance while remaining commercially permissive and widely accessible via Hugging Face.

Model Overview and Architecture

Distillation from DeepSeek R1 0528 (671B)

At the heart of OpenReasoning-Nemotron lies a distillation strategy that transfers reasoning ability from DeepSeek R1—a massive 671B parameter model—into smaller architectures. The process prioritizes reasoning generalization over raw token prediction, enabling compact models to perform effectively on structured, high-cognition tasks.

The distillation dataset emphasizes mathematics, science, and programming languages, aligning model capabilities with key reasoning domains.

Model Variants and Specs

Model NameParametersIntended UseHugging Face Page
OpenReasoning-Nemotron-1.5B1.5BEntry-level reasoning and inferenceLink
OpenReasoning-Nemotron-7B7BMid-scale reasoning, good for code/mathLink
OpenReasoning-Nemotron-14B14BAdvanced reasoning capabilitiesLink
OpenReasoning-Nemotron-32B32BNear frontier-model performance in logic-intensive tasksLink

All models are compatible with transformer architectures, support FP16/INT8 quantization, and are optimized for NVIDIA GPUs and NeMo frameworks.

Performance Benchmarks

OpenReasoning-Nemotron models outperform their size-equivalent peers on a wide range of reasoning-specific benchmarks, particularly in:

ModelGSM8K AccuracyHumanEval Pass@1ARC-challengeMATH
7B66.7%34.2%77.3%40.5%
14B72.9%42.0%80.1%47.6%
32B77.5%49.5%83.9%52.3%

All metrics represent best-of evaluations under 0-shot or few-shot settings.

These results outperform LLaMA2, Mixtral, and DeepSeek-Coder at similar scales, underscoring the strength of the reasoning-focused distillation method.

Training Data and Reasoning Specialization

The training corpus is a distilled, high-quality subset of the DeepSeek R1 0528 dataset. Key features include:

This deliberate curation ensures strong alignment with real-world reasoning problems found in both academia and applied ML domains.

Open and Ecosystem Integration

All four OpenReasoning-Nemotron models are released under an open and commercially permissive license, with model cards, evaluation scripts, and inference-ready weights available on Hugging Face:

These models are designed to plug into the NVIDIA NeMo framework, and support TensorRT-LLM, ONNX, and Hugging Face Transformers toolchains, facilitating rapid deployment in production and research settings.

Key Use Cases

Conclusion

NVIDIA’s OpenReasoning-Nemotron models offer a pragmatic, open-source path toward scaling reasoning ability without frontier-scale compute costs. By distilling from the 671B DeepSeek R1 and targeting high-leverage reasoning domains, these models deliver a powerful balance of accuracy, efficiency, and accessibility.

For developers, researchers, and enterprises working on logic-intensive AI applications, OpenReasoning-Nemotron provides a compelling foundation—free from the trade-offs that often accompany proprietary or overgeneralized models.


Frequently Asked Questions (FAQs)

1. What is the difference between OpenReasoning-Nemotron and general-purpose LLMs like LLaMA or Mixtral?
OpenReasoning-Nemotron models are specifically distilled to enhance reasoning in math, science, and code. While LLaMA and Mixtral are trained on broad web corpora, OpenReasoning models emphasize symbolic and multi-step logic, outperforming general-purpose LLMs on domain-specific reasoning benchmarks.

2. How were these models distilled from the 671B DeepSeek R1 0528 model?
The distillation process used high-quality outputs from DeepSeek R1 to guide smaller models during training. This includes a curated reasoning-focused dataset and prompt-based training, allowing the smaller Nemotron variants to replicate the reasoning behavior of a much larger model.

3. Are the OpenReasoning-Nemotron models suitable for commercial use?
Yes. All models in the suite are released with commercially permissive licenses and can be deployed in enterprise environments using NVIDIA’s NeMo, TensorRT-LLM, or Hugging Face Transformers toolkits.

4. Which model size should I use for my application?


Check out the Technical details. All credit for this research goes to the researchers of this project.

Sponsorship Opportunity: Reach the most influential AI developers in US and Europe. 1M+ monthly readers, 500K+ community builders, infinite possibilities. [Explore Sponsorship]

The post NVIDIA AI Releases OpenReasoning-Nemotron: A Suite of Reasoning-Enhanced LLMs Distilled from DeepSeek R1 0528 appeared first on MarkTechPost.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

NVIDIA OpenReasoning-Nemotron 大语言模型 AI推理 模型蒸馏
相关文章