Nvidia Blog 02月16日
Accelerate DeepSeek Reasoning Models With NVIDIA GeForce RTX 50 Series AI PCs
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

DeepSeek-R1模型家族的发布为AI社区带来了新的活力,它允许爱好者和开发者在本地PC上运行最先进的推理模型,具备解决问题、数学和代码能力。NVIDIA GeForce RTX 50系列GPU拥有强大的AI算力,可以比PC市场上任何其他产品更快地运行DeepSeek系列模型。推理模型是一种新型的大型语言模型(LLM),它花费更多时间进行“思考”和“反思”来解决复杂问题,同时描述解决任务所需的步骤。RTX AI平台提供了广泛的AI工具、软件开发工具包和模型选择,用户可以通过各种软件体验DeepSeek-R1和RTX AI PC的强大功能。

💡DeepSeek-R1模型家族基于一个大型的6710亿参数的混合专家(MoE)模型。MoE模型由多个较小的专家模型组成,用于解决复杂问题。DeepSeek模型进一步细分工作,并将子任务分配给较小的专家组。

🚀DeepSeek采用了一种称为蒸馏的技术,从大型DeepSeek 6710亿参数模型构建了六个较小的学生模型——参数范围从15亿到700亿。大型DeepSeek-R1 6710亿参数模型的推理能力被传授给较小的Llama和Qwen学生模型,从而产生了强大的、更小的推理模型,可以在RTX AI PC上以快速的性能本地运行。

💻GeForce RTX 50系列GPU配备专用第五代Tensor Cores,基于与数据中心中推动世界领先的AI创新的NVIDIA Blackwell GPU架构相同的架构。RTX完全加速DeepSeek,在PC上提供最大的推理性能。

🛠️NVIDIA的RTX AI平台提供了最广泛的AI工具、软件开发工具包和模型选择,可以在全球超过1亿台NVIDIA RTX AI PC上访问DeepSeek-R1的功能,包括由GeForce RTX 50系列GPU驱动的PC。

The recently released DeepSeek-R1 model family has brought a new wave of excitement to the AI community, allowing enthusiasts and developers to run state-of-the-art reasoning models with problem-solving, math and code capabilities, all from the privacy of local PCs.

With up to 3,352 trillion operations per second of AI horsepower, NVIDIA GeForce RTX 50 Series GPUs can run the DeepSeek family of distilled models faster than anything on the PC market.

A New Class of Models That Reason

Reasoning models are a new class of large language models (LLMs) that spend more time on “thinking” and “reflecting” to work through complex problems, while describing the steps required to solve a task.

The fundamental principle is that any problem can be solved with deep thought, reasoning and time, just like how humans tackle problems. By spending more time — and thus compute — on a problem, the LLM can yield better results. This phenomenon is known as test-time scaling, where a model dynamically allocates compute resources during inference to reason through problems.

Reasoning models can enhance user experiences on PCs by deeply understanding a user’s needs, taking actions on their behalf and allowing them to provide feedback on the model’s thought process — unlocking agentic workflows for solving complex, multi-step tasks such as analyzing market research, performing complicated math problems, debugging code and more.

The DeepSeek Difference

The DeepSeek-R1 family of distilled models is based on a large 671-billion-parameter mixture-of-experts (MoE) model. MoE models consist of multiple smaller expert models for solving complex problems. DeepSeek models further divide the work and assign subtasks to smaller sets of experts.

DeepSeek employed a technique called distillation to build a family of six smaller student models — ranging from 1.5-70 billion parameters — from the large DeepSeek 671-billion-parameter model. The reasoning capabilities of the larger DeepSeek-R1 671-billion-parameter model were taught to the smaller Llama and Qwen student models, resulting in powerful, smaller reasoning models that run locally on RTX AI PCs with fast performance.

Peak Performance on RTX

Inference speed is critical for this new class of reasoning models. GeForce RTX 50 Series GPUs, built with dedicated fifth-generation Tensor Cores, are based on the same NVIDIA Blackwell GPU architecture that fuels world-leading AI innovation in the data center. RTX fully accelerates DeepSeek, offering maximum inference performance on PCs.

Throughput performance of the Deepseek-R1 distilled family of models across GPUs on the PC.

Experience DeepSeek on RTX in Popular Tools

NVIDIA’s RTX AI platform offers the broadest selection of AI tools, software development kits and models, opening access to the capabilities of DeepSeek-R1 on over 100 million NVIDIA RTX AI PCs worldwide, including those powered by GeForce RTX 50 Series GPUs.

High-performance RTX GPUs make AI capabilities always available — even without an internet connection — and offer low latency and increased privacy because users don’t have to upload sensitive materials or expose their queries to an online service.

Experience the power of DeepSeek-R1 and RTX AI PCs through a vast ecosystem of software, including Llama.cpp, Ollama, LM Studio, AnythingLLM, Jan.AI, GPT4All and OpenWebUI, for inference. Plus, use Unsloth to fine-tune the models with custom data.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

DeepSeek-R1 推理模型 RTX AI PC NVIDIA AI
相关文章