Nvidia Blog 02月16日
NVIDIA Blackwell Now Generally Available in the Cloud
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

CoreWeave率先推出基于NVIDIA GB200 NVL72的云实例,成为首家提供NVIDIA Blackwell平台的云服务商。该实例采用机架级NVIDIA NVLink,连接72个NVIDIA Blackwell GPU和36个Grace CPU,并通过NVIDIA Quantum-2 InfiniBand网络扩展至高达11万个GPU,为构建和部署下一代AI推理模型和代理提供所需的规模和性能。NVIDIA GB200 NVL72是液冷机架级解决方案,其72个GPU作为一个整体运行,显著提升了推理token的生成速度并降低服务成本。CoreWeave的托管云服务专为Blackwell构建,通过CoreWeave Kubernetes Service优化工作负载编排,并提供实时NVLink性能、GPU利用率和温度的洞察。

🚀 CoreWeave成为首家提供NVIDIA Blackwell平台的云服务商,推出基于NVIDIA GB200 NVL72的实例,为AI推理模型和代理提供强大算力支持。

💡 NVIDIA GB200 NVL72采用液冷和机架级设计,拥有72个GPU通过NVLink互联,作为一个统一的GPU运行,显著提升AI性能并降低成本。其第五代NVLink技术提供130TB/s的GPU带宽,第二代Transformer引擎支持FP4,进一步加速AI计算。

🌐 CoreWeave的托管云服务专为Blackwell优化,CoreWeave Kubernetes Service通过暴露NVLink域ID优化工作负载编排,Slurm on Kubernetes (SUNK) 支持拓扑块插件,实现跨GB200 NVL72机架的智能工作负载分配。同时,CoreWeave的Observability Platform提供对NVLink性能、GPU利用率和温度的实时监控。

🤝 NVIDIA提供全栈AI平台,结合Blackwell基础设施和先进软件,帮助企业构建快速、准确且可扩展的AI代理。NVIDIA Blueprints提供预定义的参考工作流,NVIDIA NIM提供易于使用的微服务,NVIDIA NeMo提供AI模型的训练、定制和持续改进工具。

AI reasoning models and agents are set to transform industries, but delivering their full potential at scale requires massive compute and optimized software. The “reasoning” process involves multiple models, generating many additional tokens, and demands infrastructure with a combination of high-speed communication, memory and compute to ensure real-time, high-quality results.

To meet this demand, CoreWeave has launched NVIDIA GB200 NVL72-based instances, becoming the first cloud service provider to make the NVIDIA Blackwell platform generally available.

With rack-scale NVIDIA NVLink across 72 NVIDIA Blackwell GPUs and 36 NVIDIA Grace CPUs, scaling to up to 110,000 GPUs with NVIDIA Quantum-2 InfiniBand networking, these instances provide the scale and performance needed to build and deploy the next generation of AI reasoning models and agents.

NVIDIA GB200 NVL72 on CoreWeave 

NVIDIA GB200 NVL72 is a liquid-cooled, rack-scale solution with a 72-GPU NVLink domain, which enables the six dozen GPUs to act as a single massive GPU.

NVIDIA Blackwell features many technological breakthroughs that accelerate inference token generation, boosting performance while reducing service costs. For example, fifth-generation NVLink enables 130TB/s of GPU bandwidth in one 72-GPU NVLink domain, and the second-generation Transformer Engine enables FP4 for faster AI performance while maintaining high accuracy.

CoreWeave’s portfolio of managed cloud services is purpose-built for Blackwell. CoreWeave Kubernetes Service optimizes workload orchestration by exposing NVLink domain IDs, ensuring efficient scheduling within the same rack. Slurm on Kubernetes (SUNK) supports the topology block plug-in, enabling intelligent workload distribution across GB200 NVL72 racks. In addition, CoreWeave’s Observability Platform provides real-time insights into NVLink performance, GPU utilization and temperatures.

CoreWeave’s GB200 NVL72 instances feature NVIDIA Quantum-2 InfiniBand networking that delivers 400Gb/s bandwidth per GPU for clusters up to 110,000 GPUs. NVIDIA BlueField-3 DPUs also provide accelerated multi-tenant cloud networking, high-performance data access and GPU compute elasticity for these instances.

Full-Stack Accelerated Computing Platform for Enterprise AI 

NVIDIA’s full-stack AI platform pairs cutting-edge software with Blackwell-powered infrastructure to help enterprises build fast, accurate and scalable AI agents.

NVIDIA Blueprints provides pre-defined, customizable, ready-to-deploy reference workflows to help developers create real-world applications. NVIDIA NIM is a set of easy-to-use microservices designed for secure, reliable deployment of high-performance AI models for inference. NVIDIA NeMo includes tools for training, customization and continuous improvement of AI models for modern enterprise use cases. Enterprises can use NVIDIA Blueprints, NIM and NeMo to build and fine-tune models for their specialized AI agents.

These software components, all part of the NVIDIA AI Enterprise software platform, are key enablers to delivering agentic AI at scale and can readily be deployed on CoreWeave.

Bringing Next-Generation AI to the Cloud 

The general availability of NVIDIA GB200 NVL72-based instances on CoreWeave underscores the latest in the companies’ collaboration, focused on delivering the latest accelerated computing solutions to the cloud. With the launch of these instances, enterprises now have access to the scale and performance needed to power the next wave of AI reasoning models and agents.

Customers can start provisioning GB200 NVL72-based instances through CoreWeave Kubernetes Service in the US-WEST-01 region using the gb200-4x instance ID. To get started, contact CoreWeave.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

CoreWeave NVIDIA GB200 NVL72 AI推理 Blackwell平台 云计算
相关文章