AWS Blogs 前天 05:03
New Amazon EC2 P6e-GB200 UltraServers accelerated by NVIDIA Grace Blackwell GPUs for the highest AI performance
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

亚马逊云服务(AWS)宣布推出Amazon EC2 P6e-GB200 UltraServers,由NVIDIA Grace Blackwell GPU加速,旨在为AI训练和推理提供卓越的性能。这些UltraServers通过专用互连连接多个EC2实例,实现高带宽和低延迟。P6e-GB200 UltraServers配备了NVIDIA Blackwell GPU,提供高达360 petaflops的FP8计算能力和13.4 TB的HBM3e内存,适用于大规模AI工作负载,包括混合专家模型和推理模型,并支持生成式AI应用。

⚡️Amazon EC2 P6e-GB200 UltraServers是基于NVIDIA GB200 NVL72的,旨在为AI训练和推理提供最高的GPU性能。

🚀每个NVIDIA Grace Blackwell Superchip包含两个高性能NVIDIA Blackwell Tensor Core GPU和一个基于Arm架构的NVIDIA Grace CPU,通过NVIDIA NVLink-C2C互连连接。

💻EC2 P6e-GB200 UltraServers提供高达72个NVIDIA Blackwell GPU,FP8计算能力达到360 petaflops,总带宽内存(HBM3e)达13.4 TB。

🌐P6e-GB200 UltraServers通过AWS Nitro System部署在EC2 UltraClusters中,可以安全可靠地扩展到数万个GPU。

💡用户可以通过EC2 Capacity Blocks for ML在达拉斯本地区域使用P6e-GB200 UltraServers,并可与Amazon SageMaker Hyperpod、Amazon EKS和Amazon FSx for Lustre等AWS服务无缝集成。

<section class="blog-post-content lb-rtxt"><table id="amazon-polly-audio-table"><tbody><tr><td id="amazon-polly-audio-tab"><p></p></td></tr></tbody></table><p>Today, we’re announcing the general availability of <a href="https://aws.amazon.com/ec2/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Amazon Elastic Compute Cloud (Amazon EC2)</a> <a href="https://aws.amazon.com/ec2/instance-types/p6/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;P6e-GB200 UltraServers</a>, accelerated by NVIDIA GB200 NVL72 to offer the highest GPU performance for AI training and inference. <a href="https://aws.amazon.com/ec2/ultraservers/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Amazon EC2 UltraServers</a> connect multiple EC2 instances using a dedicated, high-bandwidth, and low-latency accelerator interconnect across these instances.</p><p>The NVIDIA Grace Blackwell Superchips connect two high-performance NVIDIA Blackwell tensor core GPUs and an NVIDIA Grace CPU based on Arm architecture using the NVIDIA NVLink-C2C interconnect. Each Grace Blackwell Superchip delivers 10 petaflops of FP8 compute (without sparsity) and up to 372 GB HBM3e memory. With the superchip architecture, GPU and CPU are colocated within one compute module, increasing bandwidth between GPU and CPU significantly compared to current generation <a href="https://aws.amazon.com/ec2/instance-types/p5/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;EC2 P5en instances</a>.</p><p>With EC2 P6e-GB200 UltraServers, you can access up to 72 NVIDIA Blackwell GPUs within one NVLink domain to use 360 petaflops of FP8 compute (without sparsity) and 13.4 TB of total high bandwidth memory (HBM3e). Powered by the <a href="https://aws.amazon.com/ec2/nitro/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;AWS Nitro System</a>, P6e-GB200 UltraServers are deployed in EC2 UltraClusters to securely and reliably scale to tens of thousands of GPUs.</p><p>EC2 P6e-GB200 UltraServers deliver up to 28.8 Tbps of total <a href="https://aws.amazon.com/hpc/efa/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Elastic Fabric Adapter</a> (EFAv4) networking. EFA is also coupled with NVIDIA GPUDirect RDMA to enable low-latency GPU-to-GPU communication between servers with operating system bypass.</p><p><strong class="c6">EC2 P6e-GB200 UltraServers specifications</strong><br />EC2 P6e-GB200 UltraServers are available in sizes ranging from 36 to 72 GPUs under NVLink. Here are the specs for EC2 P6e-GB200 UltraServers:</p><table class="c10"><tbody><tr class="c8"><td class="c7"><strong>UltraServer type</strong></td><td class="c7"><strong>GPUs<br /></strong></td><td class="c7"><strong>GPU<br />memory (GB)</strong></td><td class="c7"><strong>vCPUs</strong></td><td class="c7"><strong>Instance memory<br />(GiB)</strong></td><td class="c7"><strong>Instance storage (TB)</strong></td><td class="c7"><strong>Aggregate EFA Network Bandwidth (Gbps)</strong></td><td class="c7"><strong>EBS bandwidth (Gbps)</strong></td></tr><tr class="c9"><td class="c7"><strong>u-p6e-gb200x36</strong></td><td class="c7">36</td><td class="c7">6660</td><td class="c7">1296</td><td class="c7">8640</td><td class="c7">202.5</td><td class="c7">14400</td><td class="c7">540</td></tr><tr class="c9"><td class="c7"><strong>u-p6e-gb200x72</strong></td><td class="c7">72</td><td class="c7">13320</td><td class="c7">2592</td><td class="c7">17280</td><td class="c7">405</td><td class="c7">28800</td><td class="c7">1080</td></tr></tbody></table><p>P6e-GB200 UltraServers are ideal for the most compute and memory intensive AI workloads, such as training and inference of frontier models, including mixture of experts models and reasoning models, at the trillion-parameter scale.</p><p>You can build agentic and <a href="https://aws.amazon.com/generative-ai/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;generative AI</a> applications, including question answering, code generation, video and image generation, speech recognition, and more.</p><p><strong class="c6">P6e-GB200 UltraServers in action</strong><br />You can use EC2 P6e-GB200 UltraServers in the Dallas Local Zone through <a href="https://aws.amazon.com/ec2/capacityblocks/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;EC2 Capacity Blocks for ML</a>. The Dallas Local Zone (<code>us-east-1-dfw-2a</code>) is an extension of the US East (N. Virginia) Region.</p><p>To reserve your EC2 Capacity Blocks, choose <strong>Capacity Reservations</strong> on the <a href="https://us-east-2.console.aws.amazon.com/ec2/home?region=us-east-2#CapacityReservations:&quot;&gt;Amazon EC2 console</a>. You can select <strong>Purchase Capacity Blocks for ML</strong> and then choose your total capacity and specify how long you need the EC2 Capacity Block for <strong>u-p6e-gb200x36</strong> or <strong>u-p6e-gb200x72</strong> UltraServers.</p><p>Once Capacity Block is successfully scheduled, it is charged up front and its price doesn’t change after purchase. The payment will be billed to your account within 12 hours after you purchase the EC2 Capacity Blocks. To learn more, visit <a href="https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/ec2-capacity-blocks.html?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Capacity Blocks for ML</a> in the Amazon EC2 User Guide.</p><p>To run instances within your purchased Capacity Block, you can use <a href="https://console.aws.amazon.com/ec2?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;AWS Management Console</a>, <a href="https://aws.amazon.com/cli/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;AWS Command Line Interface (AWS CLI)</a> or <a href="http://docs.aws.amazon.com/AWSJavaScriptSDK/latest/AWS/EC2.html&quot;&gt;AWS SDKs</a>. On the software side, you can start with the <a href="https://aws.amazon.com/ai/machine-learning/amis/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;AWS Deep Learning AMIs</a>. These images are preconfigured with the frameworks and tools that you probably already know and use: PyTorch, JAX, and a lot more.</p><p>You can also integrate EC2 P6e-GB200 UltraServers seamlessly with various AWS managed services. For example:</p><ul><li><a href="https://aws.amazon.com/sagemaker-ai/hyperpod/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Amazon SageMaker Hyperpod</a> provides managed, resilient infrastructure that automatically handles the provisioning and management of P6e-GB200 UltraServers, replacing faulty instances with preconfigured spare capacity within the same NVLink domain to maintain performance.</li><li><a href="https://aws.amazon.com/eks/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Amazon Elastic Kubernetes Services (Amazon EKS)</a> allows one managed node group to span across multiple P6e-GB200 UltraServers as nodes, automating their provisioning and lifecycle management within Kubernetes clusters. You can use EKS topology-aware routing for P6e-GB200 UltraServers, enabling optimal placement of tightly coupled components of distributed workloads within a single UltraServer’s NVLink-connected instances.</li><li><a href="https://aws.amazon.com/fsx/lustre/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Amazon FSx for Lustre</a> file systems provide data access for P6e-GB200 UltraServers at the hundreds of GB/s of throughput and millions of input/output operations per second (IOPS) required for large-scale HPC and AI workloads. For fast access to large datasets, you can use up to 405 TB of local NVMe SSD storage or virtually unlimited cost-effective storage with <a href="https://aws.amazon.com/s3/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Amazon Simple Storage Service (Amazon S3)</a>.</li></ul><p><strong class="c6">Now available</strong><br />Amazon EC2 P6e-GB200 UltraServers are available today in the Dallas Local Zone (<code>us-east-1-dfw-2a</code>) through <a href="https://aws.amazon.com/ec2/capacityblocks/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;EC2 Capacity Blocks for ML</a>. For more information, visit the <a href="https://aws.amazon.com/ec2/pricing/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Amazon EC2 pricing page</a>.</p><p>Give Amazon EC2 P6e-GB200 UltraServers a try in the <a href="https://console.aws.amazon.com/ec2/&quot;&gt;Amazon EC2 console</a>. To learn more, visit the <a href="https://aws.amazon.com/ec2/instance-types/p6/?trk=769a1a2b-8c19-4976-9c45-b6b1226c7d20&amp;amp;sc_channel=el&quot;&gt;Amazon EC2 P6e instance page</a> and send feedback to <a href="https://repost.aws/tags/TAO-wqN9fYRoyrpdULLa5y7g/amazon-ec-2&quot;&gt;AWS re:Post for EC2</a> or through your usual AWS Support contacts.</p><p>— <a href="https://twitter.com/channyun&quot;&gt;Channy&lt;/a&gt;&lt;/p&gt;&lt;/section&gt;&lt;aside id="Comments" class="blog-comments"><div data-lb-comp="aws-blog:cosmic-comments" data-env="prod" data-content-id="343b549a-4e53-40d6-8a7f-e1d38b67a7f1" data-title="New Amazon EC2 P6e-GB200 UltraServers accelerated by NVIDIA Grace Blackwell GPUs for the highest AI performance" data-url="https://aws.amazon.com/blogs/aws/new-amazon-ec2-p6e-gb200-ultraservers-powered-by-nvidia-grace-blackwell-gpus-for-the-highest-ai-performance/&quot;&gt;&lt;p data-failed-message="Comments cannot be loaded… Please refresh and try again.">Loading comments…</p></div></aside>

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Amazon EC2 P6e-GB200 UltraServers NVIDIA Blackwell AI 云计算
相关文章