TechCrunch News 01月08日
Nvidia CEO says his AI chips are improving faster than Moore’s Law
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

英伟达CEO黄仁勋表示,其AI芯片的性能提升速度已超越摩尔定律的历史规律。他指出,英伟达的系统进步速度更快,最新数据中心超级芯片在运行AI推理工作负载时比上一代快30倍以上。黄仁勋认为,通过同时创新架构、芯片、系统、库和算法,可以超越摩尔定律。他还强调,AI的发展并未停滞,而是存在三种活跃的扩展定律:预训练、后训练和测试时计算。英伟达致力于提高计算能力,从而降低AI推理成本,并预计AI推理模型的成本将随着性能的提升而降低。英伟达的AI芯片性能在过去十年提升了1000倍,远超摩尔定律的速度。

🚀 英伟达AI芯片性能飞跃:英伟达CEO黄仁勋指出,其AI芯片的性能提升速度已超越摩尔定律,最新数据中心超级芯片在AI推理工作负载方面比上一代快30倍以上。

💡 AI发展的三大定律:黄仁勋认为AI发展并非停滞,而是存在预训练、后训练和测试时计算三种活跃的扩展定律,这些定律推动着AI模型能力的进步。

💰 降低推理成本:英伟达致力于提高计算能力,从而降低AI推理成本,并预计随着性能的提升,AI推理模型的成本将逐渐降低。这对于AI的广泛应用至关重要。

📈 性能十年提升千倍:黄仁勋表示,英伟达的AI芯片在过去十年中性能提升了1000倍,远远超过摩尔定律的增速,展现了其在AI硬件领域的强大实力。

Nvidia CEO Jensen Huang says the performance of his company’s AI chips is advancing faster than historical rates set by Moore’s Law, the rubric that drove computing progress for decades.

“Our systems are progressing way faster than Moore’s Law,” said Huang in an interview with TechCrunch on Tuesday, the morning after he delivered a keynote to a 10,000-person crowd at CES in Las Vegas.

Coined by Intel co-founder Gordon Moore in 1965, Moore’s Law predicted that the number of transistors on computer chips would roughly double every year, essentially doubling the performance of those chips. This prediction mostly panned out, and created rapid advances in capability and plummeting costs for decades.

In recent years, Moore’s Law has slowed down. However, Huang claims that Nvidia’s AI chips are moving at an accelerated pace of their own; the company says its latest datacenter superchip is more than 30x faster for running AI inference workloads than its previous generation.

“We can build the architecture, the chip, the system, the libraries, and the algorithms all at the same time,” said Huang. “If you do that, then you can move faster than Moore’s Law, because you can innovate across the entire stack.”

The bold claim from Nvidia’s CEO comes at a time when many are questioning whether AI’s progress has stalled. Leading AI labs – such as Google, OpenAI, and Anthropic – use Nvidia’s AI chips to train and run their AI models, and advancements to these chips would likely translate to further progress in AI model capabilities.

Huang rejects the idea that AI progress is slowing. Instead he claims there are now three active AI scaling laws: pre-training, the initial training phase where AI models learn patterns from large amounts of data; post-training, which fine tunes an AI model’s answers using methods such as human feedback; and test-time compute, which occurs during the inference phase and gives an AI model more time to “think” after each question.

“Moore’s Law was so important in the history of computing because it drove down computing costs,” Huang told TechCrunch. “The same thing is going to happen with inference where we drive up the performance, and as a result, the cost of inference is going to be less.”

(Of course, Nvidia has grown to be the most valuable company on Earth by riding the AI boom, so it benefits Huang to say so.)

Nvidia CEO Jensen Huang using a gb200 nvl72 like a shield (image credits: Nvidia)

Nvidia’s H100s were the chip of choice for tech companies looking to train AI models, but now that tech companies are focusing more on inference, some have questioned whether Nvidia’s expensive chips will still stay on top.

AI models that use test-time compute are expensive to run today. There’s concern that OpenAI’s o3 model, which uses a scaled up version of test-time compute, would be too expensive for most people to use. For example, OpenAI spent nearly $20 per task using o3 to achieve human-level scores on a test of general intelligence. A ChatGPT Plus subscription costs $20 for an entire month of usage.

Huang held up Nvidia’s latest datacenter superchip, the GB200 NVL72, onstage like a shield during Monday’s keynote. This chip is 30 to 40x faster at running AI inference workloads than Nvidia’s previous best selling chips, the H100. Huang says this performance jump means that AI reasoning models like OpenAI’s o3, which uses a significant amount of compute during the inference phase, will become cheaper over time.

Huang says he’s overall focused on creating more performant chips, and that more performant chips create lower prices in the long run.

“The direct and immediate solution for test-time compute, both in performance and cost affordability, is to increase our computing capability,” Huang told TechCrunch. He noted that in the long term, AI reasoning models could be used to create better data for the pre-training and post-training of AI models.

We’ve certainly seen the price of AI models plummet in the last year, in part due to computing breakthroughs from hardware companies like Nvidia. Huang says that’s a trend he expects to continue with AI reasoning models, even though the first versions we’ve seen from OpenAI have been rather expensive.

More broadly, Huang claimed his AI chips today are 1,000x better than what it made 10 years ago. That’s a much faster pace than the standard set by Moore’s law, one Huang says he sees no sign of stopping soon.

LIVE 3 seconds ago

CES 2025, the annual consumer tech conference held in Las Vegas, is upon us — and this is where you…

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

英伟达 AI芯片 摩尔定律 AI推理 黄仁勋
相关文章