The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence) 2024年05月12日
Parallelism and Acceleration for Large Language Models with Bryan Catanzaro - #507
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Today we’re joined by Bryan Catanzaro, vice president of applied deep learning research at NVIDIA.


Most folks know Bryan as one of the founders/creators of cuDNN, the accelerated library for deep neural networks. In our conversation, we explore his interest in high-performance computing and its recent overlap with AI, his current work on Megatron, a framework for training giant language models, and the basic approach for distributing a large language model on DGX infrastructure. 

We also discuss the three different kinds of parallelism, tensor parallelism, pipeline parallelism, and data parallelism, that Megatron provides when training models, as well as his work on the Deep Learning Super Sampling project and the role it's playing in the present and future of game development via ray tracing. 


The complete show notes for this episode can be found at twimlai.com/go/507.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

相关文章