MIT Technology Review » Artificial Intelligence 04月22日 22:03
The future of AI processing
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

文章探讨了人工智能(AI)计算从云端向边缘设备转移的趋势,以及异构计算在其中的关键作用。随着AI技术的发展,推理能力逐渐向用户端转移,从而促进了AI在智能手机、汽车和工业物联网等边缘设备上的应用。异构计算通过在不同硬件上动态分配工作负载,优化了延迟、安全性和能效。文章强调了企业在管理系统复杂性和适应未来需求方面面临的挑战,并呼吁开发适应当前和未来需求的灵活架构。

📱 AI推理正向边缘转移:随着AI技术的进步,模型的推理能力不再局限于云端,而是可以部署在智能手机、汽车等边缘设备上,从而减少对云的依赖,提供更快的响应速度和更高的隐私保护。

⚙️ 异构计算赋能无处不在的AI:为了充分发挥AI的潜力,需要在合适的硬件上进行处理和计算。异构计算允许组织在CPU、GPU、NPU等不同计算核心之间动态分配工作负载,从而优化延迟、安全性及能源使用。

⚠️ 面临的挑战:企业在管理系统复杂性以及确保现有架构能够适应未来需求方面面临挑战。尽管微芯片架构有所进步,但软件和工具仍需改进,以支持无处不在的机器学习、生成式AI和新的专业化领域。

Artificial Intelligence (AI) is emerging in everyday use cases, thanks to advances in foundational models, more powerful chip technology, and abundant data. To become truly embedded and seamless, AI computation must now be distributed—and much of it will take place on device and at the edge. 

To support this evolution, computation for running AI workloads must be allocated to the right hardware based on a range of factors, including performance, latency, and power efficiency. Heterogeneous compute enables organizations to allocate workloads dynamically across various computing cores like central processing units (CPUs), graphics processing units (GPUs), neural processing units (NPUs), and other AI accelerators. By assigning workloads to the processors best suited to different purposes, organizations can better balance latency, security, and energy usage in their systems. 

Key findings from the report are as follows: 

More AI is moving to inference and the edge. As AI technology advances, inference—a model’s ability to make predictions based on its training—can now be run closer to users and not just in the cloud. This has advanced the deployment of AI to a range of different edge devices, including smartphones, cars, and industrial internet of things (IIoT). Edge processing reduces the reliance on cloud to offer faster response times and enhanced privacy. Going forward, hardware for on-device AI will only improve in areas like memory capacity and energy efficiency. 

• To deliver pervasive AI, organizations are adopting heterogeneous compute. To commercialize the full panoply of AI use cases, processing and compute must be performed on the right hardware. A heterogeneous approach unlocks a solid, adaptable foundation for the deployment and advancement of AI use cases for everyday life, work, and play. It also allows organizations to prepare for the future of distributed AI in a way that is reliable, efficient, and secure. But there are many trade-offs between cloud and edge computing that require careful consideration based on industry-specific needs. 

Companies face challenges in managing system complexity and ensuring current architectures can adapt to future needs. Despite progress in microchip architectures, such as the latest high-performance CPU architectures optimized for AI, software and tooling both need to improve to deliver a compute platform that supports pervasive machine learning, generative AI, and new specializations. Experts stress the importance of developing adaptable architectures that cater to current machine learning demands, while allowing room for technological shifts. The benefits of distributed compute need to outweigh the downsides in terms of complexity across platforms. 

Download the full report.

This content was produced by Insights, the custom content arm of MIT Technology Review. It was not written by MIT Technology Review’s editorial staff.

This content was researched, designed, and written entirely by human writers, editors, analysts, and illustrators. This includes the writing of surveys and collection of data for surveys. AI tools that may have been used were limited to secondary production processes that passed thorough human review.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

人工智能 边缘计算 异构计算 AI推理 算力
相关文章