钛媒体:引领未来商业与生活新知 07月30日 14:10
Infinigence Unveils Next-Gen AI Infrastructure Suite, Aims to Lead China’s AI Deployment Revolution
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

Infinigence公司在WAIC 2025上推出了一系列为AI部署设计的计算平台,包括Infinicloud、InfiniCore和InfiniEdge。这些平台旨在解决中国AI行业计算基础设施碎片化的问题,提供从模型调度到大规模部署的全生命周期支持。CEO夏立学表示,这些平台实现了软硬件协同设计,并能兼容异构计算环境,帮助开发者在不同芯片和架构间无缝切换,提升AI性能。Infinigence还致力于构建中国AI生态系统的“反馈循环”,促进芯片制造商和模型开发者之间的协同,并推动AI的普及化,例如通过跨区域联邦强化学习系统支持中小企业。

🌐 Infinigence推出了面向AI 2.0时代的全栈式“软硬件协同设计的计算基础设施系统”,包括Infinicloud(支持10万GPU的全球AI云平台)、InfiniCore(多千GPU集群高性能智能计算平台)和InfiniEdge(优化于终端的边缘计算解决方案),旨在解决中国AI行业计算基础设施碎片化问题,实现跨芯片、架构和工作负载的无缝迁移,释放规模化智能性能。

💡 公司强调其平台具备“通用计算语言”,能够桥接不同指令集的芯片,实现即插即用、互操作和可组合的计算资源管理。已实现对十余种国产芯片的全栈适配,通过算法和编译器优化带来50%-200%的性能提升,并支持统一调度和混合精度计算,提供具有国际竞争力的成本效益。

🤝 Infinigence致力于构建中国AI生态系统的“反馈循环”,模仿英伟达和OpenAI的模式,促进模型开发者与芯片制造商之间的紧密协作。此外,公司还通过跨区域联邦强化学习系统推动AI民主化,连接不同区域的数据中心闲置GPU资源,使中小企业也能利用消费级显卡构建和微调模型。

🚀 公司在AI推理方面,正积极推动用户从国际芯片向国产加速器迁移,其Infini-Ask系列产品包括全球首个端上内在模型Infini-Megrez 2.0,以及与联想合作的Infini-Mizar 2.0(支持AI PC异构计算),和与苏州意歌科技联合开发的低成本FPGA大模型推理引擎,显示了其在国产化和边缘计算领域的布局。

💰 Infinigence自2023年5月成立以来,已迅速获得超过10亿元人民币融资,包括2024年创纪录的5亿元人民币A轮融资,成为中国AI基础设施领域的最大单笔融资。其产品组合覆盖模型托管、云管理到边缘优化和迁移,服务于智能计算中心、模型提供商和工业领域,公司愿景是实现“无界智能,流畅计算”。

Xia Lixue, Co-founder and CEO of Infinigence

 

AsianFin -- Infinigence, an AI infrastructure startup backed by Tsinghua University, introduced a sweeping portfolio of performance-optimized computing platforms targeting the full spectrum of AI deployment at this year’s World Artificial Intelligence Conference (WAIC 2025) .

The company officially launched three flagship products under its integrated solution suite: Infinicloud, a global-scale AI cloud platform for clusters of up to 100,000 GPUs; InfiniCore, a high-performance intelligent computing platform designed for multi-thousand-GPU clusters; and InfiniEdge, a lean, edge computing solution optimized for terminal deployments with as few as one GPU.

Together, the platforms represent what CEO Xia Lixue calls a “software-hardware co-designed infrastructure system for the AI 2.0 era.” Built for compatibility across heterogeneous computing environments, the Infinigence stack offers full lifecycle support—from model scheduling and performance optimization to large-scale application deployment.

“We’re addressing a core bottleneck in China’s AI industry: fragmentation in compute infrastructure,” Xia said. “With InfiniCloud, InfiniCore, and InfiniEdge, we’re enabling AI developers to move seamlessly between different chips, architectures, and workloads—unlocking intelligent performance at scale.”

In a fast-evolving AI landscape dominated by open-source large language models such as DeepSeek, GLM-4.5, and MiniMax M1, Chinese infra startups are racing to build the backbone that powers model deployment and inference.

Early on July 29, Infinigence announced that InfiniCloud now supports Zhipu AI’s latest GLM-4.5 and GLM-4.5-air models, which currently rank third globally in performance. The move signals Infinigence’s ambition to anchor the growing synergy between Chinese model developers and domestic chipmakers.

Xia likened the trio of newly launched platforms to “three bundled boxes” that can be matched to AI workloads of any scale. “From a single smartphone to clusters of 100,000 GPUs—our system is designed to ensure resource efficiency and intelligent elasticity,” he said.

Infinigence’s platforms are already powering Shanghai ModelSpeed Space, the world’s largest AI incubator. The facility sees daily token call volumes exceed 10 billion, supports over 100 AI use cases, and reaches tens of millions of monthly active users across its applications.

A key challenge for China’s AI infrastructure sector is hardware heterogeneity. With dozens of domestic chip vendors and proprietary architectures, developers often struggle to port models across systems.

Xia emphasized that Infinigence has developed a “universal compute language” that bridges chips with disparate instruction sets. “We treat computing resources like supermarket goods—plug-and-play, interoperable, and composable,” he said.

The company’s infrastructure has already achieved full-stack adaptation for more than a dozen domestic chips, delivering 50%–200% performance gains through algorithm and compiler optimization. It also supports unified scheduling and mixed-precision computing, enabling cost-performance ratios that beat many international offerings.

“What’s missing in China’s ecosystem is a feedback loop,” Xia said. “In the U.S., NVIDIA and OpenAI form a tight cycle: model developers know what chips are coming, and chipmakers know what models are being built. We’re building that loop domestically.”

Infinigence is also targeting AI democratization with a first-of-its-kind cross-regional federated reinforcement learning system. The system links idle GPU resources from different regional AIDC centers into a unified compute cluster—allowing SMEs to build and fine-tune domain-specific inference models using consumer-grade cards.

To support this, Infinigence launched the “AIDC Joint Operations Innovation Ecosystem Initiative” in partnership with China’s three major telecom providers and 20+ AIDC institutions.

Xia noted that while training still depends heavily on NVIDIA hardware, inference workloads are rapidly migrating to domestic accelerators. “Users often start with international chips on our platform, but we help them transition to Chinese cards—many of which now deliver strong commercial value,” he said.

Infinigence has also rolled out a series of on-device and edge inference engines under its Infini-Ask line. These include:

Founded in May 2023, Infinigence has raised more than RMB 1 billion in just two years, including a record-setting RMB 500 million Series A round in 2024—the largest to date in China’s AI infrastructure sector.

Its product portfolio now spans everything from model hosting and cloud management to edge optimization and model migration—serving clients across intelligent computing centers, model providers, and industrial sectors.

The company’s broader mission, Xia said, is to balance scale, performance, and resource availability. “Our vision is to deliver ‘boundless intelligence and flawless computing’—wherever there's compute, we want Infinigence to be the intelligence that flows through it.”

IEEE Fellow and Tsinghua professor Wang Yu, also a co-founder of Infinigence, argued that the future of China’s AI economy depends on interdisciplinary collaboration. “We need people who understand chips, models, commercialization, and investment,” Wang said. “Only then can we solve the ‘last mile’ problem—connecting AI research with real-world deployment.”

As China looks to decouple from foreign hardware dependence while competing globally in next-gen AI, Infinigence is positioning itself as a vital enabler—fusing chip-level control with cloud-scale ambition.

“Every AI system runs on two forces: models and compute,” Xia said. “They cannot evolve in silos—they must move forward in sync.”

更多精彩内容,关注钛媒体微信号(ID:taimeiti),或者下载钛媒体App

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

Infinigence AI基础设施 AI计算平台 中国AI 国产化
相关文章