Unite.AI 04月10日 04:53
Google Cloud Next 2025: Doubling Down on AI with Silicon, Software, and an Open Agent Ecosystem
index_new5.html
../../../zaker_core/zaker_tpl_static/wap/tpl_guoji1.html

 

在2025年谷歌云Next大会上,谷歌云展示了其进军企业AI市场的全面战略。谷歌强调了其在AI领域的进展,发布了Ironwood TPU、Gemini 2.5 Flash和Cloud WAN等产品,并重点介绍了Agent2Agent协议。这些举措旨在推动AI从理论走向实践,构建一个开放、互通的AI生态系统,以提升企业AI应用效率。谷歌希望通过其定制芯片、全球网络基础设施和开放协议,在激烈的云服务竞争中占据更有利的位置。

🚀 谷歌推出了Ironwood TPU,这是一款专为AI推理设计的定制芯片,旨在提升性能和能效,Ironwood在单芯片算力、内存带宽和互联方面均有显著提升,能效比Trillium提升2倍。

💡 Gemini 2.5 Flash是谷歌针对低延迟和成本效益优化的AI模型,适用于高频次、实时的应用场景,它具有动态的“思考预算”,可以根据查询的复杂程度调整处理能力,平衡速度、成本和准确性。

🌐 谷歌发布了Cloud WAN,将谷歌的全球网络基础设施开放给企业用户,提供更快的性能和更低的总体拥有成本。Cloud WAN利用谷歌现有的网络基础设施,为数据中心互联和分支机构/园区网络提供高性能连接。

🤝 谷歌推出了Agent2Agent(A2A)协议,这是一个开放标准,旨在实现不同AI代理之间的互操作性,从而解决企业内部AI系统孤立的问题。A2A协议的目标是让不同框架和供应商构建的AI代理能够安全通信和协作。

Las Vegas is playing host to Google Cloud Next 2025, an event unfolding at a critical moment for the technology industry. The artificial intelligence arms race among the cloud titans – Amazon Web Services (AWS), Microsoft Azure, and Google Cloud – is escalating rapidly. Google, often cast as the third contender despite its formidable technological prowess and deep AI research roots, seized the Cloud Next stage to articulate a comprehensive and aggressive strategy aimed squarely at the enterprise AI market.

The narrative, delivered by Google Cloud CEO Thomas Kurian and echoed by Google and Alphabet CEO Sundar Pichai, centered on moving AI transformation from mere possibility to tangible reality. Google underscored its claimed momentum, citing over 3,000 product advancements in the past year, a twentyfold surge in Vertex AI platform usage since the previous Cloud Next event, more than four million developers actively building with its Gemini family of models, and showcasing over 500 customer success stories during the conference.

However, Google Cloud Next 2025 was more than a showcase of incremental updates or impressive metrics. It also unveiled a multi-pronged offensive. By launching powerful, inference-optimized custom silicon (the Ironwood TPU), refining its flagship AI model portfolio with a focus on practicality (Gemini 2.5 Flash), opening its vast global network infrastructure to enterprises (Cloud WAN), and making a significant, strategic bet on an open, interoperable ecosystem for AI agents (the Agent2Agent protocol), Google is aggressively positioning itself to define the next evolutionary phase of enterprise AI – what the company is increasingly terming the “agentic era.”



Ironwood, Gemini, and the Network Effect

Central to Google's AI ambitions is its continued investment in custom silicon. The star of Cloud Next 2025 was Ironwood, the seventh generation of Google's Tensor Processing Unit (TPU). Critically, Ironwood is presented as the first TPU designed explicitly for AI inference – the process of using trained models to make predictions or generate outputs in real-world applications. 

The performance claims for Ironwood are substantial. Google detailed configurations scaling up to an immense 9,216 liquid-cooled chips interconnected within a single pod. This largest configuration is claimed to deliver a staggering 42.5 exaflops of compute power. Google asserts this represents more than 24 times the per-pod compute power of El Capitan, currently ranked as the world's most powerful supercomputer.

While impressive, it's important to note such comparisons often involve different levels of numerical precision, making direct equivalency complex. Nonetheless, Google positions Ironwood as a greater than tenfold improvement over its previous high-performance TPU generation.

Beyond raw compute, Ironwood boasts significant advancements in memory and interconnectivity compared to its predecessor, Trillium (TPU v6).

Perhaps equally important is the emphasis on energy efficiency. Google claims Ironwood delivers twice the performance per watt compared to Trillium and is nearly 30 times more power-efficient than its first Cloud TPU from 2018. This directly addresses the growing constraint of power availability in scaling data centers for AI.

Google TPU Generation Comparison: Ironwood (v7) vs. Trillium (v6)

FeatureTrillium (TPU v6)Ironwood (TPU v7)Improvement Factor
Primary FocusTraining & InferenceInferenceSpecialization
Peak Compute/ChipNot directly comparable (diff gen)4,614 TFLOPs (FP8 likely)
HBM Capacity/Chip32 GB (estimated based on 6x claim)192 GB6x
HBM Bandwidth/Chip~1.6 Tbps (estimated based on 4.5x)7.2 Tbps4.5x
ICI Bandwidth (bidir.)~0.8 Tbps (estimated based on 1.5x)1.2 Tbps1.5x
Perf/Watt vs. Prev GenBaseline for comparison2x vs Trillium2x
Perf/Watt vs. TPU v1 (2018)~15x (estimated)Nearly 30x~2x vs Trillium


Note: Some Trillium figures are estimated based on Google's claimed improvement factors for Ironwood. Peak compute comparison is complex due to generational differences and likely precision variations.

Ironwood forms a key part of Google's “AI Hypercomputer” concept – an architecture integrating optimized hardware (including TPUs and GPUs like Nvidia's Blackwell and upcoming Vera Rubin), software (like the Pathways distributed ML runtime), storage (Hyperdisk Exapools, Managed Lustre), and networking to tackle demanding AI workloads.

On the model front, Google introduced Gemini 2.5 Flash, a strategic counterpoint to the high-end Gemini 2.5 Pro. While Pro targets maximum quality for complex reasoning, Flash is explicitly optimized for low latency and cost efficiency, making it suitable for high-volume, real-time applications like customer service interactions or rapid summarization.

Gemini 2.5 Flash features a dynamic “thinking budget” that adjusts processing based on query complexity, allowing users to tune the balance between speed, cost, and accuracy. This simultaneous focus on a high-performance inference chip (Ironwood) and a cost/latency-optimized model (Gemini Flash) underscores Google's push towards the practical operationalization of AI, recognizing that the cost and efficiency of running models in production are becoming paramount concerns for enterprises.

Complementing the silicon and model updates is the launch of Cloud WAN. Google is effectively productizing its massive internal global network – spanning over two million miles of fiber, connecting 42 regions via more than 200 points of presence – making it directly available to enterprise customers. 

Google claims this service can deliver up to 40% faster performance compared to the public internet and reduce total cost of ownership by up to 40% versus self-managed WANs, backed by a 99.99% reliability SLA. Primarily targeting high-performance connectivity between data centers and connecting branch/campus environments, Cloud WAN leverages Google's existing infrastructure, including the Network Connectivity Center. 

While Google cited Nestlé and Citadel Securities as early adopters, this move fundamentally weaponizes a core infrastructure asset. It transforms an internal operational necessity into a competitive differentiator and potential revenue stream, directly challenging both traditional telecommunication providers and the networking offerings of rival cloud platforms like AWS Cloud WAN and Azure Virtual WAN.

(Source: Google DeepMind)

The Agent Offensive: Building Bridges with ADK and A2A

Beyond infrastructure and core models, Google Cloud Next 2025 placed an extraordinary emphasis on AI agents and the tools to build and connect them. The vision presented extends far beyond simple chatbots, envisioning sophisticated systems capable of autonomous reasoning, planning, and executing complex, multi-step tasks. The focus is clearly shifting towards enabling multi-agent systems, where specialized agents collaborate to achieve broader goals.

To facilitate this vision, Google introduced the Agent Development Kit (ADK). ADK is an open-source framework, initially available in Python, designed to simplify the creation of individual agents and complex multi-agent systems. Google claims developers can build a functional agent with under 100 lines of code. 

Key features include a code-first approach for precise control, native support for multi-agent architectures, flexible tool integration (including support for the Model Context Protocol, or MCP), built-in evaluation capabilities, and deployment options ranging from local containers to the managed Vertex AI Agent Engine. ADK also uniquely supports bidirectional audio and video streaming for more natural, human-like interactions. An accompanying “Agent Garden” provides ready-to-use samples and over 100 pre-built connectors to jumpstart development.

The true centerpiece of Google's agent strategy, however, is the Agent2Agent (A2A) protocol. A2A is a new, open standard designed explicitly for agent interoperability. Its fundamental goal is to allow AI agents, regardless of the framework they were built with (ADK, LangGraph, CrewAI, etc.) or the vendor who created them, to communicate securely, exchange information, and coordinate actions. This directly tackles the significant challenge of siloed AI systems within enterprises, where agents built for different tasks or departments often cannot interact.

This push for an open A2A protocol represents a significant strategic gamble. Instead of building a proprietary, closed agent ecosystem, Google is attempting to establish the de facto standard for agent communication. This approach potentially sacrifices short-term lock-in for the prospect of long-term ecosystem leadership and, crucially, reducing the friction that hinders enterprise adoption of complex multi-agent systems.

By championing openness, Google aims to accelerate the entire agent market, positioning its cloud platform and tools as central facilitators.

How A2A works (Source: Google)

Recalibrating the Cloud Race: Google's Competitive Gambit

These announcements land squarely in the context of the ongoing cloud wars. Google Cloud, while demonstrating impressive growth often fueled by AI adoption, still holds the third position in market share, trailing AWS and Microsoft Azure. Cloud Next 2025 showcased Google's strategy to recalibrate this race by leaning heavily into its unique strengths and addressing perceived weaknesses.

Google's key differentiators were on full display. The long-term investment in custom silicon, culminating in the inference-focused Ironwood TPU, provides a distinct hardware narrative compared to AWS's Trainium/Inferentia chips and Azure's Maia accelerator. Google consistently emphasizes performance-per-watt leadership, a potentially crucial factor as AI energy demands soar. The launch of Cloud WAN weaponizes Google's unparalleled global network infrastructure, offering a distinct networking advantage.

Furthermore, Google continues to leverage its AI and machine learning heritage, stemming from DeepMind's research and manifested in the comprehensive Vertex AI platform, aligning with its market perception as a leader in AI and data analytics.

Simultaneously, Google signaled efforts to address historical enterprise concerns. The massive $32 billion acquisition of cloud security firm Wiz, announced shortly before Next, is a clear statement of intent to bolster its security posture and improve the usability and experience of its security offerings – areas critical for enterprise trust.

Continued emphasis on industry solutions, enterprise readiness, and strategic partnerships further aims to reshape market perception from a pure technology provider to a trusted enterprise partner. 

Taken together, Google's strategy appears less focused on matching AWS and Azure service-for-service across the board, and more concentrated on leveraging its unique assets – AI research, custom hardware, global network, and open-source affinity – to establish leadership in what it perceives as the next crucial wave of cloud computing: AI at scale, particularly efficient inference and sophisticated agentic systems. 

The Road Ahead for Google AI

Google Cloud Next 2025 presented a compelling narrative of ambition and strategic coherence. Google is doubling down on artificial intelligence, marshaling its resources across custom silicon optimized for the inference era (Ironwood), a balanced and practical AI model portfolio (Gemini 2.5 Pro and Flash), its unique global network infrastructure (Cloud WAN), and a bold, open approach to the burgeoning world of AI agents (ADK and A2A).

Ultimately, the event showcased a company moving aggressively to translate its deep technological capabilities into a comprehensive, differentiated enterprise offering for the AI era. The integrated strategy – hardware, software, networking, and open standards – is sound. Yet, the path ahead requires more than just innovation. 

Google's most significant challenge may lie less in technology and more in overcoming enterprise adoption inertia and building lasting trust. Converting these ambitious announcements into sustained market share gains against deeply entrenched competitors demands flawless execution, clear go-to-market strategies, and the ability to consistently convince large organizations that Google Cloud is the indispensable platform for their AI-driven future. The agentic future Google envisions is compelling, but its realization depends on navigating these complex market dynamics long after the Las Vegas spotlight has dimmed.

The post Google Cloud Next 2025: Doubling Down on AI with Silicon, Software, and an Open Agent Ecosystem appeared first on Unite.AI.

Fish AI Reader

Fish AI Reader

AI辅助创作,多种专业模板,深度分析,高质量内容生成。从观点提取到深度思考,FishAI为您提供全方位的创作支持。新版本引入自定义参数,让您的创作更加个性化和精准。

FishAI

FishAI

鱼阅,AI 时代的下一个智能信息助手,助你摆脱信息焦虑

联系邮箱 441953276@qq.com

相关标签

谷歌云 人工智能 企业AI Ironwood TPU Agent2Agent
相关文章