What happened
Google is actively promoting its custom-designed Tensor Processing Units (TPUs) to major enterprises, expanding their availability beyond Google's internal operations. The latest Trillium TPU offers a 4.7 times performance increase over its predecessor, with pods scaling to 256 units. These TPUs, which power Google's Gemini and other AI applications for over a billion users, are now being deployed in smaller cloud provider data centres. Google is also enhancing the TPU ecosystem through SparseCore improvements, increased HBM capacity, and better inter-chip interconnects, alongside its JAX AI Stack, a modular, production-ready platform. Nvidia currently holds approximately 80% of the AI accelerator market, largely due to its CUDA software.
Why it matters
The introduction of Google's TPUs into a broader market, particularly via smaller cloud providers, introduces a new dependency for organisations considering diverse AI infrastructure. Procurement and platform operators face increased due diligence requirements to evaluate the long-term viability and support ecosystem of non-Nvidia hardware, especially given Nvidia's established market share and CUDA software. This creates a potential control gap in standardising AI acceleration platforms, requiring IT architecture teams to assess compatibility and integration complexities with existing AI development workflows and toolchains.
Related Articles

Nvidia Faces AI Chip Challenge
Read more about Nvidia Faces AI Chip Challenge →
Google Challenges Nvidia's AI Dominance
Read more about Google Challenges Nvidia's AI Dominance →
China Dominates Open AI
Read more about China Dominates Open AI →
Google's AI Fuels Stock Surge
Read more about Google's AI Fuels Stock Surge →
