Skip to main content
Apr 22

Google Cloud Launches Dual AI Chips to Rival Nvidia

Google Cloud recently unveiled its eighth generation of proprietary AI chips, known as Tensor Processing Units (TPUs), which will now feature a dual-p

2 min read67 views3 tags
Originally reported bytechcrunch

Google Cloud recently unveiled its eighth generation of proprietary AI chips, known as Tensor Processing Units (TPUs), which will now feature a dual-purpose architecture. The TPU 8t is specifically engineered for AI model training, while the TPU 8i is designed for inference tasks.

Inference refers to the ongoing application of AI models, encompassing the processes that occur subsequent to user prompt submissions.

Predictably, Google Cloud highlights significant performance enhancements for these new TPUs compared to their predecessors. These include up to three times faster AI model training, an 80% improvement in performance per dollar, and the capability to integrate over one million TPUs within a single cluster. This translates to substantially more computational power with reduced energy consumption and lower costs for customers. The designation "TPU" (Tensor Processing Unit) rather than "GPU" stems from the original naming of Google's custom low-power chips as "Tensor."

However, Google's introduction of these chips does not signify a direct, comprehensive challenge to Nvidia's market position, at least for the present. Similar to other major cloud providers such as Microsoft and Amazon, Google is integrating these custom chips to augment its existing infrastructure, which largely relies on Nvidia-based systems, rather than outright replacing them. Indeed, Google has committed to making Nvidia's latest chip, Vera Rubin, available within its cloud offerings later this year.

It is conceivable that hyperscale cloud providers, including Amazon, Microsoft, and Google, who are developing their own AI chips, may eventually reduce their reliance on Nvidia. This shift could occur as enterprises increasingly migrate their AI workloads to these cloud platforms and adapt their applications to run on these proprietary chips.

Nevertheless, in the current landscape, betting against Nvidia's continued dominance proves unprofitable. As prominent chip market analyst Patrick Moore humorously noted on X, his 2016 prediction that Google's initial TPU launch could pose a threat to Nvidia (and Intel) has not materialized, given Nvidia's current market capitalization approaching $5 trillion.

From Nvidia's strategic perspective, Google's expansion as an AI cloud provider is anticipated to generate increased business for the chip manufacturer, rather than a reduction, even as a significant portion of workloads are processed on Google's proprietary chips.

Furthermore, Google has announced a collaborative agreement with Nvidia to engineer advanced computer networking solutions designed to enhance the efficiency of Nvidia-based systems within its cloud environment. Specifically, both technology leaders are focused on strengthening Falcon, a software-defined networking technology developed and open-sourced by Google in 2023 under the auspices of the Open Compute Project, a leading organization for open-source data center hardware.

ES
Editorial StaffEditor

The Editorial Staff at AIChief is a team of professional content writers with extensive experience in AI and marketing. Founded in 2025, AIChief has quickly grown into the largest free AI resource hub in the industry.

View all posts
Reader feedback

What did you think of this story?

User Comments

Filter:
No comments yet. Be the first to comment!
Continue reading
View all news