TPUs: Google's home advantage
How does TPU v7 stack up against Nvidia's latest chips – and can Google scale AI using only its own supply?
In the race to train and deploy generative AI models, companies have poured hundreds of billions of dollars into GPUs, chips that have become essential for the parallel processing needs of large language models.
Nvidia alone has forecast $500 billion in sales across 2025 and 2026, driven largely by Jensen Huang, founder and CEO at Nvidia, recently stated that “inference has become the most compute-intensive phase of AI — demanding real-time reasoning at planetary scale”.
Google is meeting these demands in its own way. Unlike other firms reliant on chips by Nvidia, AMD, and others, Google has long used its in-house ‘tensor processing units’ (TPUs) for AI training and inference.
What are the benefits and drawbacks of Google’s reliance on TPUs? And how do its chips stack up against the competition?
In this episode, Jane and Rory discuss TPUs – Google’s specialized processors for AI and ML – and how they could help the hyperscaler outcompete its rivals.
Highlights
“So one of the main advantages of TPUs for Google is that it designs them, it has them manufactured through TSMC, which means it has sole dibs on them. This avoids a lot of the supply chain bottlenecks we've seen in recent years where companies have been queuing up for Nvidia chips. Nation states have been queuing up for Nvidia chips, or they've been forced to buy them in massive quantities up front and then slowly figure out where they're going to deploy them all.”
“Nvidia really is setting the standard on enterprise AI hardware. Like you said, Intel, AMD are also major players in the space and they have their chunk of the market, but Nvidia is the one that the major AI developers come back to again and again. It's making the hundred billion dollar announcements, all of this investment that seems to be announced every month.”
Sign up today and you will receive a free copy of our Future Focus 2025 report - the leading guidance on AI, cybersecurity and other IT challenges as per 700+ senior executives
“It's a trade off for Google, because what you're doing is throwing a lot more compute at one workload, but potentially you're sacrificing some latency there. So all of that is to say that it's not as simple as just bigger numbers are better here. It really depends on the workload. But it's clear that, particularly with the latest iteration of TPUs, Google is seriously contending with the raw performance that companies like Nvidia can offer.”
“I'd say at this point it's clear that we haven't quite hit the wall that maybe some people were predicting earlier in the year, where people were saying it's incremental returns, if you just increase the size of the model – and maybe we do need to go back to the drawing board. I think Gemini 3 Pro shows that, at least for now, there is some life in just scaling things and seeing how much performance we can squeeze out of them.”
Footnotes
- What is a tensor processing unit (TPU)?
- ‘TPUs just work’: Why Google Cloud is betting big on its custom chips
- Google boasts that a single Gemini prompt uses roughly the same energy as a basic search – but that’s not painting the full picture
- Google Cloud announces major computing boost with Ironwood chip, new hypercomputer upgrades
- Google claims its AI chips are ‘faster, greener’ than Nvidia’s
Subscribe
- Subscribe to The IT Pro Podcast on Apple Podcasts
- Subscribe to The IT Pro Podcast on Spotify
- Subscribe to the IT Pro newsletter
- Join us on LinkedIn

Rory Bathgate is Features and Multimedia Editor at ITPro, overseeing all in-depth content and case studies. He can also be found co-hosting the ITPro Podcast with Jane McCallion, swapping a keyboard for a microphone to discuss the latest learnings with thought leaders from across the tech sector.
In his free time, Rory enjoys photography, video editing, and good science fiction. After graduating from the University of Kent with a BA in English and American Literature, Rory undertook an MA in Eighteenth-Century Studies at King’s College London. He joined ITPro in 2022 as a graduate, following four years in student journalism. You can contact Rory at rory.bathgate@futurenet.com or on LinkedIn.
-
Microsoft Excel is still alive and kicking at 40News A recent survey found Gen Z and Millennial finance professionals have a strong “emotional attachment” to Microsoft Excel
-
LastPass hit with ICO fine after 2022 data breach exposed 1.6 million usersNews The impact of the LastPass breach was felt by customers as late as December 2024
-
On the ground at HPE Discover Barcelona 2025ITPro Podcast This is a pivotal time for HPE, as it heralds its Juniper Networks acquisition and strengthens ties with Nvidia and AMD
-
What is a tensor processing unit (TPU)?Explainer Google's in-house AI chips are the most notable alternative to Nvidia at the enterprise scale
-
Inside a cloud outageITPro Podcast Businesses must adopt proactive planning for cloud outages – but what does that look like?
-
October rundown: AWS chaos and supercomputers surgingITPro Podcast As the dust settled on the AWS outage, the US Department of Energy announced a slew of new supercomputers for national security
-
Is all-photonics the future of networking?ITPro Podcast Using light to transmit data rather than relying on electronic components could slash latency
-
Future-proofing AI infrastructureSponsored Podcast Constructing the future of the tech sector can only be done with a strategic approach and access to the best tools
-
‘TPUs just work’: Why Google Cloud is betting big on its custom chipsNews As AI inference skyrockets, Google Cloud wants customers to choose it as the go-to partner to meet demand
-
Google shakes off tariff concerns to push on with $75 billion AI spending plans – but analysts warn rising infrastructure costs will send cloud prices sky highNews Google CEO Sundar Pichai has confirmed the company will still spend $75 billion on building out data centers despite economic concerns in the wake of US tariffs.