August 3rd, 2024

TPU transformation: A look back at 10 years of our AI-specialized chips

Google has advanced its AI capabilities with Tensor Processing Units (TPUs), specialized chips for AI workloads, enhancing performance and efficiency, and making them available through Cloud services for external developers.

Read original articleLink Icon
TPU transformation: A look back at 10 years of our AI-specialized chips

Over the past decade, Google has significantly advanced its AI capabilities through the development of Tensor Processing Units (TPUs), specialized chips designed for AI workloads. The need for these chips arose when Google's AI compute demand began to exceed existing infrastructure, particularly for speech recognition features. The first TPU, launched in 2015, was a success, leading to the production of over 100,000 units to support various applications, including Ads and self-driving technology.

TPUs are application-specific integrated circuits (ASICs) optimized for the matrix and vector mathematics essential for AI model training and inference. The latest generation, Trillium, offers over 4.7 times the compute performance of its predecessor, TPU v5e, enabling the training of advanced AI models like Gemini 1.5.

The evolution of TPUs has paralleled advancements in machine learning, with each generation focusing on improving performance and efficiency. TPU v2 introduced a supercomputing approach, while subsequent versions incorporated innovations like liquid cooling and optical circuit switches to enhance data processing capabilities.

Google has also made TPUs available through its Cloud services, allowing external developers and companies to leverage this technology for their AI projects. Today, a significant portion of generative AI startups utilize Google Cloud's AI infrastructure, indicating the widespread impact of TPUs in the AI landscape. Looking ahead, Google plans to further customize its data center designs to optimize TPU deployment, signaling ongoing innovation in AI infrastructure.

Link Icon 6 comments
By @nl - 2 months
It's crazy that Google doesn't spin-out their TPU work as a separate company.

TPUs are the second most widely used environment for training after Nvidia. It's the only environment that people build optimized kernels for outside CUDA.

If it was separate to Google then there a bunch of companies who would happily spend some money on a real, working NVidia alternative.

It might be profitable from day one, and it surely would gain substantial market capitalization - Alphabet shareholders should be agitating for this!

By @ec109685 - 2 months
Impressive: “Overall, more than 60% of funded generative AI startups and nearly 90% of gen AI unicorns use Google Cloud’s AI infrastructure, including Cloud TPUs.”
By @walterbell - 2 months
Apple Intelligence uses Google TPUs instead of GPUs.
By @alecco - 2 months
How are they connected? PCIe? Something like NVLink?
By @PedroBatista - 2 months
The real winner here is the marketing department who manage to make this article a "celebration of successes" when in fact we know the TPU is yet one more of those biggest failures of Google to have the lead by a mile and then.. squander it. And no, "it's on our cloud and Pixel phones" doesn't cut it at this level.