When you purchase through links on our site, we may earn an affiliate commission. Here’s how it works.

Google takes on Nvidia Blackwell GPUs with the new Trillium TPUs

Google Trillium TPUs

Google started developing custom AI accelerators in the name of Tensor Processing Units (TPUs) a decade ago. Early this year, Google announced Trillium, its sixth-generation TPU, that delivers better performance with efficiency than its predecessors. Today, Google announced the general availability of Trillium TPUs for Google Cloud customers. Google today also revealed that it used Trillium TPUs to train the new Gemini 2.0.

Nvidia's GPUs are incredibly popular among developers for AI workloads, not just because of their hardware capabilities but also because of their software support. To make Trillium TPUs popular among AI developers, Google has made several improvements to its software layer. It has optimized the XLA compiler and AI frameworks such as JAX, PyTorch, and TensorFlow so that developers can achieve price-performance across AI training, tuning, and serving.

Compared to the previous generation TPU, Trillium offers the following improvements:

  • Over 4x improvement in training performance
  • Up to 3x increase in inference throughput
  • A 67% increase in energy efficiency
  • An impressive 4.7x increase in peak compute performance per chip
  • Double the High Bandwidth Memory (HBM) capacity
  • Double the Interchip Interconnect (ICI) bandwidth
  • 100K Trillium chips in a single Jupiter network fabric
  • Up to 2.5x improvement in training performance per dollar and up to 1.4x improvement in inference performance per dollar

Google also claimed that Trillium TPUs can achieve 99% scaling efficiency with a deployment of 12 pods consisting of 3072 chips and can achieve 94% scaling efficiency across 24 pods with 6144 chips to pre-train gpt3-175b.

Trillium is now available in North America (US East region), Europe (West region), and Asia (Northeast region). For evaluation, Trillium is available starting at $2.7000 per chip-hour. With 1-year and 3-year commitments, it is available starting at $1.8900 per chip-hour and $1.2200 per chip-hour, respectively.

With its ability to scale to hundreds of thousands of chips and improved software support, Trillium represents a significant leap forward for Google in the cloud AI infrastructure market.

Report a problem with article
YouTube On TV Recap 2024
Next Article

YouTube on TV Recap 2024 reveals people stream a billion hours of content daily

Google Gemini 20 Flash
Previous Article

Google announces Gemini 2.0 Flash with native image and audio output

Join the conversation!

Login or Sign Up to read and post a comment.

0 Comments - Add comment