Google debuts two new custom chips in latest bid to challenge Nvidia's dominance

Dow Jones04-22 20:07

MW Google debuts two new custom chips in latest bid to challenge Nvidia's dominance

By Britney Nguyen

The new TPUs offer cost advantages and improved storage functions

Google announced its eighth generation of TPUs on Wednesday.

One of Google's big artificial-intelligence advantages is its custom chip business, and the company is stepping up the competitiveness of its products in a potential challenge to Nvidia.

Google on Wednesday introduced the eighth generation of its tensor processing units, with two distinct chips focused on AI training and inference. The TPU 8t for training and TPU 8i for inference have been used to develop AI agents and to run massive inference workloads, Google said in a blog post. Inference is the process of running AI models.

Google has co-developed its TPUs with Broadcom $(AVGO)$ for more than a decade, and the chips have been used to train its Gemini AI models, as well as to power its YouTube and search engine algorithms. The newest chips were designed alongside Google DeepMind, which is Alphabet's $(GOOGL)$ $(GOOG)$ AI research lab.

See more: Broadcom's stock is rising. Here's why its new Google and Anthropic deals are so significant.

The current era of agentic AI, where an AI model can complete tasks with little to no human prompting, requires "a new set of demands on infrastructure," Google said. The new chips were designed to handle larger, more complex workloads, and to be able to adapt to the evolving capabilities of AI models, the company added.

There's been an increased focus on inference as adoption of AI applications grows. Nvidia (NVDA) has dominated the market for AI training and is working to improve its inference capabilities, in part through a nonexclusive licensing agreement with inference-chip maker Groq in December. But it remains to be seen whether the company will have its same advantages in the inference era.

"Several years ago, we anticipated rising demand for inference from customers as frontier AI models are deployed in production and at scale," Google said.

Because agentic AI works "in continuous loops of reasoning, planning, execution and learning," Google said it determined the industry would benefit from separate, specialized chips. Therefore, TPU 8t and TPU 8i were built specifically to support agentic AI, according to the company.

Google said it redesigned the stack for TPU 8i "to eliminate the 'waiting room' effect" during inference, which refers to delays in responses that can happen when AI model usage is high.

With the improvements, Google said the inference chip's performance-per-dollar compared to the previous Ironwood TPU is 80% better, meaning users can meet nearly twice the demand for the same cost.

The training TPU aims to cut down the time it takes to develop frontier AI models from months to weeks, Google said. Now, a single TPU 8t superpod, or chip cluster, can scale to 9,600 chips and two petabytes of high-bandwidth memory - double the interchip bandwidth of Ironwood. TPU 8t can also access storage faster, the company said.

The latest TPUs are "the latest expression of our co-design philosophy," Google said, meaning its hardware and software are designed to work together to optimize performance. As part of that, Google said both chips are running on its Axion central-processing-unit host for the first time, allowing for efficiency on a system level, rather than just on a chip level. A CPU host helps an AI processor stay fully utilized by helping to manage data and tasks.

Additionally, by integrating power management into its TPUs to adjust how much power is being drawn based on demand, Google said the new chips can provide up to two times better performance-per-watt than Ironwood.

Google said both TPUs will be generally available to its cloud customers later this year.

Unlike Nvidia, Google does not sell its TPUs to external customers, though AI startup Anthropic uses the chips.

Don't miss: Google's AI lead is growing in this key area. That's good news for Alphabet's stock.

That said, Google is still a significant Nvidia customer. The company also said Wednesday that it will be among the first cloud providers to offer Nvidia's upcoming Vera Rubin AI platform later this year.

-Britney Nguyen

This content was created by MarketWatch, which is operated by Dow Jones & Co. MarketWatch is published independently from Dow Jones Newswires and The Wall Street Journal.

 

(END) Dow Jones Newswires

April 22, 2026 08:07 ET (12:07 GMT)

Copyright (c) 2026 Dow Jones & Company, Inc.

At the request of the copyright holder, you need to log in to view this content

Disclaimer: Investing carries risk. This is not financial advice. The above content should not be regarded as an offer, recommendation, or solicitation on acquiring or disposing of any financial products, any associated discussions, comments, or posts by author or other users should not be considered as such either. It is solely for general information purpose only, which does not consider your own investment objectives, financial situations or needs. TTM assumes no responsibility or warranty for the accuracy and completeness of the information, investors should do their own research and may seek professional advice before investing.

Comments

We need your insight to fill this gap
Leave a comment