Google debuts two new custom chips in latest bid to challenge Nvidia's dominance

Dow Jones
04/22

MW Google debuts two new custom chips in latest bid to challenge Nvidia's dominance

By Britney Nguyen

The new TPUs offer cost advantages and improved storage functions

Google announced its eighth generation of TPUs on Wednesday.

One of Google's big artificial-intelligence advantages is its custom chip business, and the company is stepping up the competitiveness of its products in a potential challenge to Nvidia.

Google on Wednesday introduced the eighth generation of its tensor processing units, with two distinct chips focused on AI training and inference. The TPU 8t for training and TPU 8i for inference have been used to develop AI agents and to run massive inference workloads, Google said in a blog post. Inference is the process of running AI models.

Google has co-developed its TPUs with Broadcom $(AVGO)$ for more than a decade, and the chips have been used to train its Gemini AI models, as well as to power its YouTube and search engine algorithms. The newest chips were designed alongside Google DeepMind, which is Alphabet's $(GOOGL)$ $(GOOG)$ AI research lab.

See more: Broadcom's stock is rising. Here's why its new Google and Anthropic deals are so significant.

The current era of agentic AI, where an AI model can complete tasks with little to no human prompting, requires "a new set of demands on infrastructure," Google said. The new chips were designed to handle larger, more complex workloads, and to be able to adapt to the evolving capabilities of AI models, the company added.

There's been an increased focus on inference as adoption of AI applications grows. Nvidia (NVDA) has dominated the market for AI training and is working to improve its inference capabilities, in part through a nonexclusive licensing agreement with inference-chip maker Groq in December. But it remains to be seen whether the company will have its same advantages in the inference era.

"Several years ago, we anticipated rising demand for inference from customers as frontier AI models are deployed in production and at scale," Google said.

Because agentic AI works "in continuous loops of reasoning, planning, execution and learning," Google said it determined the industry would benefit from separate, specialized chips. Therefore, TPU 8t and TPU 8i were built specifically to support agentic AI, according to the company.

Google said it redesigned the stack for TPU 8i "to eliminate the 'waiting room' effect" during inference, which refers to delays in responses that can happen when AI model usage is high.

With the improvements, Google said the inference chip's performance-per-dollar compared to the previous Ironwood TPU is 80% better, meaning users can meet nearly twice the demand for the same cost.

The training TPU aims to cut down the time it takes to develop frontier AI models from months to weeks, Google said. Now, a single TPU 8t superpod, or chip cluster, can scale to 9,600 chips and two petabytes of high-bandwidth memory - double the interchip bandwidth of Ironwood. TPU 8t can also access storage faster, the company said.

The latest TPUs are "the latest expression of our co-design philosophy," Google said, meaning its hardware and software are designed to work together to optimize performance. As part of that, Google said both chips are running on its Axion central-processing-unit host for the first time, allowing for efficiency on a system level, rather than just on a chip level. A CPU host helps an AI processor stay fully utilized by helping to manage data and tasks.

Additionally, by integrating power management into its TPUs to adjust how much power is being drawn based on demand, Google said the new chips can provide up to two times better performance-per-watt than Ironwood.

Google said both TPUs will be generally available to its cloud customers later this year.

Unlike Nvidia, Google does not sell its TPUs to external customers, though AI startup Anthropic uses the chips.

Don't miss: Google's AI lead is growing in this key area. That's good news for Alphabet's stock.

That said, Google is still a significant Nvidia customer. The company also said Wednesday that it will be among the first cloud providers to offer Nvidia's upcoming Vera Rubin AI platform later this year.

-Britney Nguyen

This content was created by MarketWatch, which is operated by Dow Jones & Co. MarketWatch is published independently from Dow Jones Newswires and The Wall Street Journal.

 

(END) Dow Jones Newswires

April 22, 2026 08:07 ET (12:07 GMT)

Copyright (c) 2026 Dow Jones & Company, Inc.

应版权方要求,你需要登录查看该内容

免责声明:投资有风险,本文并非投资建议,以上内容不应被视为任何金融产品的购买或出售要约、建议或邀请,作者或其他用户的任何相关讨论、评论或帖子也不应被视为此类内容。本文仅供一般参考,不考虑您的个人投资目标、财务状况或需求。TTM对信息的准确性和完整性不承担任何责任或保证,投资者应自行研究并在投资前寻求专业建议。

热议股票

  1. 1
     
     
     
     
  2. 2
     
     
     
     
  3. 3
     
     
     
     
  4. 4
     
     
     
     
  5. 5
     
     
     
     
  6. 6
     
     
     
     
  7. 7
     
     
     
     
  8. 8
     
     
     
     
  9. 9
     
     
     
     
  10. 10