Nvidia outperformed all expectations, reporting hovering income Wednesday because of its graphics processing items that excel at AI workloads. However extra classes of AI chips are gaining floor.
Customized ASICs, or application-specific built-in circuits, are actually being designed by all the main hyperscalers, from Google’s TPU to Amazon’s Trainium and OpenAI’s plans with Broadcom. These chips are smaller, cheaper, accessible and will cut back these corporations’ reliance on Nvidia GPUs. Daniel Newman of the Futurum Group informed CNBC that he sees customized ASICs “rising even quicker than the GPU market over the subsequent few years.”
Moreover GPUs and ASICs, there are additionally field-programmable gate arrays, which may be reconfigured with software program after they’re made to be used in all types of purposes, like sign processing, networking and AI. There’s additionally a whole group of AI chips that energy AI on gadgets somewhat than within the cloud. Qualcomm, Apple and others have championed these on-device AI chips.
CNBC talked to consultants and insiders on the Large Tech corporations to interrupt down the crowded house and the varied sorts of AI chips on the market.
GPUs for basic compute
As soon as used primarily for gaming, GPUs made Nvidia the world’s most beneficial public firm after their use shifted towards AI workloads. Nvidia shipped some 6 million current-generation Blackwell GPUs over the previous yr.
Nvidia senior director of AI infrastructure Dion Harris exhibits CNBC’s Katie Tarasov how 72 Blackwell GPUs work collectively as one in a GB200 NVL72 rack-scale server system for AI at Nvidia headquarters in Santa Clara, California, on November 12, 2025.
Marc Ganley
The shift from gaming to AI began round 2012, when Nvidia’s GPUs had been utilized by researchers to construct AlexNet, what many think about to be fashionable AI’s huge bang second. AlexNet was a device that was entered right into a distinguished picture recognition contest. Whereas others within the contest used central processing items for his or her purposes, AlexNet reliance on GPUs offered unbelievable accuracy and obliterated its competitors.
AlexNet’s creators found that the identical parallel processing that helps GPUs render lifelike graphics was additionally nice for coaching neural networks, during which a pc learns from information somewhat than counting on a programmer’s code. AlexNet showcased the potential of GPUs.
At the moment, GPUs are sometimes paired with CPUs and bought in server rack programs to be positioned in information facilities, the place they run AI workloads within the cloud. CPUs have a small variety of highly effective cores working sequential general-purpose duties, whereas GPUs have 1000’s of smaller cores extra narrowly targeted on parallel math like matrix multiplication.
As a result of GPUs can carry out many operations concurrently, they’re splendid for the 2 primary phases of AI computation: coaching and inference. Coaching teaches the AI mannequin to study from patterns in massive quantities of knowledge, whereas inference makes use of the AI to make selections primarily based on new data.
GPUs are the general-purpose workhorses of Nvidia and its prime competitor, Superior Micro Units. Software program is a significant differentiator between the 2 GPU leaders. Whereas Nvidia GPUs are tightly optimized round CUDA, Nvidia’s proprietary software program platform, AMD GPUs use a largely open-source software program ecosystem.
AMD and Nvidia promote their GPUs to cloud suppliers like Amazon, Microsoft, Google, Oracle and CoreWeave. These corporations then hire the GPUs to AI corporations by the hour or minute. Anthropic’s $30 billion deal with Nvidia and Microsoft, for instance, contains 1 gigawatt of compute capability on Nvidia GPUs. AMD has additionally lately landed huge commitments from OpenAI and Oracle.
Nvidia additionally sells on to AI corporations, like a current deal to promote no less than 4 million GPUs to OpenAI, and to international governments, together with South Korea, Saudi Arabia and the U.Ok.
The chipmaker informed CNBC that it costs round $3 million for one in all its server racks with 72 Blackwell GPUs appearing as one, and ships about 1,000 every week.
Dion Harris, Nvidia’s senior director of AI infrastructure, informed CNBC he could not have imagined this a lot demand when he joined Nvidia over eight years in the past.
“After we had been speaking to folks about constructing a system that had eight GPUs, they thought that was overkill,” he mentioned.
ASICs for customized cloud AI
Coaching on GPUs has been key within the early increase days of enormous language fashions, however inference is changing into extra essential because the fashions mature. Inference can occur on much less highly effective chips which are programmed for extra particular duties. That is the place ASICs are available in.
Whereas a GPU is sort of a Swiss Military Knife in a position to do many sorts of parallel math for various AI workloads, an ASIC is sort of a single-purpose device. It is very environment friendly and quick, however hard-wired to do the precise math for one kind of job.
Google launched its seventh era TPU, Ironwood, in November 2025, a decade after making its first customized ASIC for AI in 2015.
“You possibly can’t change them as soon as they’re already carved into silicon, and so there is a commerce off when it comes to flexibility,” mentioned Chris Miller, writer of “Chip Conflict.”
Nvidia’s GPUs are versatile sufficient for adoption by many AI corporations, however they price as much as $40,000 and may be onerous to get. Nonetheless, startups depend on GPUs as a result of designing a customized ASIC has a fair greater up-front price, beginning at tens of hundreds of thousands of {dollars}, in response to Miller.
For the largest cloud suppliers who can afford them, analysts say customized ASICs repay within the long-run.
“They need to have just a little bit extra management over the workloads that they construct,” Newsom mentioned. “On the identical time, they are going to proceed to work very intently with Nvidia, with AMD, as a result of additionally they want the capability. The demand is so insatiable.”
Google was the primary Large Tech firm to make a customized ASIC for AI acceleration, coining the time period Tensor Processing Unit when its first ASIC got here out in 2015. Google mentioned it thought-about making a TPU way back to 2006, however the state of affairs grew to become “pressing” in 2013 because it realized AI was going to double its variety of information facilities. In 2017, the TPU additionally contributed to Google’s invention of the Transformer, the structure powering virtually all fashionable AI.
A decade after its first TPU, Google launched its seventh era TPU in November. Anthropic introduced it is going to practice its LLM Claude on as much as 1 million TPUs. Some folks assume TPUs are technically on par or superior to Nvidia’s GPUs, Miller mentioned.
“Historically, Google has solely used them for in-house functions,” Miller mentioned. “There’s quite a lot of hypothesis that within the longer run, Google would possibly open up entry to TPUs extra broadly.”
Amazon Net Providers was the subsequent cloud supplier to design its personal AI chips, after buying Israeli chip startup Annapurna labs in 2015. AWS introduced Inferentia in 2018, and it launched Trainium in 2022. AWS is anticipated to announce Trainium’s third era as quickly December.
Ron Diamant, Trainium’s head architect, informed CNBC that Amazon’s ASIC has 30% to 40% higher value efficiency in comparison with different {hardware} distributors in AWS.
“Over time, we have seen that Trainium chips can serve each inference and coaching workloads fairly properly,” Diamant mentioned.
CNBC’s Katie Tarasov holds Amazon Net Providers’ Trainium 2 AI chip that fill its new AI information heart in New Carlisle, Indiana, on October 8, 2025.
Erin Black
In October, CNBC went to Indiana for the first on-camera tour of Amazon’s greatest AI information heart, the place Anthropic is coaching its fashions on half one million Trainium2 chips. AWS fills its different information facilities with Nvidia GPUs to satisfy the demand from AI prospects like OpenAI.
Constructing ASICs is not simple. This is the reason corporations flip to chip designers Broadcom and Marvell. They “present the IP and the know-how and the networking” to assist their shoppers construct their ASICs, Miller mentioned.
“So you’ve got seen Broadcom specifically be one of many greatest beneficiaries of the AI increase,” Miller mentioned.
Broadcom helped construct Google’s TPUs and Meta‘s Coaching and Inference Accelerator launched in 2023, and has a brand new deal to assist OpenAI construct its personal customized ASICs beginning in 2026.
Microsoft can also be stepping into the ASIC recreation, telling CNBC that its in-house Maia 100 chips are at the moment deployed in its information facilities within the jap U.S. Others embody Qualcomm with the A1200, Intel with its Gaudi AI accelerators and Tesla with its AI5 chip. There’s additionally a slew of start-ups going all in on customized AI chips, together with Cerebras, which makes enormous full-wafer AI chips, and Groq, with inference-focused language processing items.
In China, Huawei, ByteDance, and Alibaba are making customized ASICs, though export controls on essentially the most superior gear and AI chips pose a problem.
Edge AI with NPUs and FPGAs
The ultimate huge class of AI chips are these made to run on gadgets, somewhat than within the cloud. These chips are sometimes constructed into a tool’s primary System on a Chip, SoC. Edge AI chips, as they’re known as, allow gadgets to have AI capabilities whereas serving to them save battery life and house for different elements.
“You’ll try this proper in your telephone with very low latency, so you do not have to have communication all the best way again to a knowledge heart,” mentioned Saif Khan, former White Home AI and semiconductor coverage advisor. “And you’ll protect privateness of your information in your telephone.”
Neural processing items are a significant kind of edge AI chip. Qualcomm, Intel and AMD are making NPUs that allow AI capabilities in private computer systems.
Though Apple would not use the time period NPU, the in-house M-series chips inside its MacBooks embody a devoted neural engine. Apple additionally constructed neural accelerators into the most recent iPhone A-series chips.
“It’s environment friendly for us. It’s responsive. We all know that we’re way more in management over the expertise,” Tim Millet, Apple platform structure vice chairman, informed CNBC in an unique September interview.
The newest Android telephones even have NPUs constructed into their major Qualcomm Snapdragon chips, and Samsung has its personal NPU on its Galaxy telephones, too. NPUs by corporations like NXP and Nvidia energy AI embedded in automobiles, robots, cameras, good dwelling gadgets and extra.
“A lot of the {dollars} are going in the direction of the information heart, however over time that is going to vary as a result of we’ll have AI deployed in our telephones and our automobiles and wearables, all types of different purposes to a a lot better diploma than in the present day,” Miller mentioned.
Then there’s field-programmable gate arrays, or FPGAs, which may be reconfigured with software program after they’re made. Though much more versatile than NPUs or ASICs, FPGAs have decrease uncooked efficiency and decrease vitality effectivity for AI workloads.
AMD grew to become the biggest FPGA maker after buying Xilinx for $49 billion in 2022, with Intel in second because of its $16.7 billion buy of Altera in 2015.
These gamers designing AI chips depend on a single firm to fabricate all of them: Taiwan Semiconductor Manufacturing Firm.
TSMC has an enormous new chip fabrication plant in Arizona, the place Apple has dedicated to transferring some chip manufacturing. In October, Nvidia CEO Jensen Huang mentioned Blackwell GPUs had been in “full manufacturing” in Arizona, too.
Though the AI chip house is crowded, dethroning Nvidia will not come simply.
“They’ve that place as a result of they’ve earned it and so they’ve spent the years constructing it,” Newman mentioned. “They’ve received that developer ecosystem.”
Watch the video to see a breakdown of how all of the AI chips work: https://www.cnbc.com/video/2025/11/21/nvidia-gpus-google-tpus-aws-trainium-comparing-the-top-ai-chips.html

