Amazon's AWS is sharpening its AI edge with custom chips—an upgraded Graviton4 CPU and a forthcoming Trainium3 GPU—that could start chipping away at Nvidia's (NVDA, Financial) market stronghold in AI training and inference.
CNBC reports AWS will soon launch a Graviton4 update boasting 600 Gbps of network bandwidth, courtesy of its Annapurna Labs design, with availability expected by month's end.
Later this year, AWS plans to roll out Trainium3, promising 50% better energy efficiency versus Trainium2, which underpins Anthropic's Claude Opus 4 model. While Nvidia's Blackwell GPU retains a raw–performance lead, Trainium2 already offers superior cost-performance ratios, according to AWS Senior Director Gadi Hutt.
Developers eyeing Trainium will need to retool workloads away from Nvidia's CUDA ecosystem and validate model–accuracy parity on AWS's frameworks.
Nvidia has dominated AI compute thanks to unmatched throughput and the ubiquity of CUDA in developer toolchains.
By delivering strong price-performance and energy gains via Graviton4 and Trainium3, AWS aims to lure hyperscalers and cost-sensitive enterprises that run massive inference fleets or large-scale training jobs.
If AWS can minimize migration friction and prove equivalent accuracy, it could open the door for a meaningful shift in AI infrastructure spend.
The real test will come when Graviton4 benchmarks are published and Trainium3 previews hit developer hands.
Watch for cloud–native AI workloads running on non-CUDA stacks and for enterprise case studies highlighting total cost-of-ownership savings. Those signals will reveal whether AWS can genuinely erode Nvidia's GPU hegemony.