AWS Is stepping on the accelerator for AI infrastructure
The latest datapoint regarding how the hyperscalers are feeling about AI’s potential came this week at AWS re:Invent 2025. AWS CEO Matt Garman put a stake in the ground that we are still early in AI.
The big question going into re:Invent was whether he would signal any cooling in AI demand or a pause in spending. Matt Garman answered that at the top of his keynote:
“Demand keeps skyrocketing. We’ve added 3.8 gigawatts of new data center capacity in the last 12 months alone. And we’re not slowing anything down. We’re only speeding that up.”
That sets the tone for the broader AI infrastructure spending outlook. AWS is spending on both Nvidia and its own silicon. Garman highlighted his long running message that AWS is the best place to run Nvidia GPUs.
Alongside Nvidia, AWS is pushing hard on its own Trainium chip. The latest version, Trainium 3 UltraServers, has more than 4x the compute compared to the prior generation, with their internal Project Rainier already using hundreds of thousands of Trainium chips to run Anthropic’s Claude models. AWS also teased Trainium 4, which underscores the company is determined to quickly rev the product.
For some context, Amazon’s Trainium chips are AWS’s in house alternative to Nvidia, designed specifically for AI training and inference in its data centers. In practice, most large AI customers use Nvidia GPUs when they want maximum performance and compatibility with the broader AI ecosystem, and consider using Trainium when they want to lower costs.
On top of raw compute, AWS announced AI Factories. These let large customers deploy dedicated AWS AI racks inside their own data centers, effectively creating a private mini region that still runs AWS hardware and software. For investors and enterprises, the message is simple, AWS sees AI infrastructure as a long runway and is building as if demand will stay strong.
