A recent report suggests Amazon Web Services (AWS) has placed a massive order with Nvidia for one million next-generation GPUs, set for delivery starting in 2026.
While this deal is still an unconfirmed report from Digitimes, its sheer scale is staggering. One million GPUs translates to nearly 14,000 server racks, consuming about 1.67 gigawatts of power—equivalent to several large-scale data center campuses. The hardware cost alone could be in the range of $40 to $47 billion. This isn't just a purchase; it's a foundational investment in the future of AI.
So, what's driving such a colossal move? This isn't about simply buying more chips. It's about building what the industry calls an 'AI Factory'. The reported deal isn't just for GPUs; it includes a full suite of Nvidia's networking hardware (ConnectX, Spectrum-X) and even specialized chips for AI inference from a company called Groq. This shows AWS is strategically procuring an entire, integrated system designed for peak AI performance.
This event didn't happen in a vacuum. We can trace the causes back through recent developments. First, Nvidia's GTC 2026 conference just weeks ago was crucial. They officially unveiled the Rubin platform, which integrates everything from GPUs to networking and even partner hardware like Groq's LPU. This announcement provided the technical blueprint for the very multi-chip architecture AWS is reportedly buying into.
Second, the competitive landscape is heating up. Reports of rivals like Meta planning to buy millions of GPUs created immense pressure on hyperscalers like AWS to secure their supply chain. In a market where demand far outstrips supply, locking in a long-term, large-volume deal is a critical strategic advantage to avoid being left behind.
And third, the demand is already secured. Late last year, AWS announced a $38 billion deal with OpenAI, which requires hundreds of thousands of Nvidia GPUs. This pre-existing commitment provides a solid justification for AWS to make such a large upstream order with Nvidia, as they already have a major customer waiting for that capacity.
In conclusion, this reported deal transforms the narrative from 'AWS buys GPUs' to 'AWS builds a comprehensive AI Factory.' It reflects a strategic race to secure not just chips, but entire ecosystems of computing, networking, and power to lead the next wave of artificial intelligence.
- Glossary -
- AI Factory: A term for a data center or computing system designed end-to-end specifically for developing, training, and deploying AI models at a massive scale.
- Hyperscaler: A large cloud services provider that can offer massive-scale computing resources. Key examples include AWS, Google Cloud, and Microsoft Azure.
- LPU (Language Processing Unit): A specialized processor made by Groq, designed to run AI language models with extremely low latency, making it ideal for real-time inference tasks.
