Meta secures millions of Amazon AI CPUs — the chip race just got more complex
In a move that underscores the accelerating fragmentation of the AI infrastructure landscape, Meta announced this week a strategic agreement to procure millions of Amazon Web Services' proprietary AI CPUs, specifically designed for agentic workloads. The deal, estimated at approximately $4.5 billion over three years according to internal sources cited by industry analysts, marks the first time a major hyperscaler has committed to such large-scale deployment of non-GPU accelerators for production AI systems.
The implications extend far beyond corporate balance sheets. Meta's decision to bet on Amazon's custom Trainium and Graviton processors instead of the NVIDIA GPU ecosystem that dominates today's AI training pipelines signals a fundamental shift in how the industry thinks about inference optimization, cost efficiency, and supply chain resilience.
How Amazon's CPUs differ from traditional AI accelerators
Unlike the NVIDIA H100 and H200 GPUs that have become synonymous with large language model training, Amazon's approach combines high-performance CPU compute with specialized AI inference accelerators. The Trainium2 chip, fabricated on TSMC's 3nm process, delivers approximately 425 TFLOPS of FP8 performance while consuming significantly less power than comparable GPU configurations.
What makes Meta's choice particularly strategic is the workload profile. Agentic AI systems — autonomous agents that reason, plan, and execute multi-step tasks — require sustained, low-latency compute that differs fundamentally from the massive parallel processing needed for foundation model training.
"The agentic paradigm changes everything. You don't need a supercomputer for 90% of the compute time. You need efficient, scalable inference at the edge of your infrastructure." — Dr. Ana Paula Silva, Director of AI Infrastructure at FGV (Fundação Getulio Vargas)
Amazon's architecture excels in horizontal scaling — adding more processing units efficiently as demand grows, rather than requiring ever-larger monolithic accelerators. Meta's infrastructure handles over 3.5 billion daily active users across its family of apps, making this scalability crucial.
Market implications: NVIDIA's dominance faces its first serious challenge
The global AI chip market was valued at $80.2 billion in 2024 and is projected to reach $317.9 billion by 2032, according to McKinsey estimates. NVIDIA currently commands approximately 80% of the data center AI accelerator market, but the Meta-Amazon deal represents the largest strategic defection yet from the CUDA ecosystem.
Three factors are driving this diversification:
- Cost optimization: Trainium2 chips reportedly offer 40-60% better cost-per-inference compared to H100 configurations for certain workloads
- Supply chain resilience: The 2023 GPU shortage, which extended lead times to 40+ weeks, exposed the risk of single-source dependency
- Custom silicon momentum: Google (TPU v5), Microsoft (Maia 100), and Meta itself (MTIA) are all investing in proprietary silicon
Impact on Latin America
For Latin American enterprises and governments, this shift carries significant weight. Cloud infrastructure costs directly impact the viability of local AI implementations. Major LATAM cloud providers including Telefónica's Genius, Totvs Cloud, and Mercado Libre's infrastructure arm are closely watching how hyperscaler pricing evolves.
Brazil's Ministry of Science and Technology has allocated R$ 23.7 billion ($4.8 billion) for AI infrastructure development through 2027, with chip procurement strategy under active review. Colombia and Chile have launched similar initiatives, collectively representing a $12 billion market opportunity through 2030.
"We can't afford to be naive about geopolitical supply chain risks. The Meta-Amazon deal shows that even the biggest players are hedging. LATAM should take notes." — Carlos Mendez, Senior Analyst, IDC Latin America
What to expect
The Meta-Amazon agreement sets several dynamics in motion:
- NVIDIA response: Expect accelerated roadmap announcements for next-generation Blackwell Ultra and future Rubin architecture, likely featuring enhanced inference optimization
- Hyperscaler silicon proliferation: Microsoft and Google will likely expand custom chip programs in response to Amazon's traction
- LATAM data center expansion: Amazon, Microsoft, and Google collectively announced $8.2 billion in LATAM data center investments for 2025-2027
- Regulatory attention: The EU's AI Act and potential LATAM frameworks may scrutinize chip supply agreements for competitive implications
The deal's true significance may not be immediately apparent. Meta isn't abandoning GPUs — it continues to purchase hundreds of thousands of H100s for training. Rather, this is a bet that inference infrastructure will be where the next battle is won, and that CPU-based AI acceleration offers advantages that GPU-centric architectures cannot match for agentic workloads.
As AI systems transition from reactive tools to proactive agents, the underlying hardware must evolve. Meta's $4.5 billion wager suggests the company believes Amazon's architecture is best positioned for that future.




