Meta has signed a major agreement with Amazon Web Services to deploy millions of Amazon’s custom AI CPUs, according to TechCrunch AI, marking one of the most significant partnerships between Big Tech rivals in the AI infrastructure race. The deal represents a strategic departure from GPU-centric approaches that have dominated AI computing since the generative AI boom began.
The agreement centres on Amazon’s Graviton processors optimised for AI inference workloads, particularly the agentic AI systems that require sustained reasoning rather than the burst training performance that GPUs excel at. Meta will integrate these chips across its infrastructure to support AI agents capable of multi-step task execution, a workload profile that differs substantially from the large language model training that has driven unprecedented demand for Nvidia’s H100 and H200 GPUs.
The partnership arrives as enterprises confront mounting costs for GPU-based infrastructure. Whilst training frontier models remains GPU-dependent, the operational phase—where AI systems serve billions of user requests—presents different economic calculus. CPUs designed for AI inference offer lower power consumption and cost per query, critical factors as companies scale from experimental deployments to production systems handling sustained traffic.
“This signals recognition that the AI chip market is bifurcating,” said one semiconductor analyst familiar with the deal. “Training and inference have different optimal architectures, and we’re seeing infrastructure strategies reflect that reality.”
For Amazon, the agreement validates years of investment in custom silicon development through its Annapurna Labs division. AWS has positioned Graviton chips as cost-effective alternatives to traditional x86 processors, claiming up to 40 per cent better price-performance for certain workloads. Securing Meta as a customer provides crucial validation and scale economics that could accelerate AWS’s challenge to established chip vendors.
Meta gains supply chain diversification at a moment when GPU scarcity remains acute despite easing from 2024 peaks. The company has committed to building AI infrastructure supporting hundreds of millions of users across Facebook, Instagram, and WhatsApp, with CEO Mark Zuckerberg publicly stating the company would spend over $60 billion on capital expenditure in 2025, much of it AI-related. Reducing dependence on any single chip vendor strengthens Meta’s negotiating position and operational resilience.
The deal carries implications for Nvidia, whose dominance in AI accelerators has driven its valuation past $2 trillion. Whilst training workloads remain firmly in GPU territory, the emergence of viable CPU alternatives for inference could constrain Nvidia’s total addressable market as AI deployments mature. The company has responded by developing inference-optimised products, but faces intensifying competition from custom silicon efforts at Amazon, Google, and Microsoft.
Intel and AMD also face pressure as hyperscalers increasingly design their own chips. Intel’s struggles to capture AI market share despite decades of data centre CPU dominance illustrate how quickly specialised workloads can upend established positions. AMD’s MI300 series has gained traction as a GPU alternative, but custom silicon from cloud providers represents a different competitive threat—one that targets cost structure rather than raw performance.
The technical bet centres on agentic AI architectures that chain multiple model calls together, maintaining context across interactions whilst executing complex tasks. These workloads involve more memory bandwidth and sustained computation than the parallel processing bursts that suit GPU architectures. If agentic AI becomes the dominant paradigm—as many in the industry expect—CPU-based inference infrastructure could claim significant market share.
Industry observers will watch whether other Meta-scale AI deployers follow suit. Google already uses its own TPUs extensively, whilst Microsoft has announced custom AI chips. If the hyperscaler pattern of vertical integration extends to major AI-native companies like OpenAI or Anthropic, the chip market could fragment significantly from today’s Nvidia-centric concentration.
The deal underscores how AI infrastructure competition is evolving beyond the initial GPU land grab into more nuanced architectural choices. As workloads diversify and economic pressures mount, the chip landscape appears set for greater heterogeneity than the training-focused era suggested.













