Meta has announced a landmark multibillion-dollar agreement with Amazon Web Services (AWS) to deploy tens of millions of Amazon’s latest Graviton5 ARM CPU cores across its data centers. The move is aimed at supporting the growing demand for AI compute power, particularly for agentic AI workloads that require significant CPU-intensive processing for tasks like real-time reasoning and multi-step decision-making.
Graviton5 Chips for the Next Generation of AI
Unlike traditional AI accelerators such as GPUs or TPUs, the Graviton5 chips are general-purpose processors designed to handle the complex orchestration and inference tasks that underpin agentic AI systems. These systems are capable of executing multiple steps in response to user input, requiring substantial CPU power for reasoning and task management. By leveraging AWS’s scalable infrastructure and Graviton5 technology, Meta aims to enhance its AI capabilities and reduce latency in real-time applications.
Strategic Move Amid Rising AI Demand
This procurement deal reflects the broader industry trend of companies racing to secure compute resources for AI development. Meta’s decision comes as its AI research and deployment efforts expand, particularly in the area of agentic AI, which is increasingly seen as a critical frontier in the next generation of AI applications. The $135 billion capex budget that Meta previously set for AI infrastructure is now being stretched to meet the surge in demand, making strategic deals like this one essential for maintaining competitive edge.
Implications for the AI Ecosystem
The collaboration between Meta and AWS not only underscores the importance of ARM-based processors in AI infrastructure but also highlights the shift in focus from purely GPU-centric AI hardware to a more balanced approach. As AI models grow more complex, the demand for efficient, scalable CPUs is rising. This deal could influence other tech firms to consider similar partnerships, potentially reshaping how AI compute resources are procured and deployed across the industry.
Overall, the agreement signals a pivotal moment in the evolution of AI infrastructure, where the emphasis is shifting toward flexible, high-performance compute solutions that can support the next wave of AI innovation.



