Akamai Technologies has taken a significant leap forward in artificial intelligence by launching a global-scale implementation of the NVIDIA AI Grid reference architecture. By integrating advanced infrastructure from NVIDIA into its ecosystem, Akamai is actively reshaping how enterprises deploy AI—shifting from centralized AI factories to a distributed inference model powered by intelligent orchestration.
To begin with, this innovation builds on Akamai’s Inference Cloud, introduced recently, and expands its capabilities across more than 4,400 global edge locations. As a result, enterprises can now deploy AI workloads closer to users, ensuring faster response times and improved performance. Additionally, the rollout includes thousands of NVIDIA RTX PRO 6000 Blackwell Server Edition GPUs, enabling support for both agentic and physical AI applications at scale.
“AI factories have been purpose-built for training and frontier model workloads — and centralized infrastructure will continue to deliver the best tokenomics for those use cases,” said Adam Karon, Chief Operating Officer and General Manager, Cloud Technology Group, Akamai. “But real-time video, physical AI, and highly concurrent personalized experiences demand inference at the point of contact, not a round trip to a centralized cluster. Our AI Grid intelligent orchestration gives AI factories a way to scale inference outward — leveraging the same distributed architecture that revolutionized content delivery to route AI workloads across 4,400 locations, at the right cost, at the right time.”
At the core of this system lies an intelligent orchestrator that dynamically manages AI requests. Not only does it optimize “tokenomics,” but it also improves cost efficiency, latency, and throughput. Furthermore, enterprises can benefit from semantic caching and intelligent routing, which allocate workloads to the most suitable compute resources automatically.
For instance, gaming companies can now deliver immersive AI-driven NPC interactions in milliseconds. Similarly, financial institutions can execute real-time fraud detection and personalized recommendations instantly after user login. Meanwhile, broadcasters can leverage the platform for real-time content transcoding and dubbing across global audiences.
Importantly, the infrastructure combines edge and core computing. While edge locations ensure ultra-low latency, centralized GPU clusters handle high-density workloads such as large language models and multimodal AI processing. This hybrid approach enables a seamless continuum of compute from core to edge.
“New AI-native applications demand predictable latency and better cost efficiency at planetary scale,” said Chris Penrose, Global VP – Business Development – Telco at NVIDIA. “By operationalizing the NVIDIA AI Grid, Akamai is building the connective tissue for generative, agentic, and physical AI, moving intelligence directly to the data to unlock the next wave of real-time applications.”
Moreover, early adoption across industries such as gaming, financial services, retail, and media highlights strong demand for distributed AI capabilities. In fact, a major $200 million, four-year agreement further validates the platform’s enterprise value.
Ultimately, Akamai’s AI Grid represents a major shift in AI infrastructure strategy. By combining distributed networking, intelligent orchestration, and scalable compute, the company is paving the way for a future where AI operates as a globally distributed utility rather than isolated systems.
Recommended Cyber Technology News:
- Advanced Persistent Threat and Why it’s Dangerous
- Adversarial AI in Cybersecurity: How Hackers Manipulate AI-Powered Security
- Account Takeover (ATO): Why Identity Is the New Perimeter
To participate in our interviews, please write to our CyberTech Media Room at info@intentamplify.com





