Cloud GPUs on Akamai

Access on‑demand NVIDIA GPUs with predictable pricing, low egress costs, and fast provisioning to accelerate AI training and inference, HPC, rendering, and media workloads.

Why GPUs on Akamai Cloud

Available GPU options

Need larger clusters or edge‑native inference? Akamai Inference Cloud supports high‑performance inference stacks, including multi‑GPU configurations (up to 8x RTX PRO 6000 Blackwell Server Edition GPUs per node), NVIDIA BlueField DPUs, and high‑memory/NVMe profiles to optimize TTFT and TPS. Explore Akamai Inference Cloud

Performance and cost highlights

Learn how to optimize cost and performance. Download the AI inference cost optimization white paper

Platform capabilities

Provisioning: from zero to GPU in minutes

  1. Create an Akamai Cloud account. Sign up
  2. Choose a region and GPU plan (RTX PRO 6000 Blackwell, RTX 4000 Ada, or Quadro RTX 6000). See pricing and regions
  3. Select your OS image and deploy the instance.
  4. Install NVIDIA drivers and CUDA (or use a prepared image). Follow the CUDA guide
  5. Attach Block Storage and configure backups/snapshots as needed.
  6. Configure networking (private networking, firewall rules, and load balancer).
  7. Optional: Add GPU node pools to an LKE cluster for scalable inference/training.
  8. Deploy your application and monitor performance via the UI, API, or your observability stack.

Popular workloads

Next steps

Have a specific model or pipeline in mind? We can help map it to the right GPU, storage, and network profile so you can deploy with confidence.