GPU Plans by Amazon AWS

GPU plans sold by Amazon AWS with their specifications and prices. Each plan includes one or more GPU plus CPU, memory, disk and data transfer.

GPU Types

Cloud providers offer a variety of GPUs from vendors like NVIDIA and AMD, each optimized for different workloads. From AI training to graphics rendering, these powerful processors enable diverse applications. Here's a look at some common GPU types:

  • A10: A versatile data center GPU, balancing AI inference and graphics rendering. Offers strong performance for diverse workloads, including virtual workstations and AI-powered video processing. See also A10G
  • A40: A professional workstation GPU built for demanding graphics and AI tasks. It delivers exceptional performance for visual effects, 3D rendering, and AI-accelerated workflows in professional environments.
  • A100: A high-performance data center GPU designed for accelerating AI training and high-performance computing. Delivers exceptional computational power for complex simulations and large-scale deep learning models.
  • H100: NVIDIA's next-generation AI GPU, providing a significant leap in performance over the A100. Engineered for massive AI workloads, with improved transformer engine performance, and increased memory bandwidth.
  • H200: An enhanced version of the H100, designed to tackle the most demanding AI workloads. It offers increased memory capacity and bandwidth, enabling faster processing of massive datasets for large language models and generative AI.
  • L4: An energy-efficient GPU optimized for AI video and inference workloads in the data center. It excels at tasks like video transcoding, AI-powered video analysis, and real-time inference, while maintaining a low power footprint.
  • T4: An entry-level inference GPU, widely used in cloud environments for AI inference and graphics virtualization. Provides a cost-effective solution for deploying AI models and delivering virtual desktops.
  • L40S: A powerful data center GPU designed for professional visualization and demanding AI workloads. Ideal for rendering complex 3D models, running simulations, and accelerating AI-driven design and content creation.
  • NVIDIA V100: A previous-generation high-performance GPU, still widely used for AI training and scientific computing. It offers substantial computational power and memory bandwidth for demanding workloads. See also NVIDIA V100S.
  • AMD Radeon Pro V520: A professional workstation GPU designed for visualization and graphics-intensive applications. It delivers reliable performance for tasks like 3D modeling, rendering, and video editing.
  • Nvidia RTX 4000: The NVIDIA RTX 4000 Ada Generation is a powerful professional GPU with 20GB GDDR6 ECC memory. Featuring 6144 CUDA cores, 192 Tensor cores, and 48 RT cores, it excels in demanding creative, design, and AI workflows. Its single-slot, power-efficient design delivers high performance for complex tasks.
  • Nvidia Quadro RTX 6000: The Radeon Pro RTX 6000 is a high-end professional workstation graphics card. It boasts 48GB of GDDR6 ECC memory, 18,176 CUDA cores, 568 Tensor cores, and 142 RT cores, delivering exceptional performance for demanding tasks like 3D rendering, AI, and data science. Its 960 GB/s memory bandwidth and advanced features like DLSS 3 and SER accelerate workflows, making it a top choice for professionals.

This is not a comprehensive list and prices may vary before VPSBenchmarks can update them.

AWS Amazon AWS
GPUs - G4

Amazon EC2 G4 instances are the industry’s most cost-effective and versatile GPU instances for deploying machine learning models such as image classification, object detection, and speech recognition, and for graphics-intensive applications such as remote graphics workstations, game streaming, and graphics rendering. G4 instances are available with a choice of NVIDIA GPUs (G4dn) or AMD GPUs (G4ad).

Plan GPU Type GPU RAM vCPUs RAM Storage Price
g4dn.xlarge 1 x NVIDIA T4 16 GB 4 16 GB 125 GB $0.53/hr
g4dn.metal 8 x NVIDIA T4 16 GB 96 384 GB 1800 GB $7.82/hr
g4ad.xlarge 1 x AMD Radeon Pro V520 8 GB 4 16 GB 150 GB $0.38/hr
g4ad.2xlarge 1 x AMD Radeon Pro V520 8 GB 8 32 GB 300 GB $0.54/hr
g4ad.8xlarge 2 x AMD Radeon Pro V520 8 GB 32 128 GB 1200 GB $1.73/hr
g4ad.16xlarge 4 x AMD Radeon Pro V520 8 GB 64 256 GB 2400 GB $3.47/hr
g4dn.12xlarge 4 x Nvidia T4 16 GB 48 192 GB 900 GB $3.91/hr
GPUs - P4d

Amazon Elastic Compute Cloud (Amazon EC2) P4d instances deliver high performance for machine learning (ML) training and high performance computing (HPC) applications in the cloud. P4d instances are powered by NVIDIA A100 Tensor Core GPUs and deliver industry-leading high throughput and low-latency networking. These instances support 400 Gbps instance networking. P4d instances provide up to 60% lower cost to train ML models, including an average of 2.5x better performance for deep learning models compared to previous-generation P3 and P3dn instances.

Plan GPU Type GPU RAM vCPUs RAM Storage Price
p4d.24xlarge 8 x A100 320 GB 96 1152 GB 8000 GB $32.77/hr
p4de.24xlarge 8 x A100 640 GB 96 1152 GB 8000 GB $40.96/hr
GPUs - G5

G5 instances feature up to 8 NVIDIA A10G Tensor Core GPUs and second generation AMD EPYC processors. They also support up to 192 vCPUs, up to 100 Gbps of network bandwidth, and up to 7.6 TB of local NVMe SSD storage.

Plan GPU Type GPU RAM vCPUs RAM Storage Price
g5.xlarge 1 x Nvidia A10G 24 GB 4 16 GB 250 GB $1.01/hr
g5.16xlarge 1 x Nvidia A10G 24 GB 64 256 GB 1900 GB $4.10/hr
g5.12xlarge 4 x Nvidia A10G 96 GB 48 192 GB 3800 GB $5.67/hr
g5.48xlarge 8 x Nvidia A10G 192 GB 192 768 GB 7600 GB $16.29/hr
g5.2xlarge 1 x Nvidia A10G 24 GB 8 32 GB 450 GB $1.21/hr
GPUs - P5

P5 instances provide up to 8 NVIDIA H100 GPUs with a total of up to 640 GB HBM3 GPU memory per instance. P5e and P5en instances provide up to 8 NVIDIA H200 GPUs with a total of up to 1128 GB HBM3e GPU memory per instance. Both instances support up to 900 GB/s of NVSwitch GPU interconnect (total of 3.6 TB/s bisectional bandwidth in each instance), so each GPU can communicate with every other GPU in the same instance with single-hop latency.

Plan GPU Type GPU RAM vCPUs RAM Storage Price
p5en.48xlarge 8 x H200 1128 GB 192 2000 GB 30720 GB $63.30/hr
p5.48xlarge 8 x H100 640 GB 192 2000 GB 30720 GB $55.04/hr
p5.4xlarge 1 x H100 80 GB 16 256 GB 3840 GB $6.88/hr
GPUs - G6

G6 instances feature up to 8 NVIDIA L4 Tensor Core GPUs with 24 GB of memory per GPU and third generation AMD EPYC processors. They also support up to 192 vCPUs, up to 100 Gbps of network bandwidth, and up to 7.52 TB of local NVMe SSD storage.

Plan GPU Type GPU RAM vCPUs RAM Storage Price
g6.xlarge 1 x Nvidia L4 24 GB 4 16 GB 250 GB $0.80/hr
gr6.4xlarge 1 x Nvidia L4 24 GB 16 128 GB 600 GB $1.54/hr
g6.4xlarge 1 x Nvidia L4 24 GB 16 64 GB 600 GB $1.32/hr
g6.12xlarge 4 x Nvidia L4 96 GB 48 192 GB 3760 GB $4.60/hr
g6.48xlarge 8 x Nvidia L4 192 GB 192 768 GB 7520 GB $13.35/hr
GPUs - P6

Amazon EC2 P6 instances, powered by NVIDIA Blackwell GPUs, are designed for high-performance AI training and inference. Featuring 5th Generation Intel Xeon Scalable processors, these instances provide substantial leaps in GPU memory and compute throughput compared to previous generations. They are optimized for large-scale distributed AI workloads, including mixture of experts and trillion-parameter models, utilizing Elastic Fabric Adapter (EFA) networking and high-bandwidth local NVMe storage to accelerate data-intensive machine learning and high-performance computing applications efficiently in the cloud.

Plan GPU Type GPU RAM vCPUs RAM Storage Price
p6-b200.48xlarge 8 x NVIDIA B200 1432 GB 192 2048 GB 30720 GB $113.93/hr

Be the first to learn about new Best VPS rankings. Subscribe to our newsletter.