GPU Plans by Amazon AWS
GPU plans sold by Amazon AWS with their specifications and prices. Each plan includes one or more GPU plus CPU, memory, disk and data transfer.
GPU Types
Cloud providers offer a variety of GPUs from vendors like NVIDIA and AMD, each optimized for different workloads. From AI training to graphics rendering, these powerful processors enable diverse applications. Here's a look at some common GPU types:
- A10: A versatile data center GPU, balancing AI inference and graphics rendering. Offers strong performance for diverse workloads, including virtual workstations and AI-powered video processing. See also A10G
- A40: A professional workstation GPU built for demanding graphics and AI tasks. It delivers exceptional performance for visual effects, 3D rendering, and AI-accelerated workflows in professional environments.
- A100: A high-performance data center GPU designed for accelerating AI training and high-performance computing. Delivers exceptional computational power for complex simulations and large-scale deep learning models.
- H100: NVIDIA's next-generation AI GPU, providing a significant leap in performance over the A100. Engineered for massive AI workloads, with improved transformer engine performance, and increased memory bandwidth.
- H200: An enhanced version of the H100, designed to tackle the most demanding AI workloads. It offers increased memory capacity and bandwidth, enabling faster processing of massive datasets for large language models and generative AI.
- L4: An energy-efficient GPU optimized for AI video and inference workloads in the data center. It excels at tasks like video transcoding, AI-powered video analysis, and real-time inference, while maintaining a low power footprint.
- T4: An entry-level inference GPU, widely used in cloud environments for AI inference and graphics virtualization. Provides a cost-effective solution for deploying AI models and delivering virtual desktops.
- L40S: A powerful data center GPU designed for professional visualization and demanding AI workloads. Ideal for rendering complex 3D models, running simulations, and accelerating AI-driven design and content creation.
- NVIDIA V100: A previous-generation high-performance GPU, still widely used for AI training and scientific computing. It offers substantial computational power and memory bandwidth for demanding workloads. See also NVIDIA V100S.
- AMD Radeon Pro V520: A professional workstation GPU designed for visualization and graphics-intensive applications. It delivers reliable performance for tasks like 3D modeling, rendering, and video editing.
- Nvidia RTX 4000: The NVIDIA RTX 4000 Ada Generation is a powerful professional GPU with 20GB GDDR6 ECC memory. Featuring 6144 CUDA cores, 192 Tensor cores, and 48 RT cores, it excels in demanding creative, design, and AI workflows. Its single-slot, power-efficient design delivers high performance for complex tasks.
- Nvidia Quadro RTX 6000: The Radeon Pro RTX 6000 is a high-end professional workstation graphics card. It boasts 48GB of GDDR6 ECC memory, 18,176 CUDA cores, 568 Tensor cores, and 142 RT cores, delivering exceptional performance for demanding tasks like 3D rendering, AI, and data science. Its 960 GB/s memory bandwidth and advanced features like DLSS 3 and SER accelerate workflows, making it a top choice for professionals.
This is not a comprehensive list and prices may vary before VPSBenchmarks can update them.
|
|
|||||||
|---|---|---|---|---|---|---|---|
|
GPUs - G4
Amazon EC2 G4 instances are the industry’s most cost-effective and versatile GPU instances for deploying machine learning models such as image classification, object detection, and speech recognition, and for graphics-intensive applications such as remote graphics workstations, game streaming, and graphics rendering. G4 instances are available with a choice of NVIDIA GPUs (G4dn) or AMD GPUs (G4ad). |
|||||||
| Plan | GPU Type | GPU RAM | vCPUs | RAM | Storage | Price | |
| g4dn.xlarge | 1 x NVIDIA T4 | 16 GB | 4 | 16 GB | 125 GB | $0.53/hr | |
| g4dn.metal | 8 x NVIDIA T4 | 16 GB | 96 | 384 GB | 1800 GB | $7.82/hr | |
| g4ad.xlarge | 1 x AMD Radeon Pro V520 | 8 GB | 4 | 16 GB | 150 GB | $0.38/hr | |
| g4ad.2xlarge | 1 x AMD Radeon Pro V520 | 8 GB | 8 | 32 GB | 300 GB | $0.54/hr | |
| g4ad.8xlarge | 2 x AMD Radeon Pro V520 | 8 GB | 32 | 128 GB | 1200 GB | $1.73/hr | |
| g4ad.16xlarge | 4 x AMD Radeon Pro V520 | 8 GB | 64 | 256 GB | 2400 GB | $3.47/hr | |
| g4dn.12xlarge | 4 x Nvidia T4 | 16 GB | 48 | 192 GB | 900 GB | $3.91/hr | |
|
GPUs - P4d
Amazon Elastic Compute Cloud (Amazon EC2) P4d instances deliver high performance for machine learning (ML) training and high performance computing (HPC) applications in the cloud. P4d instances are powered by NVIDIA A100 Tensor Core GPUs and deliver industry-leading high throughput and low-latency networking. These instances support 400 Gbps instance networking. P4d instances provide up to 60% lower cost to train ML models, including an average of 2.5x better performance for deep learning models compared to previous-generation P3 and P3dn instances. |
|||||||
| Plan | GPU Type | GPU RAM | vCPUs | RAM | Storage | Price | |
| p4d.24xlarge | 8 x A100 | 320 GB | 96 | 1152 GB | 8000 GB | $32.77/hr | |
| p4de.24xlarge | 8 x A100 | 640 GB | 96 | 1152 GB | 8000 GB | $40.96/hr | |
|
GPUs - G5
G5 instances feature up to 8 NVIDIA A10G Tensor Core GPUs and second generation AMD EPYC processors. They also support up to 192 vCPUs, up to 100 Gbps of network bandwidth, and up to 7.6 TB of local NVMe SSD storage. |
|||||||
| Plan | GPU Type | GPU RAM | vCPUs | RAM | Storage | Price | |
| g5.xlarge | 1 x Nvidia A10G | 24 GB | 4 | 16 GB | 250 GB | $1.01/hr | |
| g5.16xlarge | 1 x Nvidia A10G | 24 GB | 64 | 256 GB | 1900 GB | $4.10/hr | |
| g5.12xlarge | 4 x Nvidia A10G | 96 GB | 48 | 192 GB | 3800 GB | $5.67/hr | |
| g5.48xlarge | 8 x Nvidia A10G | 192 GB | 192 | 768 GB | 7600 GB | $16.29/hr | |
| g5.2xlarge | 1 x Nvidia A10G | 24 GB | 8 | 32 GB | 450 GB | $1.21/hr | |
|
GPUs - P5
P5 instances provide up to 8 NVIDIA H100 GPUs with a total of up to 640 GB HBM3 GPU memory per instance. P5e and P5en instances provide up to 8 NVIDIA H200 GPUs with a total of up to 1128 GB HBM3e GPU memory per instance. Both instances support up to 900 GB/s of NVSwitch GPU interconnect (total of 3.6 TB/s bisectional bandwidth in each instance), so each GPU can communicate with every other GPU in the same instance with single-hop latency. |
|||||||
| Plan | GPU Type | GPU RAM | vCPUs | RAM | Storage | Price | |
| p5en.48xlarge | 8 x H200 | 1128 GB | 192 | 2000 GB | 30720 GB | $63.30/hr | |
| p5.48xlarge | 8 x H100 | 640 GB | 192 | 2000 GB | 30720 GB | $55.04/hr | |
| p5.4xlarge | 1 x H100 | 80 GB | 16 | 256 GB | 3840 GB | $6.88/hr | |
|
GPUs - G6
G6 instances feature up to 8 NVIDIA L4 Tensor Core GPUs with 24 GB of memory per GPU and third generation AMD EPYC processors. They also support up to 192 vCPUs, up to 100 Gbps of network bandwidth, and up to 7.52 TB of local NVMe SSD storage. |
|||||||
| Plan | GPU Type | GPU RAM | vCPUs | RAM | Storage | Price | |
| g6.xlarge | 1 x Nvidia L4 | 24 GB | 4 | 16 GB | 250 GB | $0.80/hr | |
| gr6.4xlarge | 1 x Nvidia L4 | 24 GB | 16 | 128 GB | 600 GB | $1.54/hr | |
| g6.4xlarge | 1 x Nvidia L4 | 24 GB | 16 | 64 GB | 600 GB | $1.32/hr | |
| g6.12xlarge | 4 x Nvidia L4 | 96 GB | 48 | 192 GB | 3760 GB | $4.60/hr | |
| g6.48xlarge | 8 x Nvidia L4 | 192 GB | 192 | 768 GB | 7520 GB | $13.35/hr | |
|
GPUs - P6
Amazon EC2 P6 instances, powered by NVIDIA Blackwell GPUs, are designed for high-performance AI training and inference. Featuring 5th Generation Intel Xeon Scalable processors, these instances provide substantial leaps in GPU memory and compute throughput compared to previous generations. They are optimized for large-scale distributed AI workloads, including mixture of experts and trillion-parameter models, utilizing Elastic Fabric Adapter (EFA) networking and high-bandwidth local NVMe storage to accelerate data-intensive machine learning and high-performance computing applications efficiently in the cloud. |
|||||||
| Plan | GPU Type | GPU RAM | vCPUs | RAM | Storage | Price | |
| p6-b200.48xlarge | 8 x NVIDIA B200 | 1432 GB | 192 | 2048 GB | 30720 GB | $113.93/hr | |