The most deployed data center GPU for AI. Ampere architecture delivers proven performance for training and inference with 80GB HBM2e memory and excellent price-to-performance ratio.
Flexible pricing options to match your workload requirements.
Pay as you go with no commitment
Save 15% with monthly commitment
Maximum savings with annual commitment
Battle-tested GPU powering AI infrastructure at leading tech companies.
Partition into up to 7 isolated GPU instances for multi-workload efficiency.
600 GB/s GPU-to-GPU bandwidth for multi-GPU training workloads.
Best price-to-performance ratio for most AI and ML workloads.
Train computer vision, NLP, and recommendation models efficiently.
Fine-tune foundation models like Llama, Mistral, and Falcon.
Deploy models at scale with MIG partitioning for cost efficiency.
Accelerate scientific simulations, genomics, and climate modeling.