The world's most advanced GPU for AI and HPC. Experience unprecedented performance with Transformer Engine.
SHA-256-C7E8976BBAF2Save 5%Train and deploy massive LLMs with Transformer Engine acceleration and FP8 precision.
Deploy production AI services with industry-leading throughput and latency.
Accelerate HPC workloads with massive memory bandwidth and compute power.
| Specification | H100 80GB | A100 80GB | H100 SXM | 
|---|---|---|---|
| Memory | 80 GB HBM3 | 80 GB HBM2e | 80 GB HBM3 | 
| Memory Bandwidth | 3,350 GB/s | 2,039 GB/s | 3,350 GB/s | 
| CUDA Cores | 16,896 | 6,912 | 16,896 | 
| FP32 Performance | 67 TFLOPS | 19.5 TFLOPS | 67 TFLOPS | 
| Architecture | Hopper | Ampere | Hopper | 
| Price/Hour | From $3.47 | From $2.49 | From $4.99 | 
The H100 features the new Hopper architecture with Transformer Engine, providing up to 9x faster AI training and 30x faster inference. It includes FP8 precision, 80GB HBM3 memory with 3.35TB/s bandwidth, and fourth-generation Tensor Cores.
The Transformer Engine uses FP8 precision with FP16 accuracy to deliver up to 5x faster training for large language models. It automatically manages precision conversion and scaling for optimal performance.
Yes, H100 supports Multi-Instance GPU (MIG) technology, allowing you to partition a single H100 into up to 7 isolated GPU instances with dedicated resources.
H100 instances come with CUDA 12.0+, PyTorch 2.1+, TensorFlow 2.14+, JAX, and support for Transformer Engine optimizations. Custom Docker images are fully supported.
Join leading AI teams using next-generation GPU compute