The H100 successor with 76% more VRAM. 141GB HBM3e for frontier models and research.
76% more VRAM (141GB vs 80GB) and 43% more memory bandwidth (4,800 vs 3,350 GB/s). Same compute, much more memory for larger models.
Yes, with 8x H200 you get 1,128 GB total VRAM — enough for LLaMA 405B inference without quantization.
HBM3e is the latest high-bandwidth memory technology, offering 4,800 GB/s bandwidth per GPU — 43% faster than HBM3 on H100.
VoltageGPU bills per second with no minimum commitment. Run for 5 minutes or 5 months — pay only for what you use.
$5 free credit. No credit card required. Deploy in under 60 seconds.