Rent NVIDIA H200 141GB HBM3e 8-GPU clusters from $26.60/hr. 1.1 TB total VRAM for large-scale LLM training, 200B+ model serving, and frontier AI research. VoltageGPU cloud.
Starting from
$26.60/hr
~$638.40/day
~$19,152/month (24/7)
Per-minute billing · No commitment
VRAM
8×141 GB HBM3e
Memory Type
HBM3e
Memory Bandwidth
4,800 GB/s
CUDA Cores
14,592
Tensor Cores
456
FP16 Performance
989.5 TFLOPS
FP32 Performance
67 TFLOPS
TDP
700W (SXM)
Architecture
Hopper
Interconnect
NVLink 4.0 / PCIe 5.0
Included Storage
1 TB NVMe SSD
vCPUs
48 vCPUs
System RAM
384 GB DDR5 ECC
Manufacturer
NVIDIA
See how VoltageGPU compares to other cloud GPU providers.
| Provider | Hourly Rate | Est. Monthly | vs VoltageGPU |
|---|---|---|---|
| VoltageGPUYou | $26.60 | $19,152 | — |
| RunPod | $29.90 | $21,528 | 11% cheaper |
| Vast.ai | $28.50 | $20,520 | 7% cheaper |
| Lambda | $32.00 | $23,040 | 17% cheaper |
| AWS (p5e equivalent) | $42.00 | $30,240 | 37% cheaper |
Competitor pricing sourced from public pages as of March 2026. Prices may vary.
Popular workloads and use cases for NVIDIA H200 141GB cloud instances.
Train 70B–200B parameter models with massive VRAM across 8 GPUs. The 141 GB per GPU (1.1 TB total) eliminates memory bottlenecks.
Serve 70B models unquantized or 400B+ models with quantization across the 8-GPU configuration for production inference.
Train and serve vision-language models, video generation models, and other multi-modal architectures that require massive memory.
Dedicated compute for AI research labs. Combine multiple H200 nodes for frontier model development.
Relative performance scores across common workload categories (B200 = 100).
Programmatically launch a H200 141GB instance with a single API call.
curl -X POST https://api.voltagegpu.com/v1/pods \
-H "Authorization: Bearer YOUR_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"gpu": "h200-141gb",
"gpu_count": 8,
"template": "pytorch-2.2",
"storage_gb": 1000,
"name": "my-h200-cluster"
}'Train and fine-tune the largest AI models on flagship GPUs like H200 and B200.
Run the largest open-source LLM on flagship GPU clusters with multi-GPU support.
Deploy multi-GPU clusters for large-scale AI training and research.
See how VoltageGPU pricing and features compare to RunPod for GPU cloud.
Compare GPU cloud pricing, features, and availability against Vast.ai.
View pricing for all GPU models and AI inference API with no hidden fees.
See real-time GPU availability and deploy a pod in under 60 seconds.
Deploy a H200 141GB instance in 30 seconds. No upfront costs, no long-term contracts. Per-minute billing starting at $26.60/hr.