
H200 Specifications
Nearly double the memory capacity of H100 with 4,800 GB/s HBM3e bandwidth, enabling larger models, longer context windows, and faster inference without compromising precision.
141 GB
HBM3e Memory
Run 70B+ parameter models on a single GPU.
900 GB/s
NVLink Bandwidth
Fourth-gen NVLink for multi-GPU scaling.
1,979
TFLOPS FP16
3,958 TFLOPS at FP8 with Transformer Engine.
9X
Faster Pre-Training
vs A100 on large language models.
NVIDIA H200 on Denvr AI Cloud

Distributed Training
Scale across 8-GPU NVLink nodes and multi-node clusters with 3,200G InfiniBand. Bare metal access for maximum training throughput with no virtualization overhead.

LLM Inference
Serve >70B parameter models on a single GPU, or scale to 8 GPUs with 1,536 GB VRAM for >1T+ parameter models at FP8 precision.

Production AI
H200 delivers 2x faster inference vs H100 with 1.8x the memory. For FP4 and next-gen training, consider the Blackwell GPUs with 180 GB HBM3e.

Managed Storage
High-performance Weka filesystem and local NVMe available. No external storage to provision for datasets, checkpoints, or model artifacts.
Platform
GPUs
GPU VRAM
vCPUs
Memory
Local Storage
Interconnect
On-Demand
NVIDIA H200 SXM
8
141 GB
208
2048 GB
6x 3.8TB NVMe
RoCE 3200G
Reserved only
Configurations
Per-minute billing with on-demand and reserved options. All configurations available as bare metal, VM, or model endpoints.
Related GPUs
Compare Denvr GPU options by workload and performance requirements.
Optimized For
Large model training, high-throughput inference
Extended context, large batch inference
Next-gen training with FP4 precision
VRAM
80 GB
141 GB
180 GB
Memory Bandwidth
3,350 GB/s
4,800 GB/s
8,000 GB/s
FP64/FP32
67 TFLOPS
67 TFLOPS
40 TFLOPS
FP16
1,979 TFLOPS
1,979 TFLOPS
2,250 TFLOPS
FP8
3,958 TFLOPS
3,958 TFLOPS
4,500 TFLOPS
NVLink
900 GB/s
900 GB/s
1.8 TB/s
On-Demand Pricing
$2.10 / GPU
Reserved only
Reserved only
Infrastructure you can trust at scale
As an NVIDIA Cloud Partner we build and operate AI clusters following NVIDIA Reference Architectures. Your models and data are supported via strict privacy safeguards and SOC 2 Type 2 security practices.









