H200 GPU
The H200 is designed for large-scale AI training and inference, offering improved precision and throughput for transformer models. It handles complex AI workloads with better energy efficiency than previous generations.
Starting Price
$1.90/hr
Available on 8 cloud providers

Key Specifications
๐พMemory
141GB VRAM
๐๏ธArchitecture
Hopper
โ๏ธCompute Units
N/A
๐งฎTensor Cores
528
Technical Specifications
Hardware Details
ManufacturerNVIDIA
ArchitectureHopper
CUDA Cores16896
Tensor Cores528
RT CoresN/A
Compute UnitsN/A
GenerationN/A
Memory & Performance
VRAM141GB
Memory InterfaceN/A
Memory Bandwidth4800 GB/s
FP32 Performance67 TFLOPS
FP16 Performance990 TFLOPS
INT8 Performance1979 TOPS
Performance
Computing Power
CUDA Cores16,896
Tensor Cores528
Computational Performance
FP32 (TFLOPS)67
FP16 (TFLOPS)990
INT8 (TOPS)1,979
Common Use Cases
Generative AI and large language models (LLMs) training. Advanced scientific computing for HPC workloads.
Machine Learning & AI
- Training large language models and transformers
- Computer vision and image processing
- Deep learning model development
- High-performance inference workloads
Graphics & Compute
- 3D rendering and visualization
- Scientific simulations
- Data center graphics virtualization
- High-performance computing (HPC)
Market Context
The H200 sits within NVIDIA's Hopper architecture lineup, positioned in the ultra performance tier.
Cloud Availability
Available across 8 cloud providers with prices ranging from $1.90/hr. Pricing and availability may vary by region and provider.
Market Position
Released in 2023, this GPU is positioned for professional workloads.