H200 GPU

The H200 is designed for large-scale AI training and inference, offering improved precision and throughput for transformer models. It handles complex AI workloads with better energy efficiency than previous generations.

Starting Price
$1.90/hr
Available on 8 cloud providers
H200 GPU

Key Specifications

๐Ÿ’พMemory

141GB VRAM

๐Ÿ—๏ธArchitecture

Hopper

โš™๏ธCompute Units

N/A

๐ŸงฎTensor Cores

528

Technical Specifications

Hardware Details

ManufacturerNVIDIA
ArchitectureHopper
CUDA Cores16896
Tensor Cores528
RT CoresN/A
Compute UnitsN/A
GenerationN/A

Memory & Performance

VRAM141GB
Memory InterfaceN/A
Memory Bandwidth4800 GB/s
FP32 Performance67 TFLOPS
FP16 Performance990 TFLOPS
INT8 Performance1979 TOPS

Performance

Computing Power

CUDA Cores16,896
Tensor Cores528

Computational Performance

FP32 (TFLOPS)67
FP16 (TFLOPS)990
INT8 (TOPS)1,979

Common Use Cases

Generative AI and large language models (LLMs) training. Advanced scientific computing for HPC workloads.

Machine Learning & AI

  • Training large language models and transformers
  • Computer vision and image processing
  • Deep learning model development
  • High-performance inference workloads

Graphics & Compute

  • 3D rendering and visualization
  • Scientific simulations
  • Data center graphics virtualization
  • High-performance computing (HPC)

Market Context

The H200 sits within NVIDIA's Hopper architecture lineup, positioned in the ultra performance tier.

Cloud Availability

Available across 8 cloud providers with prices ranging from $1.90/hr. Pricing and availability may vary by region and provider.

Market Position

Released in 2023, this GPU is positioned for professional workloads.

Current Pricing

ProviderHourly PriceSource
Deep Infra
$3.00/hr
CoreWeave
$6.31/hr
RunPod
$3.59/hr
Amazon AWS
$4.33/hr
Vast.ai
$3.29/hr
White Fiber
$1.90/hr
Genesis Cloud
$2.95/hr
Datacrunch
$3.13/hr

Prices are updated regularly. Last updated: 6/6/2025