ultraData Center

H100 SXM GPU

The H100 SXM is the flagship Hopper GPU with HBM3 memory and NVLink 4.0, designed for maximum AI training and HPC performance in data centers.

VRAM 80GB
CUDA Cores 14,592
Tensor Cores 456
TDP 700W
From
$0.80/hr
across 30 providers
H100 SXM GPU

Cloud Pricing

Cheapest on Verda 71% below avg
ProviderConfigPrice / hrUpdatedSource
1×2×4×8×
$0.80/hr
4/21/2026
1×
$1.50/hr
4/21/2026
1×
$1.50/hr
4/15/2026
1×
$1.57/hr
4/21/2026
8×
$1.60/hr12mo
4/18/2026
1×
$1.61/hr6mo
3/30/2026
1×
$1.70/hr3mo
3/30/2026
8×
$1.75/hr9mo
4/18/2026
1×
$1.79/hr
4/20/2026
1×
$1.79/hr
4/18/2026
1×
$1.79/hr1mo
3/31/2026
1×
$1.79/hr
3/30/2026
8×
$1.79/hr
4/21/2026
1×
$1.80/hr1mo
3/30/2026
1×
$1.83/hr
4/21/2026
1×2×4×
$1.85/hr
4/21/2026
1×2×4×8×
$1.90/hr
4/21/2026
8×
$1.97/hr
4/21/2026
1×
$1.98/hr
4/21/2026
1×2×4×8×
$1.99/hr
4/21/2026
8×
$1.99/hr
4/12/2026
8×
$1.99/hr24mo
4/10/2026
4×5×8×
$1.99/hr
4/21/2026
2×
$2.00/hr
4/21/2026
2×
$2.00/hr
4/21/2026
4×8×
$2.00/hr
4/21/2026
2×4×
$2.09/hr
4/21/2026
2×
$2.10/hr
4/21/2026
8×
$2.19/hr
4/18/2026
1×
$2.22/hr
4/21/2026
1×
$2.24/hr36mo
4/16/2026
4×
$2.24/hr
4/21/2026
1×
$2.25/hr3mo
3/30/2026
1×
$2.25/hr6mo
3/30/2026
1×
$2.25/hr
4/16/2026
1×2×4×
$2.29/hr
4/15/2026
1×2×4×8×
$2.29/hr
4/21/2026
1×2×4×
$2.40/hr
4/21/2026
1×8×
$2.49/hr36mo
4/17/2026
1×8×
$2.59/hr24mo
4/17/2026
1×
$2.69/hr
4/21/2026
1×
$2.69/hr
4/21/2026
1×8×
$2.69/hr12mo
4/17/2026
8×
$2.75/hr
4/21/2026
1×8×
$2.79/hr6mo
4/17/2026
1×
$2.95/hr
4/21/2026
2×
$2.98/hr
4/21/2026
8×
$2.99/hr
4/21/2026
1×
$2.99/hr
4/2/2026
1×
$2.99/hr1mo
4/19/2026
1×8×
$2.99/hr
4/17/2026
8×
$3.00/hr
4/21/2026
2×
$3.04/hr
4/21/2026
1×
$3.08/hr
4/21/2026
1×
$3.29/hr
4/21/2026
8×
$3.39/hr
4/21/2026
1×
$3.39/hr
4/21/2026
4×
$3.41/hr
4/21/2026
1×
$3.90/hr
4/20/2026
8×
$3.99/hr
4/21/2026
1×4×8×
$3.99/hr
4/21/2026
4×
$4.09/hr
4/21/2026
1×2×4×8×
$5.49/hr
4/16/2026
1×2×4×8×
$5.74/hr
4/21/2026
8×
$5.95/hr
4/21/2026
1×
$5.99/hr
4/21/2026
8×
$6.16/hr
4/17/2026
8×
$6.88/hr
4/21/2026
1×
$10.00/hr
4/21/2026
Direct from providerVia marketplace

Prices updated daily. Last check: 4/21/2026

Performance

FP16
990 TFLOPS
FP32
67 TFLOPS
Bandwidth
3350 GB/s

Strengths & Limitations

  • 80GB HBM3 memory capacity supports large model training and inference
  • 3,350 GB/s memory bandwidth enables efficient data movement for memory-intensive workloads
  • 4th Generation Tensor Cores with FP8 precision support via Transformer Engine
  • 990 TFLOPS FP16 performance for AI training and inference acceleration
  • Multi-Instance GPU (MIG) technology allows workload partitioning and isolation
  • NVIDIA Confidential Computing provides hardware-level security for sensitive workloads
  • NVLink interconnect supports high-bandwidth multi-GPU configurations
  • 700W TDP requires substantial power delivery and cooling infrastructure
  • SXM form factor limits deployment to compatible server platforms
  • Previous-generation architecture compared to current GB300 series
  • High computational power may be excessive for smaller AI models or basic inference tasks
  • Memory capacity constraints for the largest current language models exceeding 70B parameters

Key Features

4th Generation Tensor Cores
Transformer Engine with FP8 precision
Multi-Instance GPU (MIG)
NVIDIA Confidential Computing
NVIDIA Magnum IO
NVLink interconnect
HBM3 memory technology
PCIe Gen5 support

About H100 SXM

The NVIDIA H100 SXM is a datacenter GPU based on the Hopper architecture, representing NVIDIA's previous-generation high-performance computing and AI accelerator before the current GB300 series. This server-class GPU delivers substantial computational power for large-scale AI training and inference workloads, featuring 80GB of HBM3 memory and 14,592 CUDA cores paired with 456 4th Generation Tensor Cores. Key technical specifications include 3,350 GB/s memory bandwidth, 990 TFLOPS of FP16 performance, and support for FP8 precision through the Transformer Engine. The H100 SXM incorporates Multi-Instance GPU (MIG) technology for workload isolation and includes NVIDIA Confidential Computing capabilities for secure processing. With a 700W TDP, this GPU requires robust cooling infrastructure typical of enterprise server deployments. In cloud environments, the H100 SXM serves demanding AI research, large language model training, and high-performance computing applications where substantial memory capacity and computational throughput are essential. The SXM form factor enables high-density server configurations with NVLink interconnect for multi-GPU scaling.

Common Use Cases

The H100 SXM targets large-scale AI training workloads, particularly for language models up to 70 billion parameters where its 80GB memory capacity and high memory bandwidth prove essential. Its 990 TFLOPS FP16 performance and Transformer Engine make it well-suited for training and fine-tuning transformer-based models, while the substantial CUDA core count supports traditional HPC simulations and scientific computing. The MIG capability enables cloud providers to partition the GPU for multiple concurrent workloads, making it valuable for multi-tenant AI inference serving and development environments.

Full Specifications

Hardware

Manufacturer
NVIDIA
Architecture
Hopper
CUDA Cores
14,592
Tensor Cores
456
TDP
700W

Memory & Performance

VRAM
80GB
Memory Bandwidth
3350 GB/s
FP32
67 TFLOPS
FP16
990 TFLOPS
FP64
34 TFLOPS
Release
2022

Frequently Asked Questions

How much does an H100 SXM cost per hour in the cloud?

H100 SXM pricing varies by provider, region, and commitment level. Check the pricing table above for current rates across all providers.

What is the H100 SXM best used for?

The H100 SXM excels at large language model training and inference, particularly for models up to 70 billion parameters. Its 80GB memory capacity and high bandwidth make it suitable for AI research, fine-tuning large models, and high-performance computing workloads requiring substantial memory and computational resources.

How does the H100 SXM compare to the current GB300 series?

The H100 SXM represents NVIDIA's previous-generation Hopper architecture, while GB300 series uses the newer Blackwell Ultra architecture. The H100 SXM offers 80GB HBM3 memory and 990 TFLOPS FP16 performance, whereas newer GB300 GPUs typically provide higher performance and updated architectural features, though specific comparisons depend on the exact GB300 model.