ultraData Center

B100 GPU

The NVIDIA B100 is a Blackwell architecture GPU designed for data center AI and HPC workloads, offering high performance with lower power consumption than B200.

VRAM 192GB
TDP 700W
Contact providers for pricing
B100 GPU

Cloud Pricing

No pricing data available for this GPU at the moment.

Prices updated daily. Last check: 4/8/2026

Performance

FP16
3500 TFLOPS
FP32
60 TFLOPS
BF16
1750 TFLOPS
FP8
3500 TFLOPS
Bandwidth
8000 GB/s

Strengths & Limitations

  • 192 GB VRAM capacity supports large model training and inference workloads
  • 8,000 GB/s memory bandwidth enables high-throughput data processing
  • Second-Generation Transformer Engine with NVFP4 quantization optimizes AI model performance
  • Fifth-generation NVLink supports scaling up to 576 GPUs with 130TB/s aggregate bandwidth
  • Ultra Tensor Cores deliver 3,500 TFLOPS FP16 performance for AI computations
  • NVIDIA Confidential Computing provides security features for sensitive workloads
  • Dedicated RAS Engine enhances system reliability in data center deployments
  • 700W TDP requires substantial power infrastructure and cooling systems
  • High memory capacity may be excessive for smaller AI models and inference workloads
  • Server-only form factor limits deployment to data center environments
  • Blackwell architecture ecosystem may have limited software optimization compared to mature platforms

Key Features

Second-Generation Transformer Engine
Ultra Tensor Cores
NVFP4 quantization
Fifth-generation NVLink
NVIDIA Confidential Computing
RAS Engine
Decompression Engine
NVLink Switch compatibility

About B100

The NVIDIA B100 is a data center GPU built on the Blackwell architecture, positioned as part of NVIDIA's current-generation lineup for enterprise AI and high-performance computing workloads. With 192 GB of high-bandwidth memory and manufactured using TSMC's 4NP process, the B100 delivers substantial compute capabilities for large-scale AI deployments and data analytics applications. The B100 features 192 GB of VRAM with 8,000 GB/s memory bandwidth, delivering 3,500 TFLOPS of FP16 performance for AI workloads. Key technical differentiators include Ultra Tensor Cores with Second-Generation Transformer Engine support, NVFP4 quantization capabilities, and fifth-generation NVLink connectivity that enables scaling up to 576 GPUs in a single domain. The GPU operates with a 700W TDP and incorporates a dedicated RAS Engine for system reliability in enterprise deployments. In cloud environments, the B100 targets large language model training, generative AI applications, and high-throughput inference workloads that require substantial memory capacity and compute performance. The combination of large VRAM capacity and advanced interconnect makes it suitable for distributed training scenarios and memory-intensive AI applications that exceed the capabilities of consumer-grade hardware.

Common Use Cases

The B100 is designed for enterprise AI workloads requiring substantial memory capacity and compute performance. Its 192 GB VRAM makes it suitable for training and serving large language models, particularly trillion-parameter models that exceed the memory constraints of smaller GPUs. The high memory bandwidth and Ultra Tensor Cores optimize performance for generative AI applications, while NVLink scaling capabilities support distributed training across multiple nodes. The GPU also serves data analytics workloads that benefit from large memory capacity and the dedicated Decompression Engine for data processing acceleration.

Full Specifications

Hardware

Manufacturer
NVIDIA
Architecture
Blackwell
TDP
700W

Memory & Performance

VRAM
192GB
Memory Bandwidth
8000 GB/s
FP32
60 TFLOPS
FP16
3500 TFLOPS
BF16
1750 TFLOPS
FP8
3500 TFLOPS
Release
2024

Frequently Asked Questions

How much does a B100 cost per hour in the cloud?

B100 pricing varies by provider, region, and commitment level. Check the pricing table above for current rates across all providers.

What is the B100 best used for?

The B100 excels at large language model training and inference, particularly for models requiring substantial memory capacity. Its 192 GB VRAM and Second-Generation Transformer Engine make it well-suited for generative AI applications, trillion-parameter model serving, and distributed training workloads that benefit from NVLink scaling capabilities.

How does the B100 compare to the H100 for AI workloads?

The B100 features newer Blackwell architecture with Ultra Tensor Cores and Second-Generation Transformer Engine compared to the H100's Hopper architecture. The B100 provides 192 GB VRAM versus the H100's 80 GB, offering significantly more memory for large model training. Both support similar NVLink scaling, but the B100 includes NVFP4 quantization and enhanced Transformer Engine capabilities for improved AI performance.