A100 PCIE GPU
The A100 PCIe delivers strong AI training and HPC performance in a standard PCIe form factor, supporting multi-instance GPU partitioning for flexible workload sharing. It enables large-scale AI and scientific computing with improved efficiency over older GPUs.

Cloud Pricing
Cheapest on Verda — 75% below avgPrices updated daily. Last check: 4/21/2026
Performance
Strengths & Limitations
- 40GB HBM2e memory capacity supports large model training and inference workloads
- Multi-Instance GPU (MIG) technology enables partitioning into up to seven isolated GPU instances
- Third-generation Tensor Cores with TF32, BF16, FP16, and INT8 precision support
- 1,555 GB/s memory bandwidth facilitates memory-intensive compute operations
- PCIe Gen4 form factor provides compatibility with standard server architectures
- 312 TFLOPS FP16 performance for mixed-precision AI training
- 250W TDP offers power efficiency relative to compute capability
- Limited to PCIe Gen4 bandwidth compared to NVLink connectivity in SXM variants
- 250W power consumption requires adequate cooling and power infrastructure
- Previous-generation architecture compared to current H100 and GB300 offerings
- 40GB memory capacity may be insufficient for the largest language models
- Higher cost per instance compared to consumer GPUs for development workloads
Key Features
About A100 PCIE
Common Use Cases
The A100 PCIe is suited for AI training workloads requiring substantial memory capacity, particularly for natural language processing models, computer vision training, and recommendation systems that benefit from the 40GB memory buffer. Its Multi-Instance GPU capability makes it effective for inference serving scenarios where multiple smaller models can run simultaneously on partitioned resources. High-performance computing applications including scientific simulations, computational fluid dynamics, and molecular modeling leverage its FP32 and FP64 compute capabilities, while the PCIe form factor ensures compatibility with existing data center infrastructure without requiring specialized NVLink fabric investments.
Full Specifications
Hardware
- Manufacturer
- NVIDIA
- Architecture
- Ampere
- CUDA Cores
- 6,912
- Tensor Cores
- 432
- RT Cores
- 0
- Process Node
- 7nm
- TDP
- 250W
Memory & Performance
- VRAM
- 40GB
- Memory Interface
- 5120-bit
- Memory Bandwidth
- 1555 GB/s
- FP32
- 19.5 TFLOPS
- FP16
- 312 TFLOPS
- BF16
- 312 TFLOPS
- FP64
- 9.7 TFLOPS
- INT8
- 624 TOPS
- Release
- 2020
Frequently Asked Questions
How much does an A100 PCIe cost per hour in the cloud?
A100 PCIe pricing varies by provider, region, and commitment level. Check the pricing table above for current rates across all providers.
What is the A100 PCIe best used for?
The A100 PCIe excels at AI model training and inference for medium to large-scale models, particularly those requiring substantial memory capacity up to 40GB. Its Multi-Instance GPU capability makes it effective for serving multiple inference workloads simultaneously, while its compute performance supports scientific computing and high-performance computing applications.
How does the A100 PCIe compare to the A100 SXM variant?
The A100 PCIe operates at 250W TDP compared to the SXM's 400W, resulting in lower peak performance but better power efficiency. The PCIe variant uses standard PCIe Gen4 connectivity (64 GB/s) instead of NVLink (600 GB/s), limiting multi-GPU bandwidth but providing broader server compatibility. Both variants share the same 40GB HBM2e memory capacity and Ampere architecture features.