Tesla V100 GPU
The V100 remains relevant for legacy HPC and AI training workloads, with 16GB/32GB HBM2 options and Tensor Cores. While outperformed by newer GPUs, it's still used in scientific computing and older inference pipelines.

Cloud Pricing
Cheapest on Verda — 93% below avgPrices updated daily. Last check: 4/21/2026
Performance
Strengths & Limitations
- 32GB HBM2 memory provides ample capacity for medium-scale AI models and datasets
- 900 GB/s memory bandwidth enables efficient handling of memory-intensive workloads
- 640 Tensor Cores with mixed-precision support accelerate AI training and inference
- NVLink support allows multi-GPU scaling with 300 GB/s bi-directional bandwidth
- 250W TDP offers reasonable power efficiency for data center deployment
- 5,120 CUDA cores provide strong performance for traditional HPC applications
- Established ecosystem support with mature drivers and software optimization
- 12nm manufacturing process is less efficient than newer 4nm and 5nm GPU generations
- Lacks modern features like FP8 precision support found in current Tensor Cores
- Limited to older PCIe 3.0 interface compared to PCIe 5.0 in current GPUs
- 112 TFLOPS FP16 performance significantly trails current-generation alternatives
- Missing hardware support for newer AI model architectures and optimizations
Key Features
About Tesla V100
Common Use Cases
The Tesla V100 is well-suited for entry-level AI development, smaller-scale model training, and traditional HPC workloads. Its 32GB memory capacity handles medium-sized datasets and models that don't require the latest architectural optimizations. The V100 works effectively for AI inference deployments, data science workflows, and scientific computing applications like molecular dynamics simulations. Organizations transitioning from CPU-based computing or exploring GPU acceleration for the first time often find the V100's balance of capability and accessibility appropriate for initial deployments.
Full Specifications
Hardware
- Manufacturer
- NVIDIA
- Architecture
- Volta
- CUDA Cores
- 5,120
- Tensor Cores
- 640
- Process Node
- 12nm
- TDP
- 250W
Memory & Performance
- VRAM
- 32GB
- Memory Interface
- 4096-bit
- Memory Bandwidth
- 900 GB/s
- FP32
- 14 TFLOPS
- FP16
- 112 TFLOPS
- FP64
- 7 TFLOPS
- Release
- 2017
Frequently Asked Questions
How much does a Tesla V100 cost per hour in the cloud?
Tesla V100 pricing varies by provider, region, and commitment level. Check the pricing table above for current rates across all providers.
What is the Tesla V100 best used for?
The Tesla V100 excels at entry-level AI training, inference workloads, and traditional HPC applications. Its 32GB memory and Tensor Core acceleration make it suitable for medium-scale AI development, data science workflows, and scientific computing tasks that don't require cutting-edge performance.
How does the Tesla V100 compare to newer GPUs like the H100?
The V100 offers 112 TFLOPS of FP16 performance compared to the H100's significantly higher throughput, and lacks modern features like FP8 precision and fourth-generation Tensor Cores. However, the V100's 32GB memory capacity and mature ecosystem make it viable for workloads that don't require the latest architectural improvements.