A2 GPU
The NVIDIA A2 is an entry-level data center GPU for efficient AI inference at the edge and in compact servers.
Starting Price
$0.03/hr
Available on 2 cloud providers

Key Specifications
๐พMemory
16GB VRAM
๐๏ธArchitecture
Ampere
โ๏ธCompute Units
N/A
๐งฎTensor Cores
40
Technical Specifications
Hardware Details
ManufacturerNVIDIA
ArchitectureAmpere
CUDA Cores1280
Tensor Cores40
RT CoresN/A
Compute UnitsN/A
GenerationN/A
Memory & Performance
VRAM16GB
Memory InterfaceN/A
Memory Bandwidth200 GB/s
FP32 Performance4.5 TFLOPS
FP16 Performance36 TFLOPS
INT8 PerformanceN/A
Performance
Computing Power
CUDA Cores1,280
Tensor Cores40
Computational Performance
FP32 (TFLOPS)4.5
FP16 (TFLOPS)36
Common Use Cases
AI inference, edge computing, smart video analytics
Machine Learning & AI
- Training large language models and transformers
- Computer vision and image processing
- Deep learning model development
- High-performance inference workloads
Graphics & Compute
- 3D rendering and visualization
- Scientific simulations
- Data center graphics virtualization
- High-performance computing (HPC)
Market Context
The A2 sits within NVIDIAโs Ampere architecture lineup, positioned in the entry performance tier. It’s designed specifically for data center and enterprise use.
Cloud Availability
Available across 2 cloud providers with prices ranging from $0.03/hr. Pricing and availability may vary by region and provider.
Market Position
Released in 2021, this GPU is positioned for enterprise and data center workloads.