A2 GPU

The NVIDIA A2 is an entry-level data center GPU for efficient AI inference at the edge and in compact servers.

Starting Price
$0.03/hr
Available on 2 cloud providers
A2 GPU

Key Specifications

๐Ÿ’พMemory

16GB VRAM

๐Ÿ—๏ธArchitecture

Ampere

โš™๏ธCompute Units

N/A

๐ŸงฎTensor Cores

40

Technical Specifications

Hardware Details

ManufacturerNVIDIA
ArchitectureAmpere
CUDA Cores1280
Tensor Cores40
RT CoresN/A
Compute UnitsN/A
GenerationN/A

Memory & Performance

VRAM16GB
Memory InterfaceN/A
Memory Bandwidth200 GB/s
FP32 Performance4.5 TFLOPS
FP16 Performance36 TFLOPS
INT8 PerformanceN/A

Performance

Computing Power

CUDA Cores1,280
Tensor Cores40

Computational Performance

FP32 (TFLOPS)4.5
FP16 (TFLOPS)36

Common Use Cases

AI inference, edge computing, smart video analytics

Machine Learning & AI

  • Training large language models and transformers
  • Computer vision and image processing
  • Deep learning model development
  • High-performance inference workloads

Graphics & Compute

  • 3D rendering and visualization
  • Scientific simulations
  • Data center graphics virtualization
  • High-performance computing (HPC)

Market Context

The A2 sits within NVIDIAโ€™s Ampere architecture lineup, positioned in the entry performance tier. It’s designed specifically for data center and enterprise use.

Cloud Availability

Available across 2 cloud providers with prices ranging from $0.03/hr. Pricing and availability may vary by region and provider.

Market Position

Released in 2021, this GPU is positioned for enterprise and data center workloads.

Current Pricing

ProviderHourly PriceSource
RunPod
$0.06/hr
Vast.ai
$0.03/hr

Prices are updated regularly. Last updated: 1/5/2026