highData Center

A30 GPU

The NVIDIA A30 is a versatile data center GPU for mainstream AI inference, training, and HPC workloads with Multi-Instance GPU support.

VRAM 24GB
CUDA Cores 3,584
Tensor Cores 224
TDP 165W
From
$0.25/hr
across 4 providers
A30 GPU

Cloud Pricing

Cheapest on Seeweb 48% below avg
ProviderGPUsPrice / hrUpdatedSource
1× GPU
6mo$0.25
3/30/2026
1× GPU
3mo$0.26
3/30/2026
1× GPU
1mo$0.28
3/30/2026
1× GPU
$0.34
4/8/2026
1× GPU
$0.35
4/8/2026
2× GPU
$0.35
4/8/2026
4× GPU
$0.35
4/8/2026
8× GPU
$0.35
4/8/2026
1× GPU
$0.39
4/8/2026
2× GPU
$0.39
4/8/2026
4× GPU
$0.39
4/8/2026
8× GPU
$0.39
4/8/2026
1× GPU
$1.32
4/6/2026
2× GPU
$1.32
4/6/2026
Direct from providerVia marketplace

Prices updated daily. Last check: 4/8/2026

Performance

FP16
165 TFLOPS
FP32
10.3 TFLOPS
Bandwidth
933 GB/s

Strengths & Limitations

  • 24GB HBM2 memory capacity supports large model inference and training
  • Multi-Instance GPU (MIG) enables partitioning into up to 4 isolated instances
  • 165W TDP allows for dense server configurations
  • Third-generation Tensor Cores provide optimized mixed-precision AI performance
  • Third-generation NVLink connectivity at 200 GB/s for multi-GPU scaling
  • 933 GB/s memory bandwidth supports memory-intensive workloads
  • PCIe Gen4 interface provides improved host connectivity over previous generations
  • 165W power consumption may be high for edge deployment scenarios
  • Limited to Ampere architecture features, lacking newer Hopper or Blackwell capabilities
  • 10.3 TFLOPS FP32 performance may be insufficient for large-scale HPC workloads
  • Released in 2021, representing previous-generation technology compared to current offerings
  • Dual-slot form factor reduces server density compared to single-slot alternatives

Key Features

Multi-Instance GPU (MIG) with up to 4 instances
Third-generation Tensor Cores
NVIDIA Ampere architecture
Third-generation NVLink
24GB HBM2 memory
PCIe Gen4 interface
Mixed-precision compute support
Hardware-accelerated video encoding/decoding

About A30

The NVIDIA A30 is a data center GPU based on the Ampere architecture, positioned as a mid-range accelerator in NVIDIA's server GPU lineup. Released in 2021, the A30 serves as a versatile option between entry-level and flagship data center cards, offering substantial compute capabilities while maintaining relatively modest power requirements. The A30 features 24GB of HBM2 memory with 933 GB/s of memory bandwidth, 3584 CUDA cores, and 224 third-generation Tensor Cores optimized for mixed-precision AI workloads. Key technical specifications include 165 TFLOPS of FP16 performance and 10.3 TFLOPS of FP32 performance, delivered within a 165W TDP envelope. The A30 supports Multi-Instance GPU (MIG) technology, allowing a single card to be partitioned into up to four isolated GPU instances for improved resource utilization. Third-generation NVLink provides 200 GB/s of GPU-to-GPU bandwidth, while PCIe Gen4 connectivity offers 64 GB/s host interface bandwidth. The dual-slot form factor makes it suitable for dense server deployments. In cloud environments, the A30 typically handles AI training for medium-scale models, AI inference workloads requiring substantial memory, and HPC applications that benefit from GPU acceleration. The combination of 24GB memory capacity and MIG support makes it particularly suitable for multi-tenant cloud deployments where resource sharing and isolation are important considerations.

Common Use Cases

The A30 is well-suited for AI inference serving that requires substantial memory capacity, medium-scale AI training workloads, and HPC applications that can leverage GPU acceleration. Its 24GB memory makes it capable of handling large language models and computer vision tasks that exceed the memory limits of smaller GPUs. The MIG functionality makes it particularly valuable in multi-tenant cloud environments where GPU resources need to be shared among multiple users or applications while maintaining isolation. Data analytics workloads involving large datasets benefit from the combination of memory capacity and compute performance.

Full Specifications

Hardware

Manufacturer
NVIDIA
Architecture
Ampere
CUDA Cores
3,584
Tensor Cores
224
TDP
165W

Memory & Performance

VRAM
24GB
Memory Bandwidth
933 GB/s
FP32
10.3 TFLOPS
FP16
165 TFLOPS
FP64
5.2 TFLOPS
Release
2021

Frequently Asked Questions

How much does an A30 cost per hour in the cloud?

A30 pricing varies by provider, region, and commitment level. Check the pricing table above for current rates across all providers.

What is the A30 best used for?

The A30 excels at AI inference workloads requiring substantial memory, medium-scale AI training, and HPC applications. Its 24GB memory capacity and MIG support make it particularly suitable for serving large models and multi-tenant deployments where resource isolation is important.

How does the A30 compare to the V100 for AI workloads?

The A30 offers up to 3x higher throughput than the V100 for AI training and inference tasks, primarily due to its third-generation Tensor Cores and Ampere architecture optimizations. The A30 also provides MIG functionality and 24GB of memory compared to the V100's 16GB or 32GB options, though the V100 offers higher memory bandwidth in its HBM2 configuration.