entry

Tesla T4 GPU

The T4 is a low-power data center GPU optimized for AI inference, offering mixed-precision performance in a compact design. It's commonly used in cloud deployments for cost-efficient scaling of NLP and recommendation systems.

VRAM 16GB
CUDA Cores 2,560
Tensor Cores 320
TDP 70W
Process 12nm
From
$0.16/hr
across 2 providers
Tesla T4 GPU

Cloud Pricing

Cheapest on Google Cloud 70% below avg
ProviderGPUsPrice / hrUpdatedSource
1× GPU
3mo$0.16
3/31/2026
1× GPU
1mo$0.22
3/31/2026
1× GPU
$0.35
3/31/2026
1× GPU
$0.53
4/8/2026
4× GPU
$0.98
4/8/2026
8× GPU
$0.98
4/8/2026
Direct from providerVia marketplace

Prices updated daily. Last check: 4/8/2026

Performance

FP16
65 TFLOPS
FP32
8.1 TFLOPS
INT8
130 TOPS
Bandwidth
320 GB/s

Strengths & Limitations

  • Low 70-watt TDP enables deployment in power-constrained environments
  • 16 GB GDDR6 memory provides adequate capacity for moderate model sizes
  • Turing Tensor Cores support multi-precision computing including FP16 and INT8
  • Dedicated hardware transcoding engines handle video processing workloads
  • Single-slot PCIe form factor fits in space-constrained server configurations
  • 320 GB/s memory bandwidth supports inference workloads efficiently
  • 65 TFLOPS FP16 performance accelerates mixed-precision AI tasks
  • Turing architecture lacks modern features found in Hopper and Blackwell generations
  • 8.1 TFLOPS FP32 performance insufficient for large-scale training workloads
  • 320 tensor cores provide limited throughput for transformer-based models
  • No NVLink support restricts multi-GPU scaling capabilities
  • PCIe Gen3 x16 interface may bottleneck high-bandwidth applications

Key Features

Turing Tensor Cores
Multi-precision computing
Dedicated hardware transcoding engines
GDDR6 memory subsystem
PCIe Gen3 x16 interface
NVIDIA Turing architecture
Single-slot thermal design
Hardware video decode acceleration

About Tesla T4

The Tesla T4 is NVIDIA's entry-level data center GPU based on the Turing architecture, released in 2018. As an earlier generation product in NVIDIA's data center lineup, it occupies a budget-friendly position below current Hopper and Blackwell Ultra offerings. The T4 features 16 GB of GDDR6 memory and is designed for moderate AI inference workloads where power efficiency is prioritized over raw performance. The Tesla T4 delivers 8.1 TFLOPS of FP32 performance and 65 TFLOPS of FP16 performance through its 2560 CUDA cores and 320 Turing Tensor Cores. With 320 GB/s of memory bandwidth and a 70-watt TDP, it provides a balance of inference capability and power efficiency in a single-slot PCIe form factor. The dedicated hardware transcoding engines enable video processing workloads alongside AI inference tasks. In cloud deployments, the T4 serves as an accessible entry point for organizations testing AI workloads or running inference applications that don't require the computational density of newer GPU generations. Its low power consumption makes it suitable for edge computing scenarios and cost-conscious deployments where moderate AI acceleration is sufficient.

Common Use Cases

The Tesla T4 is well-suited for AI inference workloads that require moderate computational power, such as computer vision applications, natural language processing with smaller models, and real-time recommendation systems. Its dedicated transcoding engines make it effective for video analytics pipelines that combine AI inference with media processing. The 16 GB memory capacity accommodates models up to medium complexity, while the 70-watt power envelope enables deployment in edge computing scenarios or data centers with strict power budgets. Organizations using the T4 typically run inference-focused workloads rather than training, leveraging its multi-precision capabilities for INT8 and FP16 optimized models.

Full Specifications

Hardware

Manufacturer
NVIDIA
Architecture
Turing
CUDA Cores
2,560
Tensor Cores
320
RT Cores
40
Process Node
12nm
TDP
70W

Memory & Performance

VRAM
16GB
Memory Interface
256-bit
Memory Bandwidth
320 GB/s
FP32
8.1 TFLOPS
FP16
65 TFLOPS
FP64
0.25 TFLOPS
INT8
130 TOPS
Release
2018

Frequently Asked Questions

How much does a Tesla T4 cost per hour in the cloud?

Tesla T4 pricing varies by provider, region, and commitment level. Check the pricing table above for current rates across all providers.

What is the Tesla T4 best used for?

The Tesla T4 excels at AI inference workloads, particularly computer vision, natural language processing with smaller models, and video analytics. Its 70-watt power efficiency makes it ideal for edge computing and cost-conscious deployments where moderate AI acceleration is sufficient.

How does the Tesla T4 compare to newer NVIDIA data center GPUs?

The T4's Turing architecture lacks modern features found in Hopper and Blackwell generations, such as advanced tensor formats and higher memory bandwidth. However, its 70-watt TDP and 16 GB memory make it more power-efficient and cost-effective for inference workloads that don't require the computational density of H100 or GB200 series GPUs.