LLM Pricing

Compare per-token rates across 16 providers and 171 LLM models

Updated daily

💡Top Picks Right Now

Click to filter table

Price Trends

Moving average across providers (per 1M tokens) Input Output

Cheapest Option
Qwen 3.5 27B
$0.216 / $1.12per 1M tokens
Providers: 3
Most Popular
GPT-OSS-120B
$0.124 / $0.503per 1M tokens
Providers: 7
Top Performance
Claude Opus 4.5
$3.81 / $19.05per 1M tokens
Providers: 2
Pricing
Showing 15 of 168 models (403 price records)

Understanding LLM Pricing

Input vs Output Tokens

LLM APIs charge separately for input tokens (your prompts) and output tokens (model responses). Output tokens are usually 2-5x more expensive than input tokens.

Context Windows

The context window determines how much text a model can process at once. Larger context windows allow for longer conversations and document analysis.

Open vs Proprietary Models

Open-source models (Llama, Mistral) are often cheaper but may require more tuning. Proprietary models (GPT-4, Claude) typically offer better out-of-box performance.

Batch Discounts

Many providers offer 50% discounts for batch/async API usage. Consider batch APIs for non-time-sensitive workloads to reduce costs.