LLM Pricing

Compare per-token rates across 10 providers and 45 LLM models

Updated daily

๐Ÿ’กTop Picks Right Now

Click to filter table

Price Trends

Moving average across providers (per 1M tokens) Input Output

Cheapest Option
Llama 3.1 8B
Not enough data yet.
Most Popular
Llama 3.1 405B
Not enough data yet.
Top Performance
Claude Opus 4.5
Not enough data yet.
Showing 15 of 40 models (90 price records)

Understanding LLM Pricing

Input vs Output Tokens

LLM APIs charge separately for input tokens (your prompts) and output tokens (model responses). Output tokens are usually 2-5x more expensive than input tokens.

Context Windows

The context window determines how much text a model can process at once. Larger context windows allow for longer conversations and document analysis.

Open vs Proprietary Models

Open-source models (Llama, Mistral) are often cheaper but may require more tuning. Proprietary models (GPT-4, Claude) typically offer better out-of-box performance.

Batch Discounts

Many providers offer 50% discounts for batch/async API usage. Consider batch APIs for non-time-sensitive workloads to reduce costs.