LLM Pricing
Compare per-token rates across 16 providers and 171 LLM models
Updated daily
💡Top Picks Right Now
Click to filter tablePrice Trends
Moving average across providers (per 1M tokens) Input Output
Cheapest Option
Qwen 3.5 27B
$0.216 / $1.12per 1M tokens
Providers: 3
Most Popular
GPT-OSS-120B
$0.124 / $0.503per 1M tokens
Providers: 7
Top Performance
Claude Opus 4.5
$3.81 / $19.05per 1M tokens
Providers: 2
Pricing
Showing 15 of 168 models (403 price records)
Understanding LLM Pricing
Input vs Output Tokens
LLM APIs charge separately for input tokens (your prompts) and output tokens (model responses). Output tokens are usually 2-5x more expensive than input tokens.
Context Windows
The context window determines how much text a model can process at once. Larger context windows allow for longer conversations and document analysis.
Open vs Proprietary Models
Open-source models (Llama, Mistral) are often cheaper but may require more tuning. Proprietary models (GPT-4, Claude) typically offer better out-of-box performance.
Batch Discounts
Many providers offer 50% discounts for batch/async API usage. Consider batch APIs for non-time-sensitive workloads to reduce costs.