LLM Pricing
Compare per-token rates across 10 providers and 45 LLM models
Updated daily
๐กTop Picks Right Now
Click to filter tablePrice Trends
Moving average across providers (per 1M tokens) Input Output
Cheapest Option
Llama 3.1 8B
Not enough data yet.
Most Popular
Llama 3.1 405B
Not enough data yet.
Top Performance
Claude Opus 4.5
Not enough data yet.
Showing 15 of 40 models (90 price records)
Understanding LLM Pricing
Input vs Output Tokens
LLM APIs charge separately for input tokens (your prompts) and output tokens (model responses). Output tokens are usually 2-5x more expensive than input tokens.
Context Windows
The context window determines how much text a model can process at once. Larger context windows allow for longer conversations and document analysis.
Open vs Proprietary Models
Open-source models (Llama, Mistral) are often cheaper but may require more tuning. Proprietary models (GPT-4, Claude) typically offer better out-of-box performance.
Batch Discounts
Many providers offer 50% discounts for batch/async API usage. Consider batch APIs for non-time-sensitive workloads to reduce costs.