LLM Pricing
Compare per-token rates across 16 providers and 171 LLM models
Updated daily
๐กTop Picks Right Now
Click to filter tablePrice Trends
Moving average across providers (per 1M tokens) Input Output
Cheapest Option
Granite 4.0 H Micro
Not enough data yet.
Most Popular
GPT-OSS-120B
Not enough data yet.
Top Performance
Claude Opus 4.5
Not enough data yet.
Showing 1 of 1 models (6 price records) โ filteredLlama 3.1 8B
Understanding LLM Pricing
Input vs Output Tokens
LLM APIs charge separately for input tokens (your prompts) and output tokens (model responses). Output tokens are usually 2-5x more expensive than input tokens.
Context Windows
The context window determines how much text a model can process at once. Larger context windows allow for longer conversations and document analysis.
Open vs Proprietary Models
Open-source models (Llama, Mistral) are often cheaper but may require more tuning. Proprietary models (GPT-4, Claude) typically offer better out-of-box performance.
Batch Discounts
Many providers offer 50% discounts for batch/async API usage. Consider batch APIs for non-time-sensitive workloads to reduce costs.