Fireworks AI logo

Fireworks AI

The fastest platform for open-source AI

Inference specialist🇺🇸 USinferenceopen-sourcefast

Last reviewed Mar 14, 2026

Fireworks AI is a high-performance inference platform specializing in open-source generative AI models. Founded by former Meta PyTorch team members, they provide the fastest inference engine for LLMs, vision, and audio models with enterprise-grade reliability.

14
LLM Models
$0.07
From / 1M input

LLM API Pricing

Pay-per-token pricing. Prices shown per 1M tokens.

Prices last updated: April 27, 2026

ModelCreatorContextInput/1MOutput/1MUpdated
OpenAI128K$0.070$0.3004/27/2026
OpenAI128K$0.150$0.6004/27/2026
Alibaba131K$0.150$0.6004/27/2026
Alibaba41K$0.200$0.00004/27/2026
MiniMax197K$0.300$1.204/16/2026
MiniMax205K$0.300$1.204/27/2026
MiniMax128K$0.300$1.204/27/2026
Alibaba1.0M$0.500$3.004/27/2026
DeepSeek33K$0.560$1.684/27/2026
DeepSeek64K$0.560$1.684/27/2026

Pros & Cons

Advantages

  • Lightning-fast inference with industry-leading response times
  • Easy-to-use API with excellent OpenAI compatibility
  • Wide variety of optimized open-source models
  • Competitive pricing with 50% off cached tokens and batch processing
  • Enterprise reliability with 99.99% uptime SLA
  • Up to 100 fine-tuned models deployable without extra costs

Limitations

  • Limited capacity with some serverless model limits
  • Primarily focused on language models over image/video generation
  • BYOC only available for major enterprise customers
  • Feature-rich interface can have steep learning curve

Key Features

400+ Open-Source Models

Instant access to Llama, DeepSeek, Qwen, Mixtral, FLUX, Whisper, and more

Blazing Fast Inference

Industry-leading throughput and latency processing 140B+ tokens daily

Fine-Tuning Suite

SFT, DPO, and reinforcement fine-tuning with LoRA efficiency

OpenAI-Compatible API

Drop-in replacement for easy migration from OpenAI

On-Demand GPUs

A100, H100, H200, and B200 deployments with per-second billing

Batch Processing

50% discount for async bulk inference workloads

Pricing Options

OptionDetails
Serverless pay-per-tokenToken-based pricing for small and large models with transparent per-million token rates
Cached tokens50% discount on cached input tokens
Batch processing50% discount on async bulk inference
On-demand GPUsPer-second billing for A100, H100, H200, and B200 GPU deployments

Availability & Support

Regions

18+ global regions across 8 cloud providers with multi-region deployments and BYOC support for enterprise

Support

Documentation, Discord community, status page, email support, and dedicated enterprise support with SLAs

Getting Started

  1. 1

    Explore Model Library

    Browse 400+ models at fireworks.ai/models

  2. 2

    Test in Playground

    Experiment with prompts interactively without coding

  3. 3

    Generate API Key

    Create an API key from user settings in your account

  4. 4

    Make first API call

    Use OpenAI-compatible endpoints or Fireworks SDK

  5. 5

    Scale to production

    Transition to on-demand GPU deployments for production workloads