LightweightxAI

Grok 3 Mini

Grok 3 Mini is xAI's lightweight model designed for fast, cost-effective text generation with a 131K token context window.

Context 131K
Tier Lightweight
Input from
$0.300 / 1M tokens
across 1 provider

API Pricing

ProviderInput / 1MOutput / 1MSpeedTTFTUpdated
$0.300$0.500197 t/s358ms4/14/2026

Prices updated daily. Last check: 4/14/2026

Model Details

General

Creator
xAI
Family
Grok
Tier
Lightweight
Context Window
131K
Modalities
Text

Capabilities

Tool Calling
No
Open Source
No

Strengths & Limitations

  • Fast inference speed at 197.41 tokens per second
  • Quick response initiation with 358ms time to first token
  • Large context window of 131K tokens for its lightweight tier
  • Cost-effective option within the Grok model family
  • Optimized for high-throughput applications
  • No tool calling or function execution capabilities
  • Text-only modality with no image or multimodal support
  • Proprietary model with no open-source availability
  • Limited capabilities compared to larger Grok models
  • Fewer advanced features than flagship alternatives

Key Features

131,072 token context window
Text input and output processing
Fast inference with 197+ tokens per second
Sub-400ms time to first token
Streaming response generation
API access through multiple providers

About Grok 3 Mini

Grok 3 Mini is xAI's lightweight model in the Grok family, positioned as a cost-effective option for high-volume text processing tasks. As part of xAI's model lineup, it sits below the flagship Grok models while maintaining competitive performance for applications where speed and efficiency are prioritized over maximum capability. The model features a 131,072 token context window and supports text-only input and output. Performance benchmarks show the model generates approximately 197 tokens per second with a time to first token of 358 milliseconds, indicating strong throughput characteristics for its tier. The model does not include tool calling capabilities, focusing instead on core text generation tasks. Grok 3 Mini targets use cases where developers need reliable text processing at scale without the computational overhead of larger models. Its combination of reasonable context length and fast inference makes it suitable for applications requiring quick responses across moderate-length documents or conversations.

Common Use Cases

Grok 3 Mini is well-suited for high-volume text processing applications where speed and cost efficiency are priorities. Its fast inference speed and reasonable context window make it effective for content generation, document summarization, customer service chatbots, and batch text processing tasks. The model works well for applications requiring quick responses to moderate-length inputs, such as email drafting, content moderation, or simple question-answering systems. Organizations looking to process large volumes of text data without the expense of flagship models will find Grok 3 Mini appropriate for tasks like automated content creation, text classification at scale, or powering user-facing applications where response time is critical.

Frequently Asked Questions

How much does Grok 3 Mini cost per million tokens?

Grok 3 Mini pricing varies by provider and may include different rates for input and output tokens. Check the pricing table above for current rates across all available providers offering this model.

What is Grok 3 Mini best used for?

Grok 3 Mini excels at high-volume text processing tasks where speed and cost efficiency are important. It's particularly effective for content generation, document processing, customer service applications, and batch text operations that benefit from its fast 197+ tokens per second inference speed and 131K context window.

Does Grok 3 Mini support function calling or multimodal inputs?

No, Grok 3 Mini is focused on text-only processing and does not support tool calling, function execution, or multimodal inputs like images. It's designed specifically for efficient text generation and understanding tasks.