Verified 2025-09-22 · sourced from OpenAI
o4-mini Token Calculator & Cost Guide
Estimate OpenAI o4-mini API usage in dollars before you send a single request. Standard pricing is $1.10 per million input tokens and $4.40 per million output tokens with a 200K token context window.
Context window
200,000 tokens
Input price
$1.10 / 1M
Output price
$4.40 / 1M
Cached input
Not published
Usage scenarios
Compare standard and cached pricing (where available) across common workloads.
| Scenario | Tokens in | Tokens out | Total tokens | Standard cost |
|---|---|---|---|---|
Quick chat reply Single user question with a short assistant answer | 650 | 220 | 870 | $0.0017 |
Coding assistant session Multi-turn pair programming exchange (≈6 turns) | 2,600 | 1,400 | 4,000 | $0.0090 |
Knowledge base response Retrieval-augmented answer referencing multiple passages | 12,000 | 3,000 | 15,000 | $0.0264 |
Near-max context run Large document processing approaching the 200K token limit | 176,000 | 24,000 | 200,000 | $0.299 |
Daily & monthly budgeting
Translate usage into predictable operating expenses across popular deployment sizes.
| Profile | Requests/day | Tokens/day | Daily cost | Monthly cost |
|---|---|---|---|---|
| Team pilot | 25 | 75,000 | $0.165 | $4.95 |
| Product launch | 100 | 500,000 | $1.04 | $31.35 |
| Enterprise scale | 500 | 3,000,000 | $6.60 | $198.00 |
Frequently asked questions
How much does o4-mini cost per 1,000 tokens?
At the published rates of $1.10 per million input tokens and $4.40 per million output tokens, a typical 1,000 token request (≈70% input, 30% output) costs about $0.0021.
What is the context window for o4-mini?
o4-mini supports up to 200,000 tokens (200K), allowing large prompts and retrieval-augmented payloads in a single call.
How fresh is the o4-mini pricing data?
Pricing is sourced from https://platform.openai.com/docs/pricing and was last verified on 2025-09-22. The calculator updates automatically when models.json is refreshed.
Does o4-mini have a 200K token context window?
Yes, o4-mini features a 200,000 token (200K) context window, making it suitable for processing long documents, extensive code bases, or large retrieval-augmented generation (RAG) workflows. This 200K context allows you to include substantial background information in a single API call.
How does o4-mini compare to GPT-4o mini?
o4-mini is positioned as an efficient reasoning model with a 200K context window, while GPT-4o mini (128K context) focuses on general-purpose tasks. o4-mini pricing is $1.10 input / $4.40 output per million tokens. Choose o4-mini for reasoning-heavy tasks requiring large context, or GPT-4o mini for faster general-purpose applications.