Verified 2025-09-22 · sourced from OpenAI

100K GPT-5 Chat Tokens — Cost Breakdown

Use this guide to benchmark budgets for 100,000 tokens. Standard pricing is $1.25 per million input tokens and $10.00 per million output tokens. Cached input, when available, reduces costs significantly for repeated contexts.

Scenario breakdown

Cost estimates for different input/output distributions using 100K total tokens.

ScenarioTokens inTokens outStandard costCached cost
Balanced conversation
50% input · 50% output
50,00050,000$0.563$0.506
Input-heavy workflow
80% input · 20% output
80,00020,000$0.300$0.210
Generation heavy
30% input · 70% output
30,00070,000$0.738$0.704
Cached system prompt
90% cached input · 10% fresh output
90,00010,000$0.212$0.111

Workload multipliers

Convert 100K tokens into daily and monthly run-rate budgets.

ProfileRuns/dayTokens/dayDaily costMonthly costCached dailyCached monthly
Single workload1100,000$0.563$16.88$0.111$3.34
Daily batch (10 runs)101,000,000$5.63$168.75$1.11$33.38
Team workload (100 runs)10010,000,000$56.25$1687.50$11.13$333.75

Frequently asked questions

What is the standard cost of 100K GPT-5 Chat tokens?

100K tokens in a 50/50 conversation mix cost roughly $0.563 at the published OpenAI rates.

What happens if cached input is enabled?

With cached contexts, the same 100K tokens drop to $0.506 because input costs fall to $0.125 per million tokens.

How many requests does 100K tokens cover?

If your prompts average around 3,000 tokens per call, 100K total tokens cover about 33 requests.

How fresh is the pricing information?

Prices are taken from https://platform.openai.com/docs/pricing and were last verified on 2025-09-22. models.json keeps this guide in sync with upstream changes.

Related resources