Verified 2025-09-22 · sourced from xAI

100K Grok 4 Tokens — Cost Breakdown

Use this guide to benchmark budgets for 100,000 tokens. Standard pricing is $3.00 per million input tokens and $15.00 per million output tokens. Cached input, when available, reduces costs significantly for repeated contexts.

Scenario breakdown

Cost estimates for different input/output distributions using 100K total tokens.

ScenarioTokens inTokens outStandard costCached cost
Balanced conversation
50% input · 50% output
50,00050,000$0.900$0.787
Input-heavy workflow
80% input · 20% output
80,00020,000$0.540$0.360
Generation heavy
30% input · 70% output
30,00070,000$1.14$1.07
Cached system prompt
90% cached input · 10% fresh output
90,00010,000$0.420$0.217

Workload multipliers

Convert 100K tokens into daily and monthly run-rate budgets.

ProfileRuns/dayTokens/dayDaily costMonthly costCached dailyCached monthly
Single workload1100,000$0.900$27.00$0.217$6.53
Daily batch (10 runs)101,000,000$9.00$270.00$2.17$65.25
Team workload (100 runs)10010,000,000$90.00$2700.00$21.75$652.50

Frequently asked questions

What is the standard cost of 100K Grok 4 tokens?

100K tokens in a 50/50 conversation mix cost roughly $0.900 at the published xAI rates.

What happens if cached input is enabled?

With cached contexts, the same 100K tokens drop to $0.787 because input costs fall to $0.750 per million tokens.

How many requests does 100K tokens cover?

If your prompts average around 3,000 tokens per call, 100K total tokens cover about 33 requests.

How fresh is the pricing information?

Prices are taken from https://docs.x.ai/docs/models and were last verified on 2025-09-22. models.json keeps this guide in sync with upstream changes.

Related resources