Verified 2025-09-22 · sourced from OpenAI

100K GPT-4o Tokens — Cost Breakdown

Use this guide to benchmark budgets for 100,000 tokens. Standard pricing is $2.50 per million input tokens and $10.00 per million output tokens. Cached input, when available, reduces costs significantly for repeated contexts.

Scenario breakdown

Cost estimates for different input/output distributions using 100K total tokens.

ScenarioTokens inTokens outStandard costCached cost
Balanced conversation
50% input · 50% output
50,00050,000$0.625$0.563
Input-heavy workflow
80% input · 20% output
80,00020,000$0.400$0.300
Generation heavy
30% input · 70% output
30,00070,000$0.775$0.738
Cached system prompt
90% cached input · 10% fresh output
90,00010,000$0.325$0.212

Workload multipliers

Convert 100K tokens into daily and monthly run-rate budgets.

ProfileRuns/dayTokens/dayDaily costMonthly costCached dailyCached monthly
Single workload1100,000$0.625$18.75$0.212$6.38
Daily batch (10 runs)101,000,000$6.25$187.50$2.13$63.75
Team workload (100 runs)10010,000,000$62.50$1875.00$21.25$637.50

Frequently asked questions

What is the standard cost of 100K GPT-4o tokens?

100K tokens in a 50/50 conversation mix cost roughly $0.625 at the published OpenAI rates.

What happens if cached input is enabled?

With cached contexts, the same 100K tokens drop to $0.563 because input costs fall to $1.250 per million tokens.

How many requests does 100K tokens cover?

If your prompts average around 3,000 tokens per call, 100K total tokens cover about 33 requests.

How fresh is the pricing information?

Prices are taken from https://platform.openai.com/docs/pricing and were last verified on 2025-09-22. models.json keeps this guide in sync with upstream changes.

Related resources