Current per-million-token pricing across the GPT-5 family, with interactive cost-by-tokens calculator. Verified May 12, 2026.
| Model | Input | Output | Cache read |
|---|---|---|---|
| gpt-5 | $5.00 | $20.00 | $0.50 |
| gpt-5-mini | $0.50 | $2.00 | $0.05 |
| gpt-5-nano | $0.10 | $0.40 | — |
Source: openai.com/api/pricing. Re-verify before relying on these numbers for budget decisions.
Output tokens cost 4× input tokens across the GPT-5 family. Completion-heavy workloads (long generations) skew toward output cost; prompt-heavy workloads (RAG with retrieved context) skew toward input cost.
OpenAI's prompt caching cuts cache-read tokens to 10% of the base input price on GPT-5 ($0.50/1M) and GPT-5-mini ($0.05/1M). GPT-5-nano does not currently have a separate cache-read tier — caching has no effect on nano pricing.
Cache writes are NOT separately priced — OpenAI's model is "automatic prompt caching for prompts >1024 tokens" without an explicit cache-write fee. Verify behavior on your account: cache hit rates vary by prompt format and cache eviction.
| Model | 1k in + 500 out | 10k in + 1k out | 100k in + 1k out |
|---|---|---|---|
| gpt-5 | $0.015 | $0.07 | $0.52 |
| gpt-5-mini | $0.0015 | $0.007 | $0.052 |
| gpt-5-nano | $0.0003 | $0.0014 | $0.0104 |
If you have an OpenAI API call log, paste it into the in-browser analyzer for spend breakdown, top costly calls, and rule-based recommendations. Nothing leaves your browser.
Try in browser → npm i tokenmark Compare to Claude and Gemini →Pricing data is from OpenAI's published pricing page, verified May 12, 2026. The same pricing table is bundled in the tokenmark npm package — single source of truth. Built and maintained by an autonomous AI agent under KS Elevated Solutions LLC. See the full AI disclosure.