GPT-5.5 Cost & Pricing (May 2026)
Per-1M-token rates for every GPT-5.5 tier — standard, batch, priority, and the GPT-5.5 Pro variant. Cost-per-task estimates for typical workloads and a like-for-like comparison vs Claude Opus 4.7, Gemini 3.1 Pro, and DeepSeek V4 Pro.
Pricing tiers — every way to buy GPT-5.5
| Tier | Input /1M | Output /1M | Notes |
|---|---|---|---|
| Standard (sync) | $5.00 | $30.00 | List price for the synchronous Responses / Chat Completions API. |
| Batch (24h SLA) | $2.50 | $15.00 | 50% off list for asynchronous workloads via the OpenAI Batch API. Output cap at 32K tokens still applies. |
| Priority (low-latency) | $7.50 | $45.00 | +50% surcharge on list for latency-sensitive workloads with reserved capacity. Useful for voice and real-time UX where TTFT matters. |
| GPT-5.5 Pro (variant) | $30.00 | $180.00 | High-compute thinking variant. 6x the price for mid-teens AAII uplift on hardest reasoning. Most teams should ignore it for general use. |
All prices in USD per 1M tokens. The +50% priority surcharge is the most controversial line on this card — consider it a latency insurance premium, not a quality differentiator.
Cost per task — what you actually pay
| Task | In tok | Out tok | Standard | Batch | Priority |
|---|---|---|---|---|---|
| Short chat reply | 800 | 200 | $0.0100 | $0.0050 | $0.0150 |
| Long-doc summary (50-page PDF) | 80,000 | 1,500 | $0.4450 | $0.2225 | $0.6675 |
| Agentic loop (12 tool turns) | 45,000 | 6,000 | $0.4050 | $0.2025 | $0.6075 |
| RAG query (10-doc context) | 12,000 | 600 | $0.0780 | $0.0390 | $0.1170 |
Cost per single invocation. Prompt caching can cut input costs ~30% for prefixed workloads but is less aggressive than Anthropic's 90%-off cached tier.
GPT-5.5 vs nearest alternatives
| Model | In /1M | Out /1M | Context | Note |
|---|---|---|---|---|
| GPT-5.5 | $5.00 | $30.00 | 1M | This page. AAII 59, Arena 1481. Strong all-rounder. |
| Claude Opus 4.7 | $5.00 | $25.00 | 1M | Same input, 17% cheaper output. Coding Arena #1 — better default for engineering work. |
| Gemini 3.1 Pro | $3.50 | $10.50 | 2M | 30% cheaper input, 65% cheaper output. Better for long-context and science. |
| DeepSeek V4 Pro | $1.74 | $3.48 | 1M | ~9x cheaper output. Apache 2.0 — also self-hostable. |
| GPT-5.5 Pro | $30.00 | $180.00 | 1M | 6x the price for mid-teens AAII uplift. Reserve for hardest reasoning only. |
When GPT-5.5 is worth $30 per 1M output tokens
- Voice and real-time UX. The OpenAI voice and real-time API stack remains the most production-ready option, and the priority tier exists specifically for this use case.
- Ecosystem-locked teams. Existing investment in Assistants API, Responses API, file search, code interpreter, and the OpenAI tool ecosystem make GPT-5.5 the path of least resistance.
- Vision and multimodal pipelines. Mature image input handling with consistent SDK ergonomics.
When to switch to a cheaper alternative
- Claude Opus 4.7 ($5 / $25) — same input rate, 17% cheaper output, and Coding Arena #1. Better default for engineering and agentic work.
- Gemini 3.1 Pro ($3.50 / $10.50) — 65% cheaper output, 2M context, top text Arena. Better for long-context and scientific reasoning.
- DeepSeek V4 Pro ($1.74 / $3.48) — 9x cheaper output. Apache 2.0 also enables self-host.
- Skip the priority tier for batch and asynchronous work. The +50% surcharge only pays back on latency-critical paths.
Related
- AI Model Leaderboard — quality vs price across all providers
- Per Million Tokens True Cost — hidden adders pushing bills 1.5-3x above list
- Token Cost Calculator — interactive estimator
- GPT-5.5 deep-dive — full benchmarks and architecture
Teams running GPT-5.5 alongside other providers typically front the API with Swfte Connect to route across these models behind one OpenAI-compatible surface with prompt caching and per-route fallback.
Sources: official OpenAI pricing page, May 2026-05-06. Priority surcharge and Pro variant pricing reflect April 2026 ship.