openai/gpt-5.5
1,050,000 context · $5.00/M input tokens · $30.00/M output tokens
GPT-5.5 is OpenAI's frontier model released April 23, 2026, featuring a 1M+ token context window (922K input, 128K output) with text and image support. It scores 88.7% on SWE-bench Verified and 92.4% on MMLU with 60% fewer hallucinations than GPT-5.4, excelling at agentic coding, computer use, and deep research while matching GPT-5.4 per-token latency.
Pay-per-use
No upfront costs, pay only for what you use
Use the following code examples to integrate with our API:
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="openai/gpt-5.5",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)GPT-5.5 is OpenAI's frontier model released on April 23, 2026, designed for complex professional workloads including agentic coding, computer use, and deep research. Building on GPT-5.4, it delivers stronger reasoning, higher reliability with 60% fewer hallucinations, and improved token efficiency — matching GPT-5.4 per-token latency while performing at a significantly higher level of intelligence. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs.
| Benchmark | GPT-5.4 | GPT-5.5 | Claude Opus 4.7 | Gemini 3.1 Pro |
|---|---|---|---|---|
| SWE-bench Verified | ~74% | 88.7% | 87.6% | 80.6% |
| MMLU | 91.1% | 92.4% | — | — |
| Terminal-Bench 2.0 | — | 82.7% | — | — |
| Expert-SWE | — | 73.1% | — | — |
| GDPval (44 occupations) | — | 84.9% | — | — |
| OSWorld-Verified | — | 78.7% | — | — |
| Hallucination rate | baseline | −60% | — | — |
| Specification | Value |
|---|---|
| Provider | OpenAI |
| Model Type | Large Language Model (LLM) |
| Architecture | Transformer (Frontier) |
| Context Window | 1050000 tokens |
| Max Input | 922000 tokens |
| Max Output | 128000 tokens |
| Input | Text, Image |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Thinking Mode | Supported |
| Release Date | April 23, 2026 |
Note: GPT-5.5 is 2× GPT-5.4 at the token level, but uses significantly fewer tokens to complete the same tasks. Independent testing puts the net cost increase at roughly 20% once token efficiency is factored in.
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: openai/gpt-5.5
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="openai/gpt-5.5",
messages=[{"role": "user", "content": "Hello!"}]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "openai/gpt-5.5",
"messages": [{"role": "user", "content": "Hello!"}]
}'
| Aspect | GPT-5.4 | GPT-5.5 |
|---|---|---|
| SWE-bench Verified | ~74% | 88.7% |
| MMLU | 91.1% | 92.4% |
| Hallucination rate | baseline | −60% |
| Context window | 1.05M | 1.05M (922K input) |
| API input price | $2.50/M | $5.00/M |
| API output price | $15.00/M | $30.00/M |
| Computer use | Improving | Production-grade |
| Multi-step tool chains | Single-shot preferred | Full autonomous loops |
| Token efficiency | baseline | ~40% fewer tokens on same tasks |
openai/gpt-5.5
GPT-5.5 is OpenAI's frontier model released April 23, 2026, featuring a 1M+ token context window (922K input, 128K output) with text and image support. It scores 88.7% on SWE-bench Verified and 92.4% on MMLU with 60% fewer hallucinations than GPT-5.4, excelling at agentic coding, computer use, and deep research while matching GPT-5.4 per-token latency.
Input
$5 /M
Output
$30 /M
Context
1050K
Max Output
128K
Vision
Supported
Tool Use
Supported
Access GPT 5.5 through our unified API — OpenAI-compatible, no cold starts, transparent pricing.
Open PlaygroundPricing on WaveSpeedAI: $5.00 per million input tokens and $30.00 per million output tokens. Prompt caching and batch processing are billed separately and reduce effective cost on long, repetitive workloads.
GPT 5.5 supports up to 1050K tokens of context with up to 128K tokens of output per request.
Yes. WaveSpeedAI exposes GPT 5.5 through an OpenAI-compatible endpoint at https://llm.wavespeed.ai/v1. Point the official OpenAI SDK at this base URL with your WaveSpeedAI API key — no other code changes required.
Sign in to WaveSpeedAI, create an API key in Access Keys, then send a request to https://llm.wavespeed.ai/v1/chat/completions with model id set to the value shown above. New accounts receive free credits to evaluate GPT 5.5 before paying per token.