openai/gpt-5.5
1,050,000 context · $5.00/M input tokens · $30.00/M output tokens
GPT-5.5 is OpenAI's frontier model released April 23, 2026, featuring a 1M+ token context window (922K input, 128K output) with text and image support. It scores 88.7% on SWE-bench Verified and 92.4% on MMLU with 60% fewer hallucinations than GPT-5.4, excelling at agentic coding, computer use, and deep research while matching GPT-5.4 per-token latency.
Paiement à l'usage
Aucun coût initial, payez uniquement ce que vous utilisez
Utilisez les exemples de code suivants pour intégrer notre API :
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="openai/gpt-5.5",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)GPT-5.5 is OpenAI's frontier model released on April 23, 2026, designed for complex professional workloads including agentic coding, computer use, and deep research. Building on GPT-5.4, it delivers stronger reasoning, higher reliability with 60% fewer hallucinations, and improved token efficiency — matching GPT-5.4 per-token latency while performing at a significantly higher level of intelligence. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs.
| Benchmark | GPT-5.4 | GPT-5.5 | Claude Opus 4.7 | Gemini 3.1 Pro |
|---|---|---|---|---|
| SWE-bench Verified | ~74% | 88.7% | 87.6% | 80.6% |
| MMLU | 91.1% | 92.4% | — | — |
| Terminal-Bench 2.0 | — | 82.7% | — | — |
| Expert-SWE | — | 73.1% | — | — |
| GDPval (44 occupations) | — | 84.9% | — | — |
| OSWorld-Verified | — | 78.7% | — | — |
| Hallucination rate | baseline | −60% | — | — |
| Specification | Value |
|---|---|
| Provider | OpenAI |
| Model Type | Large Language Model (LLM) |
| Architecture | Transformer (Frontier) |
| Context Window | 1050000 tokens |
| Max Input | 922000 tokens |
| Max Output | 128000 tokens |
| Input | Text, Image |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Thinking Mode | Supported |
| Release Date | April 23, 2026 |
Note: GPT-5.5 is 2× GPT-5.4 at the token level, but uses significantly fewer tokens to complete the same tasks. Independent testing puts the net cost increase at roughly 20% once token efficiency is factored in.
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: openai/gpt-5.5
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="openai/gpt-5.5",
messages=[{"role": "user", "content": "Hello!"}]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "openai/gpt-5.5",
"messages": [{"role": "user", "content": "Hello!"}]
}'
| Aspect | GPT-5.4 | GPT-5.5 |
|---|---|---|
| SWE-bench Verified | ~74% | 88.7% |
| MMLU | 91.1% | 92.4% |
| Hallucination rate | baseline | −60% |
| Context window | 1.05M | 1.05M (922K input) |
| API input price | $2.50/M | $5.00/M |
| API output price | $15.00/M | $30.00/M |
| Computer use | Improving | Production-grade |
| Multi-step tool chains | Single-shot preferred | Full autonomous loops |
| Token efficiency | baseline | ~40% fewer tokens on same tasks |
openai/gpt-5.5
GPT-5.5 is OpenAI's frontier model released April 23, 2026, featuring a 1M+ token context window (922K input, 128K output) with text and image support. It scores 88.7% on SWE-bench Verified and 92.4% on MMLU with 60% fewer hallucinations than GPT-5.4, excelling at agentic coding, computer use, and deep research while matching GPT-5.4 per-token latency.
Entrée
$5 /M
Sortie
$30 /M
Contexte
1050K
Sortie max.
128K
Vision
Pris en charge
Utilisation d'outils
Pris en charge
Accédez à GPT 5.5 via notre API unifiée — compatible OpenAI, sans démarrages à froid, prix transparents.
Ouvrir le PlaygroundTarification sur WaveSpeedAI : $5.00 par million de tokens d'entrée et $30.00 par million de tokens de sortie. Le prompt caching et le traitement par batch sont facturés séparément et réduisent le coût effectif sur les charges longues et répétitives.
GPT 5.5 prend en charge jusqu'à 1050K tokens de contexte et jusqu'à 128K tokens de sortie par requête.
Oui. WaveSpeedAI expose GPT 5.5 via un endpoint compatible OpenAI à https://llm.wavespeed.ai/v1. Pointez le SDK officiel d'OpenAI vers cette base URL avec votre clé API WaveSpeedAI — aucune autre modification de code requise.
Connectez-vous à WaveSpeedAI, créez une clé API dans Access Keys, puis envoyez une requête à https://llm.wavespeed.ai/v1/chat/completions avec l'id du modèle affiché ci-dessus. Les nouveaux comptes reçoivent des crédits gratuits pour évaluer GPT 5.5.