openai/gpt-5.4-pro
1,050,000 context · $30.00/M input tokens · $180.00/M output tokens
GPT-5.4 Pro is OpenAI's most advanced model, building on GPT-5.4's unified architecture with enhanced reasoning capabilities for complex, high-stakes tasks. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs. Optimized for step-by-step reasoning, instruction following, and accuracy, GPT-5.4 Pro excels at agentic coding, long-context workflows, and multi-step problem solving.
Paiement à l'usage
Aucun coût initial, payez uniquement ce que vous utilisez
Utilisez les exemples de code suivants pour intégrer notre API :
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="openai/gpt-5.4-pro",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)GPT-5.4 Pro is OpenAI's most advanced model, building on GPT-5.4's unified architecture with enhanced reasoning capabilities for complex, high-stakes
GPT-5.4 Pro is OpenAI's most advanced model, building on GPT-5.4's unified architecture with enhanced reasoning capabilities for complex, high-stakes tasks. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs. Optimized for step-by-step reasoning, instruction following, and accuracy, GPT-5.4 Pro excels at agentic coding, long-context workflows, and multi-step problem solving.
| Specification | Value |
|---|---|
| Provider | Openai |
| Model Type | Large Language Model (LLM) |
| Architecture | N/A |
| Context Window | 1050000 tokens |
| Max Output | 128000 tokens |
| Input | Text |
| Output | Text |
| Vision | Supported |
| Function Calling | Supported |
| Token Type | Cost per Million Tokens |
|---|---|
| Input | $30.0 |
| Output | $180.0 |
Base URL: https://llm.wavespeed.ai/v1 API Endpoint: chat/completions Model ID: openai/gpt-5.4-pro
from openai import OpenAI
client = OpenAI(
api_key="YOUR_API_KEY",
base_url="https://llm.wavespeed.ai/v1"
)
response = client.chat.completions.create(
model="openai/gpt-5.4-pro",
messages=[
{"role": "user", "content": "Hello!"}
]
)
print(response.choices[0].message.content)
curl https://llm.wavespeed.ai/v1/chat/completions \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{
"model": "openai/gpt-5.4-pro",
"messages": [{"role": "user", "content": "Hello!"}]
}'
openai/gpt-5.4-pro
GPT-5.4 Pro is OpenAI's most advanced model, building on GPT-5.4's unified architecture with enhanced reasoning capabilities for complex, high-stakes tasks. It features a 1M+ token context window (922K input, 128K output) with support for text and image inputs. Optimized for step-by-step reasoning, instruction following, and accuracy, GPT-5.4 Pro excels at agentic coding, long-context workflows, and multi-step problem solving.
Entrée
$30 /M
Sortie
$180 /M
Contexte
1050K
Sortie max.
128K
Vision
Pris en charge
Utilisation d'outils
Pris en charge
Accédez à GPT 5.4 Pro via notre API unifiée — compatible OpenAI, sans démarrages à froid, prix transparents.
Ouvrir le PlaygroundTarification sur WaveSpeedAI : $30.00 par million de tokens d'entrée et $180.00 par million de tokens de sortie. Le prompt caching et le traitement par batch sont facturés séparément et réduisent le coût effectif sur les charges longues et répétitives.
GPT 5.4 Pro prend en charge jusqu'à 1050K tokens de contexte et jusqu'à 128K tokens de sortie par requête.
Oui. WaveSpeedAI expose GPT 5.4 Pro via un endpoint compatible OpenAI à https://llm.wavespeed.ai/v1. Pointez le SDK officiel d'OpenAI vers cette base URL avec votre clé API WaveSpeedAI — aucune autre modification de code requise.
Connectez-vous à WaveSpeedAI, créez une clé API dans Access Keys, puis envoyez une requête à https://llm.wavespeed.ai/v1/chat/completions avec l'id du modèle affiché ci-dessus. Les nouveaux comptes reçoivent des crédits gratuits pour évaluer GPT 5.4 Pro.