Openai Sora 2 Text To Video
Playground
Try it on WavespeedAI!OpenAI Sora 2 is a state-of-the-art text-to-video model with realistic visuals, accurate physics, synchronized audio, and strong steerability. Ready-to-use REST inference API, best performance, no coldstarts, affordable pricing.
Features
OpenAI Sora 2 — Text-to-Video
Sora 2 is a state-of-the-art video+audio generator. It advances prior video models with more accurate physics, sharper realism, synchronized audio, stronger steerability, and a wider stylistic range—built on the original Sora foundation.
Why it looks great
- Physics-aware motion: learns contact, inertia, and momentum so objects move and collide believably.
- Temporal consistency: stable identities, minimal flicker/ghosting, and clean frame-to-frame transitions.
- Synchronized audio: lip-sync alignment, beat-aware cuts, and ambience that matches on-screen action.
- High-frequency detail: preserves fine textures (skin, fabric, foliage) without plastic over-sharpening.
- Complex scene reasoning: handles multiple subjects, occlusions, depth, and long camera moves coherently.
- Cinematic camera literacy: natural pans, push-ins, and handheld vibes without warping or jelly-artifacts.
- Wide stylistic range: from photoreal and documentary to anime, 3D, and illustrative aesthetics.
- Strong steerability: responds predictably to prompt edits and control settings (duration, fps, motion strength).
How to Use
- Prompt: describe scene, style, camera, and audio cues.
- Duration: select 4s, 8s, or 12s.
- Submit: start generation; preview and download when ready.
Pricing
| Duration | Total ($) |
|---|---|
| 4s | 0.40 |
| 8s | 0.80 |
| 12s | 1.20 |
Billing Rules: Pricing scales linearly with duration (flat $0.10/s). Durations are fixed at 4s, 8s, or 12s.
Note
Please follow the user rules from OpenAI, you can find details in the reference: What images are permitted and prohibited in Sora-2
Authentication
For authentication details, please refer to the Authentication Guide.
API Endpoints
Submit Task & Query Result
# Submit the task
curl --location --request POST "https://api.wavespeed.ai/api/v3/openai/sora-2/text-to-video" \
--header "Content-Type: application/json" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}" \
--data-raw '{
"size": "720*1280",
"duration": 4
}'
# Get the result
curl --location --request GET "https://api.wavespeed.ai/api/v3/predictions/${requestId}/result" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}"
Parameters
Task Submission Parameters
Request Parameters
| Parameter | Type | Required | Default | Range | Description |
|---|---|---|---|---|---|
| prompt | string | Yes | - | The positive prompt for the generation. | |
| size | string | No | 720*1280 | 720*1280, 1280*720 | The size of the generated media in pixels (width*height). |
| duration | integer | No | 4 | 4, 8, 12 | The duration of the generated video in seconds. |
Response Parameters
| Parameter | Type | Description |
|---|---|---|
| code | integer | HTTP status code (e.g., 200 for success) |
| message | string | Status message (e.g., “success”) |
| data.id | string | Unique identifier for the prediction, Task Id |
| data.model | string | Model ID used for the prediction |
| data.outputs | array | Array of URLs to the generated content (empty when status is not completed) |
| data.urls | object | Object containing related API endpoints |
| data.urls.get | string | URL to retrieve the prediction result |
| data.has_nsfw_contents | array | Array of boolean values indicating NSFW detection for each output |
| data.status | string | Status of the task: created, processing, completed, or failed |
| data.created_at | string | ISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”) |
| data.error | string | Error message (empty if no error occurred) |
| data.timings | object | Object containing timing details |
| data.timings.inference | integer | Inference time in milliseconds |