Hunyuan Video I2V
Playground
Try it on WavespeedAI!Hunyuan i2v turns images and text prompts into high-quality videos, generating coherent short clips from descriptive inputs. Ready-to-use REST inference API, best performance, no coldstarts, affordable pricing.
Features
Hunyuan Video I2V — wavespeed-ai/hunyuan-video/i2v
Hunyuan Video I2V is an image-to-video model that turns a single reference image into a short animated clip guided by a text prompt. Upload an image to lock in subject, composition, and style, then describe the action and camera behavior you want. The model is well-suited for cinematic motion, character-driven beats, and atmospheric scenes where you want the “still” to come alive with coherent movement.
Key capabilities
- Image-to-video generation from one reference image
- Prompt-driven motion: actions, expressions, environment changes, camera movement
- Stable composition anchored to the input image
- Good for cinematic, dramatic, and stylized shots
- Supports duration and inference-step controls for quality vs. speed tradeoffs
- Multiple output sizes (e.g., 1280×720)
Use cases
- Animate key art, posters, and character portraits into short clips
- Cinematic micro-stories: close-up → reveal, slow push-ins, mood-heavy scenes
- Atmosphere and VFX-style motion: rain, fog, embers, neon flicker, drifting particles
- Social content loops from a single still image
- Rapid previsualization for scenes before full video production
Pricing
| Output | Price |
|---|---|
| Per run | $0.40 |
Inputs
- image (required): reference image to anchor subject and style
- prompt (required): action + camera directions
Parameters
- duration: clip length in seconds
- num_inference_steps: sampling steps (higher often improves coherence/detail)
- seed: random seed (-1 for random; set for reproducible results)
- size: output resolution (e.g., 1280×720)
Prompting guide (I2V)
Write prompts like a director’s brief:
- Subject: who/what is on screen
- Action: what changes over time (gestures, expression, environment)
- Camera: push-in, pull-back, pan, tilt, handheld vs. locked-off
- Mood/lighting: candlelight, moonlight, neon, fog, rim light
- Motion constraints: “subtle movement”, “no shaky camera”, “smooth dolly”
Example prompts
- A pale vampire woman stands at a candlelit window, crimson eyes glowing. She slowly raises her hand and taps long nails against the glass. Her expression shifts from seductive to dangerous as bats flutter past outside. Slow cinematic push-in, soft candle flicker, subtle fog, smooth motion, dramatic lighting.
Authentication
For authentication details, please refer to the Authentication Guide.
API Endpoints
Submit Task & Query Result
# Submit the task
curl --location --request POST "https://api.wavespeed.ai/api/v3/wavespeed-ai/hunyuan-video/i2v" \
--header "Content-Type: application/json" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}" \
--data-raw '{
"num_inference_steps": 30,
"duration": 5,
"seed": -1,
"size": "1280*720"
}'
# Get the result
curl --location --request GET "https://api.wavespeed.ai/api/v3/predictions/${requestId}/result" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}"
Parameters
Task Submission Parameters
Request Parameters
| Parameter | Type | Required | Default | Range | Description |
|---|---|---|---|---|---|
| prompt | string | No | - | The positive prompt for the generation. | |
| image | string | Yes | - | The image to generate the video from. | |
| num_inference_steps | integer | No | 30 | 1 ~ 30 | The number of inference steps to perform. |
| duration | integer | No | 5 | 5 ~ 10 | The duration of the generated media in seconds. |
| seed | integer | No | -1 | -1 ~ 2147483647 | The random seed to use for the generation. -1 means a random seed will be used. |
| size | string | No | 1280*720 | 1280*720, 720*1280 | The size of the generated media in pixels (width*height). |
Response Parameters
| Parameter | Type | Description |
|---|---|---|
| code | integer | HTTP status code (e.g., 200 for success) |
| message | string | Status message (e.g., “success”) |
| data.id | string | Unique identifier for the prediction, Task Id |
| data.model | string | Model ID used for the prediction |
| data.outputs | array | Array of URLs to the generated content (empty when status is not completed) |
| data.urls | object | Object containing related API endpoints |
| data.urls.get | string | URL to retrieve the prediction result |
| data.has_nsfw_contents | array | Array of boolean values indicating NSFW detection for each output |
| data.status | string | Status of the task: created, processing, completed, or failed |
| data.created_at | string | ISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”) |
| data.error | string | Error message (empty if no error occurred) |
| data.timings | object | Object containing timing details |
| data.timings.inference | integer | Inference time in milliseconds |
Result Request Parameters
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
| id | string | Yes | - | Task ID |
Result Response Parameters
| Parameter | Type | Description |
|---|---|---|
| code | integer | HTTP status code (e.g., 200 for success) |
| message | string | Status message (e.g., “success”) |
| data | object | The prediction data object containing all details |
| data.id | string | Unique identifier for the prediction, the ID of the prediction to get |
| data.model | string | Model ID used for the prediction |
| data.outputs | string | Array of URLs to the generated content (empty when status is not completed). |
| data.urls | object | Object containing related API endpoints |
| data.urls.get | string | URL to retrieve the prediction result |
| data.status | string | Status of the task: created, processing, completed, or failed |
| data.created_at | string | ISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”) |
| data.error | string | Error message (empty if no error occurred) |
| data.timings | object | Object containing timing details |
| data.timings.inference | integer | Inference time in milliseconds |