Wan 2.1 I2V 720p LoRA Ultra Fast
Playground
Try it on WavespeedAI!WAN 2.1 i2v 720p is an ultra-fast Image-to-Video model that turns images into 720P videos and supports custom LoRAs for style control. Ready-to-use REST inference API, best performance, no coldstarts, affordable pricing.
Features
Wan 2.1 I2V 720p LoRA Ultra Fast — wavespeed-ai/wan-2.1/i2v-720p-lora-ultra-fast
Wan 2.1 I2V 720p LoRA Ultra Fast is a high-speed image-to-video model that turns a single reference image into a short, coherent clip guided by your prompt. It supports up to 3 LoRAs per run, making it great for fast style exploration, character looks, and branded aesthetics while keeping generation time low.
Key capabilities
- SOTA Performance: Wan2.1 consistently outperforms existing open-source models and state-of-the-art commercial solutions across multiple benchmarks.
- Multiple Tasks: Wan2.1 excels in Text-to-Video, Image-to-Video, Video Editing, Text-to-Image, and Video-to-Audio, advancing the field of video generation.
- Visual Text Generation: Wan2.1 is the first video model capable of generating both Chinese and English text, featuring robust text generation that enhances its practical applications.
- Powerful Video VAE: Wan-VAE delivers exceptional efficiency and performance, encoding and decoding 1080P videos of any length while preserving temporal information, making it an ideal foundation for video and image generation.
Use cases
- Bring key art to life: animate posters, portraits, product shots into short clips
- Style locking for content pipelines: apply one or more LoRAs for consistent aesthetics
- Character look development: combine identity/style LoRAs for repeatable outputs
- Social media loops: quick 5–10s clips from a single still image
- Marketing variations: same composition, different mood, lighting, wardrobe, brand style
Pricing
| Resolution | Duration | Price per run | Effective price per second |
|---|---|---|---|
| 720p | 5s | $0.225 | $0.045/s |
| 720p | 10s | $0.338 | $0.034/s |
Inputs
- image (required): reference image that anchors subject and composition
- prompt (required): what happens in the clip (action + camera + mood)
- loras (optional): up to 3 LoRA items (each with a path and a scale)
Parameters
- duration: clip length (commonly 5s or 10s)
- num_inference_steps: more steps can improve motion coherence and detail
- guidance_scale: higher values follow the prompt more strongly
- strength: how strongly the model follows the input image vs. deviating into new variations
- negative_prompt: describe what to avoid (blur, artifacts, jitter, low quality)
- flow_shift: adjust motion behavior (useful for smoother or more dynamic movement)
- seed: set for reproducible results (-1 for random)
- loras[].path: LoRA identifier or a public safetensors URL
- loras[].scale: LoRA weight (tune per LoRA; start around 0.8–1.0)
Prompting guide (I2V + LoRA)
Write prompts like a director’s brief:
- Subject: who/what is on screen
- Action: what changes over time (gesture, expression, environment)
- Camera: push-in, pull-back, pan, tilt, handheld vs. locked-off
- Mood/lighting: golden hour, neon rain, candlelight, fog, rim light
- Motion style: subtle, smooth, energetic, dramatic
LoRA stacking tip: keep the base prompt clean, and let LoRAs handle style/identity. If multiple LoRAs conflict, lower one LoRA scale rather than overloading the prompt.
Example prompts
- A stylish street cat wearing sunglasses holds a skateboard. The cat shifts its stance, tail sways, and the camera slowly pushes in. Cinematic lighting, crisp detail, smooth natural motion, 720p, 5 seconds.
- A fantasy portrait comes alive: soft wind moves hair and fabric, embers drift across the frame, the character slowly turns their head and smiles. Slow dolly-in, moody lighting, subtle fog, smooth motion.
Authentication
For authentication details, please refer to the Authentication Guide.
API Endpoints
Submit Task & Query Result
# Submit the task
curl --location --request POST "https://api.wavespeed.ai/api/v3/wavespeed-ai/wan-2.1/i2v-720p-lora-ultra-fast" \
--header "Content-Type: application/json" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}" \
--data-raw '{
"loras": [
{
"path": "Remade-AI/Squish",
"scale": 1
}
],
"size": "1280*720",
"num_inference_steps": 30,
"duration": 5,
"guidance_scale": 5,
"flow_shift": 5,
"seed": -1
}'
# Get the result
curl --location --request GET "https://api.wavespeed.ai/api/v3/predictions/${requestId}/result" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}"
Parameters
Task Submission Parameters
Request Parameters
| Parameter | Type | Required | Default | Range | Description |
|---|---|---|---|---|---|
| image | string | Yes | - | The image for generating the output. | |
| prompt | string | Yes | - | The positive prompt for the generation. | |
| negative_prompt | string | No | - | The negative prompt for the generation. | |
| loras | array | No | max 3 items | List of LoRAs to apply (max 3). | |
| loras[].path | string | Yes | - | Path to the LoRA model | |
| loras[].scale | float | Yes | - | 0.0 ~ 4.0 | Scale of the LoRA model |
| size | string | No | 1280*720 | 1280*720, 720*1280 | The size of the generated media in pixels (width*height). |
| num_inference_steps | integer | No | 30 | 1 ~ 40 | The number of inference steps to perform. |
| duration | integer | No | 5 | 5 ~ 10 | The duration of the generated media in seconds. |
| guidance_scale | number | No | 5 | 0.00 ~ 20.00 | The guidance scale to use for the generation. |
| flow_shift | number | No | 5 | 1.0 ~ 10.0 | The shift value for the timestep schedule for flow matching. |
| seed | integer | No | -1 | -1 ~ 2147483647 | The random seed to use for the generation. -1 means a random seed will be used. |
Response Parameters
| Parameter | Type | Description |
|---|---|---|
| code | integer | HTTP status code (e.g., 200 for success) |
| message | string | Status message (e.g., “success”) |
| data.id | string | Unique identifier for the prediction, Task Id |
| data.model | string | Model ID used for the prediction |
| data.outputs | array | Array of URLs to the generated content (empty when status is not completed) |
| data.urls | object | Object containing related API endpoints |
| data.urls.get | string | URL to retrieve the prediction result |
| data.has_nsfw_contents | array | Array of boolean values indicating NSFW detection for each output |
| data.status | string | Status of the task: created, processing, completed, or failed |
| data.created_at | string | ISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”) |
| data.error | string | Error message (empty if no error occurred) |
| data.timings | object | Object containing timing details |
| data.timings.inference | integer | Inference time in milliseconds |
Result Request Parameters
| Parameter | Type | Required | Default | Description |
|---|---|---|---|---|
| id | string | Yes | - | Task ID |
Result Response Parameters
| Parameter | Type | Description |
|---|---|---|
| code | integer | HTTP status code (e.g., 200 for success) |
| message | string | Status message (e.g., “success”) |
| data | object | The prediction data object containing all details |
| data.id | string | Unique identifier for the prediction, the ID of the prediction to get |
| data.model | string | Model ID used for the prediction |
| data.outputs | string | Array of URLs to the generated content (empty when status is not completed). |
| data.urls | object | Object containing related API endpoints |
| data.urls.get | string | URL to retrieve the prediction result |
| data.status | string | Status of the task: created, processing, completed, or failed |
| data.created_at | string | ISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”) |
| data.error | string | Error message (empty if no error occurred) |
| data.timings | object | Object containing timing details |
| data.timings.inference | integer | Inference time in milliseconds |