Browse ModelsWavespeed AIWan 2.1 I2V 720p LoRA Ultra Fast

Wan 2.1 I2V 720p LoRA Ultra Fast

Wan 2.1 I2V 720p LoRA Ultra Fast

Playground

Try it on WavespeedAI!

WAN 2.1 i2v 720p is an ultra-fast Image-to-Video model that turns images into 720P videos and supports custom LoRAs for style control. Ready-to-use REST inference API, best performance, no coldstarts, affordable pricing.

Features

Wan 2.1 I2V 720p LoRA Ultra Fast — wavespeed-ai/wan-2.1/i2v-720p-lora-ultra-fast

Wan 2.1 I2V 720p LoRA Ultra Fast is a high-speed image-to-video model that turns a single reference image into a short, coherent clip guided by your prompt. It supports up to 3 LoRAs per run, making it great for fast style exploration, character looks, and branded aesthetics while keeping generation time low.

Key capabilities

  • SOTA Performance: Wan2.1 consistently outperforms existing open-source models and state-of-the-art commercial solutions across multiple benchmarks.
  • Multiple Tasks: Wan2.1 excels in Text-to-Video, Image-to-Video, Video Editing, Text-to-Image, and Video-to-Audio, advancing the field of video generation.
  • Visual Text Generation: Wan2.1 is the first video model capable of generating both Chinese and English text, featuring robust text generation that enhances its practical applications.
  • Powerful Video VAE: Wan-VAE delivers exceptional efficiency and performance, encoding and decoding 1080P videos of any length while preserving temporal information, making it an ideal foundation for video and image generation.

Use cases

  • Bring key art to life: animate posters, portraits, product shots into short clips
  • Style locking for content pipelines: apply one or more LoRAs for consistent aesthetics
  • Character look development: combine identity/style LoRAs for repeatable outputs
  • Social media loops: quick 5–10s clips from a single still image
  • Marketing variations: same composition, different mood, lighting, wardrobe, brand style

Pricing

ResolutionDurationPrice per runEffective price per second
720p5s$0.225$0.045/s
720p10s$0.338$0.034/s

Inputs

  • image (required): reference image that anchors subject and composition
  • prompt (required): what happens in the clip (action + camera + mood)
  • loras (optional): up to 3 LoRA items (each with a path and a scale)

Parameters

  • duration: clip length (commonly 5s or 10s)
  • num_inference_steps: more steps can improve motion coherence and detail
  • guidance_scale: higher values follow the prompt more strongly
  • strength: how strongly the model follows the input image vs. deviating into new variations
  • negative_prompt: describe what to avoid (blur, artifacts, jitter, low quality)
  • flow_shift: adjust motion behavior (useful for smoother or more dynamic movement)
  • seed: set for reproducible results (-1 for random)
  • loras[].path: LoRA identifier or a public safetensors URL
  • loras[].scale: LoRA weight (tune per LoRA; start around 0.8–1.0)

Prompting guide (I2V + LoRA)

Write prompts like a director’s brief:

  • Subject: who/what is on screen
  • Action: what changes over time (gesture, expression, environment)
  • Camera: push-in, pull-back, pan, tilt, handheld vs. locked-off
  • Mood/lighting: golden hour, neon rain, candlelight, fog, rim light
  • Motion style: subtle, smooth, energetic, dramatic

LoRA stacking tip: keep the base prompt clean, and let LoRAs handle style/identity. If multiple LoRAs conflict, lower one LoRA scale rather than overloading the prompt.

Example prompts

  • A stylish street cat wearing sunglasses holds a skateboard. The cat shifts its stance, tail sways, and the camera slowly pushes in. Cinematic lighting, crisp detail, smooth natural motion, 720p, 5 seconds.
  • A fantasy portrait comes alive: soft wind moves hair and fabric, embers drift across the frame, the character slowly turns their head and smiles. Slow dolly-in, moody lighting, subtle fog, smooth motion.

Authentication

For authentication details, please refer to the Authentication Guide.

API Endpoints

Submit Task & Query Result


# Submit the task
curl --location --request POST "https://api.wavespeed.ai/api/v3/wavespeed-ai/wan-2.1/i2v-720p-lora-ultra-fast" \
--header "Content-Type: application/json" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}" \
--data-raw '{
    "loras": [
        {
            "path": "Remade-AI/Squish",
            "scale": 1
        }
    ],
    "size": "1280*720",
    "num_inference_steps": 30,
    "duration": 5,
    "guidance_scale": 5,
    "flow_shift": 5,
    "seed": -1
}'

# Get the result
curl --location --request GET "https://api.wavespeed.ai/api/v3/predictions/${requestId}/result" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}"

Parameters

Task Submission Parameters

Request Parameters

ParameterTypeRequiredDefaultRangeDescription
imagestringYes-The image for generating the output.
promptstringYes-The positive prompt for the generation.
negative_promptstringNo-The negative prompt for the generation.
lorasarrayNomax 3 itemsList of LoRAs to apply (max 3).
loras[].pathstringYes-Path to the LoRA model
loras[].scalefloatYes-0.0 ~ 4.0Scale of the LoRA model
sizestringNo1280*7201280*720, 720*1280The size of the generated media in pixels (width*height).
num_inference_stepsintegerNo301 ~ 40The number of inference steps to perform.
durationintegerNo55 ~ 10The duration of the generated media in seconds.
guidance_scalenumberNo50.00 ~ 20.00The guidance scale to use for the generation.
flow_shiftnumberNo51.0 ~ 10.0The shift value for the timestep schedule for flow matching.
seedintegerNo-1-1 ~ 2147483647The random seed to use for the generation. -1 means a random seed will be used.

Response Parameters

ParameterTypeDescription
codeintegerHTTP status code (e.g., 200 for success)
messagestringStatus message (e.g., “success”)
data.idstringUnique identifier for the prediction, Task Id
data.modelstringModel ID used for the prediction
data.outputsarrayArray of URLs to the generated content (empty when status is not completed)
data.urlsobjectObject containing related API endpoints
data.urls.getstringURL to retrieve the prediction result
data.has_nsfw_contentsarrayArray of boolean values indicating NSFW detection for each output
data.statusstringStatus of the task: created, processing, completed, or failed
data.created_atstringISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”)
data.errorstringError message (empty if no error occurred)
data.timingsobjectObject containing timing details
data.timings.inferenceintegerInference time in milliseconds

Result Request Parameters

ParameterTypeRequiredDefaultDescription
idstringYes-Task ID

Result Response Parameters

ParameterTypeDescription
codeintegerHTTP status code (e.g., 200 for success)
messagestringStatus message (e.g., “success”)
dataobjectThe prediction data object containing all details
data.idstringUnique identifier for the prediction, the ID of the prediction to get
data.modelstringModel ID used for the prediction
data.outputsstringArray of URLs to the generated content (empty when status is not completed).
data.urlsobjectObject containing related API endpoints
data.urls.getstringURL to retrieve the prediction result
data.statusstringStatus of the task: created, processing, completed, or failed
data.created_atstringISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”)
data.errorstringError message (empty if no error occurred)
data.timingsobjectObject containing timing details
data.timings.inferenceintegerInference time in milliseconds
© 2025 WaveSpeedAI. All rights reserved.