
video-to-video
Idle
Your request will cost $0.2 per run.
For $10 you can run this model approximately 50 times.
One more thing::
Wan2.2-Animate is an Alibaba Wan 2.2–based motion-transfer model that turns a single character image into a full performance. It copies body movement and facial expressions from a driving video while keeping the output stable, realistic, and identity-consistent.
Wan2.2-Animate supports two behaviors:
animate The person or object in the input image is animated to follow the motion in the driving video. Use this when you want to bring a still portrait, illustration, or character design to life.
replace The model replaces the main performer in the video with the person or object from the image, keeping the original motion, scene, and camera. Use this when you want to swap a new character into an existing shot (presenter, actor, mascot, etc.).
Pricing depends on output length and resolution.
| Resolution | Price per 5 s | Max billed length |
|---|---|---|
| 480p | 0.20 USD | 120 s |
| 720p | 0.40 USD | 120 s |
Underlying logic (no math required on your side):
Billed duration is clamped to the range 5–120 seconds
480p
720p
So, for example:
image (required) Upload a clear character image.
video (required) Upload the driving video whose pose, movement, and expressions you want to copy.
mode (required)
prompt (optional) Short guidance such as:
resolution Choose 480p for faster, cheaper previews or 720p for higher-quality output.
seed (optional)
Run generation, review the result, and adjust image, video, mode, prompt, and resolution as needed.
Match composition and pose Try to keep the image and video similar in camera angle, crop, and body pose. This reduces artifacts and identity drift, especially in replace mode.
Keep aspect ratio consistent Use the same or very similar aspect ratio for both the image and the video for more stable framing.
Use clean, well-lit faces Avoid heavy occlusion by hands, microphones, or props, and avoid strong motion blur in the driving video.
Prototype at 480p Start with 480p and short clips to find good settings, then switch to 720p and longer durations once you like the behavior.
This gives you a predictable-cost, two-mode pipeline for driving AI characters and doing high-quality character replacement in existing footage.