
video-to-video
Idle
Your request will cost $0.2 per run.
For $10 you can run this model approximately 50 times.
One more thing::
Wan2.2-Fun-Control is an advanced video generation and control model developed by the Alibaba PAI team, designed for precise and creative video synthesis. By integrating Control Codes with deep learning and multi-modal conditioning, it enables users to direct motion, structure, and scene composition — achieving controllable, high-fidelity video generation under customizable guidance.
🎛️ Multi-Modal Control Supports multiple input types for fine-grained video control:
🎬 High-Quality Video Generation Built on the Wan 2.2 architecture — delivering cinematic, high-resolution video outputs with stable motion and consistent identity.
🌍 Multi-Language Prompting Accepts both Chinese and English descriptions for flexible creative control.
đź§ Intelligent Composition Aligns user-provided references (images or frames) with pose, structure, and scene layout to ensure natural transitions and realism.
| Resolution | Cost per 5 Seconds | Max Duration |
|---|---|---|
| 480p | $0.20 | 120 seconds |
| 720p | $0.40 | 120 seconds |
🧍 Keep reference consistency: The reference image’s composition, pose, and camera angle should match the desired video framing. Major mismatches between input and control maps (e.g., OpenPose or Canny) can lead to generation instability or artifacts.
🖼️ Match aspect ratios: The aspect ratio of the input image and target video should remain identical for best results.
🔄 Control balance: Combining too many control types simultaneously may reduce creative flexibility — start with one or two controls and tune gradually.