Vidu Q2 Pro
Generate a coherent shot from just two images: a start frame and an end frame. Vidu Q2 Pro infers natural, object-aware motion between them, making it perfect for scene transitions, bridging shots, and visual storytelling. No local setup required.
Why it looks great
- Bi-frame guidance: uses both start and end frames to anchor identity, layout, and lighting for the whole clip.
- Temporal continuity: minimizes flicker and “popping,” maintaining subject integrity across frames.
- Object- & human-aware motion: preserves faces, hands, and fine details while animating clothes, hair, and props.
- Layout-smart interpolation: respects foreground/background depth, occlusions, and parallax.
- Camera-path estimation: simulates subtle pans, dolly moves, and push-ins without warping.
- Natural look: balances crisp detail with cinematic smoothness—no plastic over-processing.
Use Cases
- Storyboarding & concept animation: bring static boards to life between key beats.
- Scene interpolation in long-form content: seamless bridges between shots.
- Instructional visual sequences: demonstrate change-over-time (before → after) with smooth motion.
- Film previsualization: explore transitions, blocking, and camera moves early.
How to Use
- Upload your start frame and end frame.
- Write your prompt.
- Pick duration (e.g., 5–8 s).
- Set resolution (720p, 1080p).
- (Optional) Choose add BGM or not.
- (Optional) Choose movement_amplitude to manage the amplitude of motion in generated content
- Submit the job, or use seed to reproduce your result later.
- Preview the result and download the final video.
Price
Duration | Resolution | Total ($) | Implied $/s | Equivalent per 5s ($) |
---|
5s | 720p | 0.15 | 0.03000 | 0.1500 |
5s | 1080p | 0.35 | 0.07000 | 0.3500 |
8s | 720p | 0.25 | 0.03125 | 0.1563 |
8s | 1080p | 0.50 | 0.06250 | 0.3125 |
Accelerated Inference
Our accelerated inference approach leverages advanced optimization technology from WaveSpeedAI. This fusion technique reduces computational overhead and latency, enabling rapid generation without compromising quality. The system is tuned for large-scale workloads while keeping real-time use cases snappy and reliable. For implementation details, see our engineering blog post.
Notes
- Actual processing time depends on resolution, duration, motion settings, and current queue.
- For highly dynamic changes (big pose/layout jumps), consider shorter durations or add intermediate key frames to guide motion.
- Ensure you have rights to any images you upload; outputs inherit input content constraints.