Home/Explore/Wan 2.1 Video Models/wavespeed-ai/wan-2.1/v2v-720p-ultra-fast
video-to-video

video-to-video

Wan 2.1 V2V

wavespeed-ai/wan-2.1/v2v-720p-ultra-fast

Ultra-fast Wan 2.1 V2V generates unlimited 720P video-to-video conversions and supports custom LoRAs for personalized styles. Ready-to-use REST inference API, best performance, no coldstarts, affordable pricing.

Hint: You can drag and drop a file or click to upload

Idle

Your request will cost $0.225 per run.

For $10 you can run this model approximately 44 times.

One more thing::

ExamplesView all

README

Wan 2.1 V2V 720p Ultra Fast — wavespeed-ai/wan-2.1/v2v-720p-ultra-fast

Wan 2.1 V2V 720p Ultra Fast is a speed-optimized video-to-video model that transforms an input video using a text prompt while preserving the original motion and timing. Upload a source video, describe the desired changes (style, lighting, environment, details), and tune strength to control how closely the output follows the original footage. This variant is the non-LoRA version, built for fast, clean V2V iteration at 720p.

Key capabilities

  • Ultra-fast video-to-video transformation anchored to an input video (720p output)
  • Prompt-guided edits while keeping motion continuity and pacing
  • Strength control to balance preservation vs. transformation
  • Fine motion behavior tuning via flow_shift for smoother motion
  • Efficient for rapid A/B testing with different prompts and seeds

Use cases

  • Rapid 720p V2V restyling for social, ads, and creative iteration
  • Mood and lighting changes (cinematic grade, warm window light, neon, noir)
  • Brand-safe refresh: keep composition and timing, update textures/colors/details
  • Consistent motion preservation when you only need prompt-driven changes
  • Fast iteration before upgrading to higher resolution or LoRA workflows

Pricing

DurationPrice per video
5s$0.225
10s$0.3375

Inputs

  • video (required): source video to transform
  • prompt (required): what to change and how the result should look
  • negative_prompt (optional): what to avoid (artifacts, jitter, unwanted elements)

Parameters

  • num_inference_steps: sampling steps
  • duration: output duration (seconds)
  • strength: how strongly to transform the input video (lower = preserve more; higher = change more)
  • guidance_scale: prompt adherence strength
  • flow_shift: motion/flow behavior tuning
  • seed: random seed (-1 for random; fixed for reproducible results)

Prompting guide (V2V)

A clean structure is “preserve + edit + constraints”:

Template: Keep the original motion and timing. Change [style/lighting/environment/details]. Keep faces stable and natural. Avoid flicker, warping, and jitter.

Example prompts

  • Keep the original motion and composition. Apply a candid, cinematic look with warm sunlight, soft depth of field, and gentle film grain.
  • Preserve timing and camera movement. Restyle the scene into a clean anime look with stable shading and no flicker.
  • Keep the same scene and people. Shift the color grade to golden hour and add subtle bloom while maintaining realistic shadows.