Home/Explore/Wan 2.1 Video Models/wavespeed-ai/wan-2.1-14b-vace

image-to-video

wavespeed-ai/wan-2.1-14b-vace

VACE is an all-in-one model designed for video creation and editing. It encompasses various tasks, including reference-to-video generation (R2V), video-to-video editing (V2V), and masked video-to-video editing (MV2V), allowing users to compose these tasks freely. This functionality enables users to explore diverse possibilities and streamlines their workflows effectively, offering a range of capabilities, such as Move-Anything, Swap-Anything, Reference-Anything, Expand-Anything, Animate-Anything, and more.

Doc
preview

Hint: You can drag and drop a file or click to upload

Hint: You can drag and drop a file or click to upload

Hint: You can drag and drop a file or click to upload

Hint: You can drag and drop a file or click to upload

Hint: You can drag and drop a file or click to upload

If set to true, the safety checker will be enabled.

Idle

Your request will cost $0.3 per run.

For $10 you can run this model approximately 33 times.

One more thing:

ExamplesView all

README

Wan2.1 VACE: Three Core Capabilities Analysis

Multi-modal Information Input,Making video generation more controllable.

Traditional video generation workflows, once completed, make it difficult to adjust character postures, actions, scene transitions, and other details. Wan2.1 VACE provides powerful controllable capabilities, supporting generation based on human poses, motion flow, structural preservation, spatial movement, camera angles, and other controls, while also supporting video generation based on themes and background references.

The core technology behind this is Wan VACE's multi-modal input mechanism. Unlike traditional models that rely solely on text prompts, Wan VACE(Wan2.1 VACE) has built a unified input system that integrates text, images, videos, masks, and control signals.

For image input, Wan VACE (Wan 2.1 VACE) supports object reference images or video frames. For video input, users can use Wan VACE to regenerate content through operations such as erasing and local expansion. For local regions, users can specify editing areas through binary 0/1 signals. For control signals, Wan VACE (Wan2.1 VACE) supports depth maps, optical flow, layouts, grayscale, line drawings, and pose estimation.

Unified Single Model - One-Stop Solution for Multiple Tasks

Wan VACE (Wan2.1 VACE) supports content replacement, addition, or deletion operations in specified areas within videos. In terms of time dimension, Wan VACE can arbitrarily extend the video length at the beginning or end. In terms of spatial dimension, it supports progressive generation of backgrounds or specific regions, such as background replacement - under the premise of preserving the main subject, the background environment can be changed according to prompts.

Free Combination of Multiple Tasks - Unleashing AI Creative Boundaries

Wan VACE(Wan2.1 VACE) also supports the free combination of various single-task capabilities, breaking through the limitations of traditional expert models that work in isolation. As a unified model, it can naturally integrate capabilities such as video generation, pose control, background replacement, and local region editing. There's no need to train new models for single-function tasks separately.