WaveSpeedAI APISyncSync Lipsync 2 Pro

Sync Lipsync 2 Pro

Sync Lipsync 2 Pro

Playground

Try it on WavespeedAI!

Lipsync-2-pro is a state-of-the-art video editing model that generates studio-grade lipsync in minutes, not weeks.

Features

Lipsync-2-pro is a zero-shot model for generating realistic lip movements that match spoken audio. It works out of the box—no training or fine-tuning needed—and preserves a speaker’s unique style across different languages and video types. Whether you’re working with live-action footage, animation, or AI-generated characters, Lipsync-2-pro brings new levels of realism, control, and speed.

What it does Zero-shot: No waiting around for training. Just drop in your video and audio—Lipsync handles the rest.

Style preservation: The model picks up on how someone speaks by watching them speak. Even when translating across languages, it keeps their signature delivery.

Cross-domain support: Works with live-action humans, animated characters, and AI-generated faces.

Flexible workflows: Use it for dubbing, editing words in post, or reanimating entire performances.

Key features Temperature control: Fine-tune how expressive the lipsync is. Make it subtle or dial it up depending on the scene.

Active speaker detection: Automatically detects who’s speaking in multi-person videos and applies lipsync only when that person is talking.

Flawless animation: Handles everything from stylized 3D characters to hyperreal AI avatars. Not just for translation—this unlocks editable dialogue in post-production.

Record once, edit forever: You don’t need multiple takes. Change dialogue after the fact while keeping the original speaker’s delivery intact.

Dub any video with AI: If you can generate a video with text, you can dub it too. No need to capture everything on camera anymore.

Authentication

For authentication details, please refer to the Authentication Guide.

API Endpoints

Submit Task & Query Result


# Submit the task
curl --location --request POST "https://api.wavespeed.ai/api/v3/sync/lipsync-2-pro" \
--header "Content-Type: application/json" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}" \
--data-raw '{
    "sync_mode": "cut_off"
}'

# Get the result
curl --location --request GET "https://api.wavespeed.ai/api/v3/predictions/${requestId}/result" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}"

Parameters

Task Submission Parameters

Request Parameters

ParameterTypeRequiredDefaultRangeDescription
videostringYes-The video to be used for generation
audiostringYes--The audio to be used for generation
sync_modestringNocut_offbounce, loop, cut_off, silence, remapDefines how to handle duration mismatches between video and audio inputs. See the Media Content Tips guide https://docs.sync.so/compatibility-and-tips/media-content-tips#sync-mode-options for a brief overview, or the SyncMode enum below for detailed explanations of each option.

Response Parameters

ParameterTypeDescription
codeintegerHTTP status code (e.g., 200 for success)
messagestringStatus message (e.g., “success”)
data.idstringUnique identifier for the prediction, Task Id
data.modelstringModel ID used for the prediction
data.outputsarrayArray of URLs to the generated content (empty when status is not completed)
data.urlsobjectObject containing related API endpoints
data.urls.getstringURL to retrieve the prediction result
data.has_nsfw_contentsarrayArray of boolean values indicating NSFW detection for each output
data.statusstringStatus of the task: created, processing, completed, or failed
data.created_atstringISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”)
data.errorstringError message (empty if no error occurred)
data.timingsobjectObject containing timing details
data.timings.inferenceintegerInference time in milliseconds

Result Request Parameters

© 2025 WaveSpeedAI. All rights reserved.