Introducing Sync LipSync 1.9.0 Beta on WaveSpeedAI
Try Sync LipSync 1.9.0 Beta for FREESync Labs Lipsync 1.9.0 Beta Is Now Live on WaveSpeedAI: Zero-Shot Lip Sync That Just Works
The future of video dubbing and lip synchronization has arrived. WaveSpeedAI is excited to announce the availability of Sync Labs Lipsync 1.9.0 Beta—the zero-shot lip sync model that Sync Labs called “the biggest upgrade to our model lineup in our history” and “the most natural lipsyncing model in the world.”
Whether you’re a filmmaker localizing content for international audiences, a content creator building multilingual social media campaigns, or a developer integrating lip sync into production workflows, this model delivers studio-quality results without the traditional complexity of training data or manual adjustments.
What is Sync Labs Lipsync 1.9.0 Beta?
Sync Labs, the Y Combinator-backed team behind the foundational Wav2Lip open-source model (starred over 11,000 times on GitHub), has built Lipsync 1.9.0 Beta as a complete rethinking of how lip synchronization should work.
Unlike previous generation models that used multi-stage pipelines—where errors accumulated as video passed from one processing stage to another—Lipsync 1.9.0 Beta operates as an end-to-end monolith in a single shot. This architectural breakthrough eliminates the quality degradation that plagued earlier approaches.
The model is zero-shot, meaning you need zero training data to use it effectively. Whether you have an hour of footage or just a few seconds, you can generate natural lip movements that match any audio track—across live-action footage, stylized animation, and AI-generated video.
Key Features
-
Zero-Shot Processing: No training, fine-tuning, or reference clips required. Upload your video and audio, and the model handles everything automatically.
-
Style-Aware Editing: The model adjusts only the mouth region while preserving the speaker’s identity, lighting conditions, and background—maintaining visual continuity throughout.
-
Cross-Domain Support: Works seamlessly across live-action footage, Pixar-level CG animation, stylized characters, and AI-generated faces without switching models or adjusting parameters.
-
Flexible Timing Control: Five sync modes (loop, bounce, cut_off, silence, remap) let you control exactly how the model handles duration mismatches between your video and audio tracks.
-
Natural Facial Integration: Rather than simply replacing the lower half of the face like traditional models, Lipsync 1.9.0 Beta understands how the entire face moves during speech—because facial expressions are complex, interdependent movements.
-
Active Speaker Detection: The model’s state-of-the-art pipeline can handle long videos with multiple speakers, automatically associating each unique voice with the correct face.
Real-World Use Cases
Film and Video Production
International Distribution: Release a single film in multiple languages for global streaming platforms. The model preserves the emotional performance of your original actors while syncing to translated dialogue—no need to reassemble the cast for ADR sessions.
Post-Production Dialogue Replacement: Need to revise a script after principal photography? Record new lines and let the AI match them perfectly to existing footage, eliminating expensive reshoots.
Documentary Accessibility: Make real-world interviews accessible to broader audiences without the uncanny valley effect of traditional dubbing.
Content Creation and Marketing
Social Media Localization: Create TikToks, Reels, and YouTube Shorts that speak directly to audiences in their native language. One master video can spawn dozens of localized versions in minutes.
Personalized Video Campaigns: Swap audio tracks to tailor messages for different customer segments without shooting new content. Update spokesperson videos seasonally without bringing talent back to set.
E-Learning and Training: Translate onboarding, training, and educational videos while maintaining the natural delivery that keeps learners engaged.
Developer Integration
The model’s REST API makes integration straightforward for developers building dubbing pipelines, video editing tools, or content localization platforms. Process videos programmatically at scale with consistent, professional results.
Getting Started on WaveSpeedAI
Using Lipsync 1.9.0 Beta on WaveSpeedAI is straightforward:
-
Upload Your Video: Navigate to the model page and upload your source video. For best results, use footage with a clearly visible face—front-facing or three-quarter view works best.
-
Add Your Audio: Upload your target speech track (MP3 or WAV). The cleaner the audio, the better the results—minimize background noise for optimal synchronization.
-
Select Your Sync Mode: Choose how you want the model to handle any length mismatch between video and audio:
- Loop: Repeat the shorter stream
- Bounce: Reverse and repeat
- Cut_off: Trim to match
- Silence: Pad with silence
- Remap: Time-stretch to match
-
Run and Download: Click Run and receive your processed video with perfectly synchronized lip movements.
Why WaveSpeedAI?
WaveSpeedAI delivers the fastest inference speeds for Lipsync 1.9.0 Beta with zero cold starts—your jobs begin processing immediately without waiting for model initialization. Our infrastructure is optimized for video processing workloads, and our transparent pricing means you pay only for actual processing time.
Pricing: $0.025 per second of processed video. A 30-second clip costs just $0.75. A full minute runs $1.50.
| Clip Length | Price |
|---|---|
| 5 seconds | $0.13 |
| 10 seconds | $0.25 |
| 30 seconds | $0.75 |
| 60 seconds | $1.50 |
Tips for Best Results
- Lighting: Use clean, well-lit close-ups for the most convincing lip sync
- Framing: Avoid heavy head turns or faces partially out of frame
- Audio Quality: Clean speech with minimal background noise produces the best synchronization
- Speech Rhythm: For dubbed content, match the phrasing and pauses of your translated audio roughly to the original performance timing
Take Your Video Content Global
The AI lip sync market has evolved rapidly, with Sync Labs consistently leading in quality and realism. Lipsync 1.9.0 Beta represents the culmination of years of research from the team that created the foundational Wav2Lip model—now available through WaveSpeedAI’s fast, reliable infrastructure.
Stop choosing between quality and convenience. Stop paying for expensive voice actors and manual syncing work. Start creating multilingual content that looks and feels native to every audience.

