Kling Omni 1 is available now!Experience Now
Home/Explore/Kling O1 Models/kwaivgi/kling-video-o1/text-to-video
text-to-video

text-to-video

Kling Omni Video O1 Text-To-Video | World's First Unified Multi-Modal Video AI Model | WaveSpeedAI

kwaivgi/kling-video-o1/text-to-video

Kling Omni Video O1 is Kuaishou's first unified multi-modal video model with MVL (Multi-modal Visual Language) technology. Text-to-Video mode generates cinematic videos from text prompts with subject consistency, natural physics simulation, and precise semantic understanding. Ready-to-use REST API, best performance, no coldstarts, affordable pricing.

Idle

Your request will cost $0.56 per run.

For $10 you can run this model approximately 17 times.

One more thing::

ExamplesView all

README

Kling Omni Video O1 — Text-to-Video

Kling Omni Video O1 is Kuaishou's groundbreaking unified multi-modal video model, representing the world's first AI system that seamlessly integrates text, images, videos, and subject references into a single creative engine. The Text-to-Video mode transforms natural language prompts into stunning, cinematic video content.

🌟 Why Kling Video O1 Stands Out

Universal Creative Engine

Unlike traditional single-task models, Video O1 unifies multiple video generation capabilities:

  • Text-to-video generation
  • Image-to-video transformation
  • Reference-based video creation
  • Video editing and modification
  • Shot extension and scene continuation

Multi-Modal Visual Language (MVL)

The model interprets your instructions through a revolutionary MVL system that understands:

  • Natural language descriptions
  • Visual context and references
  • Subject identity and characteristics
  • Scene dynamics and physics

Subject Consistency

Maintains stable character, prop, and scene features across varying shots — similar to professional directing techniques used in film production.

🎬 Core Features

  • Cinematic Quality — Film-grade visual output with natural lighting and realistic motion
  • Physics Simulation — Accurate real-world physics for natural movement and dynamics
  • Semantic Understanding — Deep comprehension of complex prompts and creative intent
  • Flexible Outputs — Multiple resolution and duration options

🚀 How to Use

  1. Write Your Prompt Describe the scene, action, camera movement, and mood you want.

    Example: "A young woman walking through a neon-lit Tokyo street at night, rain reflecting city lights, cinematic tracking shot"

  2. Set Parameters Choose your preferred duration, resolution, and aspect ratio.

  3. Generate Submit your request and receive high-quality video output.

💰 Pricing

ItemPrice
Per Second$0.112

Billed per second of output video duration.

💡 Pro Tips

  • Use specific camera terms: "tracking shot," "close-up," "aerial view"
  • Describe lighting conditions: "golden hour," "neon-lit," "soft diffused light"
  • Include motion cues: "slowly walking," "rapid zoom," "gentle breeze"
  • Specify mood and atmosphere for better results