Discover and experiment with cutting-edge AI models. Create stunning visuals with our state-of-the-art image generation technology.
ALL VIDEO MODELS
All video models to boost your creativity
HUNYUAN VIDEO MODELS
Hunyuan video model collection
KLING VIDEO MODELS
Kling V1.6 Model Collection
VIDU VIDEO MODELS
Vidu 2.0 model collection
FLUX IMAGE MODELS
Flux model collection
WAN 2.1 VIDEO MODELS
WAN 2.1 T2V and I2V models collection
12
16
25
14
1
10
2
81 models
wavespeed-ai/flux-dev-lora
Rapid, high-quality image generation with FLUX.1 [dev] and LoRA support for personalized styles and brand-specific outputs
hot
image-to-image
wavespeed-ai/flux-dev-lora-ultra-fast
Rapid, high-quality image generation with FLUX.1 [dev] and LoRA support for personalized styles and brand-specific outputs, ultra fast !
wavespeed-ai/wan-2.1/i2v-480p-lora
Wan-2.1 i2v model with LoRA, generate high-quality videos with superior visual quality and motion diversity
wavespeed-ai/flux-dev
Flux-dev text to image model, 12 billion parameter rectified flow transformer
wavespeed-ai/flux-schnell
FLUX.1 [schnell] is fastest image generation model tailored for local development and personal use, a 12 billion parameter rectified flow transformer
wavespeed-ai/flux-dev-ultra-fast
Flux-dev text to image model, 12 billion parameter rectified flow transformer, ultra fast!
wavespeed-ai/flux-schnell-lora
FLUX.1 [schnell] is a 12 billion parameter flow transformer that generates high-quality images from text in 1 to 4 steps, suitable for personal and commercial use.
wavespeed-ai/instant-character
InstantCharacter creates high-quality, consistent characters from text prompts, supporting diverse poses, styles, and appearances with strong identity control.
image-to-IMGE
new
wavespeed-ai/flux-pro-redux
FLUX.1 [pro] Redux is a high-performance endpoint for the FLUX.1 [pro] model that enables rapid transformation of existing images, delivering high-quality style transfers and image modifications with the core FLUX capabilities.
wavespeed-ai/flux-redux-dev
Open-weight image variation model. Create new versions while preserving key elements of your original.
wavespeed-ai/wan-2.1/i2v-480p
The Wan2.1 14B model is an advanced image-to-video model that offers accelerated inference capabilities, enabling high-res video generation with high visual quality and motion diversity
wavespeed-ai/wan-2.1/i2v-720p
Wan2.1 I2V-14B model is capable of generating 720P high-definition videos from images
wavespeed-ai/wan-2.1/i2v-720p-lora
wavespeed-ai/wan-2.1/i2v-720p-ultra-fast
wavespeed-ai/wan-2.1/i2v-480p-ultra-fast
wavespeed-ai/wan-2.1/i2v-720p-lora-ultra-fast
wavespeed-ai/flux-kontext-pro/multi
Experimental version of FLUX.1 Kontext [pro] with multi image handling capabilities
wavespeed-ai/flux-kontext-pro
A state-of-the-art image editing model, Flux Kontext, which aims to provide comparable performance against the closed-source models like GPT-4o and Gemini2 Flash.
kwaivgi/kling-v2.0-i2v-master
Kling AI is a powerful AI Text to Video and Image to Video model family developed by Kuaishou, the company behind one of China’s largest video-sharing platforms.
image-to-video
kwaivgi/kling-v1.6-i2v-pro
Generate 5s videos in 1080p resolution from image
minimax/video-01
Generate 6s videos with prompts or images. (Also known as Hailuo). Use a subject reference to make a video with a character and the S2V-01 model.
text-to-video
wavespeed-ai/flux-kontext-max/text-to-image
FLUX.1 Kontext [max] text-to-image is a new premium model brings maximum performance across all aspects – greatly improved prompt adherence.
wavespeed-ai/flux-kontext-max/multi
Experimental version of FLUX.1 Kontext [max] with multi image handling capabilities
wavespeed-ai/flux-kontext-max
wavespeed-ai/video-upscaler
The Upscale Model API is a powerful tool designed to enhance the resolution and quality of videos. Whether you're working with low-resolution videos that need a boost or aiming to improve the clarity of existing footage, this API leverages advanced machine learning models to deliver high-quality, upscaled videos.
wavespeed-ai/wan-2.1-14b-vace
VACE is an all-in-one model designed for video creation and editing. It encompasses various tasks, including reference-to-video generation (R2V), video-to-video editing (V2V), and masked video-to-video editing (MV2V), allowing users to compose these tasks freely. This functionality enables users to explore diverse possibilities and streamlines their workflows effectively, offering a range of capabilities, such as Move-Anything, Swap-Anything, Reference-Anything, Expand-Anything, Animate-Anything, and more.
wavespeed-ai/flux-kontext-pro/text-to-image
The FLUX.1 Kontext [pro] text-to-image delivers state-of-the-art image generation results with unprecedented prompt following, photorealistic rendering, and flawless typography.
kwaivgi/kling-v2.0-t2v-master
wavespeed-ai/wan-2.1/v2v-480p
Inference for Wan 2.1 14B. Unleashing high-res 480p video-to-video prowess with cutting-edge suite of video foundation models
wavespeed-ai/hunyuan-custom-ref2v-720p
HunyuanCustom, a multi-modal, conditional, and controllable generation model centered on subject consistency, built upon the Hunyuan Video generation framework. It enables the generation of subject-consistent videos conditioned on text, images, audio, and video inputs.
vidu/start-end-to-video-2.0
Create dynamic videos using just the first and last frame images, enhanced with text descriptions for seamless storytelling.
wavespeed-ai/hidream-e1-full
HiDream-E1 is an image editing model built on HiDream-I1.
wavespeed-ai/wan-flf2v
Wan-2.1 flf2v generates dynamic videos by intelligently bridging a given first frame to a desired end frame through smooth, coherent motion sequences.
wavespeed-ai/hidream-i1-dev
HiDream-I1 is a new open-source image generative foundation model with 17B parameters that achieves state-of-the-art image generation quality within seconds.
wavespeed-ai/hidream-i1-full
wavespeed-ai/wan-14b-trainer
To train a WAN Lora, you need at least 10 images to achieve good results. The trainer outputs a Lora URL, which is a temporary storage URL that is valid for 7 days. You should download the Lora to your own storage.
wavespeed-ai/flux-dev-lora-trainer
A FLUX dev LoRA trainer for subjects and styles.
wavespeed-ai/uno
An AI model that transforms input images into new ones based on text prompts, blending reference visuals with your creative directions.
wavespeed-ai/dia-tts
Dia directly generates realistic dialogue from transcripts. Audio conditioning enables emotion control. Produces natural nonverbals like laughter and throat clearing. will cost $0.04 per 1000 character.
wavespeed-ai/mmaudio-v2
MMAudio generates synchronized audio given video and/or text inputs. It can be combined with video models to get videos with audio.
wavespeed-ai/flux-dev-fill
FLUX.1 [dev] Fill is a high-performance endpoint for the FLUX.1 [dev] model that enables rapid transformation of existing images, delivering high-quality style transfers and image modifications with the core FLUX capabilities.
wavespeed-ai/wan-2.1/t2v-480p-lora
Turbo-charged inference for Wan 2.1 14B. Unleashing high-res 480p text-to-video prowess with cutting-edge suite of video foundation models, LoRA effect added
wavespeed-ai/wan-2.1/t2v-720p-lora
Turbo-charged inference for Wan 2.1 14B. Unleashing high-res 720p text-to-video prowess with cutting-edge suite of video foundation models, LoRA effect added
wavespeed-ai/wan-2.1/t2v-480p
The Wan2.1 14B model is an advanced text-to-video model that offers accelerated inference capabilities, enabling high-res video generation with high visual quality and motion diversity
wavespeed-ai/veo2-t2v
Veo 2 creates videos with realistic motion and high quality output. Explore different styles and find your own with extensive camera controls.
wavespeed-ai/veo2-i2v
Veo 2 creates videos from images with realistic motion and very high quality output.
wavespeed-ai/imagen4
Google’s highest quality image generation model
wavespeed-ai/wan-2.1/v2v-720p-lora
Turbo-charged inference for Wan 2.1 14B. Unleashing high-res 720p video-to-video prowess with cutting-edge suite of video foundation models
wavespeed-ai/wan-2.1/v2v-480p-lora
Turbo-charged inference for Wan 2.1 14B. Unleashing high-res 480p video-to-video prowess with cutting-edge suite of video foundation models
wavespeed-ai/wan-2.1/v2v-480p-lora-ultra-fast
wavespeed-ai/wan-2.1/v2v-720p-lora-ultra-fast
wavespeed-ai/wan-2.1/v2v-720p-ultra-fast
wavespeed-ai/wan-2.1/v2v-720p
Inference for Wan 2.1 14B. Unleashing high-res 720p video-to-video prowess with cutting-edge suite of video foundation models
wavespeed-ai/wan-2.1/v2v-480p-ultra-fast
wavespeed-ai/hunyuan-custom-ref2v-480p
featured
wavespeed-ai/sdxl-lora
SDXL is a text-to-image generative AI model developed by Stability AI that creates beautiful images. It is the successor to Stable Diffusion.
wavespeed-ai/sdxl
wavespeed-ai/ltx-video-v097/i2v-720p
Generate videos from prompts and images using LTX Video-0.9.7
wavespeed-ai/ltx-video-v097/i2v-480p
wavespeed-ai/step1x-edit
Step1X-Edit transforms your photos with simple instructions into stunning, professional-quality edits—rivaling top proprietary tools.
vidu/reference-to-video-2.0
Create videos that align with reference subjects—like characters, objects, and environments—using the world’s first Multi-Entity Consistency feature.
vidu/image-to-video-2.0
Bring your images to life by turning them into dynamic videos that capture your vision and action.
wavespeed-ai/SkyReels-V1
SkyReels V1 is the first and most advanced open-source human-centric video foundation model. By fine-tuning HunyuanVideo on O(10M) high-quality film and television clips
wavespeed-ai/magi-1-24b
MAGI-1 is a video generation model with exceptional understanding of physical interactions and cinematic prompts
wavespeed-ai/framepack
Framepack is an efficient Image-to-video model that autoregressively generates videos.
wavespeed-ai/flux-control-lora-depth
FLUX Control LoRA Depth is a high-performance endpoint that uses a control image to transfer structure to the generated image, using a depth map.
wavespeed-ai/flux-control-lora-canny
FLUX Control LoRA Canny is a high-performance endpoint that uses a control image to transfer structure to the generated image, using a Canny edge map.
wavespeed-ai/real-esrgan
Real-ESRGAN with optional face correction and adjustable upscale
wavespeed-ai/ghibli
Reimagine and transform your ordinary photos into enchanting Studio Ghibli style artwork
wavespeed-ai/hunyuan-video/i2v
Hunyuan Video is an Open video generation model with high visual quality, motion diversity, text-video alignment, and generation stability. This endpoint generates videos from image and text descriptions.
wavespeed-ai/wan-2.1/i2v-480p-lora-ultra-fast
kwaivgi/kling-v1.6-i2v-standard
Generate 5s videos in 720p resolution from image
kwaivgi/kling-v1.6-t2v-standard
Generate 5s videos in 720p resolution
wavespeed-ai/wan-2.1/t2v-720p
Turbo-charged inference for Wan 2.1 14B. Unleashing high-res text-to-video prowess with cutting-edge suite of video foundation models
wavespeed-ai/hunyuan-video/t2v
Hunyuan Video is an Open video generation model with high visual quality, motion diversity, text-video alignment, and generation stability. This endpoint generates videos from text descriptions.
wavespeed-ai/wan-2.1/t2v-480p-ultra-fast
wavespeed-ai/wan-2.1/t2v-720p-ultra-fast
wavespeed-ai/wan-2.1/t2v-480p-lora-ultra-fast
wavespeed-ai/wan-2.1/t2v-720p-lora-ultra-fast
test/test-model
Please don't use this model only for the development team's debugging purposes.
wavespeed-ai/hunyuan3d-v2-multi-view
Generate 3D models from your images using Hunyuan 3D. A native 3D generative model enabling versatile and high-quality 3D asset creation.