Wan 2.2 Text To Image LoRA
Playground
Try it on WavespeedAI!WAN 2.2 generates super-detailed images from text prompts and supports custom LoRAs for fine-grained style and subject control. Ready-to-use REST inference API, best performance, no coldstarts, affordable pricing.
Features
Wan-2.2-LoRA (Text-to-Image)
Wan-2.2-LoRA builds upon the acclaimed Wan 2.2 text-to-image model by introducing full LoRA (Low-Rank Adaptation) compatibility — empowering creators to fine-tune visuals with personalized styles, characters, or aesthetics. It combines Wan’s signature cinematic rendering and world-class detail synthesis with the flexibility of custom-trained LoRAs.
Why it looks great
- LoRA-ready architecture – Import
.safetensorsLoRA weights directly from Civitai, Hugging Face. - Cinematic lighting engine – Advanced diffusion backbone that simulates depth, tone, and atmosphere with film-grade realism.
- Text rendering excellence – Handles both English and Chinese typography seamlessly within the image, not as overlays.
- Cross-style adaptability – From photorealism to anime, oil painting, 3D CG, or minimalism — one prompt can shift universes.
- Consistent composition – Retains character identity and spatial coherence across multi-prompt workflows.
Limits and Performance
- Max resolution per job: up to 1536 × 1536 pixels
- LoRA path: supports
<owner>/<model-name>or direct.safetensorsURLs - LoRA scale: adjustable from 0.1 – 1.5 (default = 1.0)
- Output formats: JPEG / PNG / WEBP
- Processing speed: ~6–9 seconds per image
- Prompt input: multi-line, bilingual, descriptive prompts supported
Pricing
- $0.025 per image Each image is billed individually.
How to Use
- Write a detailed prompt (in English or Chinese).
- Set size — width and height (up to 1024×1024).
- Add LoRA(s) – paste LoRA path or URL; adjust scale for blending strength.
- (Optional) Set a seed for reproducibility (
-1= random). - Choose output format (JPEG / PNG / WEBP).
- Run → preview result → iterate with different LoRAs or scales.
Pro tips for best quality
- Mix multiple LoRAs for hybrid aesthetics (e.g., cyberpunk + watercolor).
- Use 0.6–0.9 scale for realistic subtle blending.
- Lock seed to maintain consistent faces or characters across styles.
- Start from simple prompts; layer complexity gradually for control.
Reference
Note
- LoRAs from Civitai or Hugging Face are also supported if exported in
.safetensorsformat. - For multi-LoRA blending, ensure each LoRA file is stylistically aligned for optimal results.
Authentication
For authentication details, please refer to the Authentication Guide.
API Endpoints
Submit Task & Query Result
# Submit the task
curl --location --request POST "https://api.wavespeed.ai/api/v3/wavespeed-ai/wan-2.2/text-to-image-lora" \
--header "Content-Type: application/json" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}" \
--data-raw '{
"size": "1024*1024",
"seed": -1,
"output_format": "jpeg",
"enable_sync_mode": false,
"enable_base64_output": false
}'
# Get the result
curl --location --request GET "https://api.wavespeed.ai/api/v3/predictions/${requestId}/result" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}"
Parameters
Task Submission Parameters
Request Parameters
| Parameter | Type | Required | Default | Range | Description |
|---|---|---|---|---|---|
| prompt | string | Yes | - | The positive prompt for the generation. | |
| size | string | No | 1024*1024 | 256 ~ 1536 per dimension | The size of the generated media in pixels (width*height). |
| loras | array | No | max 3 items | List of LoRAs to apply (max 3). | |
| loras[].path | string | Yes | - | Path to the LoRA model | |
| loras[].scale | float | Yes | - | 0.0 ~ 4.0 | Scale of the LoRA model |
| high_noise_loras | array | No | - | - | List of high noise LoRAs to apply (max 3). |
| low_noise_loras | array | No | - | - | List of low noise LoRAs to apply (max 3). |
| seed | integer | No | -1 | -1 ~ 2147483647 | The random seed to use for the generation. -1 means a random seed will be used. |
| output_format | string | No | jpeg | jpeg, png, webp | The format of the output image. |
| enable_sync_mode | boolean | No | false | - | If set to true, the function will wait for the result to be generated and uploaded before returning the response. It allows you to get the result directly in the response. This property is only available through the API. |
| enable_base64_output | boolean | No | false | - | If enabled, the output will be encoded into a BASE64 string instead of a URL. This property is only available through the API. |
Response Parameters
| Parameter | Type | Description |
|---|---|---|
| code | integer | HTTP status code (e.g., 200 for success) |
| message | string | Status message (e.g., “success”) |
| data.id | string | Unique identifier for the prediction, Task Id |
| data.model | string | Model ID used for the prediction |
| data.outputs | array | Array of URLs to the generated content (empty when status is not completed) |
| data.urls | object | Object containing related API endpoints |
| data.urls.get | string | URL to retrieve the prediction result |
| data.has_nsfw_contents | array | Array of boolean values indicating NSFW detection for each output |
| data.status | string | Status of the task: created, processing, completed, or failed |
| data.created_at | string | ISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”) |
| data.error | string | Error message (empty if no error occurred) |
| data.timings | object | Object containing timing details |
| data.timings.inference | integer | Inference time in milliseconds |