SDXL — Text / Image-to-Image
SDXL is Stability AI’s flagship diffusion model for high-quality image generation. It excels at photorealism, stylized illustration, and product renders.
Key Features
- Text-to-Image (T2I): Generate fully new images from natural-language prompts.
- Image-to-Image (I2I): Transform an input image toward your prompt while preserving composition.
Modes & When to Use
- T2I: Best for concepting and fresh compositions.
- I2I: Use when you already like the framing/pose and want a new style or content consistent with the source.
Inputs & Parameters
- prompt (required): Describe subject, style, lighting, camera, mood.
- image: PNG/JPEG/WebP; used as the starting point.
- width / height: Flexible canvas; common choices 1024×1024, 1536×1536, 1024×1536, 1536×1024.
- seed:
-1 for random, or set any integer for reproducible results.
How to Use
A) Text-to-Image
- Enter a prompt with subject + context + style + lighting + camera.
- Set width/height for your target aspect ratio (e.g., 1536×1536 for square).
- (Optional) Set seed for repeatability.
- Run and iterate on wording or seed.
B) Image-to-Image
- Upload an image.
- Write a prompt describing the desired change or style.
- (Optional) Set seed for repeatability.
- Run; tweak strength or prompt until the balance feels right.
Price
Prompting Tips
- Structure prompts as: [subject] + [context] + [style/artist] + [lighting] + [camera] + [mood].
- For photorealism: “cinematic lighting, shallow depth of field, 50mm, natural skin texture.”
- For products: “studio sweep, three-point lighting, soft reflections, high detail.”
- Keep negative prompts concise: “blurry, extra fingers, watermark, text, low-res.”
- Lock a seed once you like the look to iterate predictably.
Notes
- SDXL will honor platform safety rules; if a prompt is blocked, simplify and try again.
- Commercial use follows Stability AI licensing and WaveSpeedAI terms.