Google Veo3.1 Reference To Video
Playground
Try it on WavespeedAI!Google Veo3.1 Reference-to-Video performs image-to-video generation that preserves a specific subject’s appearance and identity from provided reference images, enabling consistent character or product motion across frames. Ready-to-use REST inference API, best performance, no coldstarts, affordable pricing.
Features
Google Veo 3.1 — Reference-to-Video Model
Veo 3.1 Reference-to-Video brings static images to life by combining visual reference consistency with cinematic motion generation. Powered by Google DeepMind’s next-generation Veo 3.1 architecture, this model transforms up to three reference images into coherent 5-second videos with smooth motion, accurate visual alignment, and synchronized native audio.
🌟 Key Features
🧠 Multi-Image Reference Support
- Accepts up to three reference images to define the subject, environment, or style.
- Maintains consistent identity, lighting, and appearance across frames.
- Ideal for animating people, objects, or scenes with reliable fidelity.
🎬 Cinematic Video Generation
- Produces 5-second motion clips at 1080p or 720p resolution.
- Adds camera dynamics such as panning, zooming, or subtle perspective drift.
- Supports synchronized audio generation, matching dialogue or ambient context.
💡 Smart Prompt Adherence
- Interprets both text instructions and visual cues for precise motion storytelling.
- Automatically harmonizes character interactions, props, and backgrounds.
⚙️ Capabilities
-
Input:
- Up to 3 reference images (JPEG / PNG / WEBP)
- Text prompt describing motion, action, and scene context
-
Output:
- 8-second MP4 video (720p or 1080p)
- Optional synchronized audio
-
Negative Prompt (optional):
- Exclude unwanted artifacts or elements (e.g., “no text”, “no flicker”).
-
Seed (optional):
- Reproduce specific results for consistent creative control.
💰 Pricing
| Duration | Resolution | With Audio | Without Audio |
|---|---|---|---|
| 8 seconds | 720p | $3.20 | $1.60 |
| 8 seconds | 1080p | $3.20 | $1.60 |
✅ Commercial use allowed
🧩 How to Use
- Upload up to 3 reference images — define the subject, object, or visual style.
- Write a text prompt — describe the action, setting, and camera motion.
- (Optional) Add a negative prompt to remove unwanted details.
- Choose resolution (720p or 1080p).
- (Optional) Enable audio generation for synchronized sound.
- Click Run to generate your 5-second cinematic video.
💡 Best Practices
- Use clear, well-lit reference images with similar styles and proportions.
- Keep prompts concise but specific (e.g., “The man in image 1 waves to the penguins in image 2 under bright sunlight”).
- Avoid overly complex scenarios with many characters or fast movement.
- Enable audio for more immersive storytelling results.
📝 Notes
- Ensure uploaded images are valid and accessible URLs or uploaded locally.
- If the output looks unstable, reduce reference count or simplify the prompt.
- Follow Google’s content safety rules; modify the prompt if flagged.
- For best performance, prefer portrait-oriented subjects and balanced lighting.
Authentication
For authentication details, please refer to the Authentication Guide.
API Endpoints
Submit Task & Query Result
# Submit the task
curl --location --request POST "https://api.wavespeed.ai/api/v3/google/veo3.1/reference-to-video" \
--header "Content-Type: application/json" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}" \
--data-raw '{
"resolution": "1080p",
"generate_audio": true
}'
# Get the result
curl --location --request GET "https://api.wavespeed.ai/api/v3/predictions/${requestId}/result" \
--header "Authorization: Bearer ${WAVESPEED_API_KEY}"
Parameters
Task Submission Parameters
Request Parameters
| Parameter | Type | Required | Default | Range | Description |
|---|---|---|---|---|---|
| prompt | string | Yes | - | The positive prompt for the generation. | |
| images | array | Yes | [] | 1 ~ 3 items | The model will use the provided images as references to generate a video with consistent subjects. For fields that accept images: Accepts 1 to 3 images; Images Assets can be provided via URLs or Base64 encode; You must use one of the following codecs: PNG, JPEG, JPG, WebP; The dimensions of the images must be at least 128*128 pixels; All images are limited to 50MB; The length of the base64 decode must be under 50MB, and it must include an appropriate content type string. |
| resolution | string | No | 1080p | 720p, 1080p | Video resolution. |
| generate_audio | boolean | No | true | - | Whether to generate audio. |
| negative_prompt | string | No | - | The negative prompt for the generation. | |
| seed | integer | No | - | -1 ~ 2147483647 | The random seed to use for the generation. |
Response Parameters
| Parameter | Type | Description |
|---|---|---|
| code | integer | HTTP status code (e.g., 200 for success) |
| message | string | Status message (e.g., “success”) |
| data.id | string | Unique identifier for the prediction, Task Id |
| data.model | string | Model ID used for the prediction |
| data.outputs | array | Array of URLs to the generated content (empty when status is not completed) |
| data.urls | object | Object containing related API endpoints |
| data.urls.get | string | URL to retrieve the prediction result |
| data.has_nsfw_contents | array | Array of boolean values indicating NSFW detection for each output |
| data.status | string | Status of the task: created, processing, completed, or failed |
| data.created_at | string | ISO timestamp of when the request was created (e.g., “2023-04-01T12:34:56.789Z”) |
| data.error | string | Error message (empty if no error occurred) |
| data.timings | object | Object containing timing details |
| data.timings.inference | integer | Inference time in milliseconds |