Wan 2.2 VACE Fun Outpainting (Image-to-Video)
A video generation model that combines a source image, mask, and reference video to produce prompted videos with precise source control.
Setup your API Key
If you don’t have an API key for the AI/ML API yet, feel free to use our Quickstart guide.
How to Make a Call
API Schemas
Video Generation
This endpoint creates and sends a video generation task to the server — and returns a generation ID.
The text description of the scene, subject, or action to generate in the video.
A HTTPS URL pointing to a video or a data URI containing a video. This video will be used as a reference during generation.
The description of elements to avoid in the generated video.
letterboxing, borders, black bars, bright colors, overexposed, static, blurred details, subtitles, style, artwork, painting, picture, still, overall gray, worst quality, low quality, JPEG compression residue, ugly, incomplete, extra fingers, poorly drawn hands, poorly drawn faces, deformed, disfigured, malformed limbs, fused fingers, still picture, cluttered background, three legs, many people in the background, walking backwardsNumber of frames to generate.
81Whether to match the input video's frames per second (FPS).
Frames per second of the generated video.
16The resolution of the output video, where the number refers to the short side in pixels.
480pPossible values: The aspect ratio of the generated video.
autoPossible values: Number of inference steps for sampling. Higher values give better quality but take longer.
30Classifier-free guidance scale. Controls prompt adherence / creativity.
5Noise schedule shift parameter. Affects temporal dynamics.
5If set to true, the safety checker will be enabled.
Whether to enable prompt expansion.
Whether to preprocess the input video.
Acceleration to use for inference.
regularPossible values: The quality of the generated video.
highPossible values: The method used to write the video.
balancedPossible values: Number of frames to interpolate between the original frames.
Temporal downsample factor for the video.
The minimum frames per second to downsample the video to.
The minimum frames per second to downsample the video to.
15The model to use for interpolation. Rife, or film are available.
filmPossible values: The synchronization mode for audio and video. Loose or tight are available.
Whether to expand the video to the left
trueWhether to expand the video to the right
trueWhether to expand the video to the top
trueWhether to expand the video to the bottom
trueAmount of expansion. This is a float value between 0 and 1, where 0.25 adds 25% to the original video size on the specified sides
0.25Fetch the video
After sending a request for video generation, this task is added to the queue. This endpoint lets you check the status of a video generation task using its id, obtained from the endpoint described above.
If the video generation task status is complete, the response will include the final result — with the generated video URL and additional metadata.
Successfully generated video
Successfully generated video
Full Example: Generating and Retrieving the Video From the Server
The code below creates a video generation task, then automatically polls the server every 10 seconds until it finally receives the video URL.
Last updated
Was this helpful?