video-o1-video-to-video-edit
The model transforms an input video according to a natural-language text prompt, altering style, visual attributes, or the overall look of the scene while preserving the original motion and structural layout of the footage.
How to Make a Call
API Schemas
Generating a video using this model involves sequentially calling two endpoints:
The first one is for creating and sending a video generation task to the server (returns a generation ID).
The second one is for requesting the generated video from the server using the generation ID received from the first endpoint.
Below, you can find two corresponding API schemas and an example with both endpoint calls.
Create a video generation task and send it to the server
The text description of the scene, subject, or action to generate in the video.
A HTTPS URL pointing to a video or a data URI containing a video. This video will be used as a reference during generation.
Array of image URLs for multi-image-to-video generation.
Whether to keep the original audio from the video.
falseRetrieve the generated video from the server
After sending a request for video generation, this task is added to the queue. This endpoint lets you check the status of a video generation task using its generation_id, obtained from the endpoint described above.
If the video generation task status is complete, the response will include the final result — with the generated video URL and additional metadata.
Successfully generated video
Successfully generated video
Code Example
The code below creates a video generation task, then automatically polls the server every 15 seconds until it finally receives the video URL.
Processing time: ~ 3 min 55 sec.
Generated video (1940x1068, without sound):
Last updated
Was this helpful?