Kandinsky 5 Distill (Text-to-Video)
A diffusion model designed for fast text-to-video generation (no sound), offered as a compact variant of the Kandinsky 5 (Text-to-Video) model. A resolution is slightly above standard definition (SD).
Setup your API Key
If you don’t have an API key for the AI/ML API yet, feel free to use our Quickstart guide.
How to Make a Call
API Schemas
Create a video generation task and send it to the server
You can generate a video using this API. In the basic setup, you only need a prompt. This endpoint creates and sends a video generation task to the server — and returns a generation ID.
The text description of the scene, subject, or action to generate in the video.
The aspect ratio of the generated video.
3:2Possible values: The length of the output video in seconds.
5Possible values: Number of inference steps for sampling. Higher values give better quality but take longer.
30Retrieve the generated video from the server
After sending a request for video generation, this task is added to the queue. This endpoint lets you check the status of a video generation task using its generation_id, obtained from the endpoint described above.
If the video generation task status is complete, the response will include the final result — with the generated video URL and additional metadata.
Successfully generated video
Successfully generated video
Full Example: Generating and Retrieving the Video From the Server
The code below creates a video generation task, then automatically polls the server every 10 seconds until it finally receives the video URL.
Processing time: 53.6 sec.
Generated Video (768x512, without sound):
Last updated
Was this helpful?