Text and image to video generation
Kling 3.0 is a video generation model developed by Kling, released with a training date of February 2026. It supports both text-to-video and image-to-video workflows, accepting text prompts, image URLs, and multiple configuration options as inputs. The model is identified by the ID kling-video-v3.0-std and is available on MindStudio as part of the Kling model family.
Kling 3.0 is suited for creators and developers who need to generate video content from written descriptions or existing images. Its dual input support makes it flexible for use cases ranging from concept visualization to animating static imagery. The model accepts a context window of up to 10,000 tokens, giving users room to provide detailed prompts and configuration parameters.
What Kling 3.0 supports
Text to Video
Generates video clips from written text prompts, accepting up to 10,000 tokens of input context for detailed scene descriptions.
Image to Video
Animates a provided image URL into a video, allowing static visuals to be used as the starting frame or reference for generation.
Configurable Output
Supports multiple select-type inputs at generation time, enabling control over output parameters such as aspect ratio, duration, or style mode.
Multimodal Input
Accepts a combination of text, image URLs, and dropdown selections in a single request, supporting flexible prompt construction.
Ready to build with Kling 3.0?
Get Started FreeCommon questions about Kling 3.0
What input types does Kling 3.0 accept?
Kling 3.0 accepts image URLs, text prompts, and multiple select-type configuration inputs, supporting both text-to-video and image-to-video generation workflows.
What is the context window for Kling 3.0?
Kling 3.0 has a context window of 10,000 tokens, which applies to the text input provided when generating a video.
What is the training data cutoff for Kling 3.0?
According to the model metadata, Kling 3.0 has a training date of February 2026.
Does Kling 3.0 support image-to-video generation?
Yes, Kling 3.0 supports image-to-video generation. Users can provide an image URL as input, and the model will generate a video based on that image.
Do I need an API key to use Kling 3.0 on MindStudio?
No API key is required to use Kling 3.0 on MindStudio. The model is available directly through the MindStudio platform.
What people think about Kling 3.0
Community reception to Kling 3.0 on Reddit has been notably positive, with an example post from the official blog generating 880 upvotes and over 200 comments in the r/singularity community. Users appear engaged with the model's output quality based on the high interaction counts.
Kling 3.0 has also been discussed in direct comparisons with other video generation models including Seedance 2.0, Sora 2, and VEO 3.1, suggesting it is being evaluated by users interested in benchmarking current video generation options. These threads reflect active community interest in understanding where Kling 3.0 fits among available video generation tools.
Documentation & links
Parameters & options
Description of what to exclude from the video.
Whether sound is generated simultaneously when generating a video.
Explore similar models
Start building with Kling 3.0
No API keys required. Create AI-powered workflows with Kling 3.0 in minutes — free.