Kling O1: Precision Video-to-Video Editing & Scene Consistency on playground and API | RunComfy

kling/kling-video-o1/video-to-video/edit

Transform existing footage into cinematic new scenes with Kling O1, a multimodal video-to-video model enabling seamless editing, consistent characters, and fast creative production for filmmakers and digital artists.

Use @Element1, @Element2 to reference elements and @Image1, @Image2 to reference images in order.
Reference video URL to guide motion and scene framing. Supported formats: .mp4, .mov. Duration: 3–10 seconds. Resolution: 720p–2160p. File size: ≤200 MB.
Elements 1
The frontal image of the element (main view).Max file size: 10.0MB, Min width: 300px, Min height: 300px, Min aspect ratio: 0.40, Max aspect ratio: 2.50, Timeout: 20.0s
Image 1
Additional reference images from different angles. 1-4 images supported. At least one image is required.
Provide characters/objects to include. Reference in prompt as @Element1, @Element2, etc. Maximum 7 total across elements + reference images + start image.
Reference images for style or appearance guidance. Referenced in the prompt as @Image1, @Image2, etc. Up to 4 images total. When using video, elements and reference images combined must not exceed 4.
Whether to keep the original audio track from the input video.

Introduction to Kling O1 Video Generator

Developed by Kuaishou Technology, Kling O1 is a unified multimodal video foundation model that empowers creators and teams to transform existing footage through precise video-to-video generation and seamless editing. Designed for filmmakers, marketers, and digital artists, Kling O1 ensures consistent character, scene, and style across shots while dramatically accelerating creative workflows. For developers, Kling O1 on RunComfy can be used both in the browser and via an HTTP API, so you don’t need to host or scale the model yourself.

Examples Generated Using Kling O1

Video thumbnail
Loading...
Video thumbnail
Loading...
Video thumbnail
Loading...
Video thumbnail
Loading...
Video thumbnail
Loading...
Video thumbnail
Loading...

Related Playgrounds

Frequently Asked Questions

Can I use Kling O1 video-to-video outputs for commercial projects?

Kling O1 allows video-to-video generation and editing under a license that typically follows an Open RAIL-style non-commercial or limited-commercial framework. Using Kling O1 through RunComfy does not override or bypass the original license. Always verify the official Kling O1 license before applying generated outputs in paid or brand-affiliated projects.

What type of license governs Kling O1 video-to-video use on RunComfy?

Kling O1 is distributed under the original license specified by Kuaishou Technology (currently aligned with an Open RAIL-like model). When using Kling O1 video-to-video capabilities via RunComfy, users must still comply with the model’s terms. RunComfy’s hosting only provides managed access and does not transfer or extend commercial rights.

How does RunComfy handle performance and latency for Kling O1 video-to-video generation?

RunComfy’s managed infrastructure distributes Kling O1 video-to-video requests across multiple cloud GPUs, ensuring low latency and stable throughput for concurrent users. Local runs of Kling O1 may require A100-class GPUs and are not recommended for high-volume workloads. The platform maintains dynamic scaling to balance efficiency and responsiveness.

Are there technical limits when using Kling O1 video-to-video features?

Yes. Kling O1 enforces maximum output resolution up to 1080p and supports video durations roughly between 3–10 seconds per generation cycle. Up to 10 reference images or short clips can be used for video-to-video consistency. Prompt token limits align with the RunComfy API cap, which currently allows around 1000 characters per request.

How can I transition my Kling O1 video-to-video experiments from the Playground to API production?

To migrate Kling O1 video-to-video workflows, first finalize prototype results in the RunComfy Playground. Afterward, obtain an API key and replicate your configuration via the RunComfy REST or Python interface. The API offers the same output fidelity as the web interface but allows integration into scripts, CMS pipelines, or app backends.

What distinguishes Kling O1 video-to-video from earlier models?

Kling O1 unifies generation and editing within one multimodal engine, improving consistency of characters and scenes. Compared with prior models, Kling O1 video-to-video excels at handling scene continuity, start/end-frame control, and reference-based identity preservation. This reduces content drift often seen in earlier text-to-video systems.

Can I run Kling O1 video-to-video locally instead of through RunComfy?

Technically yes, but running Kling O1 video-to-video locally demands high-end GPUs (A100/RTX 4090 or higher) and substantial VRAM. RunComfy’s managed environment handles GPU provisioning, batching, and automatic checkpoint updates, making it more efficient and reliable for most users.

Does RunComfy’s execution of Kling O1 video-to-video consume cloud credits or local resources?

All Kling O1 video-to-video generations on RunComfy are processed using cloud GPUs. Users spend platform ‘usd’ credits per render. New accounts receive trial usd, which can be replenished through the billing menu. No local hardware resources are consumed when using the web or API services.

What should I do if Kling O1 video-to-video generation fails or times out?

If Kling O1 video-to-video generation fails, verify prompt complexity, reduce reference inputs, and ensure network stability. Occasionally high busyness or quota limits may cause task timeouts. Contact hi@runcomfy.com for assistance, providing your task ID and configuration for support escalation.