ltx/ltx-2/pro/text-to-video

Generate lifelike 4K videos from text, images, or depth inputs with synchronized audio, multi-keyframe control, and open-source flexibility for cinematic, cost-efficient creative production.

Duration of the generated video in seconds.
Resolution of the generated video.
Aspect ratio of the generated video.
Frames per second of the generated video.
Whether to generate audio for the generated video.

Introduction to LTX 2 AI Video Generator

LTX 2 is the latest open-source video foundation model from Lightricks, launched on October 23, 2025, as a major advance in AI-driven video generation. This next-generation text-to-video and image-to-video model unifies synchronized audio and visual creation in a single pass, supporting native 4K resolution at up to 50fps. With multimodal input compatibility—including text prompts, reference images, depth maps, or even partial video—and creative tools like multi-keyframe conditioning and LoRA fine-tuning, LTX 2 brings professional-grade results to personal workstations. Its improved efficiency makes it up to 50% more cost-effective than competitors, while keeping the entire stack open-source for developers and researchers. LTX 2 text-to-video generation tool empowers you to transform written ideas into lifelike cinematic video clips. It's designed for filmmakers, creators, and teams who need precise control, synchronized audio, and high-quality 4K visuals. With this tool, you quickly turn concepts into ready-to-share visual stories that match your creative intent.

Examples of LTX 2 in Action

Video thumbnail
Loading...
Video thumbnail
Loading...
Video thumbnail
Loading...
Video thumbnail
Loading...
Video thumbnail
Loading...
Video thumbnail
Loading...

What makes LTX 2 stand out

LTX 2 is a high-fidelity text-to-video model engineered for realistic motion, scene coherence, and controllable pacing. Built to translate concise prompts into cinematic sequences, LTX 2 preserves structure, maintains subject integrity, and keeps lighting and perspective consistent across frames. With 4K output, 25 or 50 fps, and optional synchronized audio, LTX 2 supports production-ready delivery while remaining cost-efficient and open-source friendly. The model prioritizes temporal stability over aggressive regeneration, so LTX 2 delivers believable results that hold up under editorial scrutiny. Key capabilities:

  • Multi-keyframe prompting: LTX 2 follows time-anchored directions to evolve scenes across segments.
  • Resolution and fps control: up to 2160p at 25 or 50 fps for broadcast-ready motion; LTX 2 preserves clarity.
  • Duration-locked renders: 6, 8, or 10 seconds for predictable pacing and editorial timing.
  • Synchronized audio toggle: enable or disable audio generation for integrated previews or silent plates.
  • Aspect-ratio consistency: fixed 16:9 framing ensures stable composition and downstream editability.
  • Structure-aware realism: maintains subject identity, depth cues, and plausible lighting throughout the clip.

Prompting guide for LTX 2

Begin with a clear narrative objective, then specify subject, setting, camera behavior, and motion intensity. State duration, resolution, fps, and whether to generate audio. LTX 2 responds well to declarative, time-scoped instructions that separate phases of action. Use concise style cues and avoid conflicting adjectives. When needed, mark time ranges to guide multi-keyframe progression so LTX 2 can stage the scene coherently. For broadcast polish, lock 2160p at 50 fps; for lighter previews, 1080p at 25 fps. LTX 2 benefits from precise verbs and concrete nouns that reduce ambiguity, and LTX 2 respects constraints when you explicitly list what to preserve. Examples:

  • Single-pass cinematic: "8s, 2160p, 25 fps, generate audio: true - sunrise over a misty forest, slow push-in, soft golden light."
  • Action focus: "6s, 1440p, 50 fps - a red sports car accelerates on wet asphalt, camera pan left, rain reflections."
  • Time-scoped cues: "0-2s: quiet alley; 2-6s: neon signs flicker, light rain intensifies, handheld wobble."
  • Mood shift: "10s, 1080p - start calm blue tones, transition to warm amber by 6s, subtle film grain."
  • Dialogue-free emphasis: "8s - silent plate, no audio, medium close-up of a chef plating microgreens." Pro tips:
  • State constraints explicitly: "preserve framing, do not change subject scale."
  • Use spatial language: left, right, foreground, background, center frame.
  • Keep adjectives focused; prefer 3-5 strong descriptors over long lists.
  • Iterate in short steps; refine nouns, verbs, and camera notes between runs.
  • Match parameters to intent: 50 fps for fast motion, audio on for timing previews with LTX 2.

Related Playgrounds

Frequently Asked Questions

What is LTX 2 and what does its text-to-video feature do?

LTX 2 is an open-source video foundation model developed by Lightricks, designed to convert text descriptions into synchronized audio-visual clips using its advanced text-to-video generation system. It supports creating short, high-quality cinematic sequences with both visuals and sound in a unified workflow.

How does LTX 2 text-to-video differ from earlier AI models?

Compared to previous versions and competitors, LTX 2 text-to-video introduces native 4K resolution videos at 50fps with integrated audio. It also includes improved frame consistency, creative keyframing controls, and efficient resource use, making it faster and more cost-effective than earlier generations.

Who can benefit most from using LTX 2 text-to-video?

LTX 2 text-to-video is ideal for filmmakers, VFX artists, advertisers, and creative professionals who need production-ready video content from text prompts. It’s also suitable for independent creators or small teams looking for professional quality without heavy hardware or software investments.

What kind of quality and output should I expect from LTX 2 text-to-video?

LTX 2 text-to-video generates 4K videos at up to 50 frames per second, offering realistic motion, synchronized dialogue and audio, and smooth scene transitions. The detail and temporal stability make it especially useful for storyboarding, short films, and commercial applications.

Is LTX 2 text-to-video free to use or does it require credits?

LTX 2 text-to-video can be accessed through Runcomfy’s AI Playground platform. While new users receive free credits or a trial, continued use requires spending credits based on generation costs as outlined on the Runcomfy website.

What input types does LTX 2 text-to-video support?

In addition to text prompts, LTX 2 text-to-video supports multimodal inputs such as images, depth maps, and short video references. This flexibility allows users to refine motion, camera perspective, and scene composition for customized visual outputs.

On which platforms can I access LTX 2 text-to-video?

You can access LTX 2 text-to-video directly on the Runcomfy website, which works well on mobile and desktop browsers. Once logged in, users can start generating videos using their available credits.

Are there any limitations or current constraints with LTX 2 text-to-video?

Currently, LTX 2 text-to-video supports clips of up to 10 seconds per generation. Longer sequences and expanded editing tools are in development, but users can already chain outputs or fine-tune clips using creative settings like LoRA and multi-keyframe conditioning.

Does LTX 2 text-to-video include synchronized audio or must it be added later?

One of the standout features of LTX 2 text-to-video is that it generates both visuals and audio—dialogue, ambience, and music—in one unified pass, eliminating the need for post-synchronization editing.