An open-source AI model that generates high-quality video content in seconds, built for speed, storytelling and creative control
Redefining What Video Models Can Do
Fast Performance
Generate videos 30 times faster than comparable models, enabling real-time creation and rapid iteration.
High-Quality Output
Utilizes advanced multiscale rendering technology to generate sharp, smooth videos with fewer artifacts.
Built for Efficiency
With 13 billion parameters and built upon a DiT-based architecture, LTXV delivers exceptional video generation capabilities.
Consumer-Friendly Hardware
Optimized to operate efficiently on consumer-grade hardware, including NVIDIA RTX 4090 and 5090 GPUs, with as little as 8GB of VRAM.
Versatile Input Support
Supports multiple input types, text-to-video & image-to-video, allowing for diverse video creation workflows.
LTXV Core features
LoRA Training
Camera Motion
Outpainting
Keyframe Control
The Power Behind the Process
Multiscale Rendering
Our advanced technology initially drafts videos with lower detail to capture the broader motion, and then refines each frame to enhance both speed and quality, providing an optimized experience.
The Fastest Way to Generate Video
LTXV-13B Distilled is our fasted generative video model yetproviding results in as little as 9.5 seconds. The following metrics were tested on H100 at 1280x720.
9.5 seconds
Distilled Pipeline
Prioritizes speed without compromising core visual quality.
20 seconds
Mixed Pipeline (Base + Distilled)
Balances runtime and quality by using the Base model to generate motion and detail, then switching to the Distilled model for faster performance.
43 seconds
Base Pipeline
Focuses on full-fidelity generation, best for final output.
Part of the LTX Studio Ecosystem

Use LTX Video standalone, or through LTX Studio, Lightricks' all-in-one platform for video production.

Automated storyboards
Character & Scene Consistency
Real-time collaboration tools
Part of the LTX Studio Ecosystem

Use LTX Video standalone, or through LTX Studio, Lightricks' all-in-one platform for video production.

Character & Scene Consistency
Automated storyboards
Real-time collaboration tools
FAQS

What is LTX Video?

LTX Video is a family of open-source video generation models built on a transformer-based latent diffusion architecture. It supports text-to-video, image-to-video, keyframe animation, sequence conditioning, and video extension. The 2B distilled version offers faster-than-real-time generation, while the full 13B model delivers higher visual quality with slightly longer, yet comparatively fast runtimes.

How fast is LTX Video?

LTX Video is extremely fast, with some models capable of generating video faster than real-time playback. Exact speeds vary by model and hardware configuration.

Does LTX Video require specialized hardware?

LTX Video runs on a wide range of hardware, from community GPUs to cutting-edge data center accelerators like NVIDIA H100s and Google TPUs. While it’s optimized for modern infrastructure to maximize performance, it’s built to be accessible for independent creators and researchers as well.

Is LTX Video open source?

Yes. The code is released under the Apache 2.0 license, and model weights are available under Lightricks' custom license (see full list here). Both are publicly available for use, research, and development. We invite the community to explore, extend, and contribute.

Can LTX Video generate longer videos?

LTX Video natively supports video extension and keyframe-based generation, allowing you to create longer and more coherent scenes by extending videos forward or backward.

Can I finetune LTX Video for custom use cases?

Yes. LTX Video is designed for customization. Whether you're building for a specific style, domain, or application, you can easily fine-une it using LoRA-based training and multi-GPU support. Ready to make it your own? Start with the official framework here.

What is the difference between the distilled and quantized versions of the model?

You can run the distilled model for fewer iterations (or steps) and still achieve results similar to the full model. This is because the distilled model is separately trained to replicate the behavior of the larger model using a technique called knowledge distillation, where a smaller "student" model learns to mimic the outputs of a larger "teacher" model. The result is a faster, lighter model that retains much of the performance of the original.

The quantized version uses the same original model but compresses its weights into lower-precision formats, reducing memory usage and speeding up inference without retraining.

Both approaches make LTX Video easier to run on limited hardware, with only slight trade-offs in quality, depending on your use case.

What is multiscale rendering, and why is it useful?

Multiscale rendering is a structured pipeline where a video is generated in multiple resolution stages, starting from a low-resolution latent representation and progressively refining it at higher resolutions, while preserving the original structure, motion, and temporal coherence throughout. Rather than generating a high-resolution video all at once, each stage focuses on a different scale of information.

Where can I get help or learn more?

LTX Studio is the future of storytelling, transforming imagination into reality with our AI-driven platform. We streamline the production process from scripting to final edits, making advanced storytelling tools accessible for creators of all levels. Designed for professionals yet intuitive enough for anyone, LTX Studio is where visions come to life, redefining the art of narrative.