Nº 059 · AI ·10 min read · March 28, 2026

I Tested LTX-Video 2 for Commercial Production. Here's My Honest Take.

Fig. 01 I Tested LTX-Video 2 for Commercial Production. Here's My Honest Take.

Why LTX-Video 2 Deserves Your Attention

I've spent the last week putting LTX-Video 2 through real production tests. Not demo prompts. Not "generate a sunset" challenges. Actual briefs of the type I used to shoot with a crew: product reveals, emotional brand moments, lifestyle footage for social campaigns. The kind of work where every second costs money and every creative decision has a client consequence.

The results surprised me — not because LTX-Video 2 is magic, but because it's genuinely different from Kling and Sora in ways that matter to working directors. Here's my unfiltered assessment.

What Is LTX-Video 2?

LTX-Video is an open-source video generation model developed by Lightricks — the company behind Facetune and other creative tools. Unlike Kling (ByteDance) or Sora (OpenAI), LTX-Video is fully open: the model weights are public, you can run it locally, and it's available cheaply through inference platforms like fal.ai.

Version 2 represents a significant quality jump from the original. The motion is smoother, the temporal consistency is better — meaning objects and people don't morph weirdly between frames — and the model handles cinematic prompts more reliably than its predecessor. The "fast" configuration, which runs through cloud APIs, produces a 5-second clip in roughly 30-60 seconds at a cost of around $0.02-0.05 per generation.

For context: Kling 2.1 costs significantly more per clip and requires Chinese platform accounts. Sora is still limited in access. LTX-Video 2 is right now the most accessible high-quality option for production-grade testing.

The Tests I Ran

Test 1: Product Reveal (Cosmetics)

Brief: A foundation bottle emerging from soft morning light, with warm golden bokeh in the background. The kind of shot you'd spend half a day setting up with a beauty specialist.

Prompt: "Product reveal shot: a sleek cosmetics bottle slowly rising from white surface, soft golden morning light from behind creating warm bokeh, shallow depth of field, clean white negative space, luxurious and minimal, slow motion, 9:16 vertical, cinematic"

Result: Surprisingly clean. The motion was smooth. The lighting rendered well. The bottle had some temporal inconsistency — the label shifted slightly between frames — but for a social cutdown or a B-roll insert, this would be usable. Not hero footage. Usable insert.

Test 2: Emotional Human Moment

Brief: A woman looking out a rain-streaked window, contemplative, warm interior light against cold exterior. Classic brand mood film material.

Prompt: "Medium shot of woman gazing through rain-streaked window, warm amber interior light, cold blue exterior, reflection visible in glass, contemplative expression, handheld subtle movement, Terrence Malick visual language, film grain, 9:16"

Result: This is where LTX-Video 2 gets complicated. The composition was good. The mood was right. But the face had the classic AI problem: micro-expressions that feel slightly mechanical, like the model is approximating human emotion rather than capturing it. In a 5-second clip where you see the back of someone's head, you'd never notice. In a close-up with the face visible, you would. The rule I've established: use AI-generated clips for atmosphere, not for faces.

Test 3: Landscape / Establishing Shot

Brief: Aerial drift over misty mountain forest at dawn. The kind of shot that costs $2,000 in drone permits and crew time.

Prompt: "Aerial slow drift over ancient forest at dawn, morning mist in valleys, golden light breaking through trees, serene and vast, cinematic scale, slow motion, Roger Deakins lighting palette"

Result: This was the strongest test. LTX-Video 2 handles landscape and nature footage significantly better than human subjects. The mist movement was organic, the light transition was beautiful, and there were no consistency artifacts. I would use this in a real project. Not as hero footage for a national broadcast, but absolutely for a digital campaign, a social video, or a pitch treatment.

LTX-Video 2 vs. Kling 2.1: The Real Comparison

I ran parallel tests on both models with identical prompts. The honest comparison:

  • Motion quality: Kling 2.1 still edges out LTX-Video 2 for complex motion, especially human movement. The difference is noticeable on close-up action.
  • Atmospheric and environmental footage: LTX-Video 2 is competitive or better. Landscapes, mist, light effects — this is where open-source has caught up.
  • Cost: LTX-Video 2 is dramatically cheaper. For a 10-clip project, the cost difference can be $2 vs. $20+. At scale, that compounds.
  • Accessibility: LTX-Video 2 wins completely. No waitlists, no Chinese platform accounts, available instantly through fal.ai or runnable locally.
  • Iteration speed: Because it's cheaper and faster, you can run 20 LTX-Video 2 variations in the time you'd run 3 Kling clips. This changes the creative workflow significantly.

My Production Workflow with LTX-Video 2

After this week of testing, here's how I'm integrating LTX-Video 2 into actual work:

Phase 1 — Concept exploration: Use LTX-Video 2 to quickly generate mood references for client presentations. 5-10 clips in 30 minutes, showing the visual direction before any real budget is committed. This is the equivalent of a rough storyboard but moving. Clients respond better to motion than static frames.

Phase 2 — B-roll and atmosphere: For projects where I'm shooting primary footage but need supplemental material — establishing shots, transitions, abstract visual elements — LTX-Video 2 fills those gaps at near-zero cost.

Phase 3 — Full AI campaigns: For social-first content where the brief is "we need 30 pieces of content per month," combining LTX-Video 2 clips with AI-generated images and a strong editorial direction produces work that is indistinguishable from low-budget production on a phone screen. Not on a broadcast monitor. On Instagram.

The key is always the same: clear creative direction before prompting. The tool amplifies your vision. It cannot replace it.

The Honest Limitations

I won't oversell this. LTX-Video 2 has real limitations:

  • Human faces in close-up remain problematic. The uncanny valley is real and visible.
  • Complex action sequences — sports, fast movement, detailed hand gestures — are inconsistent.
  • Brand-specific details (logos, specific products, custom assets) don't transfer reliably.
  • 5-second clips require creative editing to build anything longer. This is a constraint, not a dealbreaker — good editors work with constraints.

These limitations define where AI video belongs in the production stack: as a capable tool for specific applications, not as a replacement for production on projects where faces, action, and brand precision matter.

What This Means for the Next 12 Months

I've been through enough technology cycles to recognize the pattern. What LTX-Video 2 can do today is roughly equivalent to what affordable drone footage could do in 2016: impressive for the price, limited in application, but pointing clearly toward a near future where those limitations disappear.

The studios that adapted to drone footage — learning to prompt operators, integrating it into workflows, building client expectations appropriately — are the ones who benefited when the technology matured. The same dynamic is playing out now with AI video.

The question isn't whether to adopt AI video generation. It's whether to adopt it while the learning curve is steep (and the competitive advantage is real) or after it's become table stakes. I know which side I want to be on.

LTX-Video 2 is available on fal.ai now. Run a test. The cost of entry is less than a coffee.

About the author

Read the manifesto Write in