I’m the Co-founder & CEO of Lightricks. We just open-sourced LTX-2, a production-ready audio-video AI model. AMA.

I’m the Co-founder & CEO of Lightricks. We just open-sourced LTX-2, a production-ready audio-video AI model. AMA.

Hi everyone. I’m Zeev Farbman, Co-founder & CEO of Lightricks.

I’ve spent the last few years working closely with our team on LTX-2, a production-ready audio–video foundation model. This week, we did a full open-source release of LTX-2, including weights, code, a trainer, benchmarks, LoRAs, and documentation.

Open releases of multimodal models are rare, and when they do happen, they’re often hard to run or hard to reproduce. We built LTX-2 to be something you can actually use: it runs locally on consumer GPUs and powers real products at Lightricks.

I’m here to answer questions about:

  • Why we decided to open-source LTX-2
  • What it took ship an open, production-ready AI model
  • Tradeoffs around quality, efficiency, and control
  • Where we think open multimodal models are going next
  • Roadmap and plans

Ask me anything!
I’ll answer as many questions as I can, with some help from the LTX-2 team.

Verification:

Lightricks CEO Zeev Farbman

submitted by /u/ltx_model
[link] [comments]