| | My setup: RTX 3060 12GB VRAM + 48GB system RAM. I spent the last couple of days messing around with LTX-2 inside ComfyUI and had an absolute blast. I created short sample scenes for a loose spy story set in a neon-soaked, rainy Dhaka (cyberpunk/Bangla vibes with rainy streets, umbrellas, dramatic reflections, and a mysterious female lead). Workflow : https://drive.google.com/file/d/1VYrKf7jq52BIi43mZpsP8QCypr9oHtCO/view Each 8-second scene took about 12 minutes to generate (with synced audio). I queued up 70+ scenes total, often trying 3-4 prompt variations per scene to get the mood right. Some scenes were pure text-to-video, others image-to-video starting from Midjourney stills I generated for consistency. Here’s a compilation of some of my favorite clips (rainy window reflections, coffee steam morphing into faces, walking through crowded neon markets, intense close-ups in the downpour): i cleaned up the audio. it had some squeaky sounds. Strengths that blew me away:
Weaknesses / Things to avoid:
Overall verdict: I literally couldn’t believe how two full days disappeared – I was having way too much fun iterating prompts and watching the queue. LTX-2 feels like a huge step forward for local audio-video gen, especially if you lean into atmospheric/illustrative styles rather than high-action. submitted by /u/tanzim31 |
made a short video with LTX-2 using an iCloRA Flow to recreate a Space Jam…
The composition of objects and their parts, along with object-object positional relationships, provides a rich…
Today, we’re announcing structured outputs on Amazon Bedrock—a capability that fundamentally transforms how you can…
As generative AI moves from experimentation to production, platform engineers face a universal challenge for…
The government has withheld details of the investigation of Renee Good’s killing—but an unrelated case…
Inspired by the shape-shifting skin of octopuses, Penn State researchers developed a smart hydrogel that…