Categories: Image

Time-to-Move + Wan 2.2 Test

Made this using mickmumpitz’s ComfyUI workflow that lets you animate movement by manually shifting objects or images in the scene. I tested both my higher quality camera and my iPhone, and for this demo I chose the lower quality footage with imperfect lighting. That roughness made it feel more grounded, almost like the movement was captured naturally in real life. I might do another version with higher quality footage later, just to try a different approach. Here’s mickmumpitz’s tutorial if anyone is interested: https://youtu.be/pUb58eAZ3pc?si=EEcF3XPBRyXPH1BX

submitted by /u/enigmatic_e
[link] [comments]

AI Generated Robotic Content

Share
Published by
AI Generated Robotic Content
Tags: ai images

Recent Posts

Start Your Surround Sound Journey With $50 off This Klipsch Soundbar

This soundbar is just the beginning, with the option to add wireless bookshelf speakers or…

35 mins ago

Researchers pioneer next-generation AI semiconductors with ‘thermal constraining’ technique

A research team led by Professor Taesung Kim from the School of Mechanical Engineering at…

35 mins ago

3 Months later – Proof of concept for making comics with Krita AI and other AI tools

Some folks might remember this post I made a few short months ago where I…

24 hours ago

NASA Delays Launch of Artemis II Lunar Mission Once Again

A failure in the helium flow of the SLS rocket has prompted NASA to delay…

1 day ago

Jailbreaking the matrix: How researchers are bypassing AI guardrails to make them safer

A paper written by University of Florida Computer & Information Science & Engineering, or CISE,…

1 day ago

Turns out LTX-2 makes a very good video upscaler for WAN

I have had a lot of fun with LTX but for a lot of usecases…

2 days ago