Remember when hands and eyes used to be a problem? (Workflow included)

Remember when hands and eyes used to be a problem? (Workflow included)

Disclaimer: This is my second time posting this. My previous attempt had its video quality heavily compressed by Reddit’s upload process.

Remember back in the day when everyone said AI couldn’t handle hands or eyes? A couple months ago? I made this silly video specifically to put hands and eyes in the spotlight. It’s not the only theme of the video though, just prominent.

It features a character named Fabiana. She started as a random ADetailer face in Auto1111 that I right-click saved from a generation. I used that low-res face as a base in ComfyUI to generate new ones, and one of them became Fabiana. Every clip in this video uses that same image as the first frame.

The models are Wan 2.1 and Wan 2.2 low noise only. You can spot the difference: 2.1 gives more details, while 2.2 looks more natural overall. In fiction, I like to think it’s just different camera settings, a new phone, and maybe just different makeup at various points in her life.

I used the “Self-Forcing / CausVid / Accvid Lora, massive speed up for Wan2.1 made by Kijai” published by Ada321. Strength was 1.25 to 1.45 for 2.1 and 1.45 to 1.75 for 2.2. Steps: 6, CFG: 1, Shift: 3. I tried the 2.2 high noise model but stuck with low noise as it worked best without it. The workflow is basically the same for both, just adjusting the LoRa strength. My nodes are a mess, but it works for me. I’m sharing one of the workflows below. (There are all more or less identical, except from the prompts.)

Note: To add more LoRas, I use multiple Lora Loader Model Only nodes.

The music is “Funny Quirky Comedy” by Redafs Music.

LINK to Workflow (ORIGAMI)

submitted by /u/Etsu_Riot
[link] [comments]