I made a free tool to create manga/webtoon easily using 3D + AI. It supports local generation using Forge or A1111. It’s called Bonsai Studio, would love some feedback!
submitted by /u/alvaro_rami [link] [comments]
Category Added in a WPeMatico Campaign
submitted by /u/alvaro_rami [link] [comments]
Here is the Episode 3 of my AI sci-fi film experiment. Earlier episodes are posted here or you can see them on www.youtube.com/@Stellarchive This time I tried to push continuity and dialogue further. A few takeaways that might help others: Making characters talk is tough. Huge render times and often a small issue is enough …
Read more “Experimenting with Continuity Edits | Wan 2.2 + InfiniteTalk + Qwen Image Edit”
Last week I built a local pipeline where a state machine + LLM watches my security cam and yells at Amazon drivers peeing on my house. State machine is the magic: it flips the system from passive (just watching) to active (video/audio ingest + ~1s TTS out) only when a trigger hits. Keeps things deterministic …
Read more “Made a local AI pipeline that yells at drivers peeing on my house”
I made a workflow for detailing faces in videos (using Impack-Pack). Basically, it uses the Wan2.2 Low model for 1-step detailing, but depending on your preference, you can change the settings or may use V2V like Infinite Talk. Use, improve and share your results. !! Caution !! It uses loads of RAM. Please bypass Upscale …
Decided to share some examples of images I got in Qwen with my LoRA for realism. Some of them look pretty interesting in terms of anatomy. If you’re interested, you can get the workflow here. I’m still in the process of cooking up a finetune and some style LoRAs for Qwen-Image (yes, so long) submitted …
submitted by /u/CeFurkan [link] [comments]
*****Her name is “Winona Ryder” – I misspelled it in the post title thinking it was spelled like Wynonna Judd. Reddit doesn’t allow you to edit Post Titles, so my mistake is now entrenched unless I delete and repost. Oops. I guess I can correct it if I cross post this in the future. I’ve …
Installed VibeVoice using the wrapper this dude created. https://www.reddit.com/r/comfyui/comments/1n20407/wip2_comfyui_wrapper_for_microsofts_new_vibevoice/ Workflow is the multi-voice example one can find in the module’s folder. Asked GPT for a harmless talk among those 3 people, used 3 1-minute audio samples, mono, 44KHz .wav Picked the 7B model. My 3060 almost died, took 54 minutes, but she didn’t croak an …
Read more “VibeVoice is crazy good (first try, no cherry-picking)”
RTX 4090 48G Vram Model: wan2.1_i2v_720p_14B_fp8_scaled Lora: lightx2v_I2V_14B_480p_cfg_step_distill_rank256_bf16 Resolution: 1280×720 frames: 81 *49 / 3375 Rendering time: 5 min *49 / 245min Steps: 4 Vram: 36 GB ————————– Song Source: My own AI cover https://youtu.be/9ptZiAoSoBM Singer: Hiromi Iwasaki (Japanese idol in the 1970s) https://en.wikipedia.org/wiki/Hiromi_Iwasaki submitted by /u/Realistic_Egg8718 [link] [comments]
I just started learning video editing (Davinci Resolve) and Ai Video generation using Wan 2.2, LTXV, and Framepack. As a learning exercise, I thought it would be fun to throw together a morph video of some of Harrison Ford’s roles. It isn’t in any chronological order, I just picked what I thought would be a …
Read more “Starring Harrison Ford – A Wan 2.2 First Last Frame Tribute using Native Workflow.”