While experimenting with the video generation model, I had the idea of taking a picture of my room and using it in the ComfyUI workflow. I thought it could be fun.
So, I decided to take a photo with my phone and transfer it to my computer. Apart from the furniture and walls, nothing else appeared in the picture. I selected the image in the workflow and wrote a very short prompt to test: “A guy in the room.” My main goal was to see if the room would maintain its consistency in the generated video.
Once the rendering was complete, I felt the onset of a panic attack. Why? The man generated in the AI video was none other than myself. I jumped up from my chair, completely panicked and plunged into total confusion as all the most extravagant theories raced through my mind.
Once I had calmed down, though still perplexed, I started analyzing the photo I had taken. After a few minutes of investigation, I finally discovered a faint reflection of myself taking the picture.
submitted by /u/Naji128
[link] [comments]
Continuing the music video u/optimisoprimeo posted: https://www.reddit.com/r/StableDiffusion/comments/1t64gni/so_far_this_is_my_favorite_usecase_for_ltx/ submitted by /u/hidden2u [link] [comments]
One of the major differentiators unlocked by learned codecs relative to their hard-coded traditional counterparts…
As companies of various sizes adopt graphic processing units (GPU)-based machine learning (ML) training, fine-tuning…
Today, we’re thrilled to announce that Gemini 3.1 Flash-Lite, our fastest and most cost-efficient Gemini…
Leaders at the tech giant were skeptical of OpenAI—but wary of pushing it into the…