VrW0MDB1j3bbZJ 3nO FR1aoPIjBj3UUlSad ai6FA
| | Hello. This may not be news to some of you, but Wan 2.1 can generate beautiful cinematic images. I was wondering how Wan would work if I generated only one frame, so to use it as a txt2img model. I am honestly shocked by the results. All the attached images were generated in fullHD (1920x1080px) and on my RTX 4080 graphics card (16GB VRAM) it took about 42s per image. I used the GGUF model Q5_K_S, but I also tried Q3_K_S and the quality was still great. The workflow contains links to downloadable models. Workflow: [https://drive.google.com/file/d/1WeH7XEp2ogIxhrGGmE-bxoQ7buSnsbkE/view] The only postprocessing I did was adding film grain. It adds the right vibe to the images and it wouldn’t be as good without it. Last thing: For the first 5 images I used sampler euler with beta scheluder – the images are beautiful with vibrant colors. For the last three I used ddim_uniform as the scheluder and as you can see they are different, but I like the look even though it is not as striking. 🙂 Enjoy. submitted by /u/yanokusnir |
📦 : https://github.com/lovisdotio/workflow-magnify-upscale-video-comfyui-lovis I did this ComfyUI workflow for Sora 2 upscaling 🚀 ( or…
Python's flexibility with data types is convenient when coding, but it can lead to runtime…
We revisit scene-level 3D object detection as the output of an object-centric framework capable of…
Editor’s Note: This is the second in a two-part series highlighting demo sessions from AIPCon…
Generative AI has emerged as a transformative technology in healthcare, driving digital transformation in essential…
The conversation around generative AI in the enterprise is getting creative. Since launching our popular…