This post summarizes a very important livestream with a WAN engineer. It will at least be partially open (model architecture, training code and inference code). Maybe even fully open weights if the community treats them with respect and gratitude, which is also what one of their engineers basically spelled out on Twitter a few days ago, where he asked us to voice our interest in an open model but in a calm and respectful way, because any hostility makes it less likely that the company releases it openly.
The cost to train this kind of model is millions of dollars. Everyone be on your best behaviors. We’re all excited and hoping for the best! I’m already grateful that we’ve been blessed with WAN 2.2 which is already amazing.
PS: The new 1080p/10 seconds mode will probably be far outside consumer hardware reach, but the improvements in the architecture at 480/720p are exciting enough already. It creates such beautiful videos and really good audio tracks. It would be a dream to see a public release, even if we have to quantize it heavily to fit all that data into our consumer GPUs. 😅
submitted by /u/pilkyton
[link] [comments]
📦 : https://github.com/lovisdotio/workflow-magnify-upscale-video-comfyui-lovis I did this ComfyUI workflow for Sora 2 upscaling 🚀 ( or…
Python's flexibility with data types is convenient when coding, but it can lead to runtime…
We revisit scene-level 3D object detection as the output of an object-centric framework capable of…
Editor’s Note: This is the second in a two-part series highlighting demo sessions from AIPCon…
Generative AI has emerged as a transformative technology in healthcare, driving digital transformation in essential…
The conversation around generative AI in the enterprise is getting creative. Since launching our popular…