Categories: Image

Radial Attention: O(nlogn) Sparse Attention with Energy Decay for Long Video Generation

We just released RadialAttention, a sparse attention mechanism with O(nlog⁡n) computational complexity for long video generation.

🔍 Key Features:

  • ✅ Plug-and-play: works with pretrained models like #Wan, #HunyuanVideo, #Mochi
  • ✅ Speeds up both training&inference by 2–4×, without quality loss

All you need is a pre-defined static attention mask!

ComfyUI integration is in progress and will be released in ComfyUI-nunchaku!

Paper: https://arxiv.org/abs/2506.19852

Code: https://hanlab.mit.edu/projects/radial-attention

Website: https://hanlab.mit.edu/projects/radial-attention

https://reddit.com/link/1lpfhfk/video/1v2gnr929caf1/player

submitted by /u/Dramatic-Cry-417
[link] [comments]

AI Generated Robotic Content

Share
Published by
AI Generated Robotic Content
Tags: ai images

Recent Posts

Image upscale with Klein 9B

Prompt: upscale image and remove jpeg compression artifacts. Added few hours later: Please note that…

5 hours ago

KV Caching in LLMs: A Guide for Developers

Language models generate text one token at a time, reprocessing the entire sequence at each…

5 hours ago

Learnings from COBOL modernization in the real world

There’s a lot of excitement right now about AI enabling mainframe application modernization. Boards are…

5 hours ago

PayPal’s historically large data migration is the foundation for its gen AI innovation

With the dawn of the gen AI era, businesses are facing unprecedented opportunities for transformative…

5 hours ago

The Latest Repair Battlefield Is the Iowa Farmlands—Again

A new bill that would give farmers in Iowa the right to repair is a…

6 hours ago

Adaptive drafter model uses downtime to double LLM training speed

Reasoning large language models (LLMs) are designed to solve complex problems by breaking them down…

6 hours ago