Re-ttention: Ultra Sparse Visual Generation via Attention Statistical Reshape
Abstract
Re-ttention uses temporal redundancy in diffusion models to enable high sparse attention in visual generation, maintaining quality with minimal computational overhead.
Diffusion Transformers (DiT) have become the de-facto model for generating high-quality visual content like videos and images. A huge bottleneck is the attention mechanism where complexity scales quadratically with resolution and video length. One logical way to lessen this burden is sparse attention, where only a subset of tokens or patches are included in the calculation. However, existing techniques fail to preserve visual quality at extremely high sparsity levels and might even incur non-negligible compute overheads. % To address this concern, we propose Re-ttention, which implements very high sparse attention for visual generation models by leveraging the temporal redundancy of Diffusion Models to overcome the probabilistic normalization shift within the attention mechanism. Specifically, Re-ttention reshapes attention scores based on the prior softmax distribution history in order to preserve the visual quality of the full quadratic attention at very high sparsity levels. % Experimental results on T2V/T2I models such as CogVideoX and the PixArt DiTs demonstrate that Re-ttention requires as few as 3.1\% of the tokens during inference, outperforming contemporary methods like FastDiTAttn, Sparse VideoGen and MInference. Further, we measure latency to show that our method can attain over 45\% end-to-end % and over 92\% self-attention latency reduction on an H100 GPU at negligible overhead cost. Code available online here: https://github.com/cccrrrccc/Re-ttention{https://github.com/cccrrrccc/Re-ttention}
Community
Re-ttention, a sparse attention method that requires as few as 3.1% of tokens during inference, leverages the temporal redundancy of Diffusion Models to reshape attention scores in order to preserve the visual quality.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- RainFusion: Adaptive Video Generation Acceleration via Multi-Dimensional Visual Redundancy (2025)
- Training-Free Efficient Video Generation via Dynamic Token Carving (2025)
- Grouping First, Attending Smartly: Training-Free Acceleration for Diffusion Transformers (2025)
- Model Reveals What to Cache: Profiling-Based Feature Reuse for Video Diffusion Models (2025)
- Attend to Not Attended: Structure-then-Detail Token Merging for Post-training DiT Acceleration (2025)
- Turbo2K: Towards Ultra-Efficient and High-Quality 2K Video Synthesis (2025)
- Fine-Tuning Visual Autoregressive Models for Subject-Driven Generation (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper