Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

BWCache: Accelerating Video Diffusion Transformers through Block-Wise Caching

Created by
  • Haebom

Author

Hanshuai Cui, Zhiqing Tang, Zhifei Xu, Zhi Yao, Wenyi Zeng, Weijia Jia

Outline

This paper presents a method to accelerate the processing of the Diffusion Transformer (DiT), a cutting-edge technique in video generation. DiT suffers from slow processing speeds due to its sequential noise removal process, and existing acceleration methods suffer from performance degradation or difficulties in reusing intermediate features. By analyzing the feature evolution patterns of DiT blocks, we discover that intermediate stages exhibit high feature similarity. Based on this analysis, we propose Block-Wise Caching (BWCache), a novel acceleration technique that requires no learning. BWCache dynamically caches and reuses features in DiT blocks, minimizing unnecessary computation while maintaining visual quality through a similarity metric. Experimental results demonstrate up to a 2.24x speedup across multiple video diffusion models.

Takeaways, Limitations

Takeaways:
We present a novel method (BWCache) to effectively improve the processing speed of DiT-based video generation models.
A training-free method that requires no learning, increasing ease of application.
Achieving speed improvements without compromising visual quality by leveraging similarity metrics.
Experimentally verifying applicability to various video diffusion models.
Limitations:
The degree of performance improvement of BWCache may vary depending on the video diffusion model and dataset used.
Further research may be needed to optimize the caching mechanism.
Possibility of increased memory usage.
Further research may be needed to determine specific similarity thresholds.
👍