Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

TIDE: Temporal-Aware Sparse Autoencoders for Interpretable Diffusion Transformers in Image Generation

Created by
  • Haebom

Author

Victor Shea - Jay Huang, Le Zhuo, Yi Xin, Zhaokai Wang, Fu-Yun Wang, Yuchi Wang, Renrui Zhang, Peng Gao, Hongsheng Li

Outline

This paper proposes TIDE (Temporal-Aware Sparse Autoencoders for Interpretable Diffusion Transforms), a novel framework that enhances the interpretability of the less-studied Diffusion Transformer (DiT) compared to U-Net-based diffusion models. TIDE extracts sparse, interpretable activation features from DiT over time, demonstrating that DiT naturally learns hierarchical semantics (e.g., 3D structure, object classes, and detailed concepts) during a large-scale pretraining process. Experimental results demonstrate that TIDE enhances interpretability and controllability while maintaining generation quality, making it suitable for applications such as secure image editing and style transfer.

Takeaways, Limitations

Takeaways:
Improved interpretability of DiT: TIDE enables understanding and control of the inner workings of DiT.
Unraveling Hierarchical Semantic Learning: Unraveling how DiT learns hierarchical semantics during large-scale pretraining.
Introducing safe image editing and style transfer applications: Opening up new application areas through improved interpretability and controllability.
Limitations:
It is unclear how well TIDE performs compared to U-Net-based diffusion models. More extensive comparative experiments are needed.
Further research is needed to determine whether TIDE is applicable to all types of DiT.
There is a need to establish criteria for the accuracy and objective evaluation of rare and interpretable feature extraction.
👍