Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

UAR-NVC: A Unified AutoRegressive Framework for Memory-Efficient Neural Video Compression

Created by
  • Haebom

Author

Jia Wang, Xinfeng Zhang, Gai Zhang, Jun Zhu, Lv Tang, Li Zhang

Outline

This paper presents UAR-NVC (Unified AutoRegressive Framework for memory-efficient Neural Video Compression), a novel framework that applies the frame-by-frame processing of existing video compression frameworks to INRs to address the memory consumption issue in video compression using Implicit Neural Representations (INRs). UAR-NVC integrates INR-based and existing video compression frameworks from a temporal autoregressive modeling perspective by segmenting a video into multiple clips and using a different INR model instance for each clip. We design two modules to optimize the initialization, training, and compression of model parameters to reduce temporal redundancy between clips. The latency can be adjusted by varying the clip length, and experimental results show improved performance compared to various baseline models.

Takeaways, Limitations

Takeaways:
A novel approach to solving the memory problem of existing INR-based video compression is presented.
Integrating INR-based and existing video compression frameworks from a time-based autoregressive modeling perspective.
Flexible waiting time adjustment through clip length adjustment.
Presenting the possibility of efficient video compression in resource-constrained environments.
Performance improvement over previous models.
Limitations:
Further research is needed on the clip segmentation strategy and optimal clip length setting of the proposed method.
Generalization performance evaluation for various video types is needed.
Performance evaluation and optimization in real application environments are required.
👍