Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

SEDEG:Sequential Enhancement of Decoder and Encoder's Generality for Class Incremental Learning with Small Memory

Created by
  • Haebom

Author

Hongyang Chen, Shaoling Pu, Lingyu Zheng, Zhongwu Sun

Outline

This paper proposes SEDEG, a novel method that uses an encoder-decoder architecture to enhance knowledge generalization in incremental learning. SEDEG involves a two-stage training process. In the first stage, an ensemble encoder is trained using feature boosting to learn generalized representations, thereby improving the decoder's generalization ability and the balance between the classifier and the decoder's generalization ability. In the second stage, a knowledge distillation (KD) strategy is used to compress the ensemble encoder and develop a new, more generalized encoder. Effective knowledge transfer is achieved through the balanced KD approach and feature KD. Extensive experiments on three benchmark datasets demonstrate the superior performance of SEDEG, and further experiments confirm the effectiveness of each component.

Takeaways, Limitations

Takeaways:
We present an effective incremental learning method that enhances the generalization ability of both the encoder and decoder, ensuring adaptability and robustness to new data while preserving existing knowledge.
It combines feature enrichment and knowledge distillation to achieve superior performance even in memory-constrained environments.
We demonstrate practical performance improvements by applying it to Vision Transformer (ViT).
Reproducibility has been improved through open source code.
Limitations:
The proposed method is specific to ViT and may have limited applicability to other architectures.
Because the dataset used was limited, generalization performance to other datasets has not been verified.
Experiments on more diverse and complex datasets are needed.
👍