Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

P2DT: Mitigating Forgetting in task-incremental Learning with progressive prompt Decision Transformer

Created by
  • Haebom

Author

Zhiyuan Wang, Xiaoyang Qu, Jing Xiao, Bokui Chen, Jianzong Wang

Outline

This paper proposes the Progressive Prompt Decision Transformer (P2DT) as a solution to the critical forgetting problem, which causes performance degradation when encountering new tasks in intelligent agents controlled by large-scale models. P2DT promotes task-specific policies by dynamically adding decision tokens during new task learning, enhancing the Transformer-based model. This mitigates forgetting in both continuous and offline reinforcement learning scenarios. Furthermore, P2DT utilizes trajectories collected through existing reinforcement learning across all tasks and generates new task-specific tokens during learning, preserving knowledge from previous learning. Initial results demonstrate that this model effectively mitigates critical forgetting and scales well in an increasing task environment.

Takeaways, Limitations

Takeaways:
A novel method to effectively alleviate the fatal forgetting problem in transformer-based models is presented.
Applicable to both continuous and offline reinforcement learning
Effectively utilize existing learning data to learn new tasks.
Demonstrates excellent scalability even in an increasing task environment
Limitations:
Only initial results are presented, requiring further experimentation and verification.
Further research is needed on generalization performance across diverse environments and tasks.
A detailed analysis of the computational cost and efficiency of P2DT is needed.
👍