Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

GRID: Scalable Task-Agnostic Prompt-Based Continual Learning for Language Models

Created by
  • Haebom

Author

Anushka Tiwari, Sayantan Pal, Rohini K. Srihari, Kaiyi Ji

Outline

Prompt-based continuous learning (CL) offers a parameter-efficient approach for adapting large-scale language models (LLMs) to task sequences. However, most existing methods rely on task-aware inference and maintain a growing set of task-specific prompts, leading to two major challenges: (1) significant performance degradation relative to previous tasks under task-agnostic inference, and (2) limited scalability due to prompt memory accumulation as task sequences grow. In this paper, we present GRID, an integrated framework designed to address these challenges. GRID integrates a decoding mechanism that utilizes representative inputs to improve backtransfer, automatic task identification, and constrained decoding. Furthermore, it employs a gradient-based prompt selection strategy to compress less informative prompts into a single, unified representation, ensuring scalable and memory-efficient continuous learning. Extensive experiments on long-sequence and negative transfer benchmarks demonstrate that GRID improves average accuracy and backtransfer, achieves competitive backtransfer, and substantially reduces prompt memory usage.

Takeaways, Limitations

Takeaways:
Improved performance degradation over previous tasks even under task-agnostic inference.
Addresses scalability issues caused by prompt memory accumulation.
Improved average accuracy and reversibility.
Competitive power outage achieved.
Significantly reduced prompt memory usage.
Limitations:
No Limitations is mentioned in the paper. (Unknown from the paper summary alone)
👍