Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

UniErase: Towards Balanced and Precise Unlearning in Language Models

Created by
  • Haebom

Author

Miao Yu, Liang Lin, Guibin Zhang, Xinfeng Li, Junfeng Fang, Xingrui Yu, Ivor Tsang, Ningyu Zhang, Kun Wang, Yang Wang

Outline

Large-scale language models (LLMs) require iterative updates to maintain information, and LLM unlearning for selective removal becomes crucial during this process. Existing unlearning methods rely on fine-tuning, resulting in low accuracy. They also struggle to balance unlearning effectiveness and general performance in large-scale and sequential environments. In this study, we propose UniErase, a novel unlearning framework that achieves a balance between precision and performance. Unlearning Tokens are introduced to guide LLMs into the forgetting space, and Unlearning Edits efficiently link unlearning targets to these meta-tokens. UniErase demonstrates outstanding performance in batch, sequential, and precision unlearning tasks. In the TOFU benchmark, it outperforms the existing best-performing unlearning method by 4.01x in model performance and 35.96% in unlearning effectiveness.

Takeaways, Limitations

Takeaways:
Proposed Unlearning Token and Unlearning Edit for precise unlearning.
Achieving a balance between maintaining model ability and unlearning effects.
Demonstrated excellent performance on batch, sequential, and precision unlearning tasks.
Outperforms existing methods in TOFU benchmarks.
Limitations:
No specific mention of Limitations in the paper. (Only the abstract is provided, so detailed analysis is not possible.)
👍