Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

UltraEdit: Training-, Subject-, and Memory-Free Lifelong Editing in Language Models

Created by
  • Haebom

Author

Xiaojie Gu, Ziying Huang, Jia-Chen Gu, Kai Zhang

Outline

This paper proposes UltraEdit, an efficient and scalable model editing method for large-scale language models (LLMs) that adapt to continuous information updates. UltraEdit is a training-, topic-, and memory-free approach that computes parameter changes in a single step using only hidden states and gradients. Furthermore, it employs a continuous regularization strategy to adapt to distributional changes and maintain consistency. UltraEdit achieves editing speeds over 7x faster than existing state-of-the-art methods and is the only method capable of editing a 7B LLM on a 24GB GPU. We build a large-scale dataset called UltraEditBench, which contains over 2 million edit pairs and supports up to 2 million edits, demonstrating excellent performance across a variety of model editing scenarios.

Takeaways, Limitations

Takeaways:
Fast and efficient model editing capabilities (up to 7x faster than existing methods).
Editing large models even on small GPUs (editing a 7B LLM on a 24GB GPU).
Scalability demonstrated using a large dataset (UltraEditBench).
Consistently excellent performance across diverse models and datasets.
Limitations:
There is no Limitations specified in the paper. (However, since it emphasizes superior performance compared to other model editing methods, it may not be optimized for certain types of editing.)
👍