Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Tool Unlearning for Tool-Augmented LLMs

Created by
  • Haebom

Author

Jiali Cheng, Hadi Amiri

Outline

This paper presents a novel challenge: "tool unlearning," which removes learning about specific tools from tool-based LLMs. Unlike conventional unlearning, this approach requires removing knowledge itself, rather than individual samples. This approach presents challenges, including the high cost of LLM optimization and the need for a principled evaluation metric. To address these challenges, we propose ToolDelete, the first approach to effectively unlearn tools in tool-based LLMs. ToolDelete implements three key properties for effective tool unlearning and introduces a novel Membership Inference Attack (MIA) model for effective evaluation. Extensive experiments on various tool-training datasets and tool-based LLMs demonstrate that ToolDelete effectively unlearns randomly selected tools while preserving the LLM's knowledge of the remaining tools and its performance on common tasks.

Takeaways, Limitations

Takeaways:
We define a new challenge called tool unlearning in tool-based LLM and present ToolDelete, an effective method for this.
ToolDelete has been experimentally proven to be effective in removing learnings from specific tools due to security vulnerabilities, privacy regulations, or tool deprecation.
A new evaluation method for measuring the effectiveness of tool unlearning is presented through a new MIA model.
Limitations:
The performance evaluation of ToolDelete is dependent on the proposed MIA model and needs to be verified through other evaluation metrics.
Although we present experimental results on various tools and datasets, further research is needed on generalization performance in real-world environments.
Further analysis is needed to determine the potential overall performance degradation of LLM during tool unlearning.
👍