Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

On the Necessity of Output Distribution Reweighting for Effective Class Unlearning

Created by
  • Haebom

Author

Ali Ebrahimpour-Boroojeny, Yian Wang, Hari Sundaram

Outline

This paper addresses the problem of privacy leakage in class unlearning evaluations by overlooking the fundamental class geometry, and presents a simple yet effective solution. Specifically, we propose a membership inference attack (MIA-NN) that detects unlearned samples by exploiting the probabilities assigned by the model to neighboring classes. Furthermore, we propose a novel fine-tuning objective, the Tilted ReWeighting (TRW) distribution, which mitigates privacy leakage by approximating the distribution of the remaining classes generated by the retrained model. Experimental results show that TRW outperforms existing unlearning methods.

Takeaways, Limitations

Takeaways:
We point out that the possibility of personal information leakage can be overlooked in class unlearning evaluations and propose a new methodology to address this.
A proposed methodology for detecting unlearned samples through membership inference attacks (MIA-NN).
A new fine-tuning objective utilizing the Tilted ReWeighting (TRW) distribution.
Demonstrated improved performance over existing unlearning methods on several benchmarks, including CIFAR-10.
Limitations:
The generalizability of the methodology presented in this paper and its performance evaluation on other datasets are needed.
Further research is needed on the detailed implementation and optimization of the TRW method.
The need to develop defense techniques against new attack methodologies
👍