Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Inducing Uncertainty on Open-Weight Models for Test-Time Privacy in Image Recognition

Created by
  • Haebom

Author

Muhammad H. Ashiq, Peter Triantafillou, Hung Yun Tseng, Grigoris G. Chrysos

Outline

We conducted research on how to prevent AI model users from using erroneous personal data to harm others. Specifically, for open-weight models, simply masking model outputs is not sufficient to prevent harmful predictions. In this study, we introduce the concept of test-time privacy and propose an algorithm that maximizes uncertainty for protected instances while maintaining accuracy for the remaining instances. This algorithm utilizes a Pareto-optimal objective that balances test-time privacy and utility, and provides a certifiable approximation algorithm that achieves the $(\varepsilon, \delta)$ guarantee without convexity assumptions. Furthermore, we prove a tight bound characterizing the privacy-utility tradeoff induced by the algorithm. Experimental results show that the proposed method achieves at least three times stronger uncertainty control than pretraining on image recognition benchmarks without compromising accuracy.

Takeaways, Limitations

Takeaways:
A novel approach to addressing privacy concerns during AI model testing.
We propose an algorithm that balances test time privacy and utility.
Ensuring privacy through certifiable approximation algorithms.
Proving the effectiveness of the algorithm through experiments.
Providing tools to improve the safety of AI models.
Limitations:
The benchmarks tested are limited to image recognition.
Further research is needed on the generalization performance of the algorithm.
In-depth research is needed to analyze specific privacy-utility tradeoffs.
👍