Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Towards Black-Box Membership Inference Attack for Diffusion Models

Created by
  • Haebom

Author

Jingwei Li, Jing Dong, Tianxing He, Jingzhao Zhang

Outline

This paper proposes a novel membership inference attack (MIA) method to determine whether a specific image is included in the training data of a diffusion model, addressing copyright issues in AI-generated images. To overcome the limitation of existing MIA methods requiring access to the internal U-net of the model, we propose a method that uses only the image-to-image transformation API to determine whether the training data is included without accessing the internal structure of the model. This method leverages the fact that the model can more easily obtain noisy predictions for the training data. Therefore, the method averages the results using the API multiple times and compares them with the original image. Experiments on DDIM, Stable Diffusion, and Diffusion Transformer architectures demonstrate that our method outperforms existing methods.

Takeaways, Limitations

Takeaways:
A novel method for attacking membership inference on training data of diffusion models is presented.
Attacks are possible using only the API without accessing the model's internal structure.
Effective in various architectures such as DDIM, Stable Diffusion, and Diffusion Transformer.
Better performance than existing methods
Limitations:
API access must be enabled. If API access is restricted, attacks are impossible.
May depend on the specific characteristics of a specific API. Using other APIs may result in performance degradation.
The attack success rate can be affected by various factors, such as the structure of the model, the size of the training data, and the characteristics of the API.
👍