Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

On the Interaction of Compressibility and Adversarial Robustness

Created by
  • Haebom

Author

Melih Barsbey, Ant onio H. Ribeiro, Umut \c{S}im\c{s}ekli, Tolga Birdal

Outline

This paper presents a theoretical analysis and experimental verification of the interplay between the compressibility of neural networks (e.g., neuron-level sparsity, spectral compressibility, etc.) and their robustness against adversarial attacks. We show that compressed neural networks introduce a small number of high-sensitivity directions in the representation space, which can be exploited by adversarial attackers to generate effective interference. We provide simple yet informative bounds on the robustness of $L_\infty$ and $L_2$, and show that the vulnerabilities arise regardless of the compression method (e.g., regularization, structural bias, implicit learning dynamics, etc.). Experimental evaluations on synthetic and real tasks confirm the theoretical predictions, and show that these vulnerabilities persist under adversarial training and transfer learning, contributing to the emergence of universal adversarial interference. In conclusion, we reveal a fundamental tension between structural compressibility and robustness, and suggest new directions for the design of efficient and secure models.

Takeaways, Limitations

Takeaways:
Provides theoretical understanding of the interplay between compressibility and adversarial robustness of neural networks.
We reveal the mechanism of adversarial vulnerability caused by compression.
It shows that vulnerabilities exist regardless of compression method.
It presents a new direction for efficient and safe model design.
Limitations:
Analysis and experiments may be limited to specific types of compression and adversarial attacks.
The presented robustness bounds may be approximate and may not perfectly reflect the actual robustness.
Further validation of generalizability across different neural network architectures and learning strategies is needed.
👍