Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Can Less Precise Be More Reliable? A Systematic Evaluation of Quantization's Impact on CLIP Beyond Accuracy

Created by
  • Haebom

Author

Aymen Bouguerra, Daniel Montoya, Alexandra Gomez-Villa, Fabio Arnez, Chokri Mraidha

Outline

We evaluate the impact of quantization on the performance of the Vision-Language Model (VLM) CLIP at scale. By comprehensively evaluating reliability metrics as well as accuracy, we find counterintuitive results depending on the pretraining source. Quantization consistently improves the calibration of underconfident pretrained models and tends to degrade the calibration of overconfident variants. We demonstrate that out-of-distribution (OOD) detection can be improved despite the degraded calibration, and that specific quantization-aware training (QAT) methods provide simultaneous benefits in accuracy, calibration, and OOD robustness.

Takeaways, Limitations

Quantization changes the calibration performance differently depending on the calibration characteristics of the pre-trained model.
OOD detection performance can be improved even with compensation degradation.
QAT mitigates the trade-off between efficiency and performance, enabling improvements in accuracy, calibration, and OOD robustness.
Provides insights into leveraging quantization for efficient and reliable deployment of CLIP models.
Details of the specific QAT methodology used in the paper were not provided.
The generalizability of the results to specific pre-trained models and datasets requires further research.
👍