Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Investigating the Impact of Quantization Methods on the Safety and Reliability of Large Language Models

Created by
  • Haebom

Author

Artyom Kharinaev, Viktor Moskvoretskii, Egor Shvetsov, Kseniia Studenikina, Bykov Mikhail, Evgeny Burnaev

Outline

In this paper, we present the OpenMiniSafety safety dataset, consisting of 1,067 difficult questions, to evaluate the safety and reliability impact of quantization techniques for improving the efficiency of large-scale language models (LLMs). Using this dataset, we release 4,268 annotated question-answer pairs for four LLMs (both quantized and precision versions) and evaluate 66 quantized model variants using four post-training quantization (PTQ) and two quantization-aware training (QAT) methods on four safety benchmarks (including human-centered evaluation). Our results show that both PTQ and QAT can degrade safety alignment, while QAT techniques such as QLORA or STE are less safe. We highlight that no single method consistently outperforms the others across benchmarks, precision settings, or models, demonstrating the need for safety-conscious compression strategies. Also, precision-specific methods such as QUIK and AWQ (4-bit), AQLM and Q-PET (2-bit) outperform their target precision, which means that these methods are not better at compression, but rather different approaches.

Takeaways, Limitations

Takeaways:
Providing a standardized benchmark for safety and reliability evaluation of LLM quantization via the OpenMiniSafety dataset.
Presentation of in-depth analysis results on the impact of PTQ and QAT techniques on the safety of LLM.
Validation of the effectiveness of precision-specific quantization techniques.
Emphasizes the need to develop LLM compression strategies that take safety into account.
Limitations:
The types of LLM and quantization techniques used in evaluation may be limited.
Further research is needed on the versatility and generalizability of the OpenMiniSafety dataset.
Since the superiority of a specific quantization technique varies across models and benchmarks, it is difficult to suggest an optimal quantization strategy.
👍