Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

CLoQ: Enhancing Fine-Tuning of Quantized LLMs via Calibrated LoRA Initialization

Created by
  • Haebom

Author

Yanxia Deng, Aozhong Zhang, Selcuk Gurses, Naigang Wang, Zi Yang, Penghang Yin

Outline

This paper introduces the Calibrated LoRA Initialization for Quantized LLMs (CLoQ), a widely used method for fine-tuning large-scale language models (LLMs) for efficient subtasks in resource-limited environments. This approach addresses the challenges inherent in applying the Low-Rank Adaptation (LoRA) technique to quantized LLMs. CLoQ focuses on minimizing the layer-by-layer differences between the original LLM and the quantized LLM during the initialization phase. It leverages a small calibration dataset to quantize pre-trained LLMs and determine optimal LoRA components for each layer, thereby establishing a robust foundation for subsequent fine-tuning. One of the key contributions of this study is the presentation of novel theoretical results that enable the precise and closed construction of optimal LoRA components. We experimentally demonstrate the effectiveness of CLoQ across various tasks, including language generation, arithmetic reasoning, and common-sense reasoning, demonstrating its superior performance over existing LoRA fine-tuning methods for quantized LLMs, particularly at ultra-low bit widths.

Takeaways, Limitations

Takeaways:
A novel method is presented to improve the efficiency of LoRA fine-tuning in quantified LLM.
Achieve excellent performance even at ultra-low bit widths.
We present new theoretical results for the precise and closed configuration of optimal LoRA components.
Validation of effectiveness in various subtasks.
Limitations:
The performance improvement of CLoQ may depend on the size and quality of the calibration dataset.
The theoretical results presented may be valid only under certain conditions.
There is a need to examine the generalizability of various quantification techniques.
There is a possibility that the experimental results may be limited to specific tasks and datasets.
👍