Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

SC-LoRA: Balancing Efficient Fine-tuning and Knowledge Preservation via Subspace-Constrained LoRA

Created by
  • Haebom

Author

Minrui Luo, Fuhang Kuang, Yu Wang, Zirui Liu, Tianxing He

Outline

This paper focuses on the efficient use of Parameter-Efficient Fine-Tuning (PEFT) method, especially Low-Rank Adaptation (LoRA). The traditional LoRA suffers from slow convergence speed and knowledge loss issues, which we address by improving the LoRA initialization method. Unlike previous works that focus only on efficient fine-tuning or knowledge preservation of pre-trained LLMs, this paper proposes Subspace-Constrained LoRA (SC-LoRA) to achieve both goals simultaneously. SC-LoRA is designed to constrain the output of trainable LoRA adapters to a low-dimensional subspace, so that the contextual information of fine-tuning data is preserved as much as possible and the contextual information of existing knowledge is kept as minimal as possible. This allows the training weights to focus on the main features of fine-tuning data while not damaging the existing knowledge. Through theoretical analysis and experimental results on various subtasks, we demonstrate that SC-LoRA provides superior fine-tuning performance and significantly reduces knowledge loss compared to traditional LoRA initialization methods.

Takeaways, Limitations

Takeaways:
SC-LoRA presents a novel initialization framework that effectively addresses the slow convergence speed and knowledge loss issues of LoRA.
It effectively balances the preservation of contextual information in fine-tuned data with the preservation of existing knowledge.
It outperforms existing LoRA initialization methods in various subtasks.
The effectiveness of SC-LoRA is supported by theoretical analysis.
Limitations:
Additional research may be needed on setting subspace constraints for SC-LoRA. A more detailed explanation of how to find the optimal subspace may be needed.
Additional validation of the generalization performance across different LLMs and subtasks may be needed.
There may be a lack of clear comparative analysis on how efficient it is compared to existing LoRA in terms of computational cost.
👍