Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Orthogonal Finetuning Made Scalable

Created by
  • Haebom

Author

Zeju Qiu, Weiyang Liu, Adrian Weller, Bernhard Sch olkopf

Outline

In this paper, we propose OFTv2 to solve the high computational cost and memory usage issues of orthogonal fine-tuning (OFT). Instead of the weight-centric implementation of the traditional OFT, OFTv2 adopts an input-centric approach, which reduces the computational complexity from 3rd order to 2nd order by using matrix-vector multiplication. In addition, we introduce the Cayley-Neumann parameterization, an efficient orthogonal parameterization method to approximate the matrix inverse in the Cayley transform. With these improvements, OFTv2 achieves up to 10x faster training speed and 3x lower GPU memory usage without any performance degradation. In addition, it supports fine-tuning of quantized base models, and demonstrates better training stability, efficiency, and memory usage than QLoRA.

Takeaways, Limitations

Takeaways:
Presenting OFTv2 algorithm that effectively solves OFT's computational cost and memory usage issues
Reduced computational complexity by using input-centric approach using matrix-vector multiplication
Efficient orthogonal parameterization via Cayley-Neumann parameterization
Support for fine-tuning quantized base models and improved performance compared to QLoRA
Achieve 10x faster learning speed and 3x lower GPU memory usage
Limitations:
Since Cayley-Neumann parameterization is a method to approximate the matrix inverse, there is a possibility of performance degradation due to approximation error.
The effectiveness of the proposed method may vary depending on the specific problem and model.
Experimental results for large-scale models may be limited.
👍