Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Steering LLM Reasoning Through Bias-Only Adaptation

Created by
  • Haebom

Author

Viacheslav Sinii, Alexey Gorbatovski, Artem Cherepanov, Boris Shaposhnikov, Nikita Balagansky, Daniil Gavrilov

Outline

This paper presents a method that uses reinforcement learning to train a single $d$-dimensional steering vector per layer, while keeping the basis weights fixed. This method achieves performance comparable to that of a fully RL-tuned inference model on a mathematical reasoning task. The additional parameterization is only about 0.0016% of the 8 billion-parameter model, and the performance is reproducible across a variety of basis models and mathematical reasoning benchmarks. These results narrow the upper bound on the parameter budget required for high-dimensional thought chain inference, suggesting that millions of adapter weights are unnecessary. The minimal trainable space reduces communication between the optimizer memory and the GPU, lowering the overall cost of fine-tuning. Furthermore, logit-lens analysis demonstrates that the learned vectors amplify consistent token orientations, providing clear insight into the model's internal computation.

Takeaways, Limitations

Takeaways: An efficient parameter learning method for high-dimensional chain inference is presented. Compared to existing methods, it achieves equivalent performance with a minimal number of additional parameters. This reduces fine-tuning costs. This also enhances understanding of the model's internal workings.
Limitations: Further research is needed to determine the generalizability of the proposed method. Performance evaluations on various types of reasoning tasks are needed. The results may be limited to specific types of mathematical reasoning.
👍