Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Function Induction and Task Generalization: An Interpretability Study with Off-by-One Addition

Created by
  • Haebom

Author

Qinyuan Ye, Robin Jia, Xiang Ren

Outline

This paper explores the ability of large-scale language models (LLMs) to perform new tasks through contextual learning. This paper explores the generalization mechanism within LLMs using off-by-one additions (1+1=3, 2+2=5, 3+3=?). Using circuit-style analysis techniques, we analyze the model's internal computation and uncover the principle by which the model generalizes from standard addition to off-by-one addition. Specifically, we discover the +1 function induction mechanism, the +1 function generation through multiple attention heads in parallel, and the reuse of this mechanism across various tasks (e.g., shifted multiple-choice QA and octal addition).

Takeaways, Limitations

Takeaways:
Provides insight into the reusable and configurable internal structure that enables LLM task-level generalization.
+1 Explain the generalization process of the model through the function induction mechanism.
A method for generating +1 functions using parallel attention heads.
Verify the reusability of this mechanism across various tasks.
Limitations:
The specific Limitations is not specified in the paper. (It is impossible to determine based on the provided information alone.)
👍