Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Latent Chain-of-Thoughts? Decoding the Depth-Recurrent Transformer

Created by
  • Haebom

Author

Wenquan Lu, Yuechuan Yang, Kyle Lee, Yanshu Li, Enqi Liu

Outline

We investigate whether the Huginn-3.5B, a depth-recurrent Transformer model, exhibits an interpretable latent Chain-of-Thought (CoT) inference structure. We examine the model's internal workings for arithmetic operations using various probing techniques, including Logit Lens and Coda Lens. By tracing the rank trajectories of final and intermediate result tokens, we find limited evidence of an interpretable latent CoT. Furthermore, we demonstrate that significant probing inconsistencies exist between recursion blocks, and that the interpretability of the hidden state varies significantly depending on the layer index and decoding method. We empirically demonstrate that increasing recursion depth yields only marginal benefits, falling short of models that explicitly externalize the inference step.

Takeaways, Limitations

Takeaways:
Huginn-3.5B shows limited interpretable latent CoT inference.
There is a mismatch in probing between recursive blocks, and the interpretability of the hidden state depends on the layer index and the decoding method.
Increasing recursion depth has limited performance gains.
Limitations:
There are limits to the effectiveness of recursive transformers in capturing potential CoTs.
Further research is needed on the consistency and interpretability of internal representations.
There is a performance gap with models that externalize the explicit inference step.
👍