[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Layerwise Recall and the Geometry of Interwoven Knowledge in LLMs

Created by
  • Haebom

Author

Ge Lei, Samuel J. Cooper

Outline

This study explores how large-scale language models (LLMs) encode interconnected scientific knowledge using chemical elements and the LLaMA family of models as case studies. Our results reveal a three-dimensional helical structure in the hidden state that matches the conceptual structure of the periodic table, suggesting that LLMs can reflect the geometric organization of scientific concepts learned from text. Linear probing revealed that the middle layers encode continuous and nested properties that enable indirect recall, while deeper layers clarify categorical distinctions and integrate linguistic context. These results suggest that LLMs represent symbolic knowledge not as isolated facts but as structured geometric manifolds that interweave semantic information across layers. This study is expected to inspire further exploration of how LLMs represent and infer scientific knowledge, especially in fields such as materials science.

Takeaways, Limitations

Takeaways:
We demonstrate that LLM can encode knowledge by reflecting the geometric organization of scientific concepts.
It reveals that different layers of the LLM represent and process knowledge in different ways (middle layer: continuous properties, deep layer: categorical distinctions and linguistic context).
LLM presents a representation of symbolic knowledge as a structured geometric manifold.
Presenting new directions for research on scientific knowledge representation and inference, including materials science.
Limitations:
Generalization is limited based on only case studies on the LLaMA series models and chemical elements.
Further research is needed on the universality of the three-dimensional helical structure and its extendibility to other scientific fields.
A deeper understanding of the knowledge representation mechanisms of LLM is needed.
👍