Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Emergent Semantics Beyond Token Embeddings: Transformer LMs with Frozen Visual Unicode Representations

Created by
  • Haebom

Author

A. Bochkov

Outline

Understanding the semantic representation of large language models (LLMs) is crucial for interpretability and architectural innovation. This paper challenges the conventional view that trainable input embeddings serve as fundamental "semantic vectors." In this study, we constructed a Transformer model using fixed, nonsemantic, precomputed visual embeddings derived from the visual structure of Unicode glyphs, rather than the data itself. These models outperformed models using trainable embeddings on the MMLU inference benchmark. We attribute this to "representational interference," where the embedding layer in existing models is burdened by the burden of learning both structural and semantic features. Our findings suggest that high-level semantics are not inherent in the input embeddings, but rather a property resulting from the Transformer's compositional architecture and data scale.

Takeaways, Limitations

Takeaways:
In LLM, meaning is not fixed in the input embeddings, but is the result of the Transformer architecture.
Trainable embeddings can cause “representation interference”, which can degrade model performance.
LLM learns effectively and achieves competitiveness in inference benchmarks with only fixed embeddings based on visual structure.
Redefining the role of embeddings from “semantic containers” to “structural primitives.”
Limitations:
Further evaluation of the general language understanding ability of the model presented in the paper may be necessary.
Research is needed on how Unicode-based embeddings can be applied to other language model architectures and tasks.
A more in-depth analysis of the specific mechanisms of expressive interference is needed.
👍