Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

LLM Probability Concentration: How Alignment Shrinks the Generative Horizon

Created by
  • Haebom

Author

Chenghao Yang, Ari Holtzman

Outline

We investigated the lack of diversity in the outputs of aligned large-scale language models (LLMs) from the perspective of the concentration of their probability distributions. To quantify this, we introduced the Branching Factor (BF), a token invariant that measures the number of possible tokens in the next stage. Experimental results show that (1) the BF decreases as the generation progresses, making the LLM more predictable, and (2) alignment tuning significantly narrows the model's output distribution, reducing the BF. This finding explains why aligned models are less sensitive to decoding strategies and why aligned CoT models generate long inference chains and produce stable outputs. Alignment tuning does not fundamentally change the model's behavior, but rather induces the model to select style tokens that open up low-entropy trajectories already present in the base model.

Takeaways, Limitations

Takeaways:
BF is a useful indicator for understanding and controlling LLM output.
Alignment tuning reduces diversity.
CoT promotes stable output.
The base model can control diversity using style tokens.
Limitations:
The specific Limitations is not specified in the paper.
👍