Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Promote, Suppress, Iterate: How Language Models Answer One-to-Many Factual Queries

Created by
  • Haebom

Author

Tianyi Lorena Yan, Robin Jia

Outline

This paper analyzes how a language model (LM) answers one-to-many factual questions (e.g., a list of cities in a given country). Using a variety of datasets, models, and prompts, we demonstrate that the LM uses a "promote-then-suppress" mechanism to first recall all answers and then suppress already generated ones. Specifically, the LM uses both the subject and previous answer tokens to recall knowledge, with attention propagating subject information and the multi-level neural network (MLP) promoting answers. Attention then suppresses previous answer tokens by attending to them, and the MLP amplifies the suppression signal. We demonstrate this mechanism through experimental evidence using token lens and knockout techniques. Ultimately, we provide new insights into how the internal components of the LM interact with various input tokens to support complex factual recall.

Takeaways, Limitations

Takeaways:
We have identified a "facilitation-inhibition" mechanism, an internal mechanism that allows language models to answer one-to-many factual questions.
By specifically analyzing the roles of attention and MLP, we clarify the processes of knowledge recall and answer suppression.
We present novel analysis techniques such as token lens and knockout technique.
It provides a new understanding of the complex factual recall process of language models.
Limitations:
The analysis is limited to a specific type of question (one-to-many factual question).
Further research is needed to determine whether the proposed mechanism can be applied to all language models and all types of questions.
The generalizability of the dataset and model used in the analysis needs to be reviewed.
👍