Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

SelfReflect: Can LLMs Communicate Their Internal Answer Distribution?

Created by
  • Haebom

Author

Michael Kirchhof, Luca F uger, Adam Goli nski, Eeshan Gunesh Dhekane, Arno Blaas, Seong Joon Oh, Sinead Williamson

Outline

A common way to convey uncertainty in large-scale language models (LLMs) is to add percentages or euphemisms to the answers. In this paper, we argue that LLMs should output a summary of all possible options and their probabilities, reflecting their internal belief distributions. To test whether LLMs possess this ability, we developed a metric called SelfReflect, which measures the information-theoretic distance between the answer distribution and the summary. Experiments show that while current LLMs fail to reveal uncertainty, sampling multiple outputs and re-introducing them into context allows them to produce faithful summaries of uncertainty.

Takeaways, Limitations

Takeaways:
We propose a new method to quantitatively evaluate the uncertainty representation ability of LLM through the SelfReflect index.
We have now found that LLM does not effectively reveal uncertainty.
We found that we could improve the uncertainty representation capability of LLM by simply sampling multiple outputs and feeding them back into context.
Limitations:
The performance of the SelfReflect metric may vary depending on the internal structure of the LLM and the training data.
Further research is needed to determine how effective sampling multiple outputs is in real-world settings.
Further validation is needed to determine whether the proposed method is applicable to all types of uncertainty.
👍