Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Verbalized Sampling: How to Mitigate Mode Collapse and Unlock LLM Diversity

Created by
  • Haebom

Author

Jiayi Zhang, Simon Yu, Derek Chong, Anthony Sicilia, Michael R. Tomz, Christopher D. Manning, Weiyan Shi

Outline

This paper reveals that the fundamental cause of mode collapse, a phenomenon in which diversity in LLM decreases during the post-training alignment process, lies not in algorithmic limitations, but in typicality bias within the preference data. The researchers theoretically established this finding and empirically verified it on a preference dataset. Based on this analysis, they propose Verbalized Sampling (VS), a simple, training-free prompting strategy that circumvents mode collapse. VS prompts the model to verbalize the probability distribution over a series of responses. Through various experiments, they demonstrate that VS significantly improves performance in various domains, including creative writing, conversation simulation, open-ended question-answering, and synthetic data generation, and particularly significantly increases diversity in the creative writing domain. Furthermore, they found that the effect of VS tends to be greater for better-performing models.

Takeaways, Limitations

Takeaways:
We present a new data-driven perspective on the mode collapse phenomenon in LLM.
We propose Verbalized Sampling, a practical, inference-time solution that can be applied without training, improving on pre-trained generative diversity.
The effectiveness of VS has been proven in various fields such as creation, conversation simulation, and open Q&A.
We observed that as the model performance improved, the effect of VS tended to increase.
Limitations:
The effectiveness of VS may vary with different prompt strategies or different types of datasets.
There is no guarantee that VS can perfectly resolve mode collapse in all cases.
Further analysis may be needed to determine why VS shows greater advantages in certain models.
👍