Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Learning to summarize user information for personalized reinforcement learning from human feedback

Created by
  • Haebom

Author

Hyunji Nam, Yanming Wan, Mickel Liu, Jianxun Lian, Peter Ahnn, Natasha Jaques

Outline

PLUS (Preference Learning Using Summarization) is a novel framework developed for personalized responses from LLM AI assistants. It overcomes the limitations of reinforcement learning from human feedback (RLHF) and generates personalized responses for each user by summarizing each user's preferences, characteristics, and past conversations. PLUS operates through an online co-adaptation loop that simultaneously trains a user summary model and a reward model. It delivers robust performance on new users and conversation topics, zero-shot personalization comparable to models like GPT-4, flexible user context learning, and interpretable user representations.

Takeaways, Limitations

Takeaways:
Offering personalized LLM responses tailored to user preferences
Powerful performance even for new users and conversation topics
Zero-shot personalization support for the latest models, including GPT-4.
Gain flexibility by learning from diverse user contexts
Enhance transparency and user control through interpretability of user expressions.
Limitations:
There is no direct mention of Limitations in this paper.
👍