Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Meta-Learning for Cold-Start Personalization in Prompt-Tuned LLMs

Created by
  • Haebom

Author

Yushang Zhao, Huijie Shen, Dannier Li, Lu Chang, Chengrui Zhou, Yinuo Yang

Outline

This paper proposes a meta-learning framework to address the problem that generative, explainable, and flexible recommender systems based on large-scale language models (LLMs) are not suitable for cold-start user situations (i.e., situations where there is little or no interaction history). Existing supervised learning fine-tuning and collaborative filtering methods are effective when there is a large amount of user-item data, but they suffer from high maintenance and update costs. In this paper, we propose a meta-learning framework that treats each user as a task and learns soft prompt embeddings using Reptile and MAML optimization. The learned vectors are added to the input tokens as differentiable control variables representing user behavior priors. We meta-optimize the prompts by episode sampling, inner-loop adaptation, and outer-loop generalization. We demonstrate that the proposed model outperforms existing methods in terms of NDCG@10, HR@10, and MRR metrics on MovieLens-1M, Amazon Reviews, and Recbole datasets, and achieves real-time processing speeds of less than 300ms on a consumer GPU. Additionally, this method also supports personalization for users without a history, and its adaptation speed of 275ms suggests that it can be applied to real-time risk profiling in financial systems, contributing to improving the stability of payment networks.

Takeaways, Limitations

Takeaways:
Presentation of a meta-learning framework that effectively solves the cold start problem of LLM-based recommender systems
Fast and efficient personalization possible through parameter-efficient prompt tuning
Achieve real-time (less than 300ms) processing speed
Personalized support for users without a history
Applicable to real-time risk profiling of financial systems and improved payment network stability
Limitations:
Further research is needed on the generalization performance of the proposed meta-learning framework.
Need to verify applicability to various datasets and recommendation systems
Consideration of the complexity and computational cost of the meta-learning process is necessary.
👍