Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Enhancing Temporal Sensitivity of Large Language Model for Recommendation with Counterfactual Tuning

Created by
  • Haebom

Author

Yutian Liu, Zhengyi Yang, Jiancan Wu, Xiang Wang

Outline

This paper discusses recent research trends in applying large-scale language models (LLMs) to sequential recommendation. Existing LLM-based methods fail to fully utilize the rich temporal information inherent in a user's past interaction sequences. This is because LLM's self-attention mechanism inherently lacks sequence information and relies on positional embeddings, which are less suitable for user interaction sequences than natural language. To address these limitations, we propose a Counterfactual Enhanced Temporal Framework for LLM-Based Recommendation (CETRec) , which separates and measures the influence of temporal information based on causal inference principles . CETRec effectively enhances LLM's understanding of both absolute order (interaction times with items) and relative order (sequential relationships between items) by utilizing counterfactual adjustments derived from causal analysis. We demonstrate the effectiveness of CETRec through extensive experiments on real-world datasets.

Takeaways, Limitations

Takeaways:
A novel approach to improving the performance of LLM-based sequential recommender systems.
Clarify and leverage the importance of temporal information using causal inference.
Accurately capture user preference changes by considering both absolute and relative time-order information.
Validation of effectiveness through experimental results using actual datasets
Limitations:
Further verification of the generalization performance of the proposed method is needed.
Applicability studies for other types of recommender systems or datasets are needed.
Further analysis of computational cost and complexity is needed.
👍