Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

SSFO: Self-Supervised Faithfulness Optimization for Retrieval-Augmented Generation

Created by
  • Haebom

Author

Xiaqiang Tang, Yi Wang, Keyu Hu, Rui Xu, Chuang Li, Weigao Sun, Jian Li, Sihong Xie

Outline

To address model hallucination in Retrieval-Augmented Generation (RAG) systems, we propose Self-Supervised Faithfulness Optimization (SSFO). SSFO constructs preference data pairs by contrasting model outputs with and without context, and uses Direct Preference Optimization (DPO) to improve model accuracy without labeling costs or additional inference overhead. SSFO leverages the "likelihood displacement" phenomenon, which transfers probability mass to context-aligned tokens, and proposes an improved DPO loss function based on this. SSFO achieves state-of-the-art performance on multiple datasets, outperforming existing methods and preserving generalization and direction-following capabilities in multilingual environments.

Takeaways, Limitations

Takeaways:
A novel approach to addressing the hallucination phenomenon in the RAG system is presented.
Improve model accuracy without labeling costs through self-supervised learning.
Align models without additional inference burden by leveraging DPO.
A new mechanism is proposed to improve model accuracy by utilizing the "likelihood displacement" phenomenon.
Achieving SOTA performance on diverse datasets and demonstrating generalization ability in multilingual environments.
Limitations:
There is no specific mention of Limitations in the paper.
👍