Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Retrieval-Augmented Generation with Conflicting Evidence

Created by
  • Haebom

Author

Han Wang, Archiki Prasad, Elias Stengel-Eskin, Mohit Bansal

Outline

This paper presents a novel approach that simultaneously addresses ambiguous user queries, conflicting information, and inaccurate information that arise when leveraging Retrieval Augmented Generation (RAG) to enhance the realism of Large-Scale Language Model (LLM) agents. Unlike previous studies that tackled each problem individually, this paper proposes a new dataset, RAMDocs, which mimics realistic scenarios containing ambiguity, misinformation, and noise. We then present MADAM-RAG, a multi-agent approach that resolves ambiguity and removes misinformation and noise through multi-round discussions among LLM agents. Experimental results demonstrate that MADAM-RAG significantly outperforms existing RAG baseline models on the AmbigDocs and FaithEval datasets, but demonstrates that there is still room for improvement, especially when the imbalance between evidence and misinformation is severe.

Takeaways, Limitations

Takeaways:
We present RAMDocs, a new dataset for evaluating realistic RAG systems that simultaneously considers ambiguity, misinformation, and noise.
We propose the MADAM-RAG technique to handle conflicting information and increase realism through multi-agent discussion.
Experimentally demonstrating performance improvements over existing RAG baseline models on AmbigDocs and FaithEval datasets.
Limitations:
The performance gains of MADAM-RAG are limited when the RAMDocs dataset has a high imbalance of evidence and misinformation.
There is still room for improvement in the performance of MADAM-RAG.
👍