Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

GRADA: Graph-based Reranking against Adversarial Documents Attack

Created by
  • Haebom

Author

Jingjie Zheng, Aryo Pradipta Gema, Giwon Hong, Xuanli He, Pasquale Minervini, Youcheng Sun, Qiongkai Xu

Outline

The Retrieval Augmented Generation (RAG) framework improves the accuracy of LLMs by retrieving external documents, but it is vulnerable to adversarial attacks that manipulate the retrieval process. In this paper, we propose GRADA, a graph-based reranking framework for adversarial document attacks. GRADA aims to maintain retrieval quality while mitigating the impact of adversarial documents. We conducted experiments on five LLMs and three datasets—GPT-3.5-Turbo, GPT-4o, Llama3.1-8b, Llama3.1-70b, and Qwen2.5-7b—and achieved up to an 80% reduction in attack success rate on the Natural Questions dataset.

Takeaways, Limitations

Takeaways: GRADA demonstrates that it can effectively mitigate the vulnerability of RAG systems to adversarial attacks. Its performance is validated on various LLMs and datasets. On the Natural Questions dataset, it significantly reduces the attack success rate while minimizing accuracy degradation.
Limitations: Currently, this has only been evaluated on a specific dataset and LLM. Therefore, generalizability to other datasets and LLMs requires further research. A more detailed analysis of the factors contributing to GRADA's improved performance may be necessary. GRADA's robustness against various types of adversarial attacks should also be further evaluated.
👍