Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Distilling a Small Utility-Based Passage Selector to Enhance Retrieval-Augmented Generation

Created by
  • Haebom

Author

Hengran Zhang, Keping Bi, Jiafeng Guo, Jiaming Zhang, Shuaiqiang Wang, Dawei Yin, Xueqi Cheng

Efficient Model Distillation for Utility-Based Selection in Retrieval-Augmented Generation (RAG)

Outline

This paper proposes a method to distill the utility judgment capabilities of a large-scale language model (LLM) into a smaller model to enhance the efficiency of utility-based retrieval in Augmented Search Generation (RAG). Specifically, we apply a utility-based selection approach to generate answers to complex questions, enabling dynamic paragraph selection rather than a fixed threshold. By training student models (RankQwen1.7B, UtilityQwen1.7B) that learn fake answer generation and utility judgment from a teacher LLM (Qwen3-32B), we improve answer quality while reducing computational costs. We plan to release relevance ranking and utility-based selection annotations for the MS MARCO dataset to support research in this area.

Takeaways, Limitations

Takeaways:
Utility-based selection is effective in improving the performance of generating answers to complex questions.
The utility judgment capabilities of LLM can be distilled into smaller models to reduce computational costs.
Dynamic paragraph selection allows you to flexibly select paragraphs to suit specific questions.
Open annotation of the MS MARCO dataset facilitates further research.
Limitations:
Currently Limitations is not specified in the paper.
👍