[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

SPAR: Scholar Paper Retrieval with LLM-based Agents for Enhanced Academic Search

Created by
  • Haebom

Author

Xiaofeng Shi, Yuduo Li, Qian Kou, Longbin Yu, Jinxin Xie, Hua Zhou

Outline

In this paper, we propose SPAR, a multi-agent framework that integrates RefChain-based query decomposition and query evolution to overcome the limitations of academic literature retrieval systems utilizing large-scale language models (LLMs). It improves the rigid pipeline and limited inference capability of existing systems, enabling more flexible and effective retrieval. In addition, we construct SPARBench, a demanding benchmark that includes expert-annotated relevance labels, to facilitate systematic evaluation. Experimental results show that SPAR outperforms existing state-of-the-art benchmark models, achieving up to +56% F1 improvement in AutoScholar and +23% F1 improvement in SPARBench. SPAR and SPARBench provide a foundation for the advancement of academic retrieval research with scalability, interpretability, and high performance.

Takeaways, Limitations

Takeaways:
Presenting the possibility of flexible and effective academic literature retrieval through RefChain-based query decomposition and query evolution
New benchmark SPARBench with expert annotations enables systematic performance evaluation
Significant performance improvement over previous top-performing models (AutoScholar +56% F1, SPARBench +23% F1)
Contributing to the advancement of scholarly search research with scalability, interpretability, and high performance
Limitations:
Lack of detailed description of the specific configuration and scale of SPARBench in the paper
Further validation of the generalizability of the experimental results is needed.
Further performance evaluations are needed for various types of academic literature and queries.
Further research is needed on performance and efficiency in real-world environments.
👍