Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Retro*: Optimizing LLMs for Reasoning-Intensive Document Retrieval

Created by
  • Haebom

Author

Junwei Lan, Jianlyu Chen, Zheng Liu, Chaofan Li, Siqi Bao, Defu Lian

Outline

With the growing popularity of LLM agents and RAGs, retrieving documents essential to solving a task, even when the connection to the task is indirect or implicit, is becoming increasingly important. Addressing this challenge requires fine-grained inference to accurately assess the relevance between the task and each candidate document. However, this capability poses significant challenges to existing IR techniques. In this paper, we propose Retro, a novel approach for inference-based document retrieval. Retro introduces a rubric-based relevance scoring mechanism that allows the model to infer the relationships between tasks and documents based on explicitly defined criteria, generating fine-grained and interpretable relevance scores. Furthermore, Retro combines multiple inference trajectories through score aggregation, supporting test time extension and generating more reliable relevance estimates. To optimize Retro's inference capabilities, we introduce a novel reinforcement learning algorithm tailored to the relevance scoring mechanism that utilizes two compound rewards to maximize the utilization of the trajectories of each training sample. Experimental results show that Retro* outperforms existing document retrieval methods on the BRIGHT benchmark, achieving state-of-the-art performance.

Takeaways, Limitations

Takeaways:
Retro* overcomes the limitations of existing IR technologies by enabling fine-grained inference through a rubric-based relevance scoring mechanism.
Supports extended test time through score consolidation, providing more reliable relevance estimates.
We present a novel reinforcement learning algorithm to optimize the inference ability of Retro*.
It has proven its effectiveness by achieving cutting-edge performance in the BRIGHT benchmark.
Limitations:
The paper alone does not provide detailed information on the specific implementation of Retro* or the rubric design method.
Generalizability is difficult to assess because detailed information about the experimental environment or dataset is not provided.
There is a lack of detailed description of the comparative analysis and performance differences with other related technologies.
👍