Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

A Tale of LLMs and Induced Small Proxies: Scalable Agents for Knowledge Mining

Created by
  • Haebom

Author

Sipeng Zhang, Longfei Yun, Zilong Wang, Jingbo Shang, Letian Peng

Outline

Falconer is a collaborative framework that combines LLM's agent-like inference with lightweight proxy models for knowledge mining tasks that extract structured information from large-scale unstructured text based on user instructions. LLM acts as a planner, decomposing user instructions into executable pipelines and generating supervision for training small proxies. Falconer integrates classification and extraction into two atomic operations (get label and get span), allowing a single instruction-following model to replace multiple task-specific components. New benchmarks evaluate the consistency between the proxy model and annotations provided by humans and larger models. Falconer approaches the state-of-the-art LLM in instruction-following accuracy, while reducing inference costs by up to 90% and accelerating large-scale knowledge mining by more than 20x, providing an efficient and scalable foundation for deep research.

Takeaways, Limitations

Takeaways:
Combining the powerful inference capabilities of LLM with the efficiency of lightweight models to ensure scalability for large-scale knowledge mining.
Reduce complexity and increase flexibility by handling classification and extraction tasks with a single model.
Significantly reduce inference costs and accelerate large-scale knowledge mining, increasing practical applicability.
Objectively evaluate model performance with new benchmarks.
Limitations:
No specific Limitations mentioned in the paper.
👍