Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

ToolACE-R: Model-aware Iterative Training and Adaptive Refinement for Tool Learning

Created by
  • Haebom

Author

Xingshan Zeng, Weiwen Liu, Xu Huang, Zezhong Wang, Lingzhi Wang, Liangyou Li, Yasheng Wang, Lifeng Shang, Xin Jiang, Ruiming Tang, Qun Liu

Outline

This paper discusses tool learning, which has emerged as a promising approach to extend the capabilities of large-scale language models (LLMs). Existing tool learning approaches have primarily focused on data synthesis to fine-tune LLMs to effectively invoke tools, but have largely neglected methods to fully exploit the model's potential. This paper proposes ToolACE-R, a novel framework that incorporates both model-aware iterative learning and adaptive improvement. ToolACE-R features a model-aware iterative learning procedure that incrementally adjusts training samples based on the model's evolving capabilities to maximize its potential. Furthermore, it incorporates a self-improving training corpus, highlighting the LLM's ability to iteratively optimize tool invocation without external feedback. Furthermore, we introduce an adaptive self-improvement mechanism for efficient testing time extension, enabling the trained model to autonomously decide when to stop the iterative self-improvement process. Extensive experiments on multiple benchmark datasets demonstrate that ToolACE-R achieves competitive performance compared to advanced API-based models. Adaptive self-improvement can efficiently further enhance tool invocation performance. These results highlight the effectiveness and generalizability of ToolACE-R and suggest promising directions for more efficient and scalable tool learning.

Takeaways, Limitations

Takeaways:
We demonstrate that the potential of LLM can be maximized through a model-aware iterative learning procedure that adjusts training samples according to the model's evolving capabilities.
We demonstrate the effectiveness of a self-improving training corpus that optimizes LLM's tool invocation performance without external feedback.
We propose that test time extension can be efficiently performed through an adaptive self-improvement mechanism.
We experimentally demonstrate that ToolACE-R achieves competitive performance compared to advanced API-based models.
Limitations:
This paper lacks a detailed description of the specific algorithms and implementation details of ToolACE-R.
Further research is needed on generalization performance across different types of tools and tasks.
Further research is needed to determine the optimal parameters of the adaptive self-improvement mechanism.
Disclosure of code and data is required to ensure reproducibility of experimental results.
👍