Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

SPFT-SQL: Enhancing Large Language Model for Text-to-SQL Parsing by Self-Play Fine-Tuning

Created by
  • Haebom

Author

Yuhao Zhang, Shaoming Duan, Jinhang Su, Chuanyi Liu, Peiyi Han

Outline

This paper discusses the development of self-play fine-tuning (SPIN), which transforms weak LLMs into robust models, and highlights its challenges in text-to-SQL tasks. SPIN fails to generate new information, and the large number of correct SQL queries generated by the adversary model degrades the master model's ability to generate accurate SQL queries. To address these issues, this paper proposes SPFT-SQL, a novel self-play fine-tuning method specifically designed for text-to-SQL tasks. SPFT-SQL introduces a validation-based iterative fine-tuning approach that iteratively synthesizes high-quality fine-tuning data based on database schema and validation feedback before self-play. This approach improves model performance and builds a model base with diverse capabilities. In the self-play fine-tuning phase, an error-based loss method is proposed that encourages the adversary model's incorrect output, enabling the master model to distinguish between correct and incorrect SQL generated by the adversary model. Through extensive experiments and in-depth analysis on six open-source LLMs and five widely used benchmarks, we demonstrate that the proposed method outperforms existing state-of-the-art (SOTA) methods.

Takeaways, Limitations

Takeaways:
We present a novel self-playing fine-tuning method (SPFT-SQL) that overcomes the limitations of SPIN in text-to-SQL tasks.
Generating high-quality fine-tuning data through a validation-based iterative fine-tuning approach.
Improve the main model's ability to generate accurate SQL using error-based loss methods.
Achieving SOTA performance through various LLMs and benchmarks.
Limitations:
The specific Limitations is not directly mentioned within the paper.
👍