Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Middo: Model-Informed Dynamic Data Optimization for Enhanced LLM Fine-Tuning via Closed-Loop Learning

Created by
  • Haebom

Author

Zinan Tang, Xin Gao, Qizhi Pei, Zhuoshi Pan, Mengzhang Cai, Jiang Wu, Conghui He, Lijun Wu

Outline

This paper proposes Middo, a novel framework for supervised learning fine-tuning (SFT) of large-scale language models (LLMs). To overcome the limitations of existing static dataset-based approaches, Middo builds a dynamic data optimization system that evolves based on model performance. Using loss patterns, embedding cluster dynamics, and self-alignment scores, Middo identifies inefficient samples and transforms them into educationally valuable ones. This process continuously improves the dataset as the model's capabilities improve, achieving an average accuracy improvement of 7.15% across multiple benchmark experiments. This presents a new paradigm for sustainable LLM learning through the dynamic co-evolution of data and models.

Takeaways, Limitations

Takeaways:
We significantly improve the efficiency of LLM learning by presenting a dynamic data optimization system that continuously evolves as model performance improves.
We overcome the limitations of existing static dataset-based SFT and propose a more effective LLM learning method.
We present a novel approach to assessing and improving data quality by leveraging various model signals.
We verified the effectiveness of Middo through experimental results showing an average accuracy improvement of 7.15%.
We present a sustainable LLM learning paradigm through the mutual evolution of data and models.
Limitations:
Currently, the dataset, model, and code are not publicly available, making it difficult to verify reproducibility.
Generalization performance across various LLMs and tasks has not been sufficiently validated.
There is a lack of analysis of the complexity and computational cost of the Middo framework.
👍