Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Learning Like Humans: Advancing LLM Reasoning Capabilities via Adaptive Difficulty Curriculum Learning and Expert-Guided Self-Reformulation

Created by
  • Haebom

Author

Enci Zhang, Xingang Yan, Wei Lin, Tianxiang Zhang, Qianchun Lu

Outline

This paper proposes two novel strategies for improving the complex problem-solving ability of large-scale language models (LLMs): Adaptive Difficulty Curriculum Learning (ADCL) and Expert-Guided Self-Reconfiguration (EGSR). ADCL is a curriculum learning strategy that addresses the phenomenon of difficulty variation during training by re-evaluating the difficulty as the model's capabilities change. EGSR is a reinforcement learning strategy that encourages expert solutions to be restructured within the model's conceptual framework, bridging the gap between imitation learning and pure exploration, thereby promoting deeper understanding and knowledge absorption. Experimental results based on the Qwen2.5-7B model demonstrate that the two proposed strategies synergistically improve performance, achieving a 10% performance improvement on the AIME24 benchmark and a 16.6% performance improvement on the AIME25 benchmark.

Takeaways, Limitations

Takeaways:
We demonstrate that complex problem-solving skills in LLM can be improved through novel learning strategies (ADCL, EGSR) that mimic human learning strategies.
The combination of ADCL and EGSR brings significant performance improvements over existing methods.
The proposed strategy has high applicability not only to mathematical reasoning but also to other complex problem-solving fields.
Limitations:
The effectiveness of the proposed strategy is limited to specific models (Qwen2.5-7B) and benchmarks (AIME24, AIME25), requiring further research on generalizability.
Further research is needed on parameter tuning and optimization of ADCL and EGSR.
Further validation is needed on the performance and generalization ability when applied to other types of problems or larger-scale models.
👍