Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Self-Evolving LLMs via Continual Instruction Tuning

Created by
  • Haebom

Author

Jiazheng Kang, Le Huang, Cheng Hou, Zhe Zhao, Zhenxiang Yan, Ting Bai

Outline

MoE-CL is a parameter-efficient adversarial mixed-expert framework for continuous learning of LoRa models (LLMs) that self-evolves to adapt to dynamic data distributions in large-scale industrial environments. MoE-CL utilizes a dual-expert design, consisting of dedicated LoRA experts that preserve task-specific knowledge and shared LoRA experts that enable cross-task transfer. Through adversarial learning, the shared experts acquire generalized representations, while the dedicated experts retain task-specific details, achieving a balance between knowledge retention and cross-task generalization. Extensive experiments on the MTL5 benchmark and the industry Tencent3 benchmark demonstrate the effectiveness of MoE-CL, and in A/B testing for content compliance review on the Tencent Video platform, it reduced manual review costs by 15.3%.

Takeaways, Limitations

Takeaways:
Presenting practical solutions for the continued adaptation of LLMs, where continuous learning and reliable transfer are crucial in large-scale industrial environments.
Mitigating the catastrophic forgetting problem through a dual-expert design that balances task-specific knowledge retention and cross-task generalization.
Enhancing cross-task transfer effectiveness by ensuring task-related information transfer from shared experts through adversarial learning.
Performance validation in real-world industrial environments (reducing manual review costs through A/B testing on the Tencent Video platform).
Limitations:
No specific Limitations mentioned in the paper.
👍