Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Self-Evolving LLMs via Continual Instruction Tuning

Created by
  • Haebom

Author

Jiazheng Kang, Le Huang, Cheng Hou, Zhe Zhao, Zhenxiang Yan, Chuan Shi, Ting Bai

Outline

MoE-CL is a parameter-efficient adversarial mixed-expert framework for continuous training of large-scale language models (LLMs) that must self-evolve to adapt to dynamic data distributions in large-scale industrial environments. MoE-CL utilizes a dual-expert design, preserving task-specific knowledge through LoRA experts for each task to mitigate forgetting, while shared LoRA experts enable inter-task transfer. To prevent task-related noise transmission through shared paths, a task-aware discriminator is integrated within the GAN.

Takeaways, Limitations

MoE-CL utilizes dedicated LoRA experts with independent parameters for task-specific knowledge maintenance and shared LoRA experts for inter-task transfer.
Through adversarial learning, a shared expert learns generalized representations that mimic the discriminator, while a dedicated expert maintains task-specific details, balancing knowledge preservation and cross-task generalization.
The effectiveness of MoE-CL has been proven in the MTL5 public benchmark and the industry Tencent3 benchmark.
In an A/B test for content compliance review on the Tencent Video platform, MoE-CL reduced manual review costs by 15.3%.
This paper does not make any specific reference to MoE-CL's Limitations.
👍