[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Federated Continuous Instruction Tuning

Created by
  • Haebom

Author

Haiyang Guo, Fanhu Zeng, Fei Zhu, Wenzhuo Liu, Da-Han Wang, Jian Xu, Xu-Yao Zhang, Cheng-Lin Liu

Outline

In this paper, we present a federated learning (FL)-based continuous instruction fine-tuning (FCIT) benchmark to address the challenges of collecting massive data and computational costs required for instruction fine-tuning of large-scale multimodal models (LMMs). Unlike existing FL methods that assume a fixed number of tasks, FCIT models real-world situations where clients continuously acquire new knowledge and struggle to maintain existing tasks. To this end, we construct a benchmark that includes two realistic scenarios, four settings, and 12 instruction fine-tuning datasets, and propose a method to address various data heterogeneities and forgetting issues through dynamic knowledge construction and subspace selective activation. Experimental results show that the proposed method significantly improves model performance. The code and datasets are publicly available.

Takeaways, Limitations

Takeaways:
We present a new benchmark and method to reduce the cost of directive fine-tuning of large-scale multimodal models by leveraging federated learning.
Provides realistic benchmarks that reflect continuous learning situations in real environments.
Effectively solve the forgetting problem of traditional federated learning, Limitations, through dynamic knowledge construction and subspace selective activation techniques.
Contribute to the advancement of future research through open code and datasets.
Limitations:
The performance evaluation of the proposed method may be limited to specific datasets and settings.
It may not fully reflect the complexity of the real environment.
Experiments with more diverse types of LMM and FL algorithms are needed.
👍