Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Federated Cross-Training Learners for Robust Generalization under Data Heterogeneity

Created by
  • Haebom

Author

Zhuang Qi, Lei Meng, Ruohan Zhang, Yu Wang, Xin Qi, Xiangxu Meng, Han Yu, Qiang Yang

Outline

This paper presents a novel method, Federated Cross-Training (FedCT), that improves cross-training strategies in federated learning. To address the limitations of existing cross-training, such as mismatched optimization objectives and feature space heterogeneity due to data distribution differences, it leverages knowledge distillation from both local and global perspectives. Specifically, it consists of three modules: a consistency-aware knowledge propagation module, a multi-perspective knowledge-guided representation learning module, and a mixup-based feature augmentation module. These modules preserve local knowledge, maintain consistency between local and global knowledge, and increase feature space diversity, thereby improving performance. Experimental results using four datasets show that FedCT outperforms existing state-of-the-art methods.

Takeaways, Limitations

Takeaways:
A novel cross-training strategy is presented to effectively address the issues caused by data distribution differences in federated learning.
Mitigating knowledge loss and improving performance through local and global knowledge distillation.
Implementing an efficient federated learning process through a combination of various modules.
Excellent performance verification on various datasets
Limitations:
Lack of detailed analysis of the computational complexity and amount of operations of the proposed method.
Further research is needed on generalization performance across diverse federated learning environments and datasets.
Lack of detailed explanations of optimal hyperparameter settings for specific datasets.
👍