Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Feature Distillation is the Better Choice for Model-Heterogeneous Federated Learning

Created by
  • Haebom

Author

Yichen Li, Xiuying Wang, Wenchao Xu, Haozhao Wang, Yining Qi, Jiahua Dong, Ruixuan Li

Outline

This paper proposes Feature Distillation for model-heterogeneous Federated Learning (FedFD), a novel method for improving knowledge aggregation in heterogeneous model federated learning (Hetero-FL). Existing Hetero-FL utilizes ensemble distillation techniques to improve the performance of a global model using logit distillation, but suffers from the limitation of not being able to compensate for knowledge biases arising from heterogeneous models. To address this issue, FedFD proposes a feature-based ensemble federated knowledge distillation paradigm that improves knowledge integration across heterogeneous models by aligning feature information through orthogonal projections. The server's global model maintains projection layers for each client model architecture to individually align features, and orthogonal techniques are used to reparameterize the projection layers to mitigate knowledge biases and maximize distilled knowledge. Experimental results demonstrate that FedFD outperforms existing state-of-the-art methods.

Takeaways, Limitations

Takeaways:
A Novel Approach to Addressing Knowledge Bias in Heterogeneous Model Federated Learning
Overcoming the limitations of conventional log-it distillation through feature-based distillation
A Proposal of a Stable and Efficient Knowledge Integration Method Using Orthogonal Projection
Demonstrated superior performance over existing state-of-the-art methods
Limitations:
Further analysis of the computational cost and complexity of the proposed method is needed.
The need to evaluate generalization performance across diverse heterogeneous models and datasets.
Further research is needed on potential problems and solutions that may arise when applying this to real-world environments.
👍