Fan
Federated learning with hierarchical clustering of local updates to improve training on non-IID data
Fan
Authors
Abstract
Federated learning (FL) is a well established method for performing machine learning tasks over massively distributed data. However in settings where data is distributed in a non-iid (not independent and identically distributed) fashion -- as is typical in real world situations -- the joint model produced by FL suffers in terms of test set accuracy and/or communication costs compared to training on iid data. We show that learning a single joint model is often not optimal in the presence of certain types of non-iid data. In this work we present a modification to FL by introducing a hierarchical clustering step (FL+HC) to separate clusters of clients by the similarity of their local updates to the global joint model. Once separated, the clusters are trained independently and in parallel on specialised models. We present a robust empirical analysis of the hyperparameters for FL+HC for several iid and non-iid settings. We show how FL+HC allows model training to converge in fewer communication rounds (significantly so under some non-iid settings) compared to FL without clustering. Additionally, FL+HC allows for a greater percentage of clients to reach a target accuracy compared to standard FL. Finally we make suggestions for good default hyperparameters to promote superior performing specialised models without modifying the the underlying federated learning communication protocol.
Citation
Fan. (2020). Federated learning with hierarchical clustering of local updates to improve training on non-IID data. International Joint Conference on Neural Networks, https://doi.org/10.1109/IJCNN48605.2020.9207469
Acceptance Date | Jul 19, 2020 |
---|---|
Publication Date | Sep 28, 2020 |
Journal | 2020 International Joint Conference on Neural Networks (IJCNN) |
Print ISSN | 2161-4393 |
Publisher | Institute of Electrical and Electronics Engineers (IEEE) |
Series Title | 2020 International Joint Conference on Neural Networks (IJCNN) |
DOI | https://doi.org/10.1109/IJCNN48605.2020.9207469 |
Publisher URL | https://ieeexplore.ieee.org/document/9207469 |
Files
IJCNN-FL+HC-paper-final.pdf
(441 Kb)
PDF
Publisher Licence URL
https://creativecommons.org/licenses/by-nc/4.0/
You might also like
The role of ‘living laboratories’ in accelerating the energy system decarbonization
(2022)
Journal Article
A Review of Privacy-preserving Federated Learning for the Internet-of-Things
(2021)
Book Chapter
Downloadable Citations
About Keele Repository
Administrator e-mail: research.openaccess@keele.ac.uk
This application uses the following open-source libraries:
SheetJS Community Edition
Apache License Version 2.0 (http://www.apache.org/licenses/)
PDF.js
Apache License Version 2.0 (http://www.apache.org/licenses/)
Font Awesome
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2024
Advanced Search