Federated Learning is a machine learning paradigm where a global model is trained in-situ across a large number of distributed edge devices. While this technique avoids the cost of transferring data to a central location and achieves a strong degree of privacy, it presents additional challenges due to the heterogeneous hardware resources available for training. Furthermore, data is not independent and identically distributed (IID) across all edge devices, resulting in statistical heterogeneity across devices. Due to these constraints, client selection strategies play an important role for timely convergence during model training. Existing strategies ensure that each individual device is included, at least periodically, in the training process. In this work, we propose HACCS, a Heterogeneity-Aware Clustered Client Selection system that identifies and exploits the statistical heterogeneity by representing all distinguishable data distributions instead of individual devices in the training process. HACCS is robust to individual device dropout, provided other devices in the system have similar data distributions. We propose privacy-preserving methods for estimating these client distributions and clustering them. We also propose strategies for leveraging these clusters to make scheduling decisions in a federated learning system. Our evaluation on real-world datasets suggests that our framework can provide 18% -38% reduction in time to convergence compared to the state of the art without any compromise in accuracy.
|Original language||English (US)|
|Title of host publication||Proceedings - 2022 IEEE 36th International Parallel and Distributed Processing Symposium, IPDPS 2022|
|Publisher||Institute of Electrical and Electronics Engineers Inc.|
|Number of pages||11|
|State||Published - 2022|
|Event||36th IEEE International Parallel and Distributed Processing Symposium, IPDPS 2022 - Virtual, Online, France|
Duration: May 30 2022 → Jun 3 2022
|Name||Proceedings - 2022 IEEE 36th International Parallel and Distributed Processing Symposium, IPDPS 2022|
|Conference||36th IEEE International Parallel and Distributed Processing Symposium, IPDPS 2022|
|Period||5/30/22 → 6/3/22|
Bibliographical noteFunding Information:
ACKNOWLEDGEMENT This research was supported in part by the NSF under grant CNS-1717834.
© 2022 IEEE.
- Federated Learning
- Non-IID data