Loading [MathJax]/extensions/MathZoom.js
Partial Training Mechanism to Handle the Impact of Stragglers in Federated Learning with Heterogeneous Clients | IEEE Conference Publication | IEEE Xplore

Partial Training Mechanism to Handle the Impact of Stragglers in Federated Learning with Heterogeneous Clients


Abstract:

Federated Learning (FL) allows distributed devices, known as clients, to train Machine Learning (ML) models collaboratively without sharing sensitive data. A characterist...Show More

Abstract:

Federated Learning (FL) allows distributed devices, known as clients, to train Machine Learning (ML) models collaboratively without sharing sensitive data. A characteristic of FL for mobile and IoT environments is system heterogeneity among clients, which can vary from low-end devices with constrained communication and computing resources to powerful devices with high-speed network access and dedicated GPUs. As the server must wait for all the clients to communicate their updates, slow clients (a.k.a. stragglers) will significantly increase the training time. To tackle this problem, we propose FedPulse, a Partial Training (PT) based mechanism to mitigate the effect of stragglers in FL. The idea is to reduce the training time by dynamically allocating smaller submodels to resource-constrained clients. Experimental results on famous classification datasets show that the proposed solution outperforms other submodel allocation mechanisms and reduces the training time by up to 58% with an accuracy loss of less than 1% when compared to FedAvg.
Date of Conference: 26-29 June 2024
Date Added to IEEE Xplore: 31 October 2024
ISBN Information:

ISSN Information:

Conference Location: Paris, France
References is not available for this document.

I. Introduction

The popularization of mobile and IoT devices allows the collection of a large amount of data at the network’s edge, which enables the training of large Machine Learning (ML) models for a wide range of applications, such as next-word prediction and object detection. Traditionally, the edge devices send their local data to a powerful central server, which trains the model in a centralized manner. However, this mechanism raises several privacy concerns as the local data might be sensitive, and regulations such as the General Data Protection Regulation (GDPR) might restrict data collection from those devices [1], making centralized training unfeasible.

Select All
1.
C. Tankard, "What the gdpr means for businesses", Network Security, vol. 2016, no. 6, pp. 5-8, 2016.
2.
B. McMahan, E. Moore, D. Ramage, S. Hampson and B. A. y Arcas, "Communication-efficient learning of deep networks from decentralized data", Artificial intelligence and statistics, pp. 1273-1282, 2017.
3.
A. Imteaj, U. Thakker, S. Wang, J. Li and M. H. Amini, "A survey on federated learning for resource-constrained iot devices", IEEE Internet of Things Journal, vol. 9, no. 1, pp. 1-24, 2021.
4.
T. Li, A. K. Sahu, A. Talwalkar and V. Smith, "Federated learning: Challenges methods and future directions", IEEE signal processing magazine, vol. 37, no. 3, pp. 50-60, 2020.
5.
C. Xie, S. Koyejo and I. Gupta, Asynchronous federated optimization, 2019.
6.
W. Wu, L. He, W. Lin, R. Mao, C. Maple and S. Jarvis, "Safa: A semiasynchronous protocol for fast federated learning with low overhead", IEEE Transactions on Computers, vol. 70, no. 5, pp. 655-668, 2020.
7.
B. Fan, S. Jiang, X. Su and P. Hui, Model-heterogeneous federated learning for internet of things: Enabling technologies and future directions, 2023.
8.
S. Caldas, J. Konečny, H. B. McMahan and A. Talwalkar, Expanding the reach of federated learning by reducing client resource requirements, 2018.
9.
N. Bouacida, J. Hou, H. Zang and X. Liu, "Adaptive federated dropout: Improving communication efficiency and generalization for federated learning", CoRR, vol. abs/2011.04050, 2020.
10.
E. Diao, J. Ding and V. Tarokh, Heterofl: Computation and communication efficient federated learning for heterogeneous clients, 2020.
11.
K. Pfeiffer, M. Rapp, R. Khalili and J. Henkel, "Federated learning for computationally-constrained heterogeneous devices: A survey", ACM Computing Surveys, 2023.
12.
T. Nishio and R. Yonetani, "Client selection for federated learning with heterogeneous resources in mobile edge", IEEE international conference on communications (ICC), pp. 1-7, 2019.
13.
Z. Chai, A. Ali, S. Zawad, S. Truex, A. Anwar, N. Baracaldo, et al., "Tifl: A tier-based federated learning system", 29th international symposium on high-performance parallel and distributed computing, pp. 125-136, 2020.
14.
A. Reisizadeh, I. Tziotis, H. Hassani, A. Mokhtari and R. Pedarsani, "Straggler-resilient federated learning: Leveraging the interplay between statistical accuracy and system heterogeneity", IEEE Journal on Selected Areas in Information Theory, vol. 3, no. 2, pp. 197-205, 2022.
15.
P. Han, S. Wang and K. K. Leung, "Adaptive gradient sparsification for efficient federated learning: An online learning approach", IEEE 40th international conference on distributed computing systems (ICDCS), pp. 300-310, 2020.
16.
M. M. Amiri, D. Gunduz, S. R. Kulkarni and H. V. Poor, Federated learning with quantized global model updates, 2020.
17.
S. Alam, L. Liu, M. Yan and M. Zhang, "Fedrolex: Modelheterogeneous federated learning with rolling sub-model extraction", Advances in Neural Information Processing Systems, vol. 35, pp. 29677-29690, 2022.
18.
D. J. Beutel, T. Topal, A. Mathur, X. Qiu, J. Fernandez-Marques, Y. Gao, L. Sani, K. H. Li, T. Parcollet, P. P. B. de Gusmão et al., Flower: A friendly federated learning research framework, 2020.

Contact IEEE to Subscribe

References

References is not available for this document.