Loading [MathJax]/extensions/MathZoom.js
Optimizing Parameter Mixing Under Constrained Communications in Parallel Federated Learning | IEEE Journals & Magazine | IEEE Xplore

Optimizing Parameter Mixing Under Constrained Communications in Parallel Federated Learning


Abstract:

In vanilla Federated Learning (FL) systems, a centralized parameter server (PS) is responsible for collecting, aggregating and distributing model parameters with decentra...Show More

Abstract:

In vanilla Federated Learning (FL) systems, a centralized parameter server (PS) is responsible for collecting, aggregating and distributing model parameters with decentralized clients. However, the communication link of a single PS can be easily overloaded by concurrent communications with a massive number of clients. To overcome this drawback, multiple PSes can be deployed to form a parallel FL (PFL) system, in which each PS only communicates with a subset of clients and its neighbor PSes. On one hand, each PS conducts iterations with clients in its subset. On the other hand, PSes communicate with each other periodically to mix their parameters so that they can finally reach a consensus. In this paper, we propose a novel parallel federated learning algorithm called Fed-PMA, which optimizes such parallel FL under constrained communications by conducting parallel parameter mixing and averaging with theoretic guarantees. We formally analyze the convergence rate of Fed-PMA with convex loss, and further derive the optimal number of times each PS should mix with its neighbor PSes so as to maximize the final model accuracy within a fixed span of training time. Theoretical study manifests that PSes should mix their parameters more frequently if the connection between PSes is sparse or the time cost of mixing is low. Inspired by our analysis, we propose the Fed-APMA algorithm that can adaptively determine the near-optimal number of mixing times with non-convex loss under dynamic communication conditions. Extensive experiments with realistic datasets are carried out to demonstrate that both Fed-PMA and its adaptive version Fed-APMA significantly outperform the state-of-the-art baselines.
Published in: IEEE/ACM Transactions on Networking ( Volume: 31, Issue: 6, December 2023)
Page(s): 2640 - 2652
Date of Publication: 27 March 2023

ISSN Information:

Funding Agency:


I. Introduction

Recently, federated learning (FL) has received tremendous research attention due to its capability in protecting data privacy for machine learning. In vanilla FL, a parameter server (PS) is in charge of collection, aggregation and distribution of model parameters with scattered clients. Clients conduct local iterations to update model parameters for multiple rounds, but never expose their original samples to the PS [1], [2], [3], [4], [5].

Contact IEEE to Subscribe

References

References is not available for this document.