I. Introduction
Artificial intelligence-based services pose data privacy, communication, and computational challenges that warrant prevailing concerns. As a distributed machine learning architecture, the proposal of federated learning (FL) [1]–[3] provides a new idea to address the above challenges. However, studies, such as those of the optimization of FL in real networks [4], are at an exploratory stage. At present, the federated averaging (FedAvg) algorithm proposed in the paper [5] assumes that all client data are random samples of real data distribution. Although the FedAvg algorithm obtains excellent training results in this case, the heterogeneous computing resources and different communication capabilities of clients may invalidate this assumption. Therefore, improving training efficiency for the successful deployment of FL in edge and mobile devices under resource-constrained and client heterogeneous environments is crucial [6], [7].