I. Introduction
Federated learning (FL) has emerged as an attractive framework in edge learning to train models when the data is distributed among edge devices and must remain local due to resource constraints and/or privacy concerns. The edge-device networks in FL could comprise millions of clients [1] whose feedback might include model updates that are on the order of 100 Mb. For example, neural network for image recognition tasks VGG-16 [2] has 160 M parameters and weights resulting in updates of size 526Mb when using 32 b encoding.