I. Evolution of Federated Learning
Federated Learning (FL) was introduced around 2016 as a privacy enhancing technique that directly applies the principle of data minimization by focused collection and immediate aggregation [36], which "enables mobile phones to collaboratively learn a shared prediction model while keeping all the training data on device, decoupling the ability to do machine learning from the need to store the data in the cloud." [27], [48] FL quickly became a widely-acknowledged paradigm of distributed learning from decentralized data, and has been adopted in various applications beyond the original on-device training scenarios: for example, the FL paradigm has been applied to collaborative learning across multiple institutions (silos) with richer computation resources than mobile devices, or to learning over Internet-of-Things devices with more limited resources. In 2019, out of the discussion in the Workshop on Federated Learning and Analytics at Google, Kairouz et al. [42] proposed a broader definition of FL:
Federated learning is a machine learning setting where multiple entities (clients) collaborate in solving a machine learning problem, under the coordination of a central server or service provider. Each client’s raw data is stored locally and not exchanged or transferred; instead, focused updates intended for immediate aggregation are used to achieve the learning objective.