Loading [MathJax]/extensions/MathMenu.js
Edge-Assisted Democratized Learning Toward Federated Analytics | IEEE Journals & Magazine | IEEE Xplore

Edge-Assisted Democratized Learning Toward Federated Analytics


Abstract:

A recent take toward federated analytics (FA), which allows analytical insights of distributed data sets, reuses the federated learning (FL) infrastructure to evaluate th...Show More

Abstract:

A recent take toward federated analytics (FA), which allows analytical insights of distributed data sets, reuses the federated learning (FL) infrastructure to evaluate the summary of model performances across the training devices. However, the current realization of FL adopts single server-multiple client architecture with limited scope for FA, which often results in learning models with poor generalization, i.e., an ability to handle new/unseen data, for real-world applications. Moreover, a hierarchical FL structure with distributed computing platforms demonstrates incoherent model performances at different aggregation levels. Therefore, we need to design a robust learning mechanism than the FL that 1) unleashes a viable infrastructure for FA and 2) trains learning models with better generalization capability. In this work, we adopt the novel democratized learning (Dem-AI) principles and designs to meet these objectives. First, we show the hierarchical learning structure of the proposed edge-assisted Dem-AI mechanism, namely Edge-DemLearn, as a practical framework to empower generalization capability in support of FA. Second, we validate Edge-DemLearn as a flexible model training mechanism to build a distributed control and aggregation methodology in regions by leveraging the distributed computing infrastructure. The distributed edge computing servers construct regional models, minimize the communication loads, and ensure distributed data analytic application’s scalability. To that end, we adhere to a near-optimal two-sided many-to-one matching approach to handle the combinatorial constraints in Edge-DemLearn and solve it for fast knowledge acquisition with optimization of resource allocation and associations between multiple servers and devices. Extensive simulation results on real data sets demonstrate the effectiveness of the proposed methods.
Published in: IEEE Internet of Things Journal ( Volume: 9, Issue: 1, 01 January 2022)
Page(s): 572 - 588
Date of Publication: 02 June 2021

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Recently, the Google AI team published an article on leveraging computing mechanism of the distributed learning model training infrastructure to facilitate data analytics, namely federated analytics (FA) [1]. FA allows data scientists to derive analytical insights of distributed data sets without the need of moving data to a central computing entity. This concept gathered keen attention for a new approach to data science, and interestingly, at the time when the centralized repositories are termed “vulnerable” toward privacy for data collection, and the era of distributed computing and storage is prominent. Besides, it means that apart from considering the distributed model training processes for improving model accuracy, we can exploit such collaboration architecture to evaluate the quality of the trained model at the user-level perspectives, i.e., the model performance at the user’s end. Hence, without the learning part, we can reuse the computing scheme of the learning architecture to perform statistical analysis on local data that may lead to building better products. To elaborate this idea further, consider an example of a prediction model where the developer would be interested in finding popular content to store in a shared regional database without breaking into user’s historical content usage data. An intuitive answer to this question would be to find the frequently requested content at first, which is best done with FA. This is similar to the Now Playing feature on Google’s Pixel phones for managing regional song database [1] to show users songs playing around them. In addition, leveraging distributed multiaccess edge computing (MEC) servers further allows fast knowledge acquisition to serve user’s requests, limit privacy leakages, and improve the on-the-fly learning process. Furthermore, this also brings FA closer to where the data is collected, i.e., at the one-hop proximity of user devices. In such scenarios, the developer’s objective would be to improve the model’s accuracy and, concurrently, enhance the generalization performance of the model at the user level for maintaining regional databases using FA.

Select All
1.
D. Ramage and S. Mazzocchi, Federated Analytics: Collaborative Data Science Without Data Collection, Sep. 2020, [online] Available: https://ai.googleblog.com/2020/05/federated-analytics-collaborative-data.html.
2.
G. Zhu, D. Liu, Y. Du, C. You, J. Zhang and K. Huang, "Toward an intelligent edge: wireless communication meets machine learning", IEEE Commun. Mag., vol. 58, no. 1, pp. 19-25, Jan. 2020.
3.
A. H. Atabaki et al., "Integrating photonics with silicon nanoelectronics for the next generation of systems on a chip", Nature, vol. 556, no. 7701, pp. 349-354, Apr. 2018.
4.
B. McMahan, E. Moore, D. Ramage, S. Hampson and B. A. y Arcas, "Communication-efficient learning of deep networks from decentralized data", Proc. Int. Conf. Artif. Intell. Statist. (AISTATS), pp. 1273-1282, Apr. 2017.
5.
T. Li, A. K. Sahu, M. Zaheer, M. Sanjabi, A. Talwalkar and V. Smith, Federated optimization in heterogeneous networks, 2018, [online] Available: http://arxiv.org/abs/1812.06127.
6.
S. R. Pandey, N. H. Tran, M. Bennis, Y. K. Tun, A. Manzoor and C. S. Hong, "A crowdsourcing framework for on-device federated learning", IEEE Trans. Wireless Commun., vol. 19, no. 5, pp. 3241-3256, May 2020.
7.
P. Kairouz et al., Advances and open problems in federated learning, 2019, [online] Available: http://arXiv:1912.04977.
8.
L. U. Khan et al., "Federated learning for edge networks: Resource optimization and incentive mechanism", IEEE Commun. Mag., vol. 58, no. 10, pp. 88-93, Oct. 2020.
9.
M. Chen, H. V. Poor, W. Saad and S. Cui, Convergence time optimization for federated learning over wireless networks, 2020, [online] Available: http://arXiv:2001.07845.
10.
N. H. Tran, W. Bao, A. Zomaya, M. N. H. Nguyen and C. S. Hong, "Federated learning over wireless networks: Optimization model design and analysis", Proc. IEEE Conf. Comput. Commun. (INFOCOM), pp. 1387-1395, 2019.
11.
T. H. T. Le et al., "An incentive mechanism for federated learning in wireless cellular network: An auction approach", IEEE Trans. Wireless Commun., Mar. 2021.
12.
M. M. Amiri and D. Gündüz, "Federated learning over wireless fading channels", IEEE Trans. Wireless Commun., vol. 19, no. 5, pp. 3546-3557, May 2020.
13.
T. Nishio and R. Yonetani, "Client selection for federated learning with heterogeneous resources in mobile edge", Proc. IEEE Int. Conf. Commun. (ICC), pp. 1-7, May 2019.
14.
K. Yang, T. Jiang, Y. Shi and Z. Ding, "Federated learning via over-the-air computation", IEEE Trans. Wireless Commun., vol. 19, no. 3, pp. 2022-2035, Mar. 2020.
15.
M. Chen, Z. Yang, W. Saad, C. Yin, H. V. Poor and S. Cui, A joint learning and communications framework for federated learning over wireless networks, 2019, [online] Available: http://arXiv:1909.07972.
16.
W. Y. B. Lim et al., "Federated learning in mobile edge networks: A comprehensive survey", IEEE Commun. Surveys Tuts., vol. 22, no. 3, pp. 2031-2063, 3rd Quart. 2020.
17.
M. N. H. Nguyen et al., "Distributed and democratized learning: Philosophy and research challenges", IEEE Comput. Intell. Mag., vol. 16, no. 1, pp. 49-62, Feb. 2021.
18.
D. Gale and L. S. Shapley, "College admissions and the stability of marriage", Amer. Math. Monthly, vol. 69, no. 1, pp. 9-15, Jan. 1962.
19.
A. E. Roth and M. Sotomayor, "Two-sided matching", Handbook Game Theory Econ. Appl., vol. 1, pp. 485-541, Jan. 1992.
20.
M. N. Nguyen et al., Self-organizing democratized learning: Towards large-scale distributed learning systems, 2020, [online] Available: http://arXiv:2007.03278.
21.
A. Fallah, A. Mokhtari and A. Ozdaglar, Personalized federated learning: A meta-learning approach, 2020, [online] Available: http://arXiv:2002.07948.
22.
Y. Deng, M. M. Kamani and M. Mahdavi, Adaptive personalized federated learning, 2020, [online] Available: http://arXiv:2003.13461.
23.
M. Zinkevich, M. Weimer, A. J. Smola and L. Li, "Parallelized stochastic gradient descent", Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), pp. 2595-2603, Dec. 2010.
24.
X. Lian, C. Zhang, H. Zhang, C.-J. Hsieh, W. Zhang and J. Liu, "Can decentralized algorithms outperform centralized algorithms? A case study for decentralized parallel stochastic gradient descent", Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), pp. 5330-5340, Dec. 2017.
25.
J. Bernstein, Y.-X. Wang, K. Azizzadenesheli and A. Anandkumar, signSGD: Compressed optimisation for non-convex problems, 2018, [online] Available: http://arXiv:1802.04434.
26.
D. Alistarh, D. Grubic, J. Li, R. Tomioka and M. Vojnovic, "QSGD: Communication-efficient SGD via gradient quantization and encoding", Proc. Adv. Neural Inf. Process. Syst. (NeurIPS), pp. 1709-1720, Dec. 2017.
27.
Y. Lin, S. Han, H. Mao, Y. Wang and W. J. Dally, Deep gradient compression: Reducing the communication bandwidth for distributed training, 2017, [online] Available: http://arXiv:1712.01887.
28.
J. Konečnỳ, H. B. McMahan, F. X. Yu, P. Richtárik, A. T. Suresh and D. Bacon, Federated learning: Strategies for improving communication efficiency, 2016, [online] Available: http://arXiv:1610.05492.
29.
F. Sattler, S. Wiedemann, K.-R. Müller and W. Samek, "Robust and communication-efficient federated learning from non-IID data", IEEE Trans. Neural Netw. Learn. Syst., vol. 31, no. 9, pp. 3400-3413, Sep. 2020.
30.
S. Wang et al., "When edge meets learning: Adaptive control for resource-constrained distributed machine learning", Proc. IEEE Conf. Comput. Commun. (INFOCOM), pp. 63-71, Apr. 2018.
Contact IEEE to Subscribe

References

References is not available for this document.