Loading [MathJax]/extensions/MathMenu.js
Efficient Federated Learning for Cloud-Based AIoT Applications | IEEE Journals & Magazine | IEEE Xplore

Efficient Federated Learning for Cloud-Based AIoT Applications


Abstract:

As a promising method for central model training on decentralized device data without compromising user privacy, federated learning (FL) is becoming more and more popular...Show More

Abstract:

As a promising method for central model training on decentralized device data without compromising user privacy, federated learning (FL) is becoming more and more popular in Internet-of-Things (IoT) design. However, due to limited computing and memory resources of devices that restrict the capabilities of hosted deep learning models, existing FL approaches for artificial intelligence IoT (AIoT) applications suffer from inaccurate prediction results. To address this problem, this article presents a collaborative Big.Little branch architecture to enable efficient FL for AIoT applications. Inspired by the architecture of BranchyNet which has multiple prediction branches, our approach deploys deep neural network (DNN) models across both cloud and AIoT devices. Our Big.Little branch model has two branches, where the big branch is deployed on cloud for strengthened prediction accuracy, and the little branches are used to fit for AIoT devices. When AIoT devices cannot make the prediction with high confidence using local little branches, they will resort to the big branch for further inference. To increase both prediction accuracy and early exit rate of Big.Little branch model, we propose a two-stage training and coinference scheme, which considers the local characteristics of AIoT scenarios. Comprehensive experiment results obtained from a real AIoT environment demonstrate the efficiency and effectiveness of our approach in terms of prediction accuracy and average inference time.
Page(s): 2211 - 2223
Date of Publication: 22 December 2020

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Along with the prosperity of artificial intelligence (AI), deep learning (DL) techniques are increasingly deployed in Internet-of-Things (IoT) domains, such as commercial surveillance, autonomous driving, and robotics, where the requirements of model prediction accuracy and real-time response are of crucial importance [1]–[3]. However, due to limited computing and memory resources of AI IoT (AIoT) devices, both requirements cannot be guaranteed [4]–[6]. To extend the processing capabilities of AIoT devices for higher prediction accuracy, more and more AIoT applications rely on cloud computing by offloading some part of computation-intensive tasks to remote cloud servers. Therefore, the combination of strong cloud computing platforms and weak AIoT devices is becoming an emerging paradigm for large-scale AIoT design [7]–[9]. However, no matter whether the role of cloud servers in AIoT applications is for DL model training or inference, it is required that AIoT devices need to send their private data to cloud servers, where the concerns of user privacy and network latency cannot be neglected.

Select All
1.
F. Samie, L. Bauer and J. Henkel, "IoT technologies for embedded computing: A survey", Proc. Int. Conf. Hardw. Softw. Codesign Syst. Synth. (CODES + ISSS), pp. 1-10, 2016.
2.
R. Hadidi, J. Cao, M. S. Ryoo and H. Kim, "Robustly executing DNNs in IoT systems using coded distributed computing", Proc. ACM/IEEE Design Autom. Conf. (DAC), pp. 1-2, 2019.
3.
K. Pei, Y. Cao, J. Yang and S. Jana, "Deepxplore: Automated whitebox testing of deep learning systems", Proc. ACM Symp. Oper. Syst. Principles (SOSP), pp. 1-18, 2017.
4.
Z. Zhao, K. M. Barijough and A. Gerstlauer, "DeepThings: Distributed adaptive deep learning inference on resource-constrained IoT edge clusters", IEEE Trans. Comput.-Aided Design Integr. Circuits Syst., vol. 37, no. 11, pp. 2348-2359, Nov. 2018.
5.
M. Motamedi, D. Fong and S. Ghiasi, "Machine intelligence on resource-constrained IoT devices: The case of thread granularity optimization for CNN inference", ACM Trans. Embedded Comput. Syst., vol. 16, no. 5s, pp. 151:1-151:19, 2017.
6.
A. Xygkis, L. Papadopoulos, D. Moloney, D. Soudris and S. Yous, "Efficient winograd-based convolution kernel implementation on edge devices", Proc. Design Autom. Conf. (DAC), pp. 136:1-136:6, 2018.
7.
S. Teerapittayanon, B. McDanel and H. T. Kung, "Distributed deep neural networks over the cloud the edge and end devices", Proc. Int. Conf. Distrib. Comput. Syst. (ICDCS), pp. 328-339, 2017.
8.
Y. Chen, J. He, X. Zhang, C. Hao and D. Chen, "Cloud-DNN: An open framework for mapping DNN models to cloud FPGAs", Proc. Int. Symp. Field Program. Gate Arrays (FPGA), pp. 73-82, 2019.
9.
F. Forooghifar, A. Aminifar and D. Atienza, "Resource-aware distributed epilepsy monitoring using self-awareness from edge to cloud", IEEE Trans. Biomed. Circuits Syst., vol. 13, no. 6, pp. 1338-1350, Dec. 2019.
10.
H. B. McMahan, E. Moore, D. Ramage, S. Hampson and B. A. Y. Arcas, "Communication-efficient learning of deep networks from decentralized data", Proc. Int. Conf. Artif. Intell. Stat. (AISTATS), pp. 1273-1282, 2017.
11.
J. Konecný, H. B. McMahan, F. X. Yu, P. Richtárik, A. T. Suresh and D. Bacon, Federated learning: Strategies for improving communication efficiency, 2016, [online] Available: http://arXiv:1610.05492.
12.
P. Kairouz et al., Advances and open problems in federated learning, 2019, [online] Available: http://arXiv:1912.04977.
13.
K. Bonawitz et al., Towards federated learning at scale: System design, 2019, [online] Available: http://arXiv:1902.01046.
14.
V. Sze, Y.-H. Chen, T.-J. Yang and J. S. Emer, "Efficient processing of deep neural networks: A tutorial and survey", Proc. IEEE, vol. 105, no. 12, pp. 2295-2329, Dec. 2017.
15.
S. Teerapittayanon, B. McDanel and H. T. Kung, "BranchyNet: Fast inference via early exiting from deep neural networks", Proc. Int. Conf. Pattern Recognit. (ICPR), pp. 2464-2469, 2016.
16.
A. Botta, W. de Donato, V. Persico and A. Pescapé, "Integration of cloud computing and Internet of things: A survey", Future Gener. Comput. Syst., vol. 56, pp. 684-700, Mar. 2016.
17.
Y. Kang et al., "Neurosurgeon: Collaborative intelligence between the cloud and mobile edge", Proc. Int. Conf. Archit. Support Program. Lang. Oper. Syst. (ASPLOS), pp. 615-629, 2017.
18.
Z. Zhao et al., "A novel framework of three-hierarchical offloading optimization for MEC in industrial IoT networks", IEEE Trans. Ind. Informat., vol. 16, no. 8, pp. 5424-5434, Aug. 2020.
19.
B. Yang, X. Cao, C. Yuen and L. Qian, "Offloading optimization in edge computing for deep learning enabled target tracking by Internet-of-UAVs", IEEE Internet Things J., Aug. 2020.
20.
H. Gu et al., "A collaborative and sustainable edge-cloud architecture for object tracking with convolutional siamese networks", IEEE Trans. Sustain. Comput. early access, Nov. 2019.
21.
D. Liu et al., "FitCNN: A cloud-assisted and low-cost framework for updating CNNs on IoT devices", Future Gener. Comput. Syst., vol. 91, pp. 277-289, Feb. 2019.
22.
C. Gamanayake, L. Jayasinghe, B. K. K. Ng and C. Yuen, "Cluster Pruning: An efficient filter pruning method for edge AI vision applications", IEEE J. Sel. Topics Signal Process., vol. 14, no. 4, pp. 802-816, May 2020.
23.
T. D. Nguyen, S. Marchal, M. Miettinen, H. Fereidooni, N. Asokan and A.-R. Sadeghi, "DÏoT: A federated self-learning anomaly detection system for IoT", Proc. Int. Conf. Distrib. Comput. Syst. (ICDCS), pp. 756-767, 2019.
24.
S. Wang et al., "Adaptive federated learning in resource constrained edge computing systems", IEEE J. Sel. Areas Commun, vol. 37, no. 6, pp. 1205-1221, Jun. 2019.
25.
J. Mills, J. Hu and G. Min, "Communication-efficient federated learning for wireless edge intelligence in IoT", IEEE Internet Things J., vol. 7, no. 7, pp. 5986-5994, Jul. 2020.
26.
E. Bengio, P. Bacon, J. Pineau and D. Precup, Conditional computation in neural networks for faster models, 2015, [online] Available: http://arXiv:1511.06297.
27.
P. Panda, A. Sengupta and K. Roy, "Conditional deep learning for energy-efficient and enhanced pattern recognition", Proc. Design Autom. Test Eur. (DATE), pp. 475-480, 2016.
28.
T. Bolukbasi, J. Wang, O. Dekel and V. Saligrama, "Adaptive neural networks for efficient inference", Proc. Int. Conf. Mach. Learn. (ICML), pp. 527-536, 2017.
29.
M. Figurnov et al., "Spatially adaptive computation time for residual networks", Proc. Conf. Comput. Vis. Pattern Recognit. (CVPR), pp. 1790-1799, 2017.
30.
N. Agarwal, A. T. Suresh, F. X. Yu, S. Kumar and B. McMahan, "cpSGD: Communication-efficient and differentially-private distributed SGD", Proc. Annu. Conf. Neural Inf. Process. Syst. (NIPS), pp. 7575-7586, 2018.

Contact IEEE to Subscribe

References

References is not available for this document.