Loading web-font TeX/Math/Italic
Loci: Federated Continual Learning of Heterogeneous Tasks at Edge | IEEE Journals & Magazine | IEEE Xplore

Loci: Federated Continual Learning of Heterogeneous Tasks at Edge


Abstract:

Federated continual learning (FCL) has attracted growing attention in achieving collaborative model training among edge clients, each of which learns its local model for ...Show More

Abstract:

Federated continual learning (FCL) has attracted growing attention in achieving collaborative model training among edge clients, each of which learns its local model for a sequence of tasks. Most existing FCL approaches aggregate clients’ latest local models to exchange knowledge. This unfortunately deviates from real-world scenarios where each model is optimized independently using the client’s own dynamic data and different clients have heterogeneous tasks. These tasks not only have distinct class labels (e.g., animals or vehicles) but also differ in input feature distributions. The aggregated model thus often shifts to a higher loss value and incurs accuracy degradation. In this article, we depart from the model-grained view of aggregation and transform it into multiple task-grained aggregations. Each aggregation allows a client to learn from other clients to improve its model accuracy on one task. To this end, we propose Loci to provide abstractions for clients’ past and peer task knowledge using compact model weights, and develop a communication-efficient approach to train each client’s local model by exchanging its tasks’ knowledge with the most accuracy relevant one from other clients. Through its general-purpose API, Loci can be used to provide efficient on-device training for existing deep learning applications of graph, image, nature language processing, and multimodal data. Using extensive comparative evaluations, we show Loci improves the model accuracy by 32.48% without increasing training time, reduces communication cost by 83.6%, and achieves more improvements when scale (task/client number) increases.
Published in: IEEE Transactions on Parallel and Distributed Systems ( Volume: 36, Issue: 4, April 2025)
Page(s): 775 - 790
Date of Publication: 29 January 2025

ISSN Information:

Funding Agency:

References is not available for this document.

I. Introduction

Artificial intelligence (AI) applications are widely deployed on edge devices or clients, e.g., classifying image objects [1], [2], translating text [3], and recognizing patterns in social graphs [4]. Federated learning (FL) [5] enables edge clients to collaboratively learn a global model by aggregating their local models without exchanging the raw data. Federated continuous learning (FCL) addresses scenarios where clients incrementally train models over a sequence of tasks characterized by their data distributions. For example, Fig. 1(a) illustrates a FCL scenario of n edge clients in an image classification application. Each client learns its own sequence of heterogeneous tasks, namely distinct class labels (e.g., cars, ships, and houses in client 1, and cats, dogs, and horses in client 2) and a subspace of input feature distribution for each task. Fig. 1(c) further shows that such heterogeneous tasks widely exist in edge AI applications of different data modalities, such as graph [4], text [3], and multi-modal data [1], [2]. It is exceedingly challenging to learn AI models that can be generalized on such heterogeneous tasks, which vary across clients and over time, especially on resource strenuous edge devices.

Select All
1.
E. Kristiani, C.-T. Yang and C.-Y. Huang, "iSEC: An optimized deep learning model for image classification on edge computing", IEEE Access, vol. 8, pp. 27267-27276, 2020.
2.
S. Abdel Magid, F. Petrini and B. Dezfouli, "Image classification on IoT edge devices: Profiling and modeling", Cluster Comput., vol. 23, pp. 1025-1043, 2020.
3.
I. Haritaoglu, "InfoScope: Link from real world to digital information space", Proc. Int. Conf. Ubiquitous Comput., pp. 247-255, 2001.
4.
L. Zeng, C. Yang, P. Huang, Z. Zhou, S. Yu and X. Chen, "GNN at the edge: Cost-efficient graph neural network processing over distributed edge servers", IEEE J. Sel. Areas Commun., vol. 41, no. 3, pp. 720-739, Mar. 2023.
5.
Z. Zhou, X. Chen, E. Li, L. Zeng, K. Luo and J. Zhang, "Edge intelligence: Paving the last mile of artificial intelligence with edge computing", Proc. IEEE, vol. 107, no. 8, pp. 1738-1762, Aug. 2019.
6.
X. Yang, H. Yu, X. Gao, H. Wang, J. Zhang and T. Li, "Federated continual learning via knowledge fusion: A survey", IEEE Trans. Knowl. Data Eng., vol. 36, no. 8, pp. 3832-3850, Aug. 2024.
7.
J. Kirkpatrick et al., "Overcoming catastrophic forgetting in neural networks", Proc. Nat. Acad. Sci. USA, vol. 114, no. 13, pp. 3521-3526, 2017.
8.
S. Jung, H. Ahn, S. Cha and T. Moon, "Continual learning with node-importance based adaptive group sparse regularization", Proc. Adv. Neural Inf. Process. Syst., pp. 3647-3658, 2020.
9.
D. Abati, J. Tomczak, T. Blankevoort, S. Calderara, R. Cucchiara and B. E. Bejnordi, "Conditional channel gated networks for task-aware continual learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 3931-3940, 2020.
10.
G. Cao et al., "E2-AEN: End-to-end incremental learning with adaptively expandable network", 2022.
11.
J. Lu, A. Liu, F. Dong, F. Gu, J. Gama and G. Zhang, "Learning under concept drift: A review", IEEE Trans. Knowl. Data Eng., vol. 31, no. 12, pp. 2346-2363, Dec. 2019.
12.
B. McMahan, E. Moore, D. Ramage, S. Hampson and B. A. Y. Arcas, "Communication-efficient learning of deep networks from decentralized data", Proc. Int. Conf. Artif. Intell. Statist., pp. 1273-1282, 2017.
13.
Y. Deng, M. M. Kamani and M. Mahdavi, "Adaptive personalized federated learning", 2020.
14.
C. He, M. Annavaram and S. Avestimehr, "Group knowledge transfer: Federated learning of large CNNs at the edge", Proc. Adv. Neural Inf. Process. Syst., pp. 14068-14080, 2020.
15.
L. Collins, H. Hassani, A. Mokhtari and S. Shakkottai, "Exploiting shared representations for personalized federated learning", Proc. Int. Conf. Mach. Learn., pp. 2089-2099, 2021.
16.
J. Dong et al., "Federated class-incremental learning", Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., pp. 10164-10173, 2022.
17.
J. Yoon, W. Jeong, G. Lee, E. Yang and S. J. Hwang, "Federated continual learning with weighted inter-client transfer", Proc. Int. Conf. Mach. Learn., pp. 12073-12086, 2021.
18.
Y. Luopan, R. Han, Q. Zhang, C. H. Liu, G. Wang and L. Y. Chen, "FedKNOW: Federated continual learning with signature task knowledge integration at edge", Proc. IEEE Int. Conf. Data Eng., pp. 341-354, 2023.
19.
F. Sattler, K.-R. Müller and W. Samek, "Clustered federated learning: Model-agnostic distributed multitask optimization under privacy constraints", IEEE Trans. Neural Netw. Learn. Syst., vol. 32, no. 8, pp. 3710-3722, Aug. 2021.
20.
A. Ghosh, J. Chung, D. Yin and K. Ramchandran, "An efficient framework for clustered federated learning", Proc. Adv. Neural Inf. Process. Syst., pp. 19586-19597, 2020.
21.
G. Tong, G. Li, J. Wu and J. Li, "GradMFL: Gradient memory-based federated learning for hierarchical knowledge transferring over non-IID data", Proc. Int. Conf. Algorithms Architectures Parallel Process., pp. 612-626, 2021.
22.
Z. Wang, Y. Zhang, X. Xu, Z. Fu, H. Yang and W. Du, "Federated probability memory recall for federated continual learning", Inf. Sci., vol. 629, pp. 551-565, 2023.
23.
X. Yao and L. Sun, "Continual local training for better initialization of federated models", Proc. IEEE Int. Conf. Image Process., pp. 1736-1740, 2020.
24.
Y. Chaudhary, P. Rai, M. Schubert, H. Schütze and P. Gupta, "Federated continual learning for text classification via selective inter-client transfer", 2022.
25.
J. Dong et al., "Dual learning with dynamic knowledge distillation for partially relevant video retrieval", Proc. IEEE/CVF Int. Conf. Comput. Vis., pp. 11302-11312, 2023.
26.
M. Contributors, "MMCV: OpenMMLab computer vision foundation", 2018, [online] Available: https://github.com/open-mmlab/mmcv.
27.
T. Wolf et al., "Transformers: State-of-the-art natural language processing", Proc. Conf. Empir. Methods Natural Lang. Process., pp. 38-45, 2020.
28.
L. Peng, N. Wang, N. Dvornek, X. Zhu and X. Li, "FedNI: Federated graph learning with network inpainting for population-based disease prediction", IEEE Trans. Med. Imag., vol. 42, no. 7, pp. 2032-2043, Jul. 2023.
29.
J. Stremmel and A. Singh, "Pretraining federated text models for next word prediction", Proc. Future Inf. Commun. Conf., pp. 477-488, 2021.
30.
Y. Liu et al., "FedVision: An online visual object detection platform powered by federated learning", Proc. AAAI Conf. Artif. Intell., pp. 13172-13179, 2020.
Contact IEEE to Subscribe

References

References is not available for this document.