I. Introduction
Federated learning (FL) is an emerging collaborative learning paradigm that improves privacy by enabling clients to keep their training data locally. Particularly, it enables machine learning models to be continuously improved and optimized through distributed and localized training [1], [2]. Despite its benefits, FL has been demonstrated to be vulnerable to backdoor attacks. Backdoor attacks are a typical adversarial attack type on deep neural networks (DNNs). Gu et al. [3] first explored backdoor attacks, known as Badnets, for DNNs. During training, attackers embed hidden backdoor triggers in DNNs, enabling normal classification on benign samples. However, activating the hidden backdoor trigger causes the model to output the attacker’s desired label. The backdoor attack was first introduced in FL by Bagdasaryan et al. [4]. The goal of the adversary is to plant backdoors in their local model such that the backdoored model will be aggregated into the global model, further inheriting the backdoor logic. Since the universal participation principle and invisible training procedure, backdoor attacks in FL emerge more destructive effects [5], [6], [7], [8], [9].