Introduction
In recent years, next-generation networks, also called NextG or 5G and beyond, have been paying attention more in academia and industry along with high demand and new ways of communication need from consumers. According to the report released by the International Telecommunication Union (ITU), the mobile data traffic based on NextG will constantly increase each year and reach thousands of exabytes [1]. NextG networks aim to connect billions of devices, systems, and applications to meet high data rate and low latency requirements to support new applications, especially delay-sensitive services using the Internet, from digital twins, virtual reality, metaverse, industry 4.0, self-driving cars, online education, to eHealth services and many more [2]. Fortunately, NextG networks can meet these requirements and support these applications with advanced communication, computing, and Artificial Intelligence (AI) technologies. AI is an extraordinary contributor among them to innovative technologies in NextG networks [3]. Intelligent Reflecting Surfaces (IRS) is one of those innovative technologies, in addition to Massive Multiple-Input Multiple-Output (MIMO) and millimeter wave, to improve the performance of NextG wireless networks in terms of data rate and channel capacity. Recently, IRS has received extensive attention in the literature due to its powerful capability of reconfiguring wireless communication environments. IRS is typically composed of a large amount of low-cost passive reflecting elements [4]. By cooperatively tuning the phase shifts of all reflecting elements, the reflected signals can be constructively or destructively added to the receiver [5]. Consequently, wireless communication environments could be changed dynamically to enhance or degrade communication performance.
Inspired by the tremendous achievements of AI, AI-powered models have also been applied to IRS-driven wireless communication in NextG wireless networks to improve performance [4], [6], [7], [8]. However, the security threats (e.g., model poisoning or adversarial machine learning attacks) and mitigation methods (e.g., adversarial training or defensive distillation) have not been investigated in AI-powered applications of NextG networks due to being new, complicated, and multi-disciplinary topics (e.g., next-generation communications, cybersecurity, and AI) [9], [10].
To fill the gap, this paper will focus on AI-powered IRS applications in 5G and beyond networks, and their vulnerabilities, which have received limited attention. Vulnerabilities of an AI-powered model are one of the top security concerns and deserve a thorough investigation. For example, a trained AI model might be manipulated by adding noise to the data, i.e., targeted and non-targeted adversarial attacks. The adversarial attacks are generated by adding a perturbation to a legitimate data point, i.e., an adversarial example, to fool the AI-powered models.
The major contributions of this paper are summarized as follows:
Evaluate the vulnerabilities of an AI-powered IRS model under widely used adversarial attacks, including Fast Gradient Sign Method (FGSM), Basic Iterative Method (BIM), Projected Gradient Descent (PGD), and Momentum Iterative Method (MIM).
Propose a defensive distillation mitigation method to train a more robust model to improve the robustness of the AI-powered IRS model.
Conduct the comprehensive simulations to assess the robustness of the proposed AI-powered IRS system with undefended and defended models under the above-mentioned adversarial attacks.
The results indicate that AI-powered models used in NextG networks are vulnerable to adversarial attacks, while the models can be more secure against adversarial attacks through the proposed defensive distillation mitigation method. Note that the scope of this study is limited to one of 5G physical layer applications, i.e., AI-powered IRS, its vulnerability analysis under selected adversarial attacks, and the proposed defensive distillation mitigation method. Other attack types like the Carlini & Wagner (C&W) attack are compute-intensive and require more iterations than traditional methods. Our study uses a less compute-intensive and more efficient way to create adversarial examples.
The remainder of the paper is organized as follows: Section II provides the background information about the IRS and common adversarial attacks. Section III presents the system overview, including the AI model and defense distillation for mitigation. Section IV shows experimental results, and Section V discusses the results along with observations. Section VI concludes the paper.
Preliminaries
This section provides background information and related works, including IRS and popular adversarial attacks using FGSM, BIM, PGD, and MIM.
A. Intelligent Reflecting Surfaces (IRS)
IRS is commonly proposed to improve wireless communication quality in various applications. Consider a typical IRS-aided wireless communication system as depicted in Figure 1. The IRS is deployed to enhance the communication performance between a transmitter and a receiver. The receiver gets the Line of Sight (LOS) signal through the LOS link as well as constructive reflected signals from IRS through the IRS-Rx link at the same time such that the communication performance between the transmitter and receiver could be significantly improved.
In the literature, there are several studies on IRS and security concerns [4], [7], [8]. AI-powered models, e.g., neural networks, have been integrated into IRS-aided systems to improve wireless communication performance. Authors in [11] propose the concept of Intelligent Spectrum Learning (ISL) to optimize IRS to tackle the interfering signals by dynamically controlling the IRS elements. The ISL first employs a well-trained convolutional neural network to realize a multi-class classification for the incident signals, and then the IRS elements can be turned on/off depending on the class of that signal by using an IRS binary control. Moreover, a dynamic “think-and-decide” function allows the reflection of incident signals to be blocked or passed based on the state of the IRS element block. Therefore, the Signal-to-Interference-plus-Noise Ratio (SINR) of the overall system can be improved. The study [12] presents a novel deep learning-based channel tracking algorithm in IRS-assisted UAV communication systems. The algorithm first designs a deep neural network with off-line training for signal denoising, and then a stacked bi-directional long short-term memory is developed to track the time-varying channel. Simulations demonstrate that this algorithm improves channel tracking performance while requiring fewer overheads for pilots than the benchmark algorithm. An IRS architecture is deployed to prevent the communications of multiple legitimate users from eavesdropping in the presence of multiple eavesdroppers [13]. They propose an approach that uses deep reinforcement learning to determine the optimal beamforming policy since the system is highly dynamic and complex.
It is challenging to acquire channel knowledge to estimate the Tx-IRS and IRS-Rx channel link in an IRS-assisted system since all the reflecting elements are expected to be nearly passive. Authors in [14] propose a new IRS architecture where all elements are passive except for a few active sensing elements and adopted a deep learning technique to assist the IRS in addressing this problem. Specifically, the transmitter and receiver first transmit two orthogonal uplink pilots to the active elements of IRS, and the active elements estimate the sampled channel vectors to construct the multipath signature as the environment descriptors. Motivated by recent advances in deep learning, this paper then proposes to train a neural network to observe the environment descriptors to predict the achievable rate with each IRS interaction vector. Based on the predictions, the IRS interaction vector corresponding to the highest predicted achievable rate will be used to reflect the transmitted data from the transmitter to the receiver. In our paper, we refer to the model above as the AI-powered IRS model and will investigate and examine the vulnerability of this model and apply the defensive distillation mitigation method.
B. Adversarial Attacks
Machine Learning (ML)-based models are trained to automatically learn the underlying patterns and correlations in data using algorithms. Once an ML-based model is trained, it can be used to predict the patterns in new data. The accuracy of the trained model is essential to achieving a high performance, which can also be called a generalization. However, the trained model can be manipulated by targeted and non-targeted adversarial ML attacks to fool the models. There are various kinds of adversarial ML attacks, such as evasion attacks, data poisoning attacks, and model inversion attacks.
Liu et al. [15] conducted a comprehensive survey on adversarial ML for wireless and mobile systems. Adversarial ML approaches can be used to generate and detect adversarial samples, which are samples that have been specifically designed to deceive a machine learning model. These samples can fool a model into misclassifying an input and can be used to exploit certain blind spots in image classifiers. The article reviews the state-of-the-art adversarial ML approaches to generating and detecting adversarial samples. It provides detailed discussions highlighting the open issues and challenges these approaches face.
An evasion attack aims to cause the ML-based models to misclassify the adversarial examples as legitimate data points, i.e., targeted and non-targeted evasion attacks. Targeted attacks aim to force the models to classify the adversarial example as a specific target class. Non-targeted attacks aim to push the models to classify the adversarial example as any class other than the ground truth. Data poisoning aims to generate malicious data points to train the ML-based models to find the desired output. It can be applied to the training data, which causes the ML-based models to produce the desired outcome. Model inversion aims to generate new data points close to the original data points to find the sensitive information of the specific data points.
These adversarial attack types are given as follows.
1) Fast Gradient Sign Method (FGSM)
FGSM is one of the most popular and straightforward approaches to constructing adversarial examples. It is called one-step gradient-based attack. It is used to compute the gradient of the loss function with respect to the input,
Compute the gradient of loss function,
\nabla _{\mathbf {x}}\ell (\mathbf {x},\mathbf {y}) Add the gradient to the input data,
,\mathbf {x}_{adv} = \mathbf {x} + \epsilon \times sign(\nabla _{\mathbf {x}}\ell)
2) Basic Iterative Method (BIM)
BIM is one of the most popular attacks called an iterative gradient-based attack. This attack is derived from the FGSM attack. It is used to compute the gradient of the loss function with respect to the input,
Initialize the adversarial example as
\mathbf {x}_{adv} = \mathbf {x} Iterate
times, wherei i=0, 1, 2, 3,\ldots, N Compute the gradient of loss function,
\nabla _{\mathbf {x}}\ell (\mathbf {x}_{adv},\mathbf {y}) Add the gradient to the input data,
,\mathbf {x}_{adv} = \mathbf {x}_{adv} + \epsilon \times sign(\nabla _{\mathbf {x}}\ell)
3) Projected Gradient Descent (PGD)
PGD is one of the most popular and powerful attacks [18]. It is used to compute the gradient of the loss function with respect to the input,
Initialize the adversarial example as
\mathbf {x}_{adv} = \mathbf {x} Iterate
times, wherei i=0, 1, 2, 3,\ldots, N Compute the gradient of loss function,
\nabla _{\mathbf {x}}\ell (\mathbf {x}_{adv},\mathbf {y}) Add random noise to the gradient,
\hat {\nabla }_{\mathbf {x}}\ell (\mathbf {x}_{adv},\mathbf {y}) = \nabla _{\mathbf {x}}\ell (\mathbf {x}_{adv},\mathbf {y}) + \mathcal {U}(\epsilon) Add the gradient to the input data,
,\mathbf {x}_{adv} = \mathbf {x}_{adv} + \alpha \times sign(\hat {\nabla }_{\mathbf {x}}\ell)
4) Momentum Iterative Method (MIM)
MIM is a variant of the BIM adversarial attack, introducing momentum and integrating it into iterative attacks [19]. It is used to compute the gradient of the loss function with respect to the input,
Initialize the adversarial example
and the momentum,\mathbf {x}_{adv} = \mathbf {x} \mu = 0 Iterate
times, wherei i=0, 1, 2, 3,\ldots, N Compute the gradient of loss function,
\nabla _{\mathbf {x}}\ell (\mathbf {x}_{adv},\mathbf {y}) Update the momentum,
\mu = \mu + \frac {\eta }{\epsilon } \times \nabla _{\mathbf {x}}\ell (\mathbf {x}_{adv},\mathbf {y}) Add random noise to the gradient,
\hat {\nabla }_{\mathbf {x}}\ell (\mathbf {x}_{adv},\mathbf {y}) = \nabla _{\mathbf {x}}\ell (\mathbf {x}_{adv},\mathbf {y}) + \mathcal {U}(\epsilon) Add the gradient to the input data,
,\mathbf {x}_{adv} = \mathbf {x}_{adv} + \alpha \times sign(\hat {\nabla }_{\mathbf {x}}\ell)
Note that there are many types of adversarial attacks and defenses. The existing defenses and adversarial attacks for images can be applied to attack and defend on intelligent reflecting surfaces and other fields [20], [21], [22], [23]. The cleverly-designed adversarial examples can fool the deep neural networks with high success rates on the test images. The adversarial examples can also be transferred from one model to another model. In our experiments, we generated the adversarial inputs with untargeted attacks.
System Overview
This section presents the overall system model for the proposed AI-powered IRS system, as illustrated in Figure 2. According to the figure, it is assumed that data collected from User Equipments (UEs) is provided to the IRS prediction model. The undefended model covers only conventional training of deep neural networks, while the defended model covers the defensive distillation-based training method. The defensive distillation method covers the teacher and student models. The teacher model is typically a large deep neural network, while the student model is usually a small and shallow neural network. In the figure, the training of the prediction model (i.e., student model) is protected against adversarial ML attacks in base stations. Adversarial attacks are applied to models, i.e., undefended and defended models, to evaluate the models’ robustness under any attacks.
A. Deep Neural Networks
As we briefly discussed in Section II-A, a neural network is designed for mapping the observed environment descriptors to the predicted achievable rate in the AI-powered IRS model. This subsection introduces the neural network architecture and training details below.
Neural Network Architecture: The input of the neural network model is defined as a stack of the environment descriptors (i.e., uplink pilot signals) received from both transmitter and receiver. Since the training process is designed to build a function mapping descriptors to reflection vectors, the output target of the neural network is to be a set of predictions on the achievable rates of every possible reflection beamforming vector. The neural network is built as a Multi-Layer Perceptron (MLP) network, which is well-demonstrated as an effective universal approximator. The MLP is adopted to establish the connection between the environment descriptors and the predicted achievable rates using reflection beamforming vectors, as shown in Figure 3. The MLP is composed of four fully connected layers. ReLU activation function is adopted, and a dropout layer is added after the activation function for every layer except for the last layer. The MLP consists of the following dimensions:
(Input), [M ,M ](Layer1), [2M ,2M ](Layer2), [4M ,4M ](Layer3), [4M ,4M ](Layer4), whereM is the number of the antenna elements on IRS.M Training Details: The training dataset has 54300 data samples since the candidate receiver locations contain 54300 points as discussed in III-C. The dataset is split into two sets, namely a training set and a testing set with 85% and 15% of the points, respectively. To measure the quality of the predictions and make the predicted achievable rates close to the real achievable rates in the dataset, we define the loss function with Mean-Squared-Error (MSE) between them. In the training process, the batch size is set to 500 samples, and the training epochs is set to 20. The dropout rate is set to 50%, and a
regularization term with the factor of 10−4 is added to the loss function. The learning rate decreases by 50% every 3 epochs starting at 0.1 with Stochastic Gradient Descent (SGD) optimizer.L_{2}
The adopted neural network architecture is composed of four fully connected layers. The number of the neurons of the four layers is (
B. Defensive Distillation
As mentioned previously, in this paper, we leverage the defensive distillation mitigation method to improve the robustness of our AI-powered IRS model. Defensive distillation is a method that applies defensive knowledge distillation to train a more robust model [24]. Knowledge distillation was previously introduced by Hinton et al. [25] to compress the knowledge of a large, densely connected neural network (the teacher) into a smaller, sparsely connected neural network (the student). It has been shown that the student could achieve a similar performance as the teacher by mimicking the teacher’s output, and the teacher would be used as a soft label to train the student. Furthermore, the student could be trained to be more resistant to adversarial attacks than the teacher by using the label of the teacher as the label of the student [26].
The architecture of the defensive distillation consists of the following steps:
Step 1: Train a model with cross-entropy loss as the classification task’s base model (teacher).
Step 2: Train the same model (teacher) with defensive distillation loss (soft label + cross-entropy) to generate the respective soft label.
Step 3: Train a model with the soft label generated in step 2 as the label (student) to obtain the robust model.
The defensive distillation loss function is defined as \begin{equation*} \mathcal {L}_{D} \left ({\theta }\right) = \left ({1 - \lambda }\right) \mathcal {L}_{CE} \left ({\theta }\right) + \lambda \mathcal {L}_{KL} \left ({P_{T} \left ({y | \theta }\right), P_{T} \left ({y }\right) }\right),\quad \tag{1}\end{equation*}
Algorithm 1 Training the Defensive Distillation
Input: Training data set
Output: Defensive distillation model
Train the base model
Initialize the defensive distillation model
while
Get a batch of samples
Calculate the cross entropy loss
Calculate the defensive distillation loss
Calculate the adversarial samples
Calculate the new loss
Update the weights of the defensive distillation model
end while
return
C. Dataset Preparation
To examine the performance of the AI-powered IRS model, a publicly available ray-tracing-based DeepMIMO dataset [27] is adopted to generate the training dataset. DeepMIMO dataset is a parameterized dataset designed for constructing the MIMO channels based on ray-tracing data obtained from the accurate ray-tracing scenario simulation. Similar to the simulation setup in [14], the outdoor ray-tracing scenario ‘O1’ is selected as shown in Figure 4. Base Station 3 (BS 3) is set as an IRS, which is equipped with an UPA (Uniform Planar Array) with 32
The adopted ray-tracing scenario where the large intelligent surface (i.e., IRS) is deployed to reflect the signal from the fixed transmitter to the candidate receivers.
D. Performance Metric
This study evaluates the AI-powered IRS model through the Mean Squared Error (MSE) performance metric. MSE scores are utilized to analyze the model vulnerabilities under undefended and defended conditions. The equation regarding the MSE score is given below.\begin{equation*} MSE = {\frac {\sum ^{}{(Y_{t} - {\hat {Y}}_{t})}^{2}}{n}}\tag{2}\end{equation*}
: The actual tY_{t} instance,^{\mathrm{ th}} : The forecasted t{\hat {Y}}_{t}~ instance,^{\mathrm{ th}} n: The total number of instance
MSE score measures the average squared difference between the actual and predicted values. A high MSE score represents a high prediction error.
Experimental Results
This section analyses the results obtained from the experiments related to AI-powered IRS models against adversarial machine learning attacks. Results are represented in three ways: (1) bar plots showing the impact of each adversarial machine learning attack on the performance of undefended and defended models, i.e., MSE, (2) histogram plots showing the MSE metric values for each attack of defended and undefended models, and (3) the table showing the prediction performance results of defended and undefended models for each adversarial attack. Figure 5–6 show the bar plots, while Figure 7–10 show the histogram plots. Table 2 shows the prediction performance results of the defended and undefended AI-powered IRS models against the attacks.
MSE values of the undefended models for each adversarial machine learning attack under different attack powers (
MSE values of the defended models for each adversarial machine learning attack under different attack powers (
Distribution of MSE values for undefended and defended models under the FGSM attack.
The trained AI-powered IRS model is implemented using Python 3.7.13 and the TensorFlow 2.8.2 framework running on Google Colab Tesla T4 GPU with 16GB of memory. Adversarial inputs are generated using Cleverhans 4.0.0. library.
The adversarial attack on AI-powered models has become more popular with various attack methods. This study uses FGSM, MIM, BIM, and PGD methods to generate adversarial examples. The performance of each model is evaluated through the MSE metric.
Figure 5 shows MSE values for the selected attack methods under attack powers from
The histogram plots investigate the distribution of MSE values for undefended and defended models under adversarial attacks. In Figure 7–10, (a) represents the undefended models, while (b) represents defended models for each attack, i.e., FGSM, BIM, MIM, and PGD, respectively. According to the results, the undefended models, i.e., (a), represent a little right-skewed distribution, which has a peak to the left of the distribution and data values that taper off to the right. MSE values vary from 0.005 to 0.025 for all attack types, and around 50% percent of MSE values are between 0.006 and 0.009. It is compatible with Figure 5–6. On the other hand, it is difficult to define the histogram plots for defended models, i.e., (b). According to the results, Figure 7, 8, 10(b) represent a little right-skewed distribution like the undefended model ones, while Figure 9(b) does not represent any distribution. The most MSE values are clustered around 0.0, i.e., 30% - 60%. It means the AI-powered model can correctly predict the target values. It is also clear that the percent of the high MSE values (<0.015) is much lower than the undefended model. The defended models are more effective against FGSM and BIM attacks, as shown in Figure 7 and 8. It is obvious that the mitigation methods can dramatically improve the model robustness under FGSM attacks, i.e., 90% of MSE values are less than 0.005. On the other hand, the defended models are not successful against MIM and PGD attacks compared to FGSM and BIM, as shown in Figure 9 and 10. Although low MSE values, i.e., <0.005, are clustered around 50%, the MSE values still go up to 0.015 for MIM and PGD attacks.
Table 2 shows the impact of a specific
Discussion
This study investigates AI-powered IRS models in NextG networks and their vulnerabilities against adversarial attacks and the contribution of mitigation methods to the model robustness. The models’ vulnerabilities are studied for various adversarial attacks, i.e., FGSM, BIM, MIM, and PGD, as well as the mitigation method, i.e., defensive distillation. The results show that AI-powered IRS models are vulnerable to adversarial attacks. On the other hand, the mitigation methods can significantly improve the model robustness under adversarial attacks. According to the results, adversarial attacks on AI-powered IRS models and the use of the proposed mitigation method can be summarized as:
Observation 1: AI-powered IRS models are vulnerable to adversarial attacks, especially BIM with a high attack power (
).\epsilon >0.5 Observation 2: There is no significant impact of the attack power (
) on some adversarial attacks, i.e., FGSM.\epsilon Observation 3:The defensive distillation mitigation method significantly increases the model robustness, especially under FGSM and BIM attacks.
Observation 4:The MSE values histogram usually represents a smaller right-skewed distribution, especially for the undefended models.
Observation 5:Around 50% percent of MSE values are between 0.006 and 0.009 for the undefended models.
Observation 6:The most MSE values are clustered around 0.0, i.e., 30% - 60% for the defended model.
Observation 7:The most effective adversarial attack types are BIM and MIM for undefended and defended models, respectively.
Conclusion and Future Work
The next generation networks, i.e., NextG or 5G and beyond, have dramatically enhanced along with advanced communication, computing, and AI technologies in the last decade. AI is the most important contributor to NextG networks’ improvement in terms of performance. This paper investigates the vulnerability of AI-powered IRS models against adversarial attacks (i.e., FGSM, BIM, PGD, and MIM) and the impact of the proposed mitigation method, i.e., defensive distillation, on the improvement of models’ robustness in NextG networks. The results indicate that the AI-powered NextG networks are vulnerable to adversarial attacks. On the other hand, mitigation methods can make the models more robust against adversarial attacks. According to the overall results, the most effective adversarial attack types are BIM and MIM for undefended and defended models, respectively. The proposed mitigation method can provide better results for the attacks, including FGSM, BIM, MIM, and PGD, in terms of increasing the model robustness and reducing the vulnerability.
In future work, the authors will focus on automatic modulation classification using an AI-powered model in NextG networks and its vulnerability under adversarial attacks.