Introduction
In recent years, substantial scientific and industrial interest has been devoted to the sixth generation (6G) of wireless networks. 6G systems will foster next-generation vertical services and facilitate the establishment of an intelligent and fully connected society by seamlessly connecting people, things, data, applications, smart cities, and transportation systems. Thanks to its enhanced capabilities, 6G is poised to accommodate a multitude of devices and manage extensive data, prioritizing the delivery of services that augment the user experience and cater to emerging applications, including the Internet of Things (IoT), Augmented Reality (AR), Virtual Reality (VR), holographic communications, immersive media streaming, and smart societies [1]. 6G networks are anticipated to offer ubiquitous connectivity, low latency, high capacity, and high reliability to fulfill the various quality-of-service (QoS) requirements of the newly added services and applications. Existing terrestrial wireless networks, constrained by limited coverage and capacity, face challenges in universally delivering reliable and cost-effective wireless services, particularly in challenging locations like disaster-stricken areas, mountainous regions, and oceans [2].
Space-Air-Ground Integrated Networks (SAGINs) will play an increasingly vital role in meeting the demand for seamless connectivity in 6G systems [3]. SAGINs consist of three network segments, i.e., the communication satellites form the space sub-network; aerial communication devices such as balloons, airships, high-altitude platforms (HAPs), and unmanned aerial vehicles (UAVs) form the aerial sub-network; and terrestrial communication networks form the ground sub-network. A key advantage of SAGINs is their ability to extend the coverage of terrestrial networks and alleviate congestion caused by a growing number of connected devices. This capability is crucial as deploying emerging technologies, such as autonomous aerial fleets, cargo drones, and flying cars, becomes more prevalent [4].
Fig. 1 provides an overview of how different elements of SAGINs are distributed on Earth and in the atmosphere. The ground sub-network consists mainly of base stations (BSs), with coverage limited to nearby users. Consequently, users located further away or in regions lacking terrestrial network infrastructure may experience performance issues or a complete lack of service. A viable option to enhance ground networks, satellite communication systems became essential for enabling worldwide connectivity and closing the digital divide [5]. Next-generation SAGINs, especially those involving UAVs, are expected to be supported by satellite networks for terrestrial and aerial communications [6]. Low-Earth orbit (LEO) constellations, which are usually deployed at altitudes of 300 to 1500 km, are the key elements in the future vision of global communication and coverage and are taking the leading role towards the integration of satellites in the SAGIN architecture [7]. Geostationary (GEO) and medium-earth orbit (MEO) are yet other types of satellites that usually orbit the Earth at altitudes of 35786 km and 7000 to 25000 km, respectively [8].
Various components of SAGIN (satellites, HAPs, UAVs, and the terrestrial network) serving users in urban, remote, rural, and isolated areas.
Recent years have witnessed significant strides in satellite technology, notably with private initiatives like SpaceX, OneWeb, and Amazon leading the charge in satellite constellation design, operation, and research [9]. While efforts are underway to standardize the next generation of satellite networks, several fundamental challenges persist. These challenges include resource management, network control, network security, spectrum management, and energy usage [10]. They are mainly due to the distinctive attributes of satellites, such as rapid orbital motion, varying coverage range, and evolving topology. Meeting the demand for adaptability in a dynamically changing radio environment inherent in satellite networks necessitates the development of agile methodologies capable of rapid and near-instantaneous adaptation [11].
Apart from the space sub-network, the two main components of the aerial sub-network are the HAPSs and UAVs. HAPSs, usually deployed in the stratosphere at cloudfree atmospheric altitudes between 20 and 50 km, can vary in design features, operability (manned or unmanned), stability (aerostatic or aerodynamic), mobility (free-flying or controlled), autonomy (autonomous or tethered), and energy source (motorized or non-motorized) [12]. The high deployment altitude of HAPSs makes them more practical in providing stable access to large numbers of users in remote areas and lower transmission energy costs than traditional fixed BSs. Furthermore, HAPSs are promoted to provide mobile edge computing (MEC) services to swiftly manage computational tasks and effectively handle the surge in traffic spurred by the expanded adoption of the IoT. HAPSs can also act as relay nodes between the satellites and the terrestrial network [13]. Because of their unique benefits, such as adaptable deployment, cost efficiency, simplified maintenance, and broader service coverage, HAPSs have the potential to enhance system throughput, thereby boosting the scalability of SAGINs.
UAVs can operate at various altitudes and exhibit different design features, such as fixed-wing or multi-rotor configurations. They can be operated remotely or autonomously, and their energy source can be battery-powered or fueled [14]. UAVs can also be equipped with lightweight BS equipment and act as airborne BSs to serve users in challenging locations and when the terrestrial network is down [15]. This solution extends further to scenarios that require high capacity and enhanced service quality, including highly populated areas and hotspots [4]. UAVs can, therefore, provide on-demand support to the terrestrial network, effectively managing unexpected traffic spikes and emergency cases [16]. On the other side, UAVs can be leveraged as aerial users that use existing infrastructure (connect to either the terrestrial or the space sub-networks) to serve diverse tasks such as parcel delivery, search and rescue missions, inspections, surveillance, and event streaming [17]. In both scenarios, UAVs face several challenges that must be addressed for efficient integration in SAGIN systems. These objectives include but are not limited to optimizing service quality, reducing energy usage, guaranteeing user and core network connectivity, and mitigating interference. In this context, UAVs, for instance, must optimize the flying trajectory to meet performance requirements and provide the expected service while taking system constraints into consideration [18]. UAVs also suffer from limited flight endurance and payload capacity, impacting the range and breadth of the communication systems they can support [14]. Despite these challenges, UAVs offer the advantage of rapid deployment and mobility, making them an indispensable component of the SAGIN architecture.
By leveraging SAGINs, 6G systems can benefit from their ability to maintain reliable, uninterrupted, and low-latency communication, mainly in difficult-to-reach locations and densely populated cities. These diverse applications also highlight the need for intelligent algorithms and improved orchestration techniques in SAGINs to address the challenges in task allocation, mobility handling, and radio resource management (RRM) [19]. Optimization efforts are also mandatory to balance the different targeted objectives, including energy efficiency, coverage region, and user satisfaction. Furthermore, the intrinsic dynamics of SAGIN environments and fluctuating user demands require high adaptability and intelligence within SAGINs. Traditional optimization techniques and heuristic algorithms fail to meet these requirements due to their constraints in flexibility and adaptability, especially given the evolving and diverse nature of SAGINs [20].
In another context, Artificial Intelligence (AI) has emerged in the last few years as a collection of algorithms that mimic human behavior, using predictions based on available data. Moreover, AI has become prevalent in scientific research for decision-making and optimization problems thanks to the huge amount of available data nowadays and highperformance computing systems. AI is expected to be a key pillar for 6G networks, providing numerous benefits and driving the evolution of future wireless communications [21]. Unlike conventional, reactive methods, AI techniques empower wireless networks to learn and adapt autonomously. This self-driven intelligence allows them to analyze non-linear network behaviors and automatically adjust configurations for optimal efficiency. Recently, AI algorithms have found their way into various applications in SAGINs. Specifically, AI algorithms can address the limitations of conventional optimization approaches in SAGINs by efficiently managing resources, intelligently allocating tasks, and autonomously making decisions while accounting for the dynamic, heterogeneous, and distributed characteristics of the SAGIN environment [22].
Among the different existing AI algorithms, Reinforcement Learning (RL) has demonstrated remarkable success in optimizing various aspects of SAGINs [23]. RLbased algorithms offer an optimal and adaptive solution for dynamic SAGIN environments through their ability to discern intricate interdependencies between network parameters and environmental cues. RL is mainly used to enhance the orchestration and control of SAGINs by optimizing factors like trajectory planning and resource allocation. Furthermore, recent advancements in Machine Learning (ML), Deep Learning (DL), and Neural Networks (NNs) can effectively address scalability issues in largescale SAGINs, facilitating. As a result, the development of adaptable, efficient, and intelligent SAGINs [24]. Federated learning (FL) is yet another efficient AI algorithm that can adapt to the distributed architecture of SAGINs by involving different SAGIN layers in the learning process, further enhancing the scalability and adaptability of SAGIN systems [25].
While the notion of AI driving SAGINs prevails, the possibility of a reverse influence requires further investigation. Certainly, AI algorithms are crucial for SAGINs, extracting insights from vast data and optimizing performance. However, we propose that SAGINs can potentially enhance several AI domains. For instance, they offer a rich source of diverse, high-quality data vital for improving model accuracy. In addition, their edge computing capabilities can significantly accelerate AI training. Given the increasing importance of AI in SAGINs, there is a need to provide a comprehensive survey of the state-of-the-art research on this topic. In light of this, this paper aims to serve this need by providing a holistic overview of the role and applications of AI in future SAGINs, summarizing and identifying the key challenges for different SAGIN layers, highlighting the interplay between AI and SAGIN, and discussing potential future directions. While we acknowledge the importance of AI in ground networks, this survey primarily focuses on the non-terrestrial aspects of SAGINs, as several existing surveys have already extensively covered AI applications in terrestrial networks [26], [27].
A. Previous Surveys
With the large number of published papers considering the integration of AI technologies in SAGINs, several surveys have tried to summarize the existing literature to provide a holistic overview of state-of-the-art research in AI-enabled SAGINs [11], [22], [23], [24], [30], [34], [35]. Furthermore, several survey papers discuss SAGINs and highlight their main potential applications, challenges, and their role in 5G and 6G, envisioning diversified architecture where non-terrestrial stations seamlessly support the terrestrial infrastructure, paving the way for unprecedented connectivity [28], [31], [32], [33]. Moreover, several surveys covering 6G applications and challenges have promoted SAGINs and AI as key enabling technologies to maximize the coverage of terrestrial networks and introduce scalable and flexible management and control of the wireless network [29]. For instance, the authors in [29] provide a comprehensive survey about 6G and separately discuss SAGINs as an emerging architecture and AI as a tool for networking.
The work done in [28] outlines the advantages and challenges associated with SAGINs and mentions AI only as a tool that will be used for 6G technologies for real-time control and optimization. Another survey [31] offers a separate chapter on DL for SAGIN integration into IoT networks. Authors in [32] focus on the antenna design, challenges, and opportunities for the SAGINs. Work done in [33] presents a comprehensive overview of the challenges and opportunities the integrated satellite-terrestrial networks created in the 6G era. Finally, the work in [19] provides a comprehensive overview of the latest technical advances, challenges, and future directions in aerospace integrated networks (AINs) for empowering 6G, covering system architecture, enabling technologies, modeling, performance analysis, and optimization of AINs.
In addition to the works mentioned above, several surveys are oriented toward the applications of AI in specific layers of the SAGIN architecture. For instance, the work in [11] presents an extensive survey of the AI-based solutions proposed to address the challenges of satellite communications. Another work done in [35] expands the discussion by addressing the AI-enabled satellite-based Non-Terrestrial Networks, its associated challenges, and future directions. Furthermore, the work in [30] describes the main applications of AI in UAV networks and highlights the limitations of the existing works in this era. At the same time, another survey [24] summarizes the main RL-driven applications for efficient multi-UAV wireless networks. The work done in [34] provides an overview of the transformative potential of ML techniques in UAV operations and communications, highlighting crucial domains like UAV perception, feature collection and processing, trajectory planning, aerodynamic control, and operational management. On the other side, some existing surveys focus on specific AI algorithms and their applications in SAGINs. For instance, the authors in [23] explore RL techniques to seamlessly integrate aerial platforms, including satellites, HAPs, and UAVs, with the existing terrestrial networks. Another paper [22] sheds light on the potential role of AI in deploying and optimizing SAGIN-based industrial IoT solutions and provides a comprehensive review of the relevant research works.
Table 1 summarizes previous surveys relevant to this paper’s scope, highlights their main contributions, and compares them with ours. While many works have been published in the cross-section of the AI and SAGIN domains, there is no comprehensive survey discussing advances made by the research community. In this paper, we aim to fill this gap by providing a comprehensive survey of the state-of-the-art research in AI-enabled SAGINs, covering the main potential applications and challenges that AI can address and highlighting the research problems for seamless integration of AI in SAGINs.
B. Contributions and Organization
In this work, we aim to provide a holistic overview of state-of-the-art research on the role of AI in enhancing the performance of SAGINs. We specifically discuss the main challenges of satellites, HAPs, and UAV networks and highlight AI’s advantages in overcoming these limitations. Furthermore, we provide a comprehensive overview of the work done on using AI to optimize the integrated architectures. We finally highlight the main open issues and future trends related to the synergy between AI and SAGINs. A recurring limitation of previous surveys is their narrow focus on specific elements and layers of SAGIN and not accounting for the intrinsic challenges of the integrated architecture. Furthermore, none of these surveys highlight the interplay of SAGINs and AI and present how the SAGINs can contribute to the evolution of AI.
To the best of the authors’ knowledge, this is the first comprehensive survey that provides a systematic review of AI-enabled solutions for different components of SAGIN architecture and highlights the main AI-aided optimization techniques to fully realize the potential of AI in the context of SAGINs, thereby guiding future research efforts. Furthermore, this survey does not focus on AI for SAGIN but also highlights the potential of SAGIN architectures in advancing the development of efficient AI algorithms. This survey is finally distinct due to its comprehensiveness, self-containment, rich content, and key insights, as evidenced by the structure and content presentation in the sequel. To summarize, the main contributions of this survey are
A comprehensive overview of the AI-enabled solutions for satellites, HAPSs, and UAVs, including most papers published in the past few years. Specifically, we review around 200 technical papers: 64 of which consider satellite networks, 18 papers address AI integration in HAPs network, 84 papers focus on AI-enabled UAV systems, and 35 papers cover AI-enabled SAGINs.
A comprehensive discussion on the role of SAGINs in advancing AI algorithms, including key wireless techniques for optimized AI systems and main 6G applications that use AI in the context of SAGINs. Fifty technical papers are reviewed in this section, grouping the main works that characterize the interplay of AI and SAGIN and highlighting the gap in the literature for the full benefit of SAGINs for optimized AI algorithms.
A thorough overview of the main open issues and future trends aiming to highlight the limitations of the current AI-enabled SAGINs and to capture the gap in the literature covering optimizing AI algorithms based on specific SAGIN architectures.
The remainder of this survey is organized as shown in Fig. 2 and includes the following sections. In Section II, we present a brief introduction to AI and its main algorithms for the reader’s convenience. Section III reports the main works that aim at using AI in solving challenges present in satellite networks. In Section IV, we highlight the main research efforts that consider AI for addressing challenges of efficient HAPs deployment and functionality. Section V presents the main works that use AI to address challenges in UAV networks. Section VI discusses how AI can optimize the integration of the space, aerial, and terrestrial subnetworks together toward the full operation of SAGINs. Section VII highlights how the recent advancements in SAGINs influence AI methods and applications. In Section VIII, we outline the remaining issues and potential research directions to harness the full benefits of AI in efficiently implementing SAGINs. Finally, Section IX concludes the paper. The list of notations used throughout the paper is presented in Table 2.
An Overview of Artificial Intelligence
AI has been extensively used to address various issues for many decades. At the same time, it has proved to be a highly complex and difficult-to-define subject. While the fundamentals of AI would likely be known to many readers, this section presents a comprehensive overview for completeness. Researchers have explored and expanded this field since McCarthy et al. [36] coined the term “Artificial Intelligence” in 1955. In a very broad sense, AI is a set of techniques and algorithms designed to teach a computer how to learn - imitating human abilities such as vision, speech, decision-making, problem-solving, analytical thinking, perception, and others through various AI methods to simulate intelligent behavior [37].
AI algorithms can be classified into different categories based on their learning approach. One way to categorize AI algorithms is by their learning type, which includes supervised, unsupervised, and RL algorithms. Another way to classify AI algorithms is based on their methodology, providing ML and DL algorithms. This section briefly discusses these different subsets of AI for the reader’s convenience. Their key differences and examples of algorithms are listed in Table 3 and Fig. 3, respectively.
Different AI techniques and algorithms surveyed in this paper and their respective classification.
A. Classification by Methodology
1) Machine Learning
ML is a branch of AI associated with developing and constructing analytical models that can automatically detect hidden and previously unknown patterns in data and independently acquire the properties necessary to recognize these patterns [38]. The critical enabling factor of ML is the presence of a data set that describes the objects or processes under study and reflects their inherent properties and patterns. Such a dataset is called a training set, which can be obtained as a set of observations created by an expert or analyst based on hypotheses, analogies, personal experience, and possibly intuition.
Observations from the training set, called training examples, are sequentially presented to the trained model, and in the process, it acquires the necessary properties. This process is called model learning; it is an iterative procedure where a data object from the training sample is presented at each step, and the model parameters are adjusted by a rule called the learning algorithm. Training continues until the model can perform the required transformation with sufficient accuracy. Once trained and tested, it can analyze the never-before-seen data. Hence, ML involves algorithms that learn patterns from data without being explicitly programmed with rules for each task. It gives computers new ways to accomplish tasks previously performed by people and teaches a computer system to make correct predictions when data is entered. It accelerates the development of AI’s potential as its most crucial component.
2) Deep Learning
DL is a subset of AI that uses Artificial Neural Networks (ANNs) to learn specific tasks [39]. ANNs are inspired by biological NNs, which can be seen as a possible explanation for ANNs’ flexibility and problem-solving capacity. NN is a complicated program comprising many internal (hidden) layers with configurable parameters - the weight coefficients of artificial neurons comprising each network layer. The network gets a vector of features that characterize the object - data in the form of signals - at the first input layer. They are processed on the inner layers: the input vector is multiplied by the matrix of connections, and the vector of new features created as a result is sent to the next layer. The outcome of signal processing is transferred to the network’s output layer.
Each NN comprises many parameters that cannot be altered manually; therefore, the NN is trained automatically with the help of the data being fed to it. Throughout the training process, the weight coefficients of neurons continually change and are adjusted so that the result of calculations and signal processing becomes meaningful. The weight coefficients are updated using a basic optimization approach based on the gradient descent method, which allows modeling of how the NN’s signal processing output changes with a slight change in each weight coefficient. The NN weights are modified at each layer, which is why it is termed deep. The more sophisticated the NN, the more layers and neurons it has, and the more computational operations it executes, the better (on average) the output it produces. However, it is more difficult to understand what is going on in its underlying layers.
B. Classification by Learning Type
1) Supervised Learning
Supervised learning is a type of AI comprised of algorithms and methods that learn based on a set of examples containing known “input-output” pairs. A supervised learning algorithm must be fed with samples containing a vector of independent variables (most commonly called features) and a vector of learnable values (target output) that the model should try to produce during training. The difference between the target and actual outputs produced by the model is called a residual error. At the heart of each supervised learning algorithm is the residual error minimization process (also known as objective function optimization), which teaches the model with the help of right (small error) and wrong (significant error) examples. This residual error is used to calculate model parameter corrections at each training iteration. Classification and regression are two main types of problems solved with the help of supervised learning (see Table 3). The class label is used as a target variable in the former type, while in the latter, a numeric value is the desired output. Supervised learning has a wide range of applications in various fields. It is commonly used in text and image analysis, object recognition, predictive analysis, and modeling. The main supervised learning algorithms used in the literature are listed in Fig. 3.
2) Unsupervised Learning
Unsupervised learning is a subset of AI that does not use an objective function to correct the parameters of a trained model. In other words, in unsupervised training models, it is not necessary to have predefined model outputs. So the output error of the model on the training set is not calculated. Instead, information about the current state of the model parameters and examples of the training set are used. For instance, the Euclidean distance between the example’s feature vector and the neuron’s weight vector can control the correction of model parameters during training. The main application of unsupervised learning is building models for clustering and association. Since the cluster data structure and the underlying relation between data points are not known in advance but determined during the model’s training, it is impossible to use any target values. The main unsupervised learning algorithms used in the literature are listed in Fig. 3.
3) Reinforcement Learning
RL is a branch of AI that studies the behavior of intelligent agents operating in a particular environment and making decisions [71]. The environment’s response to the decisions made is reinforcement signals, based on which the agent is trained. Therefore, such learning is a particular case of supervised learning, where the teacher is the environment. A reinforcement system is any set of rules that can change the state of a model over time. In this system, the agents aim to maximize their performance in a particular environment. A reward the agent receives signifies that the action contributed to achieving the goal sooner, while a penalty is given when the agent fails to produce the right move. The problem narrows down to choosing a sequence of steps (or actions) necessary to maximize the reward. Each decision brings new information to the agent, leading to the development of new strategies and solutions, which in turn lead to greater rewards. Some techniques focus on maximizing immediate compensation (called “exploitation”), while others tend to prioritize finding as much information about the environment as possible (called “exploration”). The majority of RL algorithms balance these two paradigms.
Similarly to supervised and unsupervised learning, RL is an iterative process. The strategy improves as training progresses, and the agent receives more feedback. The exact way this happens solely depends on the RL algorithm. Furthermore, a list of the main RL algorithms is presented in Fig. 3. The key differences between supervised, unsupervised, and RL algorithms are presented in Table 3.
One of the most prominent RL algorithms for optimizing wireless communications is the Deep Deterministic Policy Gradient (DDPG). DDPG can handle continuous action spaces and high-dimensional state spaces arising from network data, making it a well-suited tool for large-scale heterogeneous network architectures. Its off-policy nature and use of replay buffers enhance sample efficiency, while target networks and soft updates promote stability. DDPG’s flexibility allows tailoring to specific scenarios through customizable reward functions and state-action spaces. DDPG, as well as many other RL algorithms, are adapted to various problems by tailoring the reward function, state space, and action space to the specific scenario.
AI for Satellite Communications
Satellite communications provide many advantages to enhance the user experience, such as extended coverage, good transmission quality, high communication capacity, convenient and rapid networking, and seamless global communications. On the other hand, the financial implications of deploying and maintaining satellite infrastructure require developing robust resource management and time-sharing systems. Several papers in the literature aimed at proposing solutions for the limitations of satellite networks [72]. Recently, AI tools have been considered to deal with the massive data satellite networks generate and to provide efficient solutions to enable fast integration of satellites in the SAGIN architecture. In this section, we aim to survey the primary papers that use tools from AI to address challenges and optimize satellite communications. Namely, we focus on the AI algorithms that tackle challenges such as beam hopping, anti-jamming, traffic forecasting, channel modeling, anomaly detection, interference management, ionospheric scintillation, energy management, and resource management. Table 4 and Table 5 summarize most works in the literature that use AI to address the challenges of satellite communications. Fig. 4 visualizes the main challenges of satellite communications considered in this survey.
Challenges in satellite communications addressed by AI and discussed in this survey.
A. Beam Hopping
Beam hopping emerged as a promising solution to manage diverse traffic demands by enabling satellites to reallocate capacity between beams in response to ground requests [5]. This is accomplished using time-division multiplexing with a single frequency instead of the earlier method of isolating spot beams by “color (frequency divided by polarization). Beam hopping is used to employ flexibility in dealing with unpredictable and time-varying traffic requirements in the satellite coverage region. It mainly uses a small number of active beams simultaneously to illuminate a specific cell based on the traffic demand dynamically.
The main challenge of beam hopping is the choice of the illumination pattern, that is, which and for how long each beam must be selected [73]. Furthermore, the need for high-precision position control of the laser beam in the pointing, acquisition, and tracking system is yet another challenge for efficient beam hopping systems [74]. Another critical factor is the optimization of the system parameters, such as the number of serving beams, transmit power, beam directions, and sizes [75]. Several works in the literature attempt to propose solutions for the challenges of beam hopping, and most methods have relied on classical optimization algorithms. For instance, the authors in [76] aim at minimizing the number of satellite beam positions subject to a predefined requirement on the radius of a beam. In [77], the authors discuss cognitive techniques for enhancing spectral efficiency in satellite communications. In [74], the authors research adaptive control algorithms for jitter in laser beam pointing and tracking systems.
While optimizing algorithms yields favorable outcomes in enhancing flexibility and reducing delays in beam hopping systems [78], certain challenges persist. As the number of beams increases, there is a significant expansion in the search space, making it inherently challenging to devise the optimal parameters amid numerous local optima. This issue becomes particularly pronounced in the case of complex satellite systems employing extensive beamforming capabilities, with hundreds or even thousands of individual beams, as computational demands of conventional optimization algorithms can quickly escalate, exceeding acceptable time constraints and hindering real-world implementation. Furthermore, classical optimization algorithms necessitate adjustments when there are moderate changes in the scenario, increasing the computational complexity. This complexity becomes impractical, especially when managing resources onboard, where adaptability is crucial.
Several recent works leverage tools from AI and ML to enhance the performance of beam hopping schemes and overcome the limitations of traditional optimization techniques. For instance, the authors in [79] apply a DL technique to optimize the beam hopping illumination pattern. A Fully connected neural network (FC-NN)-based method is proposed to create a synergy of optimization and learning. The significant advantage of the technique is in the computation time, which is substantially low in contrast to other iterative methods. Furthermore, the authors use the cardinality of beam hopping patterns as a feature to achieve improved performance and accuracy. A similar algorithm for finding the efficient beam hopping illumination pattern design is presented in [80]. Having identified the proper features necessary for good predictions, namely asymmetric traffic demands, the authors propose a learning-and-optimization algorithm combining DL and optimization benefits. Experimental validation shows that FC-NN can efficiently limit the search space of optimization problems in case studies involving multibeam-defined GEO satellites.
RL is yet another AI tool frequently used to improve the throughput of beam-hopping systems in satellite networks. In [81], researchers focus on multibeam satellite systems and aim to establish the traffic model of forward links. Thus, a deep reinforcement learning (DRL)-based algorithm is proposed to optimize the allocation of resources for beam hopping. The results show that the proposed method can reduce the transmission delay and increase the throughput. Similar results can be achieved by predicting long-term utilization rates. For instance, the authors in [82] propose a DRL-based approach that considers traffic demands in spatial and temporal domains, antenna radiation patterns, and interference. The study reveals notable reductions in transmission delay (up to 52.2%) and increases in system throughput (up to 11.4%) compared to existing methodologies, demonstrating the efficacy of the proposed DRL-based method.
An overview of different beam hopping algorithms in LEO satellite constellation networks is presented in [83]. Simulation results compare the presented algorithms in terms of performance and complexity metrics, concluding that DRL-based algorithms are the fastest to deliver results comparable with hybrid simulated annealing/particle swarm optimization (PSO) and genetic algorithm (GA)-based methods. The only problem identified is the poor robustness of the DRL models, which can be solved by pairing them with GAs. In this case, the GA, acting as an evolution algorithm, can optimize the hyperparameters and architecture of the DRL models, thereby enhancing their overall performance. Another study involving GAs is presented in [84], in which researchers show that a DRL algorithm can solve the problem of the “curse of dimensionality” in satellite communications. The authors note, however, that while the multi-objective DRL method can increase the throughput, maintain the fairness of an individual cell, and decrease the delay, the GA can achieve comparable results but with 110 times less complexity.
Another DRL-based approach called double-loop learning is adopted in [85] to ensure both delay reduction and throughput maximization while maintaining the fairness of beam hopping selection for each cell. The simulations show that the suggested technique may pursue many objectives at the same time and intelligently allocate resources based on user requirements and channel characteristics. The work in [86] focuses on achieving real-time beam pattern illumination and bandwidth allocation to satisfy non-uniform and time-varying traffic requests. In this case, deep Q-learning is adopted to produce high generalization as traffic demand grows.
B. Anti-Jamming
Due to the inherent nature of broadcasting and its immense coverage area, satellite communications are particularly prone to security threats. The security problem has been addressed in the literature through cryptographic protocols, information-theoretic approaches, and physical layer security techniques [142], [143]. Satellite jamming is one of the critical attacks on satellite networks that disrupts signals sent and received by satellites by generating noise at the same radio frequency as used by the antennas [10]. Jamming attacks are mainly initiated within a satellite network to diminish the data throughput and deteriorate the user experience. Thus, efficient anti-jamming techniques should be designed to prevent jamming and maintain secure satellite communications.
Several papers focus on reducing jamming attacks by developing effective anti-jamming solutions that rely on classical methods [10], [144], [145], [146]. For instance, the authors in [146] investigate using satellite diversity to overcome the effect of jamming. However, many classical anti-jamming methods cannot counter smart jamming attacks that can adapt and evolve through interaction and learning. AI is, therefore, a mandatory tool to provide intelligent and adaptive solutions for detecting and mitigating jamming attacks in real time while improving interference mitigation. Several papers consider using different AI algorithms to optimize anti-jamming methods in satellite networks. For instance, the synchronization of the frequency hopping signal technique, which is frequently used for its anti-jamming properties, can be achieved with the help of a long short-term memory (LSTM) network, as described in [87]. The proposed method for synchronizing frequency hopping signals significantly reduces synchronization time, thus enhancing anti-jamming abilities. Furthermore, an artificial bee colony (ABC) is proposed in [93] to solve the blind separation problem between jamming and communication signals and find the sub-optimal solution.
Among the different AI algorithms, RL has proven its efficiency in learning the optimal communication policy without having prior knowledge of the jamming conditions and the radio channel model in a dynamic environment. In [92], the authors present a multi-step prediction Markov decision process (MDP) and build up a multi-step prediction Bellman iterative equation to defend against sweep jamming attacks while considering the inherent transmission latency of an actual communication system. Probabilities of state transition are calculated to maximize the communication system’s utility. The optimal anti-jamming attack technique is investigated under the premise that the jammer can learn the strategy of the communication system. A combination of convolutional neural network (CNN)-based jamming pattern recognition and Q-learning-based online channel selection is used in [89] for effective anti-jamming communication. The authors conclude that introducing the channel switching cost improves the anti-jamming algorithm while capturing the trade-off between throughput and communication overhead. Another work [94] utilizes the same RL-type Q-learning technique to choose the channel in the high-intensity jamming scenario. At the same time, a deep Q-network (DQN) determines the signal route from source to destination.
In addition to RL, DRL is used repeatedly to address the anti-jamming challenge. For instance, the work in [90] combines game theory with DRL to learn the anti-jamming policy in a dynamic Internet of satellite environment. Specifically, a Stackelberg game models the interactions between the jammers and the satellites. A DRL-based routing algorithm is used to solve the routing selection issue while preserving an available routing subset, and another Q-learning-based algorithm is used to adapt the anti-jamming strategy based on the selected routing subset. Simulations demonstrate that the proposed method has a lower routing cost and greater performance than existing algorithms and approaches the Stackelberg equilibrium. A similar work in [91] aims to reduce energy usage in a jamming scenario. The distributed dynamic anti-jamming system consists of a hierarchical Stackelberg game, a coalition formation game, and a Q-learning algorithm to obtain the anti-jamming policy. In [88], the authors tackle the anti-jamming problem from the point of view of the jammer. They investigate scenarios where DRL-based users with varying communication modes combat DRL-based jammers using different jamming methods. The simulation results suggest that the proposed DRL-based jamming may effectively limit the performance of the DRL-based anti-jamming.
C. Traffic Forecasting
Traffic forecasting involves predicting future traffic patterns and demands in satellite networks to ensure efficient resource allocation, congestion control, and network planning. Existing terrestrial traffic forecasting models suffer from several problems, such as high computational complexity, which makes them unsuitable for satellites with limited onboard computing resources [147]. In the literature, several works attempt to perform traffic forecasting in satellite networks using classical methods. For instance, a traffic prediction-based dynamic routing technique is proposed in [148] for LEO/GEO satellite networks and evaluated regarding the end-to-end delay and the packet loss rate. As satellite traffic is self-similar and demonstrates long-range dependence, the proposed traffic forecasting models must be more adaptive and achieve sufficient accuracy. AI can, therefore, be helpful because it can reduce the complexity and provide intelligent and automated solutions for accurate and efficient traffic forecasting in satellite communications. AI techniques, such as ML algorithms, can analyze historical traffic data, identify patterns, and predict future traffic demands, enabling better resource allocation and network planning [149].
Supervised learning is widely used in performing traffic forecasting in satellite networks. In [100], a new hybridization of extreme learning machine and multitask beetle antenna search (MBAS-ELM) algorithm-based distributed routing is developed for LEO satellite networks. Practical and trustworthy routing for LEO satellite networks is challenging because of changing topology, connection modifications, and uneven communication load. The suggested model selects routes based on traffic forecasts concerning the level of traffic circulation on the Earth. The results are tested using different simulation times and data sensing rates. The collected results demonstrate that the suggested MBAS-ELM model outperforms previous techniques.
A combined forecasting model for satellite network traffic based on the extreme learning machine (ELM) is proposed in [97]. This algorithm aims to empirically divide long-range-dependent self-similar satellite network traffic into several short-range dependent components. This allows for faster and more precise forecasting with reduced complexity. A similar work relying on ELM can be found in [96]. The work in [98] proposes a two-stage approach using XG Boost and recurrent neural network (RNN) algorithms. This work compares the performance of these algorithms, concluding that applying either one at default settings shows an improvement up to 9.5% for XG Boost and 9.2% for RNNs. In [95], short-term traffic loads are predicted using principal component analysis (PCA) and a generalized regression NN. The presented results demonstrate that the proposed method outperforms previous state-of-the-art algorithms regarding forecasting accuracy, training time, and robustness, making it the go-to approach for real-time traffic forecasting in satellite networks.
The authors of [102] combine a transfer learning-based gated recurrent unit (GRU) NN and particle filter online training algorithm to predict traffic faster and more accurately with insufficient online traffic data. The work done in [99] depicts satellite network traffic’s spatial and temporal features using a graph convolutional network (GCN) and GRU model with improved forecast accuracy. Apart from supervised learning, RL once again proves its usability in performance optimization tasks. For instance, the work in [101] introduces a DRL-based framework to improve the performance of satellite network routing algorithms. The fully distributed framework is created to mitigate the issues that emerge due to centralized training and execution found in other methods.
D. Channel Modeling and Estimation
Accurate channel models are essential for evaluating the performance of satellite systems and, in turn, improving coverage for existing deployments. Channel models can also predict signal propagation within planned deployment scenarios, enabling efficient planning and assessment before actual deployment. Channel modeling has been traditionally done with ray tracing simulations and extensive outdoor experiments to estimate the values of the channel parameters. However, with the need for real experiments in satellite networks and the difficulty of applying ray tracing, new tools are required for efficient channel modeling and estimation. Furthermore, with the integration of massive multiple-input multiple-output (MIMO) techniques in satellite systems, obtaining effective channel state information (CSI) to construct efficient hybrid beamforming mechanisms is extremely challenging due to the dynamic nature of these systems, long delay times, and low payloads [150]. Thus, several works in the literature propose schemes for efficient satellite channel modeling and estimation without relying on AI. For example, using ray tracing simulations, the authors in [150] characterize the satellite-terrestrial channel in a high-speed railway environment. In [151], the authors investigate the performance of satellite communication systems using the space shift keying modulation technique over shadowed Rician land mobile satellite links. In [75], a traffic simulator is developed for multibeam satellite communication systems to address the challenge of setting optimal system parameters.
AI techniques, on the other side, can process real-time environmental information, predict channel conditions, and optimize modulation and coding schemes to maximize system throughput [152]. AI can also enhance the accuracy and efficiency of channel modeling, leading to improved performance and reliability in satellite communications. Early attempts to use AI for channel modeling and estimation in terrestrial networks rely on traditional ML algorithms such as support vector machines (SVM) [153] and decision trees [154]. However, DL algorithms are more effective for channel modeling and estimation in satellite networks. For instance, using a deep CNN, the authors in [103] estimate channel characteristics (path loss exponent and standard deviation of shadowing) from 2D satellite photos. In this work, the authors present a computationally efficient and trustworthy alternative to ray tracing simulations, with experimental findings demonstrating that at 900 MHz, the prediction accuracy is equal to or higher than 76%. The work done in [104] presents a similar approach for predicting path loss distributions from 2D satellite pictures. The authors conclude that accurate path loss prediction can be achieved in real time and for various frequencies and heights.
Since establishing efficient instantaneous CSI is challenging due to the changing environment and substantial transmission delays in satellite networks, the authors in [105] offer a DL-based prediction technique to address this issue using correlations in different channels via a collection of LSTM units. The predictor is trained offline before being utilized online to extract channel information and forecast future CSI in LEO satellite settings. The findings indicate that the proposed solution efficiently alleviates channel aging issues in massive MIMO systems of LEO satellites. In [106], authors propose using deep neural networks (DNNs) to realize downlink CSI acquirement and hybrid beamforming design. The proposed schemes can predict future downlink CSI from observed uplink CSI without estimating and generate beamformers from predicted downlink CSI without complex optimization. The findings indicate that the methods offered can successfully support LEO systems with the massive MIMO technology.
E. Anomaly Detection
Due to the harsh space environment and the exposition to heat, vacuum, and radiation, satellite communications are vulnerable to failures and disconnections. Thus, anomaly detection via telemetry mining, which involves detecting outliers or anomalies in satellite time series data to identify signal distortion, attitude instability, adjacent objects’ proximity, and equipment failure, is essential for fault diagnosis and maintaining reliable satellite connections. Due to the vast amount of telemetry data, noise and measurement errors, and the complex and high-dimensional datasets, detecting anomalies via manual inspection is not feasible [155], [156]. Several papers in the literature address these problems in satellite telemetry mining without relying on AI techniques. For example, the authors in [156] propose a cluster-based method for anomaly detection in satellite telemetry data using an extended dominant sets clustering algorithm. This paper provides valuable insights into the challenges and potential solutions for telemetry mining in satellite communications.
However, AI techniques can effectively handle large volumes of data and complex patterns in satellite telemetry data. AI techniques can also improve the accuracy and efficiency of anomaly detection by learning from historical data and adapting to changing patterns in the telemetry data [11]. The work in [107] is one of the early works promoting ML use for LEO satellite telemetry mining. Here, the performance of various ML methods such as autoregressive integrated moving average (ARIMA), multilayer perceptron, RNN, deep LSTM RNN, and deep GRU RNN are compared in predicting metrics such as battery temperature, power bus voltage, and load current from spacecraft telemetry data. The results show that ARIMA outperforms the other techniques and provides higher accuracy regarding the root mean square error (RMSE) and the mean absolute error (MAE).
Another approach for anomaly detection is proposed in [109]. Tensor-based anomaly detection (TAD) is a recent direction in detecting and identifying the corruption in the telemetry data of a satellite. Standard spectral-based methods like PCA can detect anomalies, but ML-based TAD methods, such as SVM and NNs, are shown to provide better results. This approach is reasonable as accumulated data in satellites usually have a tensor structure, e.g., space-time measurements. Orbit prediction of various resident space objects (RSOs) is crucial for avoiding satellite collisions, according to [110]. The results show that the SVM model can enhance the orbit position estimation in most situations.
Telemetry data compression is often overlooked, with anomaly forecasting being paid more attention to. However, the authors in [108] describe a self-learning technique called classification probability calculation - window step optimization for obtaining class characteristics and deciding on individual parameters compression. Simulation results show that the algorithm correctly classifies simulation and real mission data into the appropriate base class, significantly reducing data and computational complexity. Work done in [112] employs a two-stage lossless algorithm based on LSTM to achieve the same goal. LSTM coupled with RNN can be a basis of a similar two-stage lossless data compression algorithm, as presented in [111].
F. Ionospheric Scintillation
One of the main challenges that hinder the efficient use and implementation of satellite communications is ionospheric scintillation. Ionospheric scintillation is characterized by rapid fading or distortion of radio signals as they pass through the ionosphere [157]. This can occur due to changes in the density of electrons in the atmosphere, affecting radio waves’ transmission. In satellite communications, ionospheric scintillation can lead to signal dropouts and degraded performance of communication systems, resulting in reduced data throughput and reliability [157]. This poses a significant challenge for satellite-based services such as TV broadcasting and mobile phone networks, as service disruptions can occur. Several papers address the problems associated with ionospheric scintillation in satellite communications. For instance, the work in [157] analyzes the correlation between the rate of total electron content index (ROTI) and scintillation indices, while the work in [158] develops a visual tool for exploring ionospheric scintillation monitoring data. In [159], the authors propose a low-cost GPS-based sensor for detecting ionospheric irregularities.
To assess the impact of the scintillation on the signals, researchers rely on simple techniques involving wavelet methods [160], analysis of statistical properties with histograms [161], and adaptive frequency-time techniques [162]. However, using these simple methods to evaluate the scintillation impact is only sometimes accurate due to its complexity. Additionally, it can be challenging to distinguish signal distortions caused by other factors like multi-path interference. Recently, AI has been used for scintillation detection in satellite communications thanks to its appealing features. AI can provide advanced signal processing techniques, adaptive/robust methods, and parameter estimation to mitigate the effects of ionospheric scintillation on GNSS receivers [163]. It can also contribute to developing models for describing ionospheric scintillation’s effects on GPS receivers, improving the estimation of tracking loop error [164].
One of the earlier works that applies an ML algorithm to predict the ionospheric scintillation is presented in [113]. A decision tree-based algorithm indicates the degree of scintillation despite the significant fluctuation of the ionospheric conditions that influence the emergence of such anomalies. Similar work is done in [116], in which experiment findings suggest that a decision tree-based strategy can outperform traditional approaches, achieving 98% detection accuracy and faster processing, allowing for early scintillation detection. A variation of the decision tree-based algorithm, namely the bagged decision tree, is proposed in [117] to differentiate multipath and ionospheric scintillation in monitoring data. The model classifies the data as scintillated, multipath impacted, or clean GNSS signal with 96% accuracy. Authors in [114] conclude that an SVM-based amplitude detector is enough at low latitudes, but a phase-scintillation detector is required at high latitudes. A similar approach is adopted in [115] where SVM and Gaussian SVM solve the frequency domain ionospheric scintillation detection problem.
G. Interference Management
Interference management is a broad issue affecting most wireless networks’ efficiency and has traditionally been addressed using signal processing techniques [77]. However, these techniques have limitations in managing interference at a large scale. One problem associated with interference management in satellite communications is the increasing demand for high-speed data rates for satellite multimedia and broadcasting services, coupled with spectrum scarcity in satellite bands [77]. This challenge has led to the exploration of new techniques for enhancing spectral efficiency in satellite communication. For example, cognitive communication is proposed as a promising solution [77]. Cognitive techniques such as underlay, overlay, interweave, and database-related methods have been studied to improve the efficiency of satellite communication systems [165]. While several papers in the literature aim to mitigate the impact of interference in satellite communications, AI techniques have been proven to enhance the effectiveness of interference management algorithms in large-scale satellite networks [77].
In [118], the authors propose using DL algorithms, namely LSTM, to mitigate the impact of interference in satellite networks. The work mainly illustrates interference detection and classification performance with interference with different power levels. Auto-encoding techniques are used to decide whether interference is present or not. The results show high accuracy values of the proposed algorithms in low signal-to-interference ratio regimes where interference is more substantial. Researchers also use the LSTM algorithm in [119] to detect in real time both short-term and long-term interference in the spectrum of the signal received from the satellite. The proposed method generally applies across a discrete collection of various signal spectra and can pinpoint interference in time and frequency. The work in [120] proposes a DRL-based framework that exploits software-defined networks and AI to manage spectrum in integrated satellite and terrestrial networks. Spectrum sharing becomes especially relevant to accommodate a broadcasting network within a constrained range. Even though there are some methods, such as cognitive radio, to enable the dynamic sharing of the spectrum between satellite and terrestrial networks, the proposed DRL-based spectrum management techniques show promising results that outperform traditional interference management techniques.
DRL methods for satellite interference management are further explored in [6]. This work explores the challenge of allocating resources in LEO satellite networks during dynamic inter-satellite interference. The objective is to optimize throughput while effectively handling interference. The authors suggest employing learning-based frequency allocation strategies, incorporating upper confidence bound (UCB) variants and DRL techniques. Through simulations with various reward and constraint configurations, the findings indicate that DQN outperforms other techniques, particularly in scenarios where prior knowledge of interfering satellites is lacking. Another study [121] uses a multi-agent DRL algorithm designed for interference-conscious channel allocation. This algorithm addresses the challenge of frequency sharing between uplink satellite networks and terrestrial services. The primary goal is to identify the most favorable channel index set, minimize interference to the victim system, and meet the network uplink’s signal-to-interference-and-noise ratio (SINR) requirements of the network uplink. The results indicate that the proposed DQN algorithm surpasses the performance of the existing graph coloring algorithm, demonstrating its effectiveness in single-satellite and multi-satellite non-terrestrial network scenarios.
H. Resource Management
Resource management involves allocating and scheduling resources such as bandwidth, power, frequency spectrum, transponders, and antennas to ensure efficient and effective utilization of available resources [166]. In addition to improving overall performance by allowing better control over network quality metrics such as latency and throughput, resource management and allocation enable satellite systems to maximize their efficiency by reducing interference. Additionally, it allows for more reliable communication links essential for mission-critical applications such as military operations or emergency response services. Resource management also plays a vital role in helping to reduce costs associated with operating a satellite system by ensuring that all resources are used efficiently [166].
Several specific problems are associated with resource management in satellite communications, and many papers in the literature propose solutions. The majority of the proposed solutions rely on complex traditional optimization techniques. For instance, the authors in [167] propose a demand and interference-aware adaptive resource management approach for GEO high throughput satellite systems. They formulate a multi-objective optimization problem to minimize power consumption and system bandwidth usage while matching the offered capacity with the demand per beam. Another paper [168] investigates the resource management problem for virtual network function placement in a decentralized LEO satellite network. It proposes a decentralized approach where each satellite manages neighboring sub-network resources and provides computing services for satellite applications. In [169], an integer linear programming problem is formulated and solved to efficiently allocate the resources of software-defined LEO satellite networks to provide services for terrestrial users.
With recent AI advancements, satellite communications can be optimized for resource allocation to improve efficiency and reduce costs. Leveraging AI addresses the rising demand for high data rates and flexible radio resource assignment while enabling enhanced spectral efficiencies and cost reductions per bit [166]. One of the main problems related to efficient resource management is dynamic channel allocation (DCA), which aims to maximize spectrum utilization. Unlike existing works on DCA, which ignore the intrinsic temporal correlation among the sequential channel allocation decisions, the research done in [122] proposes a DRL-based algorithm to solve the problem. It models the DCA optimization problem as an MDP. A CNN further extracts useful features from the image-like fashion reformulated system state. Simulation findings suggest that the proposed method may reduce blocking probability while increasing carried traffic and spectrum efficiency. Researchers in [126] also tackle spectrum efficiency and meeting traffic demands challenges in satellite communications. However, their solution involves a two-stage algorithm incorporating self-supervised learning and DRL to solve spectrum congestion and power consumption issues. Spectrum efficiency is studied in another work [123], in which a cooperative multi-agent DRL framework is employed in the bandwidth allocation problem. The experimental findings demonstrate that this strategy can improve transmission efficiency while remaining flexible.
A Q-learning-based dynamic distributed routing system is presented in [125] to achieve minimal end-to-end delay and low network traffic overhead burden. The experimental findings show that the scheme can discover the initial routing strategy and provide long-term quality of service (QoS) optimization. Another Q-learning-based approach is adopted in [124] to solve the capacity management challenge in a three-layer heterogeneous satellite network comprising GEO, MEO, and LEO satellites. The paper presents a low-complexity approach for immediate and long-term optimum capacity allocation to maximize system utility. Another work [127] explores the application of an enhanced DRL algorithm, namely twin-delayed DDPG, for joint sub-channel and power allocation within multibeam GEO satellite communication systems. Simulation results reveal that the proposed approach substantially improves compared to baseline schemes.
A different angle on the resource management problem in satellite communications is adopted in [128]. The authors suggest using satellite networks for offloading and computation services in intelligent transportation systems (ITS). The proposed design leverages a network of distributed MEC nodes comprising LEO and cube satellites. By employing a collaborative multi-agent proximal policy optimization (PPO) DRL framework with an attention mechanism, this architecture facilitates intelligent offloading, computation, and bandwidth allocation decisions optimized for dynamic network conditions. Comprehensive simulations demonstrate enhanced performance compared to baseline scenarios.
I. Handover Management
To maintain rapid and uninterrupted global connectivity, effective management of mobility aspects in satellite networks, such as satellite handovers, is crucial [170]. The impact of handovers in satellite networks is more prominent than in terrestrial networks due to the high mobility of satellites and their continuous orbiting around the Earth. Several solutions are proposed in the literature to optimize handover strategies in satellite networks and provide seamless service to users. For instance, in many prior research works, handover decisions are typically made using one or more predefined criteria. Specifically, the elevation angle, remaining service time, and the number of available channels are mainly considered [171], [172]. These criteria are linked to signal strength, handover frequency, and satellite workload [173]. The authors in [174] propose a seamless handover mechanism based on software-defined satellite networking and conduct physical layer simulations to evaluate its performance. Leveraging the competitive dynamics of potential games, [175] proposes a novel handover strategy for LEO satellite networks. This strategy promotes an equilibrium where each satellite handles an appropriate share of the workload, optimizing overall network performance. Another paper [176] proposes a digital twin-assisted storage strategy for satellite-terrestrial networks to address the problems of inconsistent service and reduced link utilization during satellite handover. However, these conventional methods fail to achieve comprehensive optimization, given the emergence of larger and faster satellite constellations. Consequently, there remains a need for novel AI-driven handover strategies to ensure uninterrupted connectivity, minimize interruptions, and improve the QoS and user experience.
One of the most frequently utilized techniques is Q-learning [129], which falls under the category of RL. For instance, the research presented in [129] offers a multi-agent Q-learning satellite handover approach to decrease average satellite handovers while meeting each satellite’s load limitation. In the presented RL algorithm, specific attributes of the user are designated as its state, and the handover process is defined as its action. Subsequently, the user can autonomously determine whether to execute a handover based on its state. Simulation results reveal that the suggested technique beats local handover solutions in average satellite handover and user blocking rate. Another work [132] treats load balancing and handover strategy problems jointly by incorporating the load coefficient as one of the handover choice factors. Double DQN achieves this aim and performs better in the handover decision, with mobile user access becoming more balanced. Researchers in [133] explore efficiently handling handover events within 5G network infrastructures. In their work, satellites employ a combinatorial multi-armed bandit approach to devise a resource allocation game capable of making decisions over time amid uncertain conditions. This approach enables the dynamic allocation of resources, considering interference with other channels. The simulations show a linear trend in running time and attaining a sub-optimal solution within approximately 20-30 rounds.
Apart from RL, an alternative method involves employing DNN for handover optimization. For instance, the work in [131] investigates simplifying the handover decision problem to a classification problem based on individual users’ SINR changes in 5G networks. The resulting classification problem is then solved using DNN, yielding an algorithm that outperforms traditional methods in terms of radio connection failure and ping-pong rates. A different work focuses on minimizing the average number of handovers while ensuring the signal strength [130]. Using a CNN, it models the handover process as a directed graph and determines the underlying regularity of different users’ optimal handover tactics. Numerical simulation demonstrates that the suggested handover approach may efficiently minimize the number of handovers while maintaining signal strength.
J. Energy Management
Energy management strategies are critical in satellite communications due to the limited capacity of onboard batteries. They are meant to maximize power availability for mission operations, extend the system’s lifetime, reduce operational costs, ensure reliable operation, and improve overall system efficiency [177]. Several classical solutions are proposed in the literature to mitigate the energy constraint and efficiently allocate resources in satellite systems [178], [179], [180]. An interesting approach is presented in [181] and focuses on satellite communication networks where satellite terminals are equipped with energy harvesting devices. In this work, the authors propose a distributed random access scheme considering energy constraints to maximize the average long-term network throughput. The paper adopts a game-theoretic method to approximate the optimal solution and proves the existence and uniqueness of the symmetric Nash equilibrium.
AI algorithms, on the other hand, offer advanced optimization algorithms that can be used for various energy management solutions. For instance, the work presented in [134] describes a DRL-based technique for power allocation in multibeam satellite systems. The suggested architecture models the problem as continuous state and action spaces, and the PPO method is utilized to optimize the allocation policy for the least amount of unmet system demand and power usage. Primarily focused on extending battery life, [135] employs a Q-learning algorithm to dynamically redistribute tasks among overloaded satellites and those experiencing lighter workloads within their vicinity. In [136], the authors present a DQN-based approach coupled with a sliding block scheme that simplifies the modeling of dynamic features of the LEO satellites to achieve 67.86% power consumption savings compared to conventional methods. The DQN-based technique is also used in [137] to devise an integrated method for user association and offloading decisions within MEC-augmented satellite networks. This method incorporates an optimal resource allocation mechanism, enabling informed joint decisions on user connectivity and computational task placement for enhanced network performance. The proposed methodology is shown to offer better long-term latency and energy costs. The DQN is particularly useful in solving the energy consumption optimization problem in satellite communications due to its model-free nature, ability to handle sequential decision-making, approximation of Q-values with NNs, experience replay, trade-off between exploration and exploitation, continuous learning and adaptation, and scalability.
Another work [138] explores the domain of energy-constrained online scheduling within satellite-terrestrial integrated networks. This study introduces a collaborative task scheduling and resource allocation data-driven bandit optimization-based strategy designed to minimize task latency. This scheme incorporates elements of online learning, online control, and offline historical information. Comparative analysis against baseline schemes demonstrates its superior regret value and energy consumption performance. The challenges of edge caching and energy conservation in non-geostationary orbit (NGSO) satellite constellations are addressed in [139], where the authors aim to formulate an efficient data sharing strategy among NGSO satellites to prevent the loss of cached data advantages caused by device mobility. The study utilizes a spreading dynamics model to establish the relationship between satellite speed, constellation structure, data size, and hit probability, determining a decisive threshold for effective data density. The proposed optimization strategy, incorporating improvements to water-filling and RL algorithms, effectively reduces energy consumption and service delay compared to traditional strategies in various scenarios.
Focusing on integrated terrestrial-satellite networks employing non-orthogonal multiple access (NOMA) protocols, another work [141] proposes a novel approach for joint resource allocation and cache design. This approach leverages a multi-agent DDPG algorithm to optimize resource utilization and network content delivery efficiency. The offloading strategy for energy consumption optimization is presented in [140]. The presented algorithm incorporates collaborative computing among satellites, enabling users to delegate tasks and alleviate local computing demands. Additionally, DDPG-based algorithms are cited for enhancing offloading decisions and optimizing resource allocation. This suggested method enhances rewards and diminishes energy consumption compared to conventional optimization techniques.
K. Summary and Lessons Learnt
Several AI methods are commonly employed in optimizing satellite communications, with DL and DRL techniques standing out prominently. Specifically, beam hopping systems frequently use FC-NN and DRL for efficient illumination pattern design and resource allocation. These AI techniques offer advantages in terms of computation time and adaptability, addressing the challenges posed by the increasing complexity of beam hopping as the number of beams grows. Additionally, anti-jamming techniques benefit significantly from DRL, providing intelligent and adaptive solutions to counter smart jamming attacks in real time. The combination of game theory, RL, and DRL-based approaches proves effective in optimizing anti-jamming policies, showcasing improved performance and adaptability in dynamic environments. DRL algorithms also take the spotlight in resource and energy management and handover strategy problems due to their ability to autonomously find the optimum solution based on only the data received from interacting with the environment. DL and ML algorithms are most effective in channel modeling and estimation, particularly for predicting channel conditions, optimizing modulation and coding schemes, and anomaly detection for satellite telemetry data, which is essential for fault diagnosis. TAD, SVM models, and ANNs showcase the effectiveness of AI in telemetry mining, providing efficient solutions for processing large volumes of telemetry data and detecting anomalies challenging for traditional methods. DL methods have also proved to be more effective in ionospheric scintillation and interference management challenges due to the availability of ground-truth data.
The prevalence of DL, DRL, and ML in addressing specific challenges in satellite communications is attributed to their adaptability, efficiency, and improved performance compared to traditional methods. These AI techniques offer solutions beyond classical optimization algorithms, providing the necessary intelligence and adaptability to effectively tackle satellite communication systems’ complexities and dynamic nature. Other AI methods not explicitly mentioned in this section that can be useful include GAs and swarm intelligence (SI), which have shown effectiveness in various communication-related optimization problems and could contribute to addressing challenges in satellite communications. Researchers might also succeed in filling the research gap by applying the existing methods to other challenges reflecting the complexity and dynamic nature of satellite communications. The list might include security concerns, satellite maintenance optimization, and mitigation of environmental factors.
AI for High Altitude Platforms Communications
Innovations in technology have fueled the advancement of HAPs systems, aiming to enhance broadband communication accessibility. HAPSs operate in the stratosphere and can cover a large area or supplement existing broadband services. Beyond data collection, HAPs systems offer exciting possibilities for collaborative computing and distributed ML. Their interconnected nature and high altitude provide a unique platform for decentralized intelligence, where processing power and learning algorithms can be distributed across the network to analyze data in real time and facilitate collaborative decision-making. However, challenges of HAPs communications, such as deployment and constellation management, mobility and energy constraints, resource allocation, and security considerations, remain key barriers to fully exploiting HAPs potential. In this section, we survey existing papers that utilize AI to efficiently address the challenges in deploying HAPs-assisted networks. This includes HAPs placement and topology management, resource management, caching, and computation offloading. Table 6 presents a summary of all works utilizing AI for addressing the challenges of HAPs communications, while Fig. 5 visualizes the challenges of HAPs communications highlighted in this survey.
A. Placement and Constellation Management
A significant obstacle in deploying HAPs systems lies in developing robust and efficient self-organizing network architectures. Conventional self-organizing networks incorporate key functionalities like self-configuration, optimization, and self-healing capabilities. These become crucial in aerial networks, given their higher dynamism compared to fixed cellular networks; in the former, elements’ positions might shift over time due to various factors such as alterations in user needs, atmospheric variations, coverage demands, battery status, or sudden shifts in network traffic. Thus, managing HAPs constellations has always been indispensable to ensure reliable communication by optimizing coverage, minimizing interference, and maximizing system capacity [200]. For instance, the work in [201] presents a geometry-based HAPs channel model for optimizing HAPs deployment, focusing on Line-of-sight (LoS) transmission probability and path loss. The study also provides an algorithm for maximizing deployment efficiency that has been validated through simulations. Another study [202] introduces a new wireless scheme that integrates satellite, airborne, and terrestrial networks to improve user throughput. The focus is on resource allocation and HAPs placement optimization through a two-stage approach, which includes approximated and low-complexity solutions. The study also demonstrates its advantages through numerical results. A layered architecture integrating HAPs at different altitudes into cellular networks is presented in another work [203]. It emphasizes using self-organizing network features for optimal HAPs placement to enhance coverage and capacity, with initial simulations indicating improved service for users.
Thanks to the advantages of AI techniques, they have been regarded as key enablers in designing and optimizing HAPs constellations and providing self-organizing and self-healing capabilities. AI provides significant advantages over traditional optimization approaches due to its ability to continuously adapt and improve based on real-time data and complex environmental factors. This allows for dynamic decision-making that considers changing weather conditions, traffic patterns, and network demands, ultimately leading to more efficient and resilient network operations - a challenge for static rule-based classical methods to achieve. Several papers in the literature attempt to optimize HAPs placement and constellation by relying on different AI algorithms. For instance, the performance of RL and SI algorithms in coordinating a swarm of HAPs for communication area coverage is compared in [183]. Due to its simple rules-based logic, the SI method demonstrates higher convergence speed and stable coverage.
On the other hand, the RL method achieves greater peak user coverage figures by using a dynamic epsilon-greedy approach and a decreasing learning rate. However, it causes coverage gaps due to the exploration strategy of each HAPs. Since RL-based approaches do not depend on feedback loops and cross-agents, they exhibit inherent coordination resilience. Therefore, it can be concluded that SI-based techniques are more efficient and reliable for building coordination algorithms but provide less ideal coverage outcomes, whereas RL algorithms produce superior peaks in coverage. The authors of a similar study [185] examine the efficiency of RL and SI in coordinating multiple unmanned HAPSs. The research builds upon existing work on both algorithms, intending to address the challenge posed by the continuous state space through partitioning this high-dimensional space. The findings indicate that SI continues to outperform RL across important performance metrics like mean overall user coverage and convergence rates despite RL demonstrating higher average peak user coverage. Nevertheless, its unpredictable coverage dip negates this advantage, rendering SI more appropriate.
In [182], an artificial immune algorithm enhances network capacity while optimizing a HAPs network constellation. The study assesses QoS limitations alongside user demand metrics, like signal-to-noise ratio (SNR), Bit Error Rate (BER), and bits per second coverage. In another scenario, NNs are utilized to mitigate the issue of frequent hand-offs that users at the cell boundary may face due to HAPs mobility. The proposed HAPs constellation aims to efficiently balance cost considerations while meeting end-user demands and ensuring QoS through isolation, rate integrity, and information availability. This particular constellation pattern can be leveraged for designing multiple HAPs constellations. It is demonstrated that a DRL approach can optimize the positioning of multiple HAPs in urgent situations requiring wireless connectivity, with limited coverage [184]. The complexity arises from irregular coverage due to site-specific blockage in 3D space. A two-tier design is proposed, including an initial model based on LoS and an advanced design considering LoS/non-LoS channel states. This utilizes double DQN and prioritized experience replay algorithms for decision-making, leading to significantly improved coverage rates compared to benchmark methods like basic DQN and K-means algorithms.
B. Resource Management
As in all wireless communication systems, radio resource management (RRM) is a crucial aspect of ensuring the performance of HAPs-assisted networks. RRM in HAPs networks involves efficient channel allocation, beamforming and antenna management, energy efficiency and power control, and communication resource allocation. Efficient channel allocation is crucial due to the limited spectrum availability. Moreover, meticulous energy management strategies are needed since HAPs systems depend heavily on limited-capacity batteries and variable solar energy. Antenna control involves adapting beamforming and steering strategies to maintain stable connections with ground users. Lastly, communication resource allocation addresses the dynamic distribution of bandwidth and computational resources to meet the demands of various applications and users. Traditional approaches to address these issues often involve intricate optimization and decision-making processes. For instance, the work in [204] suggests a method for optimizing an array of antennas on HAPs for cellular coverage by predicting co-channel interference based on curve-fit approximations for radiation patterns of elliptic beams and estimating optimum beam widths for each cell of a regular hexagonal layout. However, advanced AI techniques are needed to develop more intelligent solutions due to the complexity and scale of HAPs networks. AI has the advantage of fast adaptation and learning from past experiences and mistakes, making it crucial for any HAPs management system. In the rest of this subsection, we survey the main papers that use AI to address antenna control, channel allocation, power control, and communication resource allocation in HAPs-assisted networks.
1) Antenna Control
According to [189], various systems’ dynamic and interdependent properties in the communication link present a challenge for HAPs resource allocation. Modifying resource allocation for one service user interferes with the resource allocation of other users, necessitating readjustment. The authors propose an intelligent beamforming algorithm based on PSO to address this challenge by effectively managing power through game theory principles, thus dealing with interference between service users and expanding coverage zones. The importance of intelligent beamforming is also studied in [186], where it is argued that non-precise beamforming can lead to an increased capacity loss. Therefore, having an accurate HAPs location available at all times by incorporating a location prediction model based on LSTM and trained on the available two-dimensional angle data becomes crucial. Dynamic beam control is also studied in a multi-cell configuration to enhance system throughput and address environmental factors by optimizing antenna parameters using fuzzy Q-learning [191]. Another work [192] presents a DRL-based technique for dynamic antenna control in a HAPs system. It addresses the challenge of reducing low-throughput users caused by HAPs movement due to wind. It outperforms conventional evolution algorithms and RL methods, particularly in non-uniform user distribution scenarios.
2) Channel Allocation
An intelligent DQN-based method for dynamic channel allocation in the HAPs 5G communication system is proposed in [187], which can improve the system’s overall performance by assessing channel quality and service priority. The system can be paired with massive MIMO technology to address issues present in current ground communication systems, offering higher accuracy in channel allocation compared to existing solutions at varying traffic levels. The work in [188] presents an algorithm that integrates Q-learning with an ANN to enable HAPs 5G systems to learn autonomously from their environment and allocate channel resources effectively.
3) Power Control and Communication Resource Allocation
In [190], the authors propose a multi-agent DQN-based algorithm for transmission power control to minimize HAPs interference with existing systems while maintaining high-speed data communication. This approach effectively avoids overestimating action values without compromising performance and produces results comparable to those of the optimal exhaustive search algorithm in various conditions. Additionally, the authors develop a double DQN to prevent action value overestimation. Researchers in [193] discuss the challenges of providing timely communication services in rural areas through HAPs network, particularly for real-time applications like smart agriculture and digital forestry. The conducted work highlights the importance of balancing resource allocation among various services, both freshness-aware and non-freshness-aware, and explores free space optics backhaul in multi-layer HAPs networks. The study introduces static and DRL-based resource allocation schemes. It demonstrates that the dynamic PPO method significantly outperforms heuristic algorithms, improving performance by nearly 2.5 times compared to ant colony optimization methods. Another crucial scenario involving HAPs resource allocation is presented in a recent paper [205]. The paper introduces an intelligent distributed multi-agent DDPG-based resource allocation scheme for HAPs-enabled Internet of Vehicles networks. The proposed scheme optimizes the association and resource allocation strategies of vehicles and other mobile devices to maximize the utility of HAPSs.
C. Caching and Computation Offloading
Leveraging their superior computational resources compared to user terminals, HAPs can function as agile aerial data centers, facilitating efficient offloading of intensive processing tasks. This capability, combined with their high altitude vantage point enabling robust LoS connections, minimizes the risk of disconnections during offloading, fostering seamless and reliable data transfer. However, increased latency and response delays are critical issues to be addressed when using HAPSs as aerial data centers [206]. This can be achieved by optimizing how frequently and where to store content to cache in HAPs-assisted networks. However, it can be complex to determine what content to cache and where to place it in the dynamic HAPs environments due to varying user demands and mobility patterns. Thus, optimizing the distribution of computation offloading between HAPSs and other network components is crucial for balancing computational load, minimizing energy consumption, and ensuring low latency. Efficient content caching and computation offloading can significantly improve QoS, reduce latency by bypassing congested terrestrial networks, and support applications such as remote sensing, emergency response, and IoT connectivity.
AI and ML techniques can be used to develop inventive algorithms for intelligent decision-making to accommodate dynamic network conditions and user requirements and meet the challenges of efficient caching and computation offloading in HAPs network. A study of caching and computation offloading by HAPs in collaboration with terrestrial nodes and intelligent vehicles is presented in [196]. The multi-agent DQN-based solution targets optimal bandwidth and computing resource distribution, emphasizing the importance of caching at network boundaries and highlighting the advantages of HAPs for reducing delays. Another paper aims to minimize the total transmission and computation times in a HAPs-assisted Internet of Vehicles network [199]. The developed optimization problem is formulated and solved using the soft actor-critic (SAC) algorithm, with simulation results demonstrating time reduction through the application of HAPs. Additionally, it shows significant superiority compared to existing approaches. A similar scenario, where HAPs serve as a computing server that can take offloaded tasks from aerial vehicles such as drones and UAVs, is depicted in [197]. A partial offloading strategy based on a DRL method reduces total task latency and energy consumption costs.
A related work [198] introduces a HAPs-mounted MEC system in a rate-splitting multiple-access environment, enabling efficient task offloading for aerial users. It optimally designs key parameters for response latency and energy consumption minimization using a DDPG algorithm with parameter noises, showing superior performance compared to benchmark schemes in simulations. In [195], a DDPG-based algorithm is used to achieve similar objectives. The paper investigates using HAPs in B5G wireless networks to improve user device capacity and computing capabilities, focusing on partial task offloading and communication resource allocation. Employing a DDPG algorithm within an MDP framework, the study seeks to maximize the accomplishment of user tasks while minimizing the energy footprint. It demonstrates that their approach outperforms other methods in simulation experiments.
D. Summary and Lessons Learnt
AI methods are crucial in addressing numerous challenges in HAPs communications, including placement and constellation management, resource management, and caching and computation offloading. These challenges involve complex optimization and decision-making processes that AI techniques can greatly enhance. AI methods such as RL and SI algorithms optimize HAPs constellation coordination for placement and constellation management. RL approaches, while more complex, offer superior coverage peaks, making them suitable for dynamic environments and applications requiring peak performance. SI-based techniques are more efficient and reliable but may offer less ideal coverage outcomes. Resource management in HAPs communications, including channel allocation, antenna control, power regulation, and communication resource assignment, can benefit from AI techniques such as PSO and DRL. These approaches facilitate intelligent beamforming, dynamic antenna control, and autonomous resource distribution while providing substantial advantages for adapting to intricate and ever-changing HAPs network environments. Caching and computation offloading challenges in HAPs communications are addressed with DRL techniques, including DQNs, PPO, and DDPG algorithms. These methods optimize content caching and computation offloading, reducing latency and energy consumption while enhancing the quality of service. In addition to the AI methods mentioned, other techniques such as GAs, Bayesian optimization, and evolutionary algorithms could also be useful for HAPs communications. The flexibility and adaptability of AI make it a valuable tool in addressing the evolving challenges of HAPs communications, ensuring efficient and reliable connectivity for various applications and user demands. The presented AI algorithms can extend the issues addressed in the literature regarding interference management, scalability, real-time applications, link stability, security and privacy, environmental impact, and weather resilience.
AI for Unmanned Aerial Vehicles Communications
UAVs are considered an essential component of the SAGIN architecture, characterized by enhanced flexibility, reduced costs, and increased capabilities. The research community has shown a growing interest in considering UAVs to extend the coverage of terrestrial networks by transmitting information to and from edge users located in remote locations. On the other hand, the efficient integration of UAVs in SAGIN architectures requires addressing several challenges, such as UAV positioning and deployment, trajectory optimization, channel estimation, autonomous navigation, scheduling and resource management, and interference management [4]. Recently, AI tools have been considered to provide efficient solutions to enable fast integration of UAVs in the SAGIN architecture. In this section, we survey existing papers trying to use AI to address the challenges of UAV-assisted wireless networks. Fig. 6 presents the challenges of UAV communications tackled in this survey.
A. UAV Positioning and Deployment
The optimum placement of UAVs is crucial for enhancing coverage and ensuring reliable communication links between UAVs and other airborne/ground nodes [225]. By strategically specifying the UAV locations, substantial gains can be obtained in terms of improved energy efficiency, increased throughput, and reduced latency. Furthermore, optimal UAV placement is needed in multi-UAV scenarios to reduce interference, as spectrum scarcity may necessitate frequency reuse over the spatial domain. Unfortunately, due to the dynamic environment and the limited capabilities of UAVs, finding optimal locations for the UAVs is not an easy task, and several challenges must be resolved. The challenges of optimizing UAV placement become even more intensified in complex and cluttered environments [226], where external obstructions block incoming signals, causing degradation of the wireless transmission.
To optimize the placement of UAVs in wireless networks, researchers explored various approaches by employing classical optimization methods. For instance, the authors in [227] rely on the semi-definite relaxation technique to enhance the physical layer security in UAV networks. The proposed optimization algorithm focuses on adapting the three-dimensional positions of the UAVs to maximize the probability of non-zero secrecy capacity subject to airspace and obstacle constraints. In [228], the authors optimize the deployment of a cluster of UAVs for data collection from randomly distributed sensors. The main objective of the work in [228] is to maximize the network’s total capacity by employing a local interaction game model and an online learning approach to find Nash equilibrium points. In [229], the authors aim to predict the users’ distribution and the downlink data needed during an increased traffic event using a weighted expectation-maximization algorithm and later use this information to optimize UAV positioning. Additionally, the research in [230] focuses on predicting nearby flying objects’ positions and categorizing them into groups to aid control and communication protocols using expectation maximization and Bayesian inference.
Due to the dynamic features of UAVs, and as the integration of UAVs in SAGINs increases further the search space and the number of components affecting the UAV connections, classical optimization tools fail to mitigate the added complexity in obtaining the best UAV locations to maximize the overall performance of the network. At the same time, there has been a growing interest in utilizing AI to devise efficient deployment strategies for UAV networks. AI can handle complex and dynamic decision-making processes, adapt to changing scenarios, and learn from data. This leads to more effective and automated strategies for optimizing UAV positioning, reducing the need for human intervention, and minimizing reaction time in response to problems or accidents. Table 7 lists the main works in the literature that use AI to enhance UAV deployment in wireless networks.
Numerous studies propose the successful deployment of UAVs alongside ground BSs during network traffic overload. When cellular hotspot areas with increased data demand require more BSs, utilizing UAVs through predictive methods can be seen as a cost-effective and efficient approach for on-demand network communication. In [209], the authors propose an ML framework that predicts possible network congestion and determines optimal locations for UAV deployment to provide wireless service when needed. The results demonstrate that this predictive ML method, based on a K-means clustering algorithm, can significantly reduce power requirements for downlink transmission and mobility compared to deploying UAVs without any ML prediction techniques. Another study [216] examines optimizing deployment and backhauling topology in tethered UAV-assisted integrated access and backhaul networks, introducing a twin-delayed DDPG-based learning framework to minimize deployment costs. Q-learning is also utilized to efficiently size excess mixed traffic demands on terrestrial BSs and manage subsequent offloading, treating UAVs as temporary BSs [217].
Another subset of studies explores the optimization of UAV placement using advanced AI techniques to improve the throughput. The authors in [211] sequentially employ various RL and DL methods to predict the optimal positioning of UAVs in multi-UAV and multi-user scenarios, aiming to maximize throughput, while the work done in [221] uses DQN to achieve the same goal. Similarly, a DDPG-based approach for 3D deployment and power allocation is proposed in [212], demonstrating superior performance compared to other DQN and GA-based algorithms in increasing system throughput. The work in [220] combines DDPG and double DQN for UAV placement, resource allocation, and computation offloading in the terahertz (THz) band. Another method [215] incorporates DRL into an FL framework, allowing UAVs to make real-time decisions based on local observations while achieving a globally optimal solution, to maximize network throughput while ensuring user privacy. The SAC-based algorithm can also be utilized for optimizing the 3D deployment of UAVs while minimizing power requirements [218]. Prioritizing fairness and efficiency, the proposed framework strives to maximize the guaranteed minimum user rate, ensuring all users experience equitable resource allocation. Furthermore, this framework exhibits enhanced stability and greater long-term rewards compared to other DRL-based approaches. SAC can also be utilized for the joint initial deployment and trajectory design of UAVs to reduce downlink rate and improve convergence speed [222]. In [219], the authors introduce the DRL PPO algorithm, which jointly optimizes UAV deployment and power allocation, outperforming other baselines in system throughput and energy efficiency. A different approach to achieving a trade-off between energy conservation and user connectivity is proposed in [224]. This work presents a multi-agent Q-learning algorithm to optimize the placement and transmission power of UAV-BSs, aiming to minimize the number of ground users experiencing outages while minimizing network energy consumption.
To address the challenges in optimizing the deployment of UAV-aided communication networks for emergency scenarios, the authors in [213] propose an iterative two-stage multi-agent SAC approach. The main objective is to maximize the system throughput and coverage during the deployment duration in emergency communication networks. Another paper [214] proposes a scheme for quickly deploying a UAV-enabled emergency cellular network in disaster scenarios to support rescue operations. The placement of UAV aerial BSs is achieved through a modified K-means algorithm, and end-user signal quality is improved through joint coordinated multi-point transmission and reception.
Several other papers address various aspects of UAV positioning and deployment challenges. In [207], the authors propose an efficient radio map reconstruction method for autonomous positioning algorithms of UAVs using a k-nearest neighbor (KNN)-based approach. In [210], an RNN is used to predict the angle arrival for better position estimation of a UAV. The work in [208] tackles generic flying path planning using a generalized regression NN. Finally, the authors in [223] utilize a DDPG-based algorithm to optimize the deployment of UAVs for handling dynamic content requests and user mobility in UAV-assisted AR applications.
In conclusion, the deployment and optimal placement of UAVs in SAGINs represent a challenging yet crucial area of research. The various AI algorithms and techniques discussed in this survey highlight the diverse approaches taken to address main system considerations, such as power allocation and system throughput, by carefully optimizing UAV deployment. Furthermore, applying RL algorithms, such as Q-learning, DDPG, and multi-agent SAC, showcases the potential of adaptable and intelligent UAV-assisted communication systems.
B. Channel Estimation
Efficient estimates of the aerial wireless channels are required to achieve precise alignment of beams between UAV and other SAGIN components. Unfortunately, channel estimation in UAV communications entails unique challenges due to the high altitude and three-dimensional placement of UAVs [18]. Furthermore, the aerial channels are subject to the Doppler effect due to the continuous UAV navigation. The placement of UAVs and the surrounding environment also significantly impacts the propagation characteristics of UAV communications. This leads to fading and time-frequency selectivity in dynamic UAV channels [246]. Moreover, when UAVs are equipped with many antennas, the non-negligible propagation delay across the array aperture causes a beam squint effect, further burdening the estimation of UAV channels.
Traditional methods involving signal processing and numerical optimization have conventionally tackled channel estimation in UAV networks [247], [248], [249]. However, optimization algorithms typically demand significant computational complexity, creating a gap between theoretical design and real-time processing requirements. Therefore, leveraging previous dataset observations and employing AI models become valuable strategies for learning the complex mapping from compressed received pilots to channels. For instance, DNNs tackle optimization problems by learning the best parameters to reduce the error in their predictions. This application of DNNs is anticipated to significantly reduce computational complexity and processing overhead, relying on several layers performing elementary operations like matrix-vector multiplications. Table 8 highlights the main works in the literature that use AI to optimize channel estimation in UAV networks.
In [240], the authors investigate a data-driven generative NN approach for designing millimeter wave (mmWave) air-to-ground channels. The method can assess the directional properties of the channel at both transmitter and receiver ends. The proposed generative model consists of two stages: first, it predicts the link’s state and then uses this state as input to a conditional variational auto-encoder that calculates arrival and departure angles, delays, and path losses of all propagation paths. The benefits of using synthetic big data for channel estimation are explored in [234], which involves multiple ray tracing simulations with various urban landscapes to create a communication channel model of UAVs acting as a communication relay. This work studies a hybrid channel modeling approach for optimizing communication relay UAV flight paths. The results demonstrate that coupling the receding horizon (RH) with the NN outperforms the RH-only approach by approximately 10% increased probability of achieving the necessary communication strength. Other works [232], [239] also utilize ray tracing software to generate samples subsequently used by the random forest and KNN algorithms for channel estimation.
In [243], an LSTM-based technique for air-to-ground transmission is proposed to generate more accurate CSI and greater resilience compared to the least square and minimum mean square error algorithms. The input, forget, and output gates are used to learn the time correlation of UAV channels, while the forget and input gates are utilized to build a memory function. This memory function and the output gate are then used to estimate the current slot CSI. Another study [245] also employs LSTM in formulating a channel estimation method based on the LSTM-distributed conditional generative adversarial network (GAN). Several studies utilize ANNs and DNNs to forecast diverse channel parameters, such as signal strength and fading channel conditions [233], [235], [236], [242], [244]. These approaches facilitate adaptive data transmission and lower power consumption. Various ML algorithms employed for the same purpose include ensemble learning methods [237], K-means clustering [238], SVMs [231], and GANs [241].
When selecting a method, key considerations include aligning with specific channel parameters, accounting for the amount and quality of training data, evaluating computational complexity, and prioritizing methods with low latency for real-time requirements in channel estimation or prediction. LSTMs and DNNs emerge as robust choices for UAV channel estimation, especially when dealing with complex, time-varying data. K-means clustering can prove useful in simpler scenarios or when data is limited, while GANs offer versatility by aiding data augmentation or exploring potential channel variations.
C. Interference Management
The high altitude of UAVs results in predominantly strong LoS links between the UAVs and the ground nodes, and between the UAVs and the upper layers of the SAGIN, presenting both advantageous opportunities and accompanying challenges. On one side, LoS links improve communication quality with UAVs and reduce losses due to blockages. Conversely, this setup can introduce significant interference in uplink and downlink communication, posing a more complex challenge in handling interference, particularly in scenarios where terrestrial and aerial nodes coexist. Several studies tackle the issues related to interference management in UAV networks by referring to classical optimization methods to optimize resource and spectrum allocation in addition to UAV placement and path planning. For instance, the authors in [250] present a method for cooperative interference cancellation and sum rate maximization in multibeam UAV communication in the cellular uplink. Their study focuses on reducing interference from other users in the cellular network to enhance overall system performance using the SCA method. Additionally, the authors in [251] investigate vulnerability assessment of UAV networks in interference scenarios using a coupled-map-lattices-based approach. The objective is to evaluate the susceptibility of crucial nodes in the UAV network, such as the control center, under interference conditions.
At the same time, AI techniques can be particularly effective in addressing interference-related issues by efficiently processing large amounts of data to detect and avoid such interference. AI can adapt to dynamic environments, optimize communication parameters, reduce signaling overhead, and learn from data. This enables efficient and reliable communication even in complex multi-UAV scenarios. Table 9 highlights several recent papers that use AI tools to devise efficient interference management techniques in UAV networks.
One way to mitigate the degradation effect of interference in UAV networks is through careful flight path planning. Focusing on cellular-connected UAVs, the work in [252] introduces an innovative path-planning algorithm that considers potential interference between UAVs and cellular infrastructure to optimize network performance. They utilize echo state network cells and a DRL algorithm to minimize time-dependent utility functions by transferring each network state observation to an action. Simulation results demonstrate improved wireless latency and ground user rate, with fewer steps required compared to a heuristic baseline. The optimal altitude of the UAVs varies based on terrestrial network density and user data rate requirements. Height optimization aiming to minimize interference is also investigated in [257]. The method proposed, based on support vector regression, linear regression, and ANNs, is assessed using outage probability and BER and demonstrates encouraging outcomes in enhancing the performance of UAV networks. A similar approach is followed in [253], where the authors aim to improve the energy efficiency of UAVs while minimizing interference on ground users. Simulations show that the deep Q-learning technique enhances UAV energy and spectrum efficiency while minimizing disruption to terrestrial users.
In UAV communication, interference may also result from jamming attacks on the exposed link. In [254], the authors propose a CNN-based method for detecting such jamming attacks in UAV networks, making it possible for UAVs to mitigate the interference caused by them. To address the low processing capability of UAVs, they suggest incorporating a DNN with a self-attention layer that can accurately identify attacks with as little as 2 dBm power. Simulations indicate that it is easier to detect attacks with three or more attackers, fewer users, and power levels exceeding 10 dBm, particularly considering the 3D distance between the small cell and the authenticated UAV for improved recognition accuracy. Examining the UAV deployment and receiver jamming method simultaneously for optimal security while considering channel uncertainty is another approach to interference management [255]. This study introduces a data-trained DNN and DQN to approximate the best UAV deployment, enhancing transmission security performance. This algorithm enables UAVs to serve as secure relays for forwarding sensitive information.
Another DRL-driven approach is presented in [256], which addresses the challenge of inter-cell interference in three-dimensional cellular networks, particularly with UAVs. The proposed solution based on deep Q-learning effectively mitigates interference without requiring channel information. The paper also discusses methods to scale the algorithms efficiently and decentralize them using multi-agent RL, facilitating the growth of civilian UAVs. Another study [258] suggests a hybrid double dueling DQN-TD3 approach to address inter-cell interference and improve wireless transmission quality for UAVs. The proposed algorithm demonstrates superior performance compared to existing methods regarding sum rate, outage probability, and collision probability.
D. Autonomous Navigation and Trajectory Optimization
Precise trajectory planning is the cornerstone for autonomous UAV operations, enabling efficient navigation, agile maneuvering, and achieving mission objectives. Traditional path planning and trajectory optimization approaches, which rely mainly on classical optimization tools, categorize the available work into two categories. The first category focuses on autonomous UAV navigation, which includes conflict resolution mechanisms like collision avoidance, collective motion control through flocking algorithms, and efficient area coverage techniques. The second category targets enhancing the performance of different tasks, such as energy efficiency, beamforming, and throughput, via UAV trajectory optimization. Unfortunately, conventional optimization solutions for designing autonomous navigation algorithms and optimizing UAV trajectories encounter significant drawbacks. Firstly, formulating an optimization problem necessitates a radio propagation model that is both accurate and manageable. Recent studies [284], [285], [286] have commonly employed statistical models such as simplified LoS-dominated models, probabilistic LoS models, and angle-dependent Rician fading models. Yet, these models typically offer predictions only in an average sense and cannot ensure performance in the specific local environment where UAVs are deployed. Secondly, trajectory designs based on offline optimization assume the availability of precise CSI derived explicitly from a particular radio propagation model. However, acquiring perfect CSI in practical scenarios becomes challenging due to uncertainties related to the UAV’s position and the dynamic nature of the communication environment [287]. Lastly, many optimization problems related to UAV-assisted communication systems are highly non-convex and present difficulties in efficient resolution. Recently, AI techniques and, more specifically, RL algorithms have been serving as efficient solutions to optimize the trajectory of UAVs in SAGINs, thanks to their ability to design optimal strategies based on collected samples in real time when no prior information about the environment exists. Table 10 lists some of the main works in the literature that leverage AI to optimize the trajectory and enhance the autonomous navigation of UAVs.
In [269], the authors propose a deep Q-learning-based scheme for UAV navigation through massive MIMO. They treat individual UAV-ground connections as agents and compute the optimal locations for the UAVs based on signal power at reception. The authors of [260] propose a DRL-based method for UAV navigation, optimizing the energy efficiency defined as a function of communication coverage, fairness, energy consumption, and connectivity. The proposed method utilizes two robust DNNs to learn about the environment and its dynamics to make informed decisions. Simulation results demonstrate that this approach consistently outperforms commonly used baseline methods regarding coverage and fairness. Similar work is proposed in [280], where the authors present a decentralized approach combining the coalition formation game and multi-agent DRL PPO algorithm to jointly optimize the trajectories of UAVs and the ground user associations to maximize the total throughput and energy efficiency. Minimizing energy consumption while performing trajectory optimization using a PPO-based approach is also studied in [281].
In [268], the authors investigate the trajectory optimization problem in UAV-assisted networks to maximize the sum rate of users each UAV serves. Two sub-problems are identified: UAV trajectory optimization and joint power and sub-channel assignment, for which the authors develop a distributed algorithm based on Q-learning. The results of simulations show that Q-learning efficiently optimizes trajectories using reward signals from the network’s topology. Maximizing the communication rate while maintaining energy efficiency is also studied in [282]. This paper proposes a two-step iterative approach that employs a heuristic algorithm for adaptive network formation and multi-agent DDPG for trajectory optimization to optimize the multi-hop UAV network topology for data collection from ground users.
A decision tree model-based RL approach for autonomous UAV navigation is presented in [259]. The algorithm can efficiently learn a trajectory in just a few iterations and outperform Q-learning methods in terms of action reward. Another study investigates how UAVs can track ground targets while avoiding obstacles using DDPG and LSTM algorithms [270]. Additionally, extensive research is conducted on cooperative approaches to learning the environment that optimizes the number of steps needed for the Q-learning-based algorithm to converge, detailed in [261]. A multi-agent Q-learning-based approach is employed in similar works to perform joint trajectory planning and power control. It utilizes the echo state network’s predictions about users’ mobility information. Other relevant studies using Q-learning are documented in [263], [279], while those employing a Q-network are referenced in [264], [274], [275], [276]. Additionally, variants of policy gradient algorithms are examined in [265], [266], [267], [271], [272], [273], [276], [277], [278], [283] encompassing diverse applications related to UAV path design and optimization strategies.
All works summarized in Table 10 rely on RL methods that offer various advantages to optimizing UAV navigation for effective communications. DQN is a simple and efficient algorithm suitable for discrete spaces and adept at handling partial observability, while PPO and DDPG are more effective in handling continuous action spaces. Multi-agent DDPG suits scenarios with multiple cooperating UAVs, enabling decentralized decision-making. Model-based RL algorithms are more sample-efficient, allowing proactive trajectory optimization strategies. Choosing the right method depends on environment complexity, action space type, communication complexity, and computational resources. For instance, DQN might suffice for simple environments, while PPO or DDPG may be required for complex scenarios.
Hybrid approaches, combining RL with evolutionary algorithms, planning algorithms, or other ML techniques, can further enhance performance. Successful implementation necessitates designing effective reward functions, utilizing realistic simulation environments, and rigorous testing and evaluation of navigation efficiency and communication quality. Hundreds of papers tackle the trajectory optimization of UAVs using AI, and this subsection presents a representative subset, capturing recent papers covering different AI methods and objectives. However, a more detailed survey on the use of RL can be found in [24] and on the use of ML in [34].
E. Scheduling and Resource Management
The constrained resources and hardware limitations inherent in UAVs, combined with their constant movement and exposure to unpredictable environmental factors, impose several constraints on the connectivity and QoS of UAV networks [306]. Specifically, the success of UAV networks depends heavily on the effectiveness of resource allocation and optimization strategies. In particular, efficient power, spectrum, and storage resources are critical in maximizing network performance and achieving mission objectives [307]. Traditional optimization and game theory techniques are used extensively to maximize the system throughput and fairness and address the limited availability of resources while respecting the UAV hardware constraints. For instance, the authors in [308] formulate a joint resource optimization problem to maximize the energy efficiency of UAVs by optimizing communication scheduling, transmit power, and motion parameters. Another work addresses energy-efficient joint scheduling and resource management in UAV-enabled multi-cell networks [309], proposing both coordinated and uncoordinated convex optimization approaches to optimize UAV locations and resource management to maximize network energy efficiency and evaluate their performance in dynamic and stationary scenarios. For SAGINs operating in dynamic and uncertain contexts, traditional optimization methods such as game theory and convex optimization often prove to be cumbersome and suboptimal. AI, however, emerges as a more efficient and effective approach. By enabling SAGINs to learn and adapt in real time, AI empowers them to reach optimal solutions even in unpredictable situations. Moreover, AI’s capability for multi-objective optimization allows for simultaneous resource allocation across diverse system aspects, maximizing overall performance and efficiency. Additionally, AI enables proactive management while effectively addressing uncertainty to make UAV communication systems more efficient and reliable. Table 11 presents the main works considering optimizing scheduling and resource allocation in UAV networks using AI.
In [291], the authors explore the capacity of UAVs to compute their routes using real-time learned data from the surrounding environment. They propose a spatio-temporal scheduling framework for autonomous UAVs based on Q-learning, enabling them to autonomously organize their schedules to cover a maximum number of pre-scheduled events within a specific geographical area and during a set period. In unforeseen emergencies, the framework can adapt to the planned schedules. Additionally, they introduce a reward function that considers battery capacity constraints, event time frames, and navigation delays between events caused by UAVs. The spectrum interaction technology of the flight formation of UAVs is utilized to address the issue of spectrum sharing in [292]. Priority allocation determines the importance of UAV tasks, and the current state is assessed before each time slot, followed by action selection, information transmission, and policy updating (formed by combining DRL and LSTM models) at the end of each round. The suggested method converges rapidly and demonstrates strong performance in dynamic channel allocation and time slot allocation models.
In device-to-device (D2D) communications, UAVs may be deployed to enhance user experience and network performance. However, the UAVs’ constant movement, limited energy, and flight duration cause challenges for their real-time use. To solve this issue, the work in [301] offers a DRL model for determining the most efficient energy harvesting schedule in UAV-assisted D2D communications. The system considers random user movement, randomly altering channel status information in each slot, and the UAV flying around a central point. The results reveal that the proposed scheme outperforms existing schemes in processing speed using an off-the-shelf processor employing trained NNs, indicating its capabilities in handling real-time resource allocation problems in UAV networks. In another work [296], RL is used to optimize the wireless charging scheme of a drone fleet to reduce wasted energy and increase the bandwidth provided to the users. High user mobility is also studied in [304], in which the authors use a multi-agent constrained SAC RL algorithm to efficiently allocate resources in multi-UAV-assisted vehicular networks while accounting for spectrum efficiency and computing power with security constraints and attention mechanisms.
Another RL method based on the optimal policy is proposed in [293]. It aims to minimize power consumption, job loss, and delay while offering an architecture for extending 5G network slices with UAVs. This is particularly crucial in network slicing within 5G-enabled systems with stringent latency constraints that require computing resources at the edge. As data increases and distances to the network’s edge increase, this task becomes increasingly challenging. The joint optimization of UAV communication latency, energy consumption, and age of information in the context of RL-aided resource allocation in UAV-enabled Internet of Medical Things networks is also studied in [194], in which authors use a multi-agent DDPG algorithm. Another study [305] investigates the performance of MEC using UAVs, focusing on security. The approach utilizes the Dinkelbach method and DRL to optimize terminal users’ binary offloading decisions and resource management while ensuring dynamic task data queue stability and minimum secure computing requirements.
Beyond traditional static approaches, AI algorithms are revolutionizing resource management in UAV networks. Computation offloading, powered by algorithms like RL actor-critic network and LSTM in [294], [302], dynamically shifts processing tasks from resource-constrained UAVs to the ground, dramatically reducing average response times and overall energy consumption. RL, in general, is widely employed for scheduling and resource management. For instance, onboard Q-learning can empower UAVs to manage resource allocation directly, maximizing user QoS metrics as in [288]. Furthermore, DDPG-driven trajectory planning and frequency band allocation, as in [300], prioritize energy-efficient operation, extending flight time and network coverage. Task-scheduling algorithms, exemplified by a DQN [289], enable real-time adjustments to UAV task strategies based on live data, optimizing resource utilization and mission fulfillment. AI algorithms like LSTM and RNN can dynamically estimate UAV-user throughput to ensure consistent QoS, as seen in [290], [298]. Predictive beamforming algorithms, based on the same LSTM and RNN algorithms and presented in [297], [299], predictive beamforming algorithms can precisely calculate optimal communication location and angle, focusing transmissions for efficient data delivery. Finally, LSTM can also enable flexible control of transmission energy, further optimizing resource usage and network performance [303].
F. Summary and Lessons Learnt
Integrating AI, particularly DRL and DL techniques, proves to be a game-changer for UAV communications. It overcomes challenges associated with dynamic environments and network constraints while optimizing positioning, deployment, channel estimation, interference management, autonomous navigation, trajectory, scheduling, and resource management for UAVs. The discussion on AI in UAV communications highlights its role in real-time adjustments based on live data for significant performance improvements.
DRL techniques are the main focus of this section, mostly due to their ability to improve dynamic control algorithms. DRL and DL algorithms contribute to predictive positioning, path planning, and trajectory optimization for UAV deployment. This minimizes the need for manual intervention while enhancing performance. Additionally, AI methods such as DNNs and RNNs improve accuracy in channel estimation and enable real-time processing, which is essential for adapting to dynamic UAV channels. Furthermore, AI-driven approaches like deep Q-learning and CNNs can optimize interference management, adapt to changing environments, and enhance network reliability. Autonomous navigation benefits from DRL and DL by allowing UAVs to learn from data sources to optimize paths autonomously in response to dynamic situations. In scheduling and resource management domains, AI technologies like DRL empower UAV systems with improved schedules within dynamic environments while bolstering network efficiency.
AI has limitations when solving UAV communications-related challenges, especially in dynamic and unpredictable environments. While DRL shows promise in optimizing network performance, it still faces challenges in rapidly changing conditions. Due to unique obstacles, privacy, safety, and air traffic management also require specialized AI solutions. Beyond technical challenges, the lack of transparency inherent to many ML models raises concerns about interpretability and accountability in UAV communication systems. Ensuring clear and explainable decision-making processes is crucial for guaranteeing the safety and reliability of communication protocols in these critical applications.
The future of AI in UAV communications holds exciting possibilities, but certain research directions and challenges need attention. One potential avenue for advancement is the exploration of SI algorithms in the context of UAV communications. Inspired by the collective behavior of biological systems, SI algorithms could offer innovative solutions for optimizing the coordination and communication patterns among multiple UAVs in dynamic environments. This approach may enhance UAV networks’ adaptability, fault tolerance, and scalability. Moreover, developing hybrid AI methods and combining different AI techniques can also be a key focus for future research. These hybrid approaches leverage the strengths of different AI techniques and other emerging technologies, such as Blockchain and 6G, to enhance UAV communications capabilities further. FL is a promising approach that enables decentralized intelligence gathering and decision-making among multiple UAVs in SAGINs. Explainable AI and trustworthy systems are essential for ensuring transparency, interpretability, and ethical considerations in AI-driven UAV operations. Integration with edge computing, Blockchain, multi-modal perception, and human-AI collaboration can be identified as key research directions to enhance the capabilities of UAV communication systems. Open challenges include data security and privacy concerns associated with the increasing reliance on data collection in AI-powered UAV networks. Energy efficiency and resource management remain critical challenges, urging the development of energy-aware AI algorithms. Safety and reliability in dynamic environments, scalability, interoperability, and ethical considerations in regulation are key challenges that must be addressed to deploy AI-powered UAV networks responsibly.
AI for Optimizing Space-Air-Ground Integrated Networks
SAGINs, which consist of satellites, HAPs, UAVs, and terrestrial networks, represent one of the most complex network architectures globally. These components collaborate to deliver enhanced and adaptable end-to-end user services within the framework of the new 6G paradigm. Even though SAGINs are designed to attain uninterrupted coverage across all areas at all times while facilitating high-rate and dependable transmission over a larger area compared to traditional terrestrial networks [343], some performance degradation is inevitable unless optimization takes place. While each component of the SAGIN architecture has the potential to optimize its performance individually, it is not sufficient to optimize them separately. Joint optimization of these layers is essential to exploit the potential of SAGINs fully [334]. Integrating these components necessitates carefully considering factors such as orchestration and topology management, scheduling and collaborative resource management, routing and flexible mobility management, and caching and computation offloading. While these challenges are also faced in many heterogeneous complex networks, the dynamic and unpredictable nature of SAGINs, with their diverse components operating in varying environments, significantly amplifies the complexity of these challenges, thus justifying the need to use AI techniques to address them. Tables 12 and 13 provide a comprehensive summary of studies that have leveraged AI methodologies to tackle the challenges associated with optimizing SAGINs.
A. Orchestration and Topology Management
Due to the distributed nature of SAGINs, topology control is a critical aspect that significantly impacts the overall network performance. Distributed topology management schemes can minimize bottlenecks and response time but increase network complexity due to cooperative operation across different SAGIN components. On the other hand, centralized control can simplify the network layout but may lead to substantial response delays, further degrading the network performance. Additionally, SAGIN’s extremely dynamic network topology increases link changes, making data transmission via fixed traffic impossible. Thus, topology management for SAGINs requires careful consideration of user association and aerial node trajectory optimization levels [9]. User association determines the connections between users and network segments, influencing overall network topology and resource allocation. This has a substantial impact on network performance and efficiency. Simultaneously, optimizing the flight paths of aerial platforms, such as UAVs, enhances coverage and connectivity, reduces congestion, and minimizes energy consumption. Moreover, this approach ensures comprehensive networking adaptability to dynamic environmental conditions. Realizing a truly responsive and versatile network in SAGINs requires a reconfigurable architecture capable of intelligent resource allocation based on situational awareness. This architectural design should enable SAGINs to activate satellite services rapidly, optimize routing paths for long-distance communication needs, and dynamically dispatch UAVs according to the demands of emergencies. Traditional optimization tools may be the first choice when formulating architectural design and topology problems. For example, the work in [344] proposes a reconfigurable intelligent surface (RIS)-assisted UAV communication system to maximize the average achievable rate by jointly designing UAV trajectory and passive beamforming using SCA. Another paper [202] addresses optimizing HAPs locations and user association to maximize users’ throughput by employing binary linear optimization.
However, the inherent trade-off between data complexity and optimization efficiency in SAGINs presents a major obstacle for conventional topology and orchestration management tools. Thankfully, advancements in AI, particularly ML and DNNs, are paving the way for a paradigm shift. These data-driven methods offer robust and scalable network and topology optimization solutions, revolutionizing how we manage complex SAGIN infrastructures. For instance, the study conducted in [310] demonstrates how deep Q-learning can establish a framework for dynamically managing satellite-terrestrial networks while controlling caching and computing resources. Deep Q-learning is also applied in [319], which addresses user association in SAGINs with delayed CSI to maximize the sum rate. The authors utilize deep Q-learning to enable the satellite controller to decide whether a user should be associated with a HAPs or a terrestrial BS based on the network’s state and achieve higher performance than traditional methods. Additionally, another paper [318] proposes task-dynamic processing through a multi-UAV cooperative strategy for optimizing UAV deployment and allocating computing resources by offloading tasks to other UAVs and HAPs. The challenge addressed is UAVs’ limited computing resources and energy, which are resolved using the deep Q-learning algorithm for optimized task offloading while enhancing service capacity and reducing energy consumption. In [312], the authors present a trajectory control approach based on Q-learning to tackle the difficulties posed by the unpredictable and uncertain environment and the presence of malicious jamming in satellite-UAV coordination networks. Focusing on optimizing the operation of hybrid satellite-terrestrial networks augmented by UAVs, [315] explores efficient trajectory design and link selection strategies. It addresses this challenge through Q-learning to adjust UAV locations and determine optimal link selection using a graph NN. Similarly, another publication [313] deals with optimizing multi-satellite association and multi-UAV trajectories in time-varying non-terrestrial network topologies. It employs multi-agent Q-learning to integrate satellite networks efficiently with UAV relaying, aiming to maximize system throughput while minimizing energy consumption.
DQNs are widespread in addressing challenges such as trajectory optimization, link association, and topology management of SAGINs. For example, the authors in [323] propose a DQN-based method to tackle the partially observable user association problem in a satellite-HAPs-terrestrial network to maximize the sum rate while reducing CSI overhead. The proposed algorithm outperforms existing approaches and underscores the importance of incorporating action vectors in the policy network. Additionally, another publication [321] addresses SAGIN optimization considering non-terrestrial network complexity, mobility, and heterogeneity by utilizing the DQN algorithm for UAV trajectories and resource management optimization to maximize HAPs-UAVs-terrestrial network performance as well as user satisfaction.
Similarly, an approach based on policy-based DRL is employed in [322]. This study focuses on optimizing access control in aerial networks involving satellites and HAPs to maximize total network revenue while minimizing computational complexity. Another publication [317] tackles optimizing the performance of RIS-assisted satellite-UAV-terrestrial networks. The method proposed utilizes multi-objective DDPG for joint optimization of UAV trajectory, RIS configuration, and downlink beamforming to maximize the system’s achievable rate and minimize UAV energy consumption. The work done in [320] discusses optimizing association and trajectories in a time-varying topology for satellite-UAV networks. It employs a multi-agent actor-critic network to enhance communication efficiency while reducing energy consumption, resulting in notable performance enhancements. Another study by [314] focuses on minimizing the age of information in IoT data collection using SAC to optimize UAV trajectories and address the challenge of information freshness in wide-area IoT networks. Finally, LSTM can be supplied with different user profile information (position, power availability, and energy usage) as well as configuration data for aerial-terrestrial-maritime networks to enhance the network and decrease packet delay while enhancing energy effectiveness [316].
B. Scheduling and Collaborative Resource Management
SAGINs present unique challenges compared to terrestrial networks, primarily due to their constantly shifting topology, diverse network elements, and intricate relational patterns within the network architecture. Additionally, the dynamic resource constraints imposed by the high mobility of SAGIN components necessitate flexible and adaptive resource management strategies. In addition, SAGINs allow for different services with different requirements and needs. Thus, intelligent and collaborative resource scheduling is the key technology for maximizing network performance and application responsiveness, considering the dynamic network topology, varying communication ranges, and heterogeneous resource capabilities of SAGIN systems [19]. Sub-optimal scheduling by focusing on only parts of the SAGIN architecture may cause data bottlenecks and congested links, affecting critical applications such as remote surgery or emergency response. Additionally, inefficient use of resources can lead to wasting valuable bandwidth and processing power while struggling to meet demand in certain areas. Lastly, frequent switching between resources and unnecessary data relays can deplete limited onboard power, especially for UAV operations.
Classical SAGIN scheduling and resource management approaches rely on rule-based algorithms and static optimization techniques. These methods often struggle with the dynamic nature of SAGINs, leading to sub-optimal solutions. For instance, in [345], the angle-based diversity strategy is utilized in priority-based and load-balancing scheduling algorithms that use software-defined networking and network function virtualization as a foundation. Another work [346] utilizes a progressive convex approximation to enhance the system’s energy efficiency by optimizing sub-channel selection, uplink transmission power control, and UAV relay deployment in SAGINs. A joint heuristic subchannel assignment and power allocation algorithm for NOMA-enabled cognitive satellite-UAV-terrestrial networks with imperfect CSI is proposed in [347]. The resource allocation problem is formulated as a mixed integer non-linear programming problem and is solved using heuristic approaches, Taylor series, and SCA-based techniques. The work in [348] assumes a cooperative LEO satellites/HAPs network and proposes a two-tier matching algorithm to optimize the matching between users and HAPs and between HAPs and satellites. These classical approaches often face computational complexity, especially for large networks or scenarios, and limited scalability; they rely on accurate channel models. On the other hand, AI models can learn from network behavior to dynamically predict traffic patterns, enabling proactive scheduling. Also, DRL algorithms can interact with the environment in real time without any models or historical data. Thus, leveraging AI algorithms, service-driven resource management paradigms emerge as a powerful tool for SAGINs. These intelligent and collaborative strategies allow real-time adaptation to dynamically evolving service demands, ensuring flexible and efficient management of SAGIN’s diverse network offerings.
Many proposed AI-based methods heavily rely on supervised learning techniques. For instance, intelligent spectrum sharing in satellite-terrestrial networks can be achieved using supervised learning algorithms such as a combination of SVM and CNN algorithms [325]. These methods are utilized for making spectrum predictions based on wideband and narrow-band measurements, demonstrating higher spectrum management efficiency when used together. Yet another supervised learning approach is suggested to tackle the channel estimation and synchronization issues within a HAPs-LEO network [330]. Here, CNN-based estimators are employed to minimize channel equalization and carrier frequency offset, resulting in improved data rate throughput per second and superior service quality due to an agile signal reconstruction process. In [328], the authors optimize user scheduling in satellite-HAPs-terrestrial networks, focusing on these networks’ complexity and stochastic nature. The adopted approach employs unsupervised learning, particularly ensembling DNNs, to achieve higher sum rate performance with reduced computational complexity, tackling the real-time processing challenges in these networks. Another study [327] introduces an innovative approach for cross-network radio resource scheduling across all layers of SAGINs. This research contributes to the field by addressing the issues related to unified mapping and scheduling of radio resources within SAGINs. Utilizing a GA and a unified resource mapping method, this proposed scheme optimizes resource allocation and improves the performance of SAGINs, specifically in emergency logistics scenarios.
Finally, DRL is also extensively utilized to address the challenge of scheduling and collaboratively optimizing SAGINs’ resources. For instance, the authors in [329] address the SAGIN resource allocation issue using the deep Q-learning approach. Antenna pointing, mobile tracking, and resource scheduling can also be addressed with DQN, as demonstrated in [324]. A similar study is conducted in [326], which formulates the resource scheduling problem of SAGIN as a multi-domain virtual network embedding. They leverage DQN to improve the performance of the multi-domain virtual network embedding algorithm by employing basic NN modules to construct a five-layer policy network. Another paper [331] tackles the issue of enhancing RIS-assisted hybrid free space optics/radio frequency-enabled HAPs-UAVs-terrestrial networks. It uses an asymmetric LSTM-DDPG algorithm to optimize active and passive beamforming techniques simultaneously, intending to improve system performance in complex and dynamic network environments. Another study [332] focuses on energy-efficient resource allocation and privacy preservation of offloaded tasks in HAPs-UAVs-enabled MEC systems. It employs hierarchical deep federated RL to optimize resource allocation and decision-making, addressing the challenge of minimizing energy consumption while meeting diverse task requirements in dynamic environments. The optimization of beam selection for UAV- and HAPs-based networks to enhance connectivity for ground vehicles in 5G systems through RL is explored in [333]. This work addresses fluctuating traffic and environmental variations using the multi-armed bandit algorithm to autonomously select optimal beams over time.
C. Routing and Flexible Mobility Management
Routing and mobility management in SAGINs pose unique challenges due to the heterogeneous nature of network components and their mobility. Adaptive routing protocols are required to handle dynamic topology and energy constraints. At the same time, seamless mobility and handover mitigation strategies must be developed to maintain uninterrupted communication during the continuous movement of satellites and aerial nodes [349]. Furthermore, the expanding scale and diverse node types within SAGINs demand innovative solutions for inter-layer mobility management. While ultra-dense satellite constellations offer additional capacity, they also introduce intricate handover dynamics with increased overhead and potential delays. Navigating these complexities and maintaining consistent performance, routing, and mobility protocols require significant advancements in decision-making speed, resource efficiency, and intelligent adaptation to network conditions. The traditional methods used to address the routing and mobility management challenges of SAGINs, like rule-based routing algorithms, do not have the adaptability to handle the dynamic characteristics of SAGINs and often produce less than optimal routes [350]. Pre-determined handover triggers based only on signal strength may not completely capture the context of network conditions, resulting in unnecessary handovers [351]. However, AI-enabled solutions, including ML and DRL, can optimize routing and handover decisions based on real-time network conditions and environmental factors, ensuring seamless and reliable SAGIN communication. For instance, CNN is utilized in [334] to address the routing issue. The objective is for CNN to learn traffic patterns within multiple SAGIN layers and use this information to optimize routing paths. Similarly, the work in [130] also employs CNN technology to tackle handover management challenges in satellite-terrestrial networks. Reference signal received power (RSRP) values are inputted into the CNN model to learn sub-optimal handover decisions.
D. Caching and Computation Offloading
The heterogeneous platforms that constitute the SAGIN architecture, including satellites, HAPs, UAVs, and ground stations, have different storage capacities, processing power, and network bandwidth. This dynamic environment requires smart strategies for caching content and offloading computation to maintain application responsiveness while optimizing network efficiency. For instance, hierarchical computing and offloading must be considered to account for the different storage and computation capabilities of the SAGIN elements in addition to the locations of users and the reliability of the wireless environment [352]. Traditional caching and offloading methods frequently do not meet the needs of SAGINs, which require collaboration and organization of caching and offloading decisions across different layers. Fixed caching strategies have difficulty adjusting to changing traffic patterns, resulting in unnecessary data transfers and higher latency [353], [354]. Centralized data offloading techniques lead to congestion and bottlenecks [355]. In addition, constraints in onboard storage and processing capabilities for platforms such as UAVs require resource-conscious offloading approaches. AI becomes a potent solution for providing collaborative caching and computation offloading schemes to SAGINs. AI-based caching and computation offloading methods can analyze network patterns in real time, allowing for proactive resource allocation and adaptive adjustments. These algorithms can forecast user needs and content popularity, strategically storing data throughout the different network layers to shorten retrieval times and decrease backhaul traffic. This may entail placing frequently accessed data nearer to users on platforms such as UAVs or edge servers, thereby reducing latency and enhancing user satisfaction. Additionally, AI models can predict computational demands and preemptively transfer tasks to the most appropriate platforms based on their available resources and prevailing network conditions.
Many studies in the literature use DRL to optimize caching and computation offloading in SAGINs. For instance, the authors in [338] tackle the issue of optimizing task offloading decisions made by UAVs in satellite-UAVs-assisted IoT networks. The proposed scheme applies the multi-agent actor-critic approach to empower UAVs to learn efficient offloading strategies, maximizing task completion while minimizing energy usage. Another study [339] discusses the issue of enhancing energy-efficient task offloading from distant IoT settings to satellites or UAVs by using federated DQNs for immediate offloading decisions and reducing energy usage while considering task processing delay. A comparable approach based on deep Q-learning is utilized in another work [342] to tackle the problem of maximizing utility efficiency in hybrid offloading for IoT devices within satellite-UAV-assisted networks. In addition to DRLs, policy-based DRL methods are also extensively used in the literature. For instance, the authors in [337] address the issue of task offloading to satellites and UAVs in hierarchical SAGINs, employing multi-agent DDPG networks to minimize energy consumption and task execution delay for IoT devices. Another study [340] addresses the challenge of task offloading in HAPs-UAVs-terrestrial networks, utilizing a multi-agent DDPG approach to maximize service satisfaction while minimizing energy consumption for IoT in underprivileged areas. Another publication [341] confronts the challenge of maximizing the sum rate of terrestrial users in dynamic satellite- HAPs network environments by using multi-agent PPO to optimize computation offloading and resource allocation, resulting in improved performance and battery backup management.
Supervised learning models can also be beneficial in optimizing the decision-making for collaborative caching and computation offloading in SAGINs. For instance, the work in [336] deals with the problem of energy-efficient computation offloading in satellite-UAV-based IoT systems, utilizing the LSTM model to enhance task success rate and energy dynamics for efficient offloading. Another study [302] addresses the optimization of computation offloading in satellite-UAV-integrated IoT systems, employing LSTM to boost the completion rate of tasks delivered to edge devices while simultaneously improving the overall processing capacity of the system. The research in [335] focuses on decreasing content access delay in HAPs-assisted multi-UAV networks by predicting content popularity. In this paper, DNNs are used to predict popular content accurately for caching, thereby minimizing content access delay or reduction.
E. Summary and Lessons Learnt
SAGINs pose a range of complex challenges due to their diverse composition. They consist of satellites, HAPs, UAVs, and terrestrial networks, requiring dynamic optimization across multiple layers. These challenges involve dealing with a constantly changing topology influenced by the mobility of aerial platforms, handling potential link fluctuations, and managing constraints like limited storage, processing power, and network bandwidth. Quick decision-making is crucial in this scenario, as network conditions and user requirements can change rapidly. This calls for algorithms capable of learning and adapting in real time.
AI-based solutions, including DRL and supervised learning methods, are becoming increasingly important. Traditional approaches face challenges in dealing with the complex dynamics of SAGINs, making AI essential for improving network performance and user experience. Additionally, it becomes clear that joint optimization is significant – isolated optimization of individual SAGIN components is not enough. Optimizing topology, scheduling, resource allocation, routing, and mobility improves overall performance. Real-time learning and adaptation become fundamental principles, with AI algorithms constantly learning from network data to adapt strategies in response to changing conditions. Moreover, considering the limited resources on aerial platforms, efficient solutions are essential for caching, computation offloading, and scheduling. Acknowledging the diversity of network components, with optimization algorithms customized to accommodate different capabilities and constraints, is crucial.
Several important areas for research and development drive the advancement of SAGINs. This includes a focus on enhancing AI algorithms by exploring more advanced DRL and supervised learning techniques tailored specifically for SAGINs. For instance, the sheer volume of data inherent in SAGIN mobility management poses a considerable obstacle to traditional techniques. Tensor-based approaches and learning algorithms represent a compelling direction, potentially enabling data compression with techniques like Tucker decomposition and accelerating convergence through learning algorithms. However, bridging the gap between intelligent decision-making and its practical application in multi-layered, large-scale SAGINs with aerial components remains an open research question. Further exploration and experimentation in this area is crucial. Furthermore, harnessing the potential of emerging technologies like edge computing and network function, virtualization offers promising avenues for improved performance and enhanced flexibility in SAGIN mobility management. Standardization efforts to establish protocols are crucial for seamless collaboration between different SAGIN components. Addressing security and privacy concerns is vital for robust protection against vulnerabilities while safeguarding user privacy within AI-powered SAGINs.
Finally, how to determine the hyper-parameter settings under different neural networks in different SAGIN scenarios is a crucial question. Hyperparameter tuning is influential in any AI, but it is especially intricate in RL, where the algorithm must learn through trial and error in a dynamic environment. While relying on references to previous works can provide some insights, it is often insufficient due to the unique characteristics and dynamics of each SAGIN environment. Therefore, it is essential to explore adaptive methods for hyperparameter tuning that can adjust to varying conditions and optimize the performance of AI algorithms in real time.
Leveraging SAGINs for AI: Applications and Techniques
The evolution of SAGINs and AI are intrinsically linked. While AI algorithms are crucial for tackling current limitations in SAGIN deployment and functionality, effectively managing these complex networks relies on intelligent orchestration and optimization tools. Conversely, SAGINs provide a unique platform for AI advancement. Their distributed architecture facilitates edge computing and data aggregation capabilities, fostering wider device connectivity and richer data collection. This, in turn, fuels cutting-edge data analytics and AI development, further empowering SAGINs for improved performance and adaptability. In this section, we shed light on the synergy between AI and SAGINs while focusing mainly on how SAGINs can accelerate the development of efficient AI algorithms. Specifically, we focus on FL applications and the benefits of SAGINs in allowing more devices to contribute to FL. Furthermore, we highlight some wireless techniques for optimized AI systems, i.e., analog over-the-air computation and digital RRM optimized for AI. Finally, we highlight the main SAGIN-enabled new applications relying on AI, such as autonomous cars and vehicular networks, IoT, VR/AR, digital twins, and semantic communications. Fig. 7 illustrates a SAGIN incorporating the previously mentioned applications to serve user demands in different environments.
SAGIN-aided AI applications: Internet of Things, Autonomous vehicles, Virtual/Augmented reality, Digital twins, and Semantic communications.
A. Motivation
Terrestrial networks can provide the necessary infrastructure for AI development, such as high-speed Internet access, fast transfer to reliable data storage capabilities, and access to computer processing power. Additionally, they can provide secure connections to help protect sensitive data from malicious attacks. These networks are also cost-effective solutions that allow businesses to save money on expensive hardware investments while providing flexibility for scaling or expanding operations. However, with the increasing popularity of mobile devices and the exponential growth of IoT applications, terrestrial networks are heavily overloaded with the amount of traffic generated. Furthermore, the widespread adoption of new applications such as VR/AR is burdening the capacity of terrestrial wireless networks. Such limitations add more constraints and halt the full benefit of terrestrial wireless networks for AI applications.
SAGINs, on the other hand, hold immense promise in augmenting the capabilities of AI systems and overcoming the limitations of terrestrial networks. Here, we delve into the benefits these networks bring to the field of AI:
Enhanced Data Accessibility: SAGINs provide AI developers with access to extensive real-world and near real-time data from many sources. This data diversity enables the construction of more complex and diverse datasets, enriching AI training. AI algorithms, thus exposed to various scenarios and situations, become better equipped to identify patterns and make informed decisions, even in challenging or unfamiliar environments.
Scalability without Boundaries: Unlike traditional terrestrial networks, SAGINs offer unparalleled scalability. These networks do not depend on physical infrastructure or wiring, making adding new nodes and expanding coverage relatively easy. This scalability is essential as AI applications grow in complexity and data demands.
Advanced Problem Solving: SAGINs enable AI systems to work collaboratively by sharing data and resources. This collaborative approach allows AI systems to tackle complex problems from different angles and leverage each other’s strengths. The result is more efficient problem-solving and higher-quality outcomes.
Near Real-Time Communication: SAGINs facilitate faster and more efficient data transmission while considering network congestion and hotspots. This capability enables near real-time communication between interconnected AI systems, a critical requirement for applications demanding timely decision-making and responsiveness.
Innovation Catalyst: SAGINs serve as catalysts for innovation, particularly in AI-driven technologies like smart cities and autonomous vehicles. These innovations heavily rely on AI capabilities for their operation and can benefit from the collaborative power and extensive data access facilitated by SAGINs.
Privacy-Preserving AI Development: SAGINs can provide a secure platform for testing new AI technologies without compromising privacy or overloading terrestrial networks. Developers can explore AI’s potential without exposing sensitive information, making them ideal environments for developing and testing data-intensive applications like autonomous vehicles and facial recognition software.
To address these challenges, new wireless methods must be designed to carry the data needed for AI tasks. These methods aim not to deliver bits as efficiently as possible but to distill the intelligence carried within the data; thus, they are more aligned with the AI objective. Two major solutions arise among the new methods proposed to adapt SAGINs with AI objectives: analog over-the-air aggregation and RRM optimized for AI. The details of these methods are discussed in the next sections.
B. SAGINs for Edge Intelligence and Federated Learning
Considering the demands of the advancing 6G technology, there is a growing emphasis on edge intelligence to help create the vision of ubiquitous intelligence [356]. Edge intelligence represents a promising approach to harnessing intelligence by gathering, processing, and analyzing the vast data traffic generated at the network’s edge. One of the key prerequisites for these smart services is to maintain the privacy and efficiency of data exchange, which restricts the direct transmission or collection of raw data by users [356].
FL has emerged as an increasingly promising ML paradigm suited to address these challenges. FL enables multiple devices to train ML models collaboratively without sharing raw data [357]. Each device trains a local model using its respective data in FL and transmits the updated model parameters to a central server for global model aggregation. Fig. 8 illustrates the FL algorithm where a set of devices participate in the FL and send their local updates to the global server. FL stands out for its ability to provide robust security, reduce network congestion, and optimize energy consumption by integrating training capabilities throughout the network nodes [358]. Consequently, when viewed from this perspective, deploying FL at the edge represents a viable paradigm to realize the much-needed pervasive edge intelligence in the context of 6G.
Sample FL scenario. Three devices collect local data, train the global model, and send the updated local models back to the server, where aggregation occurs.
Terrestrial networks have been used conventionally to support edge FL scenarios. However, in remote regions distant from well-established communication hubs (such as rural or mountainous areas) or during critical situations (like large gatherings or military exercises), the limitations of terrestrial communication infrastructure can significantly impact the performance of edge FL. Fortunately, the SAGIN architecture introduces a new dimension of possibilities for FL, offering a range of benefits that can significantly enhance the efficiency and capabilities of FL in various applications. For instance, UAVs have the potential to supplant terrestrial BSs to deliver both communication and computing services to terrestrial devices, thus establishing what is referred to as air-ground integrated FL [359]. Several papers in the literature have considered the advantages of UAV networks in promoting efficient FL schemes. For example, the authors in [360] optimize the trajectory of a UAV, acting as a mobile orchestrator, in an FL setting where several communities with a specific task for each exist. Graph theory and convex optimization tools were used, and the simulations highlight the benefits of UAV mobility and illustrate the out-performance of the proposed solution when compared to static UAV scenarios. In [361], the authors consider an air-ground integrated FL scenario and jointly optimize the UAV location and resource allocation to minimize the terrestrial users’ energy consumption and the tradeoff between energy consumption and FL training latency. Leveraging UAVs as edge servers also offers global accessibility to diverse datasets, which enables the training of more robust and generalized FL models.
Apart from global data accessibility and remote coverage, SAGINs, with their potential for near real-time communication, can expedite the aggregation process of local model updates from edge devices, allowing for faster model convergence and decision-making. Furthermore, SAGINs often incorporate edge computing capabilities, allowing for local processing of FL tasks. This distributed approach reduces the need to transmit extensive datasets to a central server, enhancing privacy and security while minimizing bandwidth usage. SAGINs’ scalability also aligns with the distributed nature of FL. As the number of participating devices increases, SAGINs can seamlessly accommodate additional nodes, ensuring that FL can scale to meet the demands of large-scale applications without significant infrastructure changes [362]. The authors in [363] promote using HAPs to solve the issues of slow convergence and high communication delay due to limited client participation and multi-hop communications of FL implementations in terrestrial networks. Thanks to the high altitude and size of HAPs, it allows the participation of more devices with LoS links, and it has powerful computational capabilities to act as the server for local update aggregation. In [363], the authors develop a joint client selection and resource allocation algorithm to minimize the FL delay and a communication and computation resource-aware algorithm to achieve the target FL accuracy.
Furthermore, SAGINs align with the security and privacy-preserving paradigm envisioned by FL. For instance, satellite and HAPs communication technologies provide an additional layer of security due to their inherent physical barriers, making it difficult for hackers or malicious actors to access sensitive information within the network itself [364]. Incorporating SAGINs into the FL ecosystem enhances the framework’s adaptability, scalability, and reach. This synergy empowers FL to address a broader spectrum of applications, ranging from space exploration and environmental monitoring to smart cities and healthcare, by leveraging the unique capabilities of space, aerial, and ground networks. Several papers in the literature have considered the synergy between different components of SAGINs and FL [362], [365], [366], [367]. Yet remaining challenges must be addressed to achieve the full benefits of SAGINs for efficient and optimized FL.
C. Analog Over-the-Air Computation
Several solutions are proposed in traditional wireless networks to mitigate the negative impact of interference. However, most of these solutions introduce significant reductions in the available resources for each user. Conversely, the analog over-the-air computation technique promotes interference by allocating the same time and frequency resources to multiple users. This causes the analog combining of the users’ signals in the air (see Fig. 9). The design of efficient precoding schemes at the transmitters allows for calculating useful functions (i.e., sum or average) from the superposed analog signals over the air. However, extracting the individual signals becomes nearly impossible for the receivers. While some traditional data transmissions cannot benefit from the analog over-the-air computation technique as they require the reception of individual signals, distributed ML and AI algorithms can benefit from it as they are mainly based on the computation of a weighted sum of the updates [368]. A typical example is the FL algorithm since the individual local models are not needed at any point, and the model aggregation step consists of transmitting multiple local models from users to the BS or server and then computing the weighted mean of these updates [369].
UAVs and HAP are sensing the signal, resulting from analog over-the-air summation of signals coming from individual devices.
While the analog over-the-air computation technique offers several benefits to distributed AI algorithms, it also introduces distinctive challenges in terms of strong demands on CSI [370], stringent synchronization requirements [371], and limited peak transmission power [372]. A notable limitation of this technique in FL is the straggler problem: the need to align local gradients on the server side. In an FL setup, the server receives a linear combination of local gradients with specific coefficients. This leads to a situation where devices with strong channel conditions must reduce their transmit power to accommodate devices with weaker channel conditions, commonly referred to as “stragglers.” This adjustment is necessary to ensure accurate aggregation. Such biasing between strong and weak users can be solved by increasing the transmit power of weak users. However, with practical transmitters and large-scale terrestrial networks, the maximum transmit power is limited, and a large difference in the path loss between nearby and far users exists.
In this context, SAGIN architectures, specifically UAVs, can help mitigate the straggler problem of analog over-the-air computation [373], [374]. UAV can, therefore, act as the server and aggregate local updates from terrestrial users. Due to the LoS nature of the aerial channels, the user-to-UAV links are generally not blocked, potentially leading to good channel conditions. Furthermore, the mobility of the UAV allows it to cover a larger area and to fly to stragglers for better services. This will improve the learning accuracy of the FL algorithm and reduce the learning performance loss. In [375], the authors design a hierarchical over-the-air computation scheme for a UAV-assisted FL system, where the UAV acts as the server to serve nearby devices and move to another location to avoid the straggler problem. A mean squared error minimization problem is formulated and solved to tune the UAV trajectory and the global aggregation coefficients. The results highlight that the proposed scheme can significantly improve complicated FL scenarios. In [365], the authors propose an over-the-air computation-based satellite FL framework in which the satellite acts as the server that collects local updates from many terrestrial users. An optimization problem is tackled to accelerate the convergence of the FL algorithm by minimizing the downlink broadcasting error and the uplink over-the-air aggregation error.
D. Radio Resource Management for AI
The widespread of ML and AI has brought several new advantages by relying on the potential of wireless communications. SAGINs, specifically, allow for enhanced data accessibility and promote global collaboration to enhance the scalability of different AI further and distributed ML frameworks. However, the problem with AI and distributed ML is that they differ from general data communications in different ways. These differences result in new constraints in terms of computational complexity, training time, training data, and more. In this context, existing data communication protocols perform poorly in satisfying AI needs, motivating the design of new RRM protocols for AI. SAGINs can be a great enabler, further enhancing the efficiency of the new AI-tolerated RRM protocols. Emerging RRM solutions prioritize optimized allocation of key radio resources like channels, power, and frequency within wireless systems. This strategic resource management aims to foster an environment that empowers AI algorithms to improve their performance, thus enhancing user experience. This is especially important for AI applications that require high bandwidth, minimal latency, and optimal use of network resources.
In traditional RRM, max-min fairness protocols usually sacrifice spectrum usage to ensure a minimum level of service for all users in the network. On the other side, users can be treated in a discriminatory manner in FL based on the importance of their data. Thus, such spectrum usage sacrifice is no longer needed for FL [368]. New RRM protocols are therefore required to recognize devices with valuable data and enough resources (battery, processing power, network bandwidth) for FL training participation, ensuring seamless training progress without overburdening resource-limited devices. According to network load and training needs, the RRM protocols can assign bandwidth and power dynamically to participating devices to optimize resource usage and reduce training time. Furthermore, the envisioned schemes must choose the best channels for communication between the central server and devices, which minimizes interference and ensures reliable data transmission for precise model updates.
Another AI-driven application, autonomous vehicles, can greatly benefit from new RRM schemes that prioritize network resources for self-driving cars, ensuring critical data such as sensor readings and control signals are transmitted with minimal delay, enabling quick and accurate decisions for safe navigation. Furthermore, as vehicles move through different network cells, new RRM protocols must be developed to facilitate seamless handovers without interrupting data flow. This ensures the continuous operation of AI algorithms for navigation and obstacle detection. AI-enabled IoT remote monitoring applications can also benefit from new RRM schemes to assign ample bandwidth and power to edge devices such as sensors and cameras used in remote monitoring, facilitating efficient data transfer for AI analysis and instantaneous decision-making. New RRM schemes can also be developed to allocate network resources primarily for data linked to critical events identified by AI algorithms, guaranteeing prompt intervention and response. Lastly, new RRM protocols are envisioned to adjust power levels according to network conditions and data requirements, prolonging the battery life of remote devices while lowering operational costs. These RRM schemes can also benefit from silent periods, in FL, for instance, to perform power transfer from server to devices, further enhancing their battery storage [368].
The introduction of SAGINs opens up new opportunities for enhanced RRM protocols that maximize the effectiveness of different AI applications, particularly when these applications involve either aerial or remote devices. Such devices may encounter challenges with limited coverage, high latency, and unreliable connections in traditional networks. However, with SAGIN, the new RRM protocols can benefit from the different SAGIN components to seamlessly transfer the connection based on the location of these devices, ensuring continuous data flow [376]. Recognizing the critical nature of real-time data from remote devices, new RRM schemes can allocate bandwidth and latency capabilities across various SAGIN communication nodes for faster AI analysis and quicker response times for crucial applications. Furthermore, by distributing traffic load among network segments, RRM protocols can optimize resource utilization and prevent congestion that could impede AI applications [377]. At the same time, utilizing SAGINs presents distinct challenges. Each segment of SAGIN has its own features and constraints, demanding flexible adjustments of the RRM schemes. UAVs, HAPs, and satellites are in constant motion, requiring real-time resource allocation and network route modifications. Moreover, all SAGIN nodes have restricted energy and computational capacity, underscoring the importance of efficient resource utilization for sustainable operation.
E. Applications
1) Autonomous Cars and Vehicular Networks
The past decade has seen significant progress in autonomous cars, often referred to as self-driving or driverless cars, with AI at the heart of their operation. AI empowers autonomous cars to perceive their surroundings using sensors like LiDAR, cameras, and radar, interpret this data to make informed decisions, and navigate complex environments without human intervention. ML algorithms enable these vehicles to learn from vast datasets and improve their driving abilities over time [378]. While AI-enabled autonomous car systems hold great promise, they also face significant challenges. Safety is one of the foremost concerns. Ensuring that these systems can reliably handle complex and unpredictable driving scenarios by efficient communication and sharing of information remains a substantial hurdle [379].
With their distinguishable features, SAGINs can provide a reliable solution to address the limitations of AI-enabled autonomous cars and vehicular networks. Non-terrestrial networks are essential for covering areas that terrestrial networks cannot reach due to geographical constraints or other obstacles [380]. Satellite networks, for instance, can extend the range of vehicular communications beyond its traditional limits, allowing autonomous vehicles and other connected cars to communicate over long distances without relying on infrastructure provided by cellular companies. This enables autonomous cars, and subsequently, the AI algorithms behind them, to operate in remote or rural areas that don’t have access to traditional terrestrial wireless services. Furthermore, SAGINs help handle unpredictable driving scenarios and reduce accidents by facilitating real-time information exchange between vehicles. Therefore, real-time traffic monitoring and management are facilitated, allowing vehicles to adapt to traffic conditions and optimize overall traffic flow.
In addition to reliable communication, SAGINs provide precise global positioning data, contributing to improved navigation and location-based services. They enable efficient data offloading, accommodating the vast amounts of data generated by autonomous vehicles and their associated sensors for advanced analytics and decision-making. For instance, the work in [196] promotes using HAPs for caching and computation offloading in ITS. As HAPSs are deployed in the stratosphere and are available to provide wide coverage and strong computational capabilities, they have been regarded as efficient enablers for coordinating terrestrial resources and storing the fundamental data associated with ITS-based applications.
Another characteristic of SAGINs is their low-latency capabilities, which are crucial for real-time data exchange and autonomous vehicle decision-making. This is especially important for safety-critical applications such as automated emergency braking systems that require fast response times. Since such decision-making algorithms are mainly AI-powered, providing access to the data necessary to make a correct prediction makes SAGINs an inarguably massive driver of the accuracy and success of these AI algorithms. Furthermore, thanks to their scalability, SAGINs can accommodate the growing number of autonomous vehicles without requiring extensive terrestrial infrastructure expansion. Furthermore, SAGINs can aid in coordinating autonomous vehicles and emergency responders in emergencies. For instance, UAVs can be sent to swiftly establish communication networks when the terrestrial network is down, and autonomous cars can exchange information through UAVs acting as relays. For instance, the work presented in [23] utilizes flying UAVs and terrestrial vehicles as relay nodes to improve vehicular communications’ coverage, throughput, and link reliability. Multiple UAVs can assist the ground vehicular network through air-to-ground communications, offering advantages over terrestrial vehicle relays [23].
Although SAGIN architectures promise to offer several advantages to accelerate autonomous cars and vehicular networks, several challenges remain to be resolved. Specifically, the impact of the severe requirements of these technologies in terms of low latency, high reliability, and high throughput must be considered in a joint communication and learning optimization framework [381]. Moreover, the impact of the quick variations in the wireless channels due to the induced mobility must be captured in the analysis of the learning convergence of the AI algorithms. Finally, resource allocation and network organization schemes must be designed to efficiently use the different SAGIN layers to optimize the performance of autonomous cars and vehicular networks.
2) Internet of Things
Integrating AI in IoT systems enables IoT devices to perform advanced tasks and deliver more intelligent and autonomous services. As IoT devices generate huge amounts of data, AI can enhance data processing, allowing devices for real-time decision-making in addition to predictive and prescriptive analytics by investigating and detecting patterns, anomalies, and correlations in the IoT data [382]. However, integrating AI with IoT poses challenges such as data privacy, model complexity, and power consumption. Furthermore, reliable communication links must be maintained to collect the IoT-generated data from the massive number of sparsely distributed devices in remote and hard-to-reach regions.
SAGINs can play a significant role in efficiently implementing AI-enabled IoT systems. SAGINs allow for greater control over the data transmission process from the IoT devices to the cloud processing platform; this is especially important if sensitive data needs to be securely transmitted across multiple nodes before reaching its destination point [383]. Additionally, many aerial solutions offer low-power consumption options ideal for battery-powered IoT devices located at remote sites or hard-to-reach places requiring minimal maintenance costs associated with recharging or replacing batteries [384], [385], [386]. Moreover, SAGINs enhance the scalability and adaptability of IoT systems by allowing a growing number of IoT devices to communicate together. This enhances further the accuracy of the AI algorithms as larger and more diverse datasets are now available. Such characteristics are mainly important for deploying smart cities that require the deployment of thousands of sensors over large regions. SAGINs also allow data fusion from various sources, such as ground sensors, aerial drones, and satellite imaging. This can enable advanced analytics and predictive modeling for sectors like agriculture, environmental monitoring, and disaster management.
As SAGIN architectures allow for a wide range of IoT devices to communicate and share data, the heterogeneity of IoT devices and sensors is one of the main barriers to efficient AI algorithms. IoT devices are mainly different in terms of computing capability, cache size, battery power, data rate, and latency requirements. Therefore, distributed AI algorithms such as FL will be biased toward devices with better characteristics as they have a higher chance to share their updates with the server. In [387], the authors characterize such bias and highlight its impact on the accuracy of the FL algorithm in a UAV-enabled network. Furthermore, they propose an efficient method of unbiasing the FL algorithm to account for each IoT device’s different requirements and capabilities. One possible solution to the massive number of heterogeneous IoT devices is to optimally use the available resources via sparsification, allowing only a subset of devices to contribute to the learning. However, the criteria for choosing the contributing devices while maintaining high accuracy and preventing bias is also challenging and must be optimally designed.
Split learning (See Fig. 10) is yet another potential scheme to mitigate the device heterogeneity issue of IoT devices as it distributes the learning vertically between different SAGIN layers [388]. Such a distributed computing approach can also effectively address resource limitations in IoT devices. This entails performing partial model training locally on the devices, with further processing being seamlessly shifted to existing ground servers or at UAVs, HAPs, and satellites if they exist. Split learning, however, requires jointly optimizing the layer selection across the different SAGIN layers, IoT device clustering and association, and efficient resource allocation to minimize the training latency and improve the convergence of the considered ML algorithm [389]. Efficient resource allocation schemes are further required to mitigate the high interference that further increases the model bias, and that is mainly caused by the massive number of IoT devices contributing to the learning and communicating with the different layers of the SAGIN [390].
Example of a split learning scenario, in which different parts of the NN are processed in distinct vertical SAGIN nodes.
3) Virtual and Augmented Reality
In VR/AR, AI tools maintain smooth virtual and real-world interactions by enabling real-time adjustments and processing. AI can also help personalize the content and create context-aware information. In addition, AI algorithms are meant to detect and minimize errors and enhance image recognition to improve performance and increase user satisfaction. In wireless networks offering VR/AR services, a sudden decline in data rate or rise in latency can impact the user experience negatively. Such interruptions in the virtual world result in breaks in presence events that can be detrimental to the users’ immersive VR/AR experience [391]. Moreover, most VR/AR applications require low latency and high resource allocation, creating a bottleneck at the air interface of wireless networks. Edge computing has been regarded as an efficient contributor to mitigating this bottleneck by reducing the transmission and processing latency significantly [392]. However, the fixed locations of the edge servers limit the resource management for VR-based applications [393]. Moreover, while 5G networks operate at high frequencies and utilize adaptable MIMO configurations and frame structures to improve throughput and minimize latency, the susceptibility of high-frequency communication links to blockages can diminish the quality of experience for VR and AR users. Offering connectivity from the sky is an efficient solution to mitigate the impact of blockages by providing alternative LoS links.
SAGINs can assist in meeting the heavy communication demands of VR/AR applications. One of the many improvements that SAGINs promise, in addition to enhancing link quality and extending coverage, is to reduce the latency and increase the data processing capabilities of terrestrial networks. Undoubtedly, the salient features of automation, flexibility, and better SNR make SAGINs a promising enabler for AI-based VR/AR applications. For instance, using UAVs is essential in reducing the latency by bringing computation and caching capabilities closer to VR/AR users [393], [394], [395]. One major challenge for UAV-assisted VR/AR systems is the tradeoff between maintaining high learning accuracy and minimal delay. As UAVs have limited computational resources, the resolution might suffer from large attenuation, thus resulting in low accuracy. Conversely, being close to the user improves the wireless channel and minimizes the propagation delay. Thus, joint optimization of delay and learning accuracy by optimizing the UAV deployment and trajectory is mandatory when populating terrestrial networks with UAVs for VR/AR applications [396]. Furthermore, joint consideration of the SAGIN wireless environment parameters and the VR/AR user-specific metrics such as user orientation, user association, and user awareness is another critical challenge for enhancing AI-enabled VR/AR systems over SAGINs.
4) Digital Twins
A digital twin is a virtual representation of a physical object, system, or process leveraging data and analytics to drive informed decision-making [397]. It combines data-driven insights and technological agility to create predictive models for business operations, product development, manufacturing processes, and customer experience management [397]. Digital twins are pivotal in enabling organizations to gain real-time visibility across the entire life cycle, from design to operation, ensuring optimized performance and risk mitigation through timely issue detection. Moreover, digital twins foster personalized experiences to develop deeper customer relationships tailored to individual preferences. Integrating AI with digital twins enhances their ability to learn, adapt, and provision valuable insights [398], rendering them powerful tools for monitoring, analyzing, and optimizing complex physical systems in manufacturing, healthcare, transportation, and smart cities. AI and ML techniques have been used with digital twins for predictive maintenance, health monitoring, fault diagnosis, adaptive control, and operation process optimization [398]. Thus, real-time sensing combined with AI technologies will be essential to realize digital twins for 6G [399].
The advancement of digital twin technology can be further propelled by integrating SAGINs, which offer extensive data accessibility [400]. This data can be used to create accurate models and simulations of physical environments, essential for creating digital twins. Leveraging satellite imagery and remote sensing technologies could provide in-depth information on terrain and land, enabling the development of robust 3D models for various environments. Additionally, the real-time data communicated over SAGINs from the embedded sensors and IoT devices empowers the development of dynamic digital twins capable of responsive adaptations to environmental shifts. Moreover, SAGINs, characterized by accelerated connection speeds that outmatch conventional terrestrial systems, are ideal for applications necessitating high-volume or frequent updates, such as real-time monitoring and control within digital twin systems [401]. Additionally, the wide coverage SAGINs provide enables more granular data capture from diverse sources, fostering the creation of intricate digital twin models.
In digital twin systems, devices will generate data and actively join the network management and optimization. While organizing terrestrial networks might be feasible with simple schemes and policies, SAGINs demand sophisticated network management and coordination strategies to optimize the performance and functionality of the various network layers. To this extent, decisions for offloading data across different layers, edge association, and resource allocation must be taken carefully based on the real-time data generated to enhance synchronization and improve the performance of such systems [402], [403]. Furthermore, the substantial computational capabilities required in digital twin systems to ensure real-time interaction and response require expanding the computing capabilities of physical devices beyond their inherent limits, leading to the adoption of split computing, an emerging technology trend in the 6G landscape. Implementing efficient split computing schemes in SAGIN architectures is yet another research gap that must be filled in the context of SAGIN-enabled digital twin frameworks. Moreover, novel physical layer security solutions, keyless transmissions, and distributed anomaly detection schemes must be designed to maintain the trustworthiness of the SAGINs supporting the digital twins. Finally, further interest must be devoted to implementing THz communications in SAGINs to support the massive information exchange between the physical and digital worlds with low latency and high reliability and benefit from THz’s precise positioning capabilities for extreme precision mapping.
5) Semantic Communications
Semantic communication surpasses simple data transmission by incorporating context and significance into the conveyed information, thanks to advancements in AI, particularly in natural language processing [404]. This allows for efficient communication by focusing on the essence of the message rather than the raw data. In traditional relaying methods like “decode-and-forward”, the relay decodes the entire message and then retransmits it, while “amplify-and-forward” boosts the signal without understanding its content [405]. The “process-and-forward” approach in semantic communications takes a different path, and is powered by AI algorithms. Instead of decoding the entire message, the relay extracts and processes only the necessary semantic meaning, resulting in a smaller, more targeted data packet for transmission. This reduces bandwidth usage and processing time, making semantic communication a powerful tool for efficient and intelligent communication.
Semantic communications in SAGINs bring numerous benefits, primarily due to the end-to-end encapsulation in terrestrial networks. However, this hinders intermediate nodes like routers from extracting and processing semantic meaning without fully decoding the message, affecting the efficiency of “process-and-forward.” In contrast, SAGINs provide more flexibility and distributed intelligence as data packets can be intercepted and processed at intermediate nodes such as UAVs, HAPs, or satellites [406]. This enables them to extract semantic meaning before forwarding the content, which is supported by AI-powered semantic interpretation frameworks. This allows for efficient extraction and transmission of only relevant semantic information with accurate interpretation while maintaining efficiency and low latency. On the other hand, this method may also pose challenges in efficiently managing communication resources and limited onboard computing capabilities. AI plays a crucial role here by developing algorithms to optimize resource allocation and processing tasks, tailoring them to the specific limitations of SAGINs. Extracting and handling semantic data also gives rise to security and privacy concerns that require robust measures for user protection and prevention of unauthorized access. AI-powered security solutions can analyze and anonymize semantic data, ensuring information integrity and user privacy.
An unsynchronized background knowledge issue arises when the transmitter and receiver lack shared context, causing misinterpretations in semantic communications. It is a specific scenario where SAGINs, empowered by AI, can be especially useful. In situations where there is a lack of synchronized background knowledge, if both parties can facilitate semantic communication, airborne BSs can interpret semantics from the received signal using the transmitter’s background knowledge, thanks to AI-based knowledge mapping techniques. Subsequently, it can encode the signal based on the receiver’s background information, lowering the burden of synchronizing background knowledge for transmission and reception and reducing semantic noise through AI-powered filtering methods. This showcases the potential of AI in bridging the gap between different knowledge bases and promoting seamless communication even in challenging scenarios.
Open Issues and Future Directions
Integrating terrestrial and non-terrestrial networks has been an area of research that has seen significant advances in recent years. With the advent of AI algorithms, this field has great potential for further development. This section explores future research directions for the interaction between AI and SAGINs.
A. Integrating SAGINs and Key 6G Enablers
It is anticipated that 6G networks will utilize terrestrial and non-terrestrial networks, thus enabling users to access services and applications from various locations [407]. Additionally, using both networks will increase network capacity and data speeds. Therefore, many potential research challenges are associated with integrating the SAGIN and key 6G enablers.
1) THz Communications
Following the successful implementation of the mmWave technology in the 5G, THz communications are envisioned as key enablers of 6G, owing to the abundant spectrum available at the THz band [408]. THz frequencies offer aerial communications with exceptionally high data rates compared to lower frequency bands due to their ultra-wide bandwidth. Thus, implementing the THz technology in SAGINs is a natural step toward achieving service continuity, ubiquitous access, network scalability, and efficient backhauling. THz aerial links exhibit reduced spectral noise, interference, and jamming compared to terrestrial links, attributed to their high attenuation and inability to penetrate the troposphere. Moreover, THz links encounter minimal molecular absorption at higher altitudes and in free space, allowing for extended communication ranges. These benefits make using THz frequencies in SAGINs a logical and practical choice.
While using THz communications in SAGINs paves the way for multiple opportunities, several challenges and open problems must be considered. First, the limited propagation range of THz waves due to the high absorption loss and sensitivity to blockage limits the use of THz communications for long-range air-to-ground links. Thus, air-to-ground communications can use sub-6 GHz and mmWave transmissions, and communications over air-to-air links can be handled over THz frequencies, allowing for higher-capacity backhaul links. Flexible spectrum-sharing techniques are therefore required to maintain suitable isolation between different network operations [409]. Second, THz propagation has a highly varying channel, i.e., its coherence time is extremely short. In the context of SAGINs, the impact of the non-stationary THz channel is even more pronounced as SAGINs components, mainly UAVs and satellites, move at high speeds, thus encountering higher propagation delays and Doppler shifts. Thus, new comprehensive channel models are required to account for such effects. These new channel models should also consider the impact of the uncertainties associated with wind, the orientation and wobbling of the antenna arrays, and the possible signal blockages by the body of the SAGIN components, calling for more multidisciplinary research efforts [17]. Finally, THz communications are characterized by pencil beam directionality, requiring accurate tuning despite the high Doppler shifts and increased speeds in THz-SAGINs. Other challenges in THz-SAGINs include initial access, computing limitations, and cross-layer power management.
Due to the lack of explicit models to extract the performance trade-offs in terms of rate, reliability, and synchronization of THz-enabled SAGINs, AI-based solutions can be exploited. Thus, collected data, including channel and QoS measurements, can be used to decide on the performance of the THz-SAGIN system. Nonetheless, the non-stationary nature of the collected data due to excessive handovers and high losses complicates the prediction and generalization of the distribution patterns. Furthermore, centralized AI algorithms fail to predict the performance of THz-SAGINs due to the distributed architecture and the low latency requirements of most of the 6G applications. Thus, multi-agent AI algorithms are preferred to collect and learn data locally [410].
Moreover, the current AI methods still require long training periods, which prevents agents from learning in real time. Performing offline training is prohibited in THz-SAGINs architectures due to the non-stationarity of available data. Thus, a real-time multi-agent learning framework is required to predict the performance of THz-SAGIN systems. Employing such a framework instead of centralized techniques allows for a specialized understanding of specific service types, mobility patterns, or subsets of users and resources. Furthermore, incorporating multi-task and meta-learning concepts ensures a more comprehensive understanding of the distribution patterns, which allows learning agents to adjust their strategies according to the specific needs while reducing the long training periods. Challenges, however, still exist in terms of meeting the low latency demands of 6G services and dealing with the complex and non-smooth performance functions. Thus, further research is required to develop novel AI methods to optimize and control the network operations of THz-SAGIN systems.
2) Optical Wireless Communications
Unleashing the potential of optical wireless communication (OWC), including free space optics (FSO) and visible light communications (VLC), is a transformative force in the evolution of 6G SAGINs. OWC offers terabit-per-second data rates surpassing radio frequency (RF)-based alternatives, providing ultra-high capacity suitable for bandwidth-intensive applications and lower latency critical for real-time applications such as remote surgery and autonomous vehicles [411]. Additionally, its enhanced energy efficiency makes it a sustainable choice for future SAGINs, consuming significantly less power than traditional wireless technologies [412]. In particular, VLC leverages light’s limited propagation range and absence of electromagnetic interference to provide inherent security advantages crucial in today’s increasing cyber threats [413], [414].
The integration of OWC with SAGINs extends coverage to under-served areas, enhances capacity by providing an additional data transmission channel, and mitigates congestion by offloading traffic from congested RF networks [415]. Despite these benefits, integrating OWC into SAGINs poses unique challenges. These challenges encompass the limited range of OWC compared to RF communication, susceptibility to weather effects like rain and fog, and the requirement for LoS, posing difficulties for mobile users [414]. Hence, dynamic channel conditions, atmospheric turbulence, and the mobility of SAGIN components necessitate real-time link optimization. Seamless RF-OWC coexistence demands intelligent coordination while addressing VLC security concerns requiring advanced physical layer security mechanisms.
AI emerges as the key to unlocking OWC’s full potential in SAGINs. Addressing mobility challenges within SAGINs and OWC-specific range, weather, and LoS limitations requires dynamic network reconfiguration, where DRL algorithms prove valuable. DRL can facilitate real-time adjustments to beam paths, transmission power, and channel allocation as SAGIN components move or undergo deployment changes, ensuring continuous connectivity. Furthermore, ML for predictive analytics can aid in achieving seamless handoffs between OWC and RF channels, predicting optimal transition times and conditions, thereby enhancing efficiency and minimizing disruptions for mobile users. AI-driven resource allocation algorithms can effectively address resource allocation and coexistence challenges. These algorithms, including GAs and PSO, can dynamically balance traffic between OWC and RF channels, optimizing network performance based on real-time considerations such as data type, user priority, and channel availability. DL models also prove beneficial in predicting interference scenarios, allowing for minimizing interference through techniques like beamforming and frequency coordination.
Integrating OWC with the SAGIN infrastructure, which faces issues like power and bandwidth limitations, can also benefit from AI solutions. ML algorithms contribute to energy-efficient communication by optimizing the energy consumption of OWC systems and adjusting transmission power levels based on predicted network demand. Additionally, AI combined with cryptography can enhance security in open environments by implementing advanced encryption algorithms and intrusion detection systems. Moreover, enhanced computation capabilities offered by various SAGIN nodes can enhance any AI-powered solution by making it possible to run in a distributed manner, focusing on optimizing each link separately. In summary, the strategic integration of DRL, ML, and DL techniques is vital for overcoming challenges in OWC-SAGIN integration to enhance adaptability, efficiency, and security in dynamic network environments.
3) Reconfigurable Intelligent Surfaces
Reconfigurable Intelligent Surface (RIS) is recognized as an appealing technology to control electromagnetic wave propagation and tailor the propagation environment for enhanced spectral efficiency, extended coverage, and robust security, all while maintaining cost-effectiveness and energy efficiency [416], [417]. Thanks to the new opportunities RIS technology provides, it has been envisioned as a promising solution for supporting SAGIN communications and meeting the ambitious goals they are promising. Simultaneously transmitting and reflecting (STAR) RISs can extend the coverage of aerial networks by assisting the transmission and reflection of superposed signals over cascaded channels and significantly reducing energy consumption [418]. RISs can also offer an energy-efficient relaying solution for low-power, small-size LEO satellite swarms deployed to maintain ubiquitous connectivity. In such scenarios, the RIS elements can be tuned to act as passive reflectors between different satellites or across the different SAGIN layers, thus maintaining low power consumption and computing overhead. Furthermore, RIS can enhance the reliability of SAGIN in-layer and cross-layer communication links by overcoming the path loss impact due to long-distance propagation by providing alternative LoS links and introducing directional beamforming. This can be achieved by efficiently tuning the passive reflective elements of the RIS to achieve the desired gains.
The path towards fully realizing the transformative potential of RIS-empowered SAGINs is paved with challenges in its deployment. This includes, for instance, the need for dynamic RIS reconfiguration due to the mobility of the satellites and UAVs, which requires sophisticated coordination schemes between the SAGIN elements in the same layer and across different layers. Furthermore, the highly dynamic nature of the SAGIN architecture results in complicated calculation processes and optimization problems. Consequently, AI techniques are expected to provide efficient tools to address the challenges of RIS-empowered SAGINs. Together with the enhanced computing and storage capabilities provided by the space, air, and ground components, along with the cooperative features facilitated by the cloudification of the radio access network, AI will empower SAGINs to optimize, organize, and heal themselves autonomously [419].
AI algorithms can optimize RIS placement and phase shift design across different SAGIN layers in the context of RIS-empowered SAGIN. Furthermore, AI can optimize cell/user/RIS association decisions and promote proactive resource allocation to enhance performance and reduce overall delay [420]. In this context, distributed ML algorithms, such as FL, are preferred over centralized algorithms as they prioritize privacy-preserving model training, reducing communication overhead and latency. Distributed algorithms allow for a wide contribution of clients from different SAGIN layers, thus enabling network-wide training across vertical layers, which delivers a generalized global model that fits the heterogeneous SAGIN components with diverse QoS requirements.
AI is mainly beneficial in accurately estimating cascaded channels at different SAGIN layers, given the large number of parameters and the dynamic nature of the SAGIN network. This allows for developing efficient joint beamforming schemes accounting for the RIS, user, and the SAGIN element [421]. Several works in the literature have considered the integration of RISs in different layers of the SAGIN architecture. However, further efforts are still required to design efficient AI algorithms for real-time optimization of the RIS deployment and the user access schemes to meet the time-varying QoS requirements while considering the impact of the complex and dynamic SAGIN channel environment.
4) Rate-Splitting Multiple Access
Rate-Splitting Multiple Access (RSMA) is a promising NOMA technique that significantly boosts the spectral efficiency and flexibility of wireless networks. By dividing messages into common and private segments, RSMA facilitates simultaneous transmission to multiple users, thereby enhancing overall network capacity and user fairness [422]. This method overcomes the constraints of traditional orthogonal multiple access (OMA) by improving interference management and resource utilization. RSMA’s efficiency, flexibility, and robustness make it particularly suitable for addressing satellite communication challenges such as multibeam transmissions, frequency reuse, high interference, power constraints, large path loss, imperfect CSI, and high Doppler effects [423]. In UAV communications, RSMA can lower energy consumption by enabling UAVs to serve multiple ground users at once. However, implementing RSMA for UAVs introduces complex joint cross-layer optimization challenges, including UAV positioning, transmit power allocation, bandwidth management, and RSMA configuration, all while considering CSI availability and traffic patterns. Furthermore, RSMA and UAV positioning can be optimized using methods that predict cellular traffic based on historical data, further minimizing power consumption. In SAGINs, aggressive frequency reuse results in significant interference across different network layers, necessitating efficient multi-antenna and multiple access strategies. RSMA has demonstrated notable performance gains over traditional transmission strategies such as space-division multiple access (SDMA) and NOMA [424].
As an alternative to traditional convex optimization methods, AI and ML can address the challenges of RSMA-empowered SAGINs on various fronts. AI can optimize resource allocation, power distribution, task offloading, and beamforming design. Furthermore, RSMA can improve the performance of distributed AI algorithms in SAGINs, such as FL [425]. Unlike conventional FL, which typically relies on TDMA for uplink transmission, RSMA enables multiple users to share the uplink channel simultaneously for information upload. Integrating RSMA into the FL framework can reduce aggregation latency while maintaining high model training quality. The advantages of combining RSMA and AI further extend beyond the physical layer to include cross-layer design, such as optimizing network orchestration, to effectively boost wireless communication in SAGINs.
5) Integrated Sensing and Communication
Integrated Sensing and Communication (ISAC) is an emerging paradigm that merges communication and sensing functionalities within a single system [426]. SAGIN’s multilayered architecture, comprising satellite, aerial, and ground networks, significantly enhances ISAC by merging communication and sensing functionalities across its diverse platforms, each bringing unique advantages. For instance, UAVs and HAPs can simultaneously provide communication services and perform crucial sensing tasks such as environmental monitoring and surveillance, vital for rapid response scenarios like disaster relief and border security. However, integrating these functionalities presents challenges due to the varying communication performances across different networks.
Key challenges in ISAC within SAGIN include diverse communication performances, where satellite networks exhibit higher latency than ground networks, leading to inconsistent user experiences during handovers. Traditional approaches often involve separate designs for radar sensing and communication, increasing costs and reducing efficiency. Additionally, satellite networks face issues like signal fading and wider beams due to greater distances, which can reduce perception accuracy compared to ground and high-altitude networks. These challenges necessitate advanced solutions to fully harness the potential of ISAC within SAGIN.
AI can play a crucial role in overcoming the challenges of integrating ISAC within the SAGIN infrastructure. Various ML and DL architectures can create efficient waveforms that serve communication and sensing needs, allocate resources more effectively, and reduce interference - all handled by a joint model, ensuring both tasks perform well simultaneously. Additionally, AI can improve perception accuracy by using available data and advanced algorithms to address issues like signal fading and wide beams with attention to scenario-specific details. With its ability to adapt and make real-time decisions, AI can help ISAC systems within SAGIN dynamically optimize their performance, delivering robust and reliable solutions for various applications [427].
B. Generative Adversarial Networks for SAGIN
Channel estimation poses a significant challenge in deploying SAGINs because of the varied movement patterns observed by different nodes, the differing propagation environments among various SAGIN layers, and the diverse types of noise and disruptions encountered within each segment. Furthermore, compared to terrestrial channels, SAGIN channels include more model parameters, such as the 3D location and dynamic orientation of the SAGIN node [241]. Due to this, traditional channel estimation techniques like ray tracing and geographically estimated statistical models are not enough in SAGIN channel modeling. Data-driven methods can be employed to address the challenges of SAGIN channel estimation. ML algorithms present a viable option for achieving precise channel estimation in SAGINs in a centralized or distributed fashion. Centralized ML encounters challenges concerning intricacy, high power demands, and data privacy, and has not been adapted to the distributed characteristic of the SAGIN architecture. On the other side, distributed ML leverages local data at each SAGIN node for decentralized channel estimation, which might not be adequate to encompass all exceptional environmental circumstances, thus necessitating frequent channel estimation [245].
GANs are a promising strategy for effective distributed channel estimation [428]. Specifically, GANs are employed to produce data that closely resembles real-world scenarios, capturing the behaviors of channels and creating new samples within the dataset that represent various channel conditions [419]. By leveraging GANs, SAGIN can enhance channel data availability, improve training efficiency, and develop robust channel models adapting to diverse environmental conditions across space, air, and ground domains. Fig. 11 illustrates a SAGIN where UAVs, HAPSs, and satellites use GAN to generate data from channel estimation. Due to the ability of a generative model to understand the channel model’s application scope through the dataset’s temporal-spatial information during training, it offers a more effective learning framework in contrast to a discriminative method [429]. However, despite its potential benefits, utilizing GANs for channel estimation in SAGIN remains an active area of research. Key aspects such as the dataset size needed for precise sample generation, the adaptability of GANs in SAGIN, and the necessary resources for integrating GANs in SAGIN have yet to be comprehensively explored.
UAVs, HAP, and satellite using GANs to dynamically generate the synthesized channel data and share it with neighboring communication nodes to aid with the collaborative channel learning process.
Moreover, most of the previous research on utilizing GANs for channel estimation and modeling in SAGIN has focused mainly on local GANs models [240], and the literature still lacks fully distributed generative learning frameworks for addressing the challenges of data-driven SAGIN channel modeling. To bridge this gap, the development of fully distributed cooperative generative learning models is essential for accurately characterizing the SAGIN link environment. Specifically, multiple SAGIN nodes can cooperate to create the whole channel model by learning a standalone channel model by each node via GANs and then sharing synthetic channel samples generated from the local channel model with other agents. This fully decentralized approach does not require a controller and can accommodate different types of NNs while being resilient to local training errors. Furthermore, it avoids revealing the real measured data or the trained channel while allowing SAGIN nodes to learn cooperatively from each other’s datasets.
C. Security Guarantees
The heterogeneous and multi-dimensional nature of SAGINs brings new trust and safety considerations that must be addressed. Specifically, the wireless signal in SAGIN propagates in free space across different layers. In other words, not only legitimate users can receive the information, but malicious users can also extract secure information from the wireless signals. Furthermore, due to the distributed characteristics of SAGIN and multi-hop transmission, intrinsic trust and data reliability issues may arise in each layer or across layers. Thus, security and privacy have become essential to determine if the SAGIN can continue to evolve healthily.
1) Quantum-Empowered FL and Quantum AI
One of the main methods to enhance SAGIN’s security is AI. AI algorithms hold the potential to proactively detect and mitigate threats before they materialize, enhancing overall network resilience. In the context of SAGINs, specific AI paradigms like federated and swarm learning offer compelling benefits. These distributed approaches enable secure and privacy-preserving data analysis directly on edge devices, minimizing communication overhead and safeguarding sensitive information while contributing to robust threat detection and efficient security solutions. Although FL improves privacy, it still faces several issues. First, due to the large size and complexity of SAGINs, an increasing number of devices with limited capabilities are required to contribute to the FL across different segments. Second, due to the long-distance transmissions that wireless signals need to propagate in SAGINs, malicious nodes have higher chances of obtaining the model gradients and inferring the user’s private information.
Quantum computing and quantum communications are receiving increased attention due to advancements in quantum mechanics, as they offer solutions for addressing privacy leakage and ensuring secure transmissions [430]. Quantum communication harnesses the distinct attributes of quantum mechanics like superposition and entanglement to enable the secure transmission of information, ensuring inherent safeguards against interception and tampering. In particular, quantum teleportation, a quantum communication protocol for quantum information transfer, can be leveraged to establish secure communications in SAGIN between users and the FL server. Quantum-empowered FL can, therefore, mitigate the privacy concerns of traditional FL, facilitating its widespread adoption in SAGINs [431]. Despite Quantum-empowered FL’s great advantages for SAGINs, it still faces several challenges. For instance, existing noise and decoherence can significantly affect quantum communications, limiting the communication range and resulting in unreliable FL learning outcomes. A key feature the SAGIN architecture can provide is the abundant existence of nodes in the same layer or across layers. These can act as quantum relays due to the maneuverability and LoS transmission capabilities [432]. A potential research direction would be to study and optimize multi-hop empowered FL for secure and trustworthy SAGINs.
In addition to quantum communications, the coupling of quantum coding with AI techniques, quantum AI, is yet another technique to enhance the security of SAGINs. For instance, quantum key distribution can be used to detect eavesdroppers trying to access the transmission more accurately and rapidly as the network dynamically adapts and learns in real time. Quantum NNs, which also utilize quantum processing, are another efficient tool that can be exploited to increase the learning rate of different AI algorithms in SAGINs. However, the research in this area is still not mature, and several research directions can still be exploited to enable the fast utilization of quantum AI for SAGINs.
2) AI-Oriented Authentication
Finally, AI can also contribute to enhancing the authentication of SAGINs. The authentication process in SAGINs raises questions that should be carefully considered to maintain secure and trustworthy networks. First, the long communication range significantly increases the propagation delay, reducing the QoS of latency-sensitive scenarios. Further delay is caused by the processing of conventional cryptographic authentication methods, which requires increased communication and computation overhead to cope with the new security requirements. Furthermore, these crypto-based authentication schemes verify the legitimacy only at the beginning and are therefore susceptible to different attacks. As such, new mechanisms are required for real-time and transparent security provisioning.
Continuous authentication mechanisms have emerged recently to fulfill the security requirements and mitigate the limitations of conventional mechanisms [433]. Typically, continuous schemes verify the legitimacy of users based on behavior and physiological biometrics, which are user-specific. Due to the increased storage and computational resources of SAGINs, continuous authentication can be effectively implemented while relying on collecting and storing massive amounts of data. AI can be introduced here to extract unique user features from the collected data. Specifically, spatial-temporal features such as position, Doppler shift, and traffic volume can be used in the continuous authentication phase driven by the potential of SAGIN to provide vertical applications supported by heterogeneous types of devices. Therefore, AI techniques like DL, supervised and unsupervised learning, and RL are key to maintaining trust while adapting to the dynamic environment.
Although AI-oriented authentication has shown great potential in SAGINs, several research gaps remain to be investigated. For instance, dedicated work should be focused on optimizing the allocation of authentication tasks between the different layers of the SAGIN architecture while considering both communication and computation resources scarcity in addition to latency constraints. Furthermore, novel AI-based authentication algorithms must be designed to allow for simple feature extraction processes while satisfying the low computation and communication requirements of SAGINs. Furthermore, Blockchain technology can be leveraged to enhance the intelligence of the authentication algorithms further. With AI, Blockchain can create a distributed and tamper-proof ledger system for storing user credentials and audit trails, improving system transparency and accountability. This integration can also enable efficient revocation of compromised credentials and facilitate dynamic trust management between entities within the SAGIN network. Ultimately, the convergence of AI and Blockchain holds immense promise for revolutionizing authentication in SAGINs, paving the way for secure and trustworthy networks that cater to the ever-growing diverse demands of future communication systems. By addressing the critical research gaps and fostering continued innovation in this domain, we can unlock the full potential of AI-powered authentication for robust and resilient SAGIN infrastructures.
D. Performance Trade-Offs in AI-Enabled SAGINs
Integrating AI into SAGINs presents a classical trade-off between performance enhancement and increased system complexity. SAGIN networks are often unpredictable and intricate, prompting critical questions: What constitutes an effective AI algorithm for SAGINs? Should it be general or task-specific? Although AI algorithms have significant potential for optimizing network operations, resource allocation, and overall performance, they typically entail higher computational demands and increased energy consumption. This trade-off is especially pertinent in SAGINs, where resources on aerial platforms like UAVs and HAPSs are limited. Future research should prioritize developing AI algorithms specifically tailored for the resource-constrained environments of SAGINs to address this challenge. This includes exploring lightweight AI models, efficient training techniques, and optimized inference mechanisms to minimize computational overhead, time, and energy consumption. Future research can also investigate techniques for compressing AI models and reducing their complexity without compromising performance.
Another promising approach is to exploit the distributed nature of SAGINs by employing decentralized AI algorithms. This strategy involves distributing the computational load across multiple nodes in the network, thereby reducing the burden on individual devices and enhancing overall efficiency. For instance, federated learning enables collaborative model training across distributed devices while preserving data privacy. Additionally, future research can investigate hardware acceleration techniques, such as specialized AI chips or field-programmable gate arrays (FPGAs), to accelerate AI computations in SAGINs. These hardware solutions can significantly enhance AI algorithms’ performance and computation time, making them more suitable for deployment in resource-constrained environments.
Conclusion
This paper explores the dynamic relationship between SAGINs and AI, significantly impacting the forthcoming 6G paradigm. The convergence of AI and SAGINs has great potential for communication and information accessibility. However, the complex nature of SAGINs presents challenges such as dynamic optimization across multiple layers, management of changing topologies, and resource constraints. AI-based solutions like DRL and supervised and unsupervised learning methods offer promising ways to overcome these challenges by enhancing network performance by collectively optimizing topology, scheduling, resource allocation, routing, and mobility. Real-time learning is essential as AI algorithms adapt strategies based on network data in response to evolving conditions. An efficient AI-based approach is crucial for tasks such as caching, computation offloading, and scheduling due to the limited resources on aerial platforms. Additionally, the synergy between AI and SAGINs can accelerate the development of effective AI algorithms, i.e., by leveraging FL applications and exploiting SAGIN advantages to enhance contributions to FL. Wireless techniques like analog over-the-air computation and digital RRM increase the potential, creating new opportunities for applications reliant on AI and empowered by SAGINs. In conclusion, combining AI with SAGINs offers a significant opportunity to transform communication and information access by harnessing the full range of capabilities through AI-driven optimization and adaptation across diverse applications.
ACKNOWLEDGMENT
Fig. 7 was created by Heno Hwang, a scientific illustrator at King Abdullah University of Science and Technology (KAUST).