Introduction
A. Context and Motivation
Satellite Communications (SatCom) have become increasingly important in recent years due to the surge in global connectivity demands. With the integration of terrestrial systems like 6G (Sixth-Generation) and the pressing need to reduce the digital divide and obtain ubiquitous coverage, SatCom plays a vital role in bridging the communication gap worldwide [1]. However, the growing traffic demand in SatCom systems presents significant challenges in effectively managing the allocation of radio resources to meet Quality-of-Service (QoS) requirements, while minimizing resource utilization [2].
Conventional SatCom systems typically employ static multi-beam configurations with fixed bandwidth and power allocations. These systems are incapable of adapting to the dynamic nature of traffic demands. As a result, resources may be wasted, while user demands may be left unfulfilled. Recognizing the temporal and spatial variations in demand, software-defined payloads have emerged, offering unprecedented flexibility and adaptability in Radio Resource Management (RRM) for SatCom [3].
Software-defined payloads have revolutionized the SatCom landscape by providing fully reconfigurable systems capable of dynamically allocating power and bandwidth resources. The advent of these payloads has driven the need for effective RRM techniques to optimize resource allocation and ensure efficient utilization. While traditional optimization-based solutions have been explored for RRM, they are often computationally cumbersome and lack the required flexibility to address the diverse and dynamic traffic patterns encountered in SatCom systems [4].
In recent years, Machine Learning (ML) algorithms have emerged as a promising alternative to conventional optimization approaches for RRM in SatCom [5], [6]. ML-based solutions not only offer the potential to adaptively learn and predict traffic patterns, but also contribute to accelerate complex RRM algorithmics and bring adaptation and flexibility to static optimization solutions. However, implementing ML algorithms on board may be problematic due to the potentially high energy budgets incompatible with satellites’ available resources.
B. Contributions
To address this problem, this paper investigates neuromorphic computing as an alternative to conventional neural network-based computing platforms to enhance the efficiency and sustainability of on-board SatCom operations [7] (see Fig. 1). Neuromorphic Processors (NPs) represent a new class of computing devices inspired by the human brain architecture and computational principles [8]. They offer unique advantages in low terms of power consumption, high parallelism, and real-time processing capabilities [9], [10]. Leveraging these benefits, NPs hold significant promise for enabling energy-efficient RRM on-board satellites.
To provide empirical evidence and benchmark the performance of NPs, we conduct experiments using Intel’s cutting-edge neuromorphic processor, Loihi 2. Benchmarking is carried out by comparing the accuracy, precision, recall, and energy efficiency of NPs with conventional platforms such as Xilinx Versal VCK5000 under varying traffic demands for the problem of RRM. Particularly noteworthy are the results obtained using Spiking Neural Networks (SNNs) implemented on the Loihi processor, demonstrating superior accuracy while reducing power consumption by more than a factor of 100 compared to the reference platform based on Convolutional Neural Networks (CNNs).
The principal contributions of this paper are summarized as follows:
We investigate the use of neuromorphic computing in the context of on-board satellite operations. Our investigation is among the first to systematically assess the impact of employing NPs in SatCom for efficient RRM, addressing the key limitations of conventional neural network-based platforms in terms of power and computational resource constraints.
Through rigorous benchmarking, we demonstrate that SNNs implemented on Intel’s advanced neuromorphic processor, Loihi 2, deliver a remarkable reduction in power consumptionexceeding a factor of 100 while maintaining superior accuracy, precision, and recall metrics when compared to the conventional CNNs implemented on Xilinx Versal VCK5000 platform.
We provide empirical data and analysis, which were previously scarce, showcasing the tangible benefits of neuromorphic computing in a real-world satellite communications setting. This includes detailed comparisons under varying traffic demands, offering insights into the scalability and adaptability of NPs for RRM problems.
The results of our experiments serve as a proof-of-concept benchmark for the deployment of neuromorphic computing in space applications, potentially revolutionizing RRM in SatCom by significantly enhancing the efficiency and sustainability of on-board operations.
Lastly, our study aids in bridging the gap between the burgeoning field of neuromorphic computing and its practical application in the space sector.
C. Related Work
1) Machine Learning for RRM in SatCom
The efficient implementation of on-board RRM is crucial for optimizing performance and ensuring seamless connectivity in SatCom systems. In recent years, ML techniques have gained significant attention for resource management in various SatCom scenarios [11], [12]. A study by [13] focused on ML-based resource management in multi-beam Geostationary Earth Orbit (GEO) satellite systems. The authors analyze different ML techniques applied to systems with power, bandwidth, and/or beamwidth flexibility and systems with beam hopping capabilities. Furthermore, reference [14] proposes a combined learning and optimization approach to address a Mixed-Integer Convex Programming (MICP) problem in satellite RRM. The problem is decomposed into classification-like tasks and power control optimization, respectively solved by dual Deep Neural Networks (DNNs) and convex optimization.
Another notable work by Deng et al. [15] focuses on resource management in next-generation Heterogeneous Satellite Networks (HSNs), and introduces an innovative framework that encourages cooperation among independent satellite systems to maximize resource utilization. Deep Reinforcement Learning (DRL) optimizes resource allocation and supports intercommunication between different satellite systems. In a related study, Ferreira et al. [16] proposed a feasible solution for real-time, single-channel resource allocation problems using DRL. Their study discretized resources before allocation, which may not be optimal for continuous resources like satellite power. Luis et al. [17] addressed this issue by exploring a DRL architecture with constant, stateful action spaces for energy allocation, avoiding the need for discretization. Liu et al. [18], [19] presented a DRL-based Dynamic Channel Allocation (DRL-DCA) algorithm for multi-beam satellite systems, achieving lower blocking probabilities than traditional algorithms. Finally, Liao et al. [6] also introduced a game model and a DRL-based bandwidth allocation framework for satellite communication scenarios, dynamically allocating bandwidth in each beam. The proposed method effectively handles time-varying traffic and large-scale communication, albeit limited to managing a single resource on the satellite.
2) Neuromorphic Computing for Communications
The application of neuromorphic processors to communications has gained significant interest in recent years as a low-power alternative to traditional systems. The work [20] explores the application of SNNs for learning and inference in battery-powered devices connected over bandwidth-constrained channels. It summarizes activity on federated learning for distributed training of SNNs [21] and the integration of neuromorphic sensing, SNNs, and pulse radio technologies for low-power remote inference [22]. The recent references [23], [24] generalized wireless neuromorphic communications to multi-access channels and introduced the concept of Neuromorphic Integrated Sensing And Communications (N-ISAC). N-ISAC utilizes a common impulse radio waveform for transmitting digital information and detecting radar targets, employing an SNN for decoding and target detection. The optimization of SNN operation balances data communications and radar sensing performance metrics, showcasing the synergies and trade-offs between these applications.
The recently presented work in [25] addresses the need for innovative spectrum monitoring techniques due to the growing Internet of Things (IoT) and spectrum congestion. It presents a novel approach using SNNs centered on image segmentation. This method outperforms traditional energy detection methods in detection efficiency and performance, and future work will aim to improve it with ML for better feature extraction and segmentation.
3) Neuromorphic Computing for SatCom
While the field of neuromorphic learning for communications is still in its infancy, there is a growing interest in its applications within the space sector due to its high energy efficiency and promising performance [26]. Both academia and industry have made great strides in developing non-space applications for neuromorphic computing, which is a milestone that the space industry can leverage as a “spin-off” technology.
The work presented in [27] discusses neuromorphic systems, which mimic biological neural networks and are ideal for low-power computing. These systems are particularly well suited for space applications because of their low power requirements and resilience to space challenges, such as resource scarcity and radiation. The article highlights the potential of neuromorphic devices to improve the efficiency and effectiveness of space missions, with previous studies demonstrating their capability in tasks such as optical flow sensing for spacecraft landing.
Recent research initiatives further highlight the increasing interest in exploring the potential benefits of neuromorphic computing. U.S National Aeronautics and Space Administration (NASA) has launched of TechEdSat-13, equipped with Intel Loihi neuromorphic processor, into Low Earth Orbit (LEO) in 2022 is a testament to their commitment to testing new capabilities for future Artificial Intelligence (AI) science and engineering applications in space [28]. Additionally, the European Space Agency (ESA) initiated an Advanced Research in Telecommunications Systems (ARTES) Future Preparation (FP) activity in 2021 [29], focusing on exploring the use of neuromorphic computing for SatCom systems, underscoring the significance of this technology within the space sector.
Our previous work, published in 2022, highlights the potential use cases and applications of neuromorphic processors for SatCom [7]. Most recently, the publication [30] highlights the energy limitations in satellite networks due to the dependence on solar energy as the only power source, as well as the lack of efficient and high-performance processors to run AI algorithms, which restrict the use of AI techniques in orbit. The authors note the need for new AI algorithms such as SNNs and specialized hardware to achieve resilient LEO satellite constellations for IoT. More specifically, the authors have discussed what is involved in highly dynamic channel prediction, spectrum sensing and classification, signal detection and demodulation, satellite and inter-satellite access network optimization, and network security, while outlining future paradigms for these mechanisms in practical networks. In this sense, Neuromorphic processors, such as Loihi [31], are, in fact, particularly well-suited for processing sparse time series data since their energy requirements are proportional to the number of emitted spikes, which occur only in the presence of relevant events.
D. Organization
The document is structured as follows. Section II provides a detailed overview of the system model, covering the problem definition and the traffic demand model. Section III discusses our benchmark approach based on CNNs. In Section IV, we describe the approach based on neuromorphic computing. Section V discusses the implementation of both models on the respective hardware. Section VI presents experimental results, showing the superiority of the neuromorphic computing approach for practical workloads. Finally, Section VII offers a detailed analysis of the results and draws some conclusions.
System Model
As illustrated in Fig. 1, the system architecture comprises a flexible, Software-Defined Radio (SDR) satellite payload that allows for adjustable bandwidth and power allocation to each beam. The management of communication resources is dynamically adapted in response to changes in traffic demand. Specifically, adopting a data-driven ML-based solution, we focus on designing ML models that take the traffic demand over the service area as input, and output an optimized resource configuration. After offline training, the ML payload controller can be deployed on board the satellite for real-time inference. This data-driven architecture offers the advantage of reduced processing times [32]. Table 1 provides an overview of the main characteristics of the setting under study.
In more detail, we consider a GEO high-throughput satellite system consisting of a single multibeam satellite that provides coverage to a wide region of Earth through
A. Problem Statement
The heterogeneous distribution of traffic demands across the satellite beams and over the satellite’s lifetime motivates the use of dynamic RRM [3]. The objective of RRM is to efficiently allocate available bandwidth and power resources to minimize the discrepancy between the offered capacity
The offered capacity \begin{equation*}C_{\tau }^{b}= {\text W}_{\tau }^{b}\cdot {\mathrm{ \kappa }}_{\tau }^{b}, \tag{1}\end{equation*}
The spectral efficiency
The CINR \begin{equation*} \gamma _{N,\tau }^{b} = \frac {P_{\tau }^{b}|h^{b}|^{2}}{I_{\tau }^{b}+N_{0} {\text W}_{\tau }^{b}}, \tag{2}\end{equation*}
\begin{equation*} |h_{\tau} ^{b}|^{2} = \frac {G_{\text {SAT}} \left ({\theta ^{b}_{\tau} }\right) G_{\text {RX,max}}}{\left ({4 \pi D_{\tau }^{b}/{\lambda }}\right)^{2}L_{\tau }^{b}}, \tag{3}\end{equation*}
Inspired by [34], our goal is to use minimal power and spectral bandwidth to match the offered capacity given in (1) to the aggregated traffic demand \begin{align*} U(P_{\tau},W_{\tau})=\beta _{0}\sum _{b=1}^{B}|C_{\tau }^{b} - R_{\tau }^{b}|- \beta _{1}\sum _{b=1}^{B}P_{\tau }^{b} - \beta _{2}\sum _{b=1}^{B}{\text W}_{\tau }^{b}, \tag{4}\end{align*}
The hyperparameters
The objective function (4) is minimized under total bandwidth constraints and total system power constraints as \begin{align*} &\underset {P_{\tau},W_{\tau} }{\mathrm {minimize}} U(P_{\tau},W_{\tau}) \\ &{~\text {s.t. }} \sum \nolimits _{b=1}^{B} P_{\tau }^{b} \leq P_{\max } \\ &\hphantom {{~\text {s.t. }} } \sum \nolimits _{b=1}^{B} {\text W}_{\tau }^{b} \leq {\text W}_{\max }.~\tag{5}\end{align*}
We assume that power and bandwidth variables
B. Traffic Model
In order to model the requested capacity
The traffic demand model in Europe depends on population density, aerial, and maritime density.
Lastly, the maritime dataset accounts for the potential demand for satellite connectivity through ship communications, which exhibit significant changes over time and location. To capture this, the simulator employs a dataset obtained from Vessel Traffic Services (VTS), comprising vessel positions and maritime traffic detected by the global Automatic Identification System (AIS) [35], [38].
To prepare the collected datasets for analysis, a pre-processing unit within the traffic simulator handles tasks such as eliminating redundant and conflicting traffic records, resolving missing information, and extracting user positions. UTs are categorized and assigned to their respective service beams based on the geographic longitudes and latitudes. The simulator also considers the limited use of FSS in large urban areas, recognizing the prevalence of alternative broadband technologies in such regions.
For modeling daily hourly traffic demands, the aeronautical data traces are pre-processed by collecting and gathering flight data for one hour. Similarly, in capturing the temporal maritime traffic demand, the maritime data is analyzed, taking into account the position of each ship’s first appearance within the covered area during each hour. This approach allows for a reasonable estimation of the current and anticipated demand within an hour [35].
C. Training Data
The per-beam demand requests one formatted in a matrix form \begin{align*} \boldsymbol {R}_{\tau } = \begin{bmatrix} r_{1,1} & \quad {\dots } & \quad r_{1,n} \\ \vdots & \quad \ddots & \quad \vdots \\ r_{m,1} & \quad {\dots } & \quad r_{m,n} \\ \end{bmatrix}\, \tag{6}\end{align*}
To generate training data, we collect labels for several realizations of the traffic matrix in (6) by running an exhaustive search method on the problem (5). Although data generation is a data-intensive task, it is performed offline. Each label represents the optimal choices
In principle, the generated labels should encompass all potential payload configurations, which consist of feasible combinations of power and bandwidth pairs per-beam derived from the sets
The time it takes to generate each sample using an exhaustive search to allocate resources optimally depends on the number of possible configurations on the satellite and the characteristics of the computer on which the data is generated. We used the High Performance Computing (HPC) facilities of the University of Luxembourg [39] to generate a large amount of data, and on average, it took more than 8 hours per 500 samples generated for the parameters defined in Section V.
One important consideration is that the larger the search space, the more complex the generation of training data based on exhaustive search becomes to the point that it may become infeasible. Although it may seem an alternative to generate the training database using traditional optimization techniques, it has been shown that most of these techniques only achieve sub-optimal performance due to the non-convex and non-linear structure of the RRM problems [3]. In that sense, despite the drawbacks and limitations of the exhaustive search, we have decided to use it for the generation of the database in order to achieve the absolute optimum resource allocation and thus improve the reliability of the data used to train the ML models. On the other hand, the whole training process is assumed offline to mitigate the inconvenience of the processing time for the training data generation.
Additionally, we have made our generated database available to the general public. To access it and obtain more information about the structure of this database, please refer to [40].
Conventional ML Benchmark
In this section, we establish a CNN-based conventional ML benchmark for resource management tasks within the scope of on-board RRM systems using the Xilinx Versal VCK5000 platform. The CNN architecture, illustrated in Figure 3, is tailored for the classification of potential payload configurations.
Benchmark approach based on convolutional neural networks (CNNs) for on-board RRM.
A. Architecture
The proposed CNN takes input matrices
Our CNN comprises of convolutional layers for feature extraction, Rectified Linear Unit (ReLU) for nonlinearity, pooling layers for dimensionality reduction, and a fully-connected layer for classification. By constraining the output layer in alignment with
The architecture is optimized for performance, capitalizing on the ability of the convolutional layer to reduce parameter count and facilitate weight sharing. This efficiency is crucial in interpreting the complex input data and enabling fast, reliable classification:
Convolutional layers conduct feature mapping via the convolution operation, maintaining spatial demand relationships to highlight input characteristics.
ReLU layers implement nonlinearity, setting negative values to zero within the feature maps.
Pooling layers condense feature dimensions while preserving salient information, thus summarizing the features extracted.
The fully-connected layer harnesses these features to discern among the classes defined by the training data.
B. CNN Training
For the CNN, we refine the network parameters
We introduce a loss function \begin{equation*} \mathcal {L}_{ \boldsymbol {R}_{\tau }, S_{\tau }}(\boldsymbol {\mu }) = \sum _{b=1}^{B} \mathscr {L}\big (S_{\tau }^{b}, \mathfrak {F}_{b}(\boldsymbol {\mu }, \boldsymbol {R}_{\tau })\big), \tag{7}\end{equation*}
Aggregating these individual losses, the cumulative training loss \begin{equation*} \mathcal {L}_{\mathcal {D}}(\boldsymbol {\mu }) = \frac {1}{|\mathcal {D}|} \sum _{(\boldsymbol {R}_{\tau }, S_{\tau }) \in \mathcal {D}} \mathcal {L}_{ \boldsymbol {R}_{\tau }, S_{\tau }}(\boldsymbol {\mu }), \tag{8}\end{equation*}
\begin{equation*} \boldsymbol {\mu }^{*} = \arg\min _{ \boldsymbol {\mu }} \mathcal {L}_{\mathcal {D}}(\boldsymbol {\mu }). \tag{9}\end{equation*}
For the classification of multi-dimensional input data, we utilize a loss function that enables probabilistic interpretation of class predictions, such as categorical cross-entropy. The loss for a given data instance is then \begin{equation*} \mathcal {L}_{ \boldsymbol {R}_{\tau }, S_{\tau }}(\boldsymbol {\mu }) = -\sum _{b=1}^{B} S_{\tau }^{b} \log (\mathfrak {F}_{b}(\boldsymbol {\mu }, \boldsymbol {R}_{\tau })). \tag{10}\end{equation*}
The CNN leverages gradient-based optimization techniques like Stochastic Gradient Descent (SGD) or Adam, which adaptively estimates lower-order moments to steer the network weights
Neuromorphic Computing for SatCom RRM
ML-based algorithms have gained popularity due to their performance and flexibility, but their practical application is hindered by the substantial computational power required for training and inference. This limitation becomes particularly evident when considering scenarios such as deploying these models on board of satellites, where the use of one or several Graphics Processing Units (GPUs), which is common for modern ML architectures, is impractical due to power restrictions. To address this challenge, one promising approach involves developing more energy-efficient versions of standard ML algorithms, e.g., by employing quantization of a model’s weights. In this work, we investigate a potentially more efficient alternative, SNNs, which draw inspiration from the low-power operation of biological brains. In the following section, we provide an introduction to SNNs, including training techniques and data encoding methods. We further overview the Loihi 2 chip developed by Intel, and discuss the hardware deployment of SNNs.
A. Spiking Neural Networks
In general, an SNN is a directed, possibly cyclic, network of spiking neurons. Each spiking neuron is a dynamic system with inputs and outputs given by sequences of spikes, or binary
For the purpose of this study, we consider SNNs with fully connected layered topologies, i.e., each neuron in one layer is connected to all of the neurons in the next. Denoting as
We consider neurons following the standard Spike Response Model (SRM) [43]. At every time-step
Following the SRM, each neuron \begin{equation*} u_{k,t}=\sum _{j\in \mathcal {N}_{\ell }}w_{k,j,\ell }\cdot (\alpha _{t} * b_{j,t})+\beta _{t} * s_{k,t}, \tag{11}\end{equation*}
\begin{equation*} s_{k,t}=\Theta (u_{k,t}-\vartheta). \tag{12}\end{equation*}
B. SNN Training
The training loss over the parameter vector
We define the loss \begin{align*} \mathcal {L}_{ \boldsymbol {r}, \boldsymbol {y}}(\boldsymbol {w}) &= \sum _{t=1}^{T} \mathcal {L}_{ \boldsymbol {r}^{t}, \boldsymbol {y}_{t}}(\boldsymbol {w}) \\ &= \sum _{t=1}^{T} \sum _{k=1}^{Z} \mathscr {L}\big (y_{k,t}, f_{k}(\boldsymbol {w}, \boldsymbol {r}^{t})\big), \tag{13}\end{align*}
The training loss \begin{equation*} \mathcal {L}_{\mathcal {D}}(\boldsymbol {w}) = \frac {1}{|\mathcal {D}|} \sum _{(\boldsymbol {r}, \boldsymbol {y}) \in \mathcal {D}} \mathcal {L}_{ \boldsymbol {r}, \boldsymbol {y}}(\boldsymbol {w}), \tag{14}\end{equation*}
\begin{equation*} \min _{ \boldsymbol {w}} \mathcal {L}_{\mathcal {D}}(\boldsymbol {w}). \tag{15}\end{equation*}
We consider a spike-rate loss, whereby the SNN is given a rate \begin{align*} \mathcal {L}_{ \boldsymbol {r}, \boldsymbol {y}}(\boldsymbol {w}) = \frac {1}{2} \sum _{t=1}^{\text T} \sum _{k \neq c} (f_{k}(\boldsymbol {w}, \boldsymbol {r}^{t}) - \rho _{F}) + (f_{c}(\boldsymbol {w}, \boldsymbol {r}^{t}) - \rho). \tag{16}\end{align*}
Problem (15) cannot be directly solved using standard gradient-based methods since the spiking mechanism (12) is not differentiable in
C. Spike Encoding
Although spiking neuron models can in principle receive data in the form of an analog input current, neuromorphic processors can typically handle data only in the form of binary inputs. Consequently, the natural signals
Illustration of the encoding techniques employed for this study. Left: rate coding. Right: encoding with a time encoding machine composed of LIF neurons.
Considering the feature matrix
Rate encoding: As seen in Fig. 4, each input value, as shown using the horizontal blue bars on the left, is encoded in the spike rate of the corresponding encoding neuron: a larger input generates a large number of spikes within a fixed encoding window time.
Time encoding machine: A TEM is a system that receives as input a (bounded) natural signal
, and outputs binary spikes. We consider a TEM model based on Leaky Integrate-and-Fire (LIF) neurons, whereby a spike is emitted when the voltage of the neuron crosses a pre-defined threshold. More specifically, the TEM operates using the following set of recursive equations$x(t)$ where\begin{align*} u_{t} &= (1 - \alpha _{u}) \cdot u_{t-1} + x_{t} \\ v_{t} &= (1 - \alpha _{v}) \cdot v_{t-1} + u_{t} \\ s_{t} &= \Theta (v_{t}-\vartheta) \\ v_{t} &= v_{t} \cdot (1-s_{t}), \tag{17}\end{align*} View Source\begin{align*} u_{t} &= (1 - \alpha _{u}) \cdot u_{t-1} + x_{t} \\ v_{t} &= (1 - \alpha _{v}) \cdot v_{t-1} + u_{t} \\ s_{t} &= \Theta (v_{t}-\vartheta) \\ v_{t} &= v_{t} \cdot (1-s_{t}), \tag{17}\end{align*}
and$u_{t}$ denote the current and voltage of the neuron at time instant$v_{t}$ ;$t$ ;$x_{t}:= x(t)$ and$1 > \alpha _{u} > 0$ ;$1 > \alpha _{v} > 0$ is the threshold of the neuron;$\vartheta $ is the spike output at instant$s_{t} \in \{0,1\}$ ; and$t$ is the Heaviside step function. In practice, a single signal$\Theta (\cdot)$ can be encoded through$x$ TEMs with varying decays$N$ and$\alpha _{u}$ .$\alpha _{t}$
Chipset Implementation
In this section, we present the implementation of the CNN model and the SNN model in hardware chipsets, namely Xilinx Versal VCK5000 and Intel Loihi2, respectively.
In selecting the Xilinx Versal VCK5000 for benchmarking, we aimed to align with the cutting-edge of satellite processing technology. The ESA SPAICE project adoption of the Versal family underscores its suitability for space applications [45], offering advanced on-board data processing and AI capabilities [32]. With the first space-grade Versal AI Core devices already in use, these processors represent the next generation of satellite hardware, offering significant enhancements in signal processing and AI inferencing. Our benchmark choice thus reflects an anticipation of future industry standards and the evolving requirements of satellite technology.
First, let us define the simulation parameters. The focus of this study is on downlink analysis for the forward link. The parameters used during the simulations are listed in Table 2. We consider two different values of bandwidth,
We focus on eight beams and six configurations per beam, and hence the number of possible payload configurations is more than 40,000 options. However, after setting
30,000 samples were generated, The data was divided into two sets, with 80% used for training and 20% used for validation.
A. CNN Model
The CNN model is implemented on the VCK5000 AI accelerator, a high-performance platform based on the Xilinx 7nm Versal ACAP architecture. It utilizes matrices as input and applies convolutional layers for feature extraction, followed by fully connected layers for classification. The CNN architecture comprises Conv2D, Maxpooling2D, and Dense layers.
The specific CNN architecture employed consists of Conv2D layers with 8 filters and a kernel size of (3,3), followed by Maxpooling2D layers with a pool size of (2,2). This is followed by additional Conv2D and Maxpooling2D layers with 4 filters and a kernel size of (3,3) and (2,2), respectively. The subsequent layers include a Flatten layer, Dense layers with 512 and 256 units and ReLU activation, and a final Dense layer with 6 units and softmax activation, representing the different payload configurations. The trainable parameters sum up to a total of 3,192,058.
To implement the CNN model, we consider using the VCK5000 Versal development card, designed to provide high-throughput AI inference and signal processing compute performance. It supports popular ML frameworks such as TensorFlow, PyTorch, and Caffe, using Python or C++ APIs. The Vitis AI framework facilitates the deployment of TensorFlow/PyTorch trained models on the VCK5000 for inference. In Table 4 we summarize the main VCK5000 card features.
B. Neuromorphic Model for RMM
As seen in Fig. 5, we consider a layered SNN with
Proposed neuromorphic solution. (a) Encoding of natural signals into spikes. (b) Spiking neural networks with a layered architecture comprising three hidden layers. The prediction is given by rate decoding, that is, by selecting the index of the readout neuron producing the most spikes.
Following the approach proposed in reference [34], training is completed using a dataset
We perform inference using SNNs as described in the previous sections on Intel’s Loihi 2 chips [47]. Loihi 2 is a research neuromorphic chip that uses asynchronous spiking neurons to implement fine-grained, event-driven, adaptive, self-modifying, parallel computations. Loihi’s first iteration was fabricated on Intel’s 14 nm process and houses 128 clusters of 1,024 artificial neurons each, for a total of 131,072 simulated neurons, which is about 130 million synapses, which is still far below the 800 trillion synapses in the human brain. As members of the Intel Neuromorphic Research Community (INRC), we were given access to Loihi 1 under the Kapoho Bay form factor (see Table 5), as well as the second iteration of the chip via Intel’s cloud services. Experimental results were obtained on Loihi 2.
The Lava software library gives access to a number of metrics of interest, of which we show a summary in Fig. 6. We can hence measure the evolution of the execution time per algorithmic step, and the total power consumed by them, as well as the activity over the various cores on the chip. Activity measurements comprise the number of synaptic operations, neuron updates, and output and input spikes per core. As can be seen, the execution time per step decreases after the first few steps. In our experiments, we exploit this by initializing the network by sending all-zeros inputs up to
Summary of the metrics measured on Loihi 2. Left: Execution time per algorithmic time-step. Middle: Evolution of the total power consumption. Right: Activity per core.
C. Evaluation Metrics
We evaluate the performance of the proposed algorithm in terms of average capacity gap. This is a measure of the capacity gap between the predicted configuration, and the resource requirements. Formally, it is defined as \begin{equation*} G = \frac {1}{B|\mathcal {D}|} \sum _{ \boldsymbol {R} \in \mathcal {D}}\sum _{b = 1}^{B} \big | C_{ \boldsymbol {R}}^{b} - Y_{ \boldsymbol {R}}^{b} \big |, \tag{18}\end{equation*}
Further metrics can be employed when comparing a neuromorphic to a conventional approach. These metrics include accuracy, but also latency and energy consumption, as we detail now.
Accuracy: Accuracy measures the ability of the RRM algorithm to choose the most appropriate configuration given the current traffic requirements. It is obtained by comparing the configuration predicted by the model to the ground truth. Comparing the accuracy achieved by both approaches allows to determine which approach performs better in effectively utilizing the available resources.
Latency: Latency is a measure of the computational efficiency and responsiveness of the algorithm under study, and a key criterion in many SatCom applications. In the case of RRM, it measures how quickly the proposed algorithm can measure a change in the traffic requirement conditions, and propose an alternative configuration.
Energy Consumption: Energy consumption is a critical metric, especially in satellite systems, wherein power resources are often limited. Evaluating the energy consumed by RRM approaches allows us to assess their suitability for deployment on board of satellites. The energy consumption of the neuromorphic solution includes all the energy spent by the chip during inference, including I/O interfaces. For the conventional approach, energy consumption relates to the power consumed by the VCK chip during inference. Comparing the energy consumed by both approaches allows to inform choices pertaining to the deployment of algorithms on board of satellite systems.
Experimental Results
In this section, we present the most outstanding numerical results.
A. Benchmarking
The CNN model was trained for 25 epochs with a batch size of 128. The training for the conventional models was carried out using the HPC facilities of the University of Luxembourg [39] (see hpc.uni.lu). The training configuration used the SGD optimizer with a learning rate of 0.01, momentum of 0.9, and Nesterov acceleration. The loss function used was the categorical cross-entropy and the model’s accuracy was used as the metric.
We determined experimentally that the set of hyperparameters providing the maximal accuracy. The final accuracy of the model on the training data corresponded to max-pooling stride
Results obtained for a CNN model with max-pooling stride
Results obtained for an SNN model with max-pooling stride
The ROC plots the True Positive Rate (TPR) against the False Positive Rate (FPR) at different classification thresholds. It provides information about the trade-off between TPR and FPR, and it is useful for selecting the optimal classification threshold. The Area Under the Curve (AUC) is the area under the ROC curve. AUC ranges between 0 and 1, where a higher value indicates a better model. Accordingly, we can observe that the CNN results for the flexible payload maintain an area close to 1 for all classes.
As for the F1-score, it measures the model’s balance between precision and recall. A high F1-score indicates a good balance between the two. Table 6 shows the obtained F1-scores. The results indicate that the CNN architecture and the training configuration used were able to effectively manage the power and bandwidth of a multibeam satellite as a function of traffic demand.
Finally, we vary the size of the traffic matrix,
B. Neuromorphic Model
We determined experimentally that the set of hyperparameters providing the maximal accuracy whilst minimizing energy expenditure and execution time corresponded to max-pooling stride
In Fig. 9, we compare results in terms of average execution time and energy expenditure per example between the neuromorphic processor and Intel’s lava-dl simulator running on an A100 GPU. Measurements on the GPU are carried using the PyJoules library [48]. It can be seen that the simulator running on GPU is slower than the execution on Loihi 2, but also six orders of magnitude more costly in terms of energy. It is to be noted that execution on GPU also allows one to perform inference over a minibatch
Comparison between execution of an SNN on Loihi 2 and simulation on an NVIDIA A100 GPU with various batch sizes. Left: Average execution time per example. Right: Energy expenditure.
In the following, we extensively study how the different hyperparameters of the system affect the metrics of interest – that is, accuracy, execution time, energy and the number of synaptic operations. In all the following figures, the red data point corresponds to the model with the hyperparameters of reference described above.
We start by comparing the two encoding techniques described in Section IV-C, i.e., rate encoding and time encoding machines. As can be seen in Fig. 10, the TEM generally provides much higher accuracy, at the cost of a slightly larger execution time and energy expenditure compared with rate encoding with
Comparison between various types of encoding techniques and number of encoding time-steps. From left to right are shown: Test accuracy, average execution time per example, average energy expenditure per example, average number of synaptic operations per example. In red are shown the parameters selected for comparison with the conventional approach.
Next, we vary the size of the exogenous inputs by considering several max-pooling strides
Comparison between various encoding max-pooling strides
In Fig. 12, we demonstrate how using smaller target rates
Comparison between various target rates
Comparison between the number of encoding time-steps
Comparison between the number of TEM threshold
We now examine how the number of encoding time-steps
Lastly, varying the threshold
Overall, we explored how a variety of hyperparameters impact the operation of the system. We found a reference set of parameters that allowed us to miminize the execution time and energy expenditure while maintaining the highest accuracy we were able to obtain. We note that we were able to further reduce the processing time, even by reducing more the input size or sparsity of the signal. We suspect this is because, at this level of sparsity, the execution time is lower-bounded by the general operation time of the chip.
C. Comparison Between Non-Neuromorphic and Neuromorphic Models
We present a comparative analysis between the results obtained from inference using SNN on Loihi 2 and the VCK5000 chip for the CNN model. Fig. 15 presents a comprehensive comparison of accuracy and power consumption for both combinations. The results clearly demonstrate that the neuromorphic model consistently outperforms the non-neuromorphic approach in terms of accuracy while drastically reducing energy consumption. In fact, the energy consumption is up to two orders of magnitude lower than the models running on the VCK5000 chip.
Comparison between the execution of an SNN on Loihi 2 and the execution of a CNN on the CPU of the VCK 5000 chip. Left: Accuracy vs. Energy Consumption. Right: Solution Time Ratio.
One of the factors contributing to the superior performance of the neuromorphic approach is the deployment of the CNN model on the VCK5000 chip. To facilitate this deployment, the CNN model must undergo quantization, which involves reducing its accuracy. This quantization process inevitably leads to a loss of accuracy compared to the original values obtained in Section VI-A. Consequently, the lower accuracy of the CNN model quantized on the VCK5000 chip is one of the reasons for the difference in performance between the two approaches.
However, the primary and more significant reason behind the higher accuracy achieved through inference using SNN on Loihi 2 can be attributed to the characteristics of the traffic matrix
The ability of the neuromorphic approach to process sparse data efficiently is a key advantage that enables it to outperform the non-neuromorphic approach. The unique design of the Loihi 2 chip allows for specialized processing of sparse data patterns, leading to optimized and accurate inferences. As a result, the neuromorphic models excel in extracting valuable information from the sparse traffic matrix
In Fig. 15, we present the results of the comparative benchmarks for latency and energy in a unified view. This two-dimensional plot highlights the key advantages offered by neuromorphic hardware, such as the Loihi chip, as compared to commercially available programmable architectures. The dashed diagonal line represents the energy-delay ratio parity line [49], with benchmark points located below and to the left of this line indicating architectures that outperform Loihi. In contrast, points located above and to the right indicate superior performance for Loihi. In all the scenarios presented here, the conventional models lie above the parity line, indicating that neuromorphic algorithms running on dedicated hardware provide clear benefits to perform RRM on board of satellites.
D. Complexity Analysis
We now describe the computational capabilities and constraints of CNN and SNN models, with direct implications for on-board satellite systems. The CNN model comprises of 3,192,058 trainable parameters, while the SNN model incorporates 395,312 for
Comparing the memory complexity used for ML models with traditional optimization methods such as exhaustive search (which was used for the generation of training data), we have that for ML models the memory required involves the storage of parameters and intermediate data for the CNN and the retention of states and synaptic weights for each of the 1,280 neurons of the SNN model on the Intel Loihi2 chip. In comparison, exhaustive search requiring considerable memory for combinatorial exploration of configurations that for our scenario making to over 40,000 options.
Power consumption between both ML-based approaches contrasts the power consumption of the Versal VCK5000 board with the average of 21 W during CNN inference with the remarkable efficiency of SNN with an average of 0.12 W on Intel Loihi2. On the other hand, although not directly measured, the power overhead of exhaustive search is presumed significant due to its intensive computation.
Latency is evaluated by the time efficiency of the CNN and SNN in responding to data inputs, with a comparative illustration in Figure 15. The exhaustive search latency is inherently higher, as it encompasses the duration to traverse all configurations to produce viable training data sets.
Conclusion
SNNs excel in processing sparse temporal data due to their reliance on spike patterns. However, they face limitations with dense datasets common across domains, hindering their universal adoption over CNNs. SNNs energy efficiency benefits rely on specialized neuromorphic hardware, yet its limited availability poses a barrier to widespread implementation. Furthermore, the learning curve associated with neuromorphic computing, its evolving frameworks, and tools contrasts with the mature ecosystem surrounding CNNs. Despite these challenges, ongoing research and technological advancements offer promise for neuromorphic computing to complement traditional methods, paving the way for more energy-efficient processing in the future.
This article presents an extensive investigation into the benefits of incorporating neuromorphic computing and SNNs for on-board radio resource management in SatCom systems. By leveraging innovative approaches, we addressed the challenge of implementing on-board RRM, comparing the performance of the proposed neuromorphic computing approach with a traditional CNN model. Our experiments demonstrate that SNNs, enabled by dedicated hardware, offer higher accuracy and significantly reduce energy consumption and latency. These remarkable results underscore the potential of neuromorphic computing and SNNs in improving RRM for SatCom, leading to better efficiency and sustainability for future SatCom systems.
To advance this research further, several avenues of investigation remain open. An important aspect is the implementation of the proposed approach in a real system, taking into account factors such as radiation tolerance, which holds great significance in the space environment. Moreover, future research could focus on optimizing the SNN architecture to achieve better performance and energy efficiency, considering the specific requirements and constraints of SatCom systems.
Although our current model is suitable for GEO satellite systems [50], we recognize the dynamic nature of LEO/MEO systems [51], [52], where the Doppler effect, rapid elevation angle changes, and other factors significantly influence the analysis. In future work we intend to extend our model to address these challenges, recognizing that channel and traffic conditions vary much more rapidly due to the higher relative velocity of LEO/MEO satellites. This will require a more complex and robust model to accommodate the highly dynamic environment.
We also wish to emphasize the generalization performance of the proposed models. Rigorous testing through cross-validation on diverse datasets simulating various operational scenarios has shown that the SNN model, in particular, exhibits strong generalization capabilities. This is evidenced by its ability to maintain high accuracy and low energy consumption when exposed to unseen data, indicating its robustness in real-world deployments. We acknowledge, however, that substantial changes in the operational environment, such as a transition from GEO to LEO/MEO systems, will necessitate the adaptation of the model. To this end, our future work will focus on enhancing the model complexity to cope with the increased dynamics of LEO/MEO systems, ensuring that the generalization capabilities extend across different orbital conditions. The planned integration of continual learning mechanisms is anticipated to bolster the model adaptability further, allowing it to update its parameters in response to evolving traffic patterns and channel conditions, thereby sustaining high performance without frequent retraining. These enhancements will be pivotal for deploying SNN-based RRM in the highly variable and demanding environment of space communications.
The findings of this study lay a solid foundation for the application of neuromorphic computing and SNNs in the field of SatCom RRM. Future investigations can build upon this work to further advance the state-of-the-art in SatCom systems, leveraging the benefits and insights gained from this comprehensive study.
ACKNOWLEDGMENT
The authors gratefully acknowledge the support of Intel Labs through the Intel Neuromorphic Research Community (INRC) and Tomas Navarro as ESA Officer. Please note that the views of the authors of this article do not necessarily reflect the views of ESA.