Introduction
A. Contextualization and Background
Signal processing techniques operating over centralized or distributed network architectures have been largely studied in the past, especially for Situation Awareness (SA) applications [1], [2], [3], [4]. The main application domains include Internet of Things (IoT) [5], Connected Autonomous Vehicles (CAVs) [6], [7] and Maritime Situational Awareness (MSA) [8], [9]. These applications are critical as they require sensors (hereafter generally referred as agents) monitoring and perceiving their surroundings and making informed decisions based on the perceived information. The key aspect is the cooperation among agents which enables Cooperative Positioning (CP) techniques and enhances the perception of the environment.
The Message Passing Algorithm (MPA), also known as Belief Propagation (BP) or Sum-Product Algorithm (SPA) [10], [11], is a probabilistic iterative technique which has gained a lot of interest in the field of CP [12] given its ability of linearly scaling with the number of agents [13]. MPA has been largely employed in a different number of SA frameworks, mainly addressing the Multiple Object Tracking (MOT) problem with static or mobile sensing agents [14], [15], [16], [17], [18], [19], [20], [21], [22], [23], [24], embedding or not the measurement to target association problem [25], [26], [27], [28].
B. Related Works
MPA attains optimal performances in case of linear models and Gaussian processes, where the iteratively computed marginal posterior belief converges to the exact marginal posterior distribution. When the conditions of linearity and Gaussianity are not met, particle-based MPA can be employed, although this typically results in a notable increase of computational and communication expenses (i.e., due to particles’ sharing and aggregation). Some works tried to improve performances of particle-based MPA implementations by reducing the particle degeneracy in dense and large networks [29], [30] or by auto-tuning the parameters of time-varying system models [31]. However, they did not resolve the main issue of MPA, which is related to the convergence of the beliefs.
Since MPA involves a repeated exchange of information (i.e., an iterative message passing) over a graph that is representative of the considered problem, the intrinsic cyclic structure of graphs leads the MPA‘s outcome to be only an approximation of the true marginal posterior distribution as the algorithm converges to a local optimum [32], [33], [34], [35]. Specifically, the approximation of beliefs can be considered satisfactory if the optimization problem is locally convex. To improve the performances, Neural Enhanced Belief Propagation (NEBP) have been recently proposed [36], [37], [38], [39], wherein MPA and Message Passing Neural Network (MPNN) are combined to rectify errors caused by cycles and model mismatch.
The MPNN [40], [41] is an extension of Neural Network (NN) customized to work on graph structures. Indeed, in conventional MPNN, a NN is present in each node and edge of the graph, elaborating the input features through an iterative message passing scheme. The elaborated features, i.e., node and edge embeddings, are usually taken as input to perform a specific task, like node/edge regression or classification. Given their similarity with the message passing in MPA, they have been used within the NEBP framework to address the problems of Data Association (DA) [39], CP [37] and also MOT [38], as well as with the implicit cooperative positioning framework [42]. However, NEBP approaches require performing both iterations of MPNN and MPA, increasing the already high computational time of particle-based methods. Furthermore, it has been demonstrated that in cases where sufficient training data are available, MPNN exhibit superior performance to MPA on cyclic graphs [43], while at the same time being scalable and able to learn non-linear dependencies.
C. Contribution
In this paper, we propose an MPNN solution that can be used as an efficient alternative of MPA in high-complexity problems. We made a first attempt in this direction in [44] where we employed MPNN for solving DA in sensor networks. Here we generalize the analysis by investigating the parallelism between MPA and MPNN, and comparing their performances in the challenging context of dynamic CP.
Mobile CP systems rely on a dynamic model for describing the temporal evolution of the agent locations and a graph model for modeling the inter-agent measurements. Here we propose an NN architecture that combines a Long Short-Term Memory (LSTM) for dynamics modeling and an MPNN for the computation of the marginal likelihoods. The LSTM learns the motion model of agents in time, while the iterative update of estimates based on measurements is obtained with the MPNN.
The main contributions of this paper are as follows:
definition of a theoretical framework based on the analogy between MPA and MPNN, with focus on the definition of exchanged messages, iterative processing steps and inference prediction;
proposal of an LSTM-MPNN model which completely replaces MPA for the task of CP. The model is trained using a centralized approach, while it is able to perform a completely distributed inference after deployment;
comparison with the conventional particle-based MPA, with particular focus on positioning performances and generalization properties.
D. Paper Organization
This paper is organized as follows. Section II is devoted to the description of the adopted system model. Section III first describes the MPA for CP, giving the main steps of the algorithm, and then defines the proposed LSTM-MPNN model with a one-to-one parallelism with MPA. Lastly, it provides insights on distributed inference and centralized training procedures. Section IV presents the simulation scenario and implementation details, followed by simulation results, while Section V draws the conclusions.
System Model
We denote with \begin{equation*} \mathbf {x}_{i,n} = f^{\left ({\mathbf {x}}\right)} \left ({\mathbf {x}_{i,n-1},\mathbf {w}_{i,n-1}^{\left ({\mathbf {x}}\right)} }\right), \tag{1}\end{equation*}
Each agent has access to two types of measurements: a partial and noisy observation
CP aims at estimating the states of agents from all the aggregated measurements up to time
A compact representation of the temporal evolution of the system model is reported in Fig. 1, where two different network topologies (i.e., different measurement availability) at time
Illustration of the working principle of CP, with highlighted state vectors and measurement sets for two consecutive time instants. The figure highlights the variation of the graph
Cooperative Positioning Methods
In this section, we first review the MPA Bayesian solution for CP and then we perform a one-to-one comparison with our newly proposed LSTM-MPNN model. Lastly, a description of the inference and training procedure is given.
A. MPA-Based CP
The agent’s marginal posterior probability \begin{align*} p\left ({\mathbf {x}_{0:n}|\mathbf {z}_{1:n}^{(\mathrm {A})}, \mathbf {z}_{1:n}^{(\mathrm {A2A})}}\right) \propto \prod _{i=1}^{I_{n}} p\left ({\mathbf {x}_{i,0}}\right) \prod _{n^{\prime }=1}^{n} p\left ({\mathbf {x}_{i,n^{\prime }}| \mathbf {x}_{i,n^{\prime }-1}}\right)& \\ p\left ({\mathbf {z}_{i,n^{\prime }}^{(\mathrm {A})}| \mathbf {x}_{i,n^{\prime }}}\right) \prod _{j \in \mathcal {N}_{i,n^{\prime }}} p\left ({\mathbf {z}_{j \to i, n^{\prime }}^{(\mathrm {A2A})}| \mathbf {x}_{j,n^{\prime }}, \mathbf {x}_{i,n^{\prime }}}\right).&\tag{2}\end{align*}
Prediction message: The predicted state of agent
is represented by the message:{i} where\begin{equation*} \mu _{i, \overrightarrow {n}}\left ({\mathbf {x}_{i,n}}\right) \propto \int p\left ({\mathbf {x}_{i,n}| \mathbf {x}_{i,n-1}}\right) \, \mathbf {b}_{i,n-1}^{(T)} \mathrm {d}\mathbf {x}_{i,n-1}, \tag{3}\end{equation*} View Source\begin{equation*} \mu _{i, \overrightarrow {n}}\left ({\mathbf {x}_{i,n}}\right) \propto \int p\left ({\mathbf {x}_{i,n}| \mathbf {x}_{i,n-1}}\right) \, \mathbf {b}_{i,n-1}^{(T)} \mathrm {d}\mathbf {x}_{i,n-1}, \tag{3}\end{equation*}
is the agent’s belief computed at previous time\mathbf {b}_{i,n-1}^{(T)} aftern -1 message passing steps. Note that the beliefs are initialized at time{T} = 0 asn .\mathbf {b}_{i,0}^{(T)} \triangleq p(\mathbf {x}_{i,0}) Beliefs exchange: During message passing iteration
, each agent{t \in \{1, \ldots, T\}} broadcasts{i} and receives\mathbf {b}_{i,n}^{(t-1)} from its neighbors\mathbf {b}_{j,n}^{(t-1)} . Atj \in \mathcal {N}_{i,n} = 1, the exchanged beliefs aret .\mathbf {b}_{i,n}^{(0)} = \mu _{i, \overrightarrow {n}}(\mathbf {x}_{i,n}) Measurement messages computation: During message passing iteration
, each agentt \in \{1, \ldots, T\} computes two measurements messages (one for each type of measurement) as:{i} \begin{align*}&\mu _{i, n}^{(t)\left ({\mathrm {A}}\right)}\left ({\mathbf {x}_{i,n}}\right) \triangleq p\left ({\mathbf {z}_{i,n}^{\left ({\mathrm {A}}\right)}| \mathbf {x}_{i,n}}\right), \tag{4}\\&\mu _{j \to i, n}^{(t)\left ({\mathrm {A2A}}\right)}\left ({\mathbf {x}_{i,n}}\right) \!\propto \!\! \int p\left ({\mathbf {z}_{j \to i, n}^{\left ({\mathrm {A2A}}\right)}| \mathbf {x}_{j,n}, \mathbf {x}_{i,n}}\right) \mathbf {b}_{j,n}^{(t-1)} \mathrm {d}\mathbf {x}_{j,n} \\&\;\qquad \qquad \qquad \qquad \qquad \qquad \qquad \forall j \in \mathcal {N}_{i,n}. \tag{5}\end{align*} View Source\begin{align*}&\mu _{i, n}^{(t)\left ({\mathrm {A}}\right)}\left ({\mathbf {x}_{i,n}}\right) \triangleq p\left ({\mathbf {z}_{i,n}^{\left ({\mathrm {A}}\right)}| \mathbf {x}_{i,n}}\right), \tag{4}\\&\mu _{j \to i, n}^{(t)\left ({\mathrm {A2A}}\right)}\left ({\mathbf {x}_{i,n}}\right) \!\propto \!\! \int p\left ({\mathbf {z}_{j \to i, n}^{\left ({\mathrm {A2A}}\right)}| \mathbf {x}_{j,n}, \mathbf {x}_{i,n}}\right) \mathbf {b}_{j,n}^{(t-1)} \mathrm {d}\mathbf {x}_{j,n} \\&\;\qquad \qquad \qquad \qquad \qquad \qquad \qquad \forall j \in \mathcal {N}_{i,n}. \tag{5}\end{align*}
Beliefs update: At message passing iteration
, the beliefs are updated as:{t \in \{1, \ldots, T\}} \begin{equation*} \mathbf {b}_{i,n}^{(t)} \propto \mu _{i,\overrightarrow {n}}\left ({\mathbf {x}_{i,n}}\right) \, \mu _{i, n}^{(t)\left ({\mathrm {A}}\right)}\left ({\mathbf {x}_{i,n}}\right) \prod _{j \in \mathcal {N}_{i,n}} \mu _{j \to i, n}^{(t)\left ({\mathrm {A2A}}\right)}\left ({\mathbf {x}_{i,n}}\right). \tag{6}\end{equation*} View Source\begin{equation*} \mathbf {b}_{i,n}^{(t)} \propto \mu _{i,\overrightarrow {n}}\left ({\mathbf {x}_{i,n}}\right) \, \mu _{i, n}^{(t)\left ({\mathrm {A}}\right)}\left ({\mathbf {x}_{i,n}}\right) \prod _{j \in \mathcal {N}_{i,n}} \mu _{j \to i, n}^{(t)\left ({\mathrm {A2A}}\right)}\left ({\mathbf {x}_{i,n}}\right). \tag{6}\end{equation*}
State inference: Lastly, after
message passing steps, the state of agent{T} is estimated with the MMSE estimator as:{i} \begin{equation*} \widehat {\mathbf {x}}_{i,n} =\mathbb {E}\left [{\mathbf {b}_{i,n}^{(t)}}\right]. \tag{7}\end{equation*} View Source\begin{equation*} \widehat {\mathbf {x}}_{i,n} =\mathbb {E}\left [{\mathbf {b}_{i,n}^{(t)}}\right]. \tag{7}\end{equation*}
For graphs with a tree structure, the MPA provides exact approximation of the beliefs, which coincide with the true marginal posterior pdf [10]. However, for cyclic graphs, MPA only provides a reasonably accurate approximation of the marginal posterior with a computational complexity that linearly scales with the number of agents
In comparison, MPNN holds the same time scalability [46], it has fewer parameters and it is able to catch any linear or non-linear relationship between input-output data, outperforming BP on loopy graphs if there is a sufficient amount of training data [43]. However, MPNN does not have the knowledge of features relation between time instants, i.e., each message passing iteration
B. LSTM-MPNN-Based CP
The idea behind the proposed solution is to build an equivalent DL-based model of the MPA-based CP described in Section III-A. We start describing the overall model structure, shown in Fig. 2, and then we analyze each single model block. The proposed architecture is composed of two main components, an LSTM block and an MPNN block. Adopting the same logic of the MPA at prediction step, the LSTM at time
The MPNN runs on the same physical graph of the agent network, i.e.,
The proposed MPNN model is composed of NNs for three different functions, encoding of input features
The complete proposed LSTM-MPNN algorithm is shown in Fig. 3 and it is computed by each agent
Prediction LSTM: The LSTM model in agent
predicts the node embeddings{i} at time\mathbf {v}_{i,n}^{(t)} as:n where\begin{equation*} \mathbf {v}_{i,n}^{(0)} = g_{v}^{\left ({\mathrm {LSTM}}\right)}\left ({\widehat {\mathbf {x}}_{i,n-1}}\right), \tag{8}\end{equation*} View Source\begin{equation*} \mathbf {v}_{i,n}^{(0)} = g_{v}^{\left ({\mathrm {LSTM}}\right)}\left ({\widehat {\mathbf {x}}_{i,n-1}}\right), \tag{8}\end{equation*}
is the LSTM model. Atg_{v}^{(\mathrm {LSTM})} = 0, the inference is initialized asn . Note that the output of the LSTM coincides with the initialization of the node embeddings at message passing iteration\widehat {\mathbf {x}}_{i,n-1} \triangleq \mathbb {E}[p(\mathbf {x}_{i,0})] = 0. Observing the parallelism with MPA, the belief estimatet is replaced by the state estimate\mathbf {b}_{i,n-1}^{(T)} , while the state-transition probability pdf\widehat {\mathbf {x}}_{i,n-1} is learned by the LSTM.p(\mathbf {x}_{i,n}| \mathbf {x}_{i,n-1}) Measurements encoding: At each time
, before starting the message passing, the agent and inter-agent measurements are encoded as:n The encoding is necessary to elaborate the input features, it transforms the input measurements into a hidden representation. This is important since all features within the message passing should not belong to the original feature space, but to the hidden space for data privacy reasons. At message passing iteration\begin{align*} \mathbf {z_{h}}_{i, n}^{\left ({\mathrm {A}}\right)}=&g_{v}^{\left ({\mathrm {A}}\right)}\left ({\mathbf {z}_{i, n}^{\left ({\mathrm {A}}\right)}}\right), \tag{9}\\ \mathbf {z_{h}}_{j \to i, n}^{\left ({\mathrm {A2A}}\right)}=&g_{e}^{\left ({\mathrm {A2A}}\right)}\left ({\mathbf {z}_{j \to i, n}^{\left ({\mathrm {A2A}}\right)}}\right), \quad \forall j \in \mathcal {N}_{i,n}. \tag{10}\end{align*} View Source\begin{align*} \mathbf {z_{h}}_{i, n}^{\left ({\mathrm {A}}\right)}=&g_{v}^{\left ({\mathrm {A}}\right)}\left ({\mathbf {z}_{i, n}^{\left ({\mathrm {A}}\right)}}\right), \tag{9}\\ \mathbf {z_{h}}_{j \to i, n}^{\left ({\mathrm {A2A}}\right)}=&g_{e}^{\left ({\mathrm {A2A}}\right)}\left ({\mathbf {z}_{j \to i, n}^{\left ({\mathrm {A2A}}\right)}}\right), \quad \forall j \in \mathcal {N}_{i,n}. \tag{10}\end{align*}
= 1, the edge embeddings are initialized as:t .{\mathbf {e}_{j\to i,n}^{(0)} = \mathbf {z_{h}}_{j \to i, n}^{(\mathrm {A2A})}} Node embeddings exchange: At message passing iteration
, each agent{t \in \{1, \ldots, T\}} broadcasts{i} and receives\mathbf {v}_{i,n}^{(t-1)} from its neighbors\mathbf {v}_{j,n}^{(t-1)} . Here, the analogy with MPA is straightforward if we compare the beliefs exchange with the node embeddings exchange.j \in \mathcal {N}_{i,n} Edge and node embeddings update: At message passing iteration
, the edge embeddings are updated as:{t \in \{1, \ldots, T\}} Note that (11) is the analogous of (5). Subsequently, the node embeddings are updated as:\begin{align*}&\mathbf {e}_{j\to i,n}^{(t)} =g_{e}\left ({\mathbf {e}_{j\to i,n}^{(t-1)}, \mathbf {z_{h}}_{j \to i, n}^{\left ({\mathrm {A2A}}\right)}, \mathbf {v}_{j,n}^{(t-1)}, \mathbf {v}_{i,n}^{(t-1)}}\right), \\&\; \qquad \qquad \qquad \qquad \qquad \qquad \qquad \forall j \in \mathcal {N}_{i,n}.\quad \tag{11}\end{align*} View Source\begin{align*}&\mathbf {e}_{j\to i,n}^{(t)} =g_{e}\left ({\mathbf {e}_{j\to i,n}^{(t-1)}, \mathbf {z_{h}}_{j \to i, n}^{\left ({\mathrm {A2A}}\right)}, \mathbf {v}_{j,n}^{(t-1)}, \mathbf {v}_{i,n}^{(t-1)}}\right), \\&\; \qquad \qquad \qquad \qquad \qquad \qquad \qquad \forall j \in \mathcal {N}_{i,n}.\quad \tag{11}\end{align*}
where\begin{equation*} \mathbf {v}_{i,n}^{(t)} = g_{v}\left ({\mathbf {v}_{i,n}^{(t-1)}, \mathbf {v}_{i,n}^{(0)}, \mathbf {z_{h}}_{i, n}^{\left ({\mathrm {A}}\right)}, \Phi \left ({\left \{{\mathbf {e}_{j\to i,n}^{(t)}}\right \}_{j \in \mathcal {N}_{i,n} }}\right)}\right),\quad \tag{12}\end{equation*} View Source\begin{equation*} \mathbf {v}_{i,n}^{(t)} = g_{v}\left ({\mathbf {v}_{i,n}^{(t-1)}, \mathbf {v}_{i,n}^{(0)}, \mathbf {z_{h}}_{i, n}^{\left ({\mathrm {A}}\right)}, \Phi \left ({\left \{{\mathbf {e}_{j\to i,n}^{(t)}}\right \}_{j \in \mathcal {N}_{i,n} }}\right)}\right),\quad \tag{12}\end{equation*}
is called aggregation function, i.e., a function invariant to permutations of its inputs (e.g., element-wise summation, mean, maximum). In the node embeddings update, exactly as in the beliefs update in (6), the initial node embeddings\Phi (\cdot) are used as a short-connection from the output of the LSTM, i.e., prediction step.\mathbf {v}_{i,n}^{(0)} State inference: Lastly, after
message passing steps, the regressor NN predicts the state of agent{T} as:{i} The MMSE estimator in (7) is substituted here by the node regressor\begin{equation*} \widehat {\mathbf {x}}_{i,n} = \widehat {\mathbf {x}}_{i,n}^{(T)} = g_{v}^{\left ({\mathrm {regres}}\right)}\left ({\mathbf {v}_{i,n}^{(T)}}\right). \tag{13}\end{equation*} View Source\begin{equation*} \widehat {\mathbf {x}}_{i,n} = \widehat {\mathbf {x}}_{i,n}^{(T)} = g_{v}^{\left ({\mathrm {regres}}\right)}\left ({\mathbf {v}_{i,n}^{(T)}}\right). \tag{13}\end{equation*}
which has the objective of extracting the state prediction from the compact node embeddings.g_{v}^{(\mathrm {regres})}(\cdot)
LSTM-MPNN algorithm for CP. (a) Graph representation of the agent network with agent states and measurements. (b) LSTM prediction at time
C. Inference and Training Procedure
The proposed LSTM-MPNN model for CP, as the MPA-based CP, is suited for distributed inference as each agent
To this aim, we propose a centralized training procedure in which the NNs are firstly trained to learn the CP task and then deployed in an agent network. To compute the training loss and perform back-propagation, we employ the Residual Sum of Squares (RSS) that is estimated at each timestep \begin{equation*} \mathcal {L} = \frac {1}{N}\sum _{n = 1}^{N} \frac {1}{|\mathcal {V}_{n}|} \sum _{t = 1}^{T} \sum _{i \in \mathcal {V}_{n}} \left \|{\widehat {\mathbf {x}}_{i,n}^{(t)}-\mathbf {x}_{i,n} }\right \|_{2}^{2},\tag{14}\end{equation*}
Simulation Experiments
A. Dataset
We consider a 2D scenario in which
Performance evaluation of the proposed LSTM-MPNN for CP. (a) Scenario with 16 moving agents. (b) RMSE of position and velocity over time for the non-cooperative Kalman and particle filters, the cooperative MPA and the proposed LSTM-MPNN.
For both MPA and MPNN, we consider
For the training and testing phases of the model, we used PyTorch version 1.12 and Python version 3.7.11. These operations were conducted on a workstation equipped with an Intel Xeon Silver 4210R CPU, which operates at a frequency of 2.40 GHz. The workstation was also supported by 96 GB of RAM and a Quadro RTX 6000 GPU with 24 GB of memory. For what concerns the optimizer, we used the Adam optimization algorithm [47] with an initial learning rate of 0.0001, and momentum values of 0.9 and 0.999 for
B. Model and Implementation Details
The LSTM architecture has been inspired by [48], but here we reduced its complexity such that it is constituted by two LSTM layers and a hidden output dimension, i.e., node embeddings, of 16. The complexity reduction is motivated by considering that the state estimation in CP comprises two steps (i.e., prediction and update). For the measurement encoding, update of node and edge embeddings, and state inference, we use Multi-Layer Perceptrons (MLPs) with linear layers and Gaussian Error Linear Unitss (GELUs) activation functions [49]. The complete LSTM and Multi-Layer Perceptrons (MLPs) model structures are reported in Table I.
The selected final architecture of our model was derived upon experimentation, including varying the number of layers and neurons. However, the main rationale behind the general structures is the following. First, the NN encoders
C. Simulation Results
1) Tracking Performances:
The first test aims at assessing the performances of the proposed LSTM-MPNN model and highlighting the advantages of adopting a data-driven solution. The comparison includes two non cooperative algorithms, i.e., a Kalman Filter (KF) and a Particle Filter (PF), which only use the agent state measurements
For the particle-based methods, the number of particles is set to
The results of the comparison are reported in Fig. 4, where we show a realization of the scenario (Fig. 4a) and the RMSE of the position and velocity for each timestep (Fig. 4b) (averaged over 30 simulations). Starting from non-cooperative methods, we notice that the KF is well approximated by the particle-based MPA and reaches a positioning error of 1.62 m while tracking. The cooperative MPA permits to increase the performances by reaching an RMSE on position of 89 cm at convergence. Lastly, the proposed LSTM-MPNN method outperforms all the other methods, achieving an RMSE of 21 cm on the position. Concerning the velocities, all the methods converge at about 0.05 m/s of RMSE. Apart from regime performances, an additional important aspect to consider is the model convergence. Indeed, the LSTM-MPNN method is able to converge after few timesteps, while BP-based algorithms require more time. This feature allows the LSTM-MPNN model to fast react in case of track initialization and recovery after a sudden trajectory variation as it rapidly forgets the previous estimates, updating the state knowledge through LSTM hidden states.
2) Generalization Capabilities:
This experiment compares the performances of MPA and LSTM-MPNN under different validation conditions. In particular, we test different intensities of driving process and state-measurement noises. The MPA retains inside the true value of the motion and measurement noises, while the LSTM-MPNN has been trained with noise-free driving processes and measurement models. This is done in order to prove the efficacy of the method with a full-calibrated MPA and a completely miscalibrated LSTM-MPNN.
In a first test, we consider a zero-mean Gaussian-distributed driving noise, i.e.,
Analysis of the impact of driving noise standard deviation on the position accuracy for MPA and LSTM-MPNN.
In a second test, we consider a constant motion model and a varying state-measurement noise, i.e.,
Comparison of the impact of state-measurement noise error in terms of RMSE of the position between MPA and LSTM-MPNN.
3) Impact on Different Number of Agents:
For this last assessment, we evaluate how the different number of cooperative agents affects the performances of the two methods. To this aim, in Fig. 7, we plot the RMSE on the position varying the number of connected agents
Comparison of the impact of varying number of cooperative agents in terms of RMSE of the position between MPA and LSTM-MPNN.
4) Computational Complexity:
Given the same graph structure and same number of message passing iterations between MPA and LSTM-MPNN models, the major difference in computational complexity lies in the computation of the prediction and update steps. In order to compare one-to-one the two methods, we define with
To this aim, in Fig. 8 we show the whole prediction time of an instance of agent trajectories, i.e., 16 agents moving as shown in Fig. 4a, varying
Comparison of the impact of varying number of particles
Conclusion
This paper addressed the problem of CP by proposing an innovative LSTM-MPNN model that can be considered as a promising alternative to conventional probabilistic MPA. Besides providing for the first time a one-to-one parallelism with respect to MPA, we demonstrated the improved performance of a fully DL-based model. We detailed each part of the proposed model, starting from the need of temporal-dependence solved using an LSTM block, up to the message passing structure. The MPNN runs on the same physical graph created by the network of connected agents and it is able to perform inference in a completely distributed way. Mirroring the MPA, the messages, i.e., node embeddings, are exchanged between agents until convergence. Finally, as opposed to the MMSE estimator in MPA, the state inference is carried out through a NN at the node.
We validated the proposed approach in a synthetic network of cooperative agents moving in a scenario over straight trajectories. Numerical results showed that the proposed approach is able to address the problem of CP in an efficient and effective way by outperforming particle-based MPA in a different number of aspects. First, under peak performances point of view, the LSTM-MPNN model reaches a lower RMSE on the position by a factor of 3. Second, the LSTM-MPNN model holds a much higher speed of convergence, an order of magnitude lower computational complexity. As an example, in our experiments, the dimension of the messages exchanged by the MPNN is 16, while the number of particles exchanged by the BP is 1000. Moreover, the proposed model better handles different state-measurement noises, as well as driving noises if trained on all ranges of state feature values. Finally, the LSTM-MPNN model better exploits the power of cooperation, giving a huge improvement even with small number of cooperating agents.
The value of cooperative positioning is foreseen to dramatically grow over the next several years, especially in the context of automated and connected mobility, where dense networks of agents have to handle complex and dynamic environments. It results that an effective data-driven approach is of paramount importance to enhance positioning capabilities. Our method makes a step toward this direction, by enabling distributed and efficient cooperative inference. Future developments could be implementing not only a distributed inference but also a distributed training, maintaining at the same time the agent’s local data privacy. Moreover, applications of fully DL-based methods are foreseen for the major fields of target detection and tracking.
Code Availability Statement
The GitHub repository with the dataset and the Python code for the model, training and inference is available upon request to the corresponding author.
NOTE
Open Access provided by 'Politecnico di Milano' within the CRUI CARE Agreement