Introduction
Recent studies have shown that human factors, such as speeding, drunk driving, and distracted driving, have the strongest influence and are responsible for 80-90% of road traffic accidents [1]. Thus, replacing human drivers with more reliable solutions would potentially mitigate the majority of catastrophic traffic events by reducing human error. This potential in improving travel safety is one of the primary anticipations that motivates research and investments into autonomous vehicles (AVs) and advanced driver-assistance systems (ADAS) [2].
Most existing AV controllers are designed in a reactive manner, therefore focusing on the current state of other road users [3]. While this reactive design can suffice the goal of navigating AVs without collision under most circumstances, their inability to infer driver intent can lead to conservative driving strategy and affect traffic efficiency. Instead, if AVs were to predict surrounding traffic conditions based on past information [4], they would have more opportunities to proactively plan and execute safety manoeuvres, improving road safety and driving comfort while minimising conflicts with other road users.
Predicting trajectories accurately requires considering a multitude of factors that influence vehicle motion behaviour. These factors can be broadly categorised into two types: map-based and map-free. Both categories use historical trajectories of vehicles to predict their future paths, with the former also integrating HD maps. However, HD maps, while providing valuable contextual information, have started to reveal some issues in their application [5]. Similar to other AV systems, researchers have begun to investigate how to conduct trajectory predictions without depending on HD maps. Recent studies [6], [7] have shown that map-free methods can surpass map-based approaches in terms of both prediction accuracy and processing speed. Consequently, this paper exclusively concentrates on map-free methods.
Meanwhile, it should be noted that while human drivers’ behaviours tend to preserve some tendencies, but they are not deterministic and their reactions to the same driving scenario may differ at each time [8]. Factors affecting this behaviour include the driver’s eagerness to finish the trip, time of the day, weather conditions, among others [9]. More importantly, even if the driver maintains the same intention, the execution could differ in speed and pattern, resulting in different manoeuvres [10]. Thus, this study generates a stochastic multimodal prediction of vehicles’ future trajectory to ensure AVs’ safety operation.
While many studies have been proposed to predict vehicle trajectories [3], [4], [11], [12], [13], [14], [15], [16], [17], [18], [19], [20], [21], [22], [23], [24], [25], [26], there is a lack of a thorough comparison of these methods, especially for deep learning studies [3], [4], [12], [15], [16], [17], [18], [19], [20], [21], [22], [23], [24], [25], [26]. Meanwhile, although collision risks were estimated in some studies [11], [13], they were mainly used for the selection of candidate manoeuvres, which would restrict the influence of driver’s risk awareness on their driving decision.
Furthermore, in many existing research studies, spatial relations are commonly treated as fully connected, without distinguishing their individual influences. This approach significantly limits the interpretability of the model, making it challenging or even impossible to comprehend the results it produces. This lack of interpretability poses a significant obstacle to the safety certification and widespread adoption of AVs. Addressing this issue, Explainable Artificial Intelligence (XAI) provides a solution by enabling human users to better understand, interpret, and trust the outputs generated by AI-powered systems. By incorporating explainability, we can ensure that only meaningful variables contribute to the output, thereby guaranteeing that the models capture accurate causality. In order to enhance explainability, our research introduces risk-awareness in spatial relations, thus improving the overall explainability of the model.
The primary contributions of this study are outlined as follows:
We propose a novel approach to model the spatial-temporal interactions for vehicle trajectory prediction, which achieves an improvement of over 15% in prediction accuracy when comparing with the state-of-the-art model.
A unique kernel function that emphasises risk-awareness is developed to dynamically extract spatial dependencies between vehicles in the scene. The influence of this kernel function is evaluated in an ablation study.
We create an enhanced safety protection layer based on the stochastic vehicle trajectory predictions, which evolves with the prediction horizon to incorporate the increasing uncertainties in predicted positions.
A unified comparison analysis is conducted to evaluate the proposed method’s prediction accuracy, inference speed and distributional performance with state-of-the-art models.
This paper is structured into five sections. Section II briefly reviews some relevant literature, while Section III explains the adopted methodology. Afterwards, the obtained results and discussion of analysis are presented in Section IV. Finally, Section V discusses the findings and limitations of this study and suggests potential future works.
A. Related Work
Extensive research efforts have taken place over the last decade in the area of AV safety and trajectory prediction, underpinned by rapid advances in sensing technologies and the intense pace of AV deployment efforts. For the purposes of this study, we conducted a literature review for approaches based on deep-learning. An overview of reviewed literature is presented in Table 1.
B. Deep Learning Studies
Trajectory prediction is fundamentally a time-series classification or generation problem, which makes it particularly suitable for the application of deep learning techniques. Long Short-Term Memory (LSTM) methods, based on Recurrent Neural Network (RNN) architectures, have been especially prominent amongst the reviewed literature, as they are capable of extracting long-term relations amongst the various actors in their models. Existing LSTM-based models can consider fixed number of surrounding vehicles [3], [15], [16] or dynamically capture them over an occupancy grid [17].
However, although these models can implicitly infer the dependencies between vehicles, they can lose information of their relative positions. To compensate this deficiency, Deo and Trivedi [12] enhanced LSTM to improve the extraction of spatial relations. They added convolutional social pooling layers, which generate a context vector consisting of a compact representation of vehicle interactions. Instead of focusing only on vehicles, He et al. [19] used Multiple Layer Perceptron (MLP) and LSTM to develop a unifying framework that can predict trajectories of different road agents, such as vehicles, pedestrians and cyclists.
To better extract spatial relations, graph-based methods have been increasingly adopted for trajectory prediction. They denote vehicles as nodes, with their interactions represented using edges. Spatial information is then captured using Graph Convolutional Networks (GCNs) or Graph Attention Networks (GATs). For example, Zhao et al. [20] assumed a full connection between all vehicles in the scene and applied a set of two-layer GCNs to capture their spatial relations. However, these fully connected edges lead to equivalent interactions among vehicles, regardless of their respective positions, which is not realistic in the real-world driving context. To better address the different influences of interactions, Jeon et al. [21] used Edge-enhanced Graph Convolutional Networks, a variant of conventional GCNs, for the prediction of vehicle trajectories. This approach obtains a weighted adjacency matrix through the edge-enhanced attention mechanism, which calculates the edge feature matrix using two paired vehicles’ relative position and velocity. Another GCN-based study was implemented by Li et al. [23]. Unlike the previous two studies, they incorporated a 2D Temporal Convolutional layer after the graph operation. They also added a trainable graph to the fixed graph representation of the scene to mitigate performance degradation in urban traffic scenarios [22]. The temporal correlations in these studies were captured using LSTM [20], [21], [23] and Gated Recurrent Unit (GRU) [22].
While GATs have been used in several studies for pedestrian trajectory prediction, its application to vehicle trajectory prediction is limited, with only one publication identified [24]. Similar to [19], this model can also predict trajectories of other road agents by defining an attention circle and limiting potential conflicts only among the road users within this circle, the radius of which is determined by agent speeds, prediction times, and agent lengths. Additionally, the semantic map was used as an input that also included traffic rule information. Other attention mechanisms are also increasingly used in recent studies. For example, Chen et al. [28] introduce a novel non-autoregressive graph transformer that incorporates a self-attention module to address dynamic variations in social behaviour and a graph attention module to depict interactions between vehicles. This non-autoregressive approach enables the model to achieve both diverse trajectories and low prediction latency. Liao et al. developed two models, BAT [29] and MFTraj [30]. Both models integrate human features into trajectory prediction, incorporating a behaviour-aware module based on dynamic geometric graphs (DGGs) to capture the behavioural features of road users. In a recent study, the Human-Like Trajectory Prediction (HLTP) model integrates human cognition and decision-making processes through an advanced teacher-student knowledge distillation framework [32]. The “teacher” model, featuring an adaptive visual sector, simulates the visual processing capabilities of the human brain, specifically the occipital and temporal lobes. Meanwhile, the “student” model emphasises real-time interaction and decision-making, akin to the roles of the prefrontal and parietal cortices.
Another trend in existing deep learning studies is to use Temporal Convolutional Networks (TCNs) as an alternative for time-series prediction. TCNs employ casual convolutions and dilations to handle sequential data with their temporality and large receptive fields. As a variant of the convolutional neural networks, TCNs can alleviate RNN’s accumulated error problem while achieving comparable or even better performance when predicting sequential data [33]. Two TCNs-based studies were identified. The first study [25] focused on predicting vehicle’s long-term lane-changing behaviours and trajectories. The prediction accuracy was benchmarked with two traditional neural networks, Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN). The TCN-based method was found to achieve better performance when considering the prediction accuracy and computational cost. Another TCN-based model was developed by Strohbeck et al. [26]. This model was evaluated using the Argoverse Motion Forecasting Dataset and validated with the Argoverse Baseline and a CNN-based Multiple-Trajectory Prediction (MTP) model. Li et al. [27] consider both local and global features in their study. They utilise a social convolutional pooling layer to capture local interaction features between vehicles and a multi-head self-attention layer to capture global interaction features.
C. Performance Evaluation
While some studies were evaluated with different public datasets [24] or collected their private training data [17], Next Generation Simulation (NGSIM) is found to be the most commonly used dataset for trajectory prediction. Among the identified 20 deep learning-based studies, 11 of them published their prediction accuracy with the NGSIM dataset, with a prediction horizon ranging from 1s to 5s. The prediction accuracy of these models evaluated with NGSIM is summarised in Table 2.
Table 2 reveals that no single model achieves the best performance across all prediction horizons. In general, BAT, GISNet, and GRIP++ are the top-performing models. BAT excels in shorter prediction horizons (
Overall, graph-based approaches employing GCNs or GATs demonstrate superior ability to capture spatial relations among vehicles. The surrounding vehicles in most existing studies are either selected within a predefined area of interest [3], [12], [15], [16] or by simulating the effect of sensor detection [17]. While this selection process increases the similarity to the current traffic environment, insufficient information about the entire driving context can cause prediction errors. From human drivers’ perspective, their anticipation of driving is highly concerned with the risk of collisions. They tend to pay more attention to road agents posing higher level of risks.
Considering the temporal domain, among the three TCN-based studies [25], [26], [27], only TCN-SA [27] provides results on the NGSIM dataset, where it achieves a middle level of performance. It is noteworthy that another TCN-based model [26] achieves better prediction accuracy than the UST model [19] on the Argoverse dataset. However, as shown in Table 2, the prediction accuracy of UST is not among the top tier. Therefore, it remains challenging to determine whether LSTM-based or TCN-based models can achieve better prediction accuracy overall. Although TCNs theoretically offer faster inference speeds due to their parallelism [33], there has been limited analysis on this hypothesis within vehicle trajectory prediction studies.
This paper aims to fill these gaps by proposing a novel approach that combines TCNs for capturing temporal correlations and risk-enhanced GCNs for extracting social interactions among vehicles. By evaluating the risk of collision, surrounding vehicles are automatically selected, and their differentiated influence on other vehicles is accounted for, enhancing the accuracy and explainability of the model.
Methodology
This section consists of the dataset preparation, the formulation of the research problem, the model’s specification and the prediction of future trajectory.
A. Problem Description
Given a set of N vehicles in the scene with their past track information
B. Model Architecture
The architecture of the proposed model is illustrated in Fig. 1. It consists of three primary components: (i) a GCNs module for extracting spatial features, (ii) a Temporal Convolutional Networks (TCNs) module for capturing temporal correlations, and (iii) a decoder module for predicting future trajectories.
1) Spatial Feature Extraction
To capture spatial interactions, an undirected graph representation of vehicle trajectories is constructed at each time step t, represented as \begin{align*} e_{t}^{ij}= \begin{cases} 0, & ~ if\ i=j \\ r_{t}^{{lon}_{ij}} * r_{t}^{{lat}_{ij}}, & ~ otherwise \end{cases} \tag {1}\end{align*}
\begin{align*} r=& \begin{cases} 0, & ~ if~ d \geq {d_{min}} \geq {0} \\ 1-\frac {d-d_{min,b}}{d_{min}-d_{min,b}}, & ~ if~ d_{min} \geq {d} \\gt 0 \\ 1, & ~ otherwise \end{cases} \tag {2}\\ d_{min}=& \left [{{v_{j}\tau +\frac {1}{2}\tau ^{2}a_{max\_a}^{j}+\frac {{(v_{j}+\tau a_{max\_a}^{j})}^{2}}{2a_{min\_b}^{j}}-\frac {v_{i}^{2}}{2a_{max\_b}^{i}}}}\right ]_{+} \tag {3}\\ d_{min\_b}=& \left [{{v_{j}\tau +\frac {1}{2}\tau ^{2}a_{max\_a}^{j}+\frac {{(v_{j}+\tau a_{max\_a}^{j})}^{2}}{2a_{max\_b}^{j}}-\frac {v_{i}^{2}}{2a_{max\_b}^{i}}}}\right ]_{+} \tag {4}\end{align*}
As the graph representation is undirected, it can be noted that
The adjacency matrix can then be normalized as,\begin{equation*} \hat {A}_{t}=\Lambda _{t}^{-\frac {1}{2}}{\widetilde {A}}_{t}\Lambda _{t}^{-\frac {1}{2}} \tag {5}\end{equation*}
The output of the lth layer can be denoted as,\begin{equation*} H^{l}=\sigma \left ({{{\hat {A}}_{t}H^{l-1}W^{l}}}\right ) \tag {6}\end{equation*}
2) Temporal Feature Extraction
Temporal dependencies are captured using TCNs, which offer a robust alternative to Recurrent Neural Networks (RNNs). Unlike RNNs, TCNs avoid issues like accumulated error propagation and are computationally efficient [33]. A customised 3-layer TCN is employed to process historical trajectory data across varying temporal scales effectively. The output of the lth layer can be denoted as,\begin{equation*} H^{l} = \sigma \left ({{ \sum _{i=0}^{k-1} W^{l} \cdot H^{l-1}\left [{{t - d \cdot i}}\right ] + B^{l} }}\right ) \tag {7}\end{equation*}
At each time step, the absolute positions of all vehicles are transformed into a localised coordinate frame and embedded into fixed-length vectors. The spatial features derived from GCNs are concatenated with these positional embeddings, creating combined feature vectors. These vectors are then processed by the TCN module to capture both spatial and temporal relations among vehicles.
3) Future Trajectory Prediction
The variety loss proposed by Gupta et al. [37] is adopted to generate multiple reasonable and realistic future trajectories. A shared random noise z is concatenated with the extracted embeddings before being sent to the decoder as input. The noise is randomly sampled from \begin{equation*} h={\overline {m}}\ ||\ {z} \tag {8}\end{equation*}
These concatenated vectors are sent to the MLP-based decoder (3 layers) and twenty possible future trajectories are computed at each step.
The model is trained to minimise the variety loss, computed by choosing the trajectory with the minimum ADE to ground truth. The loss function can be denoted as,\begin{equation*} L_{variety}=\underset {k}{\min }{{\|{\hat {s}}_{i}^{k}-s_{i}\|}_{2}}, \ k\in \{1,2,\ldots , 20\} \tag {9}\end{equation*}
C. Evaluation Metrics
To evaluate the performance of the model, ADE, FDE and \begin{align*} ADE=& \frac {1}{NT_{pred}}\sum _{i=1}^{N}\sum _{t=1}^{T_{pred}}{\|{\hat {s}}_{t}^{i}-s_{t}^{i}\|}_{2} \tag {10}\\ FDE=& \frac {1}{N}\sum _{i=1}^{N}{\|{\hat {s}}_{T_{pred}}^{i}-s_{T_{pred}}^{i}\|}_{2} \tag {11}\\ KDE=& \frac {1}{N\sigma \sqrt {2\pi }}\sum _{i=0}^{N}{e^{-\frac {1}{2}\left ({{\frac {\hat {s}_{i}-s_{i}}{\sigma }}}\right )^{2}}} \tag {12}\\ NLL=& -\frac {1}{N}\sum _{i=1}^{N}{\log {P(\hat {s_{i}}|s_{i})}} \tag {13}\end{align*}
Results and Discussion
The proposed model was developed with PyTorch, an open-source machine learning Python library. The training and evaluation were implemented on a Desktop PC (CPU: Intel Core i9-9940X CPU @ 3.30GHz, GPU: 2 x NVIDIA GeForce RTX 2070 Super). The model was trained using Adam optimiser for 200 epochs with a learning rate of 0.001 and a batch size of 128. ReLU was used as the activation function of all networks, and a dropout rate of 0.2 was adopted.
This section has been divided into four parts to present the obtained results. First, an example of the constructed graph is illustrated to demonstrate the proposed risk-awareness kernel function. Second, a sample of the prediction results is presented. Afterwards, a benchmark analysis is conducted to compare the performance of this model with existing studies. Finally, an ablation study is undertaken to evaluate the proposed kernel function further.
A. Data Preparation
This study used two publicly available datasets for training and evaluation, the Highway Drone Dataset (highD) [19] and Intersection Drone Dataset [39]. Both datasets were collected by the same research group at RWTH Aachen University following the same methodology.
The highD dataset was established by recording traffic flow at six different locations on German motorways using drones. Each drone covers a road segment of approximately
Compared with the most widely used NGSIM dataset reviewed in the previous section, the highD and inD dataset prevails in several aspects: the data length, vehicle variety, and accuracy of the extracted trajectories. More importantly, it should be noted that the typical positioning error in highD and inD is less than
During preprocessing, a reconstruction of the data was carried out into the preferred structure to facilitate the proposed study, as well as a downsampling to 5 Hz to reduce the computation. This restructured dataset was divided into three subsets for training, validation and test purposes, with a ratio of 7:1.5:1.5.
B. Risk-Awareness Graph Construction
A sample scene was selected to demonstrate the proposed risk-awareness graph. As shown in Fig. 2, the plot above shows a frame of the vehicle sequence data. While there are more vehicles in this frame, this segment was selected for demonstration. There are six vehicles along this 100m road segment, and they are split into three car lanes. The average speed of these six vehicles is 96.53 km/h, with the maximum speed to be 111.42 km/h of vehicle ID9 and the minimum speed to be 77.12 km/h of vehicle 23.
The graph constructed using the risk-awareness kernel function is shown in the subplot below. The weight of edges ranges from 0 to 1, and a larger weight value indicates a more substantial social influence between vehicles. Noticeably, if the weight of an edge is zero, this edge is then removed for a clear illustration. As shown in the graph, large risk indices occurred between vehicles in the same lane. This is because the vehicles travel fast, and there are no lane-changing events in this selected sample.
Meanwhile, vehicle ID16 and ID23 are associated with larger risk values. This is because the vehicular physic parameters when computing risk index are differentiated to reflect different vehicle types. As both vehicles are trucks, they have severer restrictions on braking capability and are more prone to collisions. Thus, their associated risks are higher.
Moreover, as risk index is defined according to the limit of vehicle dynamics, a positive value can be acquired between vehicles within a more reasonable scope. While this leads to a larger graph at each frame, it ensures a more comprehensive spatial relation is extracted. Thus, AVs can have more thorough anticipations of the scene. Even if vehicles perform unusual risky manoeuvres at a distance, host AVs can still have enough time and space to avoid collisions.
C. Trajectory Prediction Example
A sample is illustrated in Fig. 3 to present the prediction results of the proposed method. The solid lines represent the 4 seconds past trajectories, the dashed lines are the stochastic predictions for the future 2 seconds, and the marked lines are the ground truth for the same period. Although the prediction errors vary among the six vehicles presented in this sample, the average ADE of these vehicles is 0.40m with a standard deviation of 0.38m. Meanwhile, the average FDE and its standard deviations are 1.02m and 0.66m, respectively. As the average travelling speed of these six vehicles is about 104.4 km/h, the prediction accuracy is very promising in this context.
D. Performance Evaluation
Benchmark Models. With the sample result presented, a benchmark analysis with some state-of-the-art approaches was conducted to evaluate the performance of the proposed model. It should be noted that some of the selected methods were commonly used in assessing the performance of newly proposed trajectory prediction approaches, and the others represent the key milestones and more recent techniques in this field of research. Therefore, these baseline models were selected to evaluate the performance of the model proposed in this study.
V-LSTM: Vanilla LSTM is one of the classical methods for time-series prediction. Its application in trajectory prediction uses a single LSTM to encode the motion history of the ego vehicle without considering its spatial interactions with surrounding vehicles.
S-LSTM [40]: Social LSTM is initially developed for pedestrian trajectory prediction. Each pedestrian is modelled using an LSTM. The hidden states of pedestrians within a specific area are pooled using fully connected social pooling to preserve spatial interactions.
CS-LSTM [12]: Convolutional Social LSTM is designed for vehicle trajectory prediction. It uses convolutional layers with social pooling and generates a multimodal prediction based on six manoeuvres.
S-GAN [37]: Social GAN models each pedestrian motion with an LSTM, and the hidden states of pedestrians are pooled using global pooling. GAN is used to generate multimodal prediction results.
STGAT [41]: Originally developed for pedestrian trajectory prediction. It is a seq2seq model which uses one LSTM for each pedestrian’s motion state. Meanwhile, the spatial interactions are extracted with Graph Attention Network (GAT), and an extra LSTM models the temporal correlations between interactions.
GRIP++ [22]: GRIP++ ranked first in the 2019 ApolloScape trajectory competition and achieved top accuracy with the NGSIM dataset, as listed in Table 2. It uses fixed and dynamic graphs for spatial relations and a two-layer GRU for trajectory prediction, applicable to various traffic agents like vehicles, pedestrians, and cyclists.
Social-STGCNN [42]: Originally developed for pedestrian trajectory prediction, Social-STGCNN uses Spatio-Temporal Graph Convolutional Neural Networks (ST-GCNN) to extract spatial and temporal relations from spatio-temporal graphs. It introduces a weighted adjacency matrix to model social influence between pedestrians, achieving notable improvements in prediction accuracy and speed.
STDAN [31]: STDAN introduces a novel spatial-temporal dynamic attention network for vehicle trajectory prediction. It incorporates a driving intention-specific feature fusion mechanism, allowing the adaptive integration of temporal and social features for maneuver-based, multi-modal trajectory prediction.
Since these baseline models were initially proposed for different purposes and were trained with various datasets, they were retrained using the same dataset in this study to facilitate the evaluation on a unified basis. Same training parameters were adopted, and the comparison was implemented with an observation horizon of 4s and a prediction horizon of 2s. The comparison mainly focuses on the accuracy of the prediction results and inference speed.
Prediction Accuracy Performance. The comparison results between the proposed method and the other eight models were summarised in Table 4. For the highD dataset, the proposed model outperforms all others, achieving the best accuracy with an ADE of 0.43m and an FDE of 0.79m. In comparison, STDAN, which previously represented the state-of-the-art, achieves an ADE of 0.47m and an FDE of 0.81m. Despite STDAN’s strong performance, the proposed model achieves an additional improvement of 8.51% in ADE and 2.47% in FDE. This improvement underscores the superior prediction accuracy of the proposed method. Among the existing models, V-LSTM exhibits the most significant prediction errors, likely due to its reliance solely on motion history and lack of spatial interaction considerations. This highlights the importance of incorporating spatial features into vehicle trajectory prediction to account for the interactions between vehicles, which can lead to more accurate results.
In the inD dataset, CS-LSTM achieves the best ADE of 0.66m, indicating its strong performance in average displacement prediction. However, Social-STGCNN records the lowest FDE of 1.28m, outperforming other models in terms of final displacement accuracy. The proposed model performs well, with an ADE of 1.07m and an FDE of 1.29m, placing it among the top models for final displacement but slightly behind in average displacement.
Social-STGCNN’s performance, while notable, requires further discussion. Although it ranks third in prediction accuracy, this model was initially designed to predict pedestrian trajectories, where it has achieved state-of-the-art performance. The fundamental principles and scales involved in pedestrian trajectory prediction differ significantly from those required for vehicle trajectory prediction, potentially limiting Social-STGCNN’s ability to generalise to vehicle scenarios. In this study, the model was used as a benchmark, and only limited adaptations were made to ensure vehicle data were appropriately scaled and transmitted. No further modifications were applied to the model’s core specifications. Thus, with more targeted adjustments and calibrations, its accuracy in vehicle trajectory prediction could potentially be improved.
A general trend observed when comparing the two datasets is the degradation in performance for most models when tested on the inD dataset. This is despite the fact that vehicles in the inD dataset generally move slower. The likely explanation lies in the more complex interactions among vehicles in the intersection scenarios found in the inD dataset. Social pooling and graph-based models, which are designed to capture pairwise interactions, struggle to extract higher-order relationships in these more complex environments. This may explain why CS-LSTM achieves better performance in the inD dataset, as its convolutional social pooling mechanism aggregates information from multiple neighboring vehicles simultaneously and captures the collective influence of all vehicles within a specific spatial area. The ability to handle these high-order interactions more effectively allows CS-LSTM to perform better in scenarios involving complex vehicle interactions, such as intersections.
Inference Speed Performance. As shown in Table 4, the advantage of extracting spatial relations with Temporal Convolutional layers over LSTM is revealed. All three models (GRIP++, Social-STGCNN and the proposed model) obtain significant improvements in inference speed than LSTM-based models. Social-STGCNN achieves the best performance among all models in comparison. It is over 73 times faster than the slowest method (S-GAN) and about 1.23 times faster than the STDAN, the fastest LSTM-based model. As the proposed model needs to construct the risk graph to determine the spatial relations at each timestep, it has not achieved the best performance in inference speed. However, it still obtains a notable improvement than most existing models and is about four times faster than V-LSTM and over 41 times faster than S-GAN.
Distributional Performance. As ADE and FDE can not compare the distributions produced by generative models, NLL is adopted to evaluate the variance and multimodality, without assumptions about the output’s distribution [38]. The state-of-the-art stochastic model (STGAT) is used for evaluation. The proposed model was evaluated over time to investigate the performance changes along the prediction horizon. Results of 1000 sampled trajectories are shown in Fig. 4, and error bars are bootstrapped at 95% confidence intervals. As the average NLL of the proposed model is smaller than STGAT at every prediction step, it indicates our model’s consistent multimodal modelling capacity.
E. Ablation Study of Risk-Awareness Kernel Function
Since the kernel function for computing the adjacency matrix represents the social influence between vehicles, it would be beneficial to evaluate if the proposed approach can effectively capture the essence in spatial relations while remaining computationally efficient. Thus, another two commonly used kernel functions were adopted to benchmark the performance of the proposed approach using the highD dataset.
The first kernel function is to treat all vehicles within the neighbourhood area equally. The spatial relation captured with this kernel function is similar to some reviewed studies [3], [15], [16], [17]. This kernel function can be represented as,\begin{align*} e_{t}^{ij}= \begin{cases} 1, & ~ if~ \|{s_{t}^{i}-s_{t}^{j}\|}_{2} \lt threshold \\ 0, & ~ otherwise \end{cases} \tag {14}\end{align*}
Meanwhile, the second kernel function is based on the relative distance between two vehicles to model their social impacts [42]. It can be easily interpretable, despite the vehicles’ dynamic relations, spatially closer vehicles tend to have a more significant influence on each other. It also correlates that human drivers tend to pay more attention to vehicles in closer proximity. This kernel function can be denoted as,\begin{equation*} e_{t}^{ij}=1-\frac {\|{s_{t}^{i}-s_{t}^{j}\|}_{2}}{max\_length} \tag {15}\end{equation*}
Both kernel functions were implemented for comparison. As listed in Table 5, the first kernel function has the worst prediction accuracy. This is because this kernel function only considers vehicles within the neighbourhood and does not differentiate the levels of their social impacts. The neighbourhood area selected with fixed thresholds can lose the essential information that determines the driver’s motivations. Moreover, assuming vehicles within the neighbourhood have the same social impacts on the ego vehicle can also lead to severe issues. Different vehicle positions and dynamic states can pose distinct levels of danger and stress on the ego driver. Meanwhile, the second kernel function performs better than the first, improving 43.3% in ADE and 21.0% in FDE. While it confirms the benefits of differentiating the social impacts between vehicles, solely depending on distance is insufficient to determine the spatial relations. This is because surrounding vehicles at the same distance can pose distinct social impacts on the ego vehicle. Thus, although the second kernel function can differentiate the social relations between most vehicles, they can lead to a false adjacency matrix in some cases. The proposed risk-awareness kernel function compensates for these drawbacks and improves 42.8% in ADE and 51.2% in FDE.
From the perspective of inference speed, while all three kernel functions achieve fast inference speed, the proposed kernel function is the slowest among them. It is mainly caused by additional variables and calculations required to construct the graph representation of the scene at each time step.
With the comparison of prediction accuracy and inference speed, it can be noted that the proposed risk-awareness kernel function can better extract the essential spatial relations between vehicles and yield improved performance in predicting the future trajectory of vehicles. Although its inference speed is slower than the other two kernel functions, it is still much faster than LSTM-based models and should be sufficient for most applications.
Conclusion
This paper presented a novel approach for vehicle trajectory prediction by modelling the spatial-temporal interactions among vehicles. The proposed model was trained and evaluated using the publicly available Highway Drone Dataset and Intersection Drone Dataset, demonstrating promising improvements in both prediction accuracy and inference speed compared to eight existing methods. An ablation study further validated the effectiveness of the risk-awareness kernel function, highlighting its contribution to enhancing the model’s explainability.
The primary contribution of this research lies in the development of a stochastic vehicle trajectory prediction method that models vehicular spatial-temporal interactions. A novel risk-awareness kernel function was introduced to construct a weighted adjacency matrix, effectively capturing the spatial relationships between vehicles. The model employs GCNs to extract spatial features and TCNs to model temporal dependencies. The combination of these features in a decoder produces a stochastic, multimodal prediction of future vehicle trajectories.
Despite the promising results, the map-free nature of the proposed approach presents limitations, particularly in more complex road geometries, such as urban environments. The use of relatively simple road structures in the datasets may have contributed to the observed high accuracy, with the model’s performance deteriorating when applied to the inD dataset, which features more intricate road layouts. Future work will focus on evaluating the model with data from more complex road segments to further test its robustness. Moreover, to address the limitations observed in complex datasets such as inD, future research will focus on integrating hypergraph representation learning into the current model. By leveraging hyperedges to capture higher-order interactions among road agents, this approach will enable the model to extract and represent spatial relations and interaction dynamics more effectively. Such advancements are expected to significantly improve the model’s ability to handle scenarios with intricate traffic dynamics, enhancing risk-awareness and prediction accuracy. Additionally, the potential trade-offs between incorporating map information and maintaining model scalability will be explored to enhance the model’s applicability in diverse traffic scenarios.