Introduction
Visualization of driving behavior can help drivers to review and understand their driving behavior. Effective review of driving behavior can contribute to improving their driving behavior and promoting safe driving. Besides, efficient visualization of driving behavior enables users, e.g., driving school teachers, security officers at transportation companies, and accident investigators, to monitor and investigate driving behavior of cars intuitively.
Visualization of driving behavior could be useful when a person needs to review his/her driving behavior because a user who wants to review their driving behaviors cannot intuitively understand their behavior by watching driving behavior data, i.e., raw time series data. In contrast, reviewing recorded video takes a long time and is not so efficient to find interesting moments in his/her trip. The effective visualization of driving behavior can help drivers to understand their good and bad driving habits. Meanwhile, to ensure safe driving, providing information about previous driving behaviors after driving to users, such as driving school teachers, security officers at transportation companies, and accident investigators is also important. For example, a driving school teacher could understand the driving behavior of students intuitively by visualizing the students’ driving behavior. In addition, security officers at a transportation company could monitor the driving behavior of a taxi or bus driver remotely by visualization, and thus, they may inform a driver promptly when a dangerous driving behavior is visualized. Moreover, we expect that the visualization of driving behavior would be useful for accident investigations. The visualization of driving behavior can allow investigators to determine whether unusual driving behavior had led to an accident. Thus, an effective visualization method of driving behavior will be useful for developing various support systems for monitoring, reviewing and analyzing driving behavior, and driving assistance systems.
Many advanced driving assistance systems (ADASs) have been focusing on assisting a driving behavior directly, i.e., vehicle assistance control [1]–[3]. They even participate in controlling the vehicle directly when the car confronts a danger [4], [5]. Driving environment detection [6], driving behavior recognition [7], prediction [8], selective darkening of the windshield [9] and determining utterance timing of driving agents [10] are also examples of ADASs. Although the visualization of driving behavior has been gathering less attention in the fields of intelligent vehicles than ADASs, the visualization of driving behavior is an extremely important issue. In this study, we propose a method that can visualize driving behavior effectively. It is shown that the method can help people to recognize distinct patterns in driving behavior.
Driving behavior can be measured and recorded by various types of sensor via a control area network (CAN), e.g., the accelerator opening rate, brake master-cylinder pressure, and steering angle. We regard driving behavior data as multi-dimensional time-series data that is formed by assembling various types of sensor information. Each type of sensor information comprises a dimension of the driving behavior data. Here, the important property of the driving behavior data is that each time series data observed by each sensor is not independent of each other. In [11], we assumed that there are a small number of essential hidden features that characterize driving behavior, and each concrete measured sensor information can be regarded as time series data that is generated by performing a nonlinear transformation to the essential hidden features. For illustrative purpose, let us assume that there are three essential hidden features here, and they relate to “acceleration”, “changes in the directions” and “velocity of the vehicle”, respectively. For example, the observed sensor information “longitudinal acceleration” is considered to be generated by hidden features related to acceleration and changes in the directions via nonlinear transformation [12]. Note that changes in the direction cause the friction between the tire and the ground and decreases longitudinal acceleration when a driver turns the steering to go right or left. Moreover, some of the observed sensor information share the same hidden feature. For example, engine speed, speed meter and speed of wheels are generated by hidden feature associated with the velocity of the vehicle. One type of sensor information can be generated by fusing several hidden features. For example, the yaw rate is generated by fusing the hidden features associated with the velocity of the vehicle and changes in the driving direction.
The important problem for visualization is how to extract the essential features from observed driving behavior data, which contain redundant information and nonlinear relationship between different sensor information. To address this problem, we require a method that automatically extracts the essential hidden features and filters out any redundant sensor information. The method should extract the necessary and sufficient features from high-dimensional driving behavior data automatically.
In this study, we propose a driving behavior visualization method based on a feature extraction method using a deep sparse autoencoder (DSAE). We call the visualization method a driving color map. We reported a brief description of this method and preliminary results in [13]. In this paper, we describe a complete description and sufficient experimental evaluation of our method and demonstrate the validity of our proposed method based on several experiments. We show that the visualization obtained by our proposed method performs better than other methods. It is shown that the proposed method can represent some complex driving behaviors more clearly than other baseline methods in particular. We also present some examples of applications of the driving color map, e.g., we detected some interesting patterns in observed driving behavior by using driving color maps. The method also successfully estimated the hidden features and visualized the driving behavior on a public road using DSAE trained by using driving behavior data measured on factory circuit. This suggests that our proposed visualization method has high generalization performance and can be used for practical use.
Background
In this section, we describe a short survey about visualization methods in the field of intelligent vehicles and feature extraction methods for multivariate time series data.
A. Visualization in Intelligent Vehicles and Intelligent Transportation Systems
Visualization of driving behavior is crucially important to improve driving skill and to investigate hazardous driving behavior. However, there have been few studies of the visualization of driving behavior data. For example, to reduce road traffic accidents, Hilton et al. presented a visualization method called SafeRoadMaps for communicating safety information to users [14], which visualizes the crash density at different locations based on a color heat map. Treiber et al. proposed an adaptive smoothing method with a visualization method, which visualizes the spatiotemporal dynamics of traffic patterns by using colors [15]. Huang et al. [16] proposed a visualization method for studying urban network centralities. The goals of the studies differ from ours.
Kilicarslan and Zheng [17] proposed a method to visualize a sequence of driving scenes by using driving videos. They manually designed a method for mapping driving videos onto a temporal profile image. This method can express driving behavior indirectly based on changes in the surrounding environment. However, it is difficult for users to infer actual driving behaviors from the compressed visual image.
Takeda et al. [18] proposed self-coaching system using on recorded driving data an Gaussian mixture model-based driver-behavior models. They developed a web-based driving feedback system. Their goal is similar to ours. However, they focused on risky driving behaviors and did not provide a visualization method of driving behavior itself.
In our study, we use driving behavior data which represents driving behaviors directly. The driving behavior data is so high-dimensional that it is difficult to visualize intuitively. Therefore, we use an unsupervised feature extraction method to extract low-dimensional hidden features from the high-dimensional time-series data automatically without any human intervention.
B. Feature Extraction and Deep Learning
Many unsupervised feature extraction methods can extract low-dimensional hidden features, i.e., latent time series data, automatically from high-dimensional time series data. Principal components analysis (PCA) is widely used for feature extraction [19]. PCA can find the principal components that correspond to an axis having the largest variance. However, PCA basically can produce good results when the input data follow a Gaussian distribution and essential hidden features are orthogonal to each other in the vector space. Independent component analysis (ICA) [20] can effectively extract the independent hidden features from multivariate signals. The method assumes that the observed signals are generated via linear transformations from source signals. PCA and ICA might not be able to extract the hidden features from driving behavior data by linear transformations because vehicle dynamics and human driving behavior involve nonlinear properties.
Schölkopf et al. [21] proposed the kernel PCA (KPCA) method. KPCA is a feature extraction method that considers non-linear transformation. It uses a nonlinear kernel function that involves a nonlinear transformation to map the data onto a high-dimensional space. After the non-linear transformation, KPCA employs PCA to find the principal axis of the high-dimensional space. However, the computational cost of the KPCA is high when there is a large volume of driving behavior data because the kernel method must compute a Gram matrix in
In recent years, feature extraction methods that employ deep learning approaches have attracted much attention. Deep learning methods employ neural networks having a deep structure, i.e., having over three or more layers. Some deep learning methods, such as a restricted Boltzmann machine (RBM [22]), autoencoder [23], and convolutional autoencoder [24], employ pre-training before fine tuning. Pre-training comprises unsupervised learning for feature extraction. Bengio indicated that an autoencoder is similar to an RBM because both have a double-layered completely undirected graph structure, although an autoencoder can be trained more easily than an RBM [25]. In the fine tuning, the neural network having a deep structure is trained throughout all of the layers.
Few studies have used deep learning methods for driving behavior analysis. Diaz et al. [26] successfully used a deep learning method based on an autoencoder to model driving behaviors to estimate the energy consumption by electric vehicles. Tagawa [27] proposed structured denoising autoencoder for fault detection and applied it to driving behavior data. Our previous study showed that DSAE could be used to extract essential hidden features from driving behavior data [11]. That suggests DSAE could be used as an information filter to decrease the redundancy of driving behavior. Therefore, in this paper, we use a DSAE to extract the hidden features for visualizing driving behavior.
Proposed Method
In this section, we describe our proposed method in detail. Our proposed method is illustrated in Fig. 1, which comprises two steps: hidden feature extraction and visualization. In the first step, the method extracts three-dimensional features from driving behavior data. It firstly uses the normalization and windowing process to pre-process the driving behavior data, which is shown in the left part of Fig. 1. The method employ a DSAE with five encoding layers1 to extract the three-dimensional hidden features from driving behavior data. Note that the detail of DSAE’s structure using in the experiment is written in IV-A. We assume that the shallower layer can extract low level features which can represent more specific driving behaviors; the deeper layer can extract higher level features which can represent more abstract driving behavior by fusing various low level features. We expect that the hidden features of different driving behaviors are represented distinctively in the feature space. The second step is the visualization which is shown on the right part of Fig. 1. After the method calculates colors by mapping the extracted three-dimensional hidden features into the RGB color space, we place these colors in their corresponding positions on the map to generate the driving color map. The goal of the method proposed in this study is to provide an intuitive color map to make different driving behaviors represented by different colors.
The proposed method includes two processing steps: hidden feature extraction by using the deep sparse autoencoder and visualization of driving behavior in the driving color map.
A. Feature Extraction by Using DSAE
The DSAE is comprised of many sparse autoencoders (SAEs). Parameters (weight matrices and bias vectors) of each SAE is optimized layer by layer in the pre-training stage. The SAE can be viewed as a three-hierarchical directed graph, i.e., two-hierarchical undirected graph, with a visible layer, hidden layer, and reconstruction layer. It is designed to ensure that the hidden layer has the property of sparseness. Each SAE encodes input data into data on its hidden layer and then decodes the data so as to reconstruct the input data. An SAE is optimized to minimize the error between the input data and reconstructed data by using a back-propagation (BP) method. After this reconstruction error converges to a sufficiently small value, pre-training for the SAE is stopped. Each SAE outputs its hidden layer’s data, i.e., hidden features, to the visible layer in the next SAE as input data. During the fine-tuning stage, the optimized parameters are initialized to train all of the layers of the DSAE. The DSAE can determine the hidden features by extracting them via the deep structure comprising tandem SAEs.
In this study, we define driving behavior data set as \begin{equation} {\mathbf{y}}_{t}=(y_{t,1}, y_{t,2}, {\dots }, y_{t,D_{Y}})^{T}\in \mathbb {R}^{D_{Y}}. \end{equation}
Before the data are given to the DSAE, normalization and windowing operation are performed. We assume that the range of observed driving behavior data in the data set \begin{align} {\mathbf{x}}_{t}=&~(x_{t,1}, x_{t,2}, {\dots }, x_{t,D_{X}})^{T}\in \mathbb {R}^{D_{Y}},\notag \\ x_{t,d}=&~2\left({\frac {y_{t,d}-{y_{d}}_{min}}{y_{d}{}_{max}-{y_{d}}_{min}}}\right)-1, \end{align}
\begin{equation} {\mathbf{v}}_{t}=({\mathbf{x}}_{t-w+1}^{T},{\mathbf{x}}_{t-w+2}^{T},\ldots, {\mathbf{x}}_{t}^{T})^{T}\in \mathbb {R}^{D_{V}}(t\geq w). \end{equation}
Next, matrix \begin{equation} {\mathbf{h}}^{(l)}_{t}=\tanh ({\mathbf{W}}_{en}^{(l)}{\mathbf{v}}^{(l)}_{t}+{\mathbf{b}}_{en}^{(l)})\in \mathbb {R}^{D_{H}^{(l)}}, \end{equation}
\begin{equation} {\mathbf{r}}^{(l)}_{t}=\tanh ({\mathbf{W}}_{de}^{(l)}{\mathbf{h}}^{(l)}_{t}+{\mathbf{b}}_{de}^{(l)})\in \mathbb {R}^{D_{V}^{(l)}}. \end{equation}
\begin{align} {O}({\mathbf{V}}^{(l)})=&~\frac {1}{2N_{V}}\sum _{t=1}^{N_{V}}||{\mathbf{r}}^{(l)}_{t}-{\mathbf{v}}^{(l)}_{t}||_{2}^{2}+\frac {\alpha }{2}(||{\mathbf{W}}_{en}^{(l)}||_{2}^{2}+||{\mathbf{W}}_{de}^{(l)}||_{2}^{2})\notag \\&+\,\beta \sum ^{D_{H}^{(l)}}_{i=1}{\mathrm{ KL}}(\omega ||{\bar {h}}_{i}^{(l)}). \end{align}
\begin{equation} {\mathrm{ KL}}(\omega ||{\bar {h}}_{i}^{(l)})=\omega \log \frac {\omega }{{\bar {h}}_{i}^{(l)}}+(1-\omega) \log \frac {1-\omega }{1-{\bar {h}}_{i}^{(l)}}, \end{equation}
\begin{equation} \bar{\boldsymbol h}^{(l)}_{i}=\frac {1}{2}\left({1+\frac {1}{N_{V}}\sum _{t=1}^{N_{V}}h^{(l)}_{t,i}}\right), \end{equation}
Finally, we use the BP method [28] to minimize the objective function and train the SAE. The BP method requires the partial differentiation of the weight matrices \begin{align} \frac {\partial {O}^{(l)}}{\partial \boldsymbol{ W_{de}^{(l)}}}=&~\frac {1}{N_{v}}\sum _{t=1}^{N_{v}}{\mathbf{h}}^{(l)}_{t}\boldsymbol {\gamma }_{t}^{T}+\alpha {\mathbf{W}}_{de}^{(l)}, \\ \frac {\partial {O}^{(l)}}{\partial \boldsymbol{ b_{de}^{(l)}}}=&~\frac {1}{N_{v}}\sum _{t=1}^{N_{v}}\boldsymbol {\gamma }_{t}, \end{align}
\begin{equation*} {\gamma }_{t}\!=\! \mathop {\mathrm {diag}}{\big (({r^{(l)}_{t,1}})^{2}\!-\!1,({r^{(l)}_{t,2}})^{2}\!-\!1,\cdots, ({r^{(l)}_{t,D^{(l)}_{v}}})^{2}\!-\!1\big) }({\mathbf{v}}^{(l)}_{t}\!-\!{\mathbf{r}}^{(l)}_{t}). \end{equation*}
\begin{align} \frac {\partial {O}^{(l)}}{\partial \boldsymbol{ W_{en}^{(l)}}}=&~\frac {1}{N_{v}}\sum _{t=1}^{N_{v}}{\mathbf{v}}_{t}^{(l)}\boldsymbol {\xi }_{t}^{T}+\alpha {\mathbf{W}}_{en}^{(l)}, \\ \frac {\partial {O}^{(l)}}{\partial \boldsymbol{ b_{en}^{(l)}}}=&~\frac {1}{N_{v}}\sum _{t=1}^{N_{v}}\boldsymbol {\xi }_{t}, \end{align}
\begin{align*} \boldsymbol {\xi }_{t}=&~\mathop {\mathrm {diag}}{\big (1-({h^{(l)}_{t,1}})^{2},1-({h^{(l)}_{t,2}})^{2},\cdots, 1-({h^{(l)}_{t,D^{(l)}_{H}}})^{2}\big) } \\&\times (\boldsymbol{ W_{de}}^{(l)}\boldsymbol {\gamma }_{t})+\beta \boldsymbol {\epsilon }, \end{align*}
\begin{equation*} \epsilon _{i}=\frac {1-\omega }{1-{\bar {h}}_{i}^{(l)}}-\frac {\omega }{{\bar {h}}_{i}^{(l)}}. \end{equation*}
The equations for the biases \begin{align} \mathbf{ W^{+}}_{en}^{(l)}\leftarrow&~{\mathbf{W}}_{en}^{(l)}-\lambda _{en}\frac {\partial {O}({\mathbf{V}}^{(l)})}{\partial \mathbf{ W_{en}^{(l)}}}, \\ \mathbf{ b^{+}}^{(l)}_{en}\leftarrow&~{\mathbf{b}}^{(l)}_{en}-\lambda _{en}\frac {\partial {O}({\mathbf{V}}^{(l)})}{\partial \mathbf{ b_{en}^{(l)}}}, \\ \mathbf{ W^{+}}_{de}^{(l)}\leftarrow&~{\mathbf{W}}_{de}^{(l)}-\lambda _{de}\frac {\partial {O}({\mathbf{V}}^{(l)})}{\partial \mathbf{ W_{de}^{(l)}}}, \\ \mathbf{ b^{+}}^{(l)}_{de}\leftarrow&~{\mathbf{b}}^{(l)}_{de}-\lambda _{de}\frac {\partial {O}({\mathbf{V}}^{(l)})}{\partial \mathbf{ b_{de}^{(l)}}}, \end{align}
During line search, we define the searching distance \begin{equation} \theta ^{+}=\begin{cases} -0.5\theta & \big ({O^{+}}({\mathbf{V}}^{(l)})>{O}({\mathbf{V}}^{(l)})\big) \\ \theta & \big ({O^{+}}({\mathbf{V}}^{(l)})\leq {O}({\mathbf{V}}^{(l)})\big) \end{cases}, \end{equation}
After pre-training phase, we stack many SAE into a DSAE, which is a deep neural network with an encoder-decoder structure. This fine-tuning stage maps the
B. Driving Behavior Visualization by the Driving Color Map
In the second step, we visualize the driving behavior by using the extracted hidden three-dimensional features. The visualization method is called the driving color map, which comprises a colored trajectory shown on a road map that represents the extracted features. In this section, we assumed that nine types of sensor information are obtained via a CAN as shown in Table I although the proposed method can be applied other to set of sensor data. It was suggested that the three-dimensional hidden features are almost sufficient to represent driving behaviors in [11]. In addition, the RGB color space is a three-dimensional space. Based on the two facts, we extract the three-dimensional hidden features to visualize driving behavior in this study. We obtain the colors in the RGB color space corresponding to the three-dimensional hidden features by a simple scaling method. When the range of the RGB color space is [0, 1]3, we normalize the three-dimensional hidden features into [0, 1]3. In summary, we map the three-dimensional hidden features to the RGB space by \begin{equation} rgb_{t,d}=\frac {h^{(final)}_{t,d}-h^{(final)}_{min_{d}}}{h^{(final)}_{max_{d}}-h^{(final)}_{min_{d}}}, \end{equation}
To monitor the visualization results, we develop a program called Deep Car Watcher3 (Fig. 2), which can simultaneously review the driving videos, extracted hidden features, driving color map, and time-series. A point indicates the current vehicle location on the driving color map.
Deep Car Watcher by which users can review the driving videos (A), and the extracted three-dimensional hidden features in the feature space (B, C, D, and E) and time space (F), which can also review the driving color map (G) simultaneously.
Visualization Experiments
A. Experimental Conditions
In this experiment, we verified that hidden features extracted using DSAE were better than those employed by other methods for visualizing driving behavior. We employed PCA, FastICA [29], KPCA with the RBF kernel and SAE as comparative methods to extract three-dimensional features from driving behavior data and generate driving color maps.
To obtain the driving behavior data, we asked a participant to drive an experimental vehicle through two courses at a company’s factory. The participant drove each course five times. During the experiment, the factory was in normal operation. Therefore, the car encountered different situations on the courses, such as pedestrians, other moving vehicles, and parked vehicles. The Circuits 1–5 correspond to the first course and Circuits 6–10 correspond to the second course. We observed 12958 frames of driving behavior data in total at a frame rate of 10 fps. Each data frame could include the nine sensor information captured via the CAN (see Table I).
As a working hypothesis, we assume that there are three essential hidden features, and they roughly correspond to “vehicle acceleration”, “vehicle velocity” and “changes in the driving direction”. We considered that the accelerator opening rate, brake master-cylinder pressure and longitudinal acceleration were mainly related to the acceleration of the vehicle; the speed of wheels, speed meter, and engine speed were mostly related to the velocity of the vehicle; the steering angle could represent a change in the driving direction; and the longitudinal acceleration and yaw rate included information about acceleration of the vehicle and changes in the driving direction. Thus, we prepared a simple three-dimensional feature called VV’S, i.e., the speed of wheels, the differential of the speed of wheels on time, and steering angle, which are expected to relate to the three essential hidden features assumed above. Then we generated driving color maps by using VV’S in addition to the above-mentioned comparative methods.
We used the VV’S and the three-dimensional hidden features extracted using PCA, FastICA, KPCA, SAE and DSAE to generate the driving color maps for comparison. These driving color maps were compared with maps obtained using the hidden features extracted by the DSAE. Except for the VV’S, each feature extraction method employed the windowing process with a time window size
B. Visualization Results
In this section, the direct observation of the visualization results is reported, first before we conduct quantitative evaluation in the following sections. The driving color maps generated by each method for Circuits 1 and 6 are shown as examples in Fig. 3. This figure shows that the colors in the driving color maps produced using the VV’S contained a significant amount of noise, and the colors were not as rich as those generated using the other methods. Thus, the driving color maps generated using the extracted hidden features with windowing processing could smooth the noise and represent driving behaviors better than the observed data (VV’S).
The driving color maps for Circuits 1 and 6 are shown as examples, which were generated based on the VV’S and the three-dimensional hidden features extracted using PCA, FastICA, KPCA, SAE, and DSAE. (a) Circuit 1 obtained using VV’S. (b) Circuit 6 obtained using VV’S. (c) Circuit 1 obtained using PCA. (d) Circuit 6 obtained using PCA. (e) Circuit 1 obtained using FastICA. (f) Circuit 6 obtained using FastICA. (g) Circuit 1 obtained using KPCA. (h) Circuit 6 obtained using KPCA. (i) Circuit 1 obtained using SAE. (j) Circuit 6 obtained using SAE. (k) Circuit 1 obtained using DSAE. (l) Circuit 6 obtained using DSAE.
We detected several simple and complex driving behaviors in the driving videos. The most obviously found driving behaviors were typical and simple driving behaviors, such as high speed forward, stopping the vehicle, accelerating forward, right rear reversing, and left rear reversing. We also found complex driving behaviors which involved several simple driving behaviors occurred simultaneously. For example, simple driving behaviors such as turning right and accelerating could be combined to yield complex driving behavior of “turning right while accelerating”. Similarly, turning left and accelerating could be combined to yield “turning left while accelerating”.
We observed and compared the driving color maps generated by each method to find their representative colors corresponding to each driving behavior. The representative color for each driving behavior is shown in Table II. Note that the representative colors were picked up manually and subjectively, and Table II is shown for illustrative purpose. The objective and quantitative evaluation is performed in the following sections.
It was observed that the proposed visualization method using DSAE generated distinctly different colors for different driving behaviors: high speed forward
In contrast, it was found that other methods generated similar colors for different driving behaviors. The similar colors for different driving behaviors are underlined in Table II. The similar colors obtained by each method are indicated by the same type of underline (
These results qualitatively suggest that the driving color maps obtained using DSAE made it easier for users to distinguish each driving behavior using the driving color maps than other methods.
C. Numerical Evaluation of Visual Separability in Driving Color Maps
We assessed whether different driving behaviors can be distinguished in each color space. We assumed that if the corresponding color of each driving behavior was linearly separable in the color space, then it could also be readily distinguished by a human. Thus, we used a binary support vector machine (SVM) with a linear kernel4 to evaluate the separability of the visualized driving behaviors.
We used the F-measure to evaluate the generalization performance of the generated colors based on leave-one-out cross-validation. In particular, we selected the driving behavior data from one circuit as the test set and the driving behavior data from the remaining nine circuits as the training set. We used the training set to train the feature extraction model and for extracting three-dimensional hidden features. Next, we used the hidden features extracted from the training set to train an SVM with a linear kernel for recognizing a specific driving behavior listed in Table III. We then inputted the test set into the trained feature extraction model to estimate the hidden features. The estimated hidden features were inputted into the trained SVM to estimate whether the estimated hidden features belonged to the specified driving behavior. We selected the driving behavior data from another circuit as the test set and repeated the training and estimation processes described above. Thus, we performed ten trials for each driving behavior shown in Table III for one feature extraction method. The segments and truth labels corresponding to each driving behavior shown in Table III was prepared in advance.
We counted the true positive (TP), false positive (FP), and false negative (FN) rates in the ten trials, and we calculated the F-measure as follows:\begin{equation} \mbox {F-measure}=\frac {2\times \mathrm {Recall}\times \mathrm {Precision}}{\mathrm {Recall}+\mathrm {Precision}}, \end{equation}
We selected several simple driving behaviors and complex driving behaviors, as shown in Table III. The F-measures obtained by the linear SVM for simple and complex driving behaviors by using DSAE and comparative methods are listed in Table III. We mark the highest F-measure by underline and bold font for each driving behavior. Meanwhile, the second peace of F-measure is marked by underline only.
First, we focus on simple driving behaviors. Even though the VV’S obtained good performance for some of the simple driving behaviors, it obtained worse results than other methods for complex driving behaviors. Thus, the VV’S can represent simple driving behaviors to some extent, but they are not suitable for complex driving behaviors. DSAE had the highest F-measure for stopping vehicle, and the second highest F-measures for turning right, and left rear reversing. The highest average F-measure was obtained by DSAE. PCA had the highest F-measures for turning right and turning left. FastICA could represent the right rear reversing and left rear reversing driving behaviors well, and it also obtained the second highest average F-measure. The results obtained using KPCA were better for accelerating forward, but it obtained the worst performance in average. When we used SAE, the F-measure was highest for high speed forward, and the F-measures for the other driving behaviors were intermediate.
Next, we focused on the complex driving behaviors. Our results clearly demonstrated that DSAE obtained better F-measures for all of the complex driving behaviors than others. Thus, it was shown that the driving color maps using DSAE could represent complex driving behaviors more clearly than the other methods.
The results of this experiment and our qualitative observation shown in the previous subsection were consistent. In the next section, we performed a subjective evaluation experiment to determine whether the driving color map based on the DSAE helps users to review and understand driving behavior intuitively or not.
D. Subjective Evaluation of the Driving Color Map
To clarify whether our proposed method allowed users to identify different driving behaviors easily or not, we conducted a subjective evaluation experiment. We asked nineteen participants to complete a questionnaire in order to evaluate the driving color maps obtained using DSAE and other methods. All, except two of these participants, had a driver’s license. To perform subjective evaluations, we developed a modified version of Deep Car Watcher, as shown in Fig. 4, where A~E indicate the names of our proposed visualization method and four other methods used for comparative purposes. In this subjective evaluation, we provided participants with the modified version of Deep Car Watcher (Fig. 4).
The modified version of Deep Car Watcher for subjective evaluation, which shows driving videos and the driving color maps obtained using PCA (A), FastICA (B), KPCA (C), SAE (D), and DSAE (E). This software does not show the name of the methods, but instead it indicates them as A~E to label the driving color maps.
We considered that each participant had a different way of distinction of driving behaviors. For example, some people may consider that accelerating from a standstill and accelerating from a constant speed cruise are different driving behaviors, whereas others consider them as the same. The experiment consisted of two parts. First, we asked the participants to identify driving behavior patterns that they could recognized by observing driving videos. The driving behaviors recognized by the participants were used in the next part of the evaluation task. Second, we focused on the quality of the visualization methods, and determined if each method help participants to distinguish the driving behavior patterns. We asked the participants to compare driving behavior patterns and colors on the driving color maps and to compare the five driving color maps. We asked them to rank the driving color maps according to the criterion of “the same color represents the same driving behavior”. It took about one hour for each participant to complete the questionnaire. Half of the time was spent searching for driving behaviors in the driving videos. All of the participants identified the simple driving behaviors, but only eight participants could identify the simple and complex driving behaviors considered in the previous experiments (see Table III). A Few participants found specific driving behaviors such as “coasting deceleration after turning right.” This suggested that different people have different understanding about distinctive driving behavior.
The results of the ranking are shown as histograms in Fig. 5. The vertical axis represents the order, and the horizontal axis represents the number of participants who selected a specific order. Figure 5 shows that the counts for the driving color map produced using DSAE ranked first compared with the others. The thick solid line in the histogram indicates the median value of the rankings. The driving color maps obtained using the SAE and DSAE had the highest median values.
Rankings for the driving color maps obtained using PCA, FastICA, KPCA, SAE, and DSAE given by nineteen subjects. The thick solid line shows the median value of the rankings.
We compared the ranking results by using statistical tests. Given that the ranking results followed an ordinal scale, we employed the Wilcoxon rank-sum test [30], which is a non-parametric statistical test. We set the null hypothesis
We found that eight participants recognized some complex driving behaviors, which showed that more than half of the participants evaluated the visual results based only on the simple driving behaviors, so these participants gave negative evaluations for the visualized complex driving behaviors. To address this issue, we divided the participant ranking results into two groups: the ranking results given by the eleven participants who only recognized simple driving behaviors and those given by the eight participants who recognized both simple and complex driving behaviors. Figure 6 shows the experimental results for the eleven participants who only recognized simple driving behaviors. There were no significant differences among the results according to the Wilcoxon rank-sum test. Next, Fig. 7 shows the experimental results for the eight participants who recognized both the simple and complex driving behaviors. They demonstrate that the driving color maps obtained using SAE had the highest median ranking value, followed by DSAE. It should be noted that the driving color maps obtained using DSAE were always ranked in the first and second places according to the results. We also used the Wilcoxon rank-sum test to verify these results. The p-values for the comparisons of the rankings for the driving color maps obtained using DSAE and PCA, and DSAE and KPCA were less than 0.001. In addition, the p-values for the comparisons of the rankings for the driving color maps obtained using DSAE and ICA, SAE and PCA, and SAE and KPCA were less than 0.05. Thus, pairs of driving color maps received significantly different rankings from the participants. We can reject
Rankings for the driving color maps obtained using PCA, FastICA, KPCA, SAE, and DSAE given by eleven subjects who only identified the simple driving behaviors in driving videos. There were no significant differences among all the methods according to the Wilcoxon rank-sum test. The thick solid line shows the median value of the rankings.
Rankings for the driving color maps obtained using PCA, FastICA, KPCA, SAE, and DSAE given by eight subjects who identified the simple and complex driving behaviors in driving videos. The thick solid line shows the median value of the rankings.
Thus, we verified that our proposed method, driving color maps using DSAE, performed well at representing different driving behaviors with different colors. It helped participants to distinguish driving behaviors from visualization results.
Applications of Driving Color Maps Using DSAE
After verifying the effectiveness of visualization of driving color maps using DSAE, we focused on the practical applications of the proposed method.
A. Detection of Interesting Patterns in Driving Behaviors
To determine interesting patterns in driving behaviors, we compared the driving color maps obtained using DSAE based on multiple circuits of the same course. One of the interesting patterns was “pedestrian on a pedestrian crossing in front of the experimental vehicle,” as shown in Fig. 8. In this figure, the red circles denote the pattern with pedestrians on a pedestrian crossing or with a car in front. In contrast, the blue boxes denote the pattern without pedestrians on the crossing. When a pedestrian was on the pedestrian crossing, a dark gray-blue color appeared suddenly on the driving color map. By contrast, the color changed in a smooth manner without a pedestrian. In the driving video of circuit 1, another car was in front of our experimental vehicle and people were on the pedestrian crossing in front of the car. This suggests that the driving color map helps users to notice actual changes in driving behaviors on the circuit. This will enable users to identify interesting and precarious driving behaviors easily, particularly when they monitor driving behaviors and investigate traffic accidents.
Samples of visualization results: the top two parts of the figure show that a pedestrian was on a pedestrian crossing or a car was in front, and the lower part of the figure shows situations where there are no pedestrians on the crossing. These results show the visualization results were clearly affected by the difference of driving behaviors caused by different situations.
B. Visualization of Driving Behaviors on a Public Road
The driving behavior data obtained on the course inside the company’s factory were used in the experiments described above. To verify if our proposed visualization method is suitable under driving conditions on a public road, and to verify its generalization performance, we asked another participant to drive the experimental vehicle around Hidaka Park in Nagoya, Japan. We obtained driving behavior data for 2064 steps, which included the nine types of sensor information that we used previously. We input the new data into our method, which was trained using the data described in previous subsections. The driving color map obtained for public road around Hidaka Park is shown in Fig. 9.
Visualization result of the driving behavior on the public road around Hidaka Park by using the trained DSAE.
We used Deep Car Watcher to identify driving behaviors in the driving videos and to assign their corresponding colors. We compared the corresponding colors generated from the data for the factory road and the data for the public road, and these colors are shown in Table IV. There was slight difference for turning left with constant speed between the colors in the two data sets, but the rest were similar. This difference occurred because most of the corners on the factory road were right angles, whereas some of the corners on the public road did not require a great turn of the steering wheel. We found that the maximum steering angle was 620 [deg] and the minimum steering angle was −624 [deg] for the factory data after examining the raw data. For the public road data, the maximum steering angle was 486 [deg] and the minimum steering angle was −297 [deg]. This explains why the colors generated from the factory data and public road data were slightly different. In general, it was suggested that our proposed visualization method is effective for a public road, and its performance was highly robust and applicable to data obtained in a new environment.
Conclusion
In this study, we proposed a visualization method for driving behaviors called the driving color map based on the hidden features extracted using DSAE which is type of unsupervised deep learning methods. We employed DSAE to extract the three-dimensional hidden features from driving behavior data. The method obtained the corresponding colors by projecting the extracted three-dimensional hidden features to the RGB color space. Diving color map was obtained by placing these colors in the corresponding positions on the map.
Based on numerical evaluations and subjective experiments, we demonstrated that the proposed visualization method, driving color map using DSAE, allows people to distinguish different driving behaviors more easily than the other comparative methods. Finally, we presented some examples of the practical application of the proposed method and verified its suitability for use on public roads.
Future challenges are as follows. The DSAE is an unsupervised feature extraction model. Therefore, there are rotational degrees of freedom between the feature spaces. This suggests that the visualization result can yield different visualization results on rotation in the color space even if it uses the same data. To clarify how to deal with the arbitrariness, i.e., the degree of freedom, appropriately is one of our future challenges. To develop practical support systems using the proposed visualization method is also our future challenge. In this paper, we did not evaluate our method from the viewpoint of a self-coaching system though one of the purposes of the visualization is to help users to review their driving behavior and improve them. Takeda et al. [18] developed a self-coaching system using driving behavior data and a GMM-based driver behavior models. They showed their system’s potential benefit through an experiment. Developing a self-coaching system using our proposed method and showing its benefit are also our future challenge.
Exploring the possible application of DSAE-based feature extraction for intelligent vehicle is another future challenge. We suggest that our proposed method can be used to provide information about previous driving behavior. For example, a teacher at a driving school could share a driving color map with a student after they have completed a driving lesson. If the teacher finds some colors that correspond to dangerous driving behaviors, they could give the student suitable guidance while showing it. As Another example, when an accident investigator need to search for clues in a large amount of raw data, he could use the driving color map to quickly and easily understand the driving behaviors to determine the cause of the accident. Our method does not tell whether driving behaviors are safe or dangerous, but it can still help users to look through and investigate driving behaviors.
In this paper, it is shown that DSAE has an excellent capability to extract hidden features from driving behavior data. Making use of the DSAE-based feature extraction method for various purposes is also our future challenge. We expect that the extracted hidden features can be used in various ways effectively, such as driving behavior segmentation, inference of driving intention, and to learn driving strategies from drivers.