Introduction
As the “lifeblood of the industry,” crude oil (CO) is the most significant strategic component raw element in contemporary industrial society, the key to prosperity and national security, and the cornerstone of civilization [1]. It is connected to how the global economy is growing, and different economic data will be more significantly impacted by changes in oil prices. Predicting how the price of crude oil will fluctuate in the future is therefore quite important. Scholars have used a range of research techniques to perform in-depth analyses and forecasts of global oil prices from the perspective of diversification. The artificial intelligence model and the econometric model, and the integrated forecasting model are roughly the three components of these research methodologies, according to the summary of the literature examined. For the models in econometrics The authors in [2] examined how well various ARIMA-GARCH (Generalized Autoregressive Conditional Heteroscedasticity) models performed in simulating and predicting the conditional mean and volatility of weekly crude oil prices. The researchers in [3] apply the findings of the ARIMA model compared with those of the Decomposition-based vector autoregressive model (DVAR), which is used to forecast the monthly price data for WTI crude oil. The authors in [4] examined the predictive power and impact of the Google index on CO price by incorporating it into the ARMA-GARCH and ARIMA models. By contrasting the suggested MMGARCH (Mixture Memory GARCH model) to other discrete unpredictable models. Klein and Walther [5] extend the literature on predictable and esteem predicting CO price returns. Stelios et al [6] compare the VAR model’s capacity for forecasting to that of the Random Walk (RW) model and the AR model. The results are listed at the top of Table 1. Generally, the economic model assumes that the data are stable, regular, and linear. Under this assumption, the economic model can accurately predict the CO price. The international crude oil market, however, exhibits complex non-linear, and multidimensional characteristics of crude oil price movements. The intricate features concealed in crude oil may be too complicated for these conventional metering methods to detect. Support vector regression (SVR), Artificial Neural Network (ANN), Random Forest (RF), and other widely used non-linear models are applied to CO price forecasting, successfully fitting the non-linear CO price series as a result of the rapid growth of artificial intelligence. For example, the researchers in [7] utilized a neural network and genetic algorithm to predict the price of WTI CO. In the same way, the authors in [8] utilized a neural network to forecast the price term structure of crude oil futures. Fan et al [9] utilize an Imperialist Competitive Algorithm and Support Vector Regression (ICA-SVR) techniques to forecast the price of crude oil. Mostafa and EI-Masry [10] projected CO price using gene expression programming (GEP); CO price is forecasted by Gao and Yalin [11] based on stream learning. The artificial intelligence model outperformed the conventional paradigm, in line with empirical research. A single artificial intelligence (AI) model cannot correctly represent the dynamic changes of complicated CO price time series responsible for the significant variations in the time series. However, an AI model can accurately anticipate non-linear and non-stationary sequences. However, the hybrid forecasting model overcomes the drawbacks of time series instability and nonlinearity and enhances the CO price prediction accuracy by combining a range of methodologies. In the past few years, integrated models for predicting the price of CO have developed quickly. Tao et al [12] suggested a more effective EMD-SBM-FNN model that can capture the intricate dynamics of the price of crude oil. Zhang et al [13] introduced EEMD-PSO-LSSVM-GRACH a novel hybrid approach to forecast CO prices. Yu et al [14] used the EEMD-DCD-LSSVR model to predict the price of CO. The authors in [15] estimated the price of CO, by using bootstrap aggregation (bagging) and Stacked Denoising Auto Encoders (SDAE). In the same way, the authors in [16] use the EEMD-RVFL model to predict the price of CO. Moreover, the authors in [17] used the EEMD-EELM-ADD model as a unique decomposition-ensemble technique for predicting CO prices. Ding [18] created a hybrid model EEMD-ANN-ADD for predicting the price of CO. The authors in [19] use the DFN-AI model to predict the CO price. Similarly, the authors in [20] use the VMDICA-ARIMA hybrid model to predict the price of CO. In the same way, Zhang et al [21] suggested an algorithm for iterated combinations to predict the CO price. The authors in [22] combine RW and ARMA to predict the CO price. Similarly, Zheng et al [23] proposed EEMD and Dynamic Artificial Neural Network (DANN) to forecast the CO price. The authors in [24] showed load prediction, based on the long short-term memory (LSTM) model, based on Back Propagation Neural Network (BPNN) and Local Mean Decomposition (BPNN-LMD-LSTM). The design is based on a fixed-time consistency algorithm with random delay to predict the economic dispatch of microgrids. The authors in [25] proposed a landslide displacement prediction model, the local mean decomposition-bidirectional long short-term memory (LMD-BiLSTM), which depends on the time-frequency analysis method. The authors Heng Sun [26] utilizes method in three steps exhibits great potential applications in the RUL prediction of rotating machines. The authors in [27] LSTM, wavelet threshold denoising (WTD), and LMD have been integrated into a novel combined model called LMD-WTD-LSTM to estimate short-term gas consumption. In the same way, the authors in [28] introduced a new model which enhanced the accuracy of the predictions. The novel technique called variational mode decomposition (VMD) and used to predict the major factor time series utilizing its secondary factors. A new technique called multiscale forecasting model is introduced that produced an optimal forecast [29]. This model outperformed the compared model to forecast the complex time series data. In the same way, the authors in [30] decomposed the data into many features via VMD. Then the mutli-features are trained with machine learning classifiers. The authors in [31] forecasted the Daily PM2.5 and PM10 data employing a Robust LMD (RLMD) and moving window ensemble technique was done using linear and nonlinear modelling frameworks. The research mentioned above claims that the hybrid model mixes single models so that the benefits of each model balance out the drawbacks of the other models. As a result, the hybrid model is superior to the single model and offers us research suggestions. From the above discussions the following research questions have been generated. How can the end-point impact be eliminated due to the complicated dynamic change of the crude oil price time series and additional information obtained on different frequencies of the crude oil price data itself? How the calculations for the hybrid model may be streamlined. How can CO price predictions be made more accurately? In the current study, we use Local Mean Decomposition (LMD) and an artificial intelligence model to forecast the price of CO: (1) Utilizing LMD to decompose the time series of CO price in an adaptive manner, removing the end-point effect, and further exploring the data’s various frequencies. (2) This work uses average mutual information (AMI) to decrease the calculation amount while taking into account the growth of the hybrid model’s calculation amount; (3) By separating the time series into random and stochastic variables, the econometric model is able to represent the volatility features of the crude oil price time series; (4) combining the outcomes of the prediction utilizing LSTM; (5) The experimental results demonstrate that the LMD-SD-ARIMA-LSTM suggested in this study outperforms a single model in terms of crude oil price prediction accuracy. They also demonstrate that the traditional econometric models can increase prediction accuracy through decomposition and aggregation. Researchers are still working on these problems. In comparison to previous studies, the novel hybrid model LMD-SD-ARIMA-LSTM has reduced the volatility and solved the issue of the overfitting problem of neural networks. The proposed hybrid technique is validated using publicly accessible data from the West Texas Intermediate (WTI), and forecast accuracy are compared using accuracy measures.
The organization of the study is as follows. Section I consists of an introduction and a literature review. Section II provides a brief description of the methods used in this study. In the same way, sections II-D and III consist of analysis and discussion along with a conclusion respectively.
Methodology
A. Local Mean Decomposition
Using adaptive time-frequency analysis, LMD is a technique for handling non-stationary signals [32]. Separating various envelope signals and purely frequency-modulated signals from the original signals is the foundation of the LMD approach. A physical significant product function (PF) component of the instantaneous frequency can be derived by multiplying the envelope signals with sole frequency modulated signals. The decomposition procedure for the initial signal x(t) can be broken down into five steps:
Select all local extremum points
of the original signal x(t) and calculate the meann_{i} of adjacent extremum pointsm_{i} ,n_{i} and envelope estimaten_{i+1} \alpha _{i}: The envelope estimate\begin{equation*} {\alpha }_{i}= \frac {\left |{ n_{i}-n_{i+1} }\right |}{2} \tag{1}\end{equation*} View Source\begin{equation*} {\alpha }_{i}= \frac {\left |{ n_{i}-n_{i+1} }\right |}{2} \tag{1}\end{equation*}
and local means\alpha _{i} are then used to smooth using the moving average tom_{i} and envelope estimate functionm_{11}(t) ;\alpha _{11}(t) Ignore the local mean function
in the original signal x(t), that is:m_{11}(t) \begin{equation*} {h}_{11}(\text {t})=\text {x(t)}-m_{11}(t) \tag{2}\end{equation*} View Source\begin{equation*} {h}_{11}(\text {t})=\text {x(t)}-m_{11}(t) \tag{2}\end{equation*}
Dividing it by
(t) is the amplitude demodulated.\alpha _{11}\left ({t }\right),h_{11} purely frequency modulated signal is Repeated iterations n times until\begin{equation*} {s}_{11}(\text {t})=h_{11}(\text {t})/\alpha _{11}(t) \tag{3}\end{equation*} View Source\begin{equation*} {s}_{11}(\text {t})=h_{11}(\text {t})/\alpha _{11}(t) \tag{3}\end{equation*}
(t). Stopping iteration should be done whens_{1n} (t) =1._{n\rightarrow \infty }\lim \alpha _{1n} The corresponding envelope
(t) and the first component\alpha _{1} are obtained:{PF}_{1}\left ({t }\right) \begin{align*} {\alpha }_{1}(\text {t})&=\alpha _{11}(\text {t})\alpha _{12}(\text {t})\ldots \alpha _{1n}(\text {t}) \\ &=\prod \nolimits _{q=1}^{n} {\alpha _{1q}\mathrm {(t)}} \tag{4}\\ {PF}_{1}(\text {t})&=a_{1}(\text {t})s_{1n}(\text {t}) \tag{5}\end{align*} View Source\begin{align*} {\alpha }_{1}(\text {t})&=\alpha _{11}(\text {t})\alpha _{12}(\text {t})\ldots \alpha _{1n}(\text {t}) \\ &=\prod \nolimits _{q=1}^{n} {\alpha _{1q}\mathrm {(t)}} \tag{4}\\ {PF}_{1}(\text {t})&=a_{1}(\text {t})s_{1n}(\text {t}) \tag{5}\end{align*}
The component
(t) is removed from x(t), the new signal{PF}_{1} is obtained and the process is repeated k times until the signalu_{1}(t) is a constant the oscillations have stopped, too. Finally, the original signal x(t) can be written asu_{k}(t) \begin{align*} x(\text {t})=\sum \nolimits _{p=1}^{k} {PF}_{p} \left ({t }\right)+u_{k}\left ({t }\right),\text {t}=1 \ldots \text {n} \tag{6}\end{align*} View Source\begin{align*} x(\text {t})=\sum \nolimits _{p=1}^{k} {PF}_{p} \left ({t }\right)+u_{k}\left ({t }\right),\text {t}=1 \ldots \text {n} \tag{6}\end{align*}
B. ARIMA
Box and Jenkins first proposed the ARIMA model in the early 1970s. The following structure of the model is said to be the autoregressive integrated moving average model, which is defined as ARIMA (p,d,q): The dependent variable must be stationary (through the I-component), and the independent variables are taken as all lags of the dependent variable (the AR-component) and/or errors lags. In general, therefore, one might consider an ARIMA model to be a specific kind of regression model (the MA component).
In general, a model with ARMA looks like this:\begin{equation*} {Y}_{t}= \alpha _{0}+ \sum \nolimits _{i=1}^{p} \alpha _{i} Y_{t-1}+ \sum \nolimits _{j=1}^{q} \beta _{i} \varepsilon _{t-j}+ \varepsilon _{t} \tag{7}\end{equation*}
C. LSTM
It is a type of RNN with the capacity to take long-term dependability into account. Scientists The authors in [34] developed the LSTM in 1997. Because LSTMs can retain information over a longer amount of time and do not have long-term dependencies, they differ from other RNN techniques. Inside, LSTMs operate similarly to other RNN methods employing neural network gates and layers. They have a chain structure. The LSTM’s construction is designed to have a cell that runs the length of the device. Gates are used to control whether or not data may be transferred into the cell state. Additionally, there are parts known as gated cells that enable the storage of data from earlier LSTM outputs; this is the place where the memory-related aspects of LSTM come into play.
An advanced soft computing technique known as LSTM was developed from the Recurrent Neural Network (RNN). One of the numerous Artificial Neural Network (ANN) techniques, the RNN, was developed to address the ANN’s weakness in handling time correlation in the data sequence and enhances neurons in the networks with canonical connections to make it possible for RNN to create a sequence-to-sequence mapping between input and output data [24]. Unfortunately, the long-range dependencies are still a challenge for traditional RNN, which have difficulty learning the long-term temporal correlations due to expanding gradients or, conversely, vanishing gradients [25]. The authors in [34] used LSTM memory cells to get around this restriction. These cells use a three-gate mechanism made up of an input gate, an output gate, and a forget gate to store the temporal state of the networks [35]. Figure 1 shows an LSTM cell with all three of those gates as well as the cell state [36].
LSTM gates are simply used to limit the amount of information that can be transferred. They typically consist of a layer of a sigmoid neural network and an action of pointwise multiplication. While the forget gate is used to use selectively forget the information in the cell state, the input gate decides what new information will be stored in the current cell state. The output gate is then utilized to determine the value that we wish to output [22]. The forget gate is the initial component of the LSTM cell. It can be expressed as follows and is used to regulate the magnitude to forget the previous cell’s concealed state:\begin{equation*} f_{t} = (W_{f}h_{t-1} + U_{f}x_{t} + b_{f}), \tag{8}\end{equation*}
\begin{align*} i_{t}& = (Wih_{t-1}+ Uix_{t}), \tag{9}\\ {C}_{t}&= \text {tanh}(W_{c}h_{t-1}+U_{c}{x}_{t}+b_{c}). \tag{10}\end{align*}
The (current) cell state (\begin{equation*} C_{t}= f_{t}\odot C_{t-1}+i_{t}\odot \tilde {C}. \tag{11}\end{equation*}
The output gate is used to determine the next hidden state in the last step (i.e., the value of the current hidden state, h). First, we use the sigmoid function as given in Eq to transmit the previously hidden state value (\begin{align*} {O}_{t}&=(W_{\mathrm {o}} h_{t-1}+U_{\mathrm {o}}x_{t}+b_{o}) \tag{12}\\ h_{t}&= O_{t} \odot \text {tanh}(C_{t}) \tag{13}\end{align*}
D. Evaluation
In this study, metrics are assessed using three different prediction error criteria: Mean Absolute Error (MAE), Root Mean Squared Error (RMSE), and Mean Absolute Percentage Error (MAPE). The last one displays the level of inaccuracy in a percentage number, whereas the first two display the error level. As shown by Shahid et al in [37] and Hansun et al in [36] and [38], All three of those requirements can be stated as:\begin{align*} MAE&=\frac {1}{n}\sum \nolimits _{t=1}^{n} {\vert \text {Yt}{-} \text {Ft} \vert } \tag{14}\\ RMSE&=\sqrt {\frac {1}{n}\sum \nolimits _{t=1}^{n} {({\mathrm {Yt {-} Ft)}}^{2}}}, \tag{15}\\ MAPE&=\frac {1}{n}\sum \nolimits _{t=1}^{n} {\frac {\mathrm {\vert Yt - Ft \vert }}{\mathrm {Yt}}}. \tag{16}\end{align*}
\begin{equation*} DM=\frac {\bar {k}}{\sqrt {Var(\bar {k})}}, \tag{17}\end{equation*}
Empirical Analysis
A. Statistical Description of Data
West Texas Intermediate (WTI) data from the U.S. Energy Information Administration (EIA, https://www.eia.doe.gov) in this study, as depicted in Figure 1. The datasets for WTI crude oil from March 12, 2018, to February 14, 2022, having a total of size 999 samples were chosen. For forecasting accuracy, the data is divided into 75/25 training and testing parts respectively. The training part of a total of 749 samples comprised from March 12, 2018, to February 23, 2021, and the testing part of size 250 is from February 24, 2021, to February 14, 2022. A dramatic downward trend in the price of WTI can be seen in Figure 1 graphical analysis up until 11 January 2022. This is followed by another downward trend in the price of WTI up until 11 May 2020, which was seen up until 30 May 2020. It is simple to spot a most pronounced upward trend from May 12, 2020, to February 14, 2022. In contrast to other months, the seasonal plot in Figure 2 demonstrates that the annual variance is still minimal from January to May for year 2020 and from September to December for the year 2018. PP and ADF statistics are shown in Table 1 demonstrating the sequence is not stationary. Jarque-Bera (J-B) statistics, however, indicate that the data is normal.
B. Data Decomposition and Reconstruction
This study proposed the reconstruction of PFs that are obtained from LMD. PFs from LMD is separated into two components, Deterministic and Stochastic. The stochastic and Deterministic PFs are modeled separately, and different models are selected for stochastic and deterministic components, ARIMA and LSTM models are then fitted for every stochastic and deterministic component all individual deterministic PFs are then combined for final forecasting. Figure 3 illustrates the suggested method’s entire framework. WTI price data is decomposed by LMD. For the LMD method, The maximum number of generating function components is 20 and the maximum number of iterations is set to 30. The decomposition results demonstrate WTI CO price sequence is composed of three PFs and one residual.
The CO price sequence is divided into three PFs and a residual series, with frequency changes from high to low and a decreasing changing trend for the residual series, as shown in the aforementioned Figure 3. Then, considering the different influences of the decomposed PFs on the original series, the PFs are reconstructed according to the mutual information of each PF. AMI is also used to recreate the IMFs. AMI is only a visual assessment of the plots that are produced for each dataset’s PFs, as shown in Figure 3, From the second PF to the fourth PF, the AMI plots are seen to follow the same pattern, so the first four PFs are shown in Figure 4, As a result, the first PF is regarded as stochastic, while the rest are all deterministic. To create two components, add from second to fourth as deterministic and the first PF for stochastic. The Mutual information is calculated according to Eq. (9), by adding all deterministic series new series is formed called virtual product functions (VPF) is obtained and displayed in Figure 5. Stochastic and deterministic graphs are displayed in Figure 6.
Here, PF1 is considered to be a stochastic component, while PF2, PF3, and PF4 are treated as deterministic. Stochastic and deterministic (SD) components are treated separately and then combined to make SD components.
1) ARIMA Model
The basic assumption to apply the ARMA model is the stationarity of the time series model. Successive differences are taken to make time-series stationery [41], [42]. ADF test is used to determine whether the time series is stationary [43]. After obtaining a stationary time series appropriate model for the time series by choosing the AR and MA words in the right order For selection, ACF and PACF plots of the best order are employed. The Adam algorithm is used in the Model for training and cross-validation. A 75:25 ratio of data is designated for training and testing, accordingly. Ljung-Box (LB) test is used for checking the model adequacy. Forecasting is done by fitting the ARIMA model to training and testing periods. By fitting data with the ARIMA model, we used R software. Figure 7 displays the residual of the WTI ARIMA fitted model with ACF lag and Residual plot.
The first and second plot of the above figure shows that residuals are uncorrelated, and data became stationary after taking lag1. Whereas the final LB statistics figure demonstrates that p-values are greater than 0.05 for all datasets [44]. A hypothesis shows no serial autocorrelation among the fitted residual is not significant. So, All of these techniques can offer the most accurate future projection. The forecasting accuracy of the models is shown below.
PF1 and PF2 test thinks that the residuals are distributed normally, and the series is stationary. However, PF3, PF4, and VPF need to be differenced. PF3 and PF4 have a difference of order 2 and the difference order number of VPF is 1. Table 3 Box test results indicate that the p-value is greater than 0.05. This amount exceeds 0.05 as a significance level and hence the hypothesis that the residuals are white noise could not be rejected. Accept the supposition that there is no residual autocorrelation. ARIMA is successfully established. All four PFs are fitted for the LMD-ARIMA model whereas PF1 (stochastic component) and VPF(deterministic component) are added and fitted for the LMD-SD-ARIMA model.
2) Stack-LSTM
The residual of LMD-ARIMA and LD-SD-ARIMA is processed in a hybrid stack LSTM model for forecasting. In this technique output of the LSTM layer becomes the sequence of vectors to be used as an input to the subsequent LSTM layer. It has been seen in the Figure 8 input layers are used again in the second layer and multiple hidden layers are stacked one on top of another hidden layer. A three-dimensional input layer is required for the LSTM layer. RETURN_SEQUENCE=TRUE enables using the buried LSTM layer’s 3D results as the next layer’s input. It makes the model deeper and more accurate as a deep learning technique. M networks are made up of many LSTM hidden layers. A Deep Recurrent Neural Network is formed by several hidden layers (DRNN). Iterative weight updating utilizing training data is crucial for training the LSTM network. The stochastic gradient descent approach employs the Adaptive Movement Estimation (Adam) algorithm for weight updating [45]. For constructing and training the LSTM model Keras neural network API [46] uses Tensorflow and [47] developed Tensorflow library written in Python. For numerical calculations, Tensorflow is employed as a machine learning framework. In contrast, Keras has a steep learning curve. Together, they deepen and improve the model’s accuracy.
In this paper, in our dataset, we represent each trading day of every month as a
Hyperparameters that will increase accuracy and reduce the chance of overfitting the data are chosen [49]. The overfitting drop-out technique, which randomly selects cells within a layer based on the probability in such a way that their output is set to 0, is used to reduce overfitting as developed by Srivastava et al [50]. As a result, the ideal dropout is chosen at 30%, which gives the lowest MSE. Epoch was selected to 100 while performing this test, with one epoch equaling one iteration of all training data that was processed by the network in [51]. Each layer’s LSTM cells were programmed to have the following values: 41, 41, 64, 1, decay to 0.2, and window length to 22. We divided the training data into a batch size, with a batch size of 32, to facilitate the propagation of the training data across the network. This indicates that the network is trained using the first 31 examples (0–31) from the training data, followed by the subsequent 31 samples (32–63). One epoch has been transmitted through the network once all samples have been propagated through it, hence the epoch continues until that happens [49].
In order to hasten learning, the loss function is used to calculate the difference between the desired output and the LSTM model output during the training. User-specified validation data, which we have defined to be 10% of the training data, is what we want as the output. We have taken the MSE as a loss function because it is frequently utilized for forecasting the time series [52]. We utilized the optimizer Adam to construct the LSTM model due to its superior results and quick convergence when other optimizers are compared, the authors in [49] gave advised us to utilize it as default. We take the decay as 0.3 while using optimizer Adam. With the dropout set to 30% and the optimizer Adam’s decay set to 0.3, Figure 10 depicts the validation loss and training of our LSTM model by using the best hyperparameter setting. Figure 10 shows that the model and data are most closely matched and that the MSE loss reduces with increasing epoch values.
Results and Discussions
The proposed new hybrid forecasting model LMD-SD-ARIMA-LSTM, which is compared with the ARIMA, LSTM, LMD-ARIMA, LMD-SD-ARIMA, LMD-ARIMA-LSTM methods for predicting the CO price of WTI. Root mean square error (RMSE), Mean absolute error (MAE), and mean absolute percentage error (MAPE), all are used to calculate WTI data and to evaluate the efficiency of the model. Table 5 provides an overview of each model’s performance accuracy.
In comparison to other models, the hybrid LMD-SD-ARIMA-LSTM achieved the lowest RMSE of 0.150; as a result, this technique is useful for predicting the price of crude oil. Second, for the conventional econometric model, accuracy measurements of the individual model (i.e., ARIMA) and decomposition-ensemble models (i.e., LMD-ARIMA and LMD-ARIMA-LSTM) are compared, and the accuracy of the latter is lower than the former in forecasting. However, for the machine learning model, the three accuracy measures of the individual model (i.e. LSTM) and decomposition-ensemble models (i.e. LMD-ARIMA-LSTM and LMD-SD-ARIMA-LSTM) are compared, Decomposition-ensemble models’ accuracy metrics are lower than those of an individual model. It shows that after decomposition and reconstruction, the performance of the conventional econometric model for predicting oil prices can be improved. LSTM model fitting of the LMD-ARIMA and LMD-SD-ARIMA residual model is shown in Figure 11. The predicting results of seven separate individual models for WTI crude oil price data are shown in Figure 12 as relative error histograms.
Twenty days of forecasting results for all seven models are shown in Figure 13 which are compared with 20 days of original WTI oil prices from
Moreover, Table 6, shows that all the models are significantly different from each other.
Figures 14 and 15 compare the proposed model’s errors to those of the other models and provide a graphic comparison of the results of oil price forecasts using all seven models. The forecasting performance of LMD-SD-ARIMA-LSTM is the best, as shown by the three figures. Additionally, compared to the short- and medium-term forecast, the LMD-SD-ARIMA-LSTM technique performs better in the long-term forecast.
For improved presentation, actual data is now compared with the function that divides up each five-day forecast. For the testing data set, the original residuals and predicted values five days ahead are compared and presented in Figure 16.
The average error for each step is analyzed separately. Usually, the tendency of RMSE decreases according to the extrapolated periods due to uncertainties. As predicted, at each step the tendency of RMSE decreases. As in
The RMSE for all five steps is presented in Figure 17.
Conclusion
Crude oil is one of the non-renewable power sources and is the lifeblood of contemporary industry. The stability of the global economic market benefits from accurate crude oil price forecasting. The LMD-SD-ARIMA-LSTM hybrid prediction approach, which is discussed in this study, is based on the LMD, ARIMA, and LSTM methodologies. The proposed hybrid technique is validated using the WTI CO prices. This study decrease the efforts by collecting the stationary and non-stationary IMFs into stochastic and deterministic components with improved accuracy. The investigation demonstrates, in comparison to the other five approaches, the novel hybrid method significantly increases the prediction accuracy of the CO price. Additionally, the results demonstrate that the conventional econometric model can enhance oil price prediction accuracy following decomposition and reconstruction. Moreover, the new hybrid forecasting system performs better when predicting the price of CO over the medium and long term. Meanwhile, accurate predictions can provide reasonable advice for relevant departments in order to make correct decisions.
A. Limitations of the Study
In this study, we only used the univariate time series data.
B. Future Recommendations
The current study can be extended using LSTM based on EEMD and other decompositions methods. Moreover, we can extend the current study into the bivariate and multivariate data.
In the future, some other traditional econometric forecasting models and other machine learning methods will be explored and studied. The factors influencing crude oil prices will also be taken into account, and it will be further investigated to see if the novel hybrid forecasting approach is appropriate for multi-variate forecasting of crude oil prices.