Introduction
As the essential medium for iron-steel making process, byproduct gaseous energy plays a pivotal role in real-world production. It is of significant importance to reasonably schedule the generation and consumption of the energy so as to sustain a secure and stable manufacturing condition. Without prediction information, the scheduling scheme can be developed on the basis of experience only, which naturally leads to the lack of efficiency and accuracy. Therefore, identifying future tendencies of the byproduct gas system becomes a primary demand stemming from the practical field, which will be beneficial for the scheduling work.
Regarding the prediction of steel industry, data-driven approaches are commonly deployed as the solution, such as nonlinear state-space modeling [1], Relevance Vector Machine [2], Gaussian Process (GP) [3], Neural Networks (NNs) [4]–[5], Kalman filtering [6]–[8], etc. However, due to the involvement of the iteration mechanism present, the models reported in these literatures can only produce a short period of forecasts, i.e., 30–60 points. While compared with the methods that focus on the single data and numeric prediction, Granular Computing (GrC) [9]–[11] concentrates on processing information granules such as fuzzy sets [12]–[13], rough sets [14]–[15], interval sets [16]–[17], etc. As such, the prediction horizon becomes extended and this offers the possibility of realizing long-term prediction. For instance, considering the semantics of industrial processes, a GrC-based long-term prediction model was proposed in [18] with the prediction length of 480, 720, and 1440 points. Adaptive data granulation coming with their intuitively perceived characteristics and combined with the collaborative conditional clustering method, [19] led to a scale-varying prediction model where the prediction horizon ranges from 60 to 180 points. In [20], the study involved structural characteristics of the energy network and established a hybrid collaborative fuzzy clustering method for predicting the storage amount of Linz-Donawitz converter Gas (LDG) realized for long horizons, i.e., 720 and 1,440 points. Although the studies above successfully produced some practical models with extended prediction horizons, the formation of final results was restricted to numeric values. With the development of various applications, the reliability of the prediction results becomes highly required. As a result, it is necessary to propose an effective approach for the construction of prediction intervals (PIs).
The common way of constructing PIs is based on the involvement of statistical vehicles, such as Mean Variance Estimation (MVE) [21]. However, the accuracy of MVE based method greatly depends on whether the related prediction model can precisely estimate the mean value of the target. The study [22] reported an instance of NN-based construction of PIs incorporating fuzzy member. Considering that the modeling considerations in this case are specialized in the field of renewable energy, the method cannot be directly applied to gaseous system of steel industry. Concentrating on electricity consumption forecasting in power system, [23] proposed a long-term probability forecasting model based on fuzzy Bayesian theory and expert prediction. While owing to the requirements on constructing accurate fuzzy membership functions, it is hardly to be deployed in energy system of steel industry for the complex pipeline network. Also, it primarily studied the data relationship among years rather than minutes in this study. Furthermore, kernel based dynamic Bayesian networks, Long Short Term Memory (LSTM) were also utilized to construct PIs, refer to [24]–[26]. Although these approaches can generate interval-like results, they failed to satisfy practical requirements of the long predicting length.
Aiming at providing accurate long-term result along with the reasonable reliability measurement, a GrC-based hybrid hierarchical method, we propose to construct PIs for byproduct gaseous energy system in steel industry. The contributions of this study can be summarized as follows.
On the one hand, the prediction horizon is successfully extended to be long; this is accomplished by modeling completed at the level of information granules instead of single numeric data. On the other hand, the PIs are constructed by vertically extending the numeric values to intervals; it offers a novel granulation method which brings bi-directional considerations.
The established hierarchical structure involving three layers of processing with the consideration of industrial periodic features of the process exhibit substantially positive impact on both efficiency and accuracy. The corresponding programming models are established at each layer of the architecture to optimize the PIs.
In order to further clarifying the importance of this paper, a comparative table is given as Table 1 to make contrast with other existed long-term PIs construction methods. Besides, the experimental study is concerned with the relationships between different parameters of the model and the performance of the proposed approach. Subsequently the optimized PIs are formed and compared with the ones being constructed by traditional methods.
The paper is organized as follows. Section II covers some preliminaries, i.e., we look at the description of the industrial problem and traditional GrC-based long-term prediction model. Section III provides the detailed elaboration on the proposed GrC- based hybrid hierarchical PIs construction method. Then experimental studies are reported for real-world data coming from steel industry; Section IV covers studies on both parameters and PIs construction to demonstrate the accuracy and efficiency of the proposed model. Finally, some conclusions and possible future topics are included in Section V.
Preliminaries
A. Industrial Problem Description
A typical structure of the byproduct gaseous system used in steel industry in China is illustrated in Fig. 1. It consists of three types of gas, i.e., Blast Furnace Gas (BFG), Coke Oven Gas (COG), and Linz-Donawitz converter Gas (LDG). Coke ovens, Blast Furnaces (BFs) and Converters (marked as LD in the figure) are the generators who produce the gases during iron-steel production. The included users are hot/cold rolling, Plate plant, etc. Besides these units, some tanks are also involved in the network for temporarily storing the gases. The pipeline along with press/mixture stations is organized for gas distribution completed across the entire system.
For a variety of reasons, imbalance always arises where the operating crews need to find a solution for fulfilling the gap between the generation and consumption. Without any estimation completed ahead, the energy scheduling work would be blind as well as inaccurate and of low efficiency. While owing to the complicated pipeline network and the large number of gas units, it is usually difficult for the staff to accurately estimate the future trend of gas flows for making a reasonable adjustment. As such, prediction of gas flows becomes necessary for steel industry.
Considering that the accumulated process data still can be sufficiently utilized, data-driven methods become a common solution for solving the complicated industrial problem. The target variables for PIs construction of this study are the generators and consumption units.
Considering that the accumulated process data still can be sufficiently utilized, data-driven methods become a common solution for solving the complicated industrial problem. The target variables for PIs construction of this study are the generators and consumption units.
B. Traditional GRC-Based Long-Term Prediction Model
Assume a time series is divided into a collection of data segments, i.e., \begin{equation*} \hat { \boldsymbol {u}}_{k}=f\left ({\boldsymbol {u}_{k-n_{I}},\cdots, \boldsymbol {u}_{k-2}, \boldsymbol {u}_{k-1} }\right)\tag{1}\end{equation*}
For building this mapping, various approaches could be considered such as polynomial fitting [30], neural networks [31] etc. Here the numeric long-term prediction result is expressed in the form \begin{equation*} \hat { \boldsymbol {s}}_{ \boldsymbol {k}}=\frac {\sum _{i=1}^{c} \hat {u}_{ki} \boldsymbol {v}_{ \boldsymbol {k}}}{\sum _{i=1}^{c} \hat {u}_{ki}}\tag{2}\end{equation*}
Obviously, the format of the final results \begin{equation*} \left [{ \mathbf {V}^{-},\mathbf {V}^{+} }\right]=[\mathbf {V}- \boldsymbol {\varepsilon },\mathbf {V}+ \boldsymbol {\varepsilon }]\tag{3}\end{equation*}
The dimensionality of
equals to the one of the final prediction results, which is typically greater than 480 samples when considering a long prediction horizon. The optimization of such scale is comparatively of low efficiency, and sometimes even not feasible which makes this approach not suitable for this real-world application.\boldsymbol {\varepsilon } The determination of the values of
requires a well-structured optimization mechanism. It should consider not only the practical aspects of the data, but also the coverage along with the specificity of the information granules into account which would be noticeably beneficial for achieving an accurate construction of PIs.\boldsymbol {\varepsilon }
GRC-Based Hybrid Hierarchical Method for PIs Construction
To overcome the shortcomings of the existing approach, a GrC-based hybrid hierarchical method is proposed for the PIs construction of gaseous energy system in steel industry. The practical requirements can be summarized with regard to the two aspects, i.e., a long prediction horizon and the interval-valued results. The long-term prediction is directly related to the modelling concerning the x-axis as well as horizontal one, in which the numeric forecasts are generated by probabilistic based approach and can be deemed as the basis of the following PIs. The construction of intervals is concerned with the vertical direction by using granulation technique again in y-axis. An illustration of this bi-directional hybrid construct is displayed in Fig. 2.
In this section, the modeling process is illustrated from the perspective of horizontal and vertical modeling first. Then the detailed computing procedures are briefly summarized.
A. Horizontal – GRC-Based Probabilistic Approach for Numeric Long-Term Prediction
In order to extend the prediction horizon, a granulation carried out with respect to the
Some definitions are given to deliver the required prerequisites.
Definition 1:
Probability of a specific prototype \begin{equation*} p\left ({\boldsymbol {v}_{ \boldsymbol {i}} }\right)\doteq \frac {\sum _{j=1}^{N} {\mathbb {I}_{j}\{u_{kj}=\mathrm {max}(\boldsymbol {u}_{ \boldsymbol {k}})\}}}{N}\tag{4}\end{equation*}
\begin{equation*} \mathbb {I}_{j}\left \{{u_{kj}=\max \left ({\boldsymbol {u}_{ \boldsymbol {k}} }\right) }\right \}=\begin{cases} 1&k=i \\ 0&k\ne i \\ \end{cases}\tag{5}\end{equation*}
Definition 2:
Probability of a data segments \begin{equation*} \boldsymbol {p}\left ({\boldsymbol {s}_{ \boldsymbol {i}} }\right)\doteq {[p\left ({\boldsymbol {v}_{\mathbf {1}} }\right)\!,p\left ({\boldsymbol {v}_{\mathbf {2}} }\right)\!,\cdots,p\left ({\boldsymbol {v}_{ \boldsymbol {c}} }\right)] }^{T}\tag{6}\end{equation*}
Definition 3:
A \begin{equation*} \mathbb {P} \doteq \{{p}(\boldsymbol {v}_{ \boldsymbol {i}}\vert \boldsymbol {v}_{ \boldsymbol {j}})\vert i=1,2,\cdots,c;j=1,2,\cdots,c\}\tag{7}\end{equation*}
Generally, the model aims at estimating the corresponding probabilities. Based on the above definitions, the mechanism of prediction can be expressed in the form \begin{equation*} \hat { \boldsymbol {p}}\left ({\boldsymbol {s}_{ \boldsymbol {k}} }\right)= \boldsymbol {p}\left ({\boldsymbol {s}_{ \boldsymbol {k}-\mathbf {1}} }\right) \boldsymbol \circ \mathbb {P}\tag{8}\end{equation*}
\begin{equation*} \hat { \boldsymbol {s}}_{ \boldsymbol {k}}=\frac {\sum _{i=1}^{c} \hat {p}_{ki} \boldsymbol {v}_{ \boldsymbol {k}}}{\sum _{i=1}^{c} \hat {p}_{ki}}\tag{9}\end{equation*}
Such an approach extends the prediction horizon. It is worth noting that this approach involves far less parameters in comparison to the number of parameters used when dealing with neural networks, polynomial fitting, etc., which makes it more suitable for real-world application.
B. Vertical – Hierarchical Structure for Optimization on the Information Granularities
Intervals can be constructed by augmenting numeric prototypes by building them around their specific numeric values. In such a way, every result of prediction comes with the corresponding upper and lower boundary. A large number of parameters would also be generated leading to the reduction of the computational efficiency.
While it should be noted that the data of byproduct gaseous system in this study always exhibits industrial features which in practical denote production procedures. Take the generation of #3 BFG as an example as shown in Fig. 3. Here the periodic patterns can be clearly distinguished here. Two typical phases are involved in each pattern, i.e., air blasting (the amplitude ranges from 400km3/h to 550km3/h in approximately 30 minutes) and temporary suspending (the amplitude ranges from 250km3/h to 450km3/h in approximately 30 minutes). Based on such type of prior knowledge, the levels of information granularities can be hierarchically optimized. This arrangement will result in higher computing efficiency.
By taking into consideration the concerns of efficiency and PI’s performance, we develop a hierarchical structure as shown in Fig. 4. The numeric data are now generalized as intervals located at the 1st layer with
The levels of information granularities are sequentially allocated starting from the bottom layer to the top one. In contrast, the optimization of the parameters is carried out in an opposite direction which means
Coverage: The coverage measure concerns the count the number of cases when predicted intervals “cover” the original experimental data. As such, Mean Coverage Error (MCE) which refers to the absolute difference between Empirical Coverage Probability (ECP) and Nominal Coverage Probability (NCP) are employed and the MCE is taken in the form \begin{equation*} MCE=\left |{ ECP-NCP }\right |\tag{10}\end{equation*}
\begin{align*} ECP=&\frac {1}{T}\sum \nolimits _{i=1}^{T} c_{i} \tag{11}\\ NCP=&\left ({1-\alpha }\right)\times 100\%\tag{12}\end{align*}
Specificity: In order to quantify the specificity of the levels of information granularities, Mean Prediction Intervals Width (MPIW), is considered here \begin{equation*} MPIW=\frac {1}{T}\sum \nolimits _{i=1}^{T} \left |{ \bar {s}_{i}-\underline {s}_{i} }\right |\tag{13}\end{equation*}
Based on the formulation expressed above, we take the two measures into account and consider their product, which is regarded as an objective function. The optimization problem is subject to the constraints imposed on the levels of information granularity. The details are presented as follows
Second layer – industrial feature-based segments
where\begin{align*}&\hspace {-0.5pc}\min \vert \frac {1}{n_{2}}\!\sum \nolimits _{i=1}^{n_{2}} c_{i,0} \!-\!\left ({1\!-\!\alpha }\right)\!\times \!100\% \vert \!\times \! \left({\frac {1}{n_{2}}\!\sum \nolimits _{i=1}^{n_{2}} \left |{ \overline s _{i,0}\!-\!\underline {s}_{i,0} }\right | }\right) \\[-2pt]&\hspace {-0.5pc}s.t.~\frac {1}{n_{2}}\sum \nolimits _{i=1}^{n_{2}} \varepsilon _{i,0} =\varepsilon \\[-2pt]&\hspace {-0.5pc}\hphantom {s.t.~} \alpha _{min}\varepsilon \le \varepsilon _{i,0}\le \alpha _{max}\varepsilon \\[-2pt]&\hspace {-0.5pc}\hphantom {s.t.~}\underline {s}_{i,0}\ge 0\tag{14}\end{align*} View Source\begin{align*}&\hspace {-0.5pc}\min \vert \frac {1}{n_{2}}\!\sum \nolimits _{i=1}^{n_{2}} c_{i,0} \!-\!\left ({1\!-\!\alpha }\right)\!\times \!100\% \vert \!\times \! \left({\frac {1}{n_{2}}\!\sum \nolimits _{i=1}^{n_{2}} \left |{ \overline s _{i,0}\!-\!\underline {s}_{i,0} }\right | }\right) \\[-2pt]&\hspace {-0.5pc}s.t.~\frac {1}{n_{2}}\sum \nolimits _{i=1}^{n_{2}} \varepsilon _{i,0} =\varepsilon \\[-2pt]&\hspace {-0.5pc}\hphantom {s.t.~} \alpha _{min}\varepsilon \le \varepsilon _{i,0}\le \alpha _{max}\varepsilon \\[-2pt]&\hspace {-0.5pc}\hphantom {s.t.~}\underline {s}_{i,0}\ge 0\tag{14}\end{align*}
is the prescribed overall level of information granularity, along with the specific one\varepsilon for each granule in 2nd layer.\varepsilon _{i,0} and\alpha _{min} are deployed to constrain\alpha _{max} from assuming excessively high or low values which are far away from the predetermined level\varepsilon _{i,0} .\varepsilon is the indicator defined asc_{i,0} in (11).c_{i} and\overline s_{i,0} refer to the upper and lower bounds of constructed PIs with\underline {s}_{i,0} , respectively.\varepsilon _{i,0} First layer – specific data points
The optimization completed for 1st layer can be illustrated as a number of \begin{align*}&\hspace {-0.5pc}\min \vert \frac {1}{n_{1}}\!\sum \nolimits _{j=1}^{n_{1}} c_{i,j} \!-\!\left ({1\!-\!\alpha }\right)\!\times \! 100\% \vert \!\times \! \left({\frac {1}{n_{1}}\sum \nolimits _{j=1}^{n_{1}} \left |{ \overline s _{i,j}-\underline {s}_{i,j} }\right | }\right) \\&\hspace {-0.5pc}s.t.~\frac {1}{n_{1}}\sum \nolimits _{j=1}^{n_{1}} \varepsilon _{i,j} =\varepsilon _{i,0} \\&\hphantom {\hspace {-0.5pc}s.t.~}\beta _{min}\varepsilon _{i,0}\le \varepsilon _{i,j}\le \beta _{max}\varepsilon _{i,0} \\&\hphantom {\hspace {-0.5pc}s.t.~}\underline {s}_{i,j}\ge 0 \\&\hphantom {\hspace {-0.5pc}s.t.~}i=1,2,\cdots,n_{2}\tag{15}\end{align*}
In view of the independence present among the problems of 1st layer, the
C. Computing Procedures
To clarify the computing procedures of the proposed GrC based hybrid hierarchical method, Fig. 5 outlines an overall framework.
The detailed processing is outlined in the form:
Step 1:
Divide the raw data into segments
concerning industrial features. Deploy unsupervised clustering method, i.e., FCM, to acquire numeric prototypes\mathbf {S}=\{ \boldsymbol {s}_{\mathbf {1}}, \boldsymbol {s}_{\mathbf {2}},\cdots, \boldsymbol {s}_{ \boldsymbol {N}} \}\!, \boldsymbol {s}_{ \boldsymbol {i}}\in \boldsymbol {R}^{ \boldsymbol {n}} and related fuzzy membership degrees\mathbf {V}=\left \{{ \boldsymbol {v}_{\mathbf {1}}, \boldsymbol {v}_{\mathbf {2}},\cdots, \boldsymbol {v}_{ \boldsymbol {c}} }\right \}\!, \boldsymbol {v}_{ \boldsymbol {i}}\in \boldsymbol {R}^{ \boldsymbol {n}} as the results of horizontal modeling.\mathbf {U}=\left \{{ \boldsymbol {u}_{\mathbf {1}}, \boldsymbol {u}_{\mathbf {2}},\cdots, \boldsymbol {u}_{ \boldsymbol {N}} }\right \}\!, \boldsymbol {u}_{ \boldsymbol {i}}\in \boldsymbol {R}^{ \boldsymbol {c}} Step 2:
Vertically extend the prototypes as
, where[v_{ij}-\varepsilon _{i,0},v_{ij}+\varepsilon _{i,0}] denotes the levels of information granularity in 2nd layer.\varepsilon _{i,0},i=1,2, \cdots,n_{2} Step 3:
Solve the optimization problem (14) with the PSO algorithm. This process involves establishment of co-occurrence matrix
for both upper and lower boundaries according to Definition 3, calculation of the corresponding forecasted probability by (8) and construction of PIs by (9), which can be regarded as a supervised learning procedure.\mathbb {P} Step 4:
Further distribute information granularities
on the prototypes which are now as\varepsilon _{i,j}, i=1,2, \cdots,n_{2};j=1,2, \cdots,n_{1} . Then optimize them via a set of problems as (15) in parallel strategy.[{v_{ij}-\varepsilon }_{i,j}-\varepsilon _{i,0},{v_{ij}+\varepsilon }_{i,j}+\varepsilon _{i,0}] Step 5:
Based on the optimized
and\varepsilon _{i,j} , the final PIs can be obtained via centroid decoding technique (9).\varepsilon _{i, 0}
Experimental Studies
The experiments reported in this study begin with PIs construction for Mackey-Glass time series. Then they are also implemented using real-world data coming from February 1st, 2017 to June 1st, 2017, of which the collection rate is 1 point/minute. Three significant objects, i.e., #3 BFG generation, #1 COG generation and BFG consumption of #2 hot blast stove are selected as representatives for experimental analysis. For retaining the continuity of the data, the training and validation set are constructed by consecutive data in this study. And the following results are all based on testing set. Parallel Computing Toolbox in MATLAB R2017a is used for the optimization of the 1st layer. The developing environment is listed as follows.
Hardware:
CPU: Intel(R) Core(TM) i7-6950X CPU @ 3.00GHz
Memory: Corsair Vengeance DDR4 3000Mhz 16Gb
GPU: Nvidia Titan xp
Motherboard: Asus ROG Rampage V Edition 10
Hard drive: Samsung NVMe SSD 960 PRO 512g
First, parametric studies of the prescribed settings of the proposed method are given to discuss the determination of the values of the parameters. Then the PIs construction experiments are conducted, in which the comparative methods are MVE, Bootstrap, Gaussian Process (GP) and traditional GrC approach. The numeric values for the MVE are predicted with the aid of the LSSVM method (including parameters: \begin{align*} PICP=&\frac {1}{n_{test}}\sum \nolimits _{i=1}^{n_{test}} c_{i} \times 100\% \tag{16}\\ PINAW=&\frac {1}{n_{test}\left ({t_{max}-t_{min} }\right)}\sum \nolimits _{i=1}^{n_{test}} \left |{ \overline s _{i,0}-\underline {s}_{i,0} }\right | \tag{17}\\&\hspace {-3.5pc}L\left ({\overline s _{i,0},\underline {s}_{i,0},t_{i} }\right) \\=&\frac {1}{n_{test}}\sum \nolimits _{i=1}^{n_{test}} l_{i} \times 100\% \\ l_{i}=&\begin{cases} -2\alpha \!\left ({\overline s_{i,0}\!-\!\underline {s}_{i,0} }\right)\!-\!4(\underline {s}_{i,0}-t_{i}),\quad if~t_{i} < \underline {s}_{i,0} \\ -2\alpha \!\left ({\overline s_{i,0}-\underline {s}_{i,0} }\right)\!,\quad if~\underline {s}_{i,0}\le t_{i}\le \overline s_{i,0} \\ -2\alpha \!\left ({\overline s_{i,0}\!-\!\underline {s}_{i,0} }\right)\!-\!4(t_{i}\!-\!\overline s_{i,0}),\quad if~t_{i}>\overline s_{i,0} \\ \end{cases}\tag{18}\end{align*}
A. Parametric Study
To carry out parametric study, we randomly select the #3 BFG generation as the representative. In order to synthetically articulate the relationship between the performance and the parameter, we integrated PICP, PINAW and Interval Score as one index \begin{equation*} I=PICP\times (1-\overline {PINAW})\times \overline {L\left ({\overline s _{i,0},\underline {s}_{i,0},t_{i} }\right)}\tag{19}\end{equation*}
Among the parameters of the proposed method, the most important one is the overall level of information granularity
Another parameter that needs to be discussed here is the number of prototypes. Similarly, Fig. 7 displays the relationship between the values of
B. Optimized PIs Results
1) Mackey-Glass
As a typical chaos time series benchmark, Mackey-Glass is deployed here to validate the proposed method. It can be depicted from the elaborative comparison in Fig. 9 that the MVE, Bootstrap and GP give unsatisfied results owing to the iterative error. Also, a considerable number of points are out of the PI constructed by the traditional GrC-based model. While the proposed hybrid hierarchical model exhibits superiority on not only accuracy including PICP, PINAW and Interval Score, but also computing efficiency, which can be also concluded from the statistics in Table 2.
PIs construction results for Mackey-Glass time series: (a) MVE (b) Bootstrap (c) GP (d) Traditional GrC (e) Proposed method.
2) #3 BFG Generation
The generation amount of BFG plays a significant role for the whole gaseous energy network, which is directly related to iron production. Experiments utilizing different methods are shown in Fig. 10. Due to the iteration mechanism, the MVE only performs well in about first half an hour. The error increases with the increase of the length of the prediction horizon so that the constructed PI turns to exhibit oscillations. Although the Bootstrap and GP both provide a series of periodic features, they fail to accurately estimate the time span and amplitude of each period. While the traditional one-layer GrC-based model exhibits an improved performance, which roughly covers the experimental data. However, it is noted that some data are out of the PIs. Comparing these three approaches, the proposed method apparently behaves much better with regard to coverage. Besides, the constructed PIs also exhibit significant specificity.
PIs construction results for #3 BFG generation: (a) MVE (b) Bootstrap (c) GP (d) Traditional GrC (e) Proposed method.
In order to further demonstrate the performance of the methods, error statistics are given in Table 3 along with the parameter settings. Although MVE exhibits some advantage over PINAW, the PICP and Interval Score are far worse than for other methods. The proposed one not only guarantees the accuracy of PIs, but also requires less computing time as forming the most efficient solution.
3) #1 Cog Generation
Another essential byproduct gaseous energy for iron-steel making process, i.e. COG, is addressed here of which the generation amount is deployed as an example. Due to the continuous burners switching in practice, this data exhibits a frequent variation, which makes it hard to be accurately predicted. Fig. 11 presents the elaborative comparison of constructed PIs along with Table 4 lists error statistics and parameters. Due to the huge amounts of parameters to be determined in only one layer, the traditional GrC exhibit inconsistent performance on the widths of PIs so that the statistics are not satisfied. The results clearly indicate the absolute superiority by the proposed method to the others on PICP, PINAW and interval score. It is also noticeable that the computational cost is reduced to a great extent because of the hierarchical structure and the parallel strategy being used.
PIs construction results for #1 COG generation: (a) MVE (b) Bootstrap (c) GP (d) Traditional GrC (e) Proposed method.
4) BFG Consumption of #3 LD Converter
The last part of the experimental study concerning a consumption unit comes from LDG network, of which the results are shown in Fig. 12. The PIs constructed by MVE and GP are far from satisfaction since it accumulated iterative errors and also ignored the industrial features during modeling process. The Bootstrap only performs well on the first 30 points and then exhibits low specificity values. Although the traditional one-layer approach successfully forecasted the general trend of this data, the coverage still behaves sometimes in a poor manner. Also, given that the computing time is about 1 minute, which could possibly exceed the required prediction frequency, it is not suitable for application in a scenario of real-world production. Whereas the proposed method exhibits superior performance, which can be also demonstrated based on the statistics shown in Table 5.
PIs construction results for BFG consumption of #3 hot blast stove: (a) MVE (b) Bootstrap (c) GP (d) Traditional GrC (e) Proposed method.
C. Practical Running Instance
On basis of the proposed hierarchical GrC method for PIs construction, a software system is developed and implemented in the same steel plant of China. The coding language is C++ for the algorithm and C# the GUI, all of which are in Microsoft™.NET Framework 4.0. The data is collected by industrial SCADA system on-site and then stored in the database with the system of IBM™ DB2 9.0.
A practical running instance is depicted in Fig. 13. The prediction objects are listed in the boxes above and the statistics beneath. A figure showing constructed PIs are given at the bottom. Based on the prediction results, a warning is consequently activated and displayed on top stating ‘Current imbalance system: BFG’. Based on such reliable guidance, the operating staff can decide to either start scheduling work or cancel this alarm.
After applied this system on-site for 6 months, the PICP, PINAW and Interval Score has been improved by around 7%, 10% and 9%, respectively. The constructed PIs successfully provided helpful information for the gaseous scheduling operation.
Conclusion
By taking into account of the practical requirements imposed on long-term PIs construction for byproduct gaseous energy system of steel industry, a GrC-based hybrid hierarchical method has been proposed where information granules were allocated at each layer of the architecture. Then the related optimization techniques were established with the consideration of both coverage and specificity criteria. The employment of industrial semantics in data granulation substantially improves the accuracy of the proposed method, while the utilization of parallel strategy effectively guarantees the efficiency. The experimental study involved various error statistics and produced the optimized PIs. It can be concluded that the proposed method came with better performance, which is helpful for energy scheduling and optimization work of steel industry.
Some possible future topics are worth pursuing. On the one hand, the optimization problem can be formulated as the multi-objective one. Under such circumstances, a suitable method is required for solving it with both satisfied accuracy and reasonable combination with industrial practice. On the other hand, the forecasted probability is related here to the one of only one-step ahead, as shown in (8). Such a mechanism can be also extended as n-1 tuple, which may lead to complex matrix operation. Besides, PSO can be replaced by other optimization alternatives, such as reinforcement learning, so as to avoid the complexity of determining parameters.