Loading web-font TeX/Math/Italic
Deep Neural Network Ensemble for the Intelligent Fault Diagnosis of Machines Under Imbalanced Data | IEEE Journals & Magazine | IEEE Xplore

Deep Neural Network Ensemble for the Intelligent Fault Diagnosis of Machines Under Imbalanced Data


This paper takes the advantages of ensemble learning and proposes an ensemble convolutional neural network (EnCNN) for intelligent diagnosis of machines using imbalanced ...

Abstract:

Imbalanced classification using deep learning has attracted much attention in intelligent fault diagnosis of machinery. However, the existing methods use individual deep ...Show More

Abstract:

Imbalanced classification using deep learning has attracted much attention in intelligent fault diagnosis of machinery. However, the existing methods use individual deep neural network to extract features and recognize the health conditions under imbalanced dataset, which may easily over-fit the mechanical data and affect the diagnosis accuracy. To deal with this problem, this paper takes the advantages of ensemble learning and proposes an ensemble convolutional neural network (EnCNN) for the intelligent fault diagnosis for machines under imbalanced data. In the proposed method, a convolutional neural network with the input of multi-sensor signals is used as the base classifier. Firstly, the mechanical imbalance dataset is first split into balanced training subsets through under-sampling strategy, and each subset is used to train a base classifier. Then the weight coefficients of each trained base classifier are calculated by G-mean score and anomalous base classifiers are screened using classifier selection. Finally, the base classifiers are integrated into EnCNN through weighted voting strategy. The proposed EnCNN is validated by the imbalanced dataset collected from a machinery fault test bench. By comparing with the related methods, the superiority of EnCNN is verified in intelligent fault diagnosis of machines under imbalanced data.
This paper takes the advantages of ensemble learning and proposes an ensemble convolutional neural network (EnCNN) for intelligent diagnosis of machines using imbalanced ...
Published in: IEEE Access ( Volume: 8)
Page(s): 120974 - 120982
Date of Publication: 03 July 2020
Electronic ISSN: 2169-3536

Funding Agency:

References is not available for this document.

CCBY - IEEE is not the copyright holder of this material. Please follow the instructions via https://creativecommons.org/licenses/by/4.0/ to obtain full-text articles and stipulations in the API documentation.
SECTION I.

Introduction

Machines are widely used in the field of aerospace, electric energy, machinery manufacturing and others [1]. In the operation stages of the machines, defects may occur on their key components, leading to the appearance of a fault. Such faults may cause unforeseen breakdown and economic loss. Therefore, intelligent fault diagnosis plays an irreplaceable role in ensuring the reliable operation of the machines [2]. Since deep learning allows deep neural networks to accomplish the tasks of feature extraction and fault classification, the research of the intelligent fault diagnosis of machines using deep learning receives much attention [3], [4].

Nowadays, deep learning based intelligent fault diagnosis methods are widely researched [5]–​[8]. For instance, Zhao et al. [9] presented a multi-scale convolutional transfer learning network for intelligent fault diagnosis of rolling bearings under variable working conditions. Tang et al. [10] developed a method using an adaptive learning rate deep belief network combined with Nesterov momentum for rotating machinery fault diagnosis. Fuan et al. [11] presented an adaptive deep convolutional neural network for rolling bearing fault diagnosis. Wang et al. [6] proposed a batch-normalized deep neural networks for feature learning and fault recognition in the intelligent fault diagnosis of machines. Through analysis of these studies, it can be found that although deep learning has been applied in the intelligent fault diagnosis, they have not considered the imbalanced distributions of mechanical health conditions. In real applications, the machines are operating normally in the most time, and the faults seldom happen during the operation. Consequently, the data samples of slight faults of the machines are harder to be collected than those of normal condition, and the data samples of serious faults are harder to be collected that those of slight faults. Thus, the data samples of normal condition are abundant while the data samples of fault conditions are relatively scarce, making the distribution of the mechanical data imbalanced. Such a distribution makes the intelligent fault diagnosis of machines suffer the imbalanced classification problem [12]. As shown in Figure 1, if the dataset is balanced, the samples of normal condition, slight fault and serious fault can be classified correctly. However, if the dataset is imbalanced, the imbalanced problem forces deep learning models to be biased towards the data of the majority health conditions such like the normal condition, and the data of the minority health conditions such like slight fault and serious fault are not adequately learned, leading to the misclassification of the data of these faults.

FIGURE 1. - Simulation of the intelligent fault diagnosis of a machine: (a) a balanced classification case, and (b) an imbalanced classification case.
FIGURE 1.

Simulation of the intelligent fault diagnosis of a machine: (a) a balanced classification case, and (b) an imbalanced classification case.

To deal with the imbalanced classification problem in the intelligent diagnosis of machines, researchers have proposed some methods. Jia et al. [13] presented a framework called deep normalized convolutional neural network for intelligent fault diagnosis, which uses normalized layers and weighted loss to overcome the imbalanced classification problem. Zhou et al. [14] designed a global optimization generative adversarial network to solve the imbalanced classification problem in the intelligent fault diagnosis. Dong et al. [15] proposed deep cost adaptive convolutional network for the imbalanced mechanical data classification. Wang et al. [16] used Wasserstein generative adversarial network to generate simulated signals and train stacked autoencoders to classify the mechanical health conditions under imbalanced data. However, these studies use individual deep network to extract features and recognize the health conditions under imbalanced dataset, which may suffer two weaknesses as follows. (1) The individual deep network may easily over-fit the training data and be biased by random factors, affecting the accuracy and stability of the intelligent fault diagnosis. (2) The hyper-parameters of the deep networks are manually selected using human empirical knowledge.

To overcome the aforementioned weaknesses, this paper takes the advantages of ensemble learning and proposes an ensemble convolutional neural network (EnCNN) for the intelligent fault diagnosis for machines under imbalanced data. In the proposed method, convolutional neural network that uses multiple sensor data as its input is applied as the base classifier in the ensemble learning. Firstly, an under-sampling strategy is used to split the imbalanced dataset of machines into several balanced sub-datasets. Then, the hyper-parameters of base classifiers are initialized by the random selection strategy, and these classifiers are trained using the balanced sub-datasets. Finally, the trained base classifiers are integrated into the EnCNN by weight voting strategy and anomalous classifier selection. The proposed EnCNN is validated by the imbalanced dataset collected from a machinery fault test bench. By comparing with the related methods, the superiority of EnCNN is verified in the intelligent diagnosis of machines under imbalanced data.

The contributions of this paper are summarized as follows.

  1. EnCNN uses the hyper-parameter random selection strategy to set the hyper-parameter for each base classifier automatically and improve the diversity of these classifiers, which is beneficial for an easy and good ensemble in EnCNN.

  2. G-mean score is used as the ensemble weight in the voting strategy and a boxplot-based classifier selection is developed to screen out the anomalous base classifiers, resulting in better classification accuracy in the intelligent fault diagnosis of machines under imbalanced data.

The rest of this paper is organized as follows. Section II details the concept of ensemble learning. In Section III, the proposed EnCNN is described in detail. In IV, the diagnosis cases of a machinery fault test bench with imbalanced data are studied using EnCNN. Finally, conclusions are drawn in Section V.

SECTION II.

The Ensemble Learning

This section introduces the concept of ensemble learning. An individual classifier may over-fit training data and affected by random factors, leading to the biased classification accuracy. Thus, ensemble learning is proposed to combine multiple individual classifiers that are called base classifiers to form an ensemble classifier [17]. It reduces the accuracy deviations made by an individual classifier and tends to overcome the overfitting phenomenon and yield better results.

The advantages of ensemble learning can be illustrated using Hoeffding’s inequality. Assuming that the base classifier h_{t} is trained by training data \boldsymbol {x}^{m} and the ideal classifier for classification is \varphi . The error of the base classifier h_{t} for classification is \varepsilon .\begin{equation*} P(h_{t} ({\boldsymbol {x}}^{m})\ne \varphi ({\boldsymbol {x}}^{m}))=\varepsilon\tag{1}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where t=\mathrm {1, 2, \ldots,}T and m=1, 2, \ldots,M .

By combining these base classifiers with the voting method, the ensemble classifier H can be represented as \begin{equation*} H({\boldsymbol {x}}^{m})=\textrm {sgn}\left ({{\sum \limits _{t=1}^{T} {h_{t} ({\boldsymbol {x}}^{m})}} }\right)\tag{2}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where \mathrm {sgn(\cdot)} is symbolic function returning −1 if the condition is less than 0, and 1 otherwise.

If the base classifiers are independent of each other, the upper limit of the classification error P for the ensemble classifier is known by the Hoeffding’s inequality \begin{equation*} P(H({\boldsymbol {x}}^{m})\ne \varphi ({\boldsymbol {x}}^{m}))\le \exp \left ({{-\frac {1}{2}T\left ({{1-2\varepsilon } }\right)^{2}} }\right).\tag{3}\end{equation*}

View SourceRight-click on figure for MathML and additional features.

It can be seen that the upper limit of the ensemble classifier error decreases exponentially as the number of base classifiers increases, and the final error tends to be zero [18]. Therefore, the ensemble classifier will achieve higher accuracy than the base classifier alone.

In ensemble learning, the base classifiers are required to be both independent each other and complement each other, namely the diversity of the base classifiers. Krogh and Vedelsby [19] developed the concept of “error-ambiguity decomposition” to analyze the effect of the diversity, which shows significance for ensemble learning. The error-ambiguity decomposition can be represented as \begin{equation*} E=\overline E -\bar {A}\tag{4}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where E is the generalization error of the ensemble classifier, \bar {E} is the average generalization error of the base classifiers, and \bar {A} is the ensemble ambiguities representing the diversity of the base classifiers. In (4), it can be seen that the higher accuracy the base classifiers are and the greater the diversity is, the better the ensemble is. Thus, the diversity of the base classifiers is important in ensemble learning [20].

SECTION III.

The Proposed Method

In this section, we introduce the proposed method ensemble convolutional neural network (EnCNN) for the intelligent fault diagnosis of machines under imbalanced data. In the proposed EnCNN, a convolutional neural network is used as the base classifier, as shown in Figure 2(a). The flowchart of EnCNN is displayed in Figure 2(b). The imbalance dataset is firstly split into balanced training subsets through under-sampling strategy, and each subset is then used to train a base classifier. Then the weight coefficients of each trained base classifier are calculated by G-mean score and anomalous base classifiers are screened using classifier selection. Finally, the base classifiers are integrated into EnCNN through weighted voting strategy.

FIGURE 2. - (a) The convolutional neural network used as the base classifier, and (b) the flowchart of the proposed EnCNN.
FIGURE 2.

(a) The convolutional neural network used as the base classifier, and (b) the flowchart of the proposed EnCNN.

A. The Base Classifier

Since CNNs are able to deal with the shift variant signals like vibration signals, we use it used as the base classifier in this paper. As shown in Figure 2(a), CNNs have four kinds of layers: input layer, convolutional layer, pooling layer and fully connected layer. The input layer is used for the input of multi-sensor signals, the convolutional layer and pooling layer are combined as a building block for the feature learning, and fully connected layer is used for the fault classification [21].

Given a training set \left \{{\boldsymbol {x}^{m},{y}^{m} }\right \}_{m=1}^{M} where \boldsymbol {x}^{m} is the m th training sample collected from the multi-sensors of the machine, its label is {y}^{m} and M indicates the number of training samples. In the feature learning stage, several feature learning blocks are stacked to achieve the feature learning purpose from the training samples. Concretely, in the l th feature learning block, its convolutional layer uses kernel \boldsymbol {k}^{l}\in \boldsymbol {R}^{J\times D\times H} to extract the feature vectors, where J is the amount of the kernels, D and H is the depth and height of the l th kernel. Suppose that \boldsymbol {x}^{m,l-1} is the feature extracted from \boldsymbol {x}^{m} in the feature learning block l-1 , the output feature \boldsymbol {u}^{m,l} of the l th convolutional layer is calculated by \begin{equation*} {\boldsymbol {u}}_{j}^{m,l} =\sigma \left ({{\sum \limits _{d} {{\boldsymbol {k}}_{j,d}^{l} \ast {\boldsymbol {x}}_{d}^{m,l-1} +b_{j}^{l}}} }\right)\tag{5}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where \boldsymbol {u}_{j}^{m,l} is the j th column of \boldsymbol {u}^{m,l} , j=1, 2, \ldots,J , \sigma (\cdot) is the activation function of rectified linear units, \boldsymbol {x}_{d}^{m,l-1} is the d th column of \boldsymbol {x}^{m,l-1} , d=1, 2, \ldots,D , \boldsymbol {k}_{j,d}^{l} is an H -dimension vector of \boldsymbol {k}^{l} used for convolution, and b_{j}^{l} is bias value.

After obtaining \boldsymbol {u}^{m,l} , the pooling layer of the l th feature learning block is used to conduct max-pooling and output its learned feature {\boldsymbol {x}}^{m, l} , which can be obtained by \begin{equation*} {\boldsymbol {x}}_{j}^{m, l} =\textrm {max}({\boldsymbol {u}}_{j}^{m, l},s)\tag{6}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where s is the pooling size that indicates the size of a non-overlapping segment of \boldsymbol {u}_{j}^{m,l} , and \textrm {max(}\cdot) represents the function that calculates the max value from each non-overlapping segment.

In the fault classification stage, fully connected layers are combined as the classification block to recognize the faults of the machine. If y^{m}\in \left \{{1, 2, \ldots,C }\right \} and p_{c}^{m} is the probability of assigning c th label to \boldsymbol {x}^{m} , the Softmax loss function of CNN is obtained by \begin{equation*} \ell =-\frac {1}{M}\sum \limits _{m=1}^{M} {\sum \limits _{c=1}^{C} {1\left \{{ {y^{m}=c} }\right \}\log \left ({{p_{c}^{m}} }\right)}}\tag{7}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where 1\left \{{\cdot }\right \} is the indicator function returning 1 if the condition is true, and 0 if the condition is false.

In (7), the Softmax loss function measures the classification error of CNN during training, and this error is used to adjust the CNN parameters. However, its loss function assumes that the contribution of each sample to the classification error is the same. When the numbers of different fault samples are different, the contribution of different fault to the classification error is different, prompting CNN to correctly classify the majority class and misclassify the minority class. Thus, traditional CNN cannot handle the imbalanced classification in intelligent fault diagnosis of machines.

B. The Balanced Subsets for Training

To deal with the training of CNN using imbalanced data, we first use under-sampling strategy [22] to randomly obtain multiple balanced subsets of samples and then these subsets are applied to train the base classifiers. This random data selection strategy could improve the data diversity in the training process of EnCNN and is described as follows.

For the training set \left \{{\boldsymbol {x}^{m},{y}^{m} }\right \}_{m=1}^{M} , we calculate the number of training samples for each health condition of machines, and the number of samples for the c th class is \begin{equation*} n_{c} =\sum \limits _{m=1}^{M} {1\left \{{{y^{m}=c} }\right \}}.\tag{8}\end{equation*}

View SourceRight-click on figure for MathML and additional features.

The number of samples for each health condition n_{\mathrm {min}} is calculated using (9) so as to obtain the sample number for the health condition having the minimum samples.\begin{equation*} n_{\min } =\textrm {min}\left ({{\{n_{c} \}_{c=1}^{C}} }\right)\tag{9}\end{equation*}

View SourceRight-click on figure for MathML and additional features.

For each health condition, 0.7n_{\mathrm {min}} samples are randomly selected from the samples of this health condition. The selected samples are combined to form a balanced training subset S_{t} and used for training the base classifier h_{t} .\begin{equation*} S_{t} =\left \{{{{\boldsymbol {x}}^{m_{t}}, y^{m_{t}}} }\right \}_{m_{t} =1}^{M_{t}}\tag{10}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where M_{t} is the sample number of the t th subset S_{t} and t=1, 2, \ldots,T .

The random data selection strategy has two advantages. (a) It reduces the sample number when balancing the samples for each health condition, and it could shorten the training time of base classifiers. (b) The strategy forms several different subsets for the training of base classifiers and increases the differences between the base classifiers during training, which improves the diversity of the base classifiers in the ensemble learning.

C. Random Hyper-Parameter Selection and Training of Base Classifiers

When training a base classifier, we should determine hyper-parameters for the classifier and these parameters include: the number of feature learning block l , the dimension of convolutional kernel \boldsymbol {k}^{l} , the type of activation function \sigma (\cdot) , the pooling size s , the dimension of the weight matrix in the fully connected layers, and the learning rate during training. Each parameter has its influence on classification performance of the base classifier and is difficult to be selected manually. Inspired by the random search method for hyper-parameter optimization [23], we could use the random hyper-parameter selection to determine the base classifiers and a sub-dataset to train each base classifier.

Random selection strategy covers a large amount of information and is computationally simple and efficient compared to traditional grid search. The probability of random selection obtaining the optimal hyper-parameter is \begin{equation*} P_{q} =1-\left ({{1-q \mathord {\left /{ {\vphantom {q Q}} }\right. } Q} }\right)^{\Gamma }\tag{11}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where q is the space for a single search, Q is the whole parameter space, q/Q is the probability of the optimal hyper-parameter in parameter space; \Gamma is the number of searches. It can be seen that if the space for a single search is fixed, the probability of obtaining the best hyper-parameter increases as the number of searches increases. After determining the hyper-parameters, the base classifier h_{t} is trained by \begin{equation*} \textrm {min}-\frac {1}{M}\sum \limits _{m_{t} =1}^{M_{t}} {\sum \limits _{c=1}^{C} {1\left \{{{y^{m_{t}}=c} }\right \}\log \left ({{p_{c}^{m_{t}}} }\right)}}.\tag{12}\end{equation*}
View SourceRight-click on figure for MathML and additional features.

The advantages of random hyper-parameter selection are as follows. (a) The manual hyper-parameter selection of a base classifier is avoided. In ensemble learning, the probability of obtaining the optimal hyper-parameters increases with the number of base classifiers gradually. (b) The independence between the random selected hyper-parameters contributes to the differences between the base classifiers, which enhances the diversity of the base classifiers.

D. Base Classifiers Selection and Ensemble

After each base classifier has been trained, these classifiers are combined to the ensemble classifier. In EnCNN, the ensemble of individual base classifiers based on the weighted voting strategy. We calculate the ensemble weights for each base classifier by using G-mean score. Since only a portion of the samples are under-sampled from the original training dataset to train a base classifier, the performance of the classifier can be tested with the original training dataset. G-mean is used to get the test performance of a base classifier and we use the performance score as the ensemble weight of this classifier.

G-mean [24] takes into account imbalanced distributions of the samples and has been a common metric to measure the classification performance under imbalanced data. It considers both the recall and specificity of the classification results, where recall is a function of true positives (TP) and false negatives (FN) and specificity is a function of true negatives (TN) and false positives (FP). \begin{equation*} recall=\frac {TP}{TP+FN},\quad specificity=\frac {TN}{TN+FP}\tag{13}\end{equation*}

View SourceRight-click on figure for MathML and additional features. where TP means true samples correctly identified, FN means false samples incorrectly identified as true samples, TN means false samples correctly identified and FP means true samples incorrectly identified as false samples. Then, G-mean score of a classifier is obtained by \begin{equation*} g=\sqrt {recall\times specificity}.\tag{14}\end{equation*}
View SourceRight-click on figure for MathML and additional features.

G-mean score reaches its best value at 1 and worst score at 0 and we use this score for the ensemble weight of the base classifiers. Due to random factors, however, a base classifier may be trained abnormally and may result in a low classification performance. To keep the anomalous base classifiers from affecting the classification performance of EnCNN, these classifiers are detected by classifier selection and are deleted before the ensemble [25]. Here, we use boxplot-based classifier selection method to find the anomalous base classifiers. Compared with other methods, boxplot does not need to assume that the data obey a particular distribution and can provide an objective criterion for identifying outliers of the data [26]. Thus, we apply it for the classifier selection of the base classifiers. The ensemble weights of anomalous base classifiers will be set to zero and the ensemble weights of other base classifiers will be reserved, as shown in (15).\begin{align*} {g}'_{t} =\begin{cases} g_{t}, &\textrm {if}~g_{t} \ge Q_{1} -1.5I \\ 0,& \textrm {if}~g_{t} < Q_{1} -1.5I \\ \end{cases}\tag{15}\end{align*}

View SourceRight-click on figure for MathML and additional features. where g_{t} is the G-mean score or ensemble weight of base classifier h_{t} , Q_{1} is the lower quartile and I is the distance between the upper and lower quartile. Then, the ensemble weights w_{t} of the base classifiers are normalized after the classifier selection.\begin{equation*} w_{t} ={g'_{t}} \mathord {\left /{ {\vphantom {{g'_{t}} {\sum \limits _{t=1}^{T} {g'_{t}}}}} }\right. } {\sum \limits _{t=1}^{T} {g'_{t}}}\tag{16}\end{equation*}
View SourceRight-click on figure for MathML and additional features.

Finally, the selected base classifiers and their ensemble weights are combined to form EnCNN H({\boldsymbol {x}}^{m}) . \begin{equation*} H({\boldsymbol {x}}^{m})=\sum \limits _{t=1}^{T} {w_{t} \cdot h_{t} ({\boldsymbol {x}}^{m})}\tag{17}\end{equation*}

View SourceRight-click on figure for MathML and additional features.

E. The Measurement of the Performance of EnCNN

In the testing stage of EnCNN, a balanced accuracy function is used to measure the performance of EnCNN, which is used to avoid the inflated performance estimating on imbalanced data. This function is the macro-average of recall scores per class. For the test sample \boldsymbol {x}^{m} , its true label is y^{m} . Balanced accuracy function uses the sample weight {w}^{m} to reduce the influence of imbalanced data. The sample weight is first adjusted to \begin{equation*} \hat {w}^{m}=\frac {w^{m}}{\sum \nolimits _{n=1}^{M} {1\left \{{{y^{n}=y^{m}} }\right \}w^{n}}}.\tag{18}\end{equation*}

View SourceRight-click on figure for MathML and additional features.

Given the predicted label \hat {y}^{m} by EnCNN for the sample \boldsymbol {x}^{m} , the balanced accuracy is defined as \begin{equation*} \text {Acc}\left ({{y,\hat {y},w} }\right)=\frac {}{\sum {\hat {w}^{m}} }\sum \limits _{m=1}^{M} {1\left \{{{\hat {y}^{m}=y^{m}} }\right \}} \hat {w}^{m}.\tag{19}\end{equation*}

View SourceRight-click on figure for MathML and additional features.

This balanced accuracy is used to test the classification performance of EnCNN in intelligent diagnosis of machines.

SECTION IV.

Intelligent Fault Diagnosis of Machines Using Imbalanced Data

In this section, we describe the experiment data used in this paper and use the data to verify the proposed method.

A. Data Description

The imbalanced dataset used in this paper is composed of the multivariate time-series acquired by a SpectraQuest’s machinery fault test bench [27], [28]. The test bench emulates the dynamic of motors with two shaft-supporting bearings and allows the study of multiple faults, as shown in Figure 3. Accelerometers were mounted on the overhang and underhand house, respectively. Axial, radial and tangential direction data were collected to represent the health conditions of the test bench. Thus, a sample contained the vibration signals from six channels. As displayed in Table 1, a variety of faults were imposed on the test bench, involving normal operation, horizontal misalignment and vertical misalignment. The misalignment of fault was induced into the test beach by shifting motor shaft horizontally or vertically with different degrees. In the experiment, different amount samples were collected for different faults and these samples composed the imbalanced dataset used in this paper. The total number of the samples are 5400.

TABLE 1 The Description of the Imbalanced Dataset
Table 1- 
The Description of the Imbalanced Dataset
FIGURE 3. - SpectraQuest’s machinery fault test bench.
FIGURE 3.

SpectraQuest’s machinery fault test bench.

The proposed ECNN method is used to identify the health condition of the test bench under the imbalanced dataset. 50% of the samples from each health condition are randomly selected to compose the training dataset and the remaining samples compose the testing dataset.

B. Classification Results

We use the training dataset to train the proposed EnCNN and use the testing dataset to verify classification performance of EnCNN under imbalanced data. Since the hyper-parameters of the base classifiers of ECNN are randomly selected, there is only one hyper-parameter should be preset manually, i.e. the ensemble number. This number indicates the number of base classifiers used to be integrated in the proposed EnCNN. The effects of this parameter on the classification accuracy and the computing time are analyzed, where the computation platform is a computer with Intel I5 CPU, GTX 1060 GPU and 8G RAM. The results are shown in Figure 4. It can be seen that the training accuracy and testing accuracy of EnCNN increase with the ensemble number increasing. When the ensemble number increases from 1 to 12, the training accuracy and testing accuracy increase rapidly and their standard deviations decrease rapidly. The training accuracy increases from 78.49% to 95.35% and the corresponding standard deviations decrease from 15.01% to 1.08%. The testing accuracy increases from 75.69% to 91.17% and the corresponding standard deviations decrease from 15.76% to 1.78%. When the ensemble number increases from 16 to 24, the training accuracy and testing accuracy become stable. The training accuracy ranges from 95.95% to 96.11% and the corresponding standard deviations range from 0.74% to 0.89%. The testing accuracy ranges from 92.05% to 92.59% and the corresponding standard deviations range from 1.15% to 1.17%. In addition, the computing time ranges from 8.4 to 175.7 s. Thus, the larger the ensemble number is, the more time the method spends. The ensemble number selection of EnCNN is actually the trade-off between the accuracy and training time. Traditionally, the larger the ensemble number of base classifiers is, the better the diversity of base classifiers is and the higher the accuracy of EnCNN achieves. However, if the ensemble number of base classifiers is too large, it not only increases the training time of ECNN but also increases the possibility of redundant base classifiers. Therefore, compared with the results in Figure 4, we choose 16 as ensemble number.

FIGURE 4. - Classification accuracy of EnCNN using different ensemble number.
FIGURE 4.

Classification accuracy of EnCNN using different ensemble number.

In order to fully inspect the health condition of machines, a sample contains six channel vibration signals. To verify the necessity of multiple channel data for the intelligent fault diagnosis of the machines, the proposed EnCNN are used to compare with the diagnosis methods based on single channel data. The classification results are shown in Table 2. The training accuracy using Channels 1 to 6 ranges from 51.69% to 80.24% and their standard deviations range from 3.12% to 1.57%. The test accuracy ranges from 36.03% to 63.48% and their standard deviations range from 3.72% to 6.57%. The results indicate that the single data cannot fully represent the health condition of the machines and effectively train the ECNN, resulting in the overfitting of ECNN during training. In contrast, the proposed method makes fully use of all the channel data to recognize the health conditions of machines under imbalanced data and finally obtain the training accuracy of 96.99% and the test accuracy of 91.58%. It illustrates the advantages of using the multiple channel data.

TABLE 2 The Classification Results Using Different Channel Data
Table 2- 
The Classification Results Using Different Channel Data

To illustrate the advantages of classifier selection used in EnCNN, the classification results are obtained by EnCNN using classifier selection and not using classifier selection, as shown in Figure 5. The G-mean scores of each base classifier of EnCNN is plotted in Figure 5(a). It can be seen that scores of these base classifiers range from 18.55% to 89.91%, where the scores of the 6th and 10th based classifiers are only 42.99% and 18.55%, respectively. From the boxplot in Figure 5(b), the scores of the two based classifiers are outlies and should be removed in the ensemble learning in EnCNN. Figure 5(c) shows the ensemble weights of the EnCNN without classifier selection and the testing accuracy of such network is the 92.98% in Figure 5(d). Figure 5(e) shows the ensemble weights of the proposed EnCNN using classifier selection and the ensemble weights of the 6th and 10th based classifiers are set to zero. The testing accuracy of EnCNN finally is improved to 93.34% as shown in Figure 5(f). Therefore, these results verify the effectiveness of the anomalous classifier selection used in EnCNN.

FIGURE 5. - The classification results of EnCNN: (a) G-mean scores of each base classifier, (b) boxplot of G-mean scores, (c) ensemble weights of base classifiers not using classifier selection, (d) testing accuracy not using classifier selection, (e) ensemble weights of base classifiers using classifier selection, and (d) testing accuracy using classifier selection.
FIGURE 5.

The classification results of EnCNN: (a) G-mean scores of each base classifier, (b) boxplot of G-mean scores, (c) ensemble weights of base classifiers not using classifier selection, (d) testing accuracy not using classifier selection, (e) ensemble weights of base classifiers using classifier selection, and (d) testing accuracy using classifier selection.

In EnCNN, a weighted voting strategy based on G-mean score is used for the ensemble of base classifiers. To illustrate the effectiveness of this strategy, we compare EnCNN using weighted voting strategy based on G-mean score and EnCNN using traditional weighted voting strategy based on accuracy scores. The confusion matrices of the testing accuracy are shown in Figure 6. It is seen that the accuracy of each health condition using EnCNN based traditional weighted voting strategy ranges from 76.71% to 98.2% in Figure 6(a). In Figure 6(b), the accuracy of vertical misalignment is improved by using the proposed EnCNN from 76.71%, 84.97% and 89.67% to 82.81%, 87.21% and 93.03%. These results verify that the used weighted voting strategy in the proposed EnCNN is superior to the traditional weighted voting strategy in the intelligent diagnosis of machines under imbalanced data.

FIGURE 6. - The confusion matrix: (a) EnCNN using traditional weighted voting strategy based on accuracy scores, and (b) the proposed EnCNN using weighted voting strategy based on G-mean score.
FIGURE 6.

The confusion matrix: (a) EnCNN using traditional weighted voting strategy based on accuracy scores, and (b) the proposed EnCNN using weighted voting strategy based on G-mean score.

C. Comparisons

To verify the effectiveness of the proposed method, some famous methods are used to classify the health conditions of the imbalanced data. These methods involve CNN, RUSboost method, SMOTE method and random forest. The input of CNN is the vibration signals and the input of the other three methods is the manual features used in [29]. RUSBoost method [30] first uses under-sampling strategy to balance the sample number of each health condition and then uses AdaBoost classifier to recognize these health conditions of machines. SMOTEBoost method [31] uses an oversampling strategy to create synthetic data samples and make the samples of each health condition balanced, and then applied ensemble classifiers for intelligent fault diagnosis. Random forest [32] selects a random sample with replacement of the training set to train multiple decision tree models, and combines the models as ensemble classifiers using with bagging strategy. The classification results of the above methods are shown in Table 3.

TABLE 3 Results of the Compared Methods
Table 3- 
Results of the Compared Methods

It is seen that the training accuracy of CNN, SMOTE method and random forest ranges from 99.73% to 100% and the corresponding standard deviations range from 0 to 0.24%. Although the training accuracy of these methods is above 99%, the test accuracy ranges from 81.77% to 88.96%, indicating that a severe overfitting occurs in the training stages. The training accuracy and testing accuracy of RUSboost method are 73.16% and 71.03%, respectively. Although overfitting is slight in its training stage, its accuracy is the lowest in the compared methods. In the proposed EnCNN, CNN is used as the base classifier, and under-sampling strategy and random hyper-parameter selection is applied to increase the diversity of the base classifiers. Thus, overfitting in the training stage of EnCNN is avoided to some extent, and its testing accuracy is higher than the other compared methods. These results verify the effectiveness of the proposed method.

SECTION V.

Conclusions

This paper proposes EnCNN for the intelligent fault diagnosis of machines under imbalanced data. In EnCNN, a convolutional neural network with multi-sensor signal input is used as the base classifier. The mechanical imbalance dataset is split into balanced training subsets through under-sampling strategy, and each subset is then used to train a base classifier. Then the weight coefficients of each trained base classifier are calculated by G-mean score and anomalous base classifiers are screened using classifier selection so as to form EnCNN through weighted voting strategy. Finally, EnCNN is applied for the classification of mechanical health conditions. An imbalanced dataset collected from the machinery fault test bench is used for validating the proposed EnCNN. The classification results show that EnCNN could select hyper-parameters of a deep network automatically and has a stable diagnosis accuracy under the imbalanced data. By comparing the related methods for imbalanced classification, the effectiveness of the proposed method is verified. Moreover, the proposed method needs to preset only one hyper-parameter, i.e. the number of base classifiers. Generally, the larger the number of base classifiers is, the better the diversity of base classifiers and the better the generalization ability of EnCNN are, but the more computing time EnCNN will spend.

Although the effectiveness of the proposed EnCNN is verified, it cannot deal with the situation that new fault may appear after training stage, namely unseen fault classification. The authors focus on this problem in the future work.

Select All
1.
L. Guo, Y. Lei, N. Li, T. Yan and N. Li, "Machinery health indicator construction based on convolutional neural networks considering trend burr", Neurocomputing, vol. 292, pp. 142-150, May 2018.
2.
Y. Lei, B. Yang, X. Jiang, F. Jia, N. Li and A. K. Nandi, "Applications of machine learning to machine fault diagnosis: A review and roadmap", Mech. Syst. Signal Process., vol. 138, Apr. 2020.
3.
R. Zhao, R. Yan, Z. Chen, K. Mao, P. Wang and R. X. Gao, "Deep learning and its applications to machine health monitoring", Mech. Syst. Signal Process., vol. 115, pp. 213-237, Jan. 2019.
4.
X. Li, H. Jiang, K. Zhao and R. Wang, "A deep transfer nonnegativity-constraint sparse autoencoder for rolling bearing fault diagnosis with few labeled data", IEEE Access, vol. 7, pp. 91216-91224, 2019.
5.
B. Hou and R. Yan, "Convolutional autoencoder model for finger-vein verification", IEEE Trans. Instrum. Meas., vol. 69, pp. 2067-2074, May 2020.
6.
J. Wang, S. Li, Z. An, X. Jiang, W. Qian and S. Ji, "Batch-normalized deep neural networks for achieving fast intelligent fault diagnosis of machines", Neurocomputing, vol. 329, pp. 53-65, Feb. 2019.
7.
S. Haidong, C. Junsheng, J. Hongkai, Y. Yu and W. Zhantao, "Enhanced deep gated recurrent unit and complex wavelet packet energy moment entropy for early fault prognosis of bearing", Knowl.-Based Syst., vol. 188, Jan. 2020.
8.
Z. He, H. Shao, X. Zhang, J. Cheng and Y. Yang, "Improved deep transfer auto-encoder for fault diagnosis of gearbox under variable working conditions with small training samples", IEEE Access, vol. 7, pp. 115368-115377, 2019.
9.
B. Zhao, X. Zhang, Z. Zhan and S. Pang, "Deep multi-scale convolutional transfer learning network: A novel method for intelligent fault diagnosis of rolling bearings under variable working conditions and domains", Neurocomputing, vol. 407, pp. 24-38, Sep. 2020.
10.
S. Tang, C. Shen, D. Wang, S. Li, W. Huang and Z. Zhu, "Adaptive deep feature learning network with Nesterov momentum and its application to rotating machinery fault diagnosis", Neurocomputing, vol. 305, pp. 1-14, Aug. 2018.
11.
W. Fuan, J. Hongkai, S. Haidong, D. Wenjing and W. Shuaipeng, "An adaptive deep convolutional neural network for rolling bearing fault diagnosis", Meas. Sci. Technol., vol. 28, no. 9, Sep. 2017.
12.
J. M. Johnson and T. M. Khoshgoftaar, "Survey on deep learning with class imbalance", J. Big Data, vol. 6, no. 1, pp. 27, Dec. 2019.
13.
F. Jia, Y. Lei, N. Lu and S. Xing, "Deep normalized convolutional neural network for imbalanced fault classification of machinery and its understanding via visualization", Mech. Syst. Signal Process., vol. 110, pp. 349-367, Sep. 2018.
14.
F. Zhou, S. Yang, H. Fujita, D. Chen and C. Wen, "Deep learning fault diagnosis method based on global optimization GAN for unbalanced data", Knowl.-Based Syst., vol. 187, Jan. 2020.
15.
X. Dong, H. Gao, L. Guo, K. Li and A. Duan, "Deep cost adaptive convolutional network: A classification method for imbalanced mechanical data", IEEE Access, vol. 8, pp. 71486-71496, 2020.
16.
J. Wang, S. Li, B. Han, Z. An, H. Bao and S. Ji, "Generalization of deep neural networks for imbalanced fault classification of machinery using generative adversarial networks", IEEE Access, vol. 7, pp. 111168-111180, 2019.
17.
Z.-H. Zhou, "Ensemble learning" in Encyclopedia of Biometrics, New York, NY, USA:Springer-Verlag, vol. 1, pp. 270-273, 2009.
18.
Z.-H. Zhou, Ensemble Methods: Foundations and Algorithms, Boca Raton, FL, USA:CRC Press, 2012.
19.
A. Krogh and J. Vedelsby, "Neural network ensembles cross validation and active learning", Proc. Adv. Neural Inf. Process. Syst., pp. 231-238, 1995.
20.
O. Sagi and L. Rokach, "Ensemble learning: A survey", Wiley Interdiscipl. Rev. Data Mining Knowl. Discovery, vol. 8, no. 4, pp. e1249, 2018.
21.
H. I. Fawaz, G. Forestier, J. Weber, L. Idoumghar and P.-A. Muller, "Deep learning for time series classification: A review", Data Mining Knowl. Discovery, vol. 33, no. 4, pp. 917-963, Jul. 2019.
22.
R. Barandela, R. M. Valdovinos, J. S. Sánchez and F. J. Ferri, "The imbalanced training sample problem: Under or over sampling?", Proc. Joint IAPR Int. Workshops Stat. Techn. Pattern Recognit. (SPR) Struct. Syntactic Pattern Recognit. (SSPR), pp. 806-814, 2004.
23.
J. Bergstra and Y. Bengio, "Random search for hyper-parameter optimization", J. Mach. Learn. Res., vol. 13, pp. 281-305, Feb. 2012.
24.
H. Guo, H. Liu, C. Wu, W. Zhi, Y. Xiao and W. She, "Logistic discrimination based on G-mean and F-measure for imbalanced problem", J. Intell. Fuzzy Syst., vol. 31, no. 3, pp. 1155-1166, Aug. 2016.
25.
L. Rokach, "Ensemble-based classifiers", Artif. Intell. Rev., vol. 33, no. 1, pp. 1-39, 2010.
26.
M. Hubert and E. Vandervieren, "An adjusted boxplot for skewed distributions", Comput. Statist. Data Anal., vol. 52, no. 12, pp. 5186-5201, 2008.
27.
M. A. Marins, F. M. L. Ribeiro, S. L. Netto and E. A. B. da Silva, "Improved similarity-based modeling for the classification of rotating-machine failures", J. Franklin Inst., vol. 355, no. 4, pp. 1913-1930, Mar. 2018.
28.
D. Pestana-Viana, R. Zambrano-Lopez, A. A. de Lima, T. De M. Prego, S. L. Netto and E. A. B. da Silva, "The influence of feature vector on the classification of mechanical faults using neural networks", Proc. IEEE 7th Latin Amer. Symp. Circuits Syst. (LASCAS), pp. 115-118, Feb. 2016.
29.
R. Razavi-Far, M. Farajzadeh-Zanjani and M. Saif, "An integrated class-imbalanced learning scheme for diagnosing bearing defects in induction motors", IEEE Trans. Ind. Informat., vol. 13, no. 6, pp. 2758-2769, Dec. 2017.
30.
C. Seiffert, T. M. Khoshgoftaar, J. Van Hulse and A. Napolitano, "RUSBoost: A hybrid approach to alleviating class imbalance", IEEE Trans. Syst. Man Cybern. A Syst. Humans, vol. 40, no. 1, pp. 185-197, Jan. 2010.

References

References is not available for this document.