The rolling bearing is one of key parts in a wind turbine drive train. Due to the harsh operating environment of the wind turbine, rolling bearing failures occur frequently. According to statistics, 30% of the rotating machinery failures are caused by rolling bearings[1] and about 80% of the wind turbine gearbox failures are caused by bearing failure[2]. Therefore, bearing fault diagnosis is essential for efficient and reliable operation of the wind turbines.
Traditionally, the fault diagnosis of wind turbine rolling bearing is based on the spectrum analysis of vibration signals[3]. The key technology of it is to extract the fault characteristic frequency from noisy signals. Methods of spectrum analysis include the Fourier transform[4], Hilbert transform along with some joint time-frequency analysis methods such as empirical mode decomposition (EMD)[5] and variational mode decomposition (VMD). The traditional methods only study the bearing vibration signal from a certain perspective, and the features are manually extracted depending on much prior knowledge about signal processing techniques and diagnostic expertise[6], which cannot meet the requirements of real-time and portability of fault diagnosis in the era of big data.
In order to comprehensively analyze the difference between faults, it is necessary to examine the vibration signal of the bearing from multiple perspectives, so as to grasp the overall state of the bearing. In this paper, both the spectrum analysis of the vibration signal and the time-frequency analysis are performed. Then, the features are extracted from the time domain, frequency domain (frequency spectrum and envelope spectrum) and time-frequency domain (EMD). Although multi-view features are highly complementary, these features tend to be redundant, which is not conducive to fault diagnosis. Therefore, feature selection and feature fusion before classification are necessary.
Feature selection is the process of selecting some of the most effective features of the original features to reduce the dimension of the feature set, and it is an important means to improve the performance of the learning algorithm[7]. The way to select features is to sort them based on certain evaluation criteria, and then determine the number of features selected according to needs. Commonly used evaluation criteria are feature missing values, feature variance, Pearson correlation coefficients, etc. In recent years, many studies have used the random forest (RF) model for feature selection[8-9]. Random forest derives the importance of features based on the performance of the training data, by calculating the importance of each feature in each tree, then taking a weighted average to achieve the final feature importance assessment.
Feature fusion is to reduce feature redundancy after feature selection. The methods can be divided into linear fusion and nonlinear fusion. Common linear fusion methods are principal components analysis (PCA) and linear discriminant analysis (LDA). Locally linear embedding (LLE) and autoencoder are typical nonlinear fusion methods. In the case of autoencoder (AE),compression and decompression of features are implemented in an unsupervised manner via a neural network[10]. Feature selection and fusion make it easy to train the model and visualization. The deeper meaning is to transform raw features into a new and concise representation[11].
The multi-view feature set can make full use of the information of the original signal to reflect the difference among states. The extraction of multi-view features and the reduction of redundancy by feature fusion are two essential components in this model.
In this paper, a new strategy for rolling bearing fault diagnosis using multi-view features is proposed. The scheme consists of two parts including feature extraction and feature fusion. Specifically, features are extracted from the time domain, frequency domain and time frequency domain. Then, the random forest model (RF) and autoencoder are employed for feature selection and feature fusion, respectively. Finally, the support vector machine (SVM)[12] is introduced to evaluate the features.
The main contributions are summarized as follows:
1) The technology of feature extraction from multiple perspectives is proposed in this work, which takes the advantages of signal processing in feature extraction.
2) Aiming at the redundancy of multi-view features, a novel feature fusion strategy based on the random forest and the autoencoder is developed.
3) The validity and superiority of the proposed method in practical applications are verified through experimental analyses.
The flow chart of the proposed method is shown in Fig.1. The main part of the method includes feature extraction and feature fusion.
Fig.1 The framework of the proposed method
In this section, the method of feature extraction of the vibration signal from the time domain, frequency domain and time frequency domain is introduced.
The original vibration signal contains the information about the normal vibration component, fault vibration component and environmental noise. In order to reduce dependence on prior knowledge, 43 statistical features are extracted from the timing waveform. Some features are as follows:
Mean
(1)
Maximum amplitude
(2)
Maximum peak
(3)
Standard deviation
(4)
Square root amplitude
(5)
Skewness
(6)
Kurtosis
(7)
Clearance factor
(8)
The data points in the time domain must first be transferred to the frequency domain by the Fourier transform and Hilbert transform, and then the signal can be analyzed in the frequency domain. The frequency domain features include all the basic 43 statistical features, and other features are constructed for the frequency, shown as follows:
Average frequency
(9)
Center frequency
(10)
Frequency root mean square
(11)
Frequency standard deviation
(12)
EMD decomposes signals into a finite number of signals defined as intrinsic mode functions (IMFs) from high frequency to low frequency of unequal bandwidth adaptively. The internal volatility of the signal is reflected in the extracted IMFs which include real physical information of the signal. Except the basic 43 statistical features, entropy features and fractal dimension are also extracted from selected IMFs.
1) Entropy feature. Entropy usually reflects the degree of chaos in the signal, which is a function that characterizes the uncertainty of the signal. Assuming that the IMF sequence is X={x1,x2,…,xn}, the probability of each data point occurring is pi=P(x(i)).
(13)
Then, the information entropy of the IMF signal can be expressed as
(14)
2) Fractal dimension. Box dimension is the degree of irregularity and complexity of characterizing fractal sets at different scales. The fractal dimension can describe the structural features of signals. Specific calculation methods can refer to Ref.[13].
In this step, a random forest is applied to the training data, then the decision of all trees in the forest is aggregated for all the data based on the most voting for the classification[14].The feature importance assessment using RF is to calculate contributes of each features to each tree in the random forest, then take the average and finally compare the contribution between different features[15].
Metrics of contribution include Gini index (or G for short) and out-of-bag data (OOB). In the evaluation method, the calculation of the Gini index is
(15)
where K is the number of categories; Pmk represents the proportion of category k in node m. The importance of the feature Xi at node m is the change in the Gini index before and after branching at node m:
(16)
where is the importance of the feature Xi at node m; Gl and Gr represent the Gini index of two new nodes after branching,respectively. If the node where feature Xi appears in decision tree i is in set m, then the importance of Xi in the i-th tree is
(17)
Assume that the forest has a total of n trees, then
(18)
All the obtained importance scores are normalized to obtain the score of the importance of the feature Xi:
(19)
Standardization before data input can effectively improve the convergence speed and training effect, since the autoencoder is a neural network model. All the feature vectors were standardized by removing the mean and scaling to unit variance. The standard score of a sample x is calculated as
(20)
where u is the mean of the training samples; and s is the standard deviation of the training samples.
The autoencoder consists of an encoder and a decoder. To learn more abstract features, feature fusion can be performed by operating on the trained encoder. The basic architecture of an autoencoder is shown in Fig.2.
The first layer is the input layer, the middle layer is the
Fig.2 Architecture of an autoencoder
hidden layer and the last layer is the output layer. The autoencoder compares the output of the feedforward network with the input and feeds the loss back into a recursive structure, which is powerful and easy to implement in an unsupervised manner[16]. The autoencoder network output layer has the following relationship with the input layer: The objective function is
(21)
where W is the weight of hidden layer nodes; b is the bias of hidden layer neurons; m is the number of hidden layer neurons.
The trained encoder can conduct a nonlinear transformation to reduce the dimensionality of high-dimensional features by back propagation and gradient descent.
Bearing vibration data is obtained from Case Western Reserve University[17]. The test stand is shown in Fig.3.
(a)
(b)
Fig.3 Schematic diagram of the experiment platform. (a) Test rig; (b) Schematic
The test stand consists of a 2-hp motor (left), a torque transducer/encoder (center), a dynamometer (right), and control electronics (not shown). The test bearings support the motor shaft. Single point faults were introduced to the test bearings using electro-discharge machining with fault diameters of 2.13, 4.26 and 6.39 mm. An accelerometer was attached to the motor housing at the drive end of the motor.
Vibration data was collected at a sampling frequency of 12 kHz and each sample contains 4 000 points. In order to more clearly demonstrate the gain of the proposed method for classification accuracy, noise is added to the bearing vibration signal used, with a SNR of -2. The details about the bearing datasets are shown in Tab.1.
Tab.1 Descriptions of bearing datasets
Number of samplesFault diameter/mmDefectClass label600Normal0302.13Inner race1302.13Balls fault2302.13Outer race@33302.13Outer race@64302.13Outer race@125304.26Inner race6304.26Balls fault7304.26Outer race@68
The dataset contains 300 signal samples covering nine different conditions, i.e., normal condition, ball fault, inner race fault,and outer race fault. When the model is trained, the entire sample set is randomly divided into the training set and the test set. The training set contains 200 samples, and the test set includes 100 samples.
The feature extraction for the vibration signal is from three aspects: time domain, frequency domain and time-frequency domain. The time domain waveform, frequency spectrum and envelope spectrum of the signal with the outer race fault are given in Fig.4.
The waveforms of the first five IMFs containing the main feature information are given in Fig.5. Time-frequency domain features are extracted from the first five IMF components obtained after EMD decomposition.
The time domain features are statistical features. The frequency domain features include statistical features and four frequency related features based on the frequency spectrum and envelope spectrum, and the time-frequency domain features include statistical features, three entropy features, and a box-dimensional feature. The dimensions of time domain features, frequency domain features and time-frequency features are 43, 94 and 235, respectively, as shown in Tab.2. Therefore, features with a dimension of 372 are extracted for each sample.
(a)
(b)
(c)
Fig.4 Analysis of measured vibration signal. (a) Time domain waveform; (b) Frequency spectrum; (c) Envelope spectrum
Fig.5 The first five IMFs yielded by EMD
Tab.2 Feature set
Time domainFrequency domainTime-frequency domain43 statistic features86 statistic features8 frequency features215 statistic features15 entropy feature5 fractal dimension43 features94 features235 features
A multi-view feature set can comprehensively reveal the conditions of the bearing from multiple perspectives. However, among them there are many features that do not change with fault conditions, which make fault recognition more difficult. Usually, a good classification result is not guaranteed if all the features are directly fed to the classifier without feature fusion.
When using the random forest model (RF) for feature selection, the features and labels of the training set are fed into the model for training. The trained RF model can give the value of each feature based on its performance on the decision tree branches. The specific parameters of RF are shown in Tab.3.
Tab.3 Parameters of random forest and autoencoder
Fusion modelHyperparameterOptimization resultsRandom forestn_estimators100learning_rate0.1min _leaf1LossDeviancemax_depth3min _split2AutoencoderNumber of hidden layers1Node per layer120-30-120RegularizerL1(10-4)LossBinary_crossentropyEpoch250Batch_size32
Fig.6 shows the importance of each feature from four different feature sets including time domain features (T), the frequency domain consisting of frequency spectrum (FFT) and envelope spectrum (HT), and the time frequency domain from the IMFs (EMD). The histogram of feature importance is shown in Fig.6(b).For the convenience of observation, the first picture has features removed that are less than 0.000 1 in importance.
(a)
(b)
Fig.6 The importance of the features given by the random forest. (a) The importance of each feature; (b) The histogram of feature importance
As can be seen from Fig.6(a), the features extracted from the original waveform and frequency spectrum are generally more important than those from the envelope spectrum and IMFs. To more clearly reveal the difference between states, it is necessary to take the important features and discard useless ones based on feature importance.
It is evident in Fig.6(b) that only a few features are of high importance. However, the optimization of the number of selected features still needs to be executed since a small number of important features may fail to guarantee a good reconstruction of fault characteristics. The optimization of feature selection is shown in Fig.7.
Fig.7 Relationship between the number of selected features and the accuracy of SVM classification
To find the optimal number of the selected features, the model is evaluated by setting the validation set,where the training set is further divided into a small training set and a validation set. The validation size is set to be 0.33. It can be found in Fig.7 that the classification accuracy is not the best when the 20 most important features are selected. Therefore, in order to improve the accuracy of the classifier, it is necessary to select some features that are not very important. In this experiment, the number of selected features is 120.
The random forest (RF) can eliminate the irrelevant features which fusion methods cannot remove completely and the autoencoder can further reduce the redundancy of the selected features.
The autoencoder model used in this experiment has a hidden layer and uses L1 regularization to prevent model overfitting. Specific parameters are shown in Tab.3.
As an unsupervised machine learning model, the autoencoder has two important parameters to optimize, the number of hidden layer nodes and the number of epochs. The results optimized by grid search are shown in Fig.8.
Fig.8 Grid research optimization results
As shown in Fig.8, there are three combinations that allow the SVM to obtain good classification results, marked as points A, B and C. The number of hidden layer nodes is 10, 20, and 30, and the epoch is 450, 250, and 450, respectively. To maximize the accuracy, point C is selected. Therefore, the number of hidden layer nodes and the number of epoches are determined as 30 and 450. The errors of autoencoder on the training set and the validation set are shown in Fig.9.
Fig.9 Autoencoder errors
To prove the superiority of the proposed feature extraction and fusion method, the SVM classification results of the proposed method and other methods are listed in Tab.4.
From the perspective of error generation, the error mainly comes from the outer race fault classification. The differences of these four outer race faults are merely the angle or depth of the crack. The time domain features cannot fully reveal the difference between the faults. After introducing the features of the frequency domain and the time-frequency domain, the classification error is reduced.
According to the results in Tab.4, compared with the SVM performance of the original features of the time domain and all the original features, the improvement of the classification accuracy is not obvious. The increase in the features in the frequency domain and the time-frequency domain will introduce a large number of irrelevant features, which cause the difference between faults to be obscured. Therefore, when random forest is used to remove a large number of less important features, the accuracy of the classifier has been greatly improved. The accuracy rate is increased from 89% to 97%. Further-more, when using the autoencoder to reduce the redundancy of the feature set, the classification accuracy is improved by 2% and achieves 99%.
To further reveal the superiority of feature dimension reduction using the RF and autoencoder, comparisons are made between three different feature fusion methods, including PCA, kernel PCA and locally linear embedding (LLE) with the proposed methods. The results are shown in Tab.4, where N represents the dimension of the fused features.
Tab.4 SVM classification accuracy of different feature sets
Feature setNNormalBalls faultInner-race faultOuter-race faultFault 0Fault 2Fault 7Fault 1Fault 6Fault 3Fault 4Fault 5Fault 8Average accuracy/%The original features of time domain4389.4710010078.5784.6260.0010010058.3387.02All the original features37210010010081.2510042.8610080.0052.9489.10RF12010010010010010085.7110080.0081.8296.71PCA4010010076.9261.9010010010050.0077.7887.59KPCA4017.9510010065.0010010010023.0890.0091.47LLE1010010085.5010010010010075.0088.5996.13RF+AE4010010010010010010010010090.0099.10
KPCA selects a better performing polynomial kernel function instead of the Gaussian kernel. As for the PCA and KPCA, parameter n_components is set to be 40. Limited by the algorithm, LLE integrates the original features into 10 dimensions and n_neighbors is set to be 100.
It can be observed that the features of PCA fusion are the worst in SVM classification, and they even fail to achieve the accuracy before fusion. In contrast, the KPCA shows the difference clearly and the accuracy increases by 2.3% compared to that of the original features. LLE demonstrates a great improvement in accuracy, mainly due to its good nonlinear mapping ability. Among these models, the proposed RF+AE model has the highest accuracy, which further illustrates the robustness of the methods for extracting effective information and reducing feature redundancy.
1) Multi-view features can fully grasp the fault state of the bearing. After feature selection and fusion, features from multiple views can clearly reveal the state difference between normal and fault conditions. Experiments show that the fault feature set can be constructed well when the features of the vibration signals are extracted from the time domain, frequency domain and the time-frequency domain.
2) Combined with the feature selection and fusion method of the random forest and autoencoder in this paper, the accuracy of bearing fault classification can be effectively improved. The classification accuracy reaches 99.10%, which exceeds the accuracy of the feature set from the single perspective and outperforms other feature fusion methods.
3) In future studies, more features will be added to achieve better classification results and the performance of the fused features can be enhanced by using a deeper autoencoder. In addition, the proposed method can be applied to the fault diagnosis of gearboxes and life prediction of rotating machinery.
[1] Hossain M, Abu-Siada A, Muyeen S. Methods for advanced wind turbine condition monitoring and early diagnosis: A literature review[J].Energies, 2018, 11(5): 1309. DOI:10.3390/en11051309.
[2] Yang W, Tavner P J, Crabtree C J, et al. Cost-effective condition monitoring for wind turbines[J].IEEE Transactions on Industrial Electronics, 2010, 57(1): 263-271. DOI:10.1109/tie.2009.2032202.
[3] Kay S M, Marple S L. Spectrum analysis: A modern perspective[J].Proceedings of the IEEE, 1981, 69(11): 1380-1419. DOI:10.1109/proc.1981.12184.
[4] Boashash B. Time-frequency signal analysis and processing: A comprehensive reference [M]. Oxford: Elsevier Science, 2016:8-9.
[5] Chen Y, Zhou C, Yuan J, et al. Application of empirical mode decomposition in random noise attenuation of seismic data [J]. Journal of Seismic Exploration, 2014, 23: 481-495.
[6] Jia F, Lei Y G, Lin J, et al. Deep neural networks: A promising tool for fault characteristic mining and intelligent diagnosis of rotating machinery with massive data[J]. Mechanical Systems and Signal Processing, 2016, 72/73: 303-315. DOI:10.1016/j.ymssp.2015.10.025.
[7] Bishop C. Pattern recognition and machine learning[M]. New York: Springer-Verlag, 2006:559-561.
[8] Cao W H, Xu J P, Liu Z T. Speaker-independent speech emotion recognition based on random forest feature selection algorithm[C]// Proceedings of the 36th Chinese Control Conference (CCC). Dalian, China,2017. DOI:10.23919/chicc.2017.8029112.
[9] Kursa M B, Rudnicki W R. Feature selection with the BorutaPackage[J].Journal of Statistical Software, 2010, 36(11):1-13. DOI:10.18637/jss.v036.i11.
[10] Liou C Y, Cheng W C, Liou J W, et al. Autoencoder for words[J].Neurocomputing, 2014, 139: 84-96. DOI:10.1016/j.neucom.2013.09.055.
[11] Zhao X P, Wu J X, Zhang Y H, et al. Fault diagnosis of motor in frequency domain signal by stacked de-noising auto-encoder[J].Computers, Materials & Continua, 2018, 57(2): 223-242. DOI:10.32604/cmc.2018.02490.
[12] Zhang X Y, Liang Y T, Zhou J Z, et al. A novel bearing fault diagnosis model integrated permutation entropy, ensemble empirical mode decomposition and optimized SVM[J].Measurement, 2015, 69: 164-179. DOI:10.1016/j.measurement.2015.03.017.
[13] Karperien A. Defining microglial morphology: Form, function, and fractal dimension[M]. Queensland, Australia: Charles Sturt University, 2004:99-102.
[14] Forouzannezhad P, Abbaspour A, Cabrerizo M, et al. Early diagnosis of mild cognitive impairment using random forest feature selection[C]//2018 IEEE Biomedical Circuits and Systems Conference (BioCAS). Cleveland, OH, USA: IEEE, 2018. DOI:10.1109/biocas.2018.8584773.
[15] Li B Q, Cai Y D, Feng K Y, et al. Prediction of protein cleavage site with feature selection by random forest[J].PLoS One, 2012, 7(9): e45854. DOI:10.1371/journal.pone.0045854.
[16] Chen Z Y, Li W H. Multisensor feature fusion for bearing fault diagnosis using sparse autoencoder and deep belief network[J].IEEE Transactions on Instrumentation and Measurement, 2017, 66(7): 1693-1702. DOI:10.1109/tim.2017.2669947.
[17] Zheng J D, Cheng J S, Yang Y. Generalized empirical mode decomposition and its applications to rolling element bearing fault diagnosis[J].Mechanical Systems and Signal Processing, 2013, 40(1): 136-153. DOI:10.1016/j.ymssp.2013.04.005.