Next Article in Journal
Pedestrian Detection with Multi-View Convolution Fusion Algorithm
Next Article in Special Issue
Molecular Weight Distribution Control for Polymerization Processes Based on the Moment-Generating Function
Previous Article in Journal
An Enhanced Full-Form Model-Free Adaptive Controller for SISO Discrete-Time Nonlinear Systems
Previous Article in Special Issue
Variance and Entropy Assignment for Continuous-Time Stochastic Nonlinear Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Sensor Vibration Signal Based Three-Stage Fault Prediction for Rotating Mechanical Equipment

1
State Key Laboratory of Nuclear Power Safety Monitoring Technology and Equipment, China Nuclear Power Engineering Co., Ltd., Shenzhen 518172, China
2
College of Electrical and Power Engineering, Taiyuan University of Technology, Taiyuan 030024, China
*
Authors to whom correspondence should be addressed.
Entropy 2022, 24(2), 164; https://doi.org/10.3390/e24020164
Submission received: 9 December 2021 / Revised: 14 January 2022 / Accepted: 18 January 2022 / Published: 21 January 2022

Abstract

:
In order to reduce maintenance costs and avoid safety accidents, it is of great significance to carry out fault prediction to reasonably arrange maintenance plans for rotating mechanical equipment. At present, the relevant research mainly focuses on fault diagnosis and remaining useful life (RUL) predictions, which cannot provide information on the specific health condition and fault types of rotating mechanical equipment in advance. In this paper, a novel three-stage fault prediction method is presented to realize the identification of the degradation period and the type of failure simultaneously. Firstly, based on the vibration signals from multiple sensors, a convolutional neural network (CNN) and long short-term memory (LSTM) network are combined to extract the spatiotemporal features of the degradation period and fault type by means of the cross-entropy loss function. Then, to predict the degradation trend and the type of failure, the attention-bidirectional (Bi)-LSTM network is used as the regression model to predict the future trend of features. Furthermore, the predicted features are given to the support vector classification (SVC) model to identify the specific degradation period and fault type, which can eventually realize a comprehensive fault prediction. Finally, the NSF I/UCR Center for Intelligent Maintenance Systems (IMS) dataset is used to verify the feasibility and efficiency of the proposed fault prediction method.

1. Introduction

In the production processes of modern industries, the performance of rotating mechanical equipment may degrade over time, even resulting in failure due to long-term operation under severe conditions such as high speed, high temperature, high pressure, and heavy loads. To ensure the safety and efficiency of the operation, health monitoring and the establishment of a maintenance strategy have become active research focus in both industry and academia [1,2,3,4]. Initially, the maintenance strategy was implemented after fault diagnosis or preventive maintenance. We know that different types of equipment faults have specific vibration frequency characteristics. Therefore, traditional signal processing methods, such as Fourier transform (FT) [5], short-time Fourier transform (STFT) [6], and wavelet transform (WT) [7], have been proposed to obtain useful features from the vibration signal to reflect the operating status of the rotating mechanical equipment. However, the above fault diagnosis methods rely heavily on expert experience. In order to solve this problem, deep learning methods, such as CNN [8], LSTM [9], and the combined CNN and LSTM [10], have been used in fault diagnosis more recently, displaying the ability of deep feature self-learning without relying on manual intervention and prior knowledge. Although these methods can achieve excellent results in fault diagnosis, they cannot provide early warnings and take recovery measures before the fault occurs. Therefore, fault prediction is gradually emerging as a preventive maintenance method.
In [11], Peng Y et al. pointed out that fault prediction involves determining the RUL or working time of the diagnostic component based on the historical condition of the component. The research on RUL prediction can be divided into two categories: model-based methods and data-driven methods [12]. Lei Y et al. used maximum likelihood estimation and a particle filter algorithm to predict the RUL of bearings [13]. The proposed method is not well applicable to abrupt degeneration trends. Model-based methods rely on prior knowledge and specific conditions, whereas the data-driven approach attempts to use deep learning to deduce the degradation of equipment based on a large amount of historical data. In [14], a spectrum-principal-energy-vector method was used to extract features first, and then a deep CNN was formulated to obtain the RUL of bearings according to the features. For the first time, Xia Mei et al. divided the monitoring data into different health stages [15]. Based on this approach, the RUL of equipment was predicted using de-noising auto-encoder-based deep neural networks (DNNs). Although the above RUL methods can estimate how long it is until a fault will occur based on historical information, they are unable to provide the exact degradation period and fault type. To solve this problem, based on gray relational analysis, Wei X U et al. used a neural network model to predict the future state of a rolling bearing [16]. In [17], Xu H et al. used two models: a regression model and a classification model, which could not only predict the stage of degradation, but also classify the type of fault that would occur. However, the traditional wavelet packet transform (WPT) method was used to extract the time-frequency domain features of the original vibration signal, which requires expertise to select the appropriate basis function. Moreover, deep learning methods rely heavily on data information. Recent studies have shown that using multi-sensor data with sensor fusion technology can improve the accuracy and robustness of fault diagnosis models [18,19].
Therefore, in this paper, CNN, LSTM, and support vector classification (SVC) are combined to establish a novel three-stage fault prediction model for rotating mechanical equipment based on vibration signals from multiple sensors. Compared with the existing fault prediction results, the main contributions of the paper are as follows:
  • More formative vibration signals, used for training the fault prediction model, are collected from multiple sensors to improve the accuracy of the prediction method;
  • Deep features of varies degradation periods and fault types can be extracted by CNN and LSTM automatically without relying on manual intervention and professional knowledge;
  • The degradation period and fault type can be predicted simultaneously in advance with high accuracy.
The rest of this paper is structured as follows. In Section 2, the proposed three-stage fault prediction framework is generally introduced. Section 3 presents the details of the combined CNN and LSTM feature extraction, the attention-bidirectional (Bi)-LSTM regression model and the SVC classification mode. In Section 4, the superiority of the proposed fault prediction method is verified by applying it to the Intelligent Maintenance Systems (IMS) dataset. Section 5 concludes this paper.

2. Problem Formulation and Main Fault Prediction Framework

To accomplish the tasks of predicting the degradation period and the type of failure, a novel three-stage fault prediction framework is presented based on the multi-sensor vibration signal, using deep learning and machine learning. The proposed architecture is illustrated in Figure 1, and can be divided into three parts: CNN-LSTM-based feature extraction, attention-Bi-LSTM-based prediction, and SVC-based classification.
As shown in Figure 1, the design of three-stage architecture is related to three objectives:
  • In the feature extraction stage, the original vibration signals collected by multiple sensors are sent to the CNN-LSTM network for the extraction of spatiotemporal features, which contain operating status information;
  • In the prediction stage, the attention-Bi-LSTM is trained to predict the trend of the features;
  • In the classification stage, based on the spatiotemporal features and their trends, the SVC model is formulated to identify the degradation period and the future fault type.
The original vibration signal from multiple sensors is collected first. In order to extract the spatiotemporal information of the obtained vibration signal, a CNN with a convolution-pooling-convolution structure and an LSTM network are combined to formulate the feature extraction model. Then, an attention-Bi-LSTM network is used to predict the feature trends, which can reflect the future health state of the rotating mechanical equipment. Finally, the predicted features are sent into the SVC for classification, which can achieve the purpose of predicting the degradation period and fault type simultaneously. The emtire process of the proposed three-stage fault prediction method is detailed in the following section.

3. Deep Learning Network-Based Three-Stage Fault Prediction

3.1. Feature Extraction Stage

Vibration signals are collected by multiple sensors is in the form of time series with noise. This is susceptible to the amplitude. Therefore, it is necessary to perform feature extraction on the time-domain vibration signals. The CNN can extract the features of the original vibration signal, reducing the amount of data and diminishing the noise. However, it only can capture the spatial features, and the temporal features are ignored. Therefore, the CNN-LSTM is used in this paper to extract the spatiotemporal features of the original vibration signal. The framework of the CNN-LSTM model is illustrated in Figure 2. First, a CNN with a convolution-pooling-convolution structure is used to extract the spatial features of the original vibration signal. Then, the deep abstract of temporal features can be obtained by adding the LSTM network after the CNN.
We can denote the originally collected vibration signal in the form of a time series from the mth sensor as V m = V t m m = 1 , 2 , , n ; t = 1 , 2 , . In order to deal with the problem of sample imbalance and obtain more detailed characteristics of the vibration signal, the data are divided into windows. At the same time, it is necessary to ensure that the window size is properly chosen. For the data in each window, 1D convolution is first used to extract the shallow features. The convolution operation formula of the τ th window is as follows:
C i 1 ( τ ) = f i = 1 L ω i 1 * V m ( τ ) + b i 1 , ( i = 1 , 2 , , L )
where C i 1 ( τ ) is named the feature map, and it denotes the ith channel output of the convolution operation; L is the total number of channels; V m ( τ ) denotes the vibration signals in the τ th window; * denotes the dot product; f ( · ) is the activation function; N represents the number of convolution kernels; ω i 1 is the ith weight parameters of the 1st layer; b i 1 represents the ith bias of the 1st layer; and ω and b are undetermined parameters to be trained.
After the 1D convolution operations, 1D maxpooling operations for the processing results C i 1 i = 1 , 2 , , L are performed, which can reduce the dimension of the feature maps and further extract key features of the vibration signal:
C i 2 = max C i , 1 1 , C i , 2 1 , , C i , d 1 , ( i = 1 , 2 , , L )
where C i , k 1 represents the kth value of the ith channel of the 1st layer, d is the number of samples participating in the maxpooling operation, and C i 2 denotes the output of the pooling layer.
Another convolution operation is performed on the processing result of the pooling operation in order to extract the deeper features of the vibration signal:
C l 3 ( τ ) = f l = 1 N ω l 3 * C 2 , l ( τ ) + b l 3
where the symbols in (3) are the same as those in (1).
Considering the time characteristics of the vibration signal, the LSTM neural network is used to extract the temporal features, processing the series data with the forgetting memory scheme while avoiding the gradient disappearance and gradient explosion problems. LSTM realizes the function of long-term and short-term state transfer through the structure of an input gate, forget gate, and output gate. After obtaining the spatial features from the CNN C l 3 l = 1 , 2 , , two LSTM networks are employed to extract the temporal features.The operation principle is as follows: the update of the hidden state h l at time l is based on the joint action of the input gate i l , forgetting gate f l , output gate o l , cell state c l , and hidden state h l 1 at time l 1 . The specific formula is as follows:
i l = σ W 1 i C l 3 + W 2 i h l 1 + b i , f l = σ W 1 f C l 3 + W 2 f h l 1 + b f , o l = σ W 1 o C l 3 + W 2 o h l 1 + b o , c l = f l c l 1 + i l tanh W 1 c C l 3 + W 2 c h l 1 + b c , h l = o l tanh c l .
where W denotes weight and b is bias. σ denotes the sigmoid activation function, and ⊙ is Hadamard product.
We can input the LSTM-processed features h l l = 1 , 2 , into the two fully connected layers to obtain the eventual features of the original vibration signal, which can be denoted as F T l l = 1 , 2 , .
F E l = f ω l · h l + b l
F T l = f ω l · F E l + b l
where ω and ω denote weights, b and b are biases, and f ( · ) is the activation function.
We can iput the obtained feature F T l l = 1 , 2 , into the Softmax layer to obtain the probability that the current feature belongs to various categories p ^ . Then, one can input the probability p ^ into the cross-entropy loss function to complete the entire back-propagation process. The cross-entropy loss function is defined as follows:
L = i = 1 N p ( i ) log p ^ ( i ) + 1 p ( i ) log 1 p ^ ( i )
where p ( i ) denotes the true probability that V m = V t m m = 1 , 2 , , n ; t = 1 , 2 , belongs to the ith category; and p ^ ( i ) denotes the corresponding probability obtained from the deep networks’ classification result. The Softmax layer and the cross-entropy loss function are used here to provide a target for the back-propagation process of the deep network. It can endow classification attributes to the extracted features.

3.2. Prediction Stage

In order to achieve the purpose of failure prediction, after obtaining the operating state characteristics of the rotating machinery, the feature trend should be predicted. Since these features have a time series correlation, attention Bi-LSTM is used as the prediction model here, as shown in Figure 3. Bi-LSTM consists of a forward LSTM and a backward LSTM, which are able to capture features of both past and upcoming time series. In addition, in order to enhance the correlation between the output results of Bi-LSTM, an attention mechanism is added, which can achieve the purpose of redistributing the weights between the output results.
The features obtained from the feature extraction stage, F T l l = 1 , 2 , are first sent into the Bi-LSTM network. The specific principle formulas are as follows:
i F l = σ W F 1 i F T F l + W F 2 i h F l 1 + b F i f F l = σ W F 1 f F T F l + W F 2 f h F l 1 + b F f o F l = σ W F 1 o F T F l + W F 2 f h F l 1 + b F o c F l = f F l c F l 1 + i F l tanh W F 1 c F T F l + W F 2 c h F l 1 + b F c h F l = o F l tanh c F l
i B l = σ W B 1 i F T B l + W B 2 i h B l + 1 + b B i f B l = σ W B 1 f F T B l + W B 2 f h B l + 1 + b B f o B l = σ W B 1 o F T B l + W B 2 f h B l + 1 + b B o c B l = f B l c B l + 1 + i B l tanh W B 1 c F T B l + W B 2 c h B l + 1 + b B c h B l = o B l tanh c B l
h l = h F l h B l
where the subscripts F and B denote forward and backward, respectively, and h l represents the concatenation of the forward output h F l and backward output h B l of Bi-LSTM.
The basic idea of adding the attention mechanism in Bi-LSTM is to calculate the correlation between the target hidden state h t a r and the output hidden states h l of Bi-LSTM, and then to output the attention vector [20]. The principle formulas are listed as follows:
score h t a r , h l = h t a r T W h l
α t s = exp score h t a r , h l l = 1 L exp score h t a r , h l
c p = s = 1 S α t s h l
a p = tanh W c c p ; h t a r
where score ( · ) is a function, α t s is the attention weight, c p denotes the context vector, a p represents the attention vector, and W and W c denote weights.
We input the a p p = 1 , 2 , into the two fully connected layers to obtain the result of the feature prediction stage, which we denote as f t p p = 1 , 2 , .
f e p = f ω p · a p + b p
f t p = f ω p · a p + b p
where ω and ω denote weights, b and b are biases, and f ( · ) is the activation function.

3.3. Classification Stage

After obtaining the future spatiotemporal features of the original vibration signal using the attention-Bi-LSTM prediction model, the degradation period and fault type should be identified by formulating a classification model. In fact, the feature types are divided into several categories by training a Softmax classifier in the feature extraction stage. However, since the prediction step is added after the feature extraction step, the classification model and the prediction model cannot perform the same backpropagation. Therefore, considering the errors of the prediction results, the more robust SVC is used as the classifier instead of the Softmax with rigorous function mapping [21]. The basic idea of the SVC model is to find a classifier that maximizes the classification interval between the hyperplane and the support vector. The principle of the SVC can be addressed as follows. For the predicted features f t p p = 1 , 2 , , we denote their failure modes as y p p = 1 , 2 , , y { 1 , 1 } . The SVC problem can be converted into the following quadratic optimization problem:
Max i = 1 n α i 1 2 i = 1 n i = 1 n α i α j y i y j K f t i p , f t j p
s . t . α i 0 , i = 1 , , n i = 1 n α i y i = 0
where α is the Lagrange multiplier. K f t i p , f t j p is the kernel function, which can map the linearly inseparable samples in the initial space to the linearly separable samples in the high-dimensional space. In this paper, we use the radial basis function (RBF) as the kernel function.
K f t i p , f t j p = e γ f i p f j p 2 , γ > 0
where γ is the width of the RBF. The output function of the category can be obtained using the following formula:
f ( x ) = sgn i = 1 n α i * y i K f t i p , f t j p + b *
where b * is the classification threshold, which is obtained by substituting support vectors.
The SVC fault diagnosis model in this paper adopts the One vs. Rest (OvR) approach to realize the multi-classification of faults. The main ideas of OvR are as follows: if N categories need to be classified, N binary classifiers should be constructed. In the training process, select one category as positive and the others as negative, and then classify them in turn. In the test process, the test samples are sequentially brought into the trained N classifiers for calculation, and then the final classification result can be calculated.

3.4. Implementing the Proposed Fault Prediction Strategy

At last, it is worth summarizing the pseudo-code of the entire fault prediction algorithm, which is shown in Algorithm 1:
Algorithm 1 Fault prediction algorithm
1:
proceduretraining process
2:
Input: original signal from sensor S 1 , ,sensor S n , and initial labels y m m = 1 , 2 , which are fault modes
3:
Output: the model parameters of trained Attention-Bi-LSTM and SVC; test dataset F T d d = 1 , 2 ,
4:
  Random initialization: The feature extraction network parameters { ω } E and { b } E ; The feature prediction network parameters { W } P and { b i } P
5:
  Split original signal from S 1 , ,sensor S n into {training set} and {test set} for feature extraction model
6:
for sensor S 1 , , sensor S n in { { training set } , { test set } } do
7:
    for k in range( t i m e s ),
t i m e s = epochs , if sensor S 1 , , sensor S n = = { training set } 1 , if sensor S 1 , , sensor S n = = { test set } do
8:
            Calculate 1 1 s t one-dimensional convolution layer for sensor S 1 as C k 1 ( S 1 ) based on ω 1 1 and b 1 1 ;…;Calculate 1 n s t one-dimensional convolution layer for sensor S n as C k 1 ( S n ) based on ω n 1 and b n 1
9:
            Calculate 2 1 n d one-dimensional maxpooling layer for C k 1 ( S 1 ) as C k 2 ( S 1 ) ;…;Calculate 2 n n d one-dimensional maxpooling layer for C k 1 ( S n ) as C k 2 ( S n )
10:
          Calculate 3 1 r d one-dimensional convolution layer for C k 2 ( S 1 ) as C k 3 ( S 1 ) based on ω 1 3 and b 1 3 ;…;Calculate 3 n r d one-dimensional convolution layer for C k 2 ( S n ) as C k 3 ( S n ) based on ω n 3 and b n 3
11:
          Calculate 4 1 t h and 5 1 t h LSTM layers for C k 3 ( S 1 ) as C k 5 ( S 1 ) based on ω 1 4 , 5 and b 1 4 , 5 ; ;Calculate 4 n t h and 5 n t h LSTM layers for C k 3 ( S n ) as C k 5 ( S n ) based on ω n 4 , 5 and b n 4 , 5
12:
          Concatenate C k 5 ( S 1 ) , , C k 5 ( S n ) as C k 6
13:
          Calculate 1 s t Dense layer for C k 6 as C k 7 and 2 n d Dense layer for C k 7 as
F T m m = 1 , 2 , = F T t r t r = 1 , 2 , , if sensor S 1 , , sensor S n = = { training set } F T t e t e = 1 , 2 , , if sensor S 1 , , sensor S n = = { test set } based on ω 7 and b 7 , then input F T m m = 1 , 2 , to softmax layer
14:
          Calculate the cross-entropy loss; Update the { ω } E and { b } E , i.e., { ω } E , { b } E ω E , b E
15:
    end for
16:
end for
17:
  Reorder F T m m = 1 , 2 , in time series
18:
  Divide F T m m = 1 , 2 , to training dataset F T train , a a = 1 , 2 , and test dataset F T test , d d = 1 , 2 ,
19:
for  e p o c h in range ( E P O C H S ) do
20:
          Calculate the Bi-LSTM model for F T train , a a = 1 , 2 , as P k 1 based on W 1 and b i 1
21:
          Calculate the attention layer for P k 1 as P k 2 based on W 2 and b i 2
22:
          Calculate the Dense layer for P k 2 as P k 3 based on W 3 and b i 3
23:
          Calculate the RMSE loss function and update the { W } P and { b i } P , i.e., { W } P , { b i } P W P , b i P
24:
end for
25:
  Use cross-validation to train the SVC model with F T m m = 1 , 2 , and y ^ m m = 1 , 2 , to obtain the parameters of the SVC { s u p p o r t · v e c t o r } C
26:
end procedure
27:
proceduretest process
28:
Input:  F T test , d d = 1 , 2 , and y d d = 1 , 2 , which are fault mode labels of F T test , d d = 1 , 2 ,
29:
Output: the labels y ^ d d = 1 , 2 , of fault prediction and their accuracy
30:
  Leading-in the parameters of trained attention-Bi-LSTM model and SVC model
31:
  Calculate attention-Bi-LSTM model prediction results for F T test , d d = 1 , 2 , as f t test , d d = 1 , 2 ,
32:
  Calculate SVC classifying results y ^ d d = 1 , 2 , for f t test , d d = 1 , 2 ,
33:
  Calculate accuracy = y ^ d = = y d y d , d = 1 , 2 , , where [ x ] denotes the number of x
34:
end procedure

4. Validating the Proposed Method

In order to verify the fault prediction method proposed in this paper, the IMS dataset was used [22]. The IMS dataset contains three data sets: dataset 1, with two acceleration sensors on each bearing, and dataset 2 and dataset 3, with one accelerometer on each bearing, respectively. In view of the fact that this experiment needs to predict the failure modes through multiple sensors, dataset 1 is used to verify the proposed method. The sampling frequency of dataset 1 is 20 kHz, and the sampling duration is 1 s. The sampling interval of the first 43 rounds of each acceleration sensor was used to collect data every 5 min, then to collect data every 10 min and generate a data file containing 20,480 sampling points. The Python programming environment was used, based on the Keras framework of version 2.3.1. All experiments were performed on Intel Xeon ES-2620 CPU.

4.1. The Description of the Dataset

At the end of the test-to-failure experiment, in dataset 1, bearing 3 displayed an inner race defect and bearing 4 displayed a roller element defect [23]. The purpose of this experiment was to classify the bearing degradation period and identify the early fault type of the bearing based on the vibration signals from multiple sensors. Therefore, bearings 1-3 and 1-4 were selected for research, as shown in Table 1. The entire life of the bearing was divided into four stages: the norm period, slight period, severe period, and failure period. In addition, there were two types of faults: the inner race defect and the roller element defect. Therefore, there were seven fault modes in the experiment, which were the norm period, the slight inner race defect, the severe inner race defect, the inner race failure, the slight roller element defect, the severe roller element defect, and the roller element failure.
The IMS dataset does not have a detailed true label of the degradation period and specific failure of the bearing. Therefore, according to the labeling method in [17,24], the threshold of each stage should be set according to actual needs.In this simulation, the root mean square (RMS) features of the 20,480 vibration signals collected per second from each sensor were first extracted. Then, expertise was involved in labeling the degradation period, which is shown in Figure 4 and Table 2. In our simulation, 20,480 samples of vibration signals were collected every 10 min, and these samples were saved in one file. There were 2156 sampling files in the whole life cycle. For example, the samples from the 2120th file to the 2151th file belonged to the severe period for bearing 1-3 H.

4.2. Feature Extraction

4.2.1. Training Process

The data need to be processed appropriately in order to extract the more detailed characteristics of the vibration signal and train the better performance of fault prediction model. Using the trial and error method, the samples in each file with 20,480 samples were divided into 80 windows.Therefore, there were 256 samples in every window.
In the feature extraction stage, the task of the CNN-LSTM network is to effectively extract the spatiotemporal features of the degradation period and fault type simultaneously. The framework of the CNN-LSTM network for feature extraction in Figure 2 was used here. The network settings are shown in Table 3. The inputs of Sensor H and Sensor V were both N × 256 × 1 , with N being the number of windows. The batch size was 512. The number of epochs was 25, and the optimizer used was Adam. The cross-entropy in (7) is employed here as the loss function.

4.2.2. Verifying the Validity of the Feature Extraction Model

In order to verify the effectiveness of the proposed multi-sensor CNN-LSTM feature extraction model, a CNN-LSTM network with a single sensor was used as the comparative model. The evaluation equation was chosen as follows:
accuracy = T P T P + F P
where T P is the number of true positives and F P is the number of false positives. The comparison results between our model and other models are shown in Table 4:
From Table 4, it can be seen that the classification accuracy based on a single-sensor CNN-LSTM feature extraction model was lower than that based on the multi-sensor model. The reason is that the large fluctuation and noise of the vibration signals from a single sensor may mislead the identification of the degradation period, such as the fluctuation of 1-4V shown in Figure 4. However, the vibration signals from multiple sensors can provide more comprehensive information, which could improve the classification accuracy.

4.3. Trend Prediction

4.3.1. Training Process

The input of the prediction model is the features obtained from the previous section. The characteristics of the normal period were not used to predict degradation trend, because the data in that period often show no trend information. In this simulation, the severe and failure periods for the inner race and the roller element were predicted, respectively. In order to illustrate the effectiveness of the proposed Attention Bi-LSTM model, only the prediction results of the failure period for roller defects are presented in the following as a representation of the model’s performance.The input of our training process was set at 1994 × w i n d o w w i d t h × 10 , and the input of the testing process was 400 × w i n d o w w i d t h × 10 . The parameter settings of the prediction model are shown in Table 5.
Sliding time window technology was used to segment the dataset. When the window moves backward, a series of sample data covering each other will be formed. In the selection of the sliding window width, we tried three, six, and nine sample points to predict the next sample point, and used the root mean square error (RMSE) as the scoring criterion for the prediction error, which is defined in (21). The results are shown in Table 6. It was finally determined that the width of the input window with the smallest prediction error was six. Therefore, six sample points were used to predict the next sample point.
R M S E = 1 n i = 1 n y ^ i y i 2
where y i is the true value and y ^ i is the predicted value.

4.3.2. Testing Results

During model training and evaluation stages, the RMSE was used as the loss function of the prediction model, and the prediction results are shown in Figure 5.
From Figure 5, it can be seen that the general feature trends can be predicted with certain errors. The reason for this is that the IMS dataset we used was designed for RUL prediction instead of degradation period prediction [25]. The final task of this paper was to classify the predicted features, which allows for an acceptable prediction error.

4.3.3. Comparison with Other Models

In order to verify the prediction effect of the proposed attention-Bi-LSTM model, LSTM, Bi-LSTM, and attention-LSTM were applied to the IMS dataset for comparison. The comparative results are shown in Table 7. Since the attention mechanism can redistribute the proportions between sequences according to the predicted target, the use of a prediction model with an attention mechanism was able to improve the accuracy. Moreover, Bi-LSTM was able to capture features of both past and upcoming time series, and the performance of Bi-LSTM was also better than that of a single LSTM. We can see that the best prediction model was attention-Bi-LSTM (1.818/1.686).

4.4. Classification

After obtaining the spatiotemporal features and the degradation trends of the bearing, the SVC was used to identify the degradation period and fault type. In this step, the modes of classification of the severe period and failure period for inner race and roller element faults are presented, which is more significant than normal mode identification. In this simulation, the data were collected every ten minutes, and our task was to realize the identification of the degradation period and fault type for the future 50 min. The classification results are shown in Figure 6.
As seen in the confusion matrix in Figure 6, the classification accuracy of the proposed SVC model combined with the attention-Bi-LSTM prediction model can reach 0.944. Furthermore, the classification accuracy of the failure mode can even reach 0.985. According to the background running data, it can be found that samples with classification errors are distributed at the connection of two adjacent periods, whereas the other samples are almost classified correctly. In summary, we can achieve short-term predictions of failure types and degradation periods through the use of our proposed fault prediction method.
Remark 1.
The time-series signals in each window have one class label. The window size is determined using trial and error with a simulation method. The final accuracy of the prediction results with different window sizes is listed in Table 8. Table 8 shows a comparison of the test accuracy, sampling time, and test time of the proposed fault prediction method with three different window sizes. From Table 8, it can be seen that under the window size of 256, the test accuracy of the proposed method was higher than that of 2048 and 4096. This is because when the window size is larger, the number of windows is fewer. This directly leads to a lack of training set data in the feature prediction stage, especially for severe and failure periods, and the key information cannot be captured during feature prediction. As a result, the final classification accuracy of the predicted features would be lower. On the other hand, although the test time with a window size of 256 was about 0.01 s more than the other two cases, the fault prediction accuracy was about 0.1 higher. Therefore, the window size was chosen to be 256 in this simulation. When the programming environment and CPU change, the fault prediction time will also change accordingly.

5. Conclusions

In this study, we divided the fault prediction task into three stages: feature extraction, feature prediction, and fault mode classification. In the first stage, the spatiotemporal features of the degradation period and fault mode are extracted through CNN-LSTM, based on vibration signals from multiple sensors. In the second stage, the features are sent to a Bi-LSTM network with an added attention mechanism to predict the feature trends. The Bi-LSTM method can take into account two-way sequences, and the attention mechanism can make the Bi-LSTM network work more efficiently by adjusting the weights. Finally, an SVC is used to classify the predicted spatiotemporal features of the deterioration period and fault type simultaneously. The IMS dataset was used to illustrate the effectiveness of the proposed fault prediction method. The simulation results show that a short-term prediction of deterioration period failure modes was achieved using the established fault prediction model. This would be helpful in arranging a maintenance plan in an industrial production setting.
The efficiency of the proposed three-stage fault prediction method is based on the premise that the training set and test set obey the same distribution, with plenty of samples. However, in real engineering scenarios, rolling bearings usually work in normal conditions under different work conditions, which leads to different distributions and few fault data. Therefore, fault prediction strategies should be investigated considering these problems in the future work.

Author Contributions

Conceptualization, H.P. and M.R.; methodology, H.P., S.W., H.L. and Y.Z.; software, S.W.; validation, H.P., S.W. and M.R.; formal analysis, M.R.; investigation, H.P.; resources, H.P. and K.G.; writing—original draft preparation, S.W. and M.R.; writing—review and editing, S.W. and M.R. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Natural Science Foundation of China (No. 61973226) and the Key Research and Development Program of Shanxi Province (No. 201903D121143).

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Baraldi, P.; Di Maio, F.; Zio, E. Unsupervised clustering for fault diagnosis in nuclear power plant components. Int. J. Comput. Intell. Syst. 2013, 6, 764–777. [Google Scholar] [CrossRef] [Green Version]
  2. Kim, K.; Bartlett, E.B. Nuclear power plant fault diagnosis using neural networks with error estimation by series association. IEEE Trans. Nucl. Sci. 1996, 43, 2373–2388. [Google Scholar]
  3. Gong, Y.; Su, X.; Qian, H.; Yang, N. Research on fault diagnosis methods for the reactor coolant system of nuclear power plant based on DS evidence theory. Ann. Nucl. Energy 2018, 112, 395–399. [Google Scholar] [CrossRef]
  4. Lu, B.; Upadhyaya, B.R. Monitoring and fault diagnosis of the steam generator system of a nuclear power plant using data-driven modeling and residual space analysis. Ann. Nucl. Energy 2005, 32, 897–912. [Google Scholar] [CrossRef]
  5. Rai, V.K.; Mohanty, A.R. Bearing fault diagnosis using FFT of intrinsic mode functions in Hilbert–Huang transform. Mech. Syst. Signal Process. 2007, 21, 2607–2615. [Google Scholar] [CrossRef]
  6. Fakhfakh, T.; Bartelmus, W.; Chaari, F.; Zimroz, R.; Haddar, M. Condition Monitoring of Machinery in Non-Stationary Operations; STFT Based Approach for Ball Bearing Fault Detection in a Varying Speed Motor; Springer: Berlin/Heidelberg, Germany, 2012; pp. 41–50. [Google Scholar]
  7. Chen, J.; Li, Z.; Pan, J.; Chen, G.; Zi, Y.; Yuan, J.; Chen, B.; He, Z. Wavelet transform based on inner product in fault diagnosis of rotating machinery: A review. Mech. Syst. Signal Process. 2016, 70, 1–35. [Google Scholar] [CrossRef]
  8. Eren, L.; Ince, T.; Kiranyaz, S. A generic intelligent bearing fault diagnosis system using compact adaptive 1D CNN classifier. J. Signal Process. Syst. 2019, 91, 179–189. [Google Scholar] [CrossRef]
  9. Zhao, H.; Sun, S.; Jin, B. Sequential fault diagnosis based on LSTM neural network. IEEE Access 2018, 6, 12929–12939. [Google Scholar] [CrossRef]
  10. Qiao, M.; Yan, S.; Tang, X.; Xu, C. Deep convolutional and LSTM recurrent neural networks for rolling bearing fault diagnosis under strong noises and variable loads. IEEE Access 2020, 8, 66257–66269. [Google Scholar] [CrossRef]
  11. Peng, Y.; Liu, D.; Peng, X. A review: Prognostics and health management. J. Electron. Meas. Instrum. 2010, 24, 1–9. [Google Scholar] [CrossRef]
  12. Liu, J.; Wang, W.; Ma, F.; Yang, Y.B.; Yang, C.S. A data-model-fusion prognostic framework for dynamic system state forecasting. Eng. Appl. Artif. Intell. 2012, 25, 814–823. [Google Scholar] [CrossRef] [Green Version]
  13. Lei, Y.; Li, N.; Gontarz, S.; Lin, J.; Radkowski, S.; Dybala, J. A model-based method for remaining useful life prediction of machinery. IEEE Trans. Reliab. 2016, 65, 1314–1326. [Google Scholar] [CrossRef]
  14. Ren, L.; Sun, Y.; Wang, H.; Zhang, L. Prediction of bearing remaining useful life with deep convolution neural network. IEEE Access 2018, 6, 13041–13049. [Google Scholar] [CrossRef]
  15. Xia, M.; Li, T.; Shu, T.; Wan, J.; De Silva, C.W.; Wang, Z. A two-stage approach for the remaining useful life prediction of bearings using deep neural networks. IEEE Trans. Ind. Inform. 2018, 15, 3703–3711. [Google Scholar] [CrossRef]
  16. Xu, W.; Liu, W.B.; Zhou, M.; Yang, J.F.; Xing, C.H. Application of Neural Network Model for Grey Relational Analysis in Bearing Fault Prediction. Bearing 2012. [Google Scholar] [CrossRef]
  17. Xu, H.; Ma, R.; Yan, L.; Ma, Z. Two-stage prediction of machinery fault trend based on deep learning for time series analysis. Digit. Signal Process. 2021, 117, 103150. [Google Scholar] [CrossRef]
  18. Park, J.W.; Sim, S.H.; Jung, H.J. Displacement estimation using multimetric data fusion. IEEE/ASME Trans. Mechatron. 2013, 18, 1675–1682. [Google Scholar] [CrossRef]
  19. Olofsson, B.; Antonsson, J.; Kortier, H.G.; Bernhardsson, B.; Robertsson, A.; Johansson, R. Sensor fusion for robotic workspace state estimation. IEEE/ASME Trans. Mechatron. 2015, 21, 2236–2248. [Google Scholar] [CrossRef]
  20. Luong, M.T.; Pham, H.; Manning, C.D. Effective approaches to attention-based neural machine translation. arXiv 2015, arXiv:1508.04025. [Google Scholar]
  21. Tang, Y. Deep learning using linear support vector machines. arXiv 2013, arXiv:1306.0239. [Google Scholar]
  22. The Bearing Dataset was Provided by the Center for Intelligent Maintenance Systems (IMS), University of Cincinnati. Available online: https://ti.arc.nasa.gov/tech/dash/groups/pcoe/prognostic-data-repository/ (accessed on 2 November 2021).
  23. Qiu, H.; Lee, J.; Lin, J.; Yu, G. Wavelet filter-based weak signature detection method and its application on rolling element bearing prognostics. J. Sound Vib. 2006, 289, 1066–1090. [Google Scholar] [CrossRef]
  24. Hong, S.; Zhou, Z.; Zio, E.; Hong, K. Condition assessment for the performance degradation of bearing based on a combinatorial feature extraction method. Digit. Signal Process. 2014, 27, 159–166. [Google Scholar] [CrossRef]
  25. Yan, M.; Xie, L.; Muhammad, I.; Yang, X.; Liu, Y. An effective method for remaining useful life estimation of bearings with elbow point detection and adaptive regression models. ISA Trans. 2021. [Google Scholar] [CrossRef]
Figure 1. The overall framework of the three-stage failure prediction method.
Figure 1. The overall framework of the three-stage failure prediction method.
Entropy 24 00164 g001
Figure 2. The model structure of the feature extraction model ( C i denotes the output of the ith layer; F T denotes the output of the last dense layer).
Figure 2. The model structure of the feature extraction model ( C i denotes the output of the ith layer; F T denotes the output of the last dense layer).
Entropy 24 00164 g002
Figure 3. Flow chart of the feature prediction stage.
Figure 3. Flow chart of the feature prediction stage.
Entropy 24 00164 g003
Figure 4. Root mean square features of each bearing.
Figure 4. Root mean square features of each bearing.
Entropy 24 00164 g004
Figure 5. Comparison of prediction results (features 1 to 10 are shown in sequence).
Figure 5. Comparison of prediction results (features 1 to 10 are shown in sequence).
Entropy 24 00164 g005
Figure 6. Confusion matrix of failure prediction results(“in” and “rl” denote inner and roller element defects, respectively; “sl”, “se” and “fa” represent slight, severe, and failure periods).
Figure 6. Confusion matrix of failure prediction results(“in” and “rl” denote inner and roller element defects, respectively; “sl”, “se” and “fa” represent slight, severe, and failure periods).
Entropy 24 00164 g006
Table 1. Introduction to dataset 1.
Table 1. Introduction to dataset 1.
DatasetBearingFault TypeSensor Number
1-1-2
dataset11-2-2
1-3inner race defect2
1-4roller element defect2
Table 2. Degradation period settings.
Table 2. Degradation period settings.
File Numbers PeriodNormSlightSevereFailure
Bearing
1-3 H1–18501851–21192120–21512152–2156
1-3 V1–18501851–21192120–21512152–2156
1-4 H1–16001601–21282129–21512152–2156
1-4 V1–16001601–21282129–21512152–2156
Table 3. Network parameter settings in the feature extraction stage.
Table 3. Network parameter settings in the feature extraction stage.
LayerTypeKernel Size/Stride/NumbersActivation FunctionPaddingBN
1-1Sensor H---N
1-2Sensor V---N
2-11D Convolution64/16/16RELUsameY
2-21D Convolution64/16/16RELUsameY
3-11D Maxpooling2/2-validN
3-21D Maxpooling2/2-validN
4-11D Convolution32/8/32RELUsameY
4-21D Convolution32/8/32RELUsameY
5-1LSTM100Tanh/Sigmoid-N
5-2LSTM100Tanh/Sigmoid-N
6-1LSTM40Tanh/Sigmoid-N
6-2LSTM40Tanh/Sigmoid-N
7Concatenate---N
8Dense1128--N
9Dense210(feature)--N
10Softmax---N
11Cross-entropy---N
Table 4. Comparison of feature extraction model with different sensor types.
Table 4. Comparison of feature extraction model with different sensor types.
Sensor TypeAccuracy
Sensor H0.892
Sensor V0.832
Sensor H and Sensor V0.928
Table 5. Predictive model network parameter settings.
Table 5. Predictive model network parameter settings.
LayerUnitsActivation Function
Input--
Bi-LSTM100Tanh/Sigmoid
Attention--
Dense75RELU
Dense1-
Table 6. The impact of different input window widths on the prediction results.
Table 6. The impact of different input window widths on the prediction results.
InputRMSE
Three inputs2.221
Six inputs1.818
Nine inputs1.906
Table 7. Comparison of prediction models.
Table 7. Comparison of prediction models.
AlgorithmRMSE
LSTM1.875
Bi-LSTM1.828
Attention-LSTM1.838
Attention-Bi-LSTM1.818
Table 8. Window size and its influence on failure prediction.
Table 8. Window size and its influence on failure prediction.
Window Size × the Number of WindowsThe Accuracy of Fault Mode PredictionSampling Time (s)Fault Prediction Time (s)
4096 × 50.80.20.244
2048 × 100.860.10.248
256 × 800.9440.01250.255
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Peng, H.; Li, H.; Zhang, Y.; Wang, S.; Gu, K.; Ren, M. Multi-Sensor Vibration Signal Based Three-Stage Fault Prediction for Rotating Mechanical Equipment. Entropy 2022, 24, 164. https://doi.org/10.3390/e24020164

AMA Style

Peng H, Li H, Zhang Y, Wang S, Gu K, Ren M. Multi-Sensor Vibration Signal Based Three-Stage Fault Prediction for Rotating Mechanical Equipment. Entropy. 2022; 24(2):164. https://doi.org/10.3390/e24020164

Chicago/Turabian Style

Peng, Huaqing, Heng Li, Yu Zhang, Siyuan Wang, Kai Gu, and Mifeng Ren. 2022. "Multi-Sensor Vibration Signal Based Three-Stage Fault Prediction for Rotating Mechanical Equipment" Entropy 24, no. 2: 164. https://doi.org/10.3390/e24020164

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop