Next Article in Journal
Heat-Induced Increase in LPG Pressure: Experimental and CFD Prediction Study
Previous Article in Journal
Phenotypic and Genotypic Analysis of Antimicrobial Resistance of Commensal Escherichia coli from Dairy Cows’ Feces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fault Diagnosis of Rotating Machinery Bearings Based on Improved DCNN and WOA-DELM

1
School of Mechanical Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450045, China
2
School of Management and Economics, North China University of Water Resources and Electric Power, Zhengzhou 450045, China
3
School of Computing, Engineering & Digital Technologies, Teesside University, Middlesbrough TS1 3BA, UK
*
Authors to whom correspondence should be addressed.
Processes 2023, 11(7), 1928; https://doi.org/10.3390/pr11071928
Submission received: 2 June 2023 / Revised: 21 June 2023 / Accepted: 25 June 2023 / Published: 26 June 2023

Abstract

:
A bearing is a critical component in the transmission of rotating machinery. However, due to prolonged exposure to heavy loads and high-speed environments, rolling bearings are highly susceptible to faults, Hence, it is crucial to enhance bearing fault diagnosis to ensure safe and reliable operation of rotating machinery. In order to achieve this, a rotating machinery fault diagnosis method based on a deep convolutional neural network (DCNN) and Whale Optimization Algorithm (WOA) optimized Deep Extreme Learning Machine (DELM) is proposed in this paper. DCNN is a combination of the Efficient Channel Attention Net (ECA-Net) and Bi-directional Long Short-Term Memory (BiLSTM). In this method, firstly, a DCNN classification network is constructed. The ECA-Net and BiLSTM are brought into the deep convolutional neural network to extract critical features. Next, the WOA is used to optimize the weight of the initial input layer of DELM to build the WOA-DELM classifier model. Finally, the features extracted by the Improved DCNN (IDCNN) are sent to the WOA-DELM model for bearing fault diagnosis. The diagnostic capability of the proposed IDCNN-WOA-DELM method was evaluated through multiple-condition fault diagnosis experiments using the CWRU-bearing dataset with various settings, and comparative tests against other methods were conducted as well. The results indicate that the proposed method demonstrates good diagnostic performance.

1. Introduction

With industrial modernization, the direction of rotating machinery development has been toward large-scale, intelligent, high-precision, and high-efficiency. Rotating machinery usually continuously works at high speed under heavy loads. Rolling bearings are used to convert sliding friction between the shaft and shaft seat into rolling friction, as a result, rolling bearings have become one of the most prone to failure parts in rotating machinery and equipment. According to relevant statistics, 40% of motor failures come from bearings [1], and therefore, in order to ensure the reliable and safe operations of rotating machinery, the fault diagnosis of rolling bearings is of great importance.
The signals used in fault diagnosis methods are usually vibration signals, acoustic signals, current signals, speed signals, temperature signals, etc. The bearing fault diagnosis relies on various sensors. One of the effective methods is based on the signals from vibration sensors [2,3]. Many mechanical failures, such as local defects in rotating machinery. The vibration signal is manifested as a series of pulse events [4]. The fault diagnosis process is generally divided into two stages: feature extraction and fault classification. The time–frequency analysis can be used to extract the information contained in the signal in the time and frequency domains, and commonly the Short-Time Fourier Transform (STFT) [5], Fast Fourier Transform (FFT) [6], Wavelet Transform (WT) [7], Variational Modal Decomposition (VMD) [8], and Ensemble Empirical Modal Decomposition (EEMD) [9], etc., are used in the feature extraction. Hou et al. [10] used EEMD to decompose the vibration signal to obtain the intrinsic modal components. Combining the permutation entropy eigenvectors of each modal component, the Linear Discriminant Analysis (LDA) method is used to process the entropy eigenvectors as the input of the clustering algorithm, which has the advantage of better intra-class clustering compactness, but the EEMD relies too much on expert experience in the decomposition process. He et al. [11] introduced the hybrid impact index (SII) as a new metric to evaluate the fault components in the VMD method and the optimal parameters of VMD were selected using an artificial bee colony algorithm. The models that can be used for fault classification include the support vector machine (SVM) [12], K-Nearest Neighbor (K-NN) [13], Artificial Neural Network (ANN) [14], etc.; Deng et al. [15] optimized the least squares support vector machine (LS-SVM) parameters using the Particle Swarm Optimization algorithm (PSO) to improve the classification accuracy. Lu et al. [16] proposed a case-based reconstruction algorithm to adaptively locate the nearest neighbors of each test sample, which can achieve the classification of bearing faults using both parameters and cases.
In recent years, research on deep learning has received more and more attention from scholars, and its applications in object recognition, image segmentation, speech recognition, machine health detection, and medical health diagnosis have become more widespread [17,18,19]. Traditional machine learning architectures are simple and difficult to automatically extract the information carried by features in higher-order samples, and only through feature engineering that relies on expert experience may more desirable classification results be obtained. Therefore, deep learning methods are more often used for bearing fault diagnosis of rotating machinery in production practice, and the commonly used deep learning methods are convolutional neural networks (CNN) [20], Deep Belief Network (DBN) [21], and Generative Adversarial Network (GAN) [22]. Jiang et al. [23] improved the feature learning capability through a layered learning structure for convolution and pooling layers by taking the multiscale characteristics of gearbox vibration signals into consideration. Gong et al. [24] proposed an improved convolutional neural network support vector machine, in which the raw data from multiple sensors were directly input to the CNN-Softmax model, and the extracted feature vectors were input to the support vector machine for fault classification. The results were better than those achieved using the SVM and K-nearest neighbor methods. Deng et al. [25] proposed a Multi-Swarm Intelligence Quantum Differential Evolution (MSIQDE) algorithm to optimize the DBN parameters to avoid premature convergence and improve the global search capability, and the experimental results showed that higher classification accuracy was achieved using the MSIQDE-DBN than other comparative methods. Zhang et al. [26] used CNN to extract features from data and then combined them with the Long Short-Term Memory (LSTM) neural network to process time series data. The Sparrow Search Algorithm (SSA) is used to optimize the parameters of the Long Short-Term Memory (LSTM) neural network and improve the accuracy and feasibility of fault diagnosis. Chen et al. [27] proposed a fault diagnosis method combining a convolutional neural network (CNN) and Extreme Learning Machine (ELM), in which, firstly, the original vibration signal was processed using the continuous wavelet transform, then advanced features were extracted with the CNN. The classification performance was improved by using the ELM as the classifier, the proposed method was able to detect different fault types and the classification accuracy was higher than other methods. Chen et al. [28] proposed a novel fault diagnosis method for rolling bearings based on hierarchical refined composite multiscale fluctuation-based dispersion entropy (HRCMFDE) and PSO-ELM. This method solves the problem of missing high-frequency signals in the process of coarse-grained decomposition and improves the anti-interference ability and computational efficiency. The extracted feature vectors can effectively describe the fault information. Finally, the PSO-ELM classifier is used to classify the fault characteristics. Experimental results show that this method has high recognition accuracy and good load migration effect. Zhou et al. [29] proposed a new adversarial Generative Adversarial Network (GAN) generator and per-discriminator, which was improved by extracting fault features through an Auto-Encoder (AE) instead depending on fault samples., and the generator training was enhanced from the original statistical overlap to the fault features and diagnostic result errors guided model. The experimental results proved the effectiveness of Zhou’s method. Mao et al. [30] used the spectral data obtained through the original signal processing as the input to the GAN, which generated synthetic samples with fewer fault classes according to the real samples. Such synthetic samples were used as the training set, and the experiment results demonstrated that the improved generalization ability was achieved.
Based on the investigation of the above research, a rotating machinery fault diagnosis method based on an improved deep convolutional neural network (DCNN) and Whale Optimization Algorithm (WOA) optimized Deep Extreme Learning Machine (DELM) is proposed in this paper. The proposed method enhances the ability of the DCNN networks to extract important features while leveraging the excellent generalization ability of the WOA-DELM model. It addresses the issues related to poor feature extraction and low diagnostic accuracy in traditional convolutional neural networks due to feature masking caused by background noise under varying operational conditions. The main contributions of this paper are as follows:
  • An improved DCNN (IDCNN) classification network with Bi-directional Long Short-Term Memory (BiLSTM) and Efficient Channel Attention Net (ECA-Net) is constructed. The BiLSTM added to DCNN can extract the deep features of the data based on the timing information. The ECA-Net in the DCNN is introduced to weight different features. Therefore, IDCNN can make expressive features play a greater role.
  • The initial weights of the first input layer of DELM are optimized by WOA to effectively improve the overall stability of DELM. Using WOA-DELM as a classifier the classification accuracy as well as the generalization performance are improved.
  • The IDCNN-WOA-DELM fault diagnosis method proposed in this paper is experimentally validated using multiple bearing data sets and its effectiveness and generalization capability for bearing fault diagnosis have been verified by comparing it with other network models.

2. Theoretical Derivation

2.1. DCNN

The deep convolutional neural network is a class of the feedforward neural network with convolutional computation and deep structure, which is one of the representative algorithms of deep learning. The convolutional layer and pooling layer are feature extraction layers, which can be set alternately. The structure of the classical DCNN network is shown in Figure 1.
In the convolutional layer, a certain size of convolutional kernels is used to convolve local regions of the input features. Each kernel convolution represents a feature map, multiple feature surfaces are output by the nonlinear activation function, and the same input feature surface and the same feature output surface use the same convolutional kernel, so as to achieve a weight-sharing network structure. The mathematical model of the convolutional layer is expressed as follows:
x n l = f i = 1 M x n l 1 × k n i l + b n l
In this equation, x n l is the nth feature mapping of the lth layer; f ( ) is the activation function; M is the number of input feature mappings; x n l 1 is the number of nth feature mappings of the l − 1th layer; k n i l is the trainable convolution kernel; and b n l is the bias.
The pooling layer is a downsampling layer, and the size of the input matrix will change in this layer, but not the depth of the matrix. The pooling layer can be used to reduce the number of nodes for the fully connected layer, which, in the neural network training parameters optimization, has a certain role to play. The pooling layer has no parameter, so there is no need for weight updates. The mathematical model of the pooling layer is expressed as follows:
x n l = f ( β n l d o w n ( x n l 1 + b n j ) )
In this formula down() represents the subsampling function.
The fully connected layer is a traditional feedforward neural network in which the neurons of the fully connected layer are connected to the neurons of the upper layer, and the features that have been convolved and pooled can continue to be expressed nonlinearly, and the inputs of the fully connected layer are all one-dimensional feature vectors. The mathematical model expression of the fully connected layer is as follows:
a l = σ ( W l a l 1 + b l )
a l is the output of the fully connected layer, W l is the weight factor, l is the network layer sequence number, b l is the bias, a l 1 is the unfolded one-dimensional vector, σ ( ) is the activation function, and the classification task usually uses the Softmax function.

2.2. DELM

ELM, proposed by Huang et al. [31], is a new training model for Single-hidden Layer Feedforward Networks without iterative tuning its network structure of the extreme learning machine consists of an input layer, an implicit layer, and an output layer. Since it contains only one implicit layer, its generalization ability is better than the classical neural network model, and in its training process, the learning parameters in the hidden nodes are randomly chosen, which are not required to be adjusted. The output weights are obtained through generalized inverse operation, and only the number of hidden nodes needs to be determined, which is no longer propagated backward during the training process. Compared with traditional deep learning models, the training speed of the ELM is significantly improved, and the model has better generalization ability.
The mathematical model of the ELM output with a sample set of x samples and l number of hidden layers is expressed as follows:
Y j = i = 1 l β i g ( w i x j + b i ) , j = 1,2 , , N
In the formula, l is the number of hidden nodes; w i and β i are the input weight and output weight between the nodes of the ith hidden layer and the output layer, respectively; b i is the threshold of the ith hidden neuron. g ( ) is the activation function. N is the arbitrary number of samples.
By setting w i x j to be the inner product of w i and x j , the above equation can be written as follows:
H β = T
where H is the output matrix of the hidden layer, and T is the desired output.
In order to make the error between the output and the desired output close to 0, let the network cost function Y T , approach minimal. Based on the ELM theory, the learning parameters of the hidden nodes can be generated randomly without considering the input data, and hence the above equation can be changed into a linear function, and the output weights can be determined by the least squares method. The mathematical model is as follows:
β ^ = H + T
H + is the Moore-Penrose generalized inverse matrix of H .
The basic unit in the DELM is the ELM-AE, which is a combination of the ELM and an Auto-Encoder (AE), which can be seen as connecting multiple ELMs into an instrument. The structure of DELM is shown in Figure 2. This allows a more comprehensive extraction of mapping relationships between data, exhibiting better performance for processing high-dimensional and nonlinear data. The mathematical model of DELM is as follows:
Y j = j = 1 L k = 0 Z β j k g j k p = 1 n w P i h x i + b P , j = 1,2 , , Q
where L is the number of hidden layer neurons; Z is the number of derived neurons corresponding to the hidden neuron, β j k is the weight vector between the jth hidden layer neuron and the output layer; g j k is the kth order derivative of the implicit layer neuron activation function of the jth implicit layer neuron activation function; n is the number of input layer neurons; w j i h is the weight vector between the input layer and the jth hidden layer neuron, and b j is the jth hidden layer node bias. Q is the number of training data sets.
The weight of the DELM input layer is an orthogonal random matrix randomly generated in the first ELM-AE pre-training stage. In DELM, the least square method can only adjust the weight parameter of the output layer. The weight of the hidden layer must be obtained through iteration. The input weight of each ELM-AE in DELM will affect the final DELM effect.

2.3. ECA-Net

ECA-Net is proposed by Wang et al. [32], which is a local cross-channel interaction network without dimensionality reduction. It can be used to reduce model complexity while maintaining performance. The attention mechanism is a method to optimize deep learning models by simulating the attention mechanism of the human brain, in the ECA-Net, mainly the SE-net module is improved so that the ECA-Net can adaptively select the size of one-dimensional convolution kernels. Only a few parameters need to be added into the model, but obvious performance gains have been achieved. The structure of ECA-Net is shown in Figure 3.
The ECA module is implemented by fast one-dimensional convolution of size K , where K represents the coverage of local cross-channel interactions. K is related to the channel dimension C , and the larger the channel dimension C , the stronger the long-term interaction. The mapping between K and C is shown below:
C = Φ ( K )
When the channel dimension is given, the size of K can be determined adaptively.
C = Φ ( K ) = l o g 2 C γ + b γ   o d d
where t o d d represents the most recent odd number; γ = 2 ; b = 1 .

2.4. BiLSTM

LSTM was developed by Hochreiterand Schmidhuber from Recurrent Neural Network (RNN) [33] in 1998. The structure of LSTM is shown in Figure 4. LSTM has a special gating mechanism, so it can learn long-term dependencies between two sequences and has long-term memory. The key to LSTM is the transmission of information from C t 1 to C t and the selective retention of desired features in the process. LSTM unit includes a forget gate f t , input gate i t and output gate o t . The mathematical expression of the gates is shown below.
i t = σ ( w i [ h t 1 , x t ] + b i )
f t = σ ( w f [ h t 1 , x t ] + b f )
o t = σ ( w o [ h t 1 , x t ] + b o )
where σ is the activation function, w x is the weight of the corresponding gate, h t 1 is the output of the previous LSTM unit, x t is the input at the current time, and b x is the bias of the corresponding gate.
The formula for calculating the cell state C t and hidden layer state h t of LSTM unit is as follows:
C t = f t C t 1 + i t t a n h ( w c [ h t 1 , x t + b c ] )
h t = o t t a n h ( C t )
where w c is the weight coefficient matrix of the current input cell state.
BiLSTM is composed of a forward LSTM and a reverse LSTM. BILSTM makes use of known time series and reverse position series, and deepens the feature extraction of the original sequence through forward and back propagation bi-directional operation. The final output of the BiLSTM neural network is the sum of LSTM output results propagated forward and back. The structure of BiLSTM is shown in Figure 5.
Forward calculation is performed from time 1 to time t in the forward layer to obtain and save the output h t of the forward hidden layer at each time. The backward calculation is performed at the backward layer along time t to time 1 to obtain and save the output h t of the backward hidden layer at each time. Finally, the final output is obtained by combining the output results O t of the corresponding moments of the forward layer and backward layer at each moment. The calculation formula is as follows:
h t = f ( w 1 x t + w 2 h t 1 )
h t = f ( w 3 x t + w 5 h t + 1 )
O t = g ( w 4 h t + w 6 h t )

2.5. WOA

The whale optimization algorithm (WOA) algorithm is a novel natural heuristic optimization algorithm proposed by Mirjalili et al. [34], and its main idea is to mimic the unique behaviors and predation strategy adopted by humpback whales when exploring prey, shrinking the encirclement, and spiral prey location updating.
At the exploration stage, whales randomly look for prey according to each other’s location during the search process, but the prey location is generally unknown, so whales need to update the position according to their own position, and the update equation is as follows:
D = C X * ( t ) X ( t )
X ( t + 1 ) = X * ( t ) A D
The current position of a whale is denoted by X ( t ) , while X * refers to the position of a randomly selected whale. The distance between the current individual and the randomly selected individual whales is represented by D . A is the coefficient vector, which is mathematically expressed as follows:
A = 2 a r 1 a
C = 2 r 2
where r 1 and r 2 are random vectors from 0 to 1. a is a vector that decreases linearly from 2 to 0.
a = 2 2 t T m a x
where T m a x is the maximum number of iterations.
There are two predation modes The first mode is to narrow down the search. It can be described using the following mathematical expression:
D = C X b e s t ( t ) X ( t )
X ( t + 1 ) = X b e s t ( t ) A D
At this point, all individuals move towards the position with the best fitness value, thus forming a contraction surround.
The second mode uses a spiral equation to update the whale’s position based on the prey’s location, and its mathematical expression is shown below:
X ( t + 1 ) = D b e s t e b l c o s ( 2 π l ) + X ( t )
D = C X b e s t ( t ) X ( t )
In Formulas (25) and (26), t is the current number of iterations, X is the current coordinate vector of the whale, X b e s t is the location of the prey found by the whale, b is the logarithmic spiral shape constant, l is a random number between −1 and 1, and D is the distance between the humpback whale and the prey.
The above two processes are synchronized in the actual predation process, and therefore it requires setting the probability control to determine the strategy to choose between the two methods, which can be explained with Equation (27).
X ( t + 1 ) = X b e s t A D , p < 0.5 X ( t ) + D b e s t e b l c o s ( 2 π l ) p 0.5
The real behavior of humpback whales is simulated by assigning equal probabilities of 50% to each of the two methods. The search is judged to be over when the number of iterations is maximized.

3. The Proposed Bearing Fault Diagnosis Method Based on IDCNN Feature Extraction and WOA-DELM

In rotating machinery, bearing fault diagnosis under variable conditions is often a challenging task due to the high noise environment, heavy load, and high speed. Such a task requires sufficient expertise and abundant experience. To address this issue, a fault diagnosis method based on a combination of IDCNN and WOA-optimized DELM is proposed in this paper. This method aims to improve the feature extraction capability of the convolutional neural network and to leverage the excellent classification capability and stable global dynamic search capability of the WOA-DELM. The overall approach is presented in the following roadmap in Figure 6.
As shown in Figure 6, the specific troubleshooting process is as follows:
Step 1: The bearing vibration signal is collected from the experimental platform, and the collected one-dimensional vibration signal is cut into small segments of 1 × 2048, and the training set and testing set are selected from the small segments to simulate the working conditions under variable working conditions.
Step 2: The IDCNN model is trained using the training set and the deep convolutional neural network combined with BiLSTM and ECA-Net is used to mine the deep features of the data so that the features with expressive power in the samples can play a greater role. The extracted fault features are then fed into the WOA-DELM model for training.
Step 3: The testing set samples are imported into the optimal IDCNN network, and the results of the fully connected layer are output as a new testing set into the trained WOA-DELM classifier. Then the diagnostic results and various evaluation indexes are combined to illustrate the effectiveness of the model. The traditional feedforward neural network using gradient descent iterative algorithm to adjust the weight parameters will lead to slow training speed, poor generalization ability, and more training parameters, which affect the effectiveness of the feedforward neural network. Therefore, using the WOA-DELM classifier instead of the commonly used Softmax classifier can effectively improve the accuracy, efficiency, and generalization ability of classification. The fault features extracted from the IDCNN network are input into the DELM classifier for classification. The features input to the DELM classifier is trained using the WOA to find the optimal weights and bias parameters for the DELM classifier to improve the diagnostic performance of the model.

WOA–DELM Classifier

WOA is used to optimize the original randomly generated weights of the first input layer in the DELM model. The optimization of DELM using the WOA is shown in Figure 7. The specific optimization steps are as follows:
Step 1: Set the number of ELM-AEs as 180, the activation function is Sigmoid, the input weights are w i , and the hidden layer bias is b i .
Step 2: Initialization operation of the WOA parameters. The population size is set to 80 and the number of iterations is set to 20.
Step 3: Initialize the individual whale with the randomly generated x, x from DELM as the initial position vector.
Step 4: Set the fitness function, which in this paper is set as the error rate of the training set, and calculate the individual fitness values in the initialized population to obtain the optimal individual.
Step 5: After randomly generating p values, A and p are used jointly to decide how to determine the formula for the location update. When A 1 , Formula (19) is chosen. When A < 1 , Formula (27) is chosen in combination with the probability p .
Step 6: Recalculate the fitness values and find a better solution.
Step 7: Examine whether the WOA algorithm meets the termination condition, and outputs the optimal result if the termination condition is met, otherwise repeat the above steps according to the set number of iterations.
Step 8: Enter the search parameters in the WOA-DELM model to start fault diagnosis.

4. Experiments and Analysis

This paper uses a bearing experimental dataset to validate the proposed method. The experimental data are a publicly available dataset from Case Western Reserve University.
The computer hardware environment was configured with Windows 11 operating system, CPU [email protected] GHz, and GPU Nvidia GeForce RTX 3060. The program deep learning framework was built and run in Tensorflow using Python 3.7 and the WOA-DELM in Matlab R2018b.

4.1. Data Description and Processing

The CWRU (Case Western Reserve University) dataset, which is a commonly used dataset for bearing fault diagnosis, consists of a motor, torque sensor, and dynamometer in its experimental platform. The experimental platform of CWRU bearing is shown in Figure 8. It encompasses four different types of bearing failures, including inner ring damage, outer ring damage, ball failure, and normal operation, with failure diameters of 0.007 inches, 0.014 inches, and 0.028 inches, respectively. Furthermore, outer ring damage is placed in three positions: 3 o’clock, 6 o’clock, and 12 o’clock. The drive end (DE) features two sampling frequencies of 12 KHz and 48 KHz, while the fan end (FE) has only a 12 KHz sampling frequency. Each fault type contains different bearing operating states, comprising four different loads of 0, 1, 2, and 3 hp, as well as four different speeds of 1797, 1772, 1750, and 1730 rpm. The CWRU (Case Western Reserve University) dataset, which is a commonly used dataset for bearing fault diagnosis, consists of a motor, torque sensor, and dynamometer in its experimental platform. It encompasses four different types of bearing failures, including inner ring damage, outer ring damage, ball failure, and normal operation, with failure diameters of 0.007 inches, 0.014 inches, and 0.028 inches, respectively. Furthermore, outer ring damage is placed in three positions: 3 o’clock, 6 o’clock, and 12 o’clock. The drive end (DE) features two sampling frequencies of 12 KHz and 48 KHz, while the fan end (FE) has only a 12 KHz sampling frequency. Each fault type contains different bearing operating states, comprising four different loads of 0, 1, 2, and 3 hp, as well as four different speeds of 1797, 1772, 1750, and 1730 rpm.
The data used in experiments 1 and 2 are sampled at 12 KHz frequency at the DE end, with speeds of 1797 rpm, 1772 rpm, 1750 rpm, and 1730 rpm. Different speeds correspond to 0 hp, 1 hp, 2 hp, and 3 hp under four kinds of load in different fault positions, including inner ring, ball, and outer ring fault states with a fault size of 0.014, as well as the normal state. The above data are divided into data sets A, B, C, and D according to the speed and load, with each data set including 4 different working conditions. Each working condition has 200 samples. The original vibration signal of each health state contains 121,048 points. In order to avoid overfitting due to the small amount of data, the data are enhanced by overlapping sampling. Each sample contains 2048 points, and each health condition yields 200 samples, totaling 800 samples for each data set. The samples of each data set are provided in Table 1.
The data used in Experiment 3 is also sampled at 12 KHz frequency at the DE end. Experiment 3 was carried out under four conditions which are 1797 rpm/0 hp, 1772 rpm/1 hp, 1750 rpm/2 hp, and 1730 rpm/3 hp. In Experiment 3, three fault types of inner race fault, outer race fault, and ball fault were selected. Each fault type included 0.07 inches, 0.014 inches, and 0.021 inches fault sizes, respectively. Therefore, the data set used in Experiment 3 contains nine different fault conditions and one normal condition. After overlapping sampling, each fault type includes 300 samples The data set used in Experiment 3 is shown in Table 2.
This study comprises three experiments, denoted Experiment 1, Experiment 2, and Experiment 3, respectively. Experiment 1 involves the use of data from two distinct working conditions as the training set, and data from a third, distinct working condition as the testing set. The sample numbers of the training set and testing set are 1600 and 686, respectively. The training set and the testing set used in Experiment 2 are, respectively, from two different working conditions. The sample numbers of the training set and testing set are 800 and 343, respectively. Since the situation of 0 load rarely occurs during the actual operation of bearings, Experiment 2 does not consider the situation of 0 hp.

4.2. Model Parameter Setting

The structure of the IDCNN feature extraction is shown in Table 3. There are 4 convolutional layers and 4 maximum pooling layers in the model. The sizes of the four convolutional filters are 16, 32, 64, and 32. The sizes of the convolutional layers are 64, 3, 3, and 3. The step sizes are 16, 1, 1, and 1, respectively. The four maximum pooling layers are behind the four convolutional layers. The kernel size and step size are both 2. The Bilstm is set after the fourth maximum pooling layer to extract the bi-directional deep features of the positive and negative extraction time series data. The ECA-Net is set after the BiLSTM layer, and this feature extraction layer multiplies the features after the convolutional pooling layer with its feature weight matrix after the ECA attention mechanism completes to achieve feature weighting. Except for the activation function of the ECA module, which uses Sigmoid, the rest of the activation functions use the ReLU function. The parameters of the WOA-DELM classifier are set as follows: the sum of the error rates of the training and testing sets is used as the fitness function during the training iterations, and the fixed parameters are calculated at the end of the optimization in the final accuracy. The hidden layer of ELM-AE is 180, while the number of populations is 80, and the number of iterations is 20; moreover, the activation function is chosen as Sigmoid.

4.3. Analysis of Experiment 1

To evaluate the classification effect, the accuracy and F1 Score are usually used as criteria, which are calculated based on multiple experiment results so that the probability of wrong conclusions can be reduced when dealing with data containing random errors.
The following basic concepts are included in the evaluation metrics: TP (True Positives) indicates that positive cases are classified as positive cases; FP (False Positives) means that negative cases are classified as positive cases; FN (false Negatives) depicts that positive cases are classified as negative cases; TN (True Negatives) represents that negative cases are classified as negative cases.
Accuracy is the ratio of the number of correctly classified samples to the total number of samples. The formula is as follows:
A C C = T P + T N T P + F P + T N + F N
F1 Score is the summed average of precision and recall. The formula is as follows:
F 1 = 2 T P T P + F P × T P T P + F N T P T P + F P + T P T P + F N
In Experiment 1, each data set in Table 1 is used. Both IDCNN and 1D-DCNN use Softmax as the classifier The network structure of the above two methods is the same as the network structure selected in this paper. Ten experiments are used during the experiments to reduce random errors.
In Experiment 1, the data from two conditions were used as the training set, and the data from the other conditions as the testing set. The average accuracy of ten experiments is shown in Figure 9. It can be seen from the experimental results that in 12 groups of experiments, the lowest average accuracy reached 96.65%, and the highest accuracy reached 99.85%. The proposed method shows high accuracy and stability.
Four of the twelve groups were selected. They are numbered Tasks 1–4. The other two methods were used to further verify the effectiveness of the proposed method. The sample situations of Tasks 1–4 are shown in Table 4.
One experiment was randomly selected from each of the four tasks, and its accuracy and F1 Score are shown in Figure 10. In the four tasks, all the indicators were higher than 97%, among which the accuracy and F1 Score of Task 4 reached 100%. The result of Task 1 was relatively poor, but the accuracy and F1 Score still reached 97.96%. The average accuracy and average F1 Score of the ten experiments are shown in Table 5. The average accuracy and average F1 Score of the proposed method in four tasks are significantly higher than those of the other two methods, which proves the effectiveness of the proposed method in fault classification and feature extraction. The training time and testing time of the three methods are shown in Table 6. The total time of the proposed method is longer than that of the other two methods, but it is within the acceptable range. This proves that the modified method can improve diagnostic performance even with a small increase in training and testing time.
The convergence curves are shown in Figure 11. The WOA-DELM classifier achieves the optimal parameter values within 20 iterations for the task types in Table 5. The proposed method has been compared with the 1D-DCNN and IDCNN methods to show its advantages in terms of classification efficiency for rotating machine bearing faults, which reflects the excellent parameter-finding ability of the WOA-DELM classifier.
In order to demonstrate the better feature extraction ability of the improved convolutional neural network, the testing set samples in the input model for the four tasks and the output samples of the fully connected layer are visualized and analyzed using the t-SNE method. Figure 12 shows the data distribution of the input testing set samples of one of the four task types and the distribution of the testing set samples of the output of the fully connected layer. It can be seen that In the four tasks, there is still some confusion about individual features but the results obtained using the proposed method in Task 2 show good diagnostic and generalization capabilities.

4.4. Analysis of Experiment 2

In Experiment 2, each data set in Table 1 has been used. A choice of 800 samples were selected from each of the four data sets as the training set, and 343 samples are randomly selected from one of the remaining data sets as the testing set. The rest of the experiment conditions, parameter settings, and evaluation criteria were the same as those used in Experiment 1. For the same reason to avoid random results, 10 experiments were conducted.
In Experiment 2, the data of one working condition are selected as the training set, and one of the remaining data sets is selected as the testing set. The average accuracy of 10 experiments is shown in Figure 13. It can be seen from the experimental results that in the 6 groups of experiments, the lowest average accuracy reached 97.73%, and the highest accuracy reached 99.42%. Under the condition that the training set is reduced, the average accuracy of the diagnosis of the same fault type still has good and stable performance.
Three of the six groups were selected. They are numbered Tasks 5–7. The other two methods are still used to compare with the proposed method. The sample situations of Tasks 5–7 are shown in Table 7.
One experiment was randomly selected from each of the four tasks, and its accuracy and F1 Score are shown in Figure 14. The accuracy and F1 Score of the proposed method in the three tasks are all over 97.37%, which is better than the other two methods. In Task 5, accuracy and F1 Score reach 98.83%, which proves the excellent fault diagnosis capability of the proposed method under variable working conditions. The average accuracy and average F1 Score of the ten experiments are shown in Table 8. The average evaluation index obtained from ten experiments in different tasks can again prove the generalization and stability of the proposed method. The training time and testing time of the three methods are shown in Table 9. The average time spent on different tasks decreases as the number of training and testing sets decreases. Although the time used is higher than the other two methods, it is still within the acceptable range.
Figure 15 shows the confusion matrix of the method proposed in this paper for one of the experiments for each of the three tasks, particularly included are the results of the classification for all four fault states. It can be seen from this figure that the accuracy reached 98.83% in Task 5, and in Task 6, three samples of fault type 3 were incorrectly diagnosed as fault type 1, whilst eight samples of fault type 3 were incorrectly diagnosed as fault type 2. The accuracy for label 1 is 91.2% and that for label 3 is 98.8%. From the signal frequency analysis, it is known that when the outer and inner ring faults occur, the low-frequency component almost disappears, and for the rolling body fault, the vibration signal is dominated by low-frequency components, that is the reason that the samples of IF and OF in Task 6 cause confusion. In Task 7, except for the accuracy of all types but label 1 reaches 100%, the accuracy for label 1 is 92.9%.

4.5. Analysis of Experiment 3

In Experiment 3, the data in Table 1 were used. A total of 3000 training set samples and 1280 testing set samples. Other experimental conditions, parameter settings and evaluation criteria are the same as in Experiment 1. By the same token, we conducted 10 experiments to avoid random results. When the working condition is 1772 rpm/1 hp, the fault status identification result is shown in Figure 16. Furthermore, the average recognition accuracy reached 100% under this condition.
In order to verify the effectiveness of the proposed method, 1D-DCNN, IDCNN, and the proposed method in this paper were, respectively, used for experiments at 1730 rpm/3 hp. Figure 17 shows the confusion matrix of the results of any of the ten experiments. It can be seen from the experimental results that the accuracy of the proposed method reaches 100% under non-variable working conditions. In addition, compared with the IDCNN method, the accuracy of the proposed method is also improved, indicating that the construction of the WOA-DELM classifier plays a certain role in improving the accuracy of fault types.
In order to verify the generalization of the method proposed in this paper. The proposed method was used to carry out experiments under four conditions, and the average accuracy and F1 Score of ten experiments were obtained. The experimental results are shown in Table 10. According to the results, it can be seen that the average accuracy and F1 Score of the proposed method reach 100% in the three working conditions. It shows that the proposed method has a high accuracy of fault identification.
Table 11 shows the accuracy and average accuracy of the three methods in ten experiments under different working conditions. According to the results, it can be seen that the proposed method has good performance in four different working conditions.
Eight commonly used fault diagnosis methods were used for Experiment 2, in which the training set and testing set were randomly selected from the corresponding data set and compared with the proposed method in this paper, and 10 experiments were conducted to find the average accuracy in order to reduce random error. The results are shown in Table 12. Apparently, under the variable working conditions, the accuracy and F1 Score of the six methods are lower than that of the proposed method.

5. Conclusions

A bearing fault diagnosis method based on the IDCNN-WOA-DELM is proposed and applied to rotating machinery fault diagnosis under variable load and varying speed operating conditions. The main findings from this research are summarized as follows:
  • By adding BiLSTM to DCNN, the deep features of timing signals can be extracted bi-directional. By using the adaptive one-dimensional convolutional kernel size determination in the ECA-Net, A certain level of performance improvement can be achieved with a small number of added parameters, and by weighting the effective features, the effective features can play even greater roles.
  • The application of the IDCNN in bearing fault features extraction under varying working conditions can reduce the dependence on expert experience and enhances the stability and integrity of the algorithm. This is demonstrated by visualizing the test samples and the fully connected layer output of the IDCNN model using t-SNE.
  • The initial weights of the DELM exhibit randomness, whereas the WOA algorithm possesses excellent global search capability. By leveraging the contraction surround mechanism and spiral advance mechanism in the WOA algorithm, the randomly generated initial weights in the DELM can be optimized, and the WOA-DELM classifier is constructed. The experiment results indicate that the WOA-DELM classifier can significantly enhance the classification performance of bearing faults under varying working conditions, demonstrating the outstanding performances of the WOA algorithm and the improved WOA-DELM classifier.
  • Various experiments were conducted using the CWRU bearing dataset to simulate bearing fault diagnosis under varying working conditions. The training and testing sets were selected randomly across multiple experiments. In Experiment 1 and Experiment 2, a total of 18 different experiments were carried out. In the 12 groups of experiments in Experiment 1, the average accuracy of the testing set reached 98.65%. In the six experiments in Experiment 2, the average accuracy of the testing set was more than 97.69%. Among the 4 groups of experiments in Experiment 3, the average accuracy of the testing set in 3 groups reached 100%. The averaged experiment results prove that the proposed method can provide better fault classification performance and generalization ability. Furthermore, the comparison with the SVM, IDCNN, MLP, DNN, DT, DCNN-BiLSTM, and other methods confirms the effectiveness of the proposed method.
In practical production, bearing faults may conform to complex patterns, leading to more intricate vibration signals. Therefore, addressing fault analysis under compound fault conditions in the following investigation will be a critical step forward. At the same time, in order to deploy deep learning models in mobile and embedded devices, the work related to lightweight models is also of great significance.

Author Contributions

Conceptualization, C.W. and D.P.; methodology, D.P. and S.J.; software, D.P.; validation, L.W. and D.P.; formal analysis, D.P.; investigation, J.S. and D.P.; resources, S.J., D.P. and L.W.; data curation, D.P.; writing—original draft preparation, D.P.; writing—review and editing, D.P.; visualization, D.P.; supervision, L.W. and J.Z.; project administration, L.W. and J.Z.; funding acquisition, L.W. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Foreign Expert Project of Ministry of Science and Technology of the People’s Republic of China (G2022026016L), “ZHONGYUAN Talent Program” (ZYYCYU202012112), Henan International Joint Laboratory of Thermo-Fluid Electro Chemical System for New Energy Vehicle (Yuke2020-23), Zhengzhou Measurement and Control Technology and Instrument Key Laboratory (121PYFZX181), and The Fund of Innovative Education Program for Graduate Students at North China University of Water Resources and Electric Power (NCWUYC-2023065).

Data Availability Statement

The data of CWRU can be obtained at: https://engineering.case.edu/bearingdatacenter/download-data-file (accessed on 1 November 2022).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gai, J.B.; Shen, J.X.; Hu, Y.F.; Wang, H. An integrated method based on hybrid grey wolf optimizer improved variational mode decomposition and deep neural network for fault diagnosis of rolling bearing. Measurement 2020, 162, 107901. [Google Scholar] [CrossRef]
  2. Nishat Toma, R.; Kim, J.-M. Bearing Fault Classification of Induction Motors Using Discrete Wavelet Transform and Ensemble Machine Learning Algorithms. Appl. Sci. 2020, 10, 5251. [Google Scholar] [CrossRef]
  3. Tang, S.N.; Yuan, S.Q.; Zhu, Y. Deep Learning-Based Intelligent Fault Diagnosis Methods Toward Rotating Machinery. IEEE Access 2020, 8, 9335–9346. [Google Scholar] [CrossRef]
  4. Wang, Y.X.; Xiang, J.W.; Markert, R.; Liang, M. Spectral kurtosis for fault detection, diagnosis and prognostics of rotating machines: A review with applications. Mech. Syst. Signal Process. 2016, 66-67, 679–698. [Google Scholar] [CrossRef]
  5. Mustafa, D.; Yicheng, Z.; Minjie, G.; Jonas, H.; Jürgen, F. Motor current based misalignment diagnosis on linear axes with short-time Fourier transform (STFT). Procedia CIRP 2022, 106, 239–243. [Google Scholar] [CrossRef]
  6. Gowid, S.; Dixon, R.; Ghani, S.; Shokry, A. Robustness analysis of the FFT-based segmentation, feature selection and machine fault identification algorithm. Insight 2019, 61, 271–278. [Google Scholar] [CrossRef]
  7. Sifuzzaman, M.; Islam, M.R.; Ali, M. Application of wavelet transform and its advantages compared to Fourier transform. Int. J. Mag. Eng. Technol. Manag. Res. 2009, 3, 1078–1083. [Google Scholar]
  8. Miao, Y.; Zhao, M.; Lin, J. Identification of mechanical compound-fault based on the improved parameter-adaptive variational mode decomposition. ISA Trans. 2019, 84, 82–95. [Google Scholar] [CrossRef]
  9. Amarouayache, I.I.E.; Saadi, M.N.; Guersi, N.; Boutasseta, N. Bearing fault diagnostics using EEMD processing and convolutional neural network methods. Int. J. Adv. Manuf. Technol. 2020, 107, 4077–4095. [Google Scholar] [CrossRef]
  10. Hou, J.; Wu, Y.; Gong, H.; Ahmad, A.S.; Liu, L. A Novel Intelligent Method for Bearing Fault Diagnosis Based on EEMD Permutation Entropy and GG Clustering. Appl. Sci. 2020, 10, 386. [Google Scholar] [CrossRef] [Green Version]
  11. He, X.Z.; Zhou, X.Q.; Yu, W.N.; Hou, Y.X.; Mechefske, C.K. Adaptive variational mode decomposition and its application to multi-fault detection using mechanical vibration signals. ISA Trans. 2021, 111, 360–375. [Google Scholar] [CrossRef] [PubMed]
  12. Wang, Z.Y.; Yao, L.G.; Cai, Y.W. Rolling bearing fault diagnosis using generalized refined composite multiscale sample entropy and optimized support vector machine. Measurement 2020, 156, 107574. [Google Scholar] [CrossRef]
  13. Zhang, S.C.; Li, X.L.; Zong, M.; Zhu, X.F.; Wang, R.L. Efficient kNN Classification With Different Numbers of Nearest Neighbors. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 1774–1785. [Google Scholar] [CrossRef]
  14. Kanai, R.A.; Desavale, R.G.; Chavan, S.P. Experimental-Based Fault Diagnosis of Rolling Bearings Using Artificial Neural Network. J. Tribol.-Trans. ASME 2016, 138, 031103. [Google Scholar] [CrossRef]
  15. Deng, W.; Yao, R.; Zhao, H.M.; Yang, X.H.; Li, G.Y. A novel intelligent diagnosis method using optimal LS-SVM with improved PSO algorithm. Soft Comput. 2019, 23, 2445–2462. [Google Scholar] [CrossRef]
  16. Lu, J.; Qian, W.; Li, S.; Cui, R. Enhanced K-Nearest Neighbor for Intelligent Fault Diagnosis of Rotating Machinery. Appl. Sci. 2021, 11, 919. [Google Scholar] [CrossRef]
  17. Hannun, A.Y.; Rajpurkar, P.; Haghpanahi, M.; Tison, G.H.; Bourn, C.; Turakhia, M.P.; Ng, A.Y. Cardiologist-level arrhythmia detection and classification in ambulatory electrocardiograms using a deep neural network. Nat. Med. 2019, 25, 65–69. [Google Scholar] [CrossRef]
  18. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.; van Ginneken, B.; Sanchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  19. Zhao, R.; Yan, R.Q.; Chen, Z.H.; Mao, K.Z.; Wang, P.; Gao, R.X. Deep learning and its applications to machine health monitoring. Mech. Syst. Signal Process. 2019, 115, 213–237. [Google Scholar] [CrossRef]
  20. Kiranyaz, S.; Avci, O.; Abdeljaber, O.; Ince, T.; Gabbouj, M.; Inman, D.J. 1D convolutional neural networks and applications: A survey. Mech. Syst. Signal Process. 2021, 151, 107398. [Google Scholar] [CrossRef]
  21. Zhang, C.; Tan, K.C.; Li, H.Z.; Hong, G.S. A Cost-Sensitive Deep Belief Network for Imbalanced Classification. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 109–122. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  22. Shao, S.Y.; Wang, P.; Yan, R.Q. Generative adversarial networks for data augmentation in machine fault diagnosis. Comput. Ind. 2019, 106, 85–93. [Google Scholar] [CrossRef]
  23. Jiang, G.Q.; He, H.B.; Yan, J.; Xie, P. Multiscale Convolutional Neural Networks for Fault Diagnosis of Wind Turbine Gearbox. IEEE Trans. Ind. Electron. 2019, 66, 3196–3207. [Google Scholar] [CrossRef]
  24. Gong, W.; Chen, H.; Zhang, Z.; Zhang, M.; Wang, R.; Guan, C.; Wang, Q. A Novel Deep Learning Method for Intelligent Fault Diagnosis of Rotating Machinery Based on Improved CNN-SVM and Multichannel Data Fusion. Sensors 2019, 19, 1693. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  25. Deng, W.; Liu, H.L.; Xu, J.J.; Zhao, H.M.; Song, Y.J. An Improved Quantum-Inspired Differential Evolution Algorithm for Deep Belief Network. IEEE Trans. Instrum. Meas. 2020, 69, 7319–7327. [Google Scholar] [CrossRef]
  26. Zhang, C.; Chen, P.; Jiang, F.; Xie, J.; Yu, T. Fault Diagnosis of Nuclear Power Plant Based on Sparrow Search Algorithm Optimized CNN-LSTM Neural Network. Energies 2023, 16, 2934. [Google Scholar] [CrossRef]
  27. Chen, Z.Y.; Gryllias, K.; Li, W.H. Mechanical fault diagnosis using Convolutional Neural Networks and Extreme Learning Machine. Mech. Syst. Signal Process. 2019, 133, 106272. [Google Scholar] [CrossRef]
  28. Chen, Y.; Yuan, Z.; Chen, J.; Sun, K. A Novel Fault Diagnosis Method for Rolling Bearing Based on Hierarchical Refined Composite Multiscale Fluctuation-Based Dispersion Entropy and PSO-ELM. Entropy 2022, 24, 1517. [Google Scholar] [CrossRef]
  29. Zhou, F.N.; Yang, S.; Fujita, H.; Chen, D.M.; Wen, C.L. Deep learning fault diagnosis method based on global optimization GAN for unbalanced data. Knowl.-Based Syst. 2020, 187, 104837. [Google Scholar] [CrossRef]
  30. Mao, W.T.; Liu, Y.M.; Ding, L.; Li, Y. Imbalanced Fault Diagnosis of Rolling Bearing Based on Generative Adversarial Network: A Comparative Study. IEEE Access 2019, 7, 9515–9530. [Google Scholar] [CrossRef]
  31. Huang, G.; Huang, G.-B.; Song, S.; You, K. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48. [Google Scholar] [CrossRef] [PubMed]
  32. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11534–11542. [Google Scholar]
  33. Greff, K.; Srivastava, R.K.; Koutník, J.; Steunebrink, B.R.; Schmidhuber, J. LSTM: A Search Space Odyssey. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2222–2232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Mirjalili, S.; Lewis, A. The Whale Optimization Algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
Figure 1. One dimensional-DCNN structure.
Figure 1. One dimensional-DCNN structure.
Processes 11 01928 g001
Figure 2. DELM structure.
Figure 2. DELM structure.
Processes 11 01928 g002
Figure 3. Structure of ECA-Net.
Figure 3. Structure of ECA-Net.
Processes 11 01928 g003
Figure 4. Structure of LSTM.
Figure 4. Structure of LSTM.
Processes 11 01928 g004
Figure 5. Structure of BiLSTM.
Figure 5. Structure of BiLSTM.
Processes 11 01928 g005
Figure 6. Overall technical roadmap for the proposed methodology.
Figure 6. Overall technical roadmap for the proposed methodology.
Processes 11 01928 g006
Figure 7. Flow chart of WOA-DELM.
Figure 7. Flow chart of WOA-DELM.
Processes 11 01928 g007
Figure 8. The experimental platform of CWRU bearing.
Figure 8. The experimental platform of CWRU bearing.
Processes 11 01928 g008
Figure 9. The accuracy of fault diagnosis when the data of two working conditions is used as the training set and the data of another different working condition is used as the testing set.
Figure 9. The accuracy of fault diagnosis when the data of two working conditions is used as the training set and the data of another different working condition is used as the testing set.
Processes 11 01928 g009
Figure 10. Results of trials; (a) results for Task 1; (b) results for Task 2; (c) results for Task 3; (d) results for Task 4.
Figure 10. Results of trials; (a) results for Task 1; (b) results for Task 2; (c) results for Task 3; (d) results for Task 4.
Processes 11 01928 g010
Figure 11. Convergence curves for the four tasks: (a) convergence curve for Task 1; (b) convergence curve for Task 2; (c) convergence curve for Task 3; (d) convergence curve for Task 4.
Figure 11. Convergence curves for the four tasks: (a) convergence curve for Task 1; (b) convergence curve for Task 2; (c) convergence curve for Task 3; (d) convergence curve for Task 4.
Processes 11 01928 g011
Figure 12. Feature visualization: (a) Task 1; (b) Task 2; (c) Task 3; (d) Task 4.
Figure 12. Feature visualization: (a) Task 1; (b) Task 2; (c) Task 3; (d) Task 4.
Processes 11 01928 g012aProcesses 11 01928 g012b
Figure 13. The accuracy of fault diagnosis when the data of one working condition are used as the training set and the data of another different working condition are used as the testing set.
Figure 13. The accuracy of fault diagnosis when the data of one working condition are used as the training set and the data of another different working condition are used as the testing set.
Processes 11 01928 g013
Figure 14. Results of trials; (a) Task 5; (b) Task 6; (c) Task 7.
Figure 14. Results of trials; (a) Task 5; (b) Task 6; (c) Task 7.
Processes 11 01928 g014
Figure 15. Confusion matrix for the three tasks: (a) Task 5; (b) Task 6; (c) Task 7.
Figure 15. Confusion matrix for the three tasks: (a) Task 5; (b) Task 6; (c) Task 7.
Processes 11 01928 g015
Figure 16. The fault identification results of the proposed method are obtained when the working condition is 1772 rpm/1 hp.
Figure 16. The fault identification results of the proposed method are obtained when the working condition is 1772 rpm/1 hp.
Processes 11 01928 g016
Figure 17. Results of fault classification by different methods at 1730 rpm/3 Hp: (a) 1D-DCNN; (b) IDCNN; (c) IDCNN-DELM.
Figure 17. Results of fault classification by different methods at 1730 rpm/3 Hp: (a) 1D-DCNN; (b) IDCNN; (c) IDCNN-DELM.
Processes 11 01928 g017aProcesses 11 01928 g017b
Table 1. Sample of data set for experiments 1 and 2.
Table 1. Sample of data set for experiments 1 and 2.
NameMotor Speed/rpmMotor Load/(hp)Fault Diameter (Inches)StatusLabelSample
A179700.014Normal IF BF OF0, 1, 2, 3800
B177210.014Normal IF BF OF0, 1, 2, 3800
C175020.014Normal IF BF OF0, 1, 2, 3800
D173030.014Normal IF BF OF0, 1, 2, 3800
Table 2. Sample of data set for Experiment 3.
Table 2. Sample of data set for Experiment 3.
LabelStatusFault Diameter (Inches)Number of Training SamplesNumber of Testing Samples
0Normal-300128
1Ball fault0.007300128
20.014300128
30.021300128
4Inner race fault0.007300128
50.014300128
60.021300128
7Outer race fault0.007300128
80.014300128
90.021300128
Table 3. Structural parameters of IDCNN.
Table 3. Structural parameters of IDCNN.
NameKernal SizeStrideInput SizeOutput SizeActivation Function
Covn164162048 × 1128 × 16ReLU
Max pool122128 × 1664 × 16-
Covn23164 × 1664 × 32ReLU
Max pool22264 × 3232 × 32-
Covn33132 × 3232 × 64ReLU
Max pool33232 × 6416 × 64-
Covn43116 × 6416 × 64ReLU
Max pool43216 × 648 × 32-
BiLSTM--8 × 328 × 32
ECA--8 × 328 × 32Sigmoid
Flatten--8 × 32256-
Fc--256128ReLU
Table 4. Sample selection for Tasks 1–4.
Table 4. Sample selection for Tasks 1–4.
Task NumberTrainSampleTestSample
1A, B800, 800C686
2A, C800, 800D686
3B, C800, 800A686
4C, D800, 800B686
Table 5. Average evaluation index for different methods in Task 1, 2, 3, 4.
Table 5. Average evaluation index for different methods in Task 1, 2, 3, 4.
Task Number1D-DCNN Average Accuracy/F1 ScoreIDCNN Average Accuracy/F1 ScoreOURS Average Accuracy/F1 Score
195.44%/95.44%96.32%/96.33%97.69%/97.68%
295.68%/95.68%96.46%/96.47%98.49%/98.49%
396.85%/96.85%98.19%/98.20%99.32%/99.30%
497.42%/97.42%98.80%/98.80%99.67%/99.66%
Table 6. Average time for different methods in Task 1, 2, 3, 4.
Table 6. Average time for different methods in Task 1, 2, 3, 4.
MethodTimeTask 1Task 2Task 3Task 4
1D-DCNNAverage Training Time (s)12.4211.8912.3012.04
Average Testing Time (s)0.360.450.390.37
IDCNNAverage Training Time (s)20.8720.4220.3620.21
Average Testing Time (s)1.041.031.011.04
OURSAverage Training Time (s)20.9920.1720.5520.18
Average Testing Time (s)10.3210.4510.1410.28
Table 7. Sample selection for Tasks 5–7.
Table 7. Sample selection for Tasks 5–7.
Task NumberTrainSampleTestSample
5B800C343
6C800D343
7D800B343
Table 8. Average evaluation index for different methods in Task 5, 6, 7.
Table 8. Average evaluation index for different methods in Task 5, 6, 7.
Task Number1D-CNN Average Accuracy/F1 ScoreIDCNN Average Accuracy/F1 ScoreOURS Average Accuracy/F1 Score
596.33%/96.33%97.32%/97.33%98.72%/98.72%
696.12%/96.11%97.11%/97.11%98.54%/98.54%
795.60%/95.60%96.68%/96.68%97.73%/97.73%
Table 9. Average time for different methods in Task 5–7.
Table 9. Average time for different methods in Task 5–7.
MethodTimeTask 5Task 6Task 7
1D-DCNNAverage Training Time (s)6.426.546.47
Average Testing Time (s)0.430.440.42
IDCNNAverage Training Time (s)11.4111.5911.62
Average Testing Time (s)0.980.960.98
OURSAverage Training Time (s)11.4711.3511.34
Average Testing Time (s)8.478.258.46
Table 10. The fault identification results of the proposed method are obtained under different load and speed conditions.
Table 10. The fault identification results of the proposed method are obtained under different load and speed conditions.
Speed/LoadStatusNumber of Training SamplesNumber of Testing SamplesAverage AccuracyAverage F1 Score
1797 rpm/0 hpNormal IF BF OF300, 900, 900, 900128, 384, 384, 384100%100%
1772 rpm/1 hpNormal IF BF OF300, 900, 900, 900128, 384, 384, 384100%100%
1750 rpm/2 hpNormal IF BF OF300, 900, 900, 900128, 384, 384, 38499.77%99.77%
1730 rpm/3 hpNormal IF BF OF300, 900, 900, 900128, 384, 384, 384100%100%
Table 11. Fault identification accuracy of different methods under different working conditions.
Table 11. Fault identification accuracy of different methods under different working conditions.
Method1797 rpm/0 hp1772 rpm/1 hp1750 rpm/2 hp1730 rpm/3 hpAverage Accuracy
1D-DCNN98.44%98.42%98.45%98.46%98.44%
IDCNN99.32%99.29%99.30%100%99.47%
OURS100%100%99.77%100%99.94%
Table 12. Fault diagnosis results of other methods.
Table 12. Fault diagnosis results of other methods.
ModelTrainTestAccuracyF1 Score
SVMBC71.5%-
CD71.7%-
DB66.9%-
IDCNNBC97.3%97.3%
CD97.1%97.1%
DB96.7%96.7%
Multilayer Perceptron (MLP)BC84.7%-
CD80.6%-
DB82.6%-
Deep Neural Networks (DNN)BC77.9%-
CD74.1%-
DB78.9%-
Decision Tree (DT)BC42.3%-
CD41.5%-
DB46.7%-
DCNN-BiLSTMBC96.7%96.7%
CD96.9%96.9%
DB97.3%97.4%
CEEMDAN-PSO-SVMBC78.6%-
CD79.2%-
DB78.3%-
HRCMFDE-PSO-ELM [28]BC92.0%-
CD98.5%-
DB84.8%-
OURSBC98.7%98.7%
CD98.5%98.5%
DB97.7%97.7%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, L.; Ping, D.; Wang, C.; Jiang, S.; Shen, J.; Zhang, J. Fault Diagnosis of Rotating Machinery Bearings Based on Improved DCNN and WOA-DELM. Processes 2023, 11, 1928. https://doi.org/10.3390/pr11071928

AMA Style

Wang L, Ping D, Wang C, Jiang S, Shen J, Zhang J. Fault Diagnosis of Rotating Machinery Bearings Based on Improved DCNN and WOA-DELM. Processes. 2023; 11(7):1928. https://doi.org/10.3390/pr11071928

Chicago/Turabian Style

Wang, Lijun, Dongzhi Ping, Chengguang Wang, Shitong Jiang, Jie Shen, and Jianyong Zhang. 2023. "Fault Diagnosis of Rotating Machinery Bearings Based on Improved DCNN and WOA-DELM" Processes 11, no. 7: 1928. https://doi.org/10.3390/pr11071928

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop