Next Article in Journal
Magnetostrictive Sensor for Blockage Detection in Pipes Subjected to High Temperatures
Previous Article in Journal
Copper(I)-Catalyzed Click Chemistry as a Tool for the Functionalization of Nanomaterials and the Preparation of Electrochemical (Bio)Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Lighted Deep Convolutional Neural Network Based Fault Diagnosis of Rotating Machinery

1
Shaanxi Engineering Laboratory for Transmissions and Controls, Northwestern Polytechnical University, Xi’an 710072, China
2
Key Laboratory of Dependable Service Computing in Cyber Physical Society Chongqing University, Chongqing 400044, China
*
Author to whom correspondence should be addressed.
Sensors 2019, 19(10), 2381; https://doi.org/10.3390/s19102381
Submission received: 23 April 2019 / Revised: 17 May 2019 / Accepted: 21 May 2019 / Published: 24 May 2019
(This article belongs to the Section Intelligent Sensors)

Abstract

:
To improve the fault diagnosis performance for rotating machinery, an efficient, noise-resistant end-to-end deep learning (DL) algorithm is proposed based on the advantages of the wavelet packet transform in vibration signal processing (the capability to extract multiscale information and more spectral distribution features) and deep convolutional neural networks (good classification performance, data-driven design and high transfer-learning ability). First, a vibration signal is subjected to pyramid wavelet packet decomposition, and each sub-band coefficient is used as the input for each channel of a deep convolutional network (DCN). Then, based on the lightweight modeling requirements and techniques, a new DCN structure is designed for the fault diagnosis. The proposed algorithm is compared with the support vector machine algorithm and the published DL algorithms based on a bearing dataset produced by Case Western Reserve University. The experimental results show that the proposed algorithm is superior to the existing algorithms in terms of accuracy, memory space, computational complexity, noise resistance, and transfer performance, producing good results.

1. Introduction

Rotating machinery systems have been extensively applied in a number of engineering fields (e.g., aviation, ships and warships, machine tools and vehicles) and play an increasingly pivotal role. Rotating machinery damage and faults not only severely affect the reliability and safety of the entire system but also cause tremendous economic losses. Therefore, researchers have been continuously conducting relevant research. One current focus of research is to extract and classify fault features, i.e., to further improve the identification accuracy and system monitorability by analyzing initial weak fault signals to achieve real-time monitoring and diagnosis.
Based on the differences in the feature extraction and fault diagnosis algorithms, rotating machinery fault diagnosis algorithms can be classified into two types, namely, vibration analysis and intelligent diagnosis [1,2,3,4,5]. In a vibration analysis, signal decomposition techniques (e.g., wavelet transform (WT) [6] and empirical mode decomposition (EMD)) are used to directly detect fault frequencies in the original data [7]. Most fault frequencies are hidden in low-frequency deterministic components and high-frequency noise components and are consequently very difficult to observe in the spectra. As a result, vibration analysis algorithms have a relatively poor practical performance. Intelligent diagnosis is a new research direction and includes artificial neural networks (ANNs) and support vector machines (SVMs) as the main algorithms. Rojas et al. [8] extracted 32-dimensional features from signals using the Fourier transform and classified them using an SVM. Xia et al. [9] proposed to use the Volterra series as a feature for describing the operating conditions of machinery systems and used a backpropagation neural network for classification and discrimination. Lei et al. [10] proposed to extract features using EMD combined with wavelet packet decomposition (WPD) and input selected sensitive features into an ANN for a fault diagnosis. Gangsar et al. [11] proposed to extract fault features from signals using the WT and classify them using an SVM. While conventional intelligent diagnosis algorithms have been extensively applied in machinery signal fault diagnosis, they have the following disadvantages [12,13,14]. (1) The accuracy of the fault diagnosis is dependent on the quality of the feature extraction. Vibration signals collected in an industrial environment are always complex, unstable and contain high levels of noise. Thus, to ensure the quality of the feature extraction, a manual design and suitable features are needed based on the characteristics of the different faults. The quality of the features directly determines the system performance. Therefore, the system feature extraction is not automatic. (2) Conventional intelligent diagnosis algorithms are based on shallow learning models, which are unable to effectively learn nonlinear relations in complex systems.
To address these deficiencies, deep learning (DL) has been gradually used in machinery signal fault diagnosis. Based on the model used in the DL, the available studies are classified into three types, namely, those involving autoencoders (AEs), convolutional neural networks (CNNs) and recurrent neural networks (RNNs) [4]. Jia et al. [15] pretrained a three-layer deep neural network (DNN) by stacking AEs and obtained final prediction results by fine-tuning the DNN. Li et al. [16] proposed a deep random forest fusion structure, which extracts signal features using the WT in conjunction with a deep Boltzmann machine and produces classification results by feature fusion using a random forest. Gan et al. [17] extracted the wavelet packet energy of the original signal as the feature input and then established a layered deep belief network-based diagnostic network, with its first layer diagnosing the fault type and the second layer determining the severity of the fault. Relying on network optimization using sparse regularization, Sun et al. [18] proposed an algorithm for predicting motor faults. This type of algorithm can be relatively easily realized and can obtain feature representation by learning, but it is slow in convergence and has a weak transfer-learning ability. The main RNN algorithms include RNNs and long short-term memory neural networks (LSTMNNs). Zhao et al. [19] proposed an LSTMNN-based fault diagnosis algorithm. This type of algorithm is effective in detecting faults in time-series data and is able to discover problems that arise as time elapses, but it has a relatively high network complexity and a weak transfer-learning ability.
There are a gradually increasing number of studies involving deep CNNs (DCNNs) in the fault diagnosis field. Abdeljaber et al. [20] established a one-dimensional (1D) LeNet5 network based on the LeNet5 architecture for detecting structural damage in machinery. DCNN models that differ in structure have also been proposed, based on 1D CNNs, for predicting faults in various types of rotating machinery [21,22,23,24]. However, most of these models use the original signal as the input but neglect its frequency-domain characteristics. DCNN models that differ in structure have been proposed based on CNNs for fault diagnosis [25,26,27]. In [25], a 1D problem was converted into a two-dimensional (2D) signal processing problem using a continuous WT scalogram. In [26,27], a 1D problem was converted into a 2D signal-processing problem by a non-overlapping equal division of the original signal, before a fault prediction was performed using a 2D convolution algorithm. Zhao et al. [28] proposed to use a wavelet packet-residual network hybrid algorithm for predicting faults in gearboxes. This type of algorithm is effective in analyzing multidimensional data and can effectively extract local features. Compared to 1D processing, 2D representation has a relatively complex structure, and 2D processing requires more time and computational resources in the training and testing processes. For example, the computational load of a 3 × 1 convolution is less than one-third that of a 3 × 3 convolution.
While the available DL algorithms are effective, the collection of original signals in an actual industrial production environment is relatively significantly affected by various components, requiring that the algorithms be highly resistant to disturbances. In addition, with the development of the industrial internet of things (IIoT), the installation of a health management unit in rotating machinery has become a trend. Constrained by costs, as well as by actual hardware and environmental conditions, the improvement of the functional intelligence level of software has high requirements for operating environment resources. As a result, the requirements for the computational complexity of DL algorithms and the space occupied by models are becoming increasingly stringent. Consequently, there are increasingly higher “small-size, lightweight, and rapidness” requirements for DL algorithms [25,29]. The complex and varied conditions of rotating machinery (with significant changes in the rotational speed and load, particularly when the system starts and shuts down) result in unstable collected data. As the conditions change and time elapses, the sample distribution no longer meets the same distribution requirement, leading to a demand for a high transfer-learning ability. The available studies are deficient in lightweight DL (as shown in Table 1), cannot meet the new requirements of the IIoT, and fail to sufficiently consider the noise resistance and transfer-learning ability of algorithms.
Hence, in this study, a lightweight network structure is proposed based on the advantages of the WT and DL. The main contributions of this study are as follows.
(1) A 1D CNN structure is proposed that can effectively improve the identification accuracy and that has a relatively high transfer-learning ability and noise resistance.
(2) The proposed network structure is relatively lightweight and has a high computational speed, and it occupies little space.
The paper is organized as follows: Section 2 elaborates the theoretical basis of the WPT and DCN for a fault diagnosis based on the proposed method; The lightweight CNN structure design is described in Section 3; Experiments and analysis are implemented based on a bearing dataset produced by Case Western Reserve University in Section 4 to verify that the proposed algorithm is superior to the existing algorithms; Finally, the concluding remarks are given in Section 5.

2. Theoretical Basis

2.1. Wavelet Packet Transform (WPT)

The framework of a wavelet packet transform is the extension of the wavelet transform [30]. The wavelet packet function is also a time-frequency function and can be described as:
w i , j n = 2 w n ( 2 j t k )
where j and k are integers; they are the index scale and translation operations, respectively. The index n is an operation modulation parameter or oscillation parameter. The first two wavelet packet functions are the scaling and mother wavelet functions:
w 0 , 0 0 ( t ) = φ ( t )
w 0 , 0 1 ( t ) = ψ ( t )
When n = 2, 3, …, the function can be defined by the following recursive relationships:
w 0 , 0 2 n ( t ) = 2 k h ( k ) w 1 , k n ( 2 t k )
w 0 , 0 2 n + 1 ( t ) = 2 k g ( k ) w 1 , k n ( 2 t k )
where h ( k ) and g ( k ) are the quadrature mirror filters (QMFs) associated with the predefined scaling function and the mother wavelet function, respectively. The wavelet packet coefficients, w i , j n are computed by the inner product < f ( t ) w i , j n >, which can be defined as:
w i , j n = f ( t ) w i , j n ( t ) d t
The framework of the WPT algorithm broken up into three resolution levels is shown in Figure 1, and the structure of WPT is a Perfect Binary Tree.
WPT can further obtain the detail wavelet coefficients of the signal in the high frequency area and provides a more detailed and comprehensive time-frequency plane tiling than the discrete wavelet transform (DWT) does [31]. As shown in Figure 2, the many advantages of WPT are used in the discrete signal processing, such as the fault diagnosis of rotating machinery [11] image processing [32] and video processing [33].

2.2. Deep Convolutional Networks (DCNs)

DCNs, proposed by LeCun, are an important branch of DL [34,35]. DCNs consist of convolutional and pooling layers as well as activation and loss functions. The convolutional layers use various types of convolution kernels and perform convolution operations on the input signals. Convolution operations are translation invariant for 1D signals and can support neurons in learning features with a relatively high robustness [36].
The pooling layers perform down-sampling operations, which use a specific value as the output value in each specific small area. Generally, the maximum or mean value is used as the output value. Down-sampling operations are performed to non-linearize 1D signals.
The activation functions stimulate neurons based on a series of input values, weights of interneuron connections and activation rules. The loss functions in a CNN are used to evaluate the difference between the output and the actual value at the training stage. Afterwards, the values of the loss functions are used to update the weights between the neurons. The purpose of the training is to minimize the values of the loss functions.
A batch normalization (BN) layer normalizes the data input into each layer of the network (normalized to a mean of 0 and a variance of 1). The BN layers not only improve the gradient flow in the network, allow a higher learning rate, and improve the training speed, but they also reduce the strong dependence on initialization.

2.3. One-by-One Convolutions

A 1 × 1 convolution filter is the same as a normal filter except that it has a size of 1 × 1 and does not consider inter-information relationships in the local areas of the previous layer. One-by-one convolutions first appeared in the Network in Network paper and were used to deepen and widen network structures by reducing the dimensionality in inception networks [37]. To reduce the computational complexity of a system, a 1 × 1 convolution can be used to reduce the dimensionality to compress the channels. In [38], the effect of a 1 × 1 convolution with a different proportion of channel numbers on the performance is discussed and verified, and a squeeze ratio of 0.5 was found to be relatively satisfactory after careful consideration. Here, a two-layer convolutional network is presented in Figure 3 as an example. Figure 3a shows a dimensionality reduction without performing a 1 × 1 convolution, and Figure 3b shows a dimensionality reduction by performing a 1 × 1 convolution.
The effect of a 1 × 1 convolution with different proportions of channel numbers on the performance is discussed and verified.

3. Lightweight CNN Structure Design

For 1D fault signal classification problems, an end-to-end network structure (as shown in Figure 4) is proposed. Based on the function, this network can be divided into two layers. The first layer performs the wavelet packet transform (WPT), with an aim to extract finer information from a frequency-domain perspective, and the second layer is the designed relatively lightweight CNN.

3.1. DCNN Design

The computational resources in a DCNN are affected by the number of channels, kernel size, data length and connection mode. To improve the ability to predict faults in the rotating machinery, different numbers of channels and structures are used in the CNN partly based on the main factors affecting the network convolutions. In this study, seven DCNs are designed for the 1D convolutional networks and their computational loads and storage space (as shown in Table 2). The number of channels and the kernel size differ among the different structures (models). In Table 2, K/C represents the kernel size/number of channels (e.g., 11 × 1/4 represents a kernel size of 11 × 1 and 4 kernels (channels)). In network design, deepening the network can enhance its ability to learn features and can reduce the number of network parameters and the training load.
As shown in Figure 5, the computational complexity and number of parameters of network structure model 2 are 16% and 9% lower, respectively, than those of model 1.
The WPT-CNN structure is described as follows.
  • The conv_1 to conv_5 convolutional layers of the network include convolution operations, rectified linear unit (ReLU) activation functions and BN (as shown in Figure 6). The use of the ReLU activation functions and BN can increase the rate of convergence and prevent a gradient explosion and vanishing problems. Maximum pooling is used in the poo1_1 to pool_4 pooling layers to allow the features that pass through the pooling layers to be translation invariant, thereby removing some noise while reducing the dimensionality of the features.
  • The lengths of the conv_1 to conv_5 convolutional layers of the network are set to 11, 3, 3, 3 and 3, respectively, using the smoothing function of the convolution kernel. The length of the conv_1 layer is set to 11 to suppress noise.
  • An alternating cascade of convolutional layers with kernel sizes of 3 × 3 and 1 × 1 is used in the conv_3 to conv_5 convolutional blocks to reduce the dimensionality while increasing the depth of the network, thereby reducing the number of training parameters and realizing a lightweight model. This method has been proven effective elsewhere [29].
  • In the second half of the network, the conv_6 and pool_6 layers are used to substitute for a commonly used fully connected layer to reduce the risk of overfitting caused by a fully connected layer while reducing the number of network parameters. A linear activation function is used in the conv_6 convolutional layer because a linear mapping is required to map the number of channels of the input feature to the same value of the classification. Other activation functions are not able to achieve this and may even slow down the convergence of the network. Global average pooling is used in the pool_5 pooling layer to allow the feature map input into each channel to correspond to each output feature, thereby improving the consistency between the feature maps and output types. The stability of the pooling process is also enhanced by the summation of the spatial information.
Here, the example shown in the network map is used for the analysis. The signal has 16 dimensions and 16 channels after passing through the pool_5 pooling layer. After the feature translation, 16 × 16 = 256 dimensions are obtained. A fully connected layer is then needed to reduce the dimensionality to 10 for the classification. Thus, a total of 256 × 10 = 2560 parameters need to be trained. However, after the conv_6 and pool_5 layers are used to substitute for the fully connected layer, only the parameters in the conv_6 layer (a total of 1 × 16 × 10 = 160 parameters) need to be trained, whereas no parameters in the pool_6 layer need to be trained. Hence, the number of parameters decreases 16-fold from 2560 to 160, thereby saving the memory space occupied by the model parameters.
5. The cross-entropy loss is used as the loss function, as shown below:
L o s s ( p , q ) = x p ( x ) log q ( x )
where p ( x ) is the label of the training set and q ( x ) is the label value predicted by the network.
In classification problems, the cross-entropy function is often used as a loss function because the gradient of the cross-entropy loss is only related to the correct classification prediction results in the model optimization process. In this way, updating the network parameters only increases the correct classification but does not affect other classifications.
6. Random number initialization
In DL, each convolution kernel is initialized with a random value, which has a certain impact on the training and system performance of the network. In this study, a Xavier initialization is used to promote the convergence of the model [39]. The weight w = R d i n × d o u t is initialized using the following equation:
w u [ 6 d i n + d o u t , + 6 d i n + d o u t ]
where u [a, b] is a uniform sample of the range [a, b].

3.2. Performance Analysis of Network Parameters

The proposed WPT-CNN hybrid network structure is a lightweight network structure that has significant advantages over some commonly used network structures in terms of the computational load and number of parameters. In other network structures, convolutional layers and fully connected layers are the parts that participate in the computation of the floating numbers and contain parameters that need to be trained. The calculation equations are as follows:
P a r a m e t e r s c o n v = K h × K w × C i n × C o u t + C o u t
P a r a m s s f c = I × O + O
F L O P c o n v = 2 × H × W × K h × K w × C i n × C o u t
F L O P s f c = 2 × I × O
where Parametersconv and FLOPconv represent the value of the number of parameters and the computation of the floating numbers in the convolutional layer; Paramssfc and FLOPsfc represent the values of the number of parameters and the computation of floating number in the fully connected layer, respectively; H , W and Cin are the height, width and number of channels of the input feature map, respectively; Kh and Kw represent the size of the convolution kernel; Cout is the number of convolution kernels; I is the dimensionality of the input; and O is the dimensionality of the output.
Various network structures, including WPT-CNN, DNN, 1D LeNet-5, residual network ResNet-18, ResNet-50 and Visual Geometry Group (VGG)-16, are compared in terms of the number of parameters and the computation of the floating numbers. DNN is the deep neural network proposed in [15] for a machinery fault diagnosis, which is pretrained by stacking AEs. DNN contains three hidden layers, each of which contains 600,200,100 neurons. 1D LeNet5 is a 1D CNN that is obtained by improving the LeNet-5 and that is used for motor fault detection. Resnet-18, Resnet-50 and VGG-16 are commonly used CNN frameworks. To accommodate 1D input data, 2D convolutions in these three networks were altered to 1D convolutions for calculation. The input length was determined by the input length of the methods reported in the literature and is an evaluation basis for the computation complexity and number of parameters (see Table 3).
The calculation results show that the proposed algorithm has the lowest computation of floating number computational load (3–500 times lower than that of the available algorithms) and the fewest number of parameters (3–2000 times fewer than those of the available algorithms). In addition, the proposed hybrid network contains 14 convolutional layers and is thus much deeper, and it has a higher ability to extract features than some of the available algorithms (e.g., [23,24,25,26,27]). See the subsequent comparison experiments for details.

4. Experiment and Analysis

The hardware and software environments used to implement the algorithm are shown in Table 4.

4.1. Experimental Data

4.1.1. Introduction of the Dataset

The bearing dataset collected by the Case Western Reserve University (CWRU)’s Bearing Data Center were used to validate the effectiveness of the proposed WPT-CNN network structure. Figure 7 shows the experimental platform for the data collection [40]. The dataset contains the data for bearings under four different conditions, that is, normal bearings, bearings with a faulty ball (ball), bearings with a faulty inner race (inner_race) and bearings with a faulty outer race (outer_race). For each type of fault, there are three fault diameters, that is, 0.007 in., 0.014 in. and 0.021 in. Thus, with the inclusion of normal bearings, this dataset contains the data for 10 classification groups of bearings (as shown in Table 5).

4.1.2. Data Preparation

Because their volume was limited, the experimental data were augmented using the overlap sampling method reported in [15,24], as shown in Figure 8. In this study, the overlap length was determined based on the data length.
In the experiments, the dataset was divided into eight training sets, one validation set, and one test set. In addition, 800 training samples, 100 test samples and 100 validation samples were randomly selected. In the training process, the validation set was used to examine the identification accuracy after every 10 epochs.
As shown in Table 5, the dataset was divided into four sub-datasets based on the number of loads. These are A, B, C and D, corresponding to the data collected under 0, 1, 2 and 3 loads, respectively. The datasets A, B, C and D each contain 800 training samples and 100 test samples of each type, totaling 8000 training samples and 1000 test samples.

4.1.3. Experimental Description

The CWRU dataset was collected in an environment with a relatively low level of ambient noise, and there are a number of noise sources in an actual environment. Therefore, in order to reflect the ability of the algorithm proposed in this paper, noise was added into the samples in the original test set.
A 10%, 30%, 50%, 70%, 90% and 100% white Gaussian noise signal was added to the data to simulate actual conditions. The signal-to-noise ratio (SNR) was used when adding the noise signal and is defined as follows:
SNR = 10 log 10 P s i g n a l / P n i o s e
where Psignal and Pniose are the intensity of the original and noise signals, respectively.
Figure 9 shows the original vibration signal from a bearing with a 0.007 in. faulty inner race under 0 load and the signals with various percentages of added noise. Note that the data with 0% noise represent the results under normal conditions.
The Daubechies-1 wavelet function was selected for all experiments. The filter of this function has various lengths and a low smoothing effect, which is favorable for the extraction of detailed signal features.
Various types of noise are present in an industrial production environment. Therefore, fault prediction algorithms must be relatively highly resistant to noise. Because the number of loads and the operating rotational speed of a machine change when it is in operation, the distribution of the collected sample data varies between different conditions. Therefore, algorithms are required to have a relatively high transfer-learning ability for various loads.
To examine the noise resistance of the proposed algorithm, the learning model with the highest accuracy for the validation set in 1000 iterations was selected. Noise-containing data were randomly generated 10 times for each sample in the training set of dataset A and used for testing and statistically analyzing the experimental results, which were represented in the form of the mean and standard deviation. Similarly, dataset A was used as a training set to train the model, and datasets B, C and D were used to test the accuracy of the algorithm (denoted by AB, AC and AD, respectively) in order to determine the transfer-learning ability for various loads.

4.2. Experiment and Analysis

Experiment 1: The effect of the different number of decompositions of WPT on the accuracy and noise resistance of the algorithm.
In this section, to determine the best number of decompositions of WPT, the network structure model 2 with three convolution kernels in the first convolutional layer is selected. The accuracy of the proposed network structure in the time domain and the different number of decompositions of WPT under different percentages of noise are obtained for the validation and test sets. Table 6 summarizes the statistical experimental results.
Because the original signal was used as the input signal, the input signal had a length of 1024. In WPT, as the decomposition level increases, the length of each sub-band decreases by a factor of 2, whereas the number of sub-bands increases by a factor of 2. For example, when decomposed at levels 3, 4 and 5, the length of each sub-band is 128, 64 and 32, and there are 8, 16 and 32 sub-bands, respectively. The experimental results show that for a 1D input, a sub-band length of 64 resulted in a satisfactory division of the frequency domain, a moderate length and a large amount of information.
Experiment 2: Effects of various sizes of convolution kernels in the first layer on the noise resistance of the algorithm.
Table 6 shows that the number of decompositions of the WPT level is 4, which has the highest noise resistance, so the WPT 4 and network structure model 2 are thus selected in this section. An experiment was conducted to compare the effects of the various sizes of convolution kernels in the first layer. Table 7 summarizes the experimental results.
The experimental data in Table 7 show that when taking the computational load and performance into account, a first-layer convolution kernel size of 11 produced a satisfactory noise-suppressing effect.
Experiment 3: Comparison of the performance of three different CNN models
The effects of the number of channels on the noise resistance and transfer-learning ability of the network structures were comparatively analyzed. The data with 0% noise represents the results under normal conditions.
As demonstrated in Table 8, model 2 was superior to model 1, with the same number of channels in each metric, suggesting that a 1 × 1 convolution can help improve the performance.
The number of channels of each model can be set to 4, 8, 16, 32, 64, and 128. When there are 4 or 8 channels, to ensure that there are sufficient channels for depicting features, a dimensionality reduction is generally not performed. When there are a large number of channels (e.g., 16, 32, 64 and 128), a dimensionality reduction can be performed but does not significantly improve the performance. This is because inter-elemental relationships in a 1D signal are simple compared to a 2D signal, and because an excessive number of channels cannot facilitate an improved performance. However, an increase in the number of channels is accompanied by a multifold increase in the number of parameters and in the computational complexity, lowering the cost effectiveness.
Experiment 4: Comparison of the performance of the network structures with different depths.
In DL, the network depth affects the performance and computational complexity of an algorithm. Here, network structures with various depths were designed based on model 2 (see Table 9) and compared in terms of the noise resistance and transferability. Table 10 summarizes the experimental results.
The experimental data show that model 2 with a five-layer structure was satisfactory in terms of the performance, number of parameters and computational complexity.
Experiment 5: Comparison of the transfer-learning ability for various levels of noise.
The transfer-learning ability of DL model 2, which had the best performance, was comparatively analyzed for various levels of noise. Table 11 summarizes the experimental results under different dataset combinations. For example, AB represents the test accuracy on sub-dataset B using a model trained by sub-dataset A.
The experimental data show that various noise intensity levels weakened the transfer-learning ability of model 2 to a certain extent (3–10%).
Experiment 6: Comparison of the noise resistance and transfer-learning ability of various algorithms.
To determine its noise resistance and transfer-learning ability, the proposed algorithm was compared with the available algorithms by using dataset A. Table 12 and Table 13 summarize the experimental results.
The experimental data show that the proposed algorithm exhibited the highest resistance to various noise intensity levels, 5–58% higher than that of the available algorithms.
The experimental data show that the proposed algorithm with the structure of model 2 exhibited the highest performance for various datasets, and its transfer-learning ability was 0.5–44% higher than that of the available algorithms.

4.3. Visualization of the Network Learning Process

It is impossible to illustrate how a CNN works because its working principle is similar to that of a black box. Therefore, the output characteristics of the network training process and the model testing process were subjected to a dimensionality reduction using the t-distributed stochastic neighbor embedding technique, and they were then visualized to facilitate the understanding of the operating process of the entire network.
Figure 10 shows the visualized network training process, demonstrating the feature distribution corresponding to the 0th to 500th iteration of the proposed algorithm on dataset A. As shown in Figure 10, as the number of times of the network training increases, the feature distance between the labels of different types increases, whereas the feature distance between the labels of the same type decreases, and they gradually form a cluster. This observation suggests that as the number of training iterations increases, the network’s ability to learn features and its classification accuracy improve.
Figure 11 shows the visualized network testing process, demonstrating the feature distribution corresponding to when the data pass through the conv_1 to conv_6 layers of the hybrid network. As shown in Figure 11, the features of all the samples are mixed and indistinguishable at the network initialization stage. As the number of network layers increases, the features gradually become distinguishable. In the conv_5 layer, the features of the same labels self-organize into one cluster.

5. Conclusions

This study investigates fault prediction in rotating machinery. Based on the ability of the wavelet packet transform to extract more frequency-domain information and based on the low computational load of 1D convolutions, a convolutional network-based lightweight deep learning fault prediction algorithm is proposed. This network can be regarded as having two layers. The first layer performs the wavelet packet transform (WPT), with an aim to extract finer information from a frequency-domain perspective, and the second layer is the designed relatively lightweight CNN (as shown in Figure 4). To validate the effectiveness of the proposed WPT-CNN network structure, the bearing dataset collected by the Case Western Reserve University (CWRU)’s Bearing Data Center were used to implement the experiments. The comparison results demonstrate that the proposed algorithm is superior to the available algorithms in noise resistance (5–58% higher) and transfer-learning ability (0.5–44% higher). In addition, the proposed algorithm also has the lowest computational complexity and memory space requirement (72.5% and 88.5% lower than those of the available algorithms, respectively). Therefore, the proposed algorithm can effectively improve the identification accuracy and is relatively lightweight.
Future research will consist in developing a new model for the deep convolution neural network optimization problem by improving the regularization term and optimization strategy with the results of this paper to predict the fault characteristics of the rotating machinery. Furthermore, to meet the actual working conditions of industrial production, future research will also focus on improving the generalization ability and the lightweight problem of the model.

Author Contributions

Conceptualization and methodology, S.M. and Z.S.; validation and data analysis, W.C. and W.L.; writing—original draft preparation, S.M., W.C. and W.L.; writing—review and editing, G.L.; funding acquisition, S.M. and G.L.

Funding

The research is supported by National Natural Science Foundation of China (Grant No. 51875458), Key Project of National Natural Science Foundation of China (Grant No. 51535009) and the 111 Project (Grant No. B13044).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, R.N.; Yang, B.Y.; Zio, E.; Chen, X.F. Artificial intelligence for fault diagnosis of rotating machinery: A review. Mech. Syst. Signal Process. 2018, 108, 33–47. [Google Scholar] [CrossRef]
  2. Yan, R.Q.; Gao, R.X.; Chen, X.F. Wavelets for fault diagnosis of rotary machines: A review with applications. Signal Process. 2014, 96, 1–15. [Google Scholar] [CrossRef]
  3. Chen, J.L.; Li, Z.P.; Pan, J.; Chen, G.G.; Zi, Y.Y.; Yuan, J.; Chen, B.Q.; He, Z.J. Wavelet transform based on inner product in fault diagnosis of rotating machinery: A review. Mech. Syst. Signal Process. 2016, 70, 1–35. [Google Scholar] [CrossRef]
  4. Khan, S.; Yairi, T. A review on the application of deep learning in system health management. Mech. Syst. Signal Process. 2018, 107, 241–265. [Google Scholar] [CrossRef]
  5. Shao, H.D.; Jiang, H.K.; Zhang, H.Z.; Duan, W.J.; Liang, T.C.; Wu, S.P. Rolling bearing fault feature learning using improved convolutional deep belief network with compressed sensing. Mech. Syst. Signal Process. 2018, 100, 743–765. [Google Scholar] [CrossRef]
  6. Adamczak, S.; Stępień, K.; Wrzochal, M. Comparative study of measurement systems used to evaluate vibrations of rolling bearings. Procedia Eng. 2017, 192, 971–975. [Google Scholar] [CrossRef]
  7. Lei, Y.G.; Lin, J.; Zuo, M.J.; He, Z.J. Condition monitoring and fault diagnosis of planetary gearboxes: A review. Measurement 2014, 48, 292–305. [Google Scholar] [CrossRef]
  8. Rojas, A.; Nandi, A.K. Practical scheme for fast detection and classification of rolling-element bearing faults using support vector machines. Mech. Syst. Signal Process. 2006, 20, 1523–1536. [Google Scholar] [CrossRef]
  9. Xia, H.; He, Y.G.; Wu, J. Fault-dictionary diagnostic method in frequency domain for nonlinear networks based on Volterra series and backward propagation neural networks (BPNN). J. Hunan Univ. Nat. Sci. 2004, 31, 41–43. [Google Scholar]
  10. Lei, Y.G.; He, Z.J.; Zi, Y.Y. Application of an intelligent classification method to mechanical fault diagnosis. Expert Syst. Appl. 2009, 36, 9941–9948. [Google Scholar] [CrossRef]
  11. Gangsar, P.; Tiwari, R. Multifault Diagnosis of induction motor at intermediate operating conditions using wavelet packet transform and support vector machine. J. Dyn. Syst-T. ASME 2018, 140, 081014-1-8. [Google Scholar] [CrossRef]
  12. Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
  13. Shao, H.D.; Jiang, H.K.; Wang, F.A.; Zhao, H.W. An enhancement deep feature fusion method for rotating machinery fault diagnosis. Knowl. Based Syst. 2017, 119, 200–220. [Google Scholar] [CrossRef]
  14. Rosa, E.D.L.; Yu, W. Randomized algorithms for nonlinear system identification with deep learning modification. Inf. Sci. 2016, 364–365, 197–212. [Google Scholar] [CrossRef]
  15. Jia, F.; Lei, Y.G.; Lin, J.; Zhou, X.; Lu, N. Deep neural networks: A promising tool for fault characteristic mining and intelligent diagnosis of rotating machinery with massive data. Mech. Syst. Signal Process. 2016, 72–73, 303–315. [Google Scholar] [CrossRef]
  16. Li, C.; Sanchez, R.V.; Zurita, G.; Cerrada, M.; Cabrera, D.; Vásquez, R.E. Gearbox fault diagnosis based on deep random forest fusion of acoustic and vibratory signals. Mech. Syst. Signal Process. 2016, 76–77, 283–293. [Google Scholar] [CrossRef]
  17. Gan, M.; Wang, C.; Zhu, C.A. Construction of hierarchical diagnosis network based on deep learning and its application in the fault pattern recognition of rolling element bearings. Mech. Syst. Signal Process. 2016, 72–73, 92–104. [Google Scholar] [CrossRef]
  18. Sun, C.; Ma, M.; Zhao, Z.B.; Chen, X.F. Sparse deep stacking network for fault diagnosis of motor. IEEE Trans. Ind. Inform. 2018, 14, 3261–3270. [Google Scholar] [CrossRef]
  19. Zhao, R.; Yan, R.Q.; Wang, J.J.; Mao, K.Z. Learning to monitor machine health with convolutional bi-directional LSTM networks. Sensors 2017, 17, 273. [Google Scholar] [CrossRef] [PubMed]
  20. Abdeljaber, O.; Avci, O.; Kiranyaz, S.; Gabbouj, M.; Inman, D.J. Real-time vibration-based structural damage detection using one-dimensional convolutional neural networks. J. Sound Vib. 2017, 388, 154–170. [Google Scholar] [CrossRef]
  21. Janssens, O.; Slavkovikj, V.; Vervisch, B.; Stockman, K.; Loccufier, M.; Verstockt, S.; Van de Walle, R.; Van Hoecke, S. Convolutional neural network based fault detection for rotating machinery. J. Sound Vib. 2016, 377, 331–345. [Google Scholar] [CrossRef]
  22. Ince, T.; Kiranyaz, S.; Eren, L.; Askar, M.; Gabbouj, M. Real-time motor fault detection by 1-D convolutional neural networks. IEEE Trans. Ind. Electron. 2016, 63, 7067–7075. [Google Scholar] [CrossRef]
  23. Zhuang, Z.L.; Qin, W. Intelligent fault diagnosis of rolling bearing using one-dimensional multi-scale deep convolutional neural network based health state classification. In Proceedings of the IEEE 15th International Conference on Networking, Sensing and Control (ICNSC), Zhuhai, China, 27–29 March 2018. [Google Scholar]
  24. Zhang, W.; Li, C.H.; Peng, G.L.; Chen, Y.H.; Zhang, Z.J. A deep convolutional neural network with new training methods for bearing fault diagnosis under noisy environment and different working load. Mech. Syst. Signal Process. 2018, 100, 439–453. [Google Scholar] [CrossRef]
  25. Wen, L.; Li, X.Y.; Gao, L.; Zhang, Y.Y. A new convolutional neural network-based data-driven fault diagnosis method. IEEE Trans. Ind. Electron. 2018, 65, 5990–5998. [Google Scholar] [CrossRef]
  26. Guo, S.; Yang, T.; Gao, W.; Zhang, C. A novel fault diagnosis method for rotating machinery based on a convolutional neural network. Sensors 2018, 18, 1429. [Google Scholar] [CrossRef]
  27. Liu, R.N.; Meng, G.T.; Yang, B.Y.; Sun, C.; Chen, X.F. Dislocated time series convolutional neural architecture: An intelligent fault diagnosis approach for electric machine. IEEE Trans. Ind. Inf. 2017, 13, 1310–1320. [Google Scholar] [CrossRef]
  28. Zhao, M.H.; Kang, M.; Tang, B.P.; Pecht, M. Deep residual networks with dynamically weighted wavelet coefficients for fault diagnosis of planetary gearboxes. IEEE Trans. Ind. Electron. 2018, 65, 4290–4300. [Google Scholar] [CrossRef]
  29. Hatcher, W.G.; Yu, W. A survey of deep learning: Platforms, applications and emerging research trends. IEEE Access 2018, 6, 24411–24432. [Google Scholar] [CrossRef]
  30. Sun, Z.; Chang, C.C. Structural damage assessment based on wavelet packet transform. J. Struct. Eng. 2002, 128, 1354–1361. [Google Scholar] [CrossRef]
  31. Stępień, K.; Makieła, W. An analysis of deviations of cylindrical surfaces with the use of wavelet transform. Metrol. Meas. Syst. 2013, 20, 139–150. [Google Scholar] [CrossRef]
  32. Dibal, P.Y.; Onwuka, E.N.; Agajo, J.; Alenoghena, C.O. Enhanced discrete wavelet packet sub-band frequency edge detection using Hilbert transform. Int. J. Wavelets Multiresolut. Inf. Process. 2018, 16, 3880–3882. [Google Scholar] [CrossRef]
  33. Zong, X.F.; Men, A.D.; Yang, B. Rate-distortion optimal wavelet packet transform for low bit rate video coding. In Proceedings of the IEEE International Workshop on Imaging Systems and Techniques, Shenzhen, China, 11–12 May 2009; pp. 359–363. [Google Scholar]
  34. Akhtar, N.; Mian, A. Threat of adversarial attacks on deep learning in computer vision: A survey. IEEE Access 2018, 6, 14410–14430. [Google Scholar] [CrossRef]
  35. LeCun, Y.; Bengio, Y.; Hinton, G.E. Review: Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  36. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  37. Szegedy, C.; Liu, W.; Jia, Y.Q.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  38. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E.H. Squeeze-and-Excitation Networks. IEEE Trans. PAMI. Available online: https://arxiv.org/abs/1709.01507 (accessed on 5 December 2018).
  39. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the 13th International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010. [Google Scholar]
  40. Data Sets. Available online: http://csegroups.case.edu/bearingdatacenter/pages/download-data-file (accessed on 23 February 2019).
Figure 1. Tree structures of the wavelet packet transform.
Figure 1. Tree structures of the wavelet packet transform.
Sensors 19 02381 g001
Figure 2. The advantages of WPT: (a) Dyadic WT time-frequency plane tiling; (b) Dyadic WPT time-frequency plane tiling.
Figure 2. The advantages of WPT: (a) Dyadic WT time-frequency plane tiling; (b) Dyadic WPT time-frequency plane tiling.
Sensors 19 02381 g002
Figure 3. Comparison of different structures: (a) Dimensionality reduction without performing a 1 × 1 convolution; (b) Dimensionality reduction by performing a 1 × 1 convolution.
Figure 3. Comparison of different structures: (a) Dimensionality reduction without performing a 1 × 1 convolution; (b) Dimensionality reduction by performing a 1 × 1 convolution.
Sensors 19 02381 g003
Figure 4. Lightweight CNN.
Figure 4. Lightweight CNN.
Sensors 19 02381 g004
Figure 5. Structure of the CNN diagram: (a) The WPT-CNN network structures of model 1; (b) The WPT-CNN network structures of model 2.
Figure 5. Structure of the CNN diagram: (a) The WPT-CNN network structures of model 1; (b) The WPT-CNN network structures of model 2.
Sensors 19 02381 g005
Figure 6. Structure of conv_i.
Figure 6. Structure of conv_i.
Sensors 19 02381 g006
Figure 7. Simulation experimental platform for rolling bearing faults.
Figure 7. Simulation experimental platform for rolling bearing faults.
Sensors 19 02381 g007
Figure 8. Schematic diagram of the data augmentation.
Figure 8. Schematic diagram of the data augmentation.
Sensors 19 02381 g008
Figure 9. Original signal and signals with various percentages of added white Gaussian noise.
Figure 9. Original signal and signals with various percentages of added white Gaussian noise.
Sensors 19 02381 g009
Figure 10. Feature distribution corresponding to the number of iterations.
Figure 10. Feature distribution corresponding to the number of iterations.
Sensors 19 02381 g010
Figure 11. Feature distribution corresponding to number of layers.
Figure 11. Feature distribution corresponding to number of layers.
Sensors 19 02381 g011
Table 1. Comparison of computational resource requirements of the main available algorithms.
Table 1. Comparison of computational resource requirements of the main available algorithms.
ReferencesMemory RequirementComputational Complexity
[28]72.35 KB1.40 × 106
[23]203.80 KB6.81 × 107
[24]258.52 KB4.69 × 106
[25]367.814 KB2.00 × 108
[27]565.16 KB8.08 × 107
[15]2696.54 KB1.37 × 106
Table 2. Number of channels and structures for various models.
Table 2. Number of channels and structures for various models.
Model 1Model 2Model 3Model 4Model 5Model 6Model 7
K/CK/CK/CK/CK/CK/CK/C
conv_111 × 1/411 × 1/411 × 1/411 × 1/411 × 1/811 × 1/1611 × 1/32
conv_23 × 1/83 × 1/83 × 1/83 × 1/83 × 1/163 × 1/323 × 1/64
conv_3_13 × 1/163 × 1/163 × 1/163 × 1/163 × 1/323 × 1/643 × 1/128
conv_3_2/1 × 1/81 × 1/81 × 1/81 × 1/161 × 1/321 × 1/64
conv_3_33 × 1/163 × 1/163 × 1/163 × 1/163 × 1/323 × 1/643 × 1/128
conv_4_13 × 1/163 × 1/163 × 1/323 × 1/323 × 1/643 × 1/1283 × 1/256
conv_4_2/1 × 1/81 × 1/161 × 1/161 × 1/321 × 1/641 × 1/128
conv_4_33 × 1/163 × 1/163 × 1/323 × 1/323 × 1/643 × 1/1283 × 1/256
conv_5_13 × 1/163 × 1/163 × 1/323 × 1/643 × 1/1283 × 1/2563 × 1/512
conv_5_2/1 × 1/81 × 1/161 × 1/321 × 1/641 × 1/1281 × 1/246
conv_5_33 × 1/163 × 1/163 × 1/323 × 1/643 × 1/1283 × 1/2563 × 1/512
conv_5_4/1 × 1/81 × 1/161 × 1/321 × 1/641 × 1/1281 × 1/256
conv_5_53 × 1/163 × 1/163 × 1/323 × 1/643 × 1/1283 × 1/2563 × 1/512
conv_61 × 1/101 × 1/101 × 1/101 × 1/101 × 1/101 × 1/101 × 1/10
The number of parameters23.77 KB19.90 KB56.93 KB142.68 KB557.07 KB2201.10 KB8750.16 KB
The computation of the floating numbers1.77 × 1051.61 × 1052.66 × 1054.41 × 1051.57 × 1065.92 × 1062.29 × 107
NoteUncompressed 4 channelsCompressed 4 channelsCompressed 4 channelsCompressed 4 channelsCompressed 8 channelsCompressed 16 channelsCompressed 32 channels
The maximum number of channels is 16The maximum number of channels is 16The maximum number of channels is 32The maximum number of channels is 64The maximum number of channels is 128The maximum number of channels is 256The maximum number of channels is 512
The WPT-CNN network structures of models 1 and 2 are shown in Figure 5a,b, respectively.
Table 3. Comparison of networks in terms of the parameters and computational load.
Table 3. Comparison of networks in terms of the parameters and computational load.
MethodNumber of ParametersComputation of Floating NumberInput Length Specification
WPT-CNN19.90 KB1.61 × 10564 × 1 × 16
Reference [28]72.35 KB1.40 × 10632 × 32 × 1
Reference [23]203.80 KB6.81 × 107400 × 1 × 1
Reference [24]258.52 KB4.69 × 1061024 × 1 × 1
Reference [27]565.16 KB8.08 × 107512 × 10 × 1
Reference [25]367.814 KB2.00 × 108128 × 128 × 1
Reference [15]2696.54 KB1.37 × 1061024 × 1 × 1
Resnet-5080,849.75 KB2.04 × 1091024 × 1 × 1
Resnet-1814,325.75 KB3.35 × 1081024 × 1 × 1
VGG-1678,544.75 KB9.24 × 1081024 × 1 × 1
1D-LeNet51349 KB1.27 × 1061024 × 1 × 1
Table 4. The hardware and software environment.
Table 4. The hardware and software environment.
Hardware EnvironmentSoftware Environment
CPUIntel Core i7-8700k, 3.7 GHz, six-core twelve threadsUbuntu 16.04 system TensorFlow 1.10 framework and Python programming language
GPUNVIDIA 1080Ti 11G
Memory32 GB
Storage2 TB
Table 5. Classification of the bearing fault datasets.
Table 5. Classification of the bearing fault datasets.
DatasetsLoad (HP)Training SamplesTest SamplesFault TypesFlaw Size (Inches)Models
A/B/C/D0/1/2/3800/800/800/800100/100/100/100normal01
800/800/800/800100/100/100/100ball0.0072
800/800/800/800100/100/100/100ball0.0143
800/800/800/800100/100/100/100ball0.0214
800/800/800/800100/100/100/100inner_race0.0075
800/800/800/800100/100/100/100inner_race0.0146
800/800/800/800100/100/100/100inner_race0.0217
800/800/800/800100/100/100/100outer_race0.0078
800/800/800/800100/100/100/100outer_race0.0149
800/800/800/800100/100/100/100outer_race0.02110
Table 6. Experimental results for the effects of the WPT level on the accuracy and noise resistance of the algorithm.
Table 6. Experimental results for the effects of the WPT level on the accuracy and noise resistance of the algorithm.
Validation SetsTitle 20%10%30%50%70%90%100%
Time domain10010094.4890.4985.7981.6277.8776.88
±0.17±0.57±0.72±1.15±1.48±0.88
WPT310010096.8093.490.7788.5686.6285.74
±0.313±0.33±0.33±0.92±0.71±0.73
WPT410010098.1897.2695.8394.5593.0792.11
±0.25±0.37±0.38±0.47±0.89±0.63
WPT510099.496.6995.7894.3092.6690.4090.16
±0.26±0.33±0.49±0.76±0.69±0.99
WPT610097.296.0294.0391.4989.0186.4385.83
±0.34±0.60±0.84±0.72±1.10±0.72
Table 7. Experimental results for the effects of the size of convolution kernels in the first layer on the noise resistance of the algorithm.
Table 7. Experimental results for the effects of the size of convolution kernels in the first layer on the noise resistance of the algorithm.
Convolution KernelsValidation Sets0%10%30%50%70%90%100%
310010098.1897.2695.8394.5593.0792.11
±0.25±0.37±0.38±0.47±0.89±0.63
510010099.0297.9496.7695.2993.5492.85
±0.19±0.32±0.57±0.39±0.76±0.42
710099.999.7399.1198.4496.9895.3195.03
±0.18±0.34±0.41±0.43±0.54±0.87
910010096.6696.0694.8894.1293.1692.61
±0.22±0.39±0.59±0.51±0.29±0.92
1110010099.3399.0298.6397.7196.5796.35
±0.12±0.26±0.24±0.45±0.43±0.52
1310010099.0698.5797.5796.7095.7495.04
±0.17±0.34±0.37±0.52±0.33±0.45
1510010099.5099.0898.3897.4996.3596.24
±0.17±0.30±0.21±0.52±0.57±0.57
2310010098.1197.2996.1395.1194.3893.89
±0.33±0.27±0.50±0.53±0.45±0.59
3110099.999.0398.2897.3096.4795.5294.70
±0.14±0.29±0.41±0.42±0.55±0.48
Table 8. Experimental results for the accuracy, noise resistance and transferability of various algorithms.
Table 8. Experimental results for the accuracy, noise resistance and transferability of various algorithms.
Model 1Model 2Model 3Model 4Model 5Model 6Model 7
Noise resistance0%100100100100100100100
10%99.7699.3399.7199.4099.9999.97100.0
±0.07±0.12±0.12±0.18±0.03±0.05±0.00
30%99.1999.0298.7899.2899.8199.7199.91
±0.27±0.26±0.25±0.27±0.14±0.14±0.09
50%98.3798.6397.7997.0198.2599.0999.52
±0.45±0.24±0.31±0.60±0.41±0.34±0.18
70%97.5797.7196.5595.8597.2598.6198.87
±0.32±0.45±0.37±0.37±0.47±0.29±0.24
90%96.4196.5795.1894.3696.7597.6498.11
±0.58±0.43±0.53±0.84±0.46±0.40±0.40
100%96.1196.3595.0994.5396.7597.3897.65
±0.44±0.52±0.54±0.42±0.46±0.20±0.32
Transfer-learning abilityAB98.810099.910098.499.299.4
AC99.110010099.999.699.897.3
AD99.399.999.599.998.799.799.2
Number of parameters23.77 KB19.90 KB56.93 KB142.68 KB557.07 KB2201.10 KB8750.16 KB
The computation of floating numbers1.77 × 1051.61 × 1052.66 × 1054.41 × 1051.57 × 1065.92 × 1062.29 × 107
Table 9. Description of network structures with various depths.
Table 9. Description of network structures with various depths.
3 Layers4 Layers5 Layers6 Layers
conv_111 × 1/411 × 1/411 × 1/411 × 1/4
conv_23 × 1/83 × 1/83 × 1/83 × 1/8
conv_3_13 × 1/163 × 1/163 × 1/163 × 1/16
conv_3_21 × 1/81 × 1/81 × 1/81 × 1/8
conv_3_33 × 1/163 × 1/163 × 1/163 × 1/16
conv_4_1 3 × 1/163 × 1/163 × 1/32
conv_4_2 1 × 1/81 × 1/81 × 1/16
conv_4_3 3 × 1/163 × 1/163 × 1/32
conv_5_1 3 × 1/163 × 1/16
conv_5_2 1 × 1/81 × 1/8
conv_5_3 3 × 1/163 × 1/16
conv_5_4 1 × 1/81 × 1/8
conv_5_5 3 × 1/163 × 1/16
conv_6_1 3 × 1/16
conv_6_2 1 × 1/8
conv_6_3 3 × 1/16
conv_6_4 1 × 1/8
conv_6_5 3 × 1/16
conv_61 × 1/101 × 1/101 × 1/101 × 1/10
Table 10. Comparison of the noise resistance and transferability of the network structures with various depths.
Table 10. Comparison of the noise resistance and transferability of the network structures with various depths.
Layers3456
Validation sets100100100100
Normal100100100100
Transfer-learning abilityAB90.999.7010099.6
AC93.599.4010099.8
AD91.598.9099.899.6
Noise resistance10%99.11 ± 0.4099.72 ± 0.1099.33 ± 0.1298.95 ± 0.17
30%97.05 ± 0.4099.32 ± 0.2699.02 ± 0.2698.32 ± 0.32
50%94.33 ± 0.6098.69 ± 0.2598.63 ± 0.2497.29 ± 0.58
70%91.87 ± 0.7497.49 ± 0.3697.71 ± 0.4595.69 ± 0.43
90%89.35 ± 0.9596.61 ± 0.6796.57 ± 0.4394.42 ± 0.51
100%87.94 ± 0.9695.88 ± 0.5996.35 ± 0.5293.97 ± 0.55
Number of parameters7.49 KB12.65 KB19.90 KB27.15 KB
The computation of floating number1.26 × 1051.47 × 1051.61 × 1051.75 × 105
Table 11. Comparison of the transfer-learning ability of the proposed algorithm in terms of noise resistance.
Table 11. Comparison of the transfer-learning ability of the proposed algorithm in terms of noise resistance.
0%10%30%50%70%90%100%
AB10099.54 ± 0.2598.82 ± 0.2497.71 ± 0.4796.96 ± 0.5096.12 ± 0.5795.53 ± 0.51
AC10099.42 ± 0.1498.78 ± 0.2598.21 ± 0.3297.38 ± 0.4596.64 ± 0.4495.94 ± 0.38
AD99.897.21 ± 0.2196.35 ± 0.4395.63 ± 0.4694.64 ± 0.6393.16 ± 0.4793.28 ± 0.88
BA98.997.07 ± 0.3396.16 ± 0.4394.42 ± 0.4893.13 ± 0.5191.79 ± 1.0090.34 ± 0.73
BC10099.47 ± 0.1599.44 ± 0.1599.19 ± 0.1898.54 ± 0.2998.11 ± 0.4297.65 ± 0.42
BD99.998.15 ± 0.2297.78 ± 0.3697.23 ± 0.2096.66 ± 0.3095.64 ± 0.3295.52 ± 0.61
CA93.492.81 ± 0.3191.40 ± 0.3889.99 ± 0.6288.59 ± 0.8286.93 ± 0.8086.31 ± 0.53
CB98.797.27 ± 0.1796.87 ± 0.3496.54 ± 0.4095.86 ± 0.3995.44 ± 0.6294.88 ± 0.27
CD10099.87 ± 0.1399.50 ± 0.1399.01 ± 0.2798.48 ± 0.3697.58 ± 0.5297.23 ± 0.42
DA95.391.84 ± 0.3290.45 ± 0.5489.18 ± 0.5688.41 ± 0.5886.90 ± 0.6086.33 ± 0.76
DB98.497.98 ± 0.1497.59 ± 0.3397.31 ± 0.2296.73 ± 0.4895.99 ± 0.3995.90 ± 0.51
DC10099.72 ± 0.0499.66 ± 0.1598.97 ± 0.1799.05 ± 0.1498.19 ± 0.3297.88 ± 0.29
Table 12. Resistance of various algorithms to various percentages of added white Gaussian noise.
Table 12. Resistance of various algorithms to various percentages of added white Gaussian noise.
Validation Sets0%10%30%50%70%90%100%
WPT-CNN10010099.3399.0298.6397.7196.5796.35
±0.12±0.26±0.24±0.45±0.43±0.52
Reference [28]10010099.8598.3395.6091.9588.0285.94
±0.08±0.27±0.44±0.52±0.55±1.13
Reference [24]10099.7099.5398.2494.5190.7286.6985.35
±0.18±0.26±0.61±0.51±0.54±0.45
Reference [23]10010080.1371.6164.5059.8857.0155.83
±0.19±0.27±0.44±0.37±0.40±0.37
Reference [26]10010088.1066.2855.2448.3843.8942.02
±0.87±0.58±1.20±0.90±1.03±0.98
Reference [27]10099.9097.4685.1470.5958.5750.9648.38
±0.52±0.80±0.75±1.03±0.47±0.73
Reference [19]10099.8099.1598.0196.3694.0391.7990.74
±0.20±0.50±0.31±0.69±0.86±0.64
Reference [15]99.197.5095.5493.8892.8690.8188.9087.49
±0.25±0.50±0.38±0.66±1.03±0.92
Reference [11]93.688.3088.0885.3876.6664.0346.8138.47
±0.330.77±0.88±0.75±1.37±0.95
Table 13. Comparison of the transfer-learning ability of various algorithms.
Table 13. Comparison of the transfer-learning ability of various algorithms.
ABACADBABCBDCACBCDDADBDCAVG
WPT-CNN10010099.898.910099.993.498.710095.398.410098.70
Reference [28]98.499.998.299.810098.996.098.399.390.493.810097.75
Reference [24]99.599.996.999.510098.296.898.499.097.197.199.198.46
Reference [23]99.998.989.899.999.997.998.699.499.881.385.094.295.40
Reference [26]89.384.573.083.698.091.585.196.995.978.088.095.388.26
Reference [27]98.798.381.798.799.998.593.297.797.489.492.094.995.03
Reference [19]71.174.872.687.687.476.677.588.079.577.778.986.779.87
Reference [15]33.744.643.643.342.051.548.447.748.342.846.341.844.50
Reference [11]40.440.339.340.555.359.841.256.357.136.956.152.647.98

Share and Cite

MDPI and ACS Style

Ma, S.; Cai, W.; Liu, W.; Shang, Z.; Liu, G. A Lighted Deep Convolutional Neural Network Based Fault Diagnosis of Rotating Machinery. Sensors 2019, 19, 2381. https://doi.org/10.3390/s19102381

AMA Style

Ma S, Cai W, Liu W, Shang Z, Liu G. A Lighted Deep Convolutional Neural Network Based Fault Diagnosis of Rotating Machinery. Sensors. 2019; 19(10):2381. https://doi.org/10.3390/s19102381

Chicago/Turabian Style

Ma, Shangjun, Wei Cai, Wenkai Liu, Zhaowei Shang, and Geng Liu. 2019. "A Lighted Deep Convolutional Neural Network Based Fault Diagnosis of Rotating Machinery" Sensors 19, no. 10: 2381. https://doi.org/10.3390/s19102381

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop