Next Article in Journal
Visual SLAM for Unmanned Aerial Vehicles: Localization and Perception
Previous Article in Journal
Transforming Poultry Farming: A Pyramid Vision Transformer Approach for Accurate Chicken Counting in Smart Farm Environments
Previous Article in Special Issue
LSTM-Autoencoder Based Anomaly Detection Using Vibration Data of Wind Turbines
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised Feature-Construction-Based Motor Fault Diagnosis

by
Tsatsral Amarbayasgalan
1 and
Keun Ho Ryu
2,*
1
Electronics and Telecommunications Research Institute, Daejeon 34129, Republic of Korea
2
Data Science Laboratory, Faculty of Information Technology, Ton Duc Thang University, Ho Chi Minh City 700000, Vietnam
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(10), 2978; https://doi.org/10.3390/s24102978
Submission received: 31 March 2024 / Revised: 5 May 2024 / Accepted: 6 May 2024 / Published: 8 May 2024
(This article belongs to the Special Issue Deep-Learning-Based Defect Detection for Smart Manufacturing)

Abstract

:
Any bearing faults are a leading cause of motor damage and bring economic losses. Fast and accurate identification of bearing faults is valuable for preventing damaging the whole equipment and continuously running industrial processes without interruption. Vibration signals from a running motor can be utilized to diagnose a bearing health condition. This study proposes a detection method for bearing faults based on two types of neural networks from motor vibration data. The proposed method uses an autoencoder neural network for constructing a new motor vibration feature and a feed-forward neural network for the final detection. The constructed signal feature enhances the prediction performance by focusing more on a fault type that is difficult to detect. We conducted experiments on the CWRU bearing datasets. The experimental study shows that the proposed method improves the performance of the feed-forward neural network and outperforms the other machine learning algorithms.

1. Introduction

Motor failures are often associated with bearing damage. This can result in high levels of vibration that can disrupt production processes, cause motor malfunctions, and lead to economic losses. Therefore, it is important to detect bearing faults early by monitoring motor vibration signals to prevent more damage to the equipment and ensure uninterrupted industrial processes [1]. In general, running motor vibration signals are collected very quickly, and data-driven learning techniques have been increasingly used to diagnose motor health conditions from these signals.
The methods for detecting bearing faults can be classified as statistical-feature-based detectors and raw-signal-based detectors. Statistical-feature-based detectors first extract statistical features from time-domain signals or frequency-domain signals, such as the maximum value of amplitude, mean value of amplitude, and Kurtosis factor. These features are then used to detect the fault type [1,2,3,4,5,6,7]. However, the efficiency of statistical features may vary based on datasets and detection models, and it requires manual feature extraction for each specific case. In contrast, deep-learning-based methods do not require feature extraction because of their enormous learnable parameters. These methods learn useful features automatically without an additional feature pre-processing stage [8,9,10,11]. Deep-learning-based detectors significantly increase detection performance from raw signals, making them more suitable for bearing fault detection.
This study proposes an unsupervised neural-network-based signal extraction for improving bearing fault detection using autoencoder (AE) neural network (AENN) and feed-forward neural network (FFNN) models. However, the process of signal extraction is distinct from statistical-based methods and involves the use of self-supervised learning. The proposed method differs from existing motor fault detection methods. It improves the detection performance by focusing more on the fault type that is misclassified the most during the training process. Specifically, the performance of the FFNN model is enhanced through the extracted signal utilizing an unsupervised AE model. The AE model learns from vibration signals belonging to a fault type that was detected with more error than the other fault types. The most misclassified fault type for the AE model is chosen using the FFNN model on the original raw signals. In general, the structure of an AE model has the same number of neurons in the input and output layers, and its output is a reconstructed version of the input. The AE model learns by minimizing the variance between input and output values. In this study, we use the difference between input and output values on the AE model, known as reconstruction error (RE), as a new feature to reduce the detection error of the FFNN model. When the inputs are signals other than those used to construct the AE model, they are reconstructed with a high difference. This is because the AE model shows a high gap between the input and output when the input is unlearned types of signals. The main contributions of this study are as follows:
  • The proposed method improves the detection performance of the FFNN model by signal extraction, which better distinguishes a fault type that is hard to detect;
  • We evaluated the proposed method on open datasets collected with different horsepower. The proposed method gave a higher performance than the compared methods.
  • The rest of this paper is organized as follows. Section 2 provides an overview of existing studies for motor bearing fault detection. The proposed method is detailed in Section 3. Section 4 demonstrates the experimental study, including experimental design, dataset, evaluation metrics, and detection performances of the compared methods. In Section 5, we conclude this study.

2. Related Work

It is beneficial to detect motor failures automatically and accurately to prevent future failures and financial losses due to sudden breakdowns and interruptions. As mentioned above, the utilization of machine learning algorithms to detect motor faults using signal data from motor vibration sensors is increasing. These methods typically build the motor fault detection model based on feature extraction and feature selection steps. Chuan et al. proposed a deep random forest fusion (DRFF) technique to diagnose gearbox faults using acoustic and vibratory signals [1]. First, they transformed signals from the gearbox by a wavelet packet transformer (WPT) and extracted features using two deep Boltzmann machines (DBMs). Finally, the random forest (RF) classifier fused the outputs of the two DBMs. The authors of [2] offered support vector machines (SVMs) and K-nearest neighbors (KNNs), and the bagged tree-based classifiers provided nearly 100 percent accuracy for motor fault diagnosis from both stator currents and vibration signals of motors. They extracted eight statistical features for the detection model: mean, median, standard deviation, and other criteria. These features were calculated from the results of two signal processing techniques: the matching pursuit (MP) and discrete wavelet transform (DWT). In [4], there were eight wavelet features extracted using a three-layer WPT transformer on the raw signal dataset to be used as the input to the classification model. Then, ensemble learning algorithms, such as adaptive boosting (AdaBoost) and RF, were suggested to diagnose crack faults in the presence of noise and small data. Another IoT-based ensemble algorithm was proposed in [5] to monitor the status of the induction motor from the motor vibration signal. Sundaram Buchaiah et al. [11] selected important features from 72 statistical features using the RF algorithm for bearing data. Then, they reduced the dimension of input features into two by dimension reduction techniques. Finally, the Bhattacharyya distance and SVM algorithms were used to verify fault diagnosis accuracy.
Recently, deep-learning-based methods have been used broadly for fault detection. The advantage of deep learning is that it does not require manual feature extraction. In [12], a deep neural network was proposed to detect multi-faults from raw sensor data without feature selection and signal processing. The authors of [9] used signals from multiple current sensors instead of vibration sensors. The current signal is accessible by low-cost sensors and not easily affected by interfering noise from the surrounding components. They diagnosed seven types of gearbox conditions, which were diagnosed by a two-dimensional convolutional neural network (CNN). Jong-Hyun Lee et al. [10] proposed the CNN model to develop a motor fault diagnosing system. It detects whether a motor condition is normal or faulty on the rotor and bearing from the vibration signals without signal preprocessing. Many studies have used the AE model for bearing fault detection [8,13,14,15,16]. The authors of [8] proposed a stacked denoising autoencoder (SDAE)-based fault detector from motor vibration signals. First, the Fourier transform produced frequency-domain signals for use as input to the SDAE. Then, several unsupervised AE models were used to extract features. The input layer of the next AE used the encoder layer of the above-level AE, and the final AE’s encoder layer was used for classification with the Softmax function. In the stacked denoising autoencoder (SDA) was investigated for the fault diagnosis of rotary machinery components when signals have ambient noise and working condition fluctuations. They stacked three AE models to obtain high-level feature representations to improve classification robustness. The authors of proposed an ensemble of deep autoencoders (EDAE) for feature learning from the vibration signals and fault diagnosis. EDAEs were constructed using fifteen AE models with various activation functions. The features from the last AE model were fed into the Softmax classifier for fault recognition.
Many machine-learning-based studies extracted fault detection features manually, based on their domain knowledge, or using some statistical approaches, such as kurtosis and skewness. In contrast, deep-learning-based methods do not require feature extraction because of their enormous learnable parameters. The proposed method in this study differs from existing studies by extracting an additional feature using the advantage of self-supervised deep neural networks. It focuses on the most misclassified fault type using the extracted signal feature to improve the fault detection model. We extract a new signal feature by reconstruction error of the input signal on the AE model that learns from only the fault type, which is the most difficult type to detect. Therefore, the value of the reconstruction error of the input signal belonging to the fault type used in the training of the AE model will be lower than the reconstruction errors of other types of signals. This characteristic of the AE trained from the single type of fault enhances distinguishing faults from the vibration signal data.

3. Proposed Method for Bearing Fault Detection

This section describes the training and testing processes of the proposed method to predict bearing faults as shown in Figure 1. The solid lines indicate the training steps that build the detection model for the bearing faults, and the dashed lines represent the testing process of fault detection on unseen signals.
The proposed method addresses the most challenging fault type for reducing detection error. Therefore, we first train an initial FFNN model from the original training signals (step 1). This initial FFNN model is used to determine the most challenging fault type from the evaluation of the validation signals (steps 2 and 3). Then, signals belonging to the selected (determined as challenging) fault type are distinguished from the whole training signals in step 4. Next, the AENN model learns from the selected type of faulty signals for feature extraction in step 5. In step 6, to generate a new feature based on the initial training signals marked as “A” in Figure 1, we feed the signals to the AENN model and obtain reconstruction errors from the AENN model. To create the final training dataset, we combine the initial training signals (marked as “A”) with the received reconstruction errors in step 7. The initial training dataset consists of input signals with 120 vibration points. After step 7, the length of the input signal becomes 121. In the last step of the training process, the second FFNN model is trained from the final prepared training signals (marked as “B”) for further detection.
The testing process (detect faults from the unseen signals) is demonstrated in steps 9–12. Before detecting bearing fault by the final FFNN model, the reconstruction error of the input signal is obtained from the AENN model (step 9). In steps 10–12, the final FFNN model detects the bearing health condition from the combination of unseen input signal and its reconstruction error.
In this study, we utilize an AE neural network model to generate a new feature from the input signal, which will enhance the detection model based on the FFNN with three hidden layers. Our proposed method involves simpler model architectures than CNNs, which are usually more effective in dealing with high-dimensional data, such as images. Therefore, we have employed the reconstruction error of the input signals on the AENN model with a single hidden layer to improve the feature representation of the FFNN.
Figure 2 shows an example of how the AE model is constructed in the proposed method. First, we separate the particular type of signals from the whole signals based on the detection results from the baseline FFNN model (original signal-based FFNN model) shown in Figure 1. For instance, the AE model trained on the selected Fault-1-type signals in Figure 2. The bottleneck hidden layer of the AE model in the proposed method has 60 neurons and transformed the received values from the preceding layer by the rectified linear unit (ReLU) activation function. To better distinguish Fault-1 from other bearing faults, a new signal feature is extracted by giving a training signal to the prepared AE model and calculating the RE. As a result, the final training signals are made by combining the original signals with the reconstruction error of the input signal.
The AE is a symmetric neural network where the numbers of neurons in the input and output layers are the same. It learns the data pattern by reducing the input dimension and then reconstructing the input from the reduced dimensional space. Due to its structure, it has been used in dimension reduction, data denoising, and synthetic data generation. For example, a bottleneck layer of the AE model that is a compressed representation of its input is used in dimension reduction. In the training process of the AE, if inputs are data with noise and outputs are the original data, the AE can be applied for data denoising. In this paper, we use RE, which is a variance between the input and output layers of the AENN model, to emphasize a particular type of motor fault. RE is defined as follows (1):
R E = j = 1 k ( x j x ^ j ) 2 k
where k is the number of neurons in the input and output layers of the AE, x j is the j-th neuron in the input layer, and x ^ j is the corresponding neuron with the reconstructed value of x j .
The proposed method predicts motor bearing faults by the FFNN model. The neural network was first introduced in 1943 [17] and has been successfully applied in various domains, such as image processing [18], natural language processing [19], and predicting motor faults [20]. The FFNN is constructed by fully connected (dense) layers that are simpler in architecture to compare convolutional neural networks (CNNs). The fully connected layers consist of neurons, and every neuron connects to all neurons in the descendant layer. CNNs are typically more effective when dealing with high-dimensional data like images in which local patterns and spatial relationships are important. Fully connected layers process the entire input signal as a single vector, allowing them to capture the relevant information that spreads across the entire signal rather than being localized in specific regions. In particular, an underlying pattern of the CWRU dataset was effectively captured with an accuracy exceeding 90% by a simple neural network with fully connected dense layers. For this study, we propose a simple fully connected feed-forward neural network with three hidden layers.
The proposed FFNN has three hidden layers, as shown in Figure 3. Neurons in each hidden layer transform their received value from the preceding layer by the ReLU activation function. The output layer uses the Softmax activation function to predict motor conditions; it returns the probability of each motor condition as a value from 0 to 1, and the sum of all probabilities is equivalent to 1. From the result outputs, the high-probability bearing fault is chosen as the final detection result.
The FFNN model is trained on the prepared training signals by combining the original and extracted signals. It can improve the detection performance by using the extracted signal based on the most misidentified fault type.

4. Experimental Study

4.1. Dataset

The Case Western Reserve University (CWRU) bearing dataset [21] is a popular benchmark dataset in machinery fault diagnosis. This dataset is relatively large, containing vibration signals gathered from bearings under various fault conditions, including inner race, outer race, and ball faults, as well as healthy conditions. There are several reasons to apply data-driven learning methods, especially neural networks, to this dataset for monitoring motor health conditions. Neural networks handle large datasets and process the entire dataset without manual operation. Moreover, neural networks are robust to noise from variations in operating conditions and other sources of variability commonly encountered in real-world industrial environments. The fault conditions represented in the CWRU dataset are common in industrial machinery, such as rotating equipment with rolling element bearings. By developing machine learning models on this dataset, researchers aim to improve the reliability and efficiency of machinery condition monitoring systems in real-world applications. We used the CWRU bearing dataset to evaluate fault detection methods.
The vibration signals were recorded from healthy bearings and three types of faulty bearings, including the inner raceway, rolling element (ball), and outer raceway, each with a failure diameter of 0.021 inches, with motor speeds from 0 to 3 horsepower. We utilized drive-end (DE) bearing vibration signals in the experimental study. There were four datasets prepared for the experimental study. Each dataset consists of four files for normal bearing conditions and three kinds of faulty bearing conditions, as shown in Figure 4. Table 1 shows the data files used in the experimental study. Figure 5 demonstrates an example of the vibration signals of bearing health conditions.

4.2. Compared Detection Methods

The proposed detection method for bearing faults has been compared with machine learning algorithms, including KNN, SVM, AdaBoost, decision tree (DT), naïve Bayes (NB), and RF. We implemented the compared detection models in Python using the Scikit-learn package [22] for machine learning classifiers and the Keras library [23] for deep neural networks. The parameter configurations of the compared algorithms are shown in Table 2. We trained several models from each algorithm using varying configurations. For instance, we trained KNN-based models with different numbers of neighbors ranging between 3 and 25. To compare the KNN-based detection model with others, we selected the best-performing model on test datasets from these KNN models. Table 3 shows the configurations of the models in the proposed method.
The proposed method employs AE and FFNN models to detect bearing faults. We generate an additional feature from the input signal via the AE model to enhance the detection performance of the FFNN model. This additional feature is generated by the reconstruction error of the input signal on the AE model. We train AE models with different structures on signals of all types of bearing conditions to select the AE model with the lowest training error. Figure 6 represents the average mean squared error of the trained AE models on datasets 1–4 shown in Table 1. The AE-3 model with a single hidden layer of 60 nodes gave the smallest average error for all bearing health conditions. Therefore, the proposed method used AE-3 for extracting the RE-based feature. The structure of the AE-3 model has an input layer with 120 nodes, a hidden layer (latent space) with 60 nodes, and an output layer with 120 nodes.
For the FFNN model used in the proposed method, we trained various FFNN models to choose the appropriate hyperparameter configurations. Figure 7 shows the average validation accuracy of the baseline FFNN and the proposed FFNN models on datasets 1–4. We used 90% of the dataset for training, and the remaining 10% was used for validation. The learning rate of 0.001 and batch size of 8 with the ReLU activation function showed higher accuracy than the other configurations.

4.3. Experimental Results

First, the baseline FFNN is learned from the original training signals. We found the most misclassified type of bearing fault from the training dataset using the baseline FFNN model to train the AE model on the most challenging fault type. Then, the proposed FFNN model was trained on the original and AE-based extracted signals. Table 4 shows the confusion matrix of baseline FFNN models on the validation set, which is 10% of the training signals. In datasets 1–3, the outer race fault was the most misclassified. However, the ball fault in Dataset-4 was more incorrectly detected than other bearing faults.
We compared the proposed detection model and baseline FFNN model to demonstrate how the proposed method improved the detection performance of the baseline FFNN model. We trained four baseline models using the original input signals of datasets 1–4 separately. Then, the proposed FFNN models were trained by the original and extracted input signals. For signal extraction for the proposed FFNN, we selected AE models learned from the most misclassified motor condition signals on each dataset based on the confusion matrix shown in Table 4. Each model was tested on three datasets, and the final performance was averaged in Table 5. We can see that the proposed AE-based FFNN models outperform the baseline FFNN models based on the initial signals without the extracted feature, and their average evaluation metrics were higher than the baseline FFNN models. Moreover, the recall and f-measure measurements of baseline FFNN models were improved by using the extracted feature from the AE model, and its average values increased by more than 1% by the proposed method.
Finally, we compared six machine learning classification models with the proposed method, including KNN, RF, AdaBoost, SVM, NB, and DT. We experimented with four different datasets gathered by different shaft speeds. Four models were trained for each algorithm using datasets listed in Table 1 for performance evaluation. Then, each model was tested on three untrained datasets. For instance, a model trained on Dataset-1 with a shaft speed of 1797 rpm was tested on Dataset-2, Dataset-3, and Dataset-4 with shaft speeds of 1772, 1750, and 1730 rpm, respectively. Table 6 shows the average performances, and detailed performances are represented in Appendix A. The proposed method showed more stable results than the compared methods on the datasets collected at different shaft speeds, as shown in Figure 8.
Table A1, Table A2, Table A3 and Table A4 show the comparison of the proposed AE-FFNN model and other machine-learning-based predictive models. We configured the input parameters of the particular machine learning models differently based on the dataset. The best values for the input parameters were selected based on the training performance. As a result of these comparisons, the proposed detection method performed better than the compared individual predictive methods.
Table A1 shows the testing results on datasets 2–4. All models learned from Dataset-1. The RF model showed higher results than the machine-learning-based models on each testing dataset. However, the proposed method outperformed the RF by increasing the average accuracy, precision, recall, f-measure, and AUC values on datasets 2–4 by 3.184%, 4.154%, 5.590%, 4.901%, and 0.509%, respectively.
For machine-learning-based compared detection models learned from Dataset-2, the RF model performed better than others on all testing datasets, including Dataset-1, Dataset-3, and Dataset-4. However, the proposed method outperformed its accuracy by 1.762%, precision by 3.023%, recall by 3.382%, f-measure by 4.024%, and AUC by 2.815% on average, as shown in Table A2.
Table A3 shows the results of the compared models trained on Dataset-3, and evaluated on datasets 1, 2, and 4. As a result of the compared models, except for the proposed model, RF and SVM models showed comparable higher results than others. However, the proposed method was superior to these models.
In Table A4, we showed the results on datasets 1–3 for detection models learned from Dataset-4. We can see that the proposed method achieves the best performance for predicting bearing faults. Its average accuracy, precision, recall, f-measure, and AUC on test datasets reached 98.18%, 96.651%, 96.007%, 95.547%, and 96.699%, respectively. These are higher than the RF model by 4.184%, 5.64%, 5.947%, and 3.766%.
Figure 8 represents distributions of detection performances across all experimented datasets of the compared models learned from datasets 1–4 by the box plot diagram. According to all evaluation measurements, RF and SVM models showed comparable results with the proposed method by giving performances that were higher than 80%. We can see that the performance of the SVM model was less spread out than the RF model, and its value was relatively lower. However, the proposed method performed better than other models according to all measurements by showing the shortest performance distribution with the highest value.

5. Conclusions

In this study, we proposed a method for bearing fault detection from motor vibration signals based on two different types of neural networks, such as the AENN for feature extraction and the FFNN for detection. The AENN is a type of neural network that reconstructs a given input into its output as similarly as possible. Mainly, it is used for data denoising by reconstructing noisy data or for generating synthetic data from its learned distribution. In the proposed method, we used AENN’s input–output difference to extract the input feature for the bearing fault detection model. By training the AENN model on only the most complicated signal types instead of all signal types, we were able to use its RE to make a feature to help distinguish fault types. In other words, the AENN model reconstructs a given input signal to the output with less loss when the input is of a learned signal type rather than its unlearned signal type.
The AENN model is learned from signals of a single type of fault to emphasize that type of fault over others. However, the detection models were designed with four outputs for ball fault, inner race fault, outer race fault, and normal condition of motor bearings. The limitation of the proposed model is that it can improve the fault detection FFNN model based on the additional generated feature that can emphasize only one fault type.
We evaluated the proposed method on the CRWU bearing open dataset and compared it with six machine-learning-based models. The presented method successfully enhanced the detection performances of the FFNN by focusing on the most challenging bearing fault type to detect using the AENN model. It increased average values of the recall and f-measure of the baseline FFNN by more than 1% on the experimented datasets. Moreover, the proposed method outperformed the average accuracy, precision, recall, f-measure, and AUC of the KNN, RF, AdaBoost, SVM, DT, and NB models in Table A1, Table A2, Table A3 and Table A4 by (10.430, 10.474, 14.298, 15.454, 10.919), (2.688, 3.632, 4.606, 4.771, 2.544), (18.013, 25.124, 29.490, 29.479, 21.619), (4.105, 5.474, 7.356, 7.971, 4.236), (16.480, 21.790, 25.573, 25.679, 21.561), and (15.171, 24.724, 28.381, 28.413, 18.236), respectively.

Author Contributions

Conceptualization, T.A. and K.H.R.; methodology, T.A.; implementation, T.A.; validation, T.A. and K.H.R.; formal analysis, T.A.; writing—original draft preparation, T.A.; supervision, K.H.R.; funding acquisition, K.H.R. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) by the Ministry of Science, Information and Communication Technology (ICT) and Future Planning in Korea under Grant No. 2020R1A2B5B02001717 in Korea.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Table A1, Table A2, Table A3 and Table A4 show the comparison of the proposed AE-FFNN model and other machine-learning-based predictive models.
Table A1. Comparison of the experimented fault detection models learned from Dataset-1.
Table A1. Comparison of the experimented fault detection models learned from Dataset-1.
Test
Dataset
PerformanceKNNRFAdaBoostSVMDTNBProposed/AENN Learned from the Outer Race Fault/
Dataset-2Accuracy87.30093.60077.50091.30081.80081.40097.739
Precision86.10090.80064.70090.70071.90070.50096.128
Recall77.80088.90062.20084.80070.50068.10096.168
F-measure80.20089.80063.30087.20071.10067.70096.133
AUC91.20099.20083.60099.10082.20093.40099.776
Dataset-3Accuracy85.10094.00079.80090.40081.70079.90097.518
Precision84.00092.00069.30090.10072.30067.20095.915
Recall73.90089.60065.70083.20070.40065.00095.724
F-measure77.30090.70067.30086.10071.10065.20095.794
AUC88.80099.30083.60099.60082.00094.20099.804
Dataset-4Accuracy83.90095.30080.30092.30080.20084.30097.196
Precision82.20092.40068.40092.00069.70073.30095.620
Recall72.00091.80066.60086.60068.90072.70095.178
F-measure75.70092.00067.40088.90068.90071.40095.276
AUC87.50099.30085.30099.60081.10095.00099.748
Table A2. Comparison of the experimented fault detection models learned from Dataset-2.
Table A2. Comparison of the experimented fault detection models learned from Dataset-2.
Test
Dataset
PerformanceKNNRFAdaBoostSVMDTNBProposed/AENN Learned from the Outer Race Fault/
Dataset-1Accuracy87.50094.30072.00090.40076.90069.10097.699
Precision88.20093.40066.70090.40072.50063.80097.130
Recall84.40092.90067.10088.00071.60064.80097.127
F-measure85.10093.10066.60088.70071.90063.40097.125
AUC94.60099.40086.00099.50082.00090.80099.707
Dataset-3Accuracy89.60095.80080.30092.80084.10080.60098.590
Precision88.40094.30070.70090.90075.10067.10097.573
Recall81.80092.70067.10087.50073.90066.60097.539
F-measure83.60093.40068.60089.00074.40066.20097.527
AUC93.50099.60085.00099.80084.20094.50099.904
Dataset-4Accuracy87.30096.80080.80095.20080.80084.80098.563
Precision85.00094.80068.90093.10070.20072.40097.683
Recall77.90094.40069.20091.50069.00073.80097.546
F-measure80.30094.60068.60092.20069.50071.90097.607
AUC91.60099.60088.00099.70081.30095.30099.894
Table A3. Comparison of the experimented fault detection models learned from Dataset-3.
Table A3. Comparison of the experimented fault detection models learned from Dataset-3.
Test
Dataset
PerformanceKNNRFAdaBoostSVMDTNBProposed/AENN Learned from the Outer Race Fault/
Dataset-1Accuracy86.50093.30071.70089.40074.30065.40096.716
Precision86.20092.10067.40089.50069.50062.50096.059
Recall83.10091.60067.30086.70069.10063.40095.910
F-measure83.90091.80067.30087.40069.30061.20095.961
AUC93.30099.00086.90098.10080.40090.50099.347
Dataset-2Accuracy90.20094.50077.00094.00082.50079.90097.527
Precision87.40091.40064.40092.00072.10068.60095.875
Recall82.90090.40063.90089.60071.20070.60095.707
F-measure84.20090.90064.00090.70071.60067.80095.733
AUC94.00099.20087.40098.20082.80093.40099.430
Dataset-4Accuracy90.10096.50079.40096.90081.80083.80098.224
Precision87.70093.90068.00095.10071.40071.70096.953
Recall82.80093.90067.60094.50070.80074.50096.901
F-measure83.90093.90067.50094.70071.10071.50096.914
AUC94.20099.50089.70099.50082.40095.40099.764
Table A4. Comparison of the experimented fault detection models learned from Dataset-4.
Table A4. Comparison of the experimented fault detection models learned from Dataset-4.
Test
Dataset
PerformanceKNNRFAdaBoostSVMDTNBProposed/AENN Learned from the Ball Fault/
Dataset-1Accuracy83.40091.10071.50086.10073.30067.60095.398
Precision84.90090.30067.00087.10068.50063.00095.064
Recall79.30088.90066.70082.60067.90064.20094.247
F-measure81.20089.50066.80083.40068.10062.40094.539
AUC91.10098.80085.40098.20079.60090.60099.455
Dataset-2Accuracy87.60093.60078.10092.30081.50080.70097.358
Precision86.30091.00066.10090.30070.60068.30096.132
Recall78.40088.80064.80086.50068.90069.20095.393
F-measure81.60089.90065.40088.20069.70067.70095.677
AUC91.10099.10087.40098.50081.30093.50099.679
Dataset-3Accuracy87.90094.10081.00092.60082.70081.30098.181
Precision87.30092.70071.40091.10073.20067.70097.199
Recall78.80089.80068.70087.10071.40068.20096.826
F-measure81.90091.10069.90089.00072.20067.20097.002
AUC91.50099.40087.10099.70082.70094.60099.881

References

  1. Li, C.; Sánchez, R.V.; Zurita, G.; Cerrada, M.; Cabrera, D.; Vásquez, R.E. Gearbox fault diagnosis based on deep random forest fusion of acoustic and vibratory signals. Mech. Syst. Signal Process. 2016, 76, 283–293. [Google Scholar] [CrossRef]
  2. Ali, M.Z.; Shabbir, N.S.K.; Liang, X.; Zhang, Y.; Hu, T. Machine learning-based fault diagnosis for single- and multi-faults in induction motors using measured stator currents and vibration signals. IEEE Trans. Ind. Appl. 2019, 55, 2378–2391. [Google Scholar] [CrossRef]
  3. Sikder, N.; Bhakta, K.; Al Nahid, A.; Islam, M.M.M. Fault diagnosis of motor bearing using ensemble learning algorithm with FFT-based preprocessing. In Proceedings of the International Conference on Robotics, Electrical and Signal Processing Techniques, Dhaka, Bangladesh, 10 January 2019. [Google Scholar] [CrossRef]
  4. Zhong, X.; Ban, H. Crack fault diagnosis of rotating machine in nuclear power plant based on ensemble learning. Ann. Nucl. Energy 2022, 168, 108909. [Google Scholar] [CrossRef]
  5. Tran, M.-Q.; Elsisi, M.; Mahmoud, K.; Liu, M.-K.; Lehtonen, M.; Darwish, M.M.F. Experimental setup for online fault diagnosis of induction machines via promising IoT and machine learning: Towards industry 4.0 empowerment. IEEE Access 2021, 9, 115429–115441. [Google Scholar] [CrossRef]
  6. Long, Z.; Zhang, X.; Zhang, L.; Qin, G.; Huang, S.; Song, D.; Shao, H.; Wu, G. Motor fault diagnosis using attention mechanism and improved adaboost driven by multi-sensor information. Measurement 2021, 170, 108718. [Google Scholar] [CrossRef]
  7. Li, B.; Chow, M.-Y.; Tipsuwan, Y.; Hung, J. Neural-network-based motor rolling bearing fault diagnosis. IEEE Trans. Ind. Electron. 2000, 47, 1060–1069. [Google Scholar] [CrossRef]
  8. Zhao, X.; Wu, J.; Zhang, Y.; Shi, Y.; Wang, L. Fault diagnosis of motor in frequency domain signal by stacked de-noising auto-encoder. Comput. Mater. Contin. 2018, 57, 223–242. [Google Scholar] [CrossRef]
  9. Azamfar, M.; Singh, J.; Bravo-Imaz, I.; Lee, J. Multisensor data fusion for gearbox fault diagnosis using 2-D convolutional neural network and motor current signature analysis. Mech. Syst. Signal Process. 2020, 144, 106861. [Google Scholar] [CrossRef]
  10. Lee, J.-H.; Pack, J.-H.; Lee, I.-S. Fault diagnosis of induction motor using convolutional neural network. Appl. Sci. 2019, 9, 2950. [Google Scholar] [CrossRef]
  11. Buchaiah, S.; Shakya, P. Bearing fault diagnosis and prognosis using data fusion based feature extraction and feature selection. Measurement 2022, 188, 110506. [Google Scholar] [CrossRef]
  12. Zhang, R.; Peng, Z.; Wu, L.; Yao, B.; Guan, Y. Fault diagnosis from raw sensor data using deep neural networks considering temporal coherence. Sensors 2017, 17, 549. [Google Scholar] [CrossRef] [PubMed]
  13. Lu, C.; Wang, Z.-Y.; Qin, W.-L.; Ma, J. Fault diagnosis of rotary machinery components using a stacked denoising autoencoder-based health state identification. Signal Process. 2017, 130, 377–388. [Google Scholar] [CrossRef]
  14. Guo, X.; Shen, C.; Chen, L. Deep fault recognizer: An integrated model to denoise and extract features for fault diagnosis in rotating machinery. Appl. Sci. 2016, 7, 41. [Google Scholar] [CrossRef]
  15. Li, C.; Zhang, W.; Peng, G.; Liu, S. Bearing fault diagnosis using fully-connected winner-take-all autoencoder. IEEE Access 2017, 6, 6103–6115. [Google Scholar] [CrossRef]
  16. Shao, H.; Jiang, H.; Lin, Y.; Li, X. A novel method for intelligent fault diagnosis of rolling bearings using ensemble deep auto-encoders. Mech. Syst. Signal Process. 2018, 102, 278–297. [Google Scholar] [CrossRef]
  17. Mcculloch, W.S.; Pitts, W.H. A logical calculus of the ideas immanent in nervous activity. Bull. Math. Biophys. 1943, 5, 115–133. [Google Scholar] [CrossRef]
  18. Touvron, H.; Bojanowski, P.; Caron, M.; Cord, M.; El-Nouby, A.; Grave, E.; Izacard, G.; Joulin, A.; Synnaeve, G.; Verbeek, J.; et al. Resmlp: Feedforward networks for image classification with data-efficient training. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5314–5321. [Google Scholar] [CrossRef]
  19. Zhou, L.; Hripcsak, G. Temporal reasoning with medical data—A review with emphasis on medical natural language processing. J. Biomed. Inform. 2007, 40, 183–202. [Google Scholar] [CrossRef]
  20. Sampaio, G.S.; Filho, A.R.d.A.V.; da Silva, L.S.; da Silva, L.A. Prediction of motor failure time using an artificial neural network. Sensors 2019, 19, 4342. [Google Scholar] [CrossRef]
  21. Case Western Reserve University Bearing Data Center Website. Available online: https://engineering.case.edu/bearingdatacenter/download-data-file (accessed on 2 October 2023).
  22. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  23. Keras. Available online: https://github.com/fchollet/keras (accessed on 1 July 2023).
  24. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  25. Prechelt, L. Early stopping-but when? Neural Netw. Tricks Trade 1998, 1254, 55–69. [Google Scholar] [CrossRef]
Figure 1. General architecture of the proposed detection method for bearing faults; FFNN: feed-forward neural network, AENN: autoencoder neural network.
Figure 1. General architecture of the proposed detection method for bearing faults; FFNN: feed-forward neural network, AENN: autoencoder neural network.
Sensors 24 02978 g001
Figure 2. Example of the AE model preparation in the proposed method.
Figure 2. Example of the AE model preparation in the proposed method.
Sensors 24 02978 g002
Figure 3. Structure of the FFNN model used in the proposed method; n is the number of input neurons, H1–H3 are the hidden layers; y is the number of output neurons.
Figure 3. Structure of the FFNN model used in the proposed method; n is the number of input neurons, H1–H3 are the hidden layers; y is the number of output neurons.
Sensors 24 02978 g003
Figure 4. Experimental datasets; DE: drive end.
Figure 4. Experimental datasets; DE: drive end.
Sensors 24 02978 g004
Figure 5. Example of the first 500 vibration signals for each bearing condition. (a) File: 98 (normal); (b) File: 210 (inner race fault); (c) File: 223: (ball fault); (d) File: 235 (outer race centered fault).
Figure 5. Example of the first 500 vibration signals for each bearing condition. (a) File: 98 (normal); (b) File: 210 (inner race fault); (c) File: 223: (ball fault); (d) File: 235 (outer race centered fault).
Sensors 24 02978 g005aSensors 24 02978 g005b
Figure 6. Average training errors with standard deviations of AE models on datasets 1–4. (a) AE models learned from the normal signals; (b) AE models learned from the inner race fault signals; (c) AE models learned from the ball fault signals; (d) AE models learned from the outer race centered fault signals. AE-1: layers [120, 60, 30, 15, 30, 60, 120]; AE-2: layers [120, 60, 30, 60, 120]; AE-3: layers [120, 60, 120]; AE-4: layers [120, 30, 120]; AE-5: layers [120, 15, 120].
Figure 6. Average training errors with standard deviations of AE models on datasets 1–4. (a) AE models learned from the normal signals; (b) AE models learned from the inner race fault signals; (c) AE models learned from the ball fault signals; (d) AE models learned from the outer race centered fault signals. AE-1: layers [120, 60, 30, 15, 30, 60, 120]; AE-2: layers [120, 60, 30, 60, 120]; AE-3: layers [120, 60, 120]; AE-4: layers [120, 30, 120]; AE-5: layers [120, 15, 120].
Sensors 24 02978 g006
Figure 7. Average validation accuracy of FFNN models with different hyperparameter configurations on experimented datasets.
Figure 7. Average validation accuracy of FFNN models with different hyperparameter configurations on experimented datasets.
Sensors 24 02978 g007
Figure 8. Performance distribution of the compared models across the experimented datasets. (a) Box plot of the accuracy; (b) box plot of the precision; (c) box plot of the recall; (d) box plot of the f-measure; (e) box plot of the AUC.
Figure 8. Performance distribution of the compared models across the experimented datasets. (a) Box plot of the accuracy; (b) box plot of the precision; (c) box plot of the recall; (d) box plot of the f-measure; (e) box plot of the AUC.
Sensors 24 02978 g008
Table 1. Data files of the experimental datasets.
Table 1. Data files of the experimental datasets.
Prepared DatasetMotor Load hpShaft Speed rpmFile NumberFault TypesFault Size# SamplesMinMaxAvgStdev
Dataset-1HP-0179797Normal*243,938−0.2870.3110.0130.073
222Ball0.021121,991−1.6181.6600.0210.134
209Inner race0.021122,136−3.3723.7880.0140.525
234Outer race centered0.021122,426−6.4126.6530.0050.583
Dataset-2HP-1177298Normal*483,983−0.3460.3180.0130.065
223Ball0.021121,701−1.4141.4750.0040.129
210Inner race0.021121,556−3.2843.6860.0030.442
235Outer race centered0.021122,281−6.6536.6530.0030.57
Dataset-3HP-2175099Normal*485,063−0.3270.3590.0120.063
224Ball0.021122,136−0.5310.6460.0050.107
211Inner race0.021121,846−3.1243.6230.0030.489
236Outer race centered0.021121,991−6.6536.6530.0040.559
Dataset-4HP-31730100Normal*485,643−0.3060.2840.0120.065
225Ball0.021122,136−0.4930.5760.0040.118
212Inner race0.021121,991−3.0873.6150.0030.449
237Outer race centered0.021121,991−6.1136.2070.0030.561
Avg: average value; Max: maximum value; Min: minimum value; Stdev: standard deviation. Please note that the symbol “*” indicates that the fault size is irrelevant.
Table 2. Input parameters of the experimented classifiers.
Table 2. Input parameters of the experimented classifiers.
ClassifiersConfigured Values
K-Nearest NeighborsThe number of neighbors was configured from 3 to 25.
Support Vector MachineThe linear, polynomial, radial basis function (rbf), and sigmoid kernels were used.
Adaptive BoostingThe number of estimators was configured from 10 to 150.
Decision TreeClassification criteria were configured by “gini” and “entropy”.
Random ForestThe number of trees was configured from 10 to 150.
Table 3. Configurations of the models in the proposed method: the proposed method consists of two neural network models for reconstruction-error-based feature generation and fault detection.
Table 3. Configurations of the models in the proposed method: the proposed method consists of two neural network models for reconstruction-error-based feature generation and fault detection.
ConfigurationsFFNNAENN
Layer: (neurons)Input: (121)
Hidden: (242, 121, 60)
Output: (4)
Input: (120)
Hidden: (60)
Output: (120)
Layer: (activation function)Hidden: (ReLU)
Output: (Softmax)
Hidden: (ReLU)
Learning rate0.0010.001
OptimizerAdam [24]Adam [24]
Loss functionCategorical Cross-EntropyMean Squared Error
Batch size88
Epochs500 early stopping [25] of 100500
Table 4. Confusion matrix of the baseline FFNN models on datasets 1–4.
Table 4. Confusion matrix of the baseline FFNN models on datasets 1–4.
Actual Actual
PredictedDataset-1NormalBall faultInner race faultOuter race faultPredictedDataset-2NormalBall faultInner race faultOuter race fault
Normal201001Normal395000
Ball fault011111Ball fault011611
Inner race fault01912Inner race fault01963
Outer race fault01099Outer race fault01193
Missed0214 Missed0224
Actual Actual
PredictedDataset-3NormalBall faultInner race faultOuter race faultPredictedDataset-4NormalBall faultInner race faultOuter race fault
Normal397000Normal403000
Ball fault09501Ball fault09603
Inner race fault001056Inner race fault001040
Outer race fault010104Outer race fault06098
Missed0107 Missed0603
Table 5. Comparison between the baseline FFNN models and the proposed AE-FFNN models on datasets 1–4.
Table 5. Comparison between the baseline FFNN models and the proposed AE-FFNN models on datasets 1–4.
PerformanceModels Were Trained on the Dataset-1, and Tested on the Datasets 2–4Models Were Trained on the Dataset-2, and Tested on the Dataset 1, 3, 4Models Were Trained on the Dataset-3, and Tested on the Datasets 1, 2, 4Models Were Trained on the Dataset-4, and Tested on the Datasets 1–3
BaselineProposedBaselineProposedBaselineProposedBaselineProposed
Accuracy96.694 (0.003)97.484 (0.002)97.724 (0.007)98.284 (0.004)96.502 (0.012)97.489 (0.006)95.526 (0.018)96.489 (0.012)
Precision95.045 (0.004)95.887 (0.002)96.833 (0.004)97.462 (0.002)95.064 (0.009)96.296 (0.005)94.974 (0.012)95.573 (0.009)
Recall94.237 (0.006)95.69 (0.004)96.574 (0.005)97.404 (0.002)94.722 (0.011)96.173 (0.005)93.339 (0.017)94.726 (0.01)
F-measure94.49 (0.006)95.735 (0.004)96.691 (0.005)97.42 (0.002)94.849 (0.01)96.203 (0.005)93.947 (0.016)95.048 (0.01)
AUC99.717 (0.001)99.776 (0)99.746 (0.001)99.835 (0.001)99.465 (0.002)99.513 (0.002)99.578 (0.002)99.606 (0.001)
Table 6. Average performance of the compared models.
Table 6. Average performance of the compared models.
PerformanceKNNRFAdaBoostSVMDTNBProposed
Accuracy87.20094.40877.45091.97580.13378.23397.559
Precision86.14292.42567.75091.02571.41768.00896.444
Recall79.42591.14266.40887.38370.30068.42596.189
F-measure81.57591.72566.89288.79270.74266.96796.274
AUC91.86799.28386.28399.12581.83393.43399.699
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Amarbayasgalan, T.; Ryu, K.H. Unsupervised Feature-Construction-Based Motor Fault Diagnosis. Sensors 2024, 24, 2978. https://doi.org/10.3390/s24102978

AMA Style

Amarbayasgalan T, Ryu KH. Unsupervised Feature-Construction-Based Motor Fault Diagnosis. Sensors. 2024; 24(10):2978. https://doi.org/10.3390/s24102978

Chicago/Turabian Style

Amarbayasgalan, Tsatsral, and Keun Ho Ryu. 2024. "Unsupervised Feature-Construction-Based Motor Fault Diagnosis" Sensors 24, no. 10: 2978. https://doi.org/10.3390/s24102978

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop