Next Article in Journal
Experimental Study on the Propagation Characteristics of Rotating Detonation Wave with Liquid Hydrocarbon/High-Enthalpy Air Mixture
Previous Article in Journal
Research and Performance Optimization of Jump-Takeoff in Autogyros
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Universal Feature Extractor Based on Self-Supervised Pre-Training for Fault Diagnosis of Rotating Machinery under Limited Data

School of Reliability and Systems Engineering, Beihang University, Beijing 100191, China
*
Author to whom correspondence should be addressed.
Aerospace 2023, 10(8), 681; https://doi.org/10.3390/aerospace10080681
Submission received: 12 July 2023 / Revised: 26 July 2023 / Accepted: 28 July 2023 / Published: 30 July 2023

Abstract

:
To address the limited data problem in real-world fault diagnosis, previous studies have primarily focused on semi-supervised learning and transfer learning methods. However, these approaches often struggle to obtain the necessary data, failing to fully leverage the potential of easily obtainable unlabeled data from other devices. In light of this, this paper proposes a novel network architecture, named Signal Bootstrap Your Own Latent (SBYOL), which utilizes unlabeled vibration signals to address the challenging issues of variable working conditions, strong noise, and limited data in rotating machinery fault diagnosis. The architecture consists of a self-supervised pre-training-based fault feature recognition network and a diagnosis network based on knowledge transfer. The fault feature recognition network uses ResNet-18 as the backbone network for self-supervised pre-training and transfers the trained fault feature extractor to the target diagnostic object. Additionally, a unique vibration signal data augmentation technique, time–frequency signal transformation (TFST), is proposed specifically for rotating machinery fault diagnosis, which addresses the key task of contrastive learning and achieves high-precision fault diagnosis with very few labeled samples. Experimental results demonstrate that the proposed diagnostic model outperforms other methods in both extremely limited sample and strong noise scenarios and can transfer unlabeled data utilization between similar and even different device types.

1. Introduction

Rotating machinery, such as aero-engines, wind turbines, and gearboxes, is widely used in industry and is prone to various failures under harsh operating conditions such as high temperatures, variable speeds, and heavy loads. Timely fault diagnosis is critical to ensure equipment safety and prevent severe failures [1].
Deep learning [2] has gained increasing attention in the field of fault diagnosis, primarily due to its exceptional feature extraction capability. Convolutional neural networks (CNNs) [3,4,5] are particularly favored for their advantages, including parameter sharing and powerful non-linear feature learning. However, data-driven fault diagnosis models based on deep learning require a large amount of high-quality raw data [6]. Obtaining large labeled datasets is impractical, as operating rotating machinery under fault conditions for extended periods is unrealistic. With only a limited amount of labeled data available, deep learning models struggle to achieve satisfactory fault diagnosis performance.
To address the above problem in fault diagnosis for rotating machinery, researchers have primarily utilized semi-supervised and transfer learning methods. Semi-supervised methods leverage a combination of unlabeled and labeled data to enhance model performance. A method called hybrid classification autoencoder was proposed by Wu et al. [7] to diagnose faults in rotating machinery using features obtained from the autoencoder. A semi-supervised deep sparse autoencoder (SSDSAE) with local and nonlocal information was proposed by Zhao et al. [8] for the intelligent fault diagnosis of rotating machinery. Transfer learning methods, on the other hand, transfer knowledge from a source domain to a target domain to improve diagnostic performance. Li et al. [9] proposed an adversarial transfer learning method based on stacked autoencoders to address fault classification under different operating conditions. Han et al. [10] combine models of joint distributed adaptation and deep networks to facilitate the diagnosis of a new but similar target task.
Although the above-mentioned methods have shown promising results, their applicability is limited to specific scenarios. Semi-supervised methods mainly focus on the case of labeled and unlabeled data from the same diagnostic object, which is difficult to obtain in practical diagnostic tasks. Similarly, transfer learning methods require labeled source domain data to improve diagnostic accuracy [11]. Both methods require computationally expensive training on additional data and target diagnostic data for different diagnostic tasks. Gaining access to a large amount of data from the specific diagnostic object or obtaining labeled data from different diagnostic objects can be challenging. However, unlabeled data from different products is often easily obtainable, raising the question of how to effectively leverage this data resource.
In contrast to the algorithms described above, the self-supervised learning methods explore a new solution that takes full advantage of unlabeled data easily available from different diagnostic objects [12]. Self-supervised learning methods can learn effective signal representations from large-scale unlabeled data, and they have better applicability than semi-supervised learning methods. Moreover, unlike transfer learning methods, self-supervised methods do not require a large amount of labeled source domain data. Furthermore, they do not need repeated training with unlabeled data for each downstream task, so they can be quickly applied to various downstream diagnostic tasks. Many self-supervised learning methods have been developed in the field of computer vision. Some methods deviate from contrastive learning and instead adopt manually designed prediction tasks like image colorization [13], image inpainting [14], image jigsaw puzzle [15], and image super-resolution [16] to learn representations. These alternative approaches have demonstrated their effectiveness. Meanwhile, contrastive learning, as the most advanced self-supervised learning method, performs representation learning by reducing the distance between representations of different augmented views of the same image (“positive pairs”) and increasing the distance between representations of views of different images (“negative pairs”) [17,18]. Contrastive learning has been shown to extract the essential features of the data and provide better representation learning than the methods described above [19,20,21].
However, research on self-supervised learning in fault diagnosis is still limited. Wang et al. [22] employed data augmentation techniques to signals and trained classification models to identify corresponding augmentation methods. Ding et al. [23] proposed self-supervised pre-training via contrast learning (SSPCL), which is based on MoCo [24] and uses momentum contrastive learning for instance-level discrimination, thus enabling feature learning. Wei et al. [25] used SimCLR [20], which first transforms the signal from one-dimensional to two-dimensional using a simple matrix operation, then applies image domain data augmentation methods before converting it back to one-dimensional for representation learning. Yan et al. [26] proposed SMoCo, which is based on MoCo with improved structure and data augmentation techniques. It introduces a novel comparison method and has been successfully applied to bearing fault diagnosis in aero-engines, even with limited data. Shul et al. [27] developed a deep neural network for detecting anomalies in washing machines based on the noise spectra generated during their operation. The main self-supervised task of their architecture is to predict future noise based on past noise. Nie et al. [28] proposed a generalized model based on self-supervised learning and sparse filtering, which employs corresponding labels assigned to signals undergoing different feature transformations for self-supervised learning.
Although the above methods attempt to apply self-supervised learning to fault diagnosis, their applicability is primarily limited to utilizing a large number of unlabeled samples and a small number of labeled samples of the same diagnostic object. Moreover, these methods have not undergone thorough investigation across different noise intensities, operating conditions, sample sizes, and other factors. As a consequence, the lack of a universal feature extractor that can be applied to various types of equipment greatly limits their application scope. The reason for this is that the data augmentation methods utilized in their study were not designed to fully exploit the unique characteristics of vibration signals. In addition, the model architectures employed exhibited several shortcomings, posing challenges to achieving high diagnostic accuracy in complex real-world settings for fault diagnosis purposes. Specifically, Wang and Nie’s method [22,28] only identifies data augmentation categories without instance-level representation learning, limiting its ability to extract robust fault features effectively. In contrast, for contrastive learning methods, Ding’s method [23] employs a time-domain transformation for vibration signal data augmentation but fails to leverage the time- and frequency-domain characteristics simultaneously. Wei’s approach [25] based on SimCLR requires large batch sizes during training, and the data augmentation method it uses is limited to the image domain.
To solve the above problem, this paper proposes a new self-supervised method signal bootstrap your own latent (SBYOL). SBYOL uses two networks including the online network and the target network, which interact and learn from each other, and uses the similarity of the output of the two networks as the loss function. Based on the characteristics of vibration signals, this paper proposes a new contrastive learning data augmentation method, time–frequency signal transformation (TFST). It consists of two parts, time–frequency contrast (TFC) and segment cross contrast (SCC), which greatly helps the model to learn the essential features of the signal. SBYOL can effectively utilize easily accessible unlabeled data with different sources from the object to be diagnosed and learn how to extract the fault characteristics of vibration signals independent from working conditions, noise, and even equipment types. Then, by transferring this capability to various downstream diagnostic tasks, it greatly solves the problems of limited data, complex signals, and strong noise in practical fault diagnosis scenarios and provides a powerful solution for the efficient utilization of industrial big data. The contributions and innovations of this paper are summarized as follows:
(1)
This paper proposes a novel data augmentation method called TFST based on the unique characteristics of vibration signals to address the key points of rotating machinery fault diagnosis, which enables the model to learn signal representations from both the time and frequency domains simultaneously.
(2)
A universal and robust automated feature extractor is constructed in this paper through pre-training on a public parallel gearbox dataset. This feature extractor can achieve excellent diagnostic accuracy using a simple classifier with limited training data, whether it is a private complex planetary gearbox or public planetary gearbox dataset, or even two public bearing products of completely different types.
(3)
The unlabeled pre-training data used by SBYOL are no longer limited to the same diagnostic object but can even be completely different types of equipment, which greatly increases its feasibility in practical tasks.
(4)
Further experiments demonstrate that SBYOL still has excellent accuracy for the target diagnostic object with extremely limited training data and strong noise, has good stability for the unlabeled pre-training dataset with smaller sampling time and data size, outperforms other state-of-the-art methods, and further proves its robustness.
The paper is structured as follows. Section 2 introduces self-supervised learning. Section 3 describes the proposed SBYOL framework. In Section 4, the performance of SBYOL is experimentally verified. Section 5 summarizes the paper and looks at future work.

2. Self-Supervised Learning

For supervised learning, given a labeled dataset D L = x i , y i i = 1 N L , x i and y i are the samples and the corresponding labels in the dataset D L , respectively. A loss function is defined to measure the distance between y i and the network output y ^ i , usually the cross-entropy loss function. The optimization objective of the network is to minimize the cross-entropy loss function, and it performs well on large-scale labeled datasets.
However, labeled data are often difficult to obtain, while self-supervised methods can learn useful representations from unlabeled data. Given an unlabeled dataset D U = x i i = 1 N U , self-supervised methods perform representation learning by constructing a suitable pretext task that minimizes a predefined loss function, which is then transferred to a downstream diagnostic task, thereby improving diagnostic accuracy. Self-supervised learning aims to keep similar samples close and different samples far away [29]. The pretext task is an important concept in self-supervised learning [12], meaning that it solves a task that is not directly needed which is a classification task in the field of fault diagnosis. It defines the task by the attributes found in unlabeled data and thus performs the implementation of unsupervised representation learning. The pretext tasks proposed by previous work mainly focused on performing time-domain transformations on the signal, enabling the model to identify different transformations of the same signal sample and distinguish between different samples.

3. Signal Bootstrap Your Own Latent (SBYOL)

To address the drawback that previous self-supervised learning methods do not make full use of the characteristics of the vibration signals to design the pretext task, we propose a fault diagnosis method based on SBYOL. It has the advantages of strong feature extraction capability, more focus on the signal itself, and weak task correlation, which makes it more suitable for fault diagnosis tasks for different models of equipment or even different types of equipment. This section details the core algorithms of SBYOL, including a novel and efficient method for signal-specific data augmentation, and how unsupervised representation learning can be performed and used for downstream diagnostic tasks.

3.1. Methodology Overview

The purpose of this study is to improve the diagnostic accuracy of the target diagnostic task by making full use of easily available task-independent unlabeled data in cases where there are limited data in practical fault diagnosis. The method allows representation learning on a pre-training dataset without any manual annotation to obtain a feature extractor capable of extracting the essential features of the vibration signal. The learned feature extractor is then transferred to a downstream diagnostic task, thereby improving diagnostic performance. Our approach is divided into three stages, and the flow chart is shown in Figure 1. In the first stage, data collection is performed and divided into an unlabeled pre-training dataset and a labeled target diagnostic dataset, where the pre-training dataset is different from the target diagnostic object. In the second stage, self-supervised learning is used to learn how to extract robust features from unlabeled vibration signal data. In the third stage, the online network encoder obtained by self-supervised learning is transferred to the target diagnosis task and used as a feature extractor with fixed weights. Finally, support vector machine (SVM) is used to classify the features extracted by the feature extractor to obtain the fault diagnosis model. Next, we will describe the second and third learning stages in detail.

3.2. Data Augmentation Based on the Time–Frequency Signal Transformation (TFST)

In contrast to supervised learning, self-supervised learning relies more on the design and optimization of data augmentation methods [20], and it is therefore necessary to design augmentation methods according to the characteristics of the vibration signal. In the actual operation of rotating machinery, there are many diagnostic difficulties, such as variable working conditions and strong noise. If a model can be independent of these factors, then this model can extract essential features from the signal, and therefore, transferring this capability to fault diagnosis tasks can greatly improve diagnostic performance. In contrast to previous approaches that focus solely on morphological changes and time-varying operations, this paper introduces a novel set of pretext tasks based on signal transformations, specifically leveraging the time–frequency characteristics of the signal. Notably, the innovative TFST method is proposed, significantly enhancing the model’s capability for representation learning. Details of how these methods transform a given vibration signal x = x 1 , x 2 , , x N are described below:
(1)
Normalize, as shown in Figure 2a. As there are differences in the measurement ranges of different sensors, this strategy normalizes the signals to a uniform measurement range, and in addition, it optimizes the convergence of the model. The formula is as follows:
x ~ = 1 + 2 × x x m i n x m a x x m i n
(2)
AddGaussian, as shown in Figure 2b. As noise is inevitable in the actual operating environment of the device, this strategy improves the model’s immunity to noise by randomly adding Gaussian noise to the input signal, which is formulated as follows:
x ~ = x + n , n ~ N 0 , σ n
where n is generated by Gaussian distribution N 0 , σ n .
(3)
Scale, as shown in Figure 2c. Since the equipment is loaded to different degrees in the actual operating environment, the sensitivity of the model to signals of different amplitudes can be improved by changing the amplitude of the signals, thus enabling the diagnosis of variable working condition faults. The strategy multiplies the input signal randomly by a factor s, which is calculated as follows:
x ~ = x × s , s ~ N 1 , σ s
where s is generated by the Gaussian distribution N 1 , σ s .
(4)
Stretch, as shown in Figure 2d. In response to the existence of different working conditions in the operation of the equipment, the signal is resampled, and its length is converted to the original length in s ~ N 1 , σ s times, simulating the speed variation of different operating conditions. Finally, equal lengths are ensured by zeroing and truncation.
(5)
Crop, as shown in Figure 2e. For the problem of missing data during fault diagnosis, this strategy randomly overwrites some signals with the following equation.
x ~ = x × m a s k
where the mask is a binary sequence with a random subsequence of zeros.
(6)
Flip, as shown in Figure 2f. The vibration signal usually oscillates up and down with an average value of zero. To improve the model from the effects of positive and negative signals, the strategy simulates this variation by randomly flipping the signal with the following equation.
x ~ = x
(7)
Time–frequency signal transformation (TFST)
The proposed time–frequency signal transformation (TFST), illustrated in Figure 3, integrates both segment cross contrast (SCC) and time–frequency contrast (TFC) techniques. This combination enables the extraction of features from both the time domain and the frequency domain simultaneously. Moreover, it retains the prior information from the time-domain signal interception, thereby greatly improving the ability of representation learning for contrastive learning. Next, we will introduce the intuition and principles of SCC and TFC in detail.
The SCC is shown on the left side of Figure 3. During signal acquisition, the state of the device does not change much over a short period, thus allowing it to be considered as the same state. After sampling, a long signal is obtained, and it is necessary to segment the long signal to obtain multiple signal instances that can be used as training data for the model. However, traditional contrastive learning only considers instances as individual classes and disregards that different instances segmented from the same long signal belong to the same class. To address this issue, we propose a strategy of comparing signal instances obtained by intercepting the same long-sampled signal with each other. This approach can increase the prior knowledge of model training, facilitate the identification and aggregation of similar samples during the representation learning process, and enhance the model’s ability to withstand time disturbances. Specifically, two different signals x 1 and x 2 are randomly selected from the same long-sampled signal interception set. They undergo a series of signal transformations, respectively, and finally, their representations are made as similar as possible.
The TFC, depicted on the right side of Figure 3, leverages both time-domain and frequency-domain features of vibration signals, which are commonly extracted as fault features, to enable effective analysis of such signals. In this paper, the original time-domain signal is compared with its fast Fourier transform (FFT) counterpart, allowing signal representation extraction from two dimensions. To achieve this, a series of data augmentation methods are applied to the time-domain signal, including Normalize, AddGaussian, Scale, Stretch, Crop, and Flip, which is referred to as time domain transformation (TDT). Additionally, the FFT, Normalize, and AddGaussian are used as data augmentation methods to generate the frequency-domain feature, referred to as frequency-domain augmentation (FDT). The overall strategy involves comparing the original signal after TDT transform with the original signal after FDT transform, to minimize the distance between the two representations.
Therefore, the overall process of TFST is that, given a long-sampled signal from which two segments of the signal x 1 and x 2 are randomly intercepted, the TDT transform of x 1 and the FDT transform of x 2 are performed, respectively, and finally, their representations are made as close as possible.

3.3. Self-Supervised Signal Representation Learning

In the self-supervised signal representation learning stage, SBYOL uses two neural networks for training: the online network and the target network, as shown in Figure 4. SBYOL obtains new and continuously enhanced representations by enabling the online network to predict the representations of the target network. This iterative process results in a sequence of representations that become increasingly refined. After completing the training, only the encoder f θ of the online network is retained for use in downstream tasks.
The online network consists of a set of weights θ and includes three components: an encoder f θ , a projector g θ , and a predictor q θ . While the original self-supervised learning methods with a deeper ResNet encoder were designed for computer vision tasks, fault diagnosis of rotating machinery is simpler compared to representation learning of images, and it can be difficult to achieve convergence with high computational complexity when using a deep network architecture. Therefore, this paper employs the encoder with a lighter architecture ResNet-18 [30], to reduce the computational complexity. Specifically, ResNet-18 consists of 18 layers in total, including 16 1D convolutional layers and 2 fully connected layers. The initial convolutional layer performs a 1 × 7 convolution with stride 2, followed by a max pooling layer. Then, four sets of residual blocks are stacked, each containing two 1D convolutional layers with 1 × 3 filters. The number of filters in each block gradually increases from 64 to 512, capturing more complex patterns as the spatial dimension reduces. Both the projector and the predictor are multilayer perceptrons (MLP), with the same output dimensions. To ensure stable training, batch normalization (BN) [31] is incorporated into the projector and predictor.
The target network has the same structure as the encoder and projector of the online network but uses different weights ξ . During training, the online network predicts the targets generated by the target network, and the predictor in the online network greatly increases flexibility, as it allows for different features in the online and target networks to be matched by the predictor, thereby improving representation learning. To prevent the target network parameters ξ from updating during training, stop-gradient (sg) is employed, and the target network is updated via exponential moving average using the online network parameter θ . This approach ensures that the two networks have different parameters so that when the online network regresses the signal features, the targets are distinct, and feature dispersion is preserved. Given the update parameter τ 0 , 1 , after each epoch, it updates the target network as follows.
ξ τ ξ + 1 τ θ
To generate the pretext task for self-supervised learning in SBYOL, two random samples ( x 1 and x 2 ) are extracted from a long signal, and data augmentation is applied to each of them using two distributions (TDT and FDT) to create two corresponding augmented time series ( v and v ). For the first augmented sequence v , the online network outputs the feature z θ = g θ ( f θ ( v ) ) and then uses the predictor to predict z θ to obtain q θ ( z θ ) . For the second augmentation sequence v , the target network outputs z ξ = g ξ ( f ξ ( v ) ) . Additionally, to prevent the scale of features from approaching zero and avoid model collapse, both q θ ( z θ ) and z ξ are normalized by l 2 , resulting in q θ ¯ ( z θ ) = q θ ( z θ ) / q θ ( z θ ) 2 and z ¯ ξ = z ξ / z ξ 2 . This normalization ensures that the model does not learn shortcuts and that the features remain well-scaled. Training is performed by minimizing the mean square error between the normalized online network prediction and the target network projection.
L θ , ξ q θ ¯ z θ z ξ ¯ 2 2 = 2 2 q θ ( z θ ) , z ξ q θ z θ 2 z ξ 2
The SBYOL network gets the symmetric loss function L ~ θ , ξ of L θ , ξ by sending v to the online network and v to the target network. Finally, the network updates the online network θ q by minimizing the loss L θ , ξ S B Y O L :
L θ , ξ S B Y O L = L θ , ξ + L ~ θ , ξ
The detailed training algorithm for self-supervised signal representation learning is shown in Algorithm 1.
Algorithm 1: Self-Supervised Signal Representation Learning
Input:
Structure of f θ , g θ , q θ , f ξ , g ξ , initial online network parameters θ , initial target network parameters ξ ,
update parameter τ , batch size M , learning rate η , optimization step N ,
distributions of transformations T D T , F D T , set of signals D
for  n = 1 to N do
    B a t c h ( x 1 i , x 2 i ) D i = 1 M
   for  ( x 1 i , x 2 i ) B a t c h do
       t T D T and t F D T
       z 1 g θ ( f θ ( t x 1 i ) ) and z 1 g ξ ( f ξ ( t x 2 i ) )
       z 2 g θ ( f θ ( t x 2 i ) ) and z 2 g ξ ( f ξ ( t x 1 i ) )
       l i 2 × ( q θ z 1 , z 1 q θ z 1 2 z 1 2 + q θ ( z 2 ) , z 2 q θ z 2 2 z 2 2 )
   end
//Back-propagation
    θ θ η 1 M i = 1 M l i θ
//Exponential moving average without back-propagation
    ξ τ ξ + 1 τ θ
end
Output: Online network encoder f θ

3.4. Fault Diagnosis Based on Knowledge Transfer

After training, the convolutional layers of the online network encoder are extracted and used for feature extraction in the downstream tasks. It is worth noting that the weights of the convolutional layers are kept fixed to handle downstream complex tasks under challenging conditions, such as limited data and significant noise. This strategy helps maintain the robustness of the model, preventing bias caused by overfitting with small training sets. Additionally, by keeping the weights fixed, the model can be readily used for downstream diagnostic tasks without requiring any additional training. This approach not only ensures efficiency but also allows for faster deployment of the model in practical applications. Since SVM is the classifier with the largest interval in the feature space and has stronger robustness in problems with limited data, this paper uses SVM to classify the extracted features to build the final fault diagnosis model.

4. Experiment Validation

Due to the small number of scenarios in which planetary gearboxes are used, they often have limited data problems, and there are almost no public data. In addition, compared to parallel gearboxes, their structure is more complex, and their signal composition is relatively complicated, making their diagnosis a difficult task. In contrast, parallel gearbox data are relatively easier to obtain, so in this paper, the unlabeled public University of Connecticut (UoC) parallel gearbox dataset [32,33] is used as the pre-training dataset. To verify the effectiveness and superiority of SBYOL, SBYOL first performs self-supervised signal representation learning on the unlabeled pre-training dataset; then, the learned feature extractor is transferred to the private Drivetrain Prognostics Simulator (DPS) planetary gearbox under limited data conditions and uses the public SEU planetary gearbox dataset [34] for further validation. To further validate the robustness and generalizability of SBYOL, this paper further increases the difficulty by using the trained feature extractors for fault diagnosis in limited data cases on two public datasets, the Paderborn University (PU) bearing dataset [35] and the Polytechnic University of Turin (PUT) aero-engine bearing dataset [36], which are characterized by completely different types of equipment than the pre-training parallel gearbox data. The data distribution between the pre-training dataset and the target diagnostic object is significantly different, and thus, it can be effectively verified that SBYOL learns a universal feature extractor that can efficiently diagnose different types of rotating machines under limited data.

4.1. Self-Supervised on the Unlabeled Pre-Training Parallel Gearbox Dataset

In this paper, the UoC parallel gearbox fault dataset [32,33] provided by the University of Connecticut is used as the unlabeled pre-training dataset, which is the most difficult public dataset with different failure modes and degrees [37]. The dataset is collected at 20 kHz and introduces nine different gear states for pinions on the input shaft, including health condition, missing tooth, root crack, spalling, and chipping tip with five different levels of severity, and they are labeled from 0 to 8, respectively. The test rig is shown in Figure 5.
Our method selects the raw vibration signal as the input data without any signal pre-processing, and 4096 was chosen as the sample length to contain enough information. In addition, the overlap length between two adjacent samples is 3036 when using the sliding segmentation method to obtain training samples, thus greatly increasing the size of the pre-training dataset. This is also a data augmentation strategy that increases the consideration of temporal offset; however, the information capacity of the dataset does not increase. To mimic the phenomenon of intercepting a signal from a long sampling signal, this paper treats approximately 9.43 s of data as a long sampling signal with 175 samples intercepted from a single sampling signal, and finally, an unlabeled pre-training dataset containing nine classes with 350 samples per class is obtained.
Since this paper uses both time-domain and frequency-domain data for learning and even contains learning between different samples, the data distribution between them is quite different; the learning task is also more complicated, and 0.1 is chosen as the initial learning rate. In this study, we adopted the numerical values of Momentum and Weight decay from [24], while the update parameter τ was derived from [18]. Regarding data augmentation methods, using larger values can enhance the model’s noise resistance capability. However, it is essential to strike a balance and avoid excessive values that may lead to information overload. The specific hyperparameter values are shown in Table 1. It is worth noting that all the data augmentation methods except TFC and SCC are implemented with a probability of 0.5 to increase the complexity of the pretext task. The learning rate is updated by the cosine learning rate scheduler with the following formula.
η t = 1 2 ( 1 + c o s ( t π T ) ) η
where η is the initial learning rate, η t is the current learning rate, T is the maximum number of epochs, and t is the current epoch.
The model has a parameter count of 8.48 million and achieves a computational efficiency of 2.81 billion floating-point operations per second (GFLOPs). The experimental environment was PyTorch 1.11 under Windows 11, running on a computer with the following configuration: i5-12400F, NVIDIA RTX 3060, 16 GB RAM. The changes in the loss values during the training process are shown in Figure 6, from which it can be found that the loss values become smooth in the later stages of training indicating that the model has reached the fitting state, and the total training time is about 6.5 h.
As a comparison, other self-supervised learning methods, SimSiam [39], SimCLR, BYOL, MoCo, and SMoCo, are also used in this paper for self-supervised pre-training on the unlabeled pre-training dataset. In addition, to further demonstrate the powerful performance of SBYOL, this paper also uses the labeled pre-training dataset for supervised pre-training, called Labeled Pre-Training. To exclude the influence of other factors, the backbone network of all methods is ResNet-18, which is trained using time-domain signals. After training, the feature extractors of all methods, i.e., the convolutional layers of ResNet-18, are used to extract features from a portion of the pre-training dataset and downscale it to two dimensions using T-SNE for visualization; the results are shown in Figure 7. Our method, SBYOL, achieves very good results with inter-class separation and intra-class aggregation, reaching the results of Labeled Pre-Training. In the face of such complex data, other self-supervised methods not only have no clustering within the class but also do not have an effective separation between classes, and only SMoCo performs better roughly achieving clustering of three classes of samples.

4.2. Application on Planetary Gearbox from Private DPS Test Rig

In this section, the pre-trained SBYOL feature extractor is used to perform fault diagnosis on the planetary gear from our Drivetrain Prognostics Simulator (DPS) test rig, which is characterized by a different device with different failure levels, different failure modes, and different working conditions compared with the unlabeled pre-training dataset.
The Drivetrain Prognostics Simulator is shown in Figure 8a which is manufactured by Spectra Quest, U.S.A. The test rig is mainly composed of the following parts: driver (control cabinet), lubrication system, drive motor, testing planetary gearbox, three load parallel gearboxes, load motor and supporting torque transducer and force transducer, etc. The testing planetary gearbox of this test rig is a one-stage drive, compared with the parallel gearbox, it has a more complex structure, including the central fixed sun wheel, planetary frame, and gear ring, and also includes four planetary wheels that change with the sun wheel rotation center at all times; the signal composition is very complex, and therefore, fault diagnosis is a more difficult task. During the data collection, we collected gear data for five states, including wear, broken teeth, missing teeth, root crack, and healthy, where the four failure modes of the gears are shown in Figure 9.
Two vibration sensors were used in the experiment, one is an end cap sensor shown in Figure 8b, and the other is a three-phase sensor mounted on the box shown in Figure 8c. The sampling frequency is 12.8 kHz. Due to the complex transmission path of the fault signal through the gear, shaft, bearing, and end cup to reach the box, the signal collected by the box vibration sensor will become weak. To verify the effectiveness of SBYOL in extreme cases, this paper selects the Z-axis box vibration signal, which is more in line with the complex environment in the actual diagnosis task. To reflect the extremely limited data situation in the actual diagnosis process, only 5 samples per class are used in the training set, and 50 samples per class are used in the testing set. The details of the DPS gear dataset are shown in Table 2, which contains two working conditions, and the length of each sample is still 4096.
To demonstrate the performance of the feature extractors learned on the unlabeled pre-training dataset, this paper uses the pre-trained feature extractors to perform feature extraction on the DPS testing set without any training and visualizes the results using T-SNE as shown in Figure 10. The features extracted by SBYOL without any adaptation of the target diagnostic object can be segmented between different classes and aggregated of the same class, far surpassing other self-supervised methods, and even Labeled Pre-Training. SMoCo also adopts the way of extracting features from the time and frequency domains simultaneously, and its feature extraction effect on the testing set data is relatively good, but it lacks unique SCC, and compared to SBYOL, labels 4, 8, and 9 are too close, so it is more prone to errors with very few training data.
To fully demonstrate the superiority of our method in the target diagnosis task, we also added MixMatch [40], ResNet-18, and FFT + SVM as comparisons for the target diagnostic dataset. Among them, MixMatch is a powerful semi-supervised method that uses both an unlabeled pre-training dataset and labeled target diagnostic training set for training. ResNet-18 uses only the targeted diagnostic training set for supervised learning and does not use the unlabeled pre-training dataset as the baseline model. FFT + SVM is a classical and effective method for fault diagnosis under limited data. The method first performs FFT transformation on the original signal and then classifies the FFT transformed features using SVM. The diagnostic accuracy of each method is shown in Table 3 and Figure 11. In addition, to ensure the fairness of the experiments, other self-supervised learning methods also use fixed weights for feature extraction, and then SVM is used as the classifier. For Labeled Pre-Training, the training is performed in the standard pre-training plus fine-tuning manner, i.e., a linear projection is trained, and fine-tuning of the convolutional layers with a small learning rate is performed. The accuracy scores are averaged and calculated ten times to eliminate computational errors, and the corresponding standard deviation (STD) is calculated to verify their robustness.
As can be seen from Table 3 and Figure 11, SBYOL achieves the best results reaching an accuracy of 97.14% in the face of a complex planetary gearbox, which greatly surpasses other methods. This is also consistent with the feature visualization performance in Figure 10, where SBYOL’s feature extractor can distinguish well between classes without training with the target diagnostic object. Consequently, only a small number of samples are required to construct a highly accurate classification boundary. Furthermore, the volatility of the feature extractor is relatively low, with a standard deviation of only 1.60. In addition, since SBYOL only acts as a fixed-weight feature extractor, its application to downstream diagnostic tasks is also very efficient, requiring only 2.43 s. For Labeled Pre-Training, although its diagnostic accuracy is improved compared to the baseline ResNet-18, its performance is far inferior to SBYOL for cross-device diagnostic problems under limited data because of its feature extractor obtained by supervised learning on the pre-training dataset. Other self-supervised learning methods perform poorly compared to SBYOL due to the lack of our unique TFST and gaps in the structure. For FFT + SVM, its performance is better than that of ResNet-18 in the case of limited data, which utilizes only time-domain features, but its diagnostic accuracy is not high in the face of complex diagnostic problems. MixMatch is trained with both target diagnostic data and unlabeled pre-training data, so it can adapt the pre-training data to the target diagnostic data to achieve good diagnostic accuracy. However, the accuracy of MixMatch is lower than SBYOL, and since it uses a large amount of unlabeled data for training at the same time, it is much slower than our method to apply it to the target diagnostic task.
This paper further explores the performance of SBYOL in the case of extremely limited target training samples and uses the other two best-performing methods, BYOL and SMoCo, as a comparison. A total of five sets of experiments are conducted with training data sizes between 1 and 5 for each class, and the results are shown in Table 4 and Figure 12. In all cases, SBYOL achieved the best results, even with only two samples per class to achieve the best performance of other methods with five samples per class. In the extreme case of one sample per class, the accuracy decreases more due to the deviation from the classification plane, but even so, the best performance is achieved. As can be seen in Figure 12b, the stability of each method increases as the number of data increases, while SBYOL achieves the smallest STD in almost all cases. The results show that SBYOL has strong diagnostic performance and robustness against limited data conditions.
In addition, the noise interference immunity of SBYOL is further validated to demonstrate its robustness and effectiveness under different signal-to-noise ratio (SNR) conditions. The two best-performing methods BYOL and SMoCo are also selected for comparison. A total of six sets of experiments with an SNR from 0 to 10 are conducted for each method using the full training set, i.e., five samples per class, and the results are shown in Table 5 and Figure 13. Compared with the two methods, SBYOL achieves the best results; even in the case of strong noise at 0 dB, it can achieve approximately 94% accuracy, proving its robustness to noise. In addition, as shown in Figure 13b, the overall standard deviation of SBYOL decreases as the noise diminishes, and is lower than the other two methods, which proves that SBYOL has good robustness to noise.

4.3. Verification on the Public Planetary Gearbox Dataset

To verify that the SBYOL pre-trained on parallel gearboxes can well solve the fault diagnosis of planetary gearboxes, the public SEU dataset is selected for further validation in this paper. The SEU dataset [34] provided by Southeast University was obtained on the drivetrain dynamic simulator (DDS), which contains two operating conditions with the rotating speed system load set to 20 Hz—0 V and 30 Hz—2 V, respectively. The test rig is shown in Figure 14. In each file, there are eight rows of vibration signals; in this paper, we use the second row of vibration signals, which means the x-axis vibration signal, and the length of each sample is still 4096. In addition, only 5 samples per class are used in the training set, and 50 samples per class are used in the testing set, and the specific information of the SEU dataset is shown in Table 6.
Similarly, feature extraction was performed on the SEU testing set using the feature extractors obtained by pre-training on the UoC dataset with fixed weights and visualized using T-SNE, and the results are shown in Figure 15. SBYOL can achieve excellent results in terms of extracted features without using any target diagnostic object for training, which is far better than other methods. SMoCo also achieves relatively good feature extraction results, but compared with SBYOL, the same category such as labels 5, 6, 7, 8, and 9 are not aggregated enough, and the two categories of labels 7 and 9 are too close to each other and are more prone to errors when there are limited training data.
The same methods were used for fault diagnosis on the SEU dataset, and the results are shown in Table 7 and Figure 16. The SBYOL method achieves the best diagnostic results in the face of a different device of the planetary gearbox, reaching an accuracy of 99.50%, which greatly exceeds other methods, and the STD of SBYOL is also only 0.47 because of its excellent feature extraction. Although SMoCo and MixMatch also achieve good diagnostic accuracy, they are still far inferior to SBYOL. In addition, since MixMatch needs to be retrained for each new diagnostic task, it is much slower than our method to apply it to the target diagnostic task.
This paper also further explores the performance of SBYOL in the case of extremely limited training samples and selects the two best-performing methods, SMoCo and MixMatch, for comparison, and the results are shown in Table 8 and Figure 17. The best results are obtained for SBYOL with different training sample sizes, even with an accuracy of 98.24% for three samples per class. MixMatch’s performance is greatly degraded under extremely limited labeled datasets due to the lack of our unique data augmentation method. As can be seen in Figure 17b, the stability of each method increases as the number of data increases, while SBYOL achieves the smallest STD in all cases.
In addition, this paper further verifies the ability of SBYOL to resist noise interference. Since MixMatch does not have the unique data augmentation possessed by the self-supervised learning methods, its noise immunity is poor; two methods with good performance and stability, BYOL and SMoCo, are selected for comparison, and the results are shown in Table 9 and Figure 18. Compared to these two methods, SBYOL achieves the best results, even at 4 dB, to reach the performance of the best-performing SMoCo at 10 dB, and its volatility is also the best in most cases. Excessive noise has a greater impact on diagnostic accuracy, but even so, it can achieve an accuracy of about 92% in the case of strong noise at 0 dB.

4.4. Verification on the Bearing Dataset

From the above experiments, the feature extractor obtained by SBYOL pre-training on the parallel gearbox well solves the fault diagnosis problem of planetary gearboxes. Therefore, to further validate that the pre-trained SBYOL learns a universal feature extractor, this section uses the PU bearing data and the PUT aero-engine high-speed bearing data as target diagnostic objects, respectively. The bearing data used have the characteristics of completely different types of equipment from the pre-training gear data.
The bearing dataset of Paderborn University (PU) was presented by Christian Lessmeier et al. [35] in 2016, and the experimental test rig is shown in Figure 19. In this dataset, there are multiple bearings divided into three main groups: 6 healthy bearings, 12 artificially damaged bearings, and 14 bearings with natural operation generating faults. The vibration signals were obtained at a sampling rate of 64 kHz and included four working conditions. It is a very difficult dataset in common datasets, which can reflect the difficulty of real fault diagnosis [37]. Ten types of real damaged bearings in the PU bearing dataset were selected as the target diagnostic dataset to better represent the actual problem, including one type of healthy bearings and two types of mixed faulty bearings. The operating condition is N15_M07_F04; specifically, the rotating speed is 1500 rpm, the loading torque is 0.7 NM, and the radial force is 400 N. The specific information is shown in Table 10. To reflect the limited data problem faced in the actual diagnosis task, 5 samples are used for each class in the training set, and 50 samples are used for each class in the testing set.
Similarly, the pre-trained feature extractors are used to extract features from the PU bearing testing set with fixed weights and visualize them using T-SNE, and the results are shown in Figure 20. The SBYOL proposed in this paper achieves amazing feature extraction results for completely different types of bearing devices without using any data of the target diagnostic object for training, which greatly surpasses other methods.
Next, fault diagnosis was performed on the PU dataset, and the results are shown in Table 11 and Figure 21. When faced with a diagnostic problem for a completely different type of product, SBYOL achieves a diagnostic accuracy of 99.54%, which greatly exceeds other methods. SBYOL is approximately 12 times faster than Labeled Pre-Training and ResNet18, which require training neural networks, and approximately 500 times faster than MixMatch, which uses both a pre-training dataset and a target diagnostic dataset, when applied to diagnostic tasks. While Labeled Pre-Training can still improve accuracy, it is far less effective than SBYOL when faced with a diagnostic problem on a completely different device and without sufficient data for fine-tuning. In addition, since SBYOL achieves very good feature extraction without further training, its std is small, as shown in Figure 21b, proving its excellent stability even in the face of different types of bearing devices.
In this paper, the robustness of SBYOL is likewise verified for extremely limited data volumes and under different noises, and two other best-performing methods, MoCo and SMoCo, are selected for comparison; the results are shown in Table 12 and Table 13 and Figure 22 and Figure 23. SBYOL can achieve very high diagnostic accuracy even when facing the diagnostic problem of a completely different type of product such as a bearing, and it can achieve excellent accuracy in the case of extremely limited data and strong noise, greatly exceeding other methods. SBYOL can achieve a diagnostic accuracy of 94.52% with just one sample per class, and with only three samples per class, it can outperform other methods with five training samples per class. Although SBYOL can achieve good performance with one sample per class, it is prone to deviation from the classification plane in this case, and only two to three samples per class are needed to greatly improve the diagnostic accuracy of SBYOL. In addition, at 0 dB of strong noise, SBYOL even achieves the performance of other best-performing methods at 10 dB, greatly exceeding the performance of other methods. As shown in Figure 22b and Figure 23b, the overall stability of all methods continues to improve as the difficulty of diagnosis decreases, i.e., more training data and less noise.
Since SBYOL can achieve extremely high accuracy in the case of strong noise even in the face of completely different types of bearing devices, we further explored the cases of 0 dB, 4 dB, and 8 dB noise, whose corresponding feature extraction results are shown in Figure 24. With the enhancement of noise, the distance between categories keeps decreasing, and there is a gradual overlap between category KI16 and category KI18, so its diagnostic performance keeps decreasing. However, even with the strong noise of 0 dB, the categories are still well distinguished from each other, and the samples of the same category are aggregated, so it can achieve excellent and stable diagnostic accuracy.
To verify the effectiveness of SBYOL more fully for completely different types of equipment, this paper uses the aero-engine high-speed bearing dataset from the Department of Mechanical and Aeronautical Engineering of the Polytechnic University of Turin (PUT), whose test rig is shown in Figure 25. For this dataset, we used the vibration acceleration data of aero-engine bearings at different speeds and different damage levels, with a sample length of 4096, and used the y-direction channel data at A1. To reflect the extremely limited data situation in the actual diagnosis process, only 3 samples per class were used in the training set, and 50 samples per class were used in the testing set. The specific dataset information is shown in Table 14.
The pre-trained feature extractors are used to extract features from the PUT aero-engine high-speed bearing testing set and are visualized using T-SNE, and the results are shown in Figure 26. The SBYOL method proposed in this paper can achieve such extremely high feature extraction results for the aero-engine high-speed bearing data without using any data of the target diagnostic object for training, which greatly surpasses other methods with large segmentation intervals between categories and the aggregation of data within classes. SMoCo has also achieved good performance, but it is less aggregated than SBYOL for the category labeled 9, and its two categories labeled 7 and 9 are too close together.
Similarly, the fault diagnosis is performed on the PUT aero-engine bearing dataset, and the results are shown in Table 15 and Figure 27. The robustness of SBYOL to the size of the training data and different levels of noise is further validated and compared with the two best-performing methods, and the results are shown in Table 16 and Table 17 and Figure 28 and Figure 29. SBYOL achieves the highest diagnostic accuracy and the smallest STD compared to other methods, even requiring only two samples per class to exceed the performance of SMoCo, which is the best performance among other methods. SBYOL also achieves the best performance in terms of noise immunity, with a diagnostic accuracy of 99.20% at 6 dB, which is comparable to the performance of SimSiam and SMoCo at 10 dB, proving that SBYOL is still robust to noise in the face of completely different devices. In addition, the STD of each method also shows a decreasing trend with the increase in data volume and the decrease in noise.
Table 15. Comparison results on PUT aero-engine bearing dataset.
Table 15. Comparison results on PUT aero-engine bearing dataset.
MethodAccuracy (%)Time (s)
SBYOL99.91 ± 0.122.59
BYOL98.52 ± 1.032.59
SimSiam99.12 ± 1.062.56
SimCLR85.08 ± 2.832.63
MoCo97.94 ± 1.002.62
SMoCo99.60 ± 0.542.60
MixMatch98.18 ± 0.84794.63
Labeled Pre-Training94.06 ± 2.0934.12
FFT + SVM94.86 ± 3.520.11
ResNet1882.83 ± 2.8830.18
Figure 27. Comparison results on PUT aero-engine bearing dataset. (a) Accuracy; (b) standard deviation.
Figure 27. Comparison results on PUT aero-engine bearing dataset. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g027
Table 16. Comparison results on PUT aero-engine bearing dataset under different volumes of training data.
Table 16. Comparison results on PUT aero-engine bearing dataset under different volumes of training data.
MethodsNumber of Samples per Class
123
SBYOL98.86 ± 0.2199.82 ± 0.1599.91 ± 0.12
SimSiam95.82 ± 2.4898.37 ± 1.3599.12 ± 1.06
SMoCo97.28 ± 1.8898.85 ±1.2399.60 ± 0.54
Figure 28. Comparison results on PUT aero-engine bearing dataset under different volumes of training data. (a) Accuracy; (b) standard deviation.
Figure 28. Comparison results on PUT aero-engine bearing dataset under different volumes of training data. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g028
Table 17. Comparison results on PUT aero-engine bearing dataset under different SNRs.
Table 17. Comparison results on PUT aero-engine bearing dataset under different SNRs.
MethodsSNR
0 dB2 dB4 dB6 dB8 dB10 dB
SBYOL96.09 ± 0.7197.83 ± 1.0598.97 ± 0.5599.20 ± 0.7999.58 ± 0.2699.78 ± 0.28
SimSiam92.65 ± 1.3195.98 ± 1.2896.86 ± 1.5598.09 ± 0.8198.67 ± 0.5198.91 ± 0.69
SMoCo96.08 ± 1.2697.71 ± 0.8698.46 ± 0.7798.74 ± 0.7199.26 ± 0.6199.46 ± 0.45
Figure 29. Comparison results on PUT aero-engine bearing dataset under different SNRs. (a) Accuracy; (b) standard deviation.
Figure 29. Comparison results on PUT aero-engine bearing dataset under different SNRs. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g029

4.5. Robustness to Sampling Time and Data Size of the Pre-Training Dataset

To further explore the requirements of SBYOL for pre-training datasets, in addition to the pre-training dataset size, the sampling time is also an influencing factor due to the unique data augmentation method TFST used by SBYOL. Therefore, this section further explores the robustness of SBYOL to the sampling time and data size of the pre-training dataset. Eight sets of experiments are conducted in this paper including SBYOL-9.43s-350, SBYOL-0.42s-350, SBYOL-9.43s-175, SBYOL-0.42s-175, SBYOL-3.86s-70, SBYOL-0.42s-70, SBYOL-2.01s-35, and SBYOL-0.42s-35. Taking SBYOL-9.43s-350 as an example, 9.43s represents the single sampling time, and 350 represents the number of samples per class in the pre-training dataset. These eight sets of experiments are self-supervised learning on the corresponding UoC pre-training datasets, and the learned feature extractors are then transferred to the DPS dataset, SEU dataset, PU dataset, and PUT dataset, respectively, where the training sets in the DPS dataset, SEU dataset, and PU dataset are still five samples per class, and the training sets in the PUT dataset are three samples per class. In addition, we also used the best-performing SMoCo and MixMatch as a comparison, which used all unlabeled pre-training datasets, i.e., 350 samples per class, and the results are shown in Table 18 and Figure 30.
SBYOL still achieves good performance with a smaller sampling time and pre-training dataset size, and the average diagnostic accuracy of SBYOL-0.42s-175 even reaches 98.23% on the four datasets. In addition, SBYOL-0.42s-35 achieves an average diagnostic accuracy of 97.49%, surpassing SMoCo and MixMatch, the best performers among other methods, while only using 10% of the dataset size compared to SMoCo and MixMatch. From comparing SBYOL-9.43s-350 with SBYOL-0.42s-350, SBYOL-9.43s-175 with SBYOL-0.42s-175, SBYOL-3.86s-70 with SBYOL-0.42s-70, and SBYOL-2.01s-35 with SBYOL-0.42s-35, with the same amount of data, a longer sampling signal allows SBYOL to perform better on average over the four datasets. From the comparison of SBYOL-0.42s-350 with SBYOL-9.43s-175, SBYOL-0.42s-175 with SBYOL-3.86s-70, and SBYOL-0.42s-70 with SBYOL-2.01s-35, SBYOL can perform better with a larger amount of data even with shorter sampling time. The performance is closer in the case of 70 and 35 samples per class compared to the difference between 350, 175, and 70 samples per class, proving that it still requires a larger variation in data volume for a larger improvement in performance.
Therefore, from the above results, SBYOL is extremely robust to the pre-training dataset, longer sampling time and larger datasets can achieve better performance, and dataset size improves performance more than sampling time when there is a large difference in dataset size.

5. Conclusions

The problem of limited data and strong noise in the rotating machinery fault diagnosis tasks under real conditions seriously affects the performance of intelligent diagnosis methods. In this paper, we innovatively propose SBYOL for the automatic feature extraction of unlabeled rotating machinery vibration signals, which incorporates the novel data augmentation method TFST proposed in this paper to retain both time domain and frequency domain information of the signal at the same time. SBYOL first performs self-supervised learning on easily available unlabeled data and then uses the trained feature extractor for downstream diagnostic tasks under limited data. In this paper, a powerful and robust universal feature extractor was constructed by self-supervised pre-training on the unlabeled UoC parallel gearbox dataset and then applied to a private DPS planetary gearbox dataset under limited data, achieving an accuracy of 97.14%. It was further validated on the SEU planetary gearbox dataset and achieved an accuracy of 99.50%. To further demonstrate that SBYOL learned a universal feature extractor, two public bearing datasets, the PU and PUT bearing datasets, with completely different types of devices compared to the pre-training dataset, were used as target diagnostic objects, both achieving diagnostic accuracy of over 99.54%. Further experiments show that SBYOL also has excellent performance for the target diagnostic object under extremely limited data and strong noise and is robust to the unlabeled pre-training dataset with a shorter sampling time and smaller data size, demonstrating its great potential and superiority for rotating machinery fault diagnosis problems.
Although SBYOL achieved excellent results, there are still some works that deserve further exploration. In the self-supervised signal representation learning stage, SBYOL takes a relatively long time to learn a good representation of the signal, and future research can be conducted on how to improve the training efficiency, such as using mixed precision training. For the structure of the encoder, future work can also try to use a transformer network that currently performs well in various fields. Finally, although SBYOL showed promising results under steady-state operating conditions, there is a lack of exploration regarding fault diagnosis under varying operating conditions. Therefore, future research can further investigate and explore more data transformation methods that are suitable for handling varying operating conditions [41,42].

Author Contributions

Conceptualization, Z.Y. and H.L.; methodology, Z.Y. and H.L.; software, Z.Y.; validation, Z.Y.; formal analysis, J.M. and Y.C.; investigation, Z.Y. and Y.C.; resources, L.T.; data curation, L.T.; writing—original draft preparation, Z.Y. and H.L.; writing—review and editing, H.L., L.T., J.M. and Y.C.; visualization, J.M.; supervision, H.L. and L.T.; project administration, H.L.; funding acquisition, H.L., L.T., J.M. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Nos. 61973011 and 61903015), Aeronautical Science Foundation of China (Grant No. ASFC-201933051001), Fundamental Research Funds for the Central Universities (Grant Nos. KG21002901, KG21003001, and YWF-22-L-516), National key Laboratory of Science and Technology on Reliability and Environmental Engineering (Grant No. WDZC2019601A304), Capital Science & Technology Leading Talent Program (Grant No. Z191100006119029), and Science and Technology Foundation of State Key Laboratory (grant number 6142004200501).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Khan, S.; Yairi, T. A review on the application of deep learning in system health management. Mech. Syst. Signal Process. 2018, 107, 241–265. [Google Scholar] [CrossRef]
  2. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  3. Chen, H.; Meng, W.; Li, Y.; Xiong, Q. An anti-noise fault diagnosis approach for rolling bearings based on multiscale CNN-LSTM and a deep residual learning model. Meas. Sci. Technol. 2023, 34, 045013. [Google Scholar] [CrossRef]
  4. Mian, T.; Choudhary, A.; Fatima, S. Vibration and infrared thermography based multiple fault diagnosis of bearing using deep learning. Nondestruct. Test. Eval. 2023, 38, 275–296. [Google Scholar] [CrossRef]
  5. Bai, R.; Xu, Q.; Meng, Z.; Cao, L.; Xing, K.; Fan, F. Rolling bearing fault diagnosis based on multi-channel convolution neural network and multi-scale clipping fusion data augmentation. Measurement 2021, 184, 109885. [Google Scholar] [CrossRef]
  6. Feng, Y.; Chen, J.; Zhang, T.; He, S.; Xu, E.; Zhou, Z. Semi-supervised meta-learning networks with squeeze-and-excitation attention for few-shot fault diagnosis. ISA Trans. 2022, 120, 383–401. [Google Scholar] [CrossRef]
  7. Wu, X.; Zhang, Y.; Cheng, C.; Peng, Z. A hybrid classification autoencoder for semi-supervised fault diagnosis in rotating machinery. Mech. Syst. Signal Process. 2021, 149, 107327. [Google Scholar] [CrossRef]
  8. Zhao, X.; Jia, M.; Liu, Z. Semisupervised Deep Sparse Auto-Encoder with Local and Nonlocal Information for Intelligent Fault Diagnosis of Rotating Machinery. IEEE Trans. Instrum. Meas. 2021, 70, 1–13. [Google Scholar] [CrossRef]
  9. Li, J.; Huang, R.; Li, W. Intelligent Fault Diagnosis for Bearing Dataset Using Adversarial Transfer Learning based on Stacked Auto-Encoder. Procedia Manuf. 2020, 49, 75–80. [Google Scholar] [CrossRef]
  10. Han, T.; Liu, C.; Yang, W.; Jiang, D. Deep transfer network with joint distribution adaptation: A new intelligent fault diagnosis framework for industry application. ISA Trans. 2020, 97, 269–281. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Zheng, H.; Wang, R.; Yang, Y.; Yin, J.; Li, Y.; Li, Y.; Xu, M. Cross-Domain Fault Diagnosis Using Knowledge Transfer Strategy: A Review. IEEE Access 2019, 7, 129260–129290. [Google Scholar] [CrossRef]
  12. Jing, L.; Tian, Y. Self-Supervised Visual Feature Learning with Deep Neural Networks: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 4037–4058. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, R.; Isola, P.; Efros, A.A. Colorful Image Colorization. In Computer Vision–ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2016; Volume 9907, pp. 649–666. [Google Scholar] [CrossRef] [Green Version]
  14. Pathak, D.; Krähenbühl, P.; Donahue, J.; Darrell, T.; Efros, A.A. Context Encoders: Feature Learning by Inpainting. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27−30 June 2016; pp. 2536–2544. [Google Scholar] [CrossRef] [Green Version]
  15. Noroozi, M.; Favaro, P. Unsupervised Learning of Visual Representations by Solving Jigsaw Puzzles. In Computer Vision–ECCV 2016; Leibe, B., Matas, J., Sebe, N., Welling, M., Eds.; Lecture Notes in Computer Science 9910; Springer International Publishing: Cham, Switzerland, 2016; pp. 69–84. [Google Scholar] [CrossRef] [Green Version]
  16. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.P.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 105–114. [Google Scholar] [CrossRef] [Green Version]
  17. Rani, V.; Nabi, S.T.; Kumar, M.; Mittal, A.; Kumar, K. Self-supervised Learning: A Succinct Review. Arch. Comput. Methods Eng. 2023, 30, 2761–2775. [Google Scholar] [CrossRef] [PubMed]
  18. Grill, J.-B.; Strub, F.; Altché, F.; Tallec, C.; Richemond, P.H.; Buchatskaya, E.; Doersch, C.; Pires, B.A.; Guo, Z.D.; Azar, M.G.; et al. Bootstrap your own latent: A new approach to self-supervised Learning. arXiv 2020, arXiv:2006.07733. [Google Scholar]
  19. Tian, Y.; Sun, C.; Poole, B.; Krishnan, D.; Schmid, C.; Isola, P. What Makes for Good Views for Contrastive Learning? In Advances in Neural Information Processing Systems; Curran Associates Inc.: Vancouver, BC, Canada, 2020; pp. 6827–6839. Available online: https://papers.nips.cc/paper/2020/hash/4c2e5eaae9152079b9e95845750bb9ab-Abstract.html (accessed on 19 April 2022).
  20. Chen, T.; Kornblith, S.; Norouzi, M.; Hinton, G. A Simple Framework for Contrastive Learning of Visual Representations. In Proceedings of the 37th International Conference on Machine Learning, PMLR, November 2020, 1597–1607; Available online: https://proceedings.mlr.press/v119/chen20j.html (accessed on 18 April 2022).
  21. Chen, T.; Kornblith, S.; Swersky, K.; Norouzi, M.; Hinton, G.E. Big Self-Supervised Models are Strong Semi-Supervised Learners. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Vancouver, BC, Canada, 2020; pp. 22243–22255. Available online: https://proceedings.neurips.cc/paper/2020/hash/fcbc95ccdd551da181207c0c1400c655-Abstract.html (accessed on 19 April 2022).
  22. Wang, H.; Liu, Z.; Ge, Y.; Peng, D. Self-supervised signal representation learning for machinery fault diagnosis under limited annotation data. Knowl.-Based Syst. 2022, 239, 107978. [Google Scholar] [CrossRef]
  23. Ding, Y.; Zhuang, J.; Ding, P.; Jia, M. Self-supervised pretraining via contrast learning for intelligent incipient fault detection of bearings. Reliab. Eng. Syst. Saf. 2022, 218, 108126. [Google Scholar] [CrossRef]
  24. He, K.; Fan, H.; Wu, Y.; Xie, S.; Girshick, R. Momentum Contrast Unsupervised Vis. Represent. Learning. arxiv 2020, arXiv:1911.05722. [Google Scholar]
  25. Wei, M.; Liu, Y.; Zhang, T.; Wang, Z.; Zhu, J. Fault Diagnosis of Rotating Machinery Based on Improved Self-Supervised Learning Method and Very Few Labeled Samples. Sensors 2021, 22, 192. [Google Scholar] [CrossRef]
  26. Yan, Z.; Liu, H. SMoCo: A Powerful and Efficient Method Based on Self-Supervised Learning for Fault Diagnosis of Aero-Engine Bearing under Limited Data. Mathematics 2022, 10, 2796. [Google Scholar] [CrossRef]
  27. Shul, Y.; Yi, W.; Choi, J.; Kang, D.-S.; Choi, J.-W. Noise-based self-supervised anomaly detection in washing machines using a deep neural network with operational information. Mech. Syst. Signal Process. 2023, 189, 110102. [Google Scholar] [CrossRef]
  28. Nie, G.; Zhang, Z.; Shao, M.; Jiao, Z.; Li, Y.; Li, L. A Novel Study on a Generalized Model Based on Self-Supervised Learning and Sparse Filtering for Intelligent Bearing Fault Diagnosis. Sensors 2023, 23, 1858. [Google Scholar] [CrossRef]
  29. Jaiswal, A.; Babu, A.R.; Zadeh, M.Z.; Banerjee, D.; Makedon, F. A Survey on Contrastive Self-Supervised Learning. Technologies 2020, 9, 2. [Google Scholar] [CrossRef]
  30. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  31. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  32. Cao, P.; Zhang, S.; Tang, J. Gear Fault Data. figshare, April 11 2018. Available online: https://figshare.com/articles/dataset/Gear_Fault_Data/6127874/1 (accessed on 22 March 2022).
  33. Cao, P.; Zhang, S.; Tang, J. Preprocessing-Free Gear Fault Diagnosis Using Small Datasets with Deep Convolutional Neural Network-Based Transfer Learning. IEEE Access 2018, 6, 26241–26253. [Google Scholar] [CrossRef]
  34. Shao, S.; McAleer, S.; Yan, R.; Baldi, P. Highly Accurate Machine Fault Diagnosis Using Deep Transfer Learning. IEEE Trans. Ind. Inf. 2019, 15, 2446–2455. [Google Scholar] [CrossRef]
  35. Lessmeier, C.; Kimotho, J.K.; Zimmer, D.; Sextro, W. Condition Monitoring of Bearing Damage in Electromechanical Drive Systems by Using Motor Current Signals of Electric Motors: A Benchmark Data Set for Data-Driven Classification. PHM Soc. Eur. Conf. 2016, 3, 1. [Google Scholar] [CrossRef]
  36. Daga, A.P.; Fasana, A.; Marchesiello, S.; Garibaldi, L. The Politecnico di Torino rolling bearing test rig: Description and analysis of open access data. Mech. Syst. Signal Process. 2019, 120, 252–273. [Google Scholar] [CrossRef]
  37. Zhao, Z.; Li, T.; Wu, J.; Sun, C.; Wang, S.; Yan, R.; Chen, X. Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Trans. 2020, 107, 224–255. [Google Scholar] [CrossRef]
  38. Zhang, X.; He, C.; Lu, Y.; Chen, B.; Zhu, L.; Zhang, L. Fault diagnosis for small samples based on attention mechanism. Measurement 2022, 187, 110242. [Google Scholar] [CrossRef]
  39. Chen, X.; He, K. Exploring Simple Siamese Representation Learning. arXiv 2020, arXiv:2011.10566. [Google Scholar]
  40. Berthelot, D.; Carlini, N.; Goodfellow, I.; Papernot, N.; Oliver, A.; Raffel, C.A. MixMatch: A Holistic Approach to Semi-Supervised Learning. In Advances in Neural Information Processing Systems; Curran Associates Inc.: Vancouver, BC, Canada, 2019; Available online: https://proceedings.neurips.cc/paper/2019/hash/1cd138d0499a68f4bb72bee04bbec2d7-Abstract.html (accessed on 21 June 2022).
  41. Luo, Z.; Tan, H.; Dong, X.; Zhu, G.; Li, J. A fault diagnosis method for rotating machinery with variable speed based on multi-feature fusion and improved ShuffleNet V2. Meas. Sci. Technol. 2023, 34, 035110. [Google Scholar] [CrossRef]
  42. Zhang, K.; Wang, J.; Shi, H.; Zhang, X. A Variable Working Condition Rolling Bearing Fault Diagnosis Method Based on Improved Triplet Loss Algorithm. Int. J. Control Autom. Syst. 2023, 21, 1361–1372. [Google Scholar] [CrossRef]
Figure 1. Flow chart of fault diagnosis based on SBYOL.
Figure 1. Flow chart of fault diagnosis based on SBYOL.
Aerospace 10 00681 g001
Figure 2. Examples of six basic data augmentation methods: (a) Normalize; (b) AddGaussian; (c) Scale; (d) Stretch; (e) Crop; (f) Flip.
Figure 2. Examples of six basic data augmentation methods: (a) Normalize; (b) AddGaussian; (c) Scale; (d) Stretch; (e) Crop; (f) Flip.
Aerospace 10 00681 g002
Figure 3. The framework of TFST.
Figure 3. The framework of TFST.
Aerospace 10 00681 g003
Figure 4. The network structure of SBYOL with ResNet-18 encoder.
Figure 4. The network structure of SBYOL with ResNet-18 encoder.
Aerospace 10 00681 g004
Figure 5. Test rig of UoC dataset [38].
Figure 5. Test rig of UoC dataset [38].
Aerospace 10 00681 g005
Figure 6. Loss curve in the self-supervised signal representation learning stage.
Figure 6. Loss curve in the self-supervised signal representation learning stage.
Aerospace 10 00681 g006
Figure 7. The visualization of pre-trained feature extractors on unlabeled UoC gear dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Figure 7. The visualization of pre-trained feature extractors on unlabeled UoC gear dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Aerospace 10 00681 g007
Figure 8. The DPS test rig. (a) Tset rig; (b) End cap vibration sensor; (c) Box vibration sensor.
Figure 8. The DPS test rig. (a) Tset rig; (b) End cap vibration sensor; (c) Box vibration sensor.
Aerospace 10 00681 g008
Figure 9. Gears with four failure modes in the DPS test rig. (a) Wear; (b) broken tooth; (c) missing tooth; (d) root crack.
Figure 9. Gears with four failure modes in the DPS test rig. (a) Wear; (b) broken tooth; (c) missing tooth; (d) root crack.
Aerospace 10 00681 g009
Figure 10. The visualization of pre-trained feature extractors on the DPS gear dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Figure 10. The visualization of pre-trained feature extractors on the DPS gear dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Aerospace 10 00681 g010
Figure 11. Comparison results on DPS gear dataset. (a) Accuracy; (b) standard deviation.
Figure 11. Comparison results on DPS gear dataset. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g011
Figure 12. Comparison results on DPS gear dataset under different volumes of training data. (a) Accuracy; (b) standard deviation.
Figure 12. Comparison results on DPS gear dataset under different volumes of training data. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g012
Figure 13. Comparison results on DPS gear dataset under different SNRs. (a) Accuracy; (b) standard deviation.
Figure 13. Comparison results on DPS gear dataset under different SNRs. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g013
Figure 14. Test rig of SEU dataset [34].
Figure 14. Test rig of SEU dataset [34].
Aerospace 10 00681 g014
Figure 15. The visualization of pre-trained feature extractors on the SEU gear dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Figure 15. The visualization of pre-trained feature extractors on the SEU gear dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Aerospace 10 00681 g015
Figure 16. Comparison results on SEU gear dataset. (a) Accuracy; (b) standard deviation.
Figure 16. Comparison results on SEU gear dataset. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g016
Figure 17. Comparison results on SEU gear dataset under different volumes of training data. (a) Accuracy; (b) standard deviation.
Figure 17. Comparison results on SEU gear dataset under different volumes of training data. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g017
Figure 18. Comparison results on SEU gear dataset under different SNRs. (a) Accuracy; (b) standard deviation.
Figure 18. Comparison results on SEU gear dataset under different SNRs. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g018
Figure 19. Test rig of PU dataset. (a) Electric motor; (b) torque-measurement shaft; (c) rolling bearing test module; (d) flywheel; (e) load motor.
Figure 19. Test rig of PU dataset. (a) Electric motor; (b) torque-measurement shaft; (c) rolling bearing test module; (d) flywheel; (e) load motor.
Aerospace 10 00681 g019
Figure 20. The visualization of pre-trained feature extractors on the PU bearing dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Figure 20. The visualization of pre-trained feature extractors on the PU bearing dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Aerospace 10 00681 g020
Figure 21. Comparison results on PU bearing dataset. (a) Accuracy; (b) standard deviation.
Figure 21. Comparison results on PU bearing dataset. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g021
Figure 22. Comparison results on PU bearing dataset under different volumes of training data. (a) Accuracy; (b) standard deviation.
Figure 22. Comparison results on PU bearing dataset under different volumes of training data. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g022
Figure 23. Comparison results on PU bearing dataset under different SNRs. (a) Accuracy; (b) standard deviation.
Figure 23. Comparison results on PU bearing dataset under different SNRs. (a) Accuracy; (b) standard deviation.
Aerospace 10 00681 g023
Figure 24. Visualization of SBYOL on noisy PU bearing datasets. (a) 0 dB; (b) 4 dB; (c) 8 dB.
Figure 24. Visualization of SBYOL on noisy PU bearing datasets. (a) 0 dB; (b) 4 dB; (c) 8 dB.
Aerospace 10 00681 g024
Figure 25. Test rig of the aero-engine bearing dataset from the Polytechnic University of Turin [36].
Figure 25. Test rig of the aero-engine bearing dataset from the Polytechnic University of Turin [36].
Aerospace 10 00681 g025
Figure 26. The visualization of pre-trained feature extractors on the PUT aero-engine bearing dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Figure 26. The visualization of pre-trained feature extractors on the PUT aero-engine bearing dataset. (a) SBYOL; (b) Labeled Pre-Training; (c) SimSiam; (d) SimCLR; (e) BYOL; (f) MoCo; (g) SMoCo.
Aerospace 10 00681 g026
Figure 30. Comparison results of different sampling times and pre-training dataset sizes.
Figure 30. Comparison results of different sampling times and pre-training dataset sizes.
Aerospace 10 00681 g030
Table 1. Hyperparameter setting.
Table 1. Hyperparameter setting.
HyperparameterValueData AugmentationValue
Batch size64Normalization/
OptimizerStochastic gradient descent (SGD)AddGaussianNoise coefficient σ n = 0.05
Learning rate0.1ScaleScale coefficient σ s = 0.05
Momentum0.9StretchStretch coefficient σ s = 0.3
Weight decay1 × 10−4CropCrop length = 100
Epochs3000Flip/
Learning rate scheduleCosine
Update parameter τ 0.996
Table 2. Labeled DPS gear dataset.
Table 2. Labeled DPS gear dataset.
Damaged ElementRotation Speed (Hz)Load (Nm)Training SamplesTesting SamplesLabel
Healthy601.25500
Healthy400.65501
Wear601.25502
Wear400.65503
Broken teeth601.25504
Broken teeth400.65505
Missing teeth601.25506
Missing teeth400.65507
Root crack601.25508
Root crack400.65509
Table 3. Comparison results on DPS gear dataset.
Table 3. Comparison results on DPS gear dataset.
MethodAccuracy (%)Time (s)
SBYOL97.14 ± 1.602.43
BYOL91.82 ± 2.332.39
SimSiam90.94 ± 4.002.41
SimCLR78.98 ± 2.412.41
MoCo90.08 ± 3.302.46
SMoCo92.80 ± 3.192.37
MixMatch90.32 ± 2.181177.21
Labeled Pre-Training86.11 ± 3.1427.42
FFT + SVM88.48 ± 1.820.10
ResNet1871.84 ± 3.9225.86
Table 4. Comparison results on DPS gear dataset under different volumes of training data.
Table 4. Comparison results on DPS gear dataset under different volumes of training data.
MethodsNumber of Samples per Class
12345
SBYOL89.40 ± 3.6492.70 ± 3.3394.20 ± 3.0695.34 ± 2.9297.14 ± 1.60
BYOL85.36 ± 4.2188.09 ± 4.2289.16 ± 4.1090.86 ± 3.5691.82 ± 2.33
SMoCo85.58 ± 4.6689.40 ± 3.0390.94 ± 3.6292.20 ± 3.1292.80 ± 3.19
Table 5. Comparison results on DPS gear dataset under different SNRs.
Table 5. Comparison results on DPS gear dataset under different SNRs.
MethodsSNR
0 dB2 dB4 dB6 dB8 dB10 dB
SBYOL93.90 ± 3.0794.20 ± 2.9294.62 ± 2.7695.22 ± 2.6895.94 ± 2.2996.52 ± 1.59
BYOL88.18 ± 3.9488.70 ± 3.5388.96 ± 3.2190.04 ± 3.7990.90 ± 2.9491.48 ± 2.70
SMoCo90.04 ± 3.5090.58 ± 3.3290.90 ± 3.1591.26 ± 2.7291.84 ± 2.9392.40 ± 2.49
Table 6. Labeled SEU gear dataset.
Table 6. Labeled SEU gear dataset.
Fault ModeRotating Speed System LoadTraining SamplesTesting SamplesLabel
Health Gear20 Hz—0 V5500
Health Gear30 Hz—2 V5501
Chipped Tooth20 Hz—0 V5502
Chipped Tooth30 Hz—2 V5503
Missing Tooth20 Hz—0 V5504
Missing Tooth30 Hz—2 V5505
Root Fault20 Hz—0 V5506
Root Fault30 Hz—2 V5507
Surface Fault20 Hz—0 V5508
Surface Fault30 Hz—2 V5509
Table 7. Comparison results on SEU gear dataset.
Table 7. Comparison results on SEU gear dataset.
MethodAccuracy (%)Time (s)
SBYOL99.50 ± 0.472.35
BYOL95.82 ± 1.292.25
SimSiam90.27 ± 2.262.61
SimCLR86.52 ± 1.942.55
MoCo93.28 ± 1.032.56
SMoCo97.36 ± 1.622.56
MixMatch97.00 ± 1.191182.17
Labeled Pre-Training87.25 ± 3.2630.56
FFT + SVM88.19 ± 5.030.11
ResNet1872.48 ± 6.6728.86
Table 8. Comparison results on SEU gear dataset under different volumes of training data.
Table 8. Comparison results on SEU gear dataset under different volumes of training data.
MethodsNumber of Samples per Class
12345
SBYOL92.38 ± 2.2496.46 ± 2.3598.24 ± 0.8999.14 ± 0.6199.50 ± 0.47
SMoCo82.84 ± 2.3691.04 ± 2.9195.12 ± 2.0896.80 ±1.2497.36 ± 1.62
MixMatch74.12 ± 6.0983.57 ± 4.0885.60 ± 3.8594.57 ± 3.9897.00 ± 1.19
Table 9. Comparison results on SEU gear dataset under different SNRs.
Table 9. Comparison results on SEU gear dataset under different SNRs.
MethodsSNR
0 dB2 dB4 dB6 dB8 dB10 dB
SBYOL91.96 ± 1.8994.72 ± 1.3096.30 ± 1.0097.44 ± 0.7198.32 ± 0.5898.70 ± 0.92
BYOL86.30 ± 1.3089.26 ± 1.6692.06 ± 1.9294.16 ± 1.03 94.26 ± 1.3694.96 ± 1.02
SMoCo89.62 ± 1.8393.00 ± 2.4795.04 ± 1.2495.62 ± 1.9896.14 ± 2.0896.82 ± 1.58
Table 10. Labeled PU bearing dataset.
Table 10. Labeled PU bearing dataset.
Bearing CodeDamaged ElementFault ModeDamage FormArrangementDamaged Extent
K001Health state////
KA04Outer ringFatigue: pittingSingle damageNo repetitionLevel 1
KA15Outer ringPlastic deform: IndentationsSingle damageNo repetitionLevel 1
KA16Outer ringFatigue: pittingRepetitive damageRandomLevel 2
KB23Outer ring and inner ringFatigue: pittingMultiple damageRandomLevel 2
KB24Outer ring and inner ringFatigue: pittingMultiple damageNo repetitionLevel 3
KI14Outer ringFatigue: pittingMultiple damageNo repetitionLevel 1
KI16Outer ringFatigue: pittingSingle damageNo repetitionLevel 3
KI17Inner ringFatigue: pittingRepetitive damageRandomLevel 1
KI18Inner ringFatigue: pittingSingle damageNo repetitionLevel 2
Table 11. Comparison results on PU bearing dataset.
Table 11. Comparison results on PU bearing dataset.
MethodAccuracy (%)Time (s)
SBYOL99.54 ± 0.382.38
BYOL89.28 ± 2.482.39
SimSiam95.64 ± 0.942.37
SimCLR93.94 ± 1.522.35
MoCo97.22 ± 0.992.37
SMoCo98.06 ± 1.492.39
MixMatch94.68 ± 1.081186.80
Labeled Pre-Training87.64 ± 3.2631.35
FFT + SVM79.62 ± 4.850.10
ResNet1871.52 ± 5.7429.04
Table 12. Comparison results on PU bearing dataset under different volumes of training data.
Table 12. Comparison results on PU bearing dataset under different volumes of training data.
MethodsNumber of Samples per Class
12345
SBYOL94.52 ± 2.2497.66 ± 1.2198.50 ± 0.9099.14 ± 0.9199.54 ± 0.38
SMoCo85.64 ± 4.0192.92 ± 2.5995.28 ± 1.7696.56 ± 1.2397.22 ± 0.99
MixMatch81.48 ± 4.1789.16 ± 3.9895.32 ± 1.6096.52 ± 2.1298.06 ± 1.49
Table 13. Comparison results on PU bearing dataset under different SNRs.
Table 13. Comparison results on PU bearing dataset under different SNRs.
MethodsSNR
0 dB2 dB4 dB6 dB8 dB10 dB
SBYOL97.08 ± 1.3798.20 ± 0.5898.68 ± 0.8298.84 ± 0.4899.14 ± 0.3999.36 ± 0.31
BYOL94.94 ± 1.0795.42 ± 1.0895.94 ± 1.0496.32 ± 1.1396.90 ± 1.0897.22 ± 1.02
SMoCo94.54 ± 2.6895.38 ± 2.1695.54 ± 1.6295.82 ± 1.2296.10 ± 2.1997.60 ± 1.00
Table 14. Labeled PUT aero-engine bearing dataset.
Table 14. Labeled PUT aero-engine bearing dataset.
Damaged ElementDiameter (μm)Rotation Speed (r/min)Load (N)Training SamplesTesting SamplesLabel
Healthy/24,00014003500
Inner ring45024,00014003501
Inner ring25024,00014003502
Inner ring15024,00014003503
Roller45024,00014003504
Roller25024,00014003505
Roller15024,00014003506
Inner ring45018,00014003507
Inner ring25018,00014003508
Inner ring15018,00014003509
Roller45018,000140035010
Roller25018,000140035011
Roller15018,000140035012
Table 18. Comparison results of different sampling times and pre-training dataset sizes.
Table 18. Comparison results of different sampling times and pre-training dataset sizes.
MethodIntercepted Sample SizeData SizeDPSSEUPUPUTAverage
SBYOL-9.43s-350175350 × 997.1499.5099.5499.9199.02
SBYOL-0.42s-3505350 × 997.2899.6199.4299.5898.97
SBYOL-9.43s-175175175 × 996.6898.6999.4999.7698.66
SBYOL-0.42s-1755175 × 995.9097.7099.5399.8098.23
SBYOL-3.86s-707070 × 993.7897.5999.2299.8597.61
SBYOL-0.42s-70570 × 993.6097.0899.5299.8297.51
SBYOL-2.01s-353535 × 993.4797.6898.8899.9797.50
SBYOL-0.42s-35535 × 993.5798.3798.3399.6997.49
SMoCo/350 × 992.8097.3698.0699.6096.96
MixMatch/350 × 990.3297.0094.6898.1895.05
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yan, Z.; Liu, H.; Tao, L.; Ma, J.; Cheng, Y. A Universal Feature Extractor Based on Self-Supervised Pre-Training for Fault Diagnosis of Rotating Machinery under Limited Data. Aerospace 2023, 10, 681. https://doi.org/10.3390/aerospace10080681

AMA Style

Yan Z, Liu H, Tao L, Ma J, Cheng Y. A Universal Feature Extractor Based on Self-Supervised Pre-Training for Fault Diagnosis of Rotating Machinery under Limited Data. Aerospace. 2023; 10(8):681. https://doi.org/10.3390/aerospace10080681

Chicago/Turabian Style

Yan, Zitong, Hongmei Liu, Laifa Tao, Jian Ma, and Yujie Cheng. 2023. "A Universal Feature Extractor Based on Self-Supervised Pre-Training for Fault Diagnosis of Rotating Machinery under Limited Data" Aerospace 10, no. 8: 681. https://doi.org/10.3390/aerospace10080681

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop