Next Article in Journal
Contextual Fine-Tuning of Language Models with Classifier-Driven Content Moderation for Text Generation
Previous Article in Journal
Network Coding-Enhanced Polar Codes for Relay-Assisted Visible Light Communication Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sample Augmentation Using Enhanced Auxiliary Classifier Generative Adversarial Network by Transformer for Railway Freight Train Wheelset Bearing Fault Diagnosis

1
School of Traffic and Transportation, Shijiazhuang Tiedao University, Shijiazhuang 050043, China
2
Hebei Province University Road Traffic Perception and Intelligent Application Technology Research and Development Center, Hebei Jiaotong Vocational and Technical College, Shijiazhuang 050011, China
3
School of Computer Science, South China Business College Guangdong University of Foreign Studies, Guangzhou 510545, China
4
College of Career Technology, Hebei Normal University, Shijiazhuang 050043, China
5
State Grid Shijiazhuang Electric Power Supply Company, Shijiazhuang 050021, China
6
School of Electrical and Electronic Engineering, Shijiazhuang Tiedao University, Shijiazhuang 050043, China
7
Hebei Provincial Collaborative Innovation Center of Transportation Power Grid Intelligent Integration Technology and Equipment, Shijiazhuang Tiedao University, Shijiazhuang 050043, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Entropy 2024, 26(12), 1113; https://doi.org/10.3390/e26121113
Submission received: 28 October 2024 / Revised: 21 November 2024 / Accepted: 12 December 2024 / Published: 20 December 2024
(This article belongs to the Section Multidisciplinary Applications)

Abstract

:
Diagnosing faults in wheelset bearings is critical for train safety. The main challenge is that only a limited amount of fault sample data can be obtained during high-speed train operations. This scarcity of samples impacts the training and accuracy of deep learning models for wheelset bearing fault diagnosis. Studies show that the Auxiliary Classifier Generative Adversarial Network (ACGAN) demonstrates promising performance in addressing this issue. However, existing ACGAN models have drawbacks such as complexity, high computational expenses, mode collapse, and vanishing gradients. Aiming to address these issues, this paper presents the Transformer and Auxiliary Classifier Generative Adversarial Network (TACGAN), which increases the diversity, complexity and entropy of generated samples, and maximizes the entropy of the generated samples. The transformer network replaces traditional convolutional neural networks (CNNs), avoiding iterative and convolutional structures, thereby reducing computational expenses. Moreover, an independent classifier is integrated to prevent the coupling problem, where the discriminator is simultaneously identified and classified in the ACGAN. Finally, the Wasserstein distance is employed in the loss function to mitigate mode collapse and vanishing gradients. Experimental results using the train wheelset bearing datasets demonstrate the accuracy and effectiveness of the TACGAN.

1. Introduction

As a crucial part of train bogies, wheelset bearings are vital for load transmission and motion conversion [1]. These bearings endure various alternating loads during train operations. Once a component’s surface sustains local damages, the fault can progressively worsen under dynamic loads, leading to a decline in the functionality of the entire transmission system and potentially causing significant accidents [2,3]. Thus, online detection and fault diagnosis of wheelset bearings are essential.
Traditional fault diagnosis methods primarily rely on signal processing to analyze bearing faults by denoising and extracting fault features from the vibration signals of wheelset bearings [4,5]. By comparing frequencies derived from the dimensions of bearing components in different damaged areas, the health status and location of damaged parts are determined. In this context, Cao H et al. [6] utilized the empirical wavelet transform (EWT) for vibration signal analysis in fault diagnosis. Moreover, Li H et al. [7] introduced an improved ensemble empirical mode decomposition (EEMD) time–frequency analysis method based on the Improved Adaptive Resonance Technique (IART), enhancing the denoising capability of fault-related pulse signals and effectively detecting rolling bearing faults. Zheng J et al. [8] introduced a rolling bearing fault diagnosis method based on composite multiscale fuzzy entropy (CMFE) and ensemble support vector machines (ESVMs), which can effectively classify fault categories and severity levels of rolling bearings. Yu W et al. [9] introduced a probabilistic ensemble learning-based Bayesian network (PEL-BN) strategy for online fault diagnosis technology. Jiao J et al. [10] introduced a favorable Sparse Local FDA (SLFDA) model, which utilizes the local data structure characteristics from both sample and variable dimensions, and significantly enhances fault performance and model explainability.
In the era of big data, deep learning exhibits unique feature extraction capabilities. Deep learning can automatically extract useful features from samples, perform end-to-end fault classification, and avoid dependence on signal-processing techniques and prior knowledge. Techniques such as convolutional neural networks (CNNs) [11,12], recurrent neural networks (RNNs) [13,14], generative adversarial networks (GANs) [15], and autoencoders (AEs) [16,17,18] are employed in this context. For instance, Yu W et al. [19] utilized a Broad Convolutional Neural Network (BCNN) with incremental learning capability, which can better capture the characteristics of the fault process and effectively update diagnostic model. N. Daldal et al. [20] utilized two CNNs with different kernel sizes for automatic signal feature extraction and employed long-term and short-term memory networks to determine fault types. Furthermore, Yang et al. [21] introduced the CGAN-2-D-CNN fusion diagnostic method for diagnosing small bearing faults.
Despite the promising results achieved by wheel bearing fault classification algorithms, several challenges persist. First, these methods require labeled samples, and it is often impractical to label wheel-bearing fault samples during collection [22,23]. Second, the methods need a sufficient and balanced number of trained samples; otherwise, the diagnostic accuracy may decrease or overfitting may occur [24]. To address the issue of unbalanced samples in bearing fault diagnosis, data augmentation methods (DAMs) like rotation, translation, scaling, flipping, and cropping are often employed to expand data samples. Li W et al. [25,26] proposed an enhanced GAN model (MACGAN) with a new framework design, incorporating a classification function to develop a new ACGAN framework. The ACGAN structure includes a CNN, providing discriminative and classification capabilities [27]. Another approach is to input noise and category labels into the generator, enabling the discriminator to identify fault types while assessing sample authenticity. Studies show that the ACGAN exhibits a promising performance for data augmentation. For instance, Zou L et al. [28] applied the ACGAN in the maritime field to enhance the resolution of sliced ground images. Moreover, Jin et al. [29] proposed a multi-layer branch ACGAN for signal expansion. However, the ACGAN faces several problems: (1) Generators in an ACGAN typically use CNNs with hierarchical structures composed of multiple convolutional and pooling layers, resulting in a cumbersome data generation process and long training times. (2) Most ACGAN methods employ Jensen Shannon (JS) divergence for their loss functions, leading to instability and vanishing gradients due to its discrete nature. (3) The ACGAN discriminator is tasked with discrimination and classification simultaneously [30,31]. In summary, addressing the limitations of the traditional ACGAN, achieving efficient and high-quality sample generation, and accurately diagnosing wheelset-bearing faults are of significant importance.
As a network structure built on a self-attention mechanism, the transformer has been extensively utilized in the field of Natural Language Processing (NLP) due to its powerful scaling capabilities and ability to learn long-range dependencies [32,33,34,35]. Recently, researchers have increasingly applied the transformer’s robust modeling capabilities to computer vision (CV) and bearing fault diagnosis [36,37,38,39]. Based on the literature survey performed, this paper proposes the transformer and auxiliary classifier generative adversarial network (TACGAN) to address issues in data augmentation and fault classification under various working conditions, aiming for balanced fault categories and high-precision fault classification. The TACGAN generates vibration signals for different fault categories of bearings and includes a transformer generator, a transformer discriminator, and a CNN classifier. The performance of the TACGAN was evaluated using collected vibration data, demonstrating its effectiveness in generating high-quality fault samples across different categories, surpassing existing ACGAN methods. The main contributions of this article can be summarized as follows:
(1)
To eliminate the cumbersome loop structure of convolutional layers; transformer networks replace CNNs in the generator and discriminator.
(2)
To avoid vanishing gradients, exploding gradients, and overfitting issues, the Wasserstein distance is introduced into the new cross-entropy loss function, ensuring stability.
(3)
To prevent the overlap of discrimination and classification tasks in the ACGAN discriminator, the classifier is separated from the discriminator as an independent component.
This article is organized as follows: Section 2 explains the basic principles of the ACGAN and transformers. Section 3 introduces the proposed TACGAN model and its training process. Section 4 presents experimental validations demonstrating the effectiveness of sample generation and fault classification. Finally, the main achievements are summarized in Section 5.

2. Basic Theory

2.1. Auxiliary Classifier Generative Adversarial Networks

The architecture of the ACGAN closely resembles that of a standard GAN, with the key distinction being its use of label information as a conditional probability for input samples. Unlike the GAN, the ACGAN incorporates a more sophisticated design by effectively utilizing the discriminator. This discriminator not only distinguishes between real and fake samples but also categorizes them. By evaluating the categories of the generated samples, the discriminator can more effectively transmit the loss function LS and classification loss LC to the generator. This dual-feedback mechanism enables the generator to more accurately learn the true distribution of samples corresponding to each label [40,41,42]. Figure 1 illustrates the similarities and differences between the ACGAN and the Conditional GAN (CGAN).
L S = E r l o g P ( s = s r | x ) + E f l o g P ( s = s f | G a , c f )
L C = E r l o g P ( c = c r | x ) + E f l o g P ( c = c f | G a , c f )
where LS denotes the loss function used to determine the authenticity of the samples, while LC represents the loss function used to classify the type of data. Moreover, Er and Ef are the expected values of x and z, which follow the distribution of real samples and noise, respectively. G(a, cf) denotes the generated samples, where a represents the noise input. cr denotes the label for the real sample x, and cf represents the label for the sample G(a, cf) generated by the generator.

2.2. Transformer Encoder

The transformer, primarily used for processing sequential data like text, was introduced by a Google research team in 2017, with a focus on machine translation scenarios [43]. Compared to previously used RNNs and CNNs, the transformer offers faster processing, requires fewer parameters, mitigates the challenge of long-distance information retention, allows for parallel feature extraction, and incorporates global feature modeling to reduce the learning cycle [44]. Consequently, the transformer model has been widely adopted and has demonstrated outstanding performance across various industries.
A typical transformer architecture consists of an encoder and a decoder, with the encoder exhibiting a robust feature extraction capability. For wheelset bearing classification, the encoder is utilized, as depicted in Figure 2. The transformer encoder is composed of N identical encoding units, each layer comprising two primary components: a multi-head self-attention (MSA) module and a feedforward network (FFN) module. Residual connections are employed between these two layers, and each layer is followed by a layer normalization (LayerNorm), structured as follows:
X M S A = L a y e r N o r m M S A X i n + X i n
X o u t = L a y e r N o r m M L P X M S A + X M S A
Firstly, through the parameter matrix, the MSA mechanism maps the input two-dimensional embedded image x, resulting in three matrices: the query Qi, the key Ki, and the value Vi. In this formulation, i ranges from 1 to H, where H represents the total number of heads. The weighted sum is then obtained by applying scaled dot product attention on V and computing the weight matrix by calculating Q and K. This process is crucial for enhancing the ability to learn global features via MSA, which is essential for the effective diagnosis of the transformer [45]. The MSA function of the i-th head is defined as shown in Figure 3.
A t t e n t i o n h e a d i Q i , K i , V i = s o f t m a x ( Q i K i T d k ) V i
where dk denotes the dimension of Q and K, and 1 / d k is a parameter introduced to balance data and model convergence. Moreover, Qi, Ki, and Vi are defined as shown below:
Q i = X W i Q , K i = X W i K , V i = X W i V
where the parameter matrices W i Q , W i K R d × d k , W i V R d × d v , and dv represents the dimension of value.
MSA gathers h individual attention sets. Each set of transformed queries, keys, and values is processed in parallel through attention aggregation. These h attention-focused outputs are then concatenated and transformed through another learnable linear projection to produce the final output. This process can be mathematically expressed as follows:
M u l t i H e a d Q , K , V = C o n c a t A t t e n t i o n h e a d 1 , A t t e n t i o n h e a d 2 , , A t t e n t i o n h e a d h W O
where W O R h . d × d represents the linear projection after concatenation, and d k = d v = d / h .
Then, the FFN module, structured as an MLP, includes a nonlinear transformation with an activation function followed by a linear transformation. In this module, the MLP first uses a nonlinear layer to perform the nonlinear dimensionality raising operation on the input and then uses the linear layer to carry out linear dimensionality reduction operation to extract features. This process can be mathematically described as follows:
M L P X M S A = A c t i v a t i o n X M S A W M L P _ 1 + b 1 W M L P _ 2 + b 2
where W M L P _ 1 R d × d M L P and W M L P _ 2 R d × d M L P are the weights, b 1 R d M L P and b 2 R d , and (·) represents the activation fuction. dMLP denotes the embedding dimension of the nonlinear transformations, dMLP > d.
To capture the relative positions between input data, a position information vector is added to the initial vector before the MSA mechanism. Finally, the token embedding and position encoding are added together. The position encoding is as follows:
P E p o s , i = s i n ( w k p o s ) , i = 2 k c o s ( w k p o s ) , i = 2 k + 1
w k = 1 1000 2 k / d , k = 1 , 2 , , d / 2

3. Proposed Framework

To address the challenges of combined identification and classification within the discriminator, this study proposes separating these functions by introducing an independent classifier. This section outlines the comprehensive architecture of the TACGAN. It provides a detailed explanation of the processes involved in sample generation, recognition, classification, and the formulation of a loss function designed for balanced sample generation. The schematic structure of the TACGAN is depicted in Figure 4.

3.1. TACGAN Generator

The TACGAN generates high-quality time–frequency diagrams of wheelset bearings by extracting deep global features from 2D time–frequency images, which are transformed from 1D vibration signals using continuous wavelet transform [46]. It is worth noting that the generated time–frequency images closely resemble real ones. Traditional GAN generators that use CNN networks suffer from computational inefficiencies due to tedious iterations with increasing depth. To address this, the CNN network in the traditional GAN is replaced with a transformer encoder in this paper. The transformer encoder, which has a simpler structure, enables parallel computing and saves GPU space. The generator structure of the TACGAN is shown in Figure 5.
In the TACGAN, the transformer’s input is a 2D time–frequency image, which is divided into tiled 2D image patches, where x R H × W × C , and x p i R z × P 1 . P 2 . C . The resolution of the input image is specified by H and W, and C represents the number of channels. The resolution of each patch is determined by P1 and P2, and Z denotes the number of patches. The index i of x p i ranges from 1 to Z, with Z = (HW)/(P1P2). A token representing the fault label is added at the beginning of the sequence during random initialization. This token can transfer information by interacting with other vibration signal tokens and positional information during the training process, facilitating learning class information from the input image. With the addition of the fault label token, the image patch sequence length becomes Z + 1. Positional encoding is used to extract the positional information of the images. Thus, the transformer’s input can be represented as follows:
X i n = x c l a s s ; x p 1 W P ; x p 2 W P ; ; x p Z W P + E p o s
where W P R P 1 . P 2 . C × d represents the linear projection. The architecture of the transformer generator is shown in Figure 5. The input consists of the fault tag token and the 2D time–frequency image, which is divided into Z patches. These inputs are then fed into the transformer encoder module 1.

3.2. TACGAN Discriminator

When training convolutional-based neural networks, performing fault diagnosis and authenticity discrimination within the same convolutional network can lead to mutual interference, reducing the network’s accuracy and classification capability. To address this issue, an independent classifier is established, allowing the discriminator to focus solely on sample authenticity analysis. In the TACGAN, the discriminator processes real samples, generates samples, and labels their authenticity. The internal structure of the transformer encoder in the discriminator is similar to that of the generator. An authenticity tag token is added to the header of the transformer encoder, which then exchanges information with the batch of 2D time–frequency images to determine the probability that the output is genuine, as illustrated in Figure 6. The input to the discriminator (D) includes a 2D time–frequency image and an authenticity tag token, resulting in an input sequence length of Z + 1. The input format for the discriminator is represented as follows:
X D _ i n = x f r ; x p 1 W P ; x p 2 W P ; ; x p Z W P + E p o s
The concrete parameters of the generator and discriminator for the TACGAN proposed are shown in Table 1.

3.3. TACGAN Classifier

The input of the TACGAN independent classifier is the generated real samples and fault tags, as shown in Figure 4, and the classifier is composed of a CNN-based model. The detailed parameters are listed in Table 2.

3.4. Loss Function of TACGAN

In the TACGAN, Wasserstein distance is used to the losses. Wasserstein distance may provide a meaningful gradient as follows:
W P r , P g = i n f γ ~ P r , P g E x , y x y
where P r , P g is a set of joint distributions combined with Pr and Pg. For each possible joint distribution γ , the real sample x, and generate sample g can be obtained from (x, g)~ γ , and the distance x g between the two sample pairs is calculated in this joint distribution γ . The loss of the TACGAN is as follows:
L C R = E x ~ P r x l o g P c = c r x
L C G = E z ~ P z z l o g P c = c g G z , c g
L D = E x ~ P r x D x E z P z z ~ D G z , c g ] ]
L G = E z ~ P z z D G z , c g + 0.5 × L C R + 0.5 × L C G
L C = λ 1 L C R + λ 2 L C G
where LD, LG and LC are the loss of D, G, and C. L C R is the loss function of C for real data, and L C G is the loss of C for generated data. λ 1 and λ 2 are the ratio factors of C loss in TACGAN. The TACGAN is needed to train, listed in Algorithm 1.
Algorithm 1: The training algorithm of the TACGAN
       Wheelset bearing fault diagnosis with the proposed model
  the Inputed labeled samples X = {(xj, aj)};
1: Initialize model parameters:
2: For j = 1 to N do:
3:   For i = 1 to 5 do:
4:    Extract a batch real datas{x}, Construct the noise with the label information{a, y};
5:    Calculate G(a, cf);
6:    LD = Ereal[D(x) − Efalse[D(G(a, cf))]];
7:   End For
8:   Extract a batch real datas{x, y}, Construct the noise with the label information{a, y};
9:    Calculate G(a, cf);
10:   LrC = Ereal[−logP(c = cr|x)]; LfC = Efalse[−logP(c = cr|x)];
11:   LG = Efalse[D(G(a, cf))] + λLfC;
12:   θC←Adam(LC);
13:   θG←RMSProp(LG);
14: End For

4. Experimental Verification

4.1. Datasets

To evaluate the performance of the TACGAN, vibration signal data from wheelset bearings, measured using a high-speed train wheelset bearing comprehensive test platform, are utilized. As shown in Figure 7, the platform includes a driving motor, loading device, bearing, and test bearing. An accelerometer for vibration signal acquisition is mounted on the test wheelset bearing, as depicted in Figure 8. This accelerometer has a sensitivity of 2.505 mV/m/s2.
During the test, the sampling frequency is set to 12.8 kHz. The wheelset bearing under test exhibits four states: normal (N), inner ring fault (I), outer ring fault (O), and rolling element fault (R). Photographs of the test wheelset bearing are shown in Figure 9, with detailed specifications provided in Table 3.
Fault samples are collected by the linear velocity of the wheelset bearing: 100 km/h, 200 km/h, and 300 km/h, corresponding to datasets A, B, and C, respectively. The operational load of the wheelset bearing is 5 tons, resulting in 12 experimental datasets for verification. Initially, the original 1D vibration signals are transformed into 2D time–frequency images using the continuous wavelet transform method, which serves as the input for the TACGAN model. The transformation employs the complex Morlet wavelet with a scale of 128, center frequency, and bandwidth of 2. The resulting 256 × 256 time–frequency images of real and generated samples are shown in Figure 10. Each sample comprises 1024 data points, and 2000 samples are used for each experimental dataset, totaling 24,000 samples from the 12 datasets. Detailed information about the train bearing datasets is provided in Table 4. The data are split into training and experimental sets in a 7:3 ratio to test the network’s generalization performance.
Experimental validation is performed using a Python 3.8 script, specifically designed for this purpose. The deep learning architecture is built on PyTorch 1.3, ensuring a robust and efficient framework. Experiments and training processes are conducted on a Windows 10 system, powered by an Intel Xeon Gold 6148 CPU, 16 GB of RAM, and an RTX 2080 Ti GPU with 4 GB of dedicated memory, facilitating high-performance computations and accelerating the training procedures.

4.2. Sample Generation and Fault Diagnosis

4.2.1. Performance of TACGAN

To evaluate the performance of the TACGAN under conditions of sample scarcity, experiments were conducted using the aforementioned samples. Figure 11a reveals that after 10,000 epochs of training, the model reached stable accuracy across all three components. Remarkably, Figure 11b demonstrates that the classifier model reached stable accuracy after approximately 20 epochs. This observation demonstrates the training process of the three modules: the generator, discriminator, and classifier. Figure 11 shows that the proposed TACGAN exhibits significant training stability, with all components working in tandem to achieve a rapid stable operational state and maintaining it until the training concludes.
A common issue with many GANs is mode collapse, where the generated samples lack diversity, causing the discriminator to be deceived by repetitive patterns. Mode collapse can be identified through visual inspection. As shown in Figure 12, the generated samples display a high level of diversity and closely resemble real samples in the temporal domain. Notably, there is no evidence of mode collapse in the generated samples.
The confusion matrices for model classification accuracy are presented in Figure 13. The classifier’s accuracy for real and test samples is 97.38% and 99.25%, respectively, indicating a high level of precision. This high accuracy allows the classifier to effectively guide the generator in producing samples of specific fault types, enhancing the overall performance and reliability of the TACGAN in fault diagnosis tasks.
Furthermore, to assess the feature-level performance, 2D feature visualization using t-SNE was conducted. The presented results in Figure 14 indicate that the samples generated by the TACGAN are well clustered in 2D space, with balanced representation across the four fault types. The clear separation of the four fault types by the TACGAN classifier further underscores the effectiveness and superiority of the TACGAN in generating high-quality samples.

4.2.2. Comparison and Analysis of Sample Generation Effect

To validate the rationality and superiority of the TACGAN, it is compared with the ACGAN and WGAN-GP [42]. The ACGAN employs CNNs and leverages tagged samples, while the WGAN-GP is a standard GAN model utilizing the Wasserstein distance, incorporating both a discriminator and a generator. The maximum mean difference (MMD) metric is used to explore the similarity between the generated samples and the actual samples. The MMD works by mapping the distributions of the generated and real samples into another space, calculating the distances between each corresponding point, and summing these distances to measure the overall similarity between the two distributions.
M M D ( X , Y ) = i = 1 n 1 φ x i j = 1 n 2 φ y i H 2 = 1 m m 1 i j m k x i , x j + 1 n n 1 i j n k y i , y j 2 m n i , j = 1 m , n k x i , x j
In this study, Φ(·) denotes a high-dimensional mapping of vectors. Equation (19) indicates that a smaller MMD reflects a closer alignment between the spatial distributions of generated samples and real samples. For each bearing fault type, five real samples were randomly selected and used for comparison across three models: TACGAN, ACGAN, and WGAN-GP. This process generated 12 MMD distances, which are illustrated in Figure 15. The results show that the TACGAN consistently achieves the smallest MMD values (0.2403, 0.2219, 0.1982, 0.2115) across various fault types, indicating that the generated samples are more similar to real samples compared to those produced by the ACGAN and WGAN-GP.
To validate the accuracy of the datasets generated by the TACGAN, two fault diagnosis models, including Convolutional Neural Networks (CNNs), Graph Convolutional Networks (GNNs), Long Short-Term Memory Networks (LSTMs) and Transformer Networks [47,48,49,50,51,52], were employed. Both models were initially trained on sufficient real samples, and their diagnostic accuracy was recorded. As depicted in Table 5 and Figure 16, the diagnostic accuracy of the models trained on TACGAN-augmented datasets is comparable to that achieved with real datasets.

5. Conclusions

This study introduces a novel model for augmenting wheelset bearing fault samples, incorporating a transformer-based generator and discriminator, along with an independent fault classifier. The proposed TACGAN model exhibits several distinct advantages as follows:
(1)
By employing a transformer network, the TACGAN bypasses the need for complex recursive structures. This approach allows for the direct extraction of both global and local features from input feature maps, thereby streamlining the model architecture and boosting computational efficiency.
(2)
The TACGAN effectively learns and replicates the distribution of real samples within a high-dimensional space. This results in generated samples that closely mirror the properties of actual data, which is particularly advantageous when addressing diverse and intricate fault types.
(3)
Testing has demonstrated that the TACGAN achieves an impressive 99% accuracy in augmenting wheelset bearing data. The fault samples produced by the TACGAN significantly enhance the dataset, improving the overall robustness and reliability of fault diagnosis systems.

Author Contributions

Conceptualization, J.Z.; methodology, J.L.; software, Z.Y.; validation, T.M.; formal analysis, Z.M.; investigation, S.L.; data curation, J.Z.; writing—original draft preparation, J.Z.; writing—review and editing, J.L. and Z.M.; visualization, J.Z.; supervision, J.L.; project administration, Z.M.; funding acquisition, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the National Natural Science Foundation of China (Grant No. 12393783), the National Natural Science Foundation of China (Grant No. 12072207), the National Natural Science Foundation of China under Grant (No. 52205571), and the Student Science Research Project of Hebei Jiaotong Vocational and Technical College (No. KJ-202305).

Data Availability Statement

The datasets presented in this article are not readily available because the data are part of an ongoing study.

Conflicts of Interest

Author Tianming Mu was employed by the company State Grid Shijiazhuang Electric Power Supply Company. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Yuan, Z.; Li, X.; Liu, S.; Ma, Z. A recursive multi-head graph attention residual network for high-speed train wheelset bearing fault diagnosis. Meas. Sci. Technol. 2023, 34, 065108. [Google Scholar] [CrossRef]
  2. Fu, W.; Jiang, X.; Li, B. Rolling bearing fault diagnosis based on 2D time-frequency images and data augmentation technique. Meas. Sci. Technol. 2023, 34, 045005. [Google Scholar] [CrossRef]
  3. Xin, G.; Li, Z.; Jia, L. Fault diagnosis of wheelset bearings in high-speed trains using logarithmic short-time continuous wavelet transformand modified self-calibrated residual network. IEEE Trans. Ind. Inform. 2021, 18, 7285–7295. [Google Scholar] [CrossRef]
  4. Yang, S.; Gu, X.; Liu, Y. A general multi-objective optimized wavelet filter and its applications in fault diagnosis of wheelset bearings. Mech. Syst. Signal Process. 2020, 145, 106914. [Google Scholar] [CrossRef]
  5. Yi, C.; Li, Y.; Huo, X. A promising new tool for fault diagnosis of railway wheelset bearings: SSO-based Kurtogram. ISA Trans. 2022, 128, 498–512. [Google Scholar] [CrossRef] [PubMed]
  6. Cao, H.; Fan, F.; Zhou, K. Wheel-bearing fault diagnosis of trains using empirical wavelet transform. Measurement 2016, 82, 439–449. [Google Scholar] [CrossRef]
  7. Li, H.; Liu, T.; Wu, X. Research on test bench bearing fault diagnosis of improved EEMD based on improved adaptive resonance technology. Measurement 2021, 185, 109986. [Google Scholar] [CrossRef]
  8. Zheng, J.; Pan, H.; Cheng, J. Rolling bearing fault detection and diagnosis based on composite multiscale fuzzy entropy and ensemble support vector machines. Mech. Syst. Signal Process. 2017, 85, 746–759. [Google Scholar] [CrossRef]
  9. Yu, W.; Zhao, C. Online fault diagnosis for industrial processes with Bayesian network-based probabilistic ensemble learning strategy. IEEE Trans. Autom. Sci. Eng. 2019, 16, 1922–1932. [Google Scholar] [CrossRef]
  10. Jiao, J.; Zhao, M.; Lin, J. Hierarchical discriminating sparse coding for weak fault feature extraction of rolling bearings. Reliab. Eng. Syst. Saf. 2019, 184, 41–54. [Google Scholar] [CrossRef]
  11. Wang, D.; Guo, Q.; Song, Y. Application of multiscale learning neural network based on CNN in bearing fault diagnosis. J. Signal Process. Syst. 2019, 91, 1205–1217. [Google Scholar] [CrossRef]
  12. An, Z.; Li, S.; Wang, J. A novel bearing intelligent fault diagnosis framework under time-varying working conditions using recurrent neural network. ISA Trans. 2020, 100, 155–170. [Google Scholar] [CrossRef] [PubMed]
  13. Zhu, J.; Jiang, Q.; Shen, Y. Application of recurrent neural network to mechanical fault diagnosis: A review. J. Mech. Sci. Technol. 2022, 36, 527–542. [Google Scholar] [CrossRef]
  14. Zhang, Y.; Zhou, T.; Huang, X. Fault diagnosis of rotating machinery based on recurrent neural networks. Measurement 2021, 171, 108774. [Google Scholar] [CrossRef]
  15. Pham, M.T.; Kim, J.M.; Kim, C.H. Rolling bearing fault diagnosis based on improved GAN and 2-D representation of acoustic emission signals. IEEE Access 2022, 10, 78056–78069. [Google Scholar] [CrossRef]
  16. Wang, X.; Jiang, H.; Wu, Z. Adaptive variational autoencoding generative adversarial networks for rolling bearing fault diagnosis. Adv. Eng. Inform. 2023, 56, 102027. [Google Scholar] [CrossRef]
  17. An, Y.; Zhang, K.; Liu, Q. Rolling bearing fault diagnosis method base on periodic sparse attention and LSTM. IEEE Sens. J. 2022, 22, 12044–12053. [Google Scholar] [CrossRef]
  18. Fu, G.; Wei, Q.; Yang, Y. Bearing fault diagnosis based on CNN-BiLSTM and residual module. Meas. Sci. Technol. 2023, 34, 125050. [Google Scholar] [CrossRef]
  19. Yu, W.; Zhao, C. Broad convolutional neural network based industrial process fault diagnosis with incremental learning capability. IEEE Trans. Ind. Electron. 2019, 67, 5081–5091. [Google Scholar] [CrossRef]
  20. Daldal, N.; Cömert, Z.; Polat, K. Automatic determination of digital modulation types with different noises using convolutional neural network based on time–frequency information. Appl. Soft Comput. 2020, 86, 105834. [Google Scholar] [CrossRef]
  21. Yang, J.; Liu, J.; Xie, J. Conditional GAN and 2-D CNN for bearing fault diagnosis with small samples. IEEE Trans. Instrum. Meas. 2021, 70, 3525712. [Google Scholar] [CrossRef]
  22. Wang, Z.; Zhou, J.; Du, W. Bearing fault diagnosis method based on adaptive maximum cyclostationarity blind deconvolution. Mech. Syst. Signal Process. 2022, 162, 108018. [Google Scholar] [CrossRef]
  23. Xiong, J.; Liu, M.; Li, C. A bearing Fault Diagnosis Method Based on Improved Mutual Dimensionless and Deep Learning. IEEE Sens. J. 2023, 23, 18338–18348. [Google Scholar] [CrossRef]
  24. Song, X.; Cong, Y.; Song, Y. A bearing fault diagnosis model based on CNN with wide convolution kernels. J. Ambient. Intell. Humaniz. Comput. 2021, 13, 4041–4056. [Google Scholar] [CrossRef]
  25. Erol, B.; Gurbuz, S.Z.; Amin, M.G. Motion classification using kinematically sifted acgan-synthesized radar micro-doppler signatures. IEEE Trans. Aerosp. Electron. Syst. 2020, 56, 3197–3213. [Google Scholar] [CrossRef]
  26. Li, W.; Zhong, X.; Shao, H. Multi-mode data augmentation and fault diagnosis of rotating machinery using modified ACGAN designed with new framework. Adv. Eng. Inform. 2022, 52, 101552. [Google Scholar] [CrossRef]
  27. Wei, H.; Zhang, Q.; Gu, Y. Remaining useful life prediction of bearings based on self-attention mechanism, multi-scale dilated causal convolution, and temporal convolution network. Meas. Sci. Technol. 2023, 34, 045107. [Google Scholar] [CrossRef]
  28. Zou, L.; Zhang, H.; Wang, C. MW-ACGAN: Generating multiscale high-resolution SAR images for ship detection. Sensors 2020, 20, 6673. [Google Scholar] [CrossRef]
  29. Cheng, Y.; Zhou, N.; Wang, Z. CFFsBD: A candidate fault frequencies-based blind deconvolution for rolling element bearings fault feature enhancement. IEEE Trans. Instrum. Meas. 2023, 72, 3506412. [Google Scholar] [CrossRef]
  30. Fu, Z.; Liu, Z.; Ping, S. TRA-ACGAN: A motor bearing fault diagnosis model based on an auxiliary classifier generative adversarial network and transformer network. ISA Trans. 2024, 149, 381–393. [Google Scholar] [CrossRef] [PubMed]
  31. Dixit, S.; Verma, N.K.; Ghosh, A.K. Intelligent fault diagnosis of rotary machines: Conditional auxiliary classifier GAN coupled with meta learning using limited data. IEEE Trans. Instrum. Meas. 2021, 70, 3517811. [Google Scholar] [CrossRef]
  32. Du, X.; Jia, L.; Haq, I.U. Fault diagnosis based on SPBO-SDAE and transformer neural network for rotating machinery. Measurement 2022, 188, 110545. [Google Scholar] [CrossRef]
  33. Ding, Y.; Jia, M.; Miao, Q. A novel time–frequency Transformer based on self–attention mechanism and its application in fault diagnosis of rolling bearings. Mech. Syst. Signal Process. 2022, 168, 108616. [Google Scholar] [CrossRef]
  34. Liu, W.; Zhang, Z.; Zhang, J. A novel fault diagnosis method of rolling bearings combining convolutional neural network and transformer. Electronics 2023, 12, 1838. [Google Scholar] [CrossRef]
  35. Wu, H.; Triebe, M.J.; Sutherland, J.W. A transformer-based approach for novel fault detection and fault classification/diagnosis in manufacturing: A rotary system application. J. Manuf. Syst. 2023, 67, 439–452. [Google Scholar] [CrossRef]
  36. Jin, Y.; Hou, L.; Chen, Y. A time series transformer based method for the rotating machinery fault diagnosis. Neurocomputing 2022, 494, 379–395. [Google Scholar] [CrossRef]
  37. Li, Z.; He, Y.; Xing, Z. Transformer fault diagnosis based on improved deep coupled dense convolutional neural network. Electr. Power Syst. Res. 2022, 209, 107969. [Google Scholar] [CrossRef]
  38. Hou, Y.; Wang, J.; Chen, Z. Diagnosisformer: An efficient rolling bearing fault diagnosis method based on improved Transformer. Eng. Appl. Artif. Intell. 2023, 124, 106507. [Google Scholar] [CrossRef]
  39. Fang, H.; An, J.; Liu, H. A lightweight transformer with strong robustness application in portable bearing fault diagnosis. IEEE Sens. J. 2023, 23, 9649–9657. [Google Scholar] [CrossRef]
  40. Gao, H.; Zhang, X.; Gao, X. ICoT-GAN: Integrated convolutional transformer GAN for rolling bearings fault diagnosis under limited data condition. IEEE Trans. Instrum. Meas. 2023, 72, 3515114. [Google Scholar] [CrossRef]
  41. Ma, Z.; Yuan, Z.; Li, X. Multirotational Speed Data Augmentation and Data Repair of High-speed Train Wheelset Bearings using Graph Speed Classifier GAN. IEEE Sens. J. 2023, 23, 30984–30997. [Google Scholar] [CrossRef]
  42. Zheng, Y.; Wang, D. An Auxiliary Classifier Generative Adversarial Network based Fault Diagnosis for Analog Circuit. IEEE Access 2023, 11, 86824–86833. [Google Scholar] [CrossRef]
  43. Hu, H.; Ma, X.; Shang, Y. A novel method for transformer fault diagnosis based on refined deep residual shrinkage network. IET Electr. Power Appl. 2022, 16, 206–223. [Google Scholar] [CrossRef]
  44. Liang, P.; Yu, Z.; Wang, B. Fault transfer diagnosis of rolling bearings across multiple working conditions via subdomain adaptation and improved vision transformer network. Adv. Eng. Inform. 2023, 57, 102075. [Google Scholar] [CrossRef]
  45. Tang, X.; Xu, Z.; Wang, Z. A novel fault diagnosis method of rolling bearing based on integrated vision transformer model. Sensors 2022, 22, 3878. [Google Scholar] [CrossRef] [PubMed]
  46. Zhao, H.; Liu, J.; Chen, H. Intelligent diagnosis using continuous wavelet transform and gauss convolutional deep belief network. IEEE Trans. Reliab. 2022, 72, 692–702. [Google Scholar] [CrossRef]
  47. Miao, Z.; Xia, Y.; Zhou, F. Fault diagnosis of wheeled robot based on prior knowledge and spatial-temporal difference graph convolutional network. IEEE Trans. Ind. Inform. 2022, 19, 7055–7065. [Google Scholar] [CrossRef]
  48. Wang, M.; Zhang, W.; Shao, M. Separation and Extraction of Compound-Fault Signal Based on Multi-Constraint Non-Negative Matrix Factorization. Entropy 2024, 26, 583. [Google Scholar] [CrossRef]
  49. Mao, Y.; Xin, J.; Zang, L. Fault Diagnosis Method for Rolling Bearings Based on Grey Relation Degree. Entropy 2024, 26, 222. [Google Scholar] [CrossRef] [PubMed]
  50. Lu, L.; Wang, W.; Kong, D. Fault diagnosis of rotating machinery using kernel neighborhood preserving embedding and a modified sparse bayesian classification model. Entropy 2023, 25, 1549. [Google Scholar] [CrossRef] [PubMed]
  51. Rostaghi, M.; Khatibi, M.M.; Ashory, M.R. Refined composite multiscale fuzzy dispersion entropy and its applications to bearing fault diagnosis. Entropy 2023, 25, 1494. [Google Scholar] [CrossRef]
  52. Gao, D.; Zhu, Y.; Ren, Z. A novel weak fault diagnosis method for rolling bearings based on LSTM considering quasi-periodicity. Knowl.-Based Syst. 2021, 231, 107413. [Google Scholar] [CrossRef]
Figure 1. Structures of GAN (a) and ACGAN (b).
Figure 1. Structures of GAN (a) and ACGAN (b).
Entropy 26 01113 g001
Figure 2. Structure of a transformer encoder network.
Figure 2. Structure of a transformer encoder network.
Entropy 26 01113 g002
Figure 3. Structure diagram of the multi-head self-attention.
Figure 3. Structure diagram of the multi-head self-attention.
Entropy 26 01113 g003
Figure 4. The framework of TACGAN.
Figure 4. The framework of TACGAN.
Entropy 26 01113 g004
Figure 5. The architecture of generator transformer network.
Figure 5. The architecture of generator transformer network.
Entropy 26 01113 g005
Figure 6. The architecture of discriminator transformer network.
Figure 6. The architecture of discriminator transformer network.
Entropy 26 01113 g006
Figure 7. Picture of wheelset bearing experiment platform.
Figure 7. Picture of wheelset bearing experiment platform.
Entropy 26 01113 g007
Figure 8. Schematic of experiment platform.
Figure 8. Schematic of experiment platform.
Entropy 26 01113 g008
Figure 9. Photo of the test bearing. (a) N; (b) I; (c) O; (d) R.
Figure 9. Photo of the test bearing. (a) N; (b) I; (c) O; (d) R.
Entropy 26 01113 g009
Figure 10. Data of wheelset fault bearings.
Figure 10. Data of wheelset fault bearings.
Entropy 26 01113 g010
Figure 11. Training process of TACGAN.
Figure 11. Training process of TACGAN.
Entropy 26 01113 g011
Figure 12. Real data and generated data of TACGAN.
Figure 12. Real data and generated data of TACGAN.
Entropy 26 01113 g012
Figure 13. Confusion matrix for classification accuracy. (a) Real samples. (b) Generated samples.
Figure 13. Confusion matrix for classification accuracy. (a) Real samples. (b) Generated samples.
Entropy 26 01113 g013
Figure 14. Feature visualization results for TACGAN.
Figure 14. Feature visualization results for TACGAN.
Entropy 26 01113 g014
Figure 15. MMD distance of the data with 100 km/h.
Figure 15. MMD distance of the data with 100 km/h.
Entropy 26 01113 g015
Figure 16. Accuracy of the different datasets.
Figure 16. Accuracy of the different datasets.
Entropy 26 01113 g016
Table 1. Details of the parameters for the generator and discriminator.
Table 1. Details of the parameters for the generator and discriminator.
ParametersValue
Encoder Stacking Depth, N5
Number of MSA Heads, h4
Hidden Dimension of MLP, dMLP256
Embedding Dimension, d64
Batch Size32
Token Sequence Length, G65
Token Sequence Length, D66
Table 2. Details of the parameters.
Table 2. Details of the parameters.
NetworksLayersParameter SettingOperation/Activation
ClassifierConv2D3 × 3 × 32BN + pooling/ReLu
Conv2D3 × 3 × 64BN + pooling/ReLu
Conv2D3 × 3 × 128BN + pooling/ReLu
Conv2D3 × 3 × 256BN + pooling/ReLu
Dense256None/ReLu
Dense128None/ReLu
DenseClass numberNone/Softmax
Table 3. Details of the test wheelset bearings.
Table 3. Details of the test wheelset bearings.
Model NumberPitch Diameter D/mmRoller Diameter D/mmContact Angle φ/°Number of Rolling Elements
197,726176.2924.768.8320
Table 4. Details of datasets.
Table 4. Details of datasets.
Fault TypeSpeed Condition/km/hSample Size
N100/200/3002000/2000/2000
I100/200/3002000/2000/2000
O100/200/3002000/2000/2000
R100/200/3002000/2000/2000
Table 5. Diagnosis accuracy of each model in 100 km/h.
Table 5. Diagnosis accuracy of each model in 100 km/h.
ModelAccuracy (Real Samples)Accuracy (TACGAN)Accuracy (ACGAN)Accuracy (WGAN-GP)
CNN97.46%97.3397.0697.25
GNN97.72%97.68%95.84%93.93%
LSTM97.82%97.56%96.12%94.38%
Transformer98.86%98.83%97.13%95.47%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, J.; Li, J.; Yuan, Z.; Mu, T.; Ma, Z.; Liu, S. Sample Augmentation Using Enhanced Auxiliary Classifier Generative Adversarial Network by Transformer for Railway Freight Train Wheelset Bearing Fault Diagnosis. Entropy 2024, 26, 1113. https://doi.org/10.3390/e26121113

AMA Style

Zhao J, Li J, Yuan Z, Mu T, Ma Z, Liu S. Sample Augmentation Using Enhanced Auxiliary Classifier Generative Adversarial Network by Transformer for Railway Freight Train Wheelset Bearing Fault Diagnosis. Entropy. 2024; 26(12):1113. https://doi.org/10.3390/e26121113

Chicago/Turabian Style

Zhao, Jing, Junfeng Li, Zonghao Yuan, Tianming Mu, Zengqiang Ma, and Suyan Liu. 2024. "Sample Augmentation Using Enhanced Auxiliary Classifier Generative Adversarial Network by Transformer for Railway Freight Train Wheelset Bearing Fault Diagnosis" Entropy 26, no. 12: 1113. https://doi.org/10.3390/e26121113

APA Style

Zhao, J., Li, J., Yuan, Z., Mu, T., Ma, Z., & Liu, S. (2024). Sample Augmentation Using Enhanced Auxiliary Classifier Generative Adversarial Network by Transformer for Railway Freight Train Wheelset Bearing Fault Diagnosis. Entropy, 26(12), 1113. https://doi.org/10.3390/e26121113

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop