You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

11 November 2023

A Fault Diagnosis Strategy for Analog Circuits with Limited Samples Based on the Combination of the Transformer and Generative Models

,
,
,
,
and
1
School of Mechanical and Electrical Engineering, Xi’an University of Architecture and Technology, Xi’an 710055, China
2
School of Civil Aviation, Northwestern Polytechnical University, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Feature Papers in Fault Diagnosis & Sensors 2023

Abstract

As a pivotal integral component within electronic systems, analog circuits are of paramount importance for the timely detection and precise diagnosis of their faults. However, the objective reality of limited fault samples in operational devices with analog circuitry poses challenges to the direct applicability of existing diagnostic methods. This study proposes an innovative approach for fault diagnosis in analog circuits by integrating deep convolutional generative adversarial networks (DCGANs) with the Transformer architecture, addressing the problem of insufficient fault samples affecting diagnostic performance. Firstly, the employment of the continuous wavelet transform in combination with Morlet wavelet basis functions serves as a means to derive time–frequency images, enhancing fault feature recognition while converting time-domain signals into time–frequency representations. Furthermore, the augmentation of datasets utilizing deep convolutional GANs is employed to generate synthetic time–frequency signals from existing fault data. The Transformer-based fault diagnosis model was trained using a mixture of original signals and generated signals, and the model was subsequently tested. Through experiments involving single and multiple fault scenarios in three simulated circuits, a comparative analysis of the proposed approach was conducted with a number of established benchmark methods, and its effectiveness in various scenarios was evaluated. In addition, the ability of the proposed fault diagnosis technique was investigated in the presence of limited fault data samples. The outcome reveals that the proposed diagnostic method exhibits a consistently high overall accuracy of over 96% in diverse test scenarios. Moreover, it delivers satisfactory performance even when real sample sizes are as small as 150 instances in various fault categories.

1. Introduction

The utilization of analog circuits is prevalent throughout numerous electronic devices, such as communication equipment and control systems. Because their components have poor tolerance, they are more susceptible to interference and influence. According to relevant surveys, approximately 80% of all electronic circuits employed within electronic equipment are digital circuits. While digital circuits account for the majority of electronic circuits, analog circuits continue to make up a small portion, around 20% of the total, resulting in excess of 80% of system faults [1]. In analog circuit faults, complex faults can directly cause the system to malfunction and are easily detected. Soft faults are mainly caused by abnormal changes in resistance, capacitance, and inductance parameters, and the measurement is complex and challenging [2]. Since analog circuits are composed of nonlinear and fault-tolerant components, the insufficient number of measurable nodes and measurement uncertainty make fault diagnosis complex. Therefore, realizing soft fault diagnosis at the component level is still challenging. Over the past few decades, numerous scholars have delved into this field, proposing various types of fault diagnostic approaches [3,4,5,6]. The field of artificial intelligence has experienced remarkable advancement in recent years, enabling the formulation and successful application of various deep learning models for fault diagnosis due to their excellent independent feature extraction capabilities and outstanding complex process generalization capabilities [7,8]. Recently, a deep neural network islanding detection technique based on statistical features was proposed in reference [9], and non-islanding disturbances were classified for hybrid systems based on synchronous and inverter distributed generators, which is a highlight in the field of fault diagnosis. The authors of [10] propose a novel technique for analog circuit fault detection using the application of image recognition, converting the power spectral density of the output signal into a two-dimensional image and inputting it into a deep convolutional neural network to achieve image classification and achieve the purpose of fault detection. Yu et al. [11] conducted research on a novel method for fault detection and diagnosis, which is predicated on a fusion of the firefly algorithm, tent chaos mapping, and extreme learning machine. The efficacy of this approach is exceptional in its ability to generalize well in the realm of fault diagnosis. Liu et al. [12] proposes a fault diagnosis method based on vibration sensor in ellipsoidal-ARTMAP network and differential evolution algorithm.
The remarkable diagnostic performance exhibited by many deep learning models is inherently tied to the acquisition of sufficient and uniformly distributed data samples. Nevertheless, in practical applications, the procurement of sufficient fault samples proves to be a formidable task owing to various reasons such as costs and safety concerns. When problems such as a small number of samples or different data distributions under different working conditions occur, the impact of a particular factor on the efficacy and precision of the neural network-based fault diagnosis system will be significant. The deep learning model requires much data support to exert its powerful capabilities in data modeling and classification identification. It also requires sufficient data coverage under different working conditions to ensure the capability of a deep learning model to extend its knowledge learned and evolve to novel situations, which is evaluated in [13]. When the limited amount of data is utilized directly for deep learning training, it will lead to extremely serious overfitting, especially not conducive to fault diagnosis. Effective training of deep learning models with small sample datasets requires preprocessing of the data, which is pivotal to achieving optimal training outcomes. The core of small sample fault diagnosis is to resolve the issue of an insufficient number of samples. In the field of fault diagnosis research, using original samples to expand the amount of data or improve data quality is an effective solution. The existing main methods to expand the target sample size are the Synthetic Minority Oversampling Technique (SMOTE), variational autoencoder (VAE) [14], autoregressive model [15], and generative adversarial network (GAN) [16,17]. Chawla et al. [18] introduced the SMOTE technique, which increases the quantity of samples available for analysis by finding similar samples near the minority class samples and synthesizing new small samples. However, it should be noted that when the samples are distributed on the edge of the classification, the samples synthesized by the SMOTE method classification boundaries may be blurred. Autoregressive models can perform density estimation on sequence data well, but their computational load is much larger than that of VAE and GAN. The VAE method is a commonly used generative model. Through the hidden features (low-dimensional hidden variables) learned in the input data, these hidden features are sampled and reconstructed, and finally, new data similar to the input data are generated, such as in reference [19], which introduces a variational autoencoder (VAE) into the fault diagnosis, facilitating data augmentation through the generation of vibration signals. Subsequently, an enhanced fault diagnosis method is proposed, wherein a convolutional neural network is combined with the aforementioned approach to realize improved accuracy and performance. Although the VAE method is trained and the training process is relatively stable, the generated pictures are relatively blurred, and the generated samples lack the capability to capture comprehensive fault information. In addition, the authors of [20] designed an adversarial transfer network comprising multiple scales based on an attention mechanism that automatically distinguishes various bearing fault states. The TrAdaBoost method is different from other methods of expanding the target sample size. It uses auxiliary datasets for training and applies them to the training of classifiers through weight adjustment. Xiao et al. [21] used CNN as a classifier, combined with the method of assigning weights to joint training datasets and the weight update of the TrAdaBoost model to achieve accurate fault diagnosis with high precision for small samples. The TrAdaBoost method is flexible and easy to operate when combined with other classifiers but requires a suitable and similar auxiliary dataset, which is difficult to obtain. In addition, classification difficulty increases when there is noise in the dataset itself.
In 2014, Goodfellow et al. [22] summarized the characteristics of the generative and discriminative models in machine learning and proposed a creative network model-generated confrontation network (GAN). A novel deep learning approach has emerged that can be utilized for the purpose of data augmentation, data generation, data modeling, and other fields. Traditional GAN uses Jensen–Shannon divergence (JS divergence) to quantify the degree of congruence between false images and authentic images. With the in-depth research on the GAN network, researchers found that the network is prone to problems such as network collapse and gradient explosion during the training process, which makes the GAN network challenging to train. In order to address the aforementioned issues, more and more variant algorithms have been proposed. Least Squares Generative Adversarial Networks (LSGAN) [23] are designed to enhance the objective function of a GAN network, specifically the discriminator, by replacing the cross-entropy loss function with a least squares loss function, making the transfer of gradients more effective and the model’s training process more stable. Liu et al. [24] advanced a technique that amalgamates variational autoencoding and GAN to glean precise features of real data samples in the context of data scarcity. Wasserstein GAN [25,26] effectively substitutes the Kullback–Leibler (KL) divergence and Jensen–Shannon (JS) divergence with the Wasserstein distance in order to accurately measure the distance between the generated distribution and the actual distribution and make the gradient calculation of the generator more accurate. The conditional generative adversarial network (CGAN) [27] proposes adding conditional information to GAN’s network structure so that GAN can accurately train and generate images of various categories in one model simultaneously and according to different datasets and generation requirements. Similarly, Dixit et al. [28] used CGAN to generate vibration signals of rotating machines. Then, they used auxiliary classifiers as GAN discriminators in these signals and performed meta-learning to improve the generalization of classifiers’ capabilities for better fault diagnosis. References [29,30,31] go further based on CGAN by adding classifiers and clusters inside GAN to partially or entirely obtain condition information. Then, they conduct model training similar to CGAN, thus partially or entirely removing the limitations of label information on CGAN training. The authors of [32] propose the use of a temporal generative adversarial network, coupled with an efficient network model, to implement a transfer learning method, displaying excellent levels of effectiveness, reliability, and generalization performance. Deep Regret Analytic Generative Adversarial Networks (DRAGAN) [33] introduce the no-regret algorithm in game theory and transform its loss function to solve the GAN collapse problem. The authors of [34] incorporate a discriminator into its network design to resolve the issue of restricted variety within the generated samples. Rad-ford et al. [35] proposed the introduction of a convolutional neural network in place of the original fully connected network within the architecture of the GAN to develop the deep convolutional generative adversarial network (DCGAN), which significantly improved the generation and generalization capabilities of the GAN network model and provided new ideas for the development of GAN networks. The authors of [36] have employed DCGAN to convert one-dimensional vibration data into grayscale images, thereby increasing the availability of fault data samples and resolving the issue of inadequate data samples.
The Transformer model [37] was proposed by Google in 2017. One of the most notable characteristics of this particular model is its departure from network structures typically utilized in RNN and CNN. The Transformer model has gained significant prominence in the realm of machine translation due to its remarkable capabilities. Recently, there has been a growing interest in applying the Transformer model to various fields, such as sequence data prediction, target detection, and image classification, leading to promising outcomes [38,39,40]. The Swin Transformer [41], proposed by Microsoft Research Asia, has attained superior performance in multiple vision tasks, indicating its state-of-the-art nature. As the number of network layers increases, the attention mechanism may collapse, resulting in a decrease in Transformer model performance rather than an increase. To address this challenge, Touvron et al. [42] conducted research based on the ViT model and designed a layer scale structure that can also converge when the network is deepened, solving the problem of variance amplification in the residual connection process. A proposed method for fault diagnosis in rotating machinery utilizes intelligent feature self-extraction and transformer neural network techniques [43]. The multi-scale SinGAN model [44] is especially adept at generating high-quality kurtosis map images, which can play a pivotal role in effectively training complex machine learning models.
The methods proposed in the above literature have shown good learning capabilities and transfer effects in various unsupervised cross-domain fault diagnosis tasks, providing an essential reference for subsequent research in this field. To enhance the precision, stability, and broad applicability of diagnostic assessments, a set of innovative enhancements have been proposed.
Contributions: Our approach is made possible by the following technical contributions:
(i)
Using continuous wavelet transform (CWT) can enhance the processing of analog circuit signals in the time–frequency domain, efficiently converting one-dimensional signals into two-dimensional time–frequency images that are advantageous for fault feature selection, leading to improved training speeds and enhanced efficiency;
(ii)
A data augmentation approach based on DCGAN has been designed, using a generative model to generate convincingly realistic new images, effectively amplifying the image set. The amplified dataset is used to support model training, improving its performance and generalization capabilities, with the aim of addressing the problem of overfitting commonly experienced with small image datasets;
(iii)
A fast and high-precision diagnosis technology based on the Transformer model is proposed, which greatly shortens the training time and achieves a good diagnosis accuracy–speed trade-off.
The rest of the content is organized as follows: In Section 2, the theoretical background and the proposed method are outlined. Section 3 illustrates the validation of the proposed method by detailing the selection of three experimental circuits and the construction and preparation of the fault dataset. Section 4 delves into the experimental outcomes and analysis. The final section, Section 5, provides a comprehensive summation of the complete paper.

3. Analog Circuit Experiment and Fault Setting

This section primarily covers topics related to analog circuit examples, fault configuration and data acquisition, signal preprocessing methods, and diagnostic model parameter settings. These are preparatory steps for the subsequent presentation and analysis of the experimental results in Section 4.
In the process of operating analog circuits, the performance of capacitors, resistors, and other components inevitably deteriorates. Deviations in the actual output signal value from the theoretical value may occur due to various factors, resulting in an uneven result. In general, the device has a normal value for its operation X, tolerance α, and threshold β. The device can still work normally under the condition of working within the tolerance range. If the device works beyond the tolerance range but does not exceed the threshold range, the circuit has a soft fault. In the event that the device parameters exceed the pre-established threshold, the circuit will cease functioning altogether. To ensure the efficacy of the suggested algorithm, three circuits, including the Sallen–Key bandpass filter circuit, the Biquad low-pass filter circuit, and the Thomas filter circuit, were subjected to experimental verification, as illustrated in the accompanying figure.

3.1. Fault Settings

In the field of circuit fault diagnosis, the selection of faulty components is typically determined based on an analysis of which components have a significant impact on the input. This determination is generally guided by the circuit’s transfer function. The specific approach involves conducting sensitivity analysis using circuit simulation software to identify components that have a substantial influence on the circuit’s output. These components are then considered as potential candidates for faults, thus laying the groundwork for subsequent research. If only one device in the analog circuit has a soft fault, it is called a single fault, and its fault settings are depicted in Table 1 and Table 2. In the event of multiple components experiencing simultaneous failures, it is referred to as a compound fault, wherein the fault signal manifests similar properties to those observed with a single fault. Although the incidence is low, its low feature recognition rate and many fault modes cannot be ignored, and its fault settings are depicted in Table 3. In an analog circuit, the component that has a significant impact on the output of the circuit is referred to as the sensitive element. The Sallen–Key bandpass filter circuit, depicted in Figure 7, exhibits a number of sensitive components, including C1, C2, R2, and R3. The Biquad low-pass filter circuit illustrated in Figure 8 features a number of vulnerable components, including C1, C2, R1, R2, R3, and R4. Lastly, the Thomas filter circuit, shown in Figure 9, possesses numerous sensitive components, including C1, C2, R3, R4, and R5. To conduct an analog circuit fault test, it is recommended to adjust the device parameter upwards (↑) or downwards (↓) by a specified amount that exceeds the standard value. This measure aims to identify any potential issues within the circuit. The tolerance of the resistance and capacitance of the circuit are 5% and 10%, respectively, and the fault threshold is set to 30%. Specifically, the range of resistance failure in the circuit is encompassed within the limits of [70% X, 95% X]∪[105% X, 130% X]. Likewise, the range of capacitor failure in the circuit is restricted within the bounds of [70% X, 90% X]∪[110% X, 130% X], where X denotes the nominal value of the component.
Table 1. Single fault of Sallen–Key bandpass filter circuit.
Table 2. Single fault of Thomas filter circuit.
Table 3. Compound faults of Thomas filter circuit.
Figure 7. The circuit schematic of Sallen–Key bandpass filter.
Figure 8. The circuit schematic of Biquad low-pass filter circuit.
Figure 9. The circuit schematic of Thomas filter circuit.

3.1.1. Sallen–Key Bandpass Filter Circuit

The single fault set has been established to encompass the states {C1↑, C1↓, C2↑, C2↓, R2↑, R2↓, R3↑, R3↓, NF} in a total of nine distinct configurations, as per 0.
Combining the single fault in Table 1 in pairs to form a compound fault, the parameters of the circuit components remain unchanged, and the complex fault set is set to {C1↑C2↑, C1↓C2↓, C2↑R2↑, R2↓R3↑, R2↓R3↓C1↑, R3↑C1↑C2↓, NF}, which can be in eight states.

3.1.2. Biquad Low-Pass Filter Circuit

Like the Sallen–Key bandpass filter circuit, in the Biquad low-pass filter circuit, the normal values of capacitors C1 and C2 are both 5.00 nF, and the fault value range is [3.50 nF, 4.50 nF]∪[5.50 nF, 6.50 nF]; the normal value of R1, R2 and R3 is 6.20 kΩ, and the fault range is [4.34 kΩ, 5.89 kΩ]∪[6.51 kΩ, 8.06 kΩ]. As previously described, the single fault set is constructed as {C1↑, C1↓, C2↑, C2↓, R1↑, R1↓, R2↑, R2↓, R3↑, R3↓, R4↑, R4↓, NF}, with a total of 13 states, and the composite fault set can be created in eight states: {C1↑C2↑, C1↓C2↓, C2↑R2↑, R2↑R5↑, R1↑R2↑C1↑, R3↑C1↑C2↓, NF}, with a total of eight states.

3.1.3. Thomas Filter Circuit

The single fault set is constructed as {C1↑, C1↓, C2↑, C2↓, R3↑, R3↓, R4↑, R4↓, R5↑, R5↓, NF}, with a total of 11 states. The specific settings are shown in Table 2.
The composite fault set is created in eight states: {C1↑C2↑, C1↓C2↓, C2↑R3↑, R3↑R5↑, R3↑R5↑C1↑, R3↑C1↑C2↓, NF}, with a total of eight states. And Table 3 lists the settings.

3.2. Signal Preprocessing

In order to enhance the training capacity of the model and minimize the number of computations required, it is imperative to undertake the process of data preprocessing prior to its collection. As shown in Figure 10, the output voltage signals of the Sallen–Key bandpass filter circuit and the Thomas filter circuit under partial fault conditions are shown.
Figure 10. The response of analog circuits under different fault states. (a) Sallen–Key bandpass filter circuit. (b) Thomas filter circuit.
A.
Standardized processing
To further standardize the collected data and better train the neural network model, the data of each sample are standardized. The standardization method adopted in this paper is min-max standardization. Its mathematical expression is as follows:
X = x i = x i x m i n x m a x x m i n
B.
Spectrogram conversion
Consider the Sallen–Key bandpass filter circuit as an example. When continuous wavelet transform processing is applied to the output signal, the resulting graph can be seen in Figure 11.
Figure 11. Continuous wavelet transform results of Sallen–Key bandpass filter circuit response under different fault states. (a) F0. (b) F2. (c) F3. (d) F5.
Upon reviewing Figure 11, it is evident that the transformations resulting from various fault categories display subtle distinctions in the accompanying images. Therefore, manual classification is difficult, and automatic recognition technology is very necessary to explore the differences between different faults.
C.
Dataset partitioning
A total of 300 samples were collected for each fault type in both single-fault and multi-fault circuit cases. The samples were subjected to the aforementioned pretreatment procedure. A training dataset of 300 samples was assembled for each fault type, as the size of the data was determined based on a combination of real samples and generated samples with varying mixing ratios. A crucial aspect of the training process is to establish the ratio of real samples to generated samples in order to optimize the performance of the fault diagnosis model.

3.3. Fault Signal Collection

The experimental platform depicted in Figure 12 demonstrates a viable method for data acquisition. The equipment utilized incorporates an Agilent 33250 arbitrary waveform generator, Agilent 54853 digital oscilloscope, power supply, LabVIEW_2023 Q1 software, National Instruments 1042q data acquisition module, and the circuit under test (CUT). The arbitrary waveform generator creates a sinusoidal scanning excitation signal, which is then recorded using LabVIEW software. The test circuit’s output signal is then collected through the input of the sinusoidal excitation signal under various faults. Since the output signal is periodic, data analysis will involve collecting the signal for 50 cycles and dividing it into 50 samples. Specifically, to obtain accurate data, the first 10ms of the output signal will be collected with the sampling frequency set at 60 kHz and the sampling point set at 600. Each fault type and non-fault state will be assigned 500 labeled samples. Subsequently, the samples will be normalized and divided evenly into five separate subsets for five-fold cross-validation.
Figure 12. Experimental platform built.

3.4. Parameter Settings Involved in the Proposed Diagnostic Strategy

As depicted in Figure 13, a pragmatic, experimental platform has been developed for data acquisition. It encompasses an Agilent 54853 digital oscilloscope, a circuit under test (CUT), LabVIEW software, an Agilent 33250 arbitrary waveform generator, a National Instruments 1042q data acquisition module, and a power supply.
Figure 13. DCGAN network structure.
The DCGAN data generation model comprises a generator and a discriminator, both of which are components of the DCGAN architecture.
The DCGAN generator uses five transposed convolutional layers to upsample a 100-dimensional noise vector. The final output is a 64 × 64 × 1 image. Each convolutional layer, when transposed, has a stride of 2 and a 4 × 4 kernel size. The output dimensions of each transposed convolutional layer are as follows: 1024 × 4 × 4, 512 × 8 × 8, 256 × 16 × 16, 128 × 32 × 32, 1 × 64 × 64.
The architecture of the DCGAN discriminator mirrors that of the generator in a symmetric manner. It takes 64 × 64 × 1 images as input and uses five convolutional layers for feature extraction, with a stride of 2 and a kernel size of 4 × 4. The output dimensions of each convolutional layer are as follows: 1 × 64 × 64, 128 × 32 × 32, 256 × 16 × 16, 512 × 8 × 8, 1024 × 4 × 4. The resultant output is remodeled in order to conform to a predetermined shape, subsequently being inputted into a fully interconnected layer, in which the ensuing output value is utilized as the basis for the discriminator’s decision.
The batch size is 64, the learning rate is 0.001, and the model is trained for 50 epochs using the Leaky ReLU activation function. The Adam optimizer is used with a β1 value of 0.5.
In the context of the Transformer diagnostic model, the parameter configuration used in the experiment is as follows: a patch size of 4 × 4, 2 encoder layers, a token dimension of 96, and 12 attention heads. The maximum number of iterations that can occur during the training phase of the model has been set to 50.

4. Result Analysis and Discussion

In Section 3, after identifying the fault types for each analog circuit test case, data were collected, and model parameters were set. Subsequently, fault diagnosis tests were conducted on the three circuits to validate the effectiveness of the proposed data generation and fault diagnosis methods. This section will primarily focus on presenting and analyzing the experimental results.

4.1. Diagnostic Results of the Proposed Method

After the fault is determined, the data are collected, and the model parameters are set, subsequent fault diagnosis tests are conducted on three circuits to validate the efficacy of the proposed data generation and fault diagnosis methodologies. Specifically, five distinct tests were conducted for the single fault and compound fault configurations of the aforementioned circuits to prevent the occurrence of arbitrary test outcomes. The sample size for each fault category in each test instance was set at 50. In addition, it is essential to note that in the training process of the Transformer for fault diagnosis, the actual sample size and the generated sample size were set at 150, respectively. Table 4 shows the specific classification results of each fault category of CUT1 single fault in a single test. For a single CUT1 circuit fault case, three indicators were calculated for each fault category, namely Precision, Recall, and Specificity. Among them, the Precision index measures the ratio of accurately predicted positive samples, while the Recall index evaluates the proportion of positively predicted samples, indicating the degree to which positive samples are identified. The Specificity index refers to the ratio of correctly identified negative samples out of all negative samples.
Table 4. Fault diagnosis results of CUT1 single fault test.
As is evident from Table 4, for the fault-free category F0, all three index values are all 1, indicating that all samples of this category are correctly identified. According to the values of the three indexes, F1 and F7 fault categories have the most serious misclassification. In addition, the confusion matrix of the single fault diagnosis result of CUT1, CUT2, and CUT3 and the composite fault diagnosis result of CUT1 is shown in Figure 14.
Figure 14. Confusion matrix of diagnostic results of the proposed method in four cases. (a) CUT1 single fault case. (b) CUT2 single fault. (c) CUT3 single fault. (d) CUT1 compound fault.
As displayed in Figure 14, in the three cases of single fault diagnosis, among the 450 test samples of CUT1 shown in Figure 14a, all the normal samples were correctly classified, and a total of 11 fault samples were misclassified. In the test samples of CUT2, as displayed in Figure 14b, all 13 samples were mismarked. A total of 16 samples were mismarked in CUT3, as shown in Figure 14c. As shown in Figure 14d, a total of 13 samples of CUT1 compound fault were misclassified. In the figure, the overall classification accuracy of these four cases has reached more than 96%. It can be observed that the proffered methodology has exhibited a commendable classification efficacy, both in instances of single faults and compound faults.

4.2. Comparison Experiments

To analyze the superiority of the proffered scheme, a comparative experiment is set up in this paper. A total of four comparison methods were set. The first comparison method involves transforming the time domain signal generated by the circuit output into a two-dimensional signal with the aim of examining and diagnosing fault signals from a different perspective. As a comparison with CWT, the subsequent DCGAN generation module and Transformer feature mining and classification module remain unchanged. This comparison method is called TGT for short. The second comparison method is to assess the merits of the proposed data generation method; it would be viable to compare the modified version with the original GAN by utilizing the DCGAN data generation module. This comparison method is called CGT for short. In addition, to examine the benefits of incorporating Transformer modules, it is necessary to analyze their various advantages so that the signal mining modules of two-dimensional CNN and ResNet 35, which have advantages in mining two-dimensional signals, are selected for comparison with the Transformer module. The third and fourth comparison methods are referred to as CGC and CGR, respectively.
It is imperative to evaluate the effectiveness of the proposed approach and compare it with the four benchmark techniques in order to assess its performance in the given task. And three cases were tested, namely CUT1 single fault, CUT2 single fault, and CUT3 compound fault. First, five tests were carried out in each of the three cases, and the accuracy of each test is shown in the bar chart in Figure 15.
Figure 15. The diagnostic results of the proposed method and the four comparison methods in 5 tests. (a) CUT1 single fault; (b) CUT2 single fault; (c) CUT3 compound fault.
As displayed in Figure 15, for the three test cases, the performance of the proposed method is the best compared with the four comparison methods. Taking CUT1 single fault as an example, the proposed method demonstrates an average accuracy of 97.8% in five distinct tests, attaining a remarkable level of performance, while the diagnostic accuracy of the three comparison methods is 86.2 (TGT), 94.2 (CGT), 95.2 (CGC), and 96.1 (CGR), respectively. In addition, for the three test cases as a whole, among the four comparison methods, comparison method 4 has the best performance, followed by comparison method 3 and 2, while comparison method 1 has the worst performance. The feature extraction of the circuit output signal through wavelet transformation exhibits a notable positive influence on the subsequent classification process. DCGAN produces samples that are more closely similar to real data, surpassing GAN in accuracy. Transformer is better at image feature mining and classification than 2D CNN and ResNet 35.
In addition, using the three test cases mentioned above as examples, Figure 16 displays box plots for the proposed method and four comparison methods across five experiments.
Figure 16. Box diagram of diagnosis results of the proposed method and the four comparison methods. (a) CUT1 single fault. (b) CUT2 single fault. (c) CUT3 compound fault.
As shown in Figure 16, the diagnostic accuracy of the proposed method is superior to that of the four comparison methods for the three test cases. In addition, in terms of the distance from 1/4 quantile to 3/4 quantile, in the three test cases, compared to the four evaluated methods, the proposed approach appears to be more compact. In the comparison methods, the first comparison method has the largest distance from the first quartile to the third quartile, while the other three comparison methods are relatively closer. Therefore, it is evident that the proposed method displays stable and robust performance, while the robustness of the comparison method 1 is the worst.

4.3. Small Sample Dataset Testing

For the purpose of evaluating the impact of a small sample size on the data quality generated by the DCGAN model and then determining the appropriate sample size suitable for the generated data, a total of five different training sets were set up, and the number of real and generated samples is diverse in each combination, but for each fault class in any circuit case, the total number of training set samples is 300. In this paper, the CUT2 compound fault is used as an illustrative example, and the specific settings of the dataset are shown in Table 5, which also gives the average accuracy results of five tests of the mentioned method.
Table 5. Comparison results of the impact of different training set compositions on the diagnosis of the proposed method.
Table 5 reveals a rising trend in the average accuracy with an augmented number of real samples in the training set and a reduced number of artificially generated samples. Given the quantities of real samples and artificially generated samples, which are both 150, the average attained level of precision is 96.65%. However, moving forward in the composition of this training dataset, the rate of improvement in diagnostic performance of the proposed algorithm noticeably decreases. Therefore, considering the difficulty of obtaining the real sample size, it is plausible to assume that the proposed technique has achieved satisfactory diagnostic performance when the number of both real and generated samples reaches 150.

4.4. Research Outcomes and Limitations of the Proposed Method

In this study, the problem of overfitting easily caused by insufficient data in analog circuits is solved, and automatic feature extraction, sample expansion, and high-precision fault classification can be realized in the CWT-DCGAN-Transformer model. How to automatically optimize hyper-parameters such as the learning rate and network structure of the model is still insufficient.

5. Conclusions

To address the challenge of limited fault samples in electronic equipment, this paper has developed a small-sample fault diagnosis approach for simulated circuits based on CWT-DCGAN-Transformer, achieving two significant breakthroughs.
(i)
In scenarios with a limited number of real fault samples, the proposed approach demonstrates satisfactory performance. For instance, in cases where the real sample size is only 150, the diagnostic accuracy of a test case exceeds 96%;
(ii)
The proposed approach eliminates the need for the manual design of fault features. Through the joint processing of continuous wavelet transformation and the Transformer model, it autonomously explores time–frequency fault features and achieves self-attention and classification of these features. This leads to efficient diagnostic performance.
On one hand, this work offers a solution to the challenging problem of small-sample fault diagnosis in practical engineering. On the other hand, it explores the potential of Transformer models in the field of fault diagnosis. In the future, it is of paramount importance to explore how to automatically optimize the different modules and networks within the proposed approach.

Author Contributions

Conceptualization, Z.J. and Q.Y.; methodology, Z.J.; software, Q.Y.; validation, Q.Y., Y.L. and Z.J.; formal analysis, Z.J.; investigation, S.W.; resources, P.X.; data curation, Q.Y.; writing—original draft preparation, Z.J.; writing—review and editing, Z.J.; visualization, Q.Y.; supervision, Z.J.; project administration, Z.L.; funding acquisition, Z.L. All authors have read and agreed to the published version of the manuscript.

Funding

This paper is funded by the National Natural Science Foundation Fund (No. 52072309).

Data Availability Statement

The data presented in this study are available in this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Binu, D.; Kariyappa, B.S. A survey on fault diagnosis of analog circuits: Taxonomy and state of the art. AEU-Int. J. Electron. Commun. 2017, 73, 68–83. [Google Scholar] [CrossRef]
  2. Gan, X.; Qu, H.; Meng, X.; Wang, C.; Zhu, J. Research on ELM soft fault diagnosis of analog circuit based on KSLPP feature extraction. IEEE Access 2019, 7, 92517–92527. [Google Scholar]
  3. Farooq, M.; Xia, L.; Hussin, F.; Malik, A. High level fault modeling and fault propagation in analog circuits using NLARX automated model generation technique. In Proceedings of the 2012 4th International Conference on Intelligent and Advanced Systems (ICIAS2012), Kuala Lumpur, Malaysia, 12–14 June 2012; Volume 2, pp. 846–850. [Google Scholar]
  4. Zhang, A.; Chen, C.; Jiang, B. Analog circuit fault diagnosis based UCISVM. Neurocomputing 2016, 173, 1752–1760. [Google Scholar] [CrossRef]
  5. Binu, D.; Kariyappa, B. Ride NN: A new rider optimization algorithm-based neural network for fault diagnosis in analog circuits. IEEE Trans. Instrum. Meas. 2018, 68, 2–26. [Google Scholar] [CrossRef]
  6. Du, T.; Zhang, H.; Wang, L. Analogue circuit fault diagnosis based on convolution neural network. Electron. Lett. 2019, 55, 1277–1279. [Google Scholar] [CrossRef]
  7. Hu, Q.; Qin, A.; Zhang, Q.; He, J.; Sun, G. Fault diagnosis based on weighted extreme learning machine with wavelet packet decomposition and KPCA. IEEE Sens. J. 2018, 18, 8472–8483. [Google Scholar] [CrossRef]
  8. Shen, J.; Li, S.; Jia, F.; Zuo, H.; Ma, J. A deep multi-label learning framework for the intelligent fault diagnosis of machines. IEEE Access 2020, 8, 113557–113566. [Google Scholar] [CrossRef]
  9. Hussain, A.; Mirza, S.; Kim, C.-H. Islanding detection and classification of non-islanding disturbance in multi-distributed generation power system using deep neural networks. Electr. Power Syst. Res. 2023, 224, 109807. [Google Scholar] [CrossRef]
  10. Shokrolahi, S.; Karimiziarani, M. A deep network solution for intelligent fault detection in analog circuit. Analog. Integr. Circuits Signal Process. 2021, 107, 597–604. [Google Scholar] [CrossRef]
  11. Yu, W.; Sui, Y.; Wang, J. The faults diagnostic analysis for analog circuit based on FA-TM-ELM. J. Electron. Test. 2016, 32, 459–465. [Google Scholar] [CrossRef]
  12. Liu, C.; Wang, G.; Xie, Q.; Zhang, Y. Vibration sensor-based bearing fault diagnosis using ellipsoid-ARTMAP and differential evolution algorithms. Sensors 2014, 14, 10598–10618. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, J.; Meng, Y.; Chen, Z. A small target detection method based on deep learning with considerate feature and effectively expanded sample size. IEEE Access 2021, 9, 96559–96572. [Google Scholar] [CrossRef]
  14. Jiao, P.; Guo, X.; Jing, X.; He, D.; Wu, X.; Pan, S.; Gong, M.; Wang, W. Temporal network embedding for link prediction via VAE joint attention mechanism. IEEE Trans. Actionson Neural Netw. Learn. Syst. 2021, 33, 7400–7413. [Google Scholar] [CrossRef] [PubMed]
  15. Huang, D.; Zhu, X.; Li, R.; Wang, H. Feature screening for network autoregression model. Stat. Sin. 2021, 31, 1239–1259. [Google Scholar] [CrossRef] [PubMed]
  16. Alankrita, A.; Mamta, M.; Gopi, B. Generative adversarial network: An overview of theory and applications. Int. J. Inf. Manag. Data Insights 2021, 1, 100004. [Google Scholar]
  17. Song, K.; Zhang, Y.; Zhang, J.; Yuan, X. A generative model for synthesizing structured datasets based on GAN. J. Comput. Res. Dev. 2019, 56, 1832–1842. [Google Scholar]
  18. Chawla, N.; Bowyer, K.; Hall, L.; Kegelmeyer, W. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  19. Zhao, D.; Liu, S.; Gu, D.; Sun, X.; Wang, L.; Wei, L.; Zhang, H. Enhanced data-driven fault diagnosis for machines with small and unbalanced data based on variational auto-encoder. Meas. Sci. Technol. 2019, 31, 03504. [Google Scholar] [CrossRef]
  20. Zhao, B.; Zhang, X.; Zhan, Z.; Wu, Q. Deep multi-scale adversarial network with attention: A novel do main adaptation method for intelligent fault diagnosis. J. Manuf. Syst. 2021, 59, 565–576. [Google Scholar] [CrossRef]
  21. Xiao, D.; Huang, Y.; Qin, C.; Liu, Z.; Li, Y.; Liu, C. Transfer learning with convolutional neural networks for small sample size problem in machinery fault diagnosis. Proc. Inst. Mech. Eng. 2019, 233, 14. [Google Scholar] [CrossRef]
  22. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  23. Mao, X.; Li, Q.; Xie, H.; Lau, R.; Wang, Z.; Smolley, S. Least squares generative adversarial networks. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2794–2802. [Google Scholar]
  24. Liu, S.; Jiang, H.; Wu, Z.; Li, X. Rolling bearing fault diagnosis using variational auto encoding generative adversarial networks with deep regret analysis. Measurement 2021, 168, 108371. [Google Scholar] [CrossRef]
  25. Arjovsky, M.; Bottou, L. Towards principled methods for training generative adversarial networks. arXiv 2017, arXiv:1701.04862. [Google Scholar]
  26. Adler, J.; Lunz, S. Banach Wasserstein GAN. In Advances in Neural Information Processing Systems; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
  27. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier GANs. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 2642–2651. [Google Scholar]
  28. Dixit, S.; Verma, N.; Ghosh, A. Intelligent fault diagnosis of rotary machines: Conditional auxiliary classifier GAN coupled with meta learning using limited data. IEEE Trans. Actionson Instrum. Meas. 2021, 70, 3517811. [Google Scholar] [CrossRef]
  29. Lučić, M.; Tschannen, M.; Ritter, M.; Zhai, X.; Bachem, O.; Gelly, S. High-fidelity image generation with fewer labels. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 4183–4192. [Google Scholar]
  30. Liu, S.; Wang, T.; Bau, D.; Zhu, J.; Torralba, A. Diverse image generation via self-conditioned GANs. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 14286–14295. [Google Scholar]
  31. Zhang, X.; Cheng, Z.; Zhang, X.; Liu, H. Posterior promoted GAN with distribution discriminator for unsupervised dimage synthesis. In Proceedings of the 2021 IEEE/CVF Conferenceon Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 6519–6528. [Google Scholar]
  32. Kodali, N.; Abernethy, J.; Hays, J.; Kira, Z. On convergence and stability of gans. arXiv 2017, arXiv:1705.07215. [Google Scholar]
  33. Li, Q.; Kailkhura, B.; Anirudh, R.; Zhou, Y.; Liang, Y.; Varshney, P. GAN: Manifold regularized generative adversarial networks. arXiv 2018, arXiv:1811.1042720. [Google Scholar]
  34. Li, J.; Liu, Y.; Li, Q. Generative adversarial network and transfer-learning-based fault detection for rotating machinery with imbalanced data condition. Meas. Sci. Technol. 2022, 33, 045103. [Google Scholar] [CrossRef]
  35. Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  36. Yang, F.; Huang, D.; Li, D.; Zhao, Y.; Lin, S.; Muyeen, S. A novel convolutional network with a self-adaptation high-pass filter for fault diagnosis of wind turbine gearboxes. Meas. Sci. Technol. 2022, 34, 025024. [Google Scholar] [CrossRef]
  37. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkorei, J.; Jones, L.; Gomez, A.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 1–11. [Google Scholar]
  38. Niki, P.; Ashish, V.; Jakob, U.; Lukasz, K.; Noam, S.; Alexander, K.; Dustin, T. Image transformer. Int. Conf. Mach. Learn. 2018, 80, 4055–4064. [Google Scholar]
  39. Carion, N.; Massa, F.; Synnaeve, G.; Usunier, N.; Kirillov, A.; Zagoruyko, S. End to end object detection with transformers. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; pp. 213–229. [Google Scholar]
  40. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T. An image is worth 16 × 16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 4 May 2021; pp. 1–22. [Google Scholar]
  41. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, L.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  42. Touvron, H.; Cord, M.; Sablayrolles, A.; Synnaeve, G.; Jégou, H. Going deeper with image transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 32–42. [Google Scholar]
  43. Du, X.; Jia, L.; Haq, I. Fault diagnosis based on SPBO-SDAE and transformer neural network for rotating machinery. Measurement 2022, 188, 110545. [Google Scholar] [CrossRef]
  44. Suthar, V.; Vakharia, V.; Patel, V.K.; Shah, M. Detection of compound faults in ball bearings using multiscale-SinGAN, heat transfer search optimization, and extreme learning machine. Machines 2022, 11, 29. [Google Scholar] [CrossRef]
  45. He, Y.; Tang, H.; Ren, Y.; Kumar, A. A semi-supervised fault diagnosis method for axial piston pump bearings based on DCGAN. Meas. Sci. Technol. 2021, 32, 125104. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.