Next Article in Journal
Design and Kinematic Analysis of a Metamorphic Mechanism-Based Robot for Climbing Wind Turbine Blades
Previous Article in Journal
A Cutting Force Prediction Model for Corner Radius End Mills Based on the Separate-Edge-Forecast Method and BP Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent Fault Diagnosis for Cross-Domain Few-Shot Learning of Rotating Equipment Based on Mixup Data Augmentation

1
School of Information and Control Engineering, China University of Mining and Technology, Xuzhou 221116, China
2
School of Mechanical Engineering and Automation, Northeastern University, Shenyang 110819, China
3
National Engineering Laboratory for Industrial Big-Data Application Technology, Chongqing Innovation Center of Industrial Big-Data Co., Ltd., Chongqing 400707, China
*
Author to whom correspondence should be addressed.
Machines 2025, 13(9), 807; https://doi.org/10.3390/machines13090807
Submission received: 29 July 2025 / Revised: 26 August 2025 / Accepted: 27 August 2025 / Published: 3 September 2025
(This article belongs to the Section Machines Testing and Maintenance)

Abstract

Existing fault diagnosis methods assume the identical distribution of training and test data, failing to adapt to source–target domain differences in industrial scenarios and limiting generalization. They also struggle to explore inter-domain correlations with scarce labeled target samples, leading to poor convergence and generalization. To address this, our paper proposes a cross-domain few-shot intelligent fault diagnosis method based on Mixup data augmentation. Firstly, a Mixup data augmentation method is used to linearly combine source domain and target domain data in a specific proportion to generate mixed-domain data, enabling the model to learn correlations and features between data from different domains and improving its generalization ability in cross-domain few-shot learning tasks. Secondly, a feature decoupling module based on the self-attention mechanism is proposed to extract domain-independent features and domain-related features, allowing the model to further reduce the domain distribution gap and effectively generalize source domain knowledge to the target domain. Then, the model parameters are optimized through a multi-task learning mechanism consisting of sample classification tasks and domain classification tasks. Finally, applications in classification tasks on multiple sets of equipment fault datasets show that the proposed method can significantly improve the fault recognition ability of the diagnosis model under the conditions of large distribution differences in the target domain and scarce labeled samples.

1. Introduction

As a core component of the modern industrial system, rotating machinery holds an irreplaceable position in strategic industries such as high-end manufacturing, aerospace, and fine chemicals [1,2,3], with its operational status directly affecting the stability and safety of production systems. Many of the degradation processes of bearings and gears, key components of rotating machinery, stem from friction phenomena at component interfaces, a process that often manifests as a vicious cycle of friction–wear–lubrication failure [4,5]. Intensified friction accelerates the wear rate, while metal debris from wear further contaminates the lubricant, impairs lubrication effectiveness, and in turn, exacerbates friction [6,7]. This cumulative damage gradually weakens the load-carrying capacity and operational precision of the components, ultimately leading to various faults, such as bearing seizure and gear tooth breakage, and posing a serious threat to the safe and stable operation of rotating machinery.
In the context of the intelligent transformation of equipment, equipment reliability has emerged as a pivotal indicator for gauging the level of industrialization [8,9]. In accordance with this trend, the construction of an active maintenance system based on condition monitoring bears particular strategic significance. Through the precise identification of abnormal operational characteristics of machinery and the establishment of a preventive maintenance mechanism, not only can this system effectively prevent major safety accidents, but it also serves as a crucial technical approach to achieve the full life-cycle management of equipment [10,11]. It is noteworthy that intelligent fault diagnosis technology, as the core module of predictive health management, has evolved into a key element in ensuring the safe operation of industry [12,13]. Traditional intelligent fault diagnosis methods, such as artificial neural networks, support vector machines, and random forests, are highly dependent on manual extraction of fault features. These methods exhibit problems such as low efficiency, strong subjectivity, and complex feature engineering, and their generalization ability is limited when dealing with high-dimensional data [14,15]. With the development of deep learning, these methods can leverage the capabilities of automatic learning and feature extraction to overcome experience-based limitations. It mines the relationship between fault features and categories from massive data, providing more efficient and accurate decision-making support for fault identification. Therefore, deep learning has been increasingly widely applied in the field of fault diagnosis [16]. Ding et al. [17] proposed a time–frequency transformer model, which significantly improved the accuracy and robustness of fault diagnosis by capturing global and local features from the time–frequency representations of vibration signals. Li et al. [18] proposed a data-driven method based on a deep convolutional neural network (CNN), which took normalized monitoring data as input and constructed samples through time windows to realize the estimation of the remaining useful life of the system. Xu et al. [19] introduced a multi-scale coarse-graining process with continuous sliding window operations and a feature attention mechanism to process raw vibration signals, which can significantly enhance diagnostic accuracy under complex working conditions. Wang et al. [20] proposed a multi-modal fusion method based on a one-dimensional CNN by fusing vibration and acoustic data, which can achieve high-precision diagnostic results superior to single-modal sensor methods in low signal-to-noise ratio environments. Niu et al. [21] proposed an approach that, by fusing multi-source vibration data with domain knowledge graphs and introducing channel attention modules, as well as non-local attention modules, achieved multi-task bearing fault diagnosis.
In the field of industrial fault diagnosis, deep learning-based methods are widely favored for their strong ability to perform automatic feature extraction. However, these approaches are highly reliant on massive datasets to guarantee model performance and generalization [22,23,24]. In practical industrial scenarios, obtaining sufficient data samples is extremely difficult, which is mainly attributed to factors such as safety restrictions on key equipment and the long cycle of fault evolution. This data scarcity leads to the insufficient robustness of deep learning models when dealing with complex fault states, making few-shot learning a crucial breakthrough. By utilizing technologies like meta-learning and transfer learning, few-shot learning can achieve efficient learning with a small number of samples, significantly reducing the reliance on data scale [25,26,27].
However, in industrial environments, data distributions vary significantly between different devices and operating conditions. Even with the adoption of few-shot learning, traditional methods still fail to handle the problem of mismatched feature distributions in cross-domain data. To address this issue, cross-domain few-shot learning has emerged. It not only inherits the characteristic of few-shot learning, i.e., low dependence on the number of samples, but also effectively bridges the gaps in cross-domain data distribution through techniques such as domain adaptation. As a result, it provides a more universal solution for complex and changing industrial fault diagnosis scenarios [28,29]. Chai et al. [30] proposed a fault prototype adaptive network that learned representative fault prototypes through an identification module based on similarity learning and designed a fault prototype adaptive module to adapt multiple fault prototypes to the target dataset, thereby realizing cross-domain fault diagnosis. Zhang et al. [31] proposed a domain-adaptive meta-learning network with a feature-oriented dropping-supplementing module, which solved problems such as severe domain distribution differences, mismatched label spaces, and scarce labeled samples within a unified framework. Wang et al. [32] proposed a cross-domain few-shot learning method for fault diagnosis, which combined the attention mechanism and 1D CNN, aligning the distribution of the source domain and the target domain through sub-domain adaptation to realize fault state recognition. Tian et al. [33] proposed a deep feature transfer fusion framework based on hybrid domain transfer learning, which transferred data features from the source domain to the target domain to solve the problem that some few-shot working conditions were difficult to diagnose due to the small amount of data in actual industrial processes.
In cross-domain few-shot learning tasks, although both source domain and target domain data are labeled, a huge difference in feature distribution exists between them. This causes the model trained on source domain data to suffer from severe limitations in generalization ability when dealing with target domain data with such large distribution differences [34]. Moreover, when labeled samples in the target domain are scarce, existing methods struggle to fully explore the potential correlations between the source domain and target domain, resulting in the model being unable to achieve effective convergence and good generalization. As a data augmentation method based on linear interpolation, Mixup generates new training samples by linearly combining two training samples. This method can effectively expand the diversity of training samples, thereby reducing the model’s dependence on the original labeled samples. In addition, the feature distribution of samples generated by Mixup through linear interpolation lies between the original samples, which effectively alleviates the problem of feature distribution differences between different domains.
As an effective data augmentation method, Mixup has spawned multiple variant structures in various fields and has been widely applied. For instance, Campbell et al. [35] proposed a Mixup augmentation method based on facial keypoints, which significantly enhanced the ability of deep learning models to recognize facial phenotypes of rare genetic diseases, especially showing outstanding performance regarding diseases not present in the training set. Zhang et al. [36] proposed the Self-Mixup mechanism. This mechanism shifts Mixup from the cross-sample level to the intra-sample level by applying two independent augmentations to the same image to generate heterogeneous perspectives and performing Beta-distribution-based linear interpolation of their feature maps at any intermediate layer of the network, thereby synthesizing pseudo-new samples in the feature space. Compatible with traditional manifold Mixup, this mechanism can improve knowledge utilization with limited samples without external data. Zhang et al. [37] enhanced the support set by linearly interpolating query samples with same-class support samples via Mixup and combined this with bidirectional feature reconstruction, effectively mitigating the support–query bias in few-shot fine-grained classification. Xiao et al. [38] proposed a cross-modal image–text Mixup method, termed MMIT-Mixup, and coupled it with robustness-invariance fine-tuning to form the semantic-guided robustness tuning framework. By leveraging textual features as domain-invariant semantic guidance and performing Mixup at intermediate layers of the Vision Transformer (ViT) [39], this method markedly improved the robustness and generalization of large-scale pre-trained models under extreme domain shifts and few-shot settings. On this basis, relevant studies applying Mixup to cross-domain few-shot tasks have also emerged. For example, Cao et al. [40] innovatively transferred Mixup to the style statistics level for the cross-domain few-shot action recognition task, which involved adversarial mixing of channel-level mean and variance. This significantly bridged the style gap between domains and effectively improved the recognition accuracy of the target domain. Zhuo et al. [41] proposed the target-guided dynamic Mixup framework, including the Mixup-3T classification network and dynamic ratio generation network. They used a bi-level meta-learning strategy for training to adapt to the target domain and current model state, aiming to solve the problems of narrowing the domain gap and knowledge transfer.
However, to date, there has been no research that applies Mixup to the task of mechanical equipment fault diagnosis in cross-domain few-shot scenarios by focusing on the characteristics of mechanical equipment condition monitoring signals. Therefore, this paper applies Mixup to cross-domain few-shot learning and proposes an intelligent fault diagnosis method for cross-domain few-shot learning based on Mixup data augmentation. In this paper, the Mixup data augmentation approach is firstly formulated to mix source domain and target domain data in a specific proportion, constructing mixed-domain data for training. This helps the model better generalize across different domains and improves its performance on the target domain dataset. Secondly, a self-attention-based feature de-coupling module is used to separate domain-independent and domain-related features, further alleviating the problem of domain distribution differences in cross-domain few-shot learning. Then, a multi-task learning mechanism consisting of sample classification tasks and domain classification tasks is employed to further optimize the network model parameters. Finally, in the experimental verification section, the proposed fault diagnosis method based on cross-domain few-shot learning is verified on both laboratory bearing datasets and multiple sets of public data.

2. Theory of Mixup Data Augmentation

Since the proposed method is built on Mixup data augmentation, this section reviews the basic theory of this technique.
Mixup, a computer vision data augmentation method proposed in ICLR 2017 [42], generates new training data by linearly interpolating and fusing the features and labels of multiple samples. The core of Mixup data augmentation technology lies in the weighted combination of input data and labels, prompting the model to learn the smooth transition characteristics between samples, thereby enhancing the generalization performance of the model. Compared with traditional data augmentation techniques, such as rotation, cropping, and flipping, which generate new samples of the same category through geometric transformations, Mixup constructs continuous data manifolds in the feature space. The generated virtual samples not only fill the gaps in the original data distribution but also effectively suppress the phenomenon of overconfidence in prediction at category boundaries through the soft label mechanism.
From the perspective of risk minimization, the optimization of machine learning models often takes Empirical Risk Minimization (ERM) as the core idea—approximating generalization performance by minimizing the average loss of the model on limited training samples. However, ERM only focuses on the local behavior of samples, which easily leads to the model overfitting to the data. This problem becomes more prominent when the scale of model parameters is close to the amount of data.
To alleviate the limitations of ERM, Vicinal Risk Minimization (VRM) attempts to expand training data by constructing a vicinal distribution to generate virtual samples. At its core, it uses the vicinal distribution to measure the probability of finding virtual feature–target pairs within the neighborhood of training feature–target pairs, and then optimizes the model by minimizing the empirical vicinal risk.
The virtual samples generated by the Mixup technique actually construct a vicinal distribution, thereby achieving VRM for the model. Specifically, Mixup can be regarded as a VRM method, and its mathematical expression is as follows [42]:
μ a , b a i , b i = 1 n j n E λ δ a = λ a i + 1 λ a j , b = λ b i + 1 λ b j
λ B e t a α , α
where the hyperparameter α 0 , determines the shape characteristics of the Beta distribution, thereby influencing the probability distribution features of the mixing coefficient λ . E λ denotes the expectation operator with respect to the parameter λ . a i , b i and a j , b j represent two feature–target pairs selected from the original dataset, and a , b represents a virtual feature–target pair.
When α 0 , Mixup degenerates to ERM as λ approaches 0 or 1. At α = 1 , λ follows a uniform distribution over [0, 1]. When α > 1 , the Beta distribution peaks at 0.5, concentrating λ values around 0.5 and producing equal-weight fusions. In short, Mixup generates virtual feature–target pairs by randomly selecting two labeled samples a i , b i and a j , b j :
a = λ a i + 1 λ a j
b = λ b i + 1 λ b j
Its core assumption is that the linear interpolation of feature vectors should lead to the linear interpolation of targets. In general, Mixup constructs a vicinal distribution through linear interpolation, enabling the model to exhibit linear behavior in the interpolation region between training samples. This linear behavior not only reduces the prediction oscillation of the model outside the training samples but also provides the model with a simpler inductive bias, which helps to improve the generalization performance of the model. Additionally, a key characteristic of Mixup is its implicit control over model complexity. As the interpolation intensity increases, the training error on real data may rise, but the generalization error decreases significantly. This indicates that Mixup effectively balances the bias-variance trade-off by introducing virtual samples, thus implicitly regulating model complexity.

3. Cross-Domain Few-Shot Intelligent Fault Diagnosis Model Based on Mixup Data Augmentation

3.1. Overall Framework of the Proposed Method

In the intelligent fault diagnosis framework based on cross-domain few-shot learning proposed in this paper, the source domain dataset with sufficient labeled samples is defined as D S = { ( x i s , p i s ) } i = 1 n s . Among them, x i s R N represents the i-th group of labeled samples in the source domain, N denotes the length of each group of labeled samples in the source domain, p i s C S stands for the label information of the i-th labeled sample in the source domain, and n S indicates the number of groups of source domain samples. The target domain dataset with a small number of labeled samples is defined as D T = { ( x i t , p i t ) } i = 1 n t . Here, x i t R N represents the i-th group of labeled samples in the target domain, p i t C T denotes the label information of the i-th labeled sample in the target domain, and the number of groups of labeled samples in the target domain is much smaller than that in the source domain, i.e., n T n S . Compared with traditional deep learning methods [14,15,17,18,19,20,21,26], where D S and D T are sampled from the same domain, the feature distributions of D S and D T in this study show significant differences. That is to say, this study attempts to identify the health status of target domain samples by means of the fault knowledge obtained from source domain samples under the conditions of severe domain distribution differences between the source domain and the target domain, as well as the scarcity of labeled samples in the target domain.
The overall framework of the proposed intelligent fault diagnosis method based on cross-domain few-shot learning is illustrated in Figure 1. This framework leverages the Mixup data augmentation technique to generate diverse training samples and enhances the model’s cross-domain generalization capability through dual-domain collaborative training. Specifically, the process is as follows:
First, two task sets (i.e., the query set and the support set) are randomly sampled from the source domain dataset and target domain dataset, respectively. The Mixup data augmentation method is employed to perform linear interpolation on the source domain query set and target domain query set at a certain ratio, thereby generating a new mixed-domain query set, while the original support sets of both domains remain unchanged.
Second, the domain-independent features and domain-related features from the source domain support set, target domain support set, and mixed-domain query set are extracted via a feature extractor and a self-attention mechanism-based feature decoupling module.
Finally, model parameters are optimized through a multi-task loss function mechanism composed of sample classification tasks and domain classification tasks, with the utilization of the domain-independent and domain-related features derived from the source domain support set, target domain support set, and mixed-domain query set.
In the subsequent sections, the main components of the proposed method, namely the Mixup data augmentation method, the feature decoupling module, and the network parameter update method, will be introduced in detail.

3.2. Mixup Data Augmentation Method

Since the source domain data and target domain data are derived from different devices, severe domain distribution differences exist between them. To alleviate this issue, this paper employs the Mixup data augmentation method to linearly combine the sample features of the source domain and target domain at a certain ratio. By making full use of these two datasets, mixed-domain training samples are generated. Specifically, during the model training process, support sets and query sets are randomly sampled and generated from the source domain dataset and target domain dataset, respectively, namely the source domain support set D s u p p o r t S = { ( x i s , p i s ) } i = 1 K , query set D q u e r y S = { ( x i s , p i s ) } i = 1 M , target domain support set D s u p p o r t T = { ( x i t , p i t ) } i = 1 K , and query set D q u e r y T = { ( x i t , p i t ) } i = 1 M . The source domain query set and target domain query set are linearly combined proportionally to form a mixed-domain query set D q u e r y m i x = { ( x ̑ i m i x ) } i = 1 M . Therefore, the samples in the mixed-domain query set contain both source domain features and target domain features. It is worth noting that the Mixup data augmentation method used in this paper only acts on the samples of the source domain query set and target domain query set, and it does not mix the labels. Its specific expression is as follows:
D q u e r y m i x = λ D q u e r y S + 1 λ D q u e r y T
where the ratio λ [ 0,1 ] is generated from the B e t a α , α distribution.

3.3. Feature Decoupling Module Based on Self-Attention

To further reduce the domain distribution difference between the source domain and the target domain, this paper proposes a self-attention-based feature decoupling module. The module aims to extract domain-independent features for cross-domain sample classification tasks, while retaining domain-related features to incorporate important inductive biases learned from each domain.
The self-attention-based feature decoupling module proposed in this paper mainly consists of a self-attention layer, a batch normalization layer, an activation function layer, and a fully connected layer, as shown in Figure 2. By modeling the long-distance dependencies between features, self-attention enhances the representation of important features and assists the module in dissociating domain-independent features and domain-related features. The normalization layer improves the stability of the model by normalizing the input features. The ReLU layer enhances the model’s ability to express complex features by introducing nonlinear transformations. FC21a and FC22a generate the mean and logarithmic variance of domain-independent features, respectively, and the domain-independent features of the source domain support set F 1 s , mixed-domain query set F 1 m i x , and target domain support set F 1 t are obtained using the reparameterization trick shown in Equations (6) and (7). FC21b and FC22b generate the mean and logarithmic variance of domain-related features, respectively, and similarly, the domain-related features of the source domain support set F 2 s , mixed-domain query set F 2 m i x , and target domain support set F 2 t are also obtained using the reparameterization trick.
σ = exp log σ 2 2
F = μ + ε σ
where log σ 2 represents the logarithmic variance, μ represents the mean, σ represents the standard deviation, and ε N 0,1 represents the standard Gaussian noise.
The design concept of this module stems from the latent variable separation principle of Variational Autoencoder (VAE), which is a probabilistic generative model that performs two key tasks. It maps input data to a latent space with a probability distribution, and this distribution is used to represent the latent features of the data. It also reconstructs the input data from the latent representations.
In this module, it achieves more accurate feature separation through an explicit supervision mechanism. However, the feature decoupling module is quite different from VAE. VAE only encodes a single latent feature, whereas the feature decoupling module explicitly separates domain-related features and domain-independent features via a dual-branch structure.

3.4. Parameter Updating

The method proposed in this paper optimizes model parameters through a multi-task loss function mechanism composed of a sample classification task and a domain classification task.

3.4.1. Sample Classification Loss

The sample classification loss, composed of the source domain classification loss and the target domain classification loss, aims to facilitate the model to achieve effective knowledge transfer between the source and target domains during cross-domain few-shot learning.
For the source domain classification loss, the source domain feature space F s = C o n c a t F 1 m i x , F 1 s is constructed using the domain-independent features of the mixed-domain query set F 1 m i x and the domain-independent features of the source domain support set F 1 s . As indicated in Section 3.2, the samples in the mixed-domain query set contain source domain query set information at a proportion of λ , but their corresponding source domain query set labels are not processed by Mixup. When the mixed-domain domain-independent features are classified in the source domain classification task, their corresponding labels are those of the source domain query set. Therefore, the labels corresponding to the constructed source domain feature space are the real labels of the source domain. By inputting F s into the classifier and using the cross-entropy loss function to calculate the difference between the classifier’s prediction results and the real labels of the source domain, the source domain classification loss is obtained, which is defined as follows:
L S = C E G c l a F s , p s
where G c l a denotes the classifier, CE denotes the cross-entropy loss function, and p s C S denotes the source domain label.
For the target domain classification loss, the target domain feature space F t = C o n c a t ( F 1 m i x , F 1 t ) is constructed using the domain-independent features of the mixed-domain query set F 1 m i x and the domain-independent features of the target domain support set F 1 t . As stated in Section 3.2, the mixed-domain query samples contain target domain query set information at a proportion of 1 λ , and similarly, their corresponding target domain query set labels are not processed by Mixup. When the mixed-domain domain-independent features are classified in the target domain classification task, their corresponding labels are those of the target domain query set. Therefore, the labels corresponding to the constructed target domain feature space are the real labels of the target domain. By inputting F t into the classifier and using the cross-entropy loss function to calculate the difference between the classifier’s prediction results and the real labels of the target domain, the target domain classification loss is obtained, which is defined as follows:
L T = C E G c l a F t , p t
where p t C T denotes the target domain label.
Since the mixed-domain query set is formed by mixing the source domain query set and the target domain query set in proportion λ , this proportion λ represents the proportion of the source domain query set and the target domain query set in the mixed-domain query set. The model allocates the weights of the source domain classification loss and the target domain classification loss according to this proportion, thereby achieving cross-domain knowledge transfer. The final expression of the sample classification loss is as follows:
L C = λ L S + 1 λ L T

3.4.2. Domain Classification Loss

The domain classification loss comprises two components: the loss based on domain-independent features and the loss based on domain-related features. It is designed to effectively tackle the problem of domain distribution discrepancy in cross-domain few-shot learning. Herein, the domain discriminator, composed of fully connected layers, maps the input features to the scores corresponding to the two domains, thereby determining whether the input features originate from the source domain or the target domain.
For the loss related to domain-independent features, the goal is to confuse the domain discriminator, making it unable to correctly distinguish between the source domain and the target domain. The network parameters of the feature decoupling module are optimized by minimizing the following loss function, thereby effectively removing the domain-related interference information in the domain-independent features. The specific loss function is defined as
L D o m 1 = 1 3 [ K L G d o m F 1 s , p 1 s + K L G d o m F 1 t , p 1 t + K L G d o m F 1 m i x , p 1 m i x ]
where G d o m denotes the discriminator, KL represents the Kullback–Leibler divergence loss function [43], and p 1 s , p 1 t , and p 1 m i x represent the corresponding domain classification labels, which are all set to [0.5, 0.5] to confuse the domain discriminator.
For the loss related to domain-related features, the model expects the domain discriminator to correctly determine to which domain the input features belong. The network parameters of the feature decoupling module and the domain classifier are optimized by minimizing the following loss function, thereby effectively separating the domain-related features. Herein, 1 and 0 are used to represent the domain categories of the source domain and the target domain, respectively. The specific loss function is defined as
L D o m 2 = 1 3 [ C E G d o m F 2 s , p 2 s + C E G d o m F 2 t , p 2 t + λ C E G d o m F 2 m i x , p 2 m i x 1 + 1 λ C E G d o m F 2 m i x , p 2 m i x 2 ]
where p 2 s and p 2 m i x 1 are set to 1, and p 2 t and p 2 m i x 2 are set to 0.

3.4.3. Integrated Loss Function

By optimizing both the sample classification loss and the domain classification loss simultaneously, the model can learn domain-independent features and domain-related features, reducing the distribution difference between domains. The expression of the final integrated loss function is as follows:
l o s s = L C + L D o m 1 + L D o m 2

3.5. Algorithm Implement Process

Algorithm 1 outlines the implementation procedure of the proposed method. In steps 3 and 4, the source domain support set, query set, target domain support set, and query set are obtained by sampling from datasets in both the source and target domains. In step 5, the Mixup data augmentation method is employed to linearly combine samples from the source and target domain query sets at a specific ratio, generating new training samples for the mixed-domain query set. In steps 6 and 7, a self-attention-based feature decoupling module is used to generate domain-independent features and domain-related features. In step 9, the source domain classification loss is computed using the domain-independent features of the source domain support set and those of the mixed-domain query set. In step 10, the target domain classification loss is also calculated using the domain-independent features of the target domain support set and the mixed-domain query set. In step 11, the sample classification loss is calculated according to the Mixup blending ratio λ to facilitate effective knowledge transfer between the source and target domains during cross-domain few-shot learning. In step 12, the domain classification loss is computed using domain-independent features to confuse the domain discriminator, preventing it from accurately distinguishing between the source and target domains. Conversely, in step 13, the domain classification loss is calculated using domain-related features, aiming to enable the domain discriminator to correctly identify the domain of input features. Finally, model parameters are updated by minimizing the integrated loss.
It should be further noted that the proposed method leverages the ViT network, which exhibits excellent performance in classification tasks, primarily because it introduces the self-attention mechanism from the Transformer network into the field of image processing [39]. In this study, we modify the ViT network for one-dimensional monitoring data according to the methods of reference [44]. Since the original ViT is designed to process two-dimensional image signals, while this paper applies it to handle one-dimensional vibration signals, we improve the patch embedding of the original ViT network. The specific improvement process is as follows: first, we truncate the one-dimensional vibration signal into several segments with equal intervals; second, we arrange these truncated segments into a matrix form; finally, we use a fully connected network to project the reshaped matrix, thereby constructing a patch embedding tailored to one-dimensional vibration signals. We then employ this modified ViT network as the feature extractor. At the same time, we use fully connected neural networks to construct both the classifier and domain discriminator.
Algorithm 1: Cross-domain few-shot intelligent fault diagnosis algorithm based on Mixup data augmentation
Input:Source domain dataset D S = { ( x i s , p i s ) } i = 1 n s ;
Target domain dataset D T = { ( x i t , p i t ) } i = 1 n t
Output:Trained feature extractor V , Feature decoupling module F ,
Domain discriminator G d o m , Classifier G c l a
1: θ v , θ f , θ g c l a , θ g d o m = N e t i n i t i a l i z e V , F , G c l a , G d o m
//Network model parameter initialization
2:While not done do
3: D s u p p o r t S = { ( x i s , p i s ) } i = 1 K , D q u e r y S = { ( x i s , p i s ) } i = 1 M
//Sample the support set and query set from the source domain dataset
4: D s u p p o r t T = { ( x i t , p i t ) } i = 1 K , D q u e r y T = { ( x i t , p i t ) } i = 1 M
//Sample the support set and query set from the target domain dataset
5: D q u e r y m i x = λ D q u e r y S + 1 λ D q u e r y T
//Obtain the mixed-domain query set D q u e r y m i x = { ( x ̑ i m i x ) } i = 1 M
6: { ( F 1 s ) } i = 1 K , { ( F 1 t ) } i = 1 K , { ( F 1 m i x ) } i = 1 M
//Extract domain-independent features according to Equation (7)
7: { ( F 2 s ) } i = 1 K , { ( F 2 t ) } i = 1 K , { ( F 2 m i x ) } i = 1 M
//Extract domain-related features according to Equation (7)
9: L S Calculate the source domain classification loss using { ( F 1 s ) } i = 1 K , { ( F 1 m i x ) } i = 1 M according to Equation (8)
10: L T Calculate the target domain classification loss using { ( F 1 t ) } i = 1 K , { ( F 1 m i x ) } i = 1 M according to Equation (9)
11: L C = λ L S + 1 λ L T //Calculate the sample classification loss
12: L D o m 1 Calculate the loss using { ( F 1 s ) } i = 1 K , { ( F 1 t ) } i = 1 K , { ( F 1 m i x ) } i = 1 M according to Equation (11)
13: L D o m 2 Calculate the loss using   { ( F 2 s ) } i = 1 K , { ( F 2 t ) } i = 1 K , { ( F 2 m i x ) } i = 1 M according to Equation (12)
14: l o s s = L C + L D o m 1 + L D o m 2
15: θ v , θ f , θ g d o m , θ g c l a = N e t u p d a t e l o s s , θ v , θ f , θ g d o m , θ g c l a
16:end while

4. Experimental Verification

4.1. Experimental Setup

The operating environment for the experimental validation includes an Intel Core i9-12900K CPU processor, 32 GB memory with a 3.60 GHz frequency, an NVIDIA GeForce RTX 3060 GPU, the Windows 10 operating system, and the PyTorch 1.7.1 deep learning framework.
The Adam optimization algorithm is used to update the network weights, with the learning rate set to 0.0001 [45]. To reduce the randomness of the experimental results, each group of experiments is repeated five times.

4.2. Comparison Methods

To validate the effectiveness of the proposed method, this test case presents the following three comparative diagnostic methods:
(1)
Comparison Method 1 (Single ViT): This method uses the same feature extractor and classifier as the proposed method but updates network parameters exclusively using labeled samples from the source domain.
(2)
Comparison Method 2 (Prototypical Network): Leveraging a prototypical network for fault diagnosis, this method employs the same base network models for the feature extractor and classifier as the proposed method. It updates network parameters using labeled target domain samples.
(3)
Comparison Method 3 (Without Mixup): This comparative method excludes both the Mixup data augmentation and the subsequent feature processing operations used in the proposed method. It employs the same feature extractor and classifier as the proposed method, and updates network parameters using a large number of labeled samples from the source domain and a small number of labeled samples from the target domain, respectively.

4.3. Test Case 1

4.3.1. Dataset Description of Test Case 1

  • Laboratory Bearing Dataset
The laboratory rotating equipment test rig, as shown in Figure 3, primarily consists of a motor, rotor system, load block, and supporting bearings. The experimental bearings were installed in the bearing housing on the right side of the platform. During the experiment, the motor speed was set at 900 r/min and 1500 r/min, and acceleration sensors mounted horizontally on the bearing housing were used to collect bearing fault data. The sampling frequency was set at 10 kHz. The collected bearing fault data covered vibration signals for four fault types: normal state, inner ring fault, ball fault, and outer ring fault. For each bearing fault state, 100 groups of data were collected as source domain training data, with each group of fault data having a length of 10 K.
2.
Jiangnan University (JNU) Bearing Dataset
The Jiangnan University (JNU) bearing dataset [46] was collected from a fault diagnosis experiment conducted on a centrifugal fan system driven by a Mitsubishi induction motor. During the experiment, the test bearings were artificially wire-cut to simulate normal state, outer ring fault, inner ring fault, and ball fault. During the collection of fault data, the sampling frequency was set at 50 kHz. In this test case, fault data at two rotational speeds (600 r/min and 1000 r/min) were selected. For each fault state, 100 groups of data were collected, with each group of fault data having a length of 10 K.

4.3.2. Experimental Results and Analysis of Test Case 1

In this test case, a series of cross-domain diagnostic tasks are constructed from the laboratory bearing dataset to the JNU bearing dataset to evaluate the effectiveness of the proposed method. Detailed information on the experimental working condition settings is presented in Table 1.
As shown in Table 1, during the training process, the source domain data contains 400 sample groups. The purpose of this setup is to enable the model to fully learn the feature distribution of different types of fault data in the source domain. For cross-domain few-shot learning tasks, the target domain data is set to 20 or 40 sample groups to align with the actual scenario of scarce target domain data in such tasks. Additionally, all other target domain data beyond the aforementioned sets are used as testing data. This sample configuration not only covers the basic feature distribution of the four types of faults but also prevents the task from deviating from the core setting of few-shot learning due to excessive data volume.
During the network training process, the training batch size is set to 20. To construct the support set and query set, it is necessary to partition the samples within each batch. The experiment follows the principle that the support set guides feature learning, and the query set verifies generalization ability. Specifically, 5 samples are selected from each batch to form the support set, which means that K equals 5. This selection is intended to balance the diversity of the support set and its adaptability to real-world scenarios. The remaining 15 samples are used to form the query set, which means that M equals 15.
Table 2 and Table 3 present the recognition results of the proposed method and various comparative methods in test case 1, respectively. Specifically, Table 2 shows the results when 5 labeled samples per fault state are available in the target domain training set, while Table 3 corresponds to the scenario with 10 labeled target domain samples per fault state. The diagnostic results indicate that reducing the number of target domain training samples affects each diagnostic method to varying degrees, yet the proposed method demonstrates relatively stable recognition performance.
Comparison method 1 exhibits the lowest average accuracy, confirming that traditional deep learning methods struggle to converge effectively with limited labeled data. In contrast, comparison method 2, which employs a meta-learning approach, shows an increase in the average accuracy of fault state recognition and a further reduction in the fluctuation of fault state recognition accuracy under data scarcity. However, its performance still lags significantly behind that of the proposed method. When compared with comparison method 3, the Mixup data augmentation employed in the proposed method enhances the model’s generalization ability on the target domain dataset, leading to further improvements in average accuracy. This result indirectly validates the effectiveness of the Mixup technique.
To intuitively observe the feature distribution patterns of training samples and testing samples, Figure 4 presents the t-SNE feature visualization results of source domain samples and target domain samples during the training process, as well as testing samples during the testing process, in the first task when the number of target domain samples is 10. It can be seen from Figure 4a,b that during the training process, both the source domain samples and the target domain samples are correctly classified, and the distinguishability of features of different categories is obvious. It can be seen from Figure 4c that during the testing process, the clustering relationship of features of different fault categories of test samples is clear, with only very few samples being misclassified. For comparison, Figure 5 presents the feature visualization results of testing samples obtained by other comparative methods. It can be observed that the feature classification boundaries obtained by comparison method 3 are relatively clearer than those obtained by comparison methods 1 and 2. However, a small number of misclassification phenomena still occur near the classification boundaries of comparison method 3. The classification boundary corresponding to the testing samples of the proposed diagnostic method is significantly better than that of other comparative diagnostic methods, which proves that the distinguishability of the diagnostic results of the proposed method is significantly better than that of other comparative diagnostic methods.

4.4. Test Case 2

4.4.1. Dataset Description of Test Case 2

  • Case Western Reserve University (CWRU) Bearing Dataset
In the CWRU bearing dataset [47,48], bearing faults were created via electrical discharge machining to simulate bearing inner ring faults, outer ring faults, and ball faults. Different fault sizes were introduced for each fault state, specifically 0.007 inches, 0.014 inches, and 0.021 inches. Additionally, data under normal conditions were included, covering a total of 10 types. During the experiment, the sampling frequency was set to 12 kHz, and rotational speeds were adjusted by applying different motor loads of 0–3 HP, corresponding to motor speeds ranging from 1797 r/min to 1720 r/min. Data under four different working conditions were collected for each fault damage diameter. In this test case, data for the normal state under motor loads of 1 HP and 3 HP were selected, along with data for three fault states under motor loads of 1 HP and 3 HP when the fault damage diameter was 0.007 inches. For each fault state, 100 groups of fault data were collected as source domain data, with each group of fault data samples having a length of 10 K.
2.
Huazhong University of Science and Technology (HUST) Bearing Dataset
In the HUST bearing fault dataset [49], the sampling frequency was set to 25.6 kHz. Vibration signals of different fault states under 11 different rotational speeds were obtained by adjusting the motor speed, among which 10 were constant speeds and 1 was a time-varying speed. In this test case, vibration signals of four fault types (normal state, severe inner ring fault, severe outer ring fault, and severe ball fault) under constant speed conditions of 40 Hz and 60 Hz were selected as target domain data. For each faulty bearing, 100 groups of data were collected, with each group of fault data having a length of 10 K. Two sets of experiments were conducted by randomly selecting 5 and 10 groups of data as target domain training data, respectively, and the remaining data were used as testing data.

4.4.2. Experimental Results and Analysis of Test Case 2

In this test case, a series of cross-domain diagnostic tasks from the CWRU bearing dataset to the HUST bearing dataset were constructed to further evaluate the applicability of the proposed method under scenarios involving different faulty equipment. Detailed information regarding the experimental working condition settings is presented in Table 4.
Since the amounts of source domain data and target domain data during the training process in this case are the same as those in the previous case, when partitioning the support set and query set within each batch, K is still set to 5, and M is still set to 15.
Table 5 and Table 6 present the diagnostic results of the proposed diagnostic method and different comparative diagnostic methods in test case 2. Among them, Table 5 shows the recognition results when there are 5 labeled samples available for each fault state in the target domain training process, and Table 6 shows the recognition results when there are 10 labeled samples available for each fault state in the target domain training process. It can be seen that when the number of available labeled samples for each fault state in the target domain is reduced, the proposed diagnostic method still maintains the highest average accuracy of diagnostic results, even under different cross-domain diagnostic tasks. Since the proposed diagnostic method has formulated a variety of strategies for the problem of domain differences and the scarcity of labeled samples in the target domain, by comparing the diagnostic results of the proposed diagnostic method with those of the other three comparative diagnostic methods, it is found that the proposed method exhibits the highest average accuracy of fault diagnosis and the smallest fluctuation of results, which further verifies the applicability of the proposed diagnostic method in scenarios involving different faulty equipment.
To further validate the effectiveness of the proposed method, three quantitative evaluation indicators are introduced, i.e., precision, recall, and F-score. Table 7, Table 8 and Table 9 present the results of these three indicators for different fault states in test case 2, comparing the proposed diagnostic method with other baseline methods under the first task when the number of target domain samples is 10. Higher values of these indicators signify better classifier performance. The results demonstrate that the proposed method outperforms all comparative methods across all three indicators, confirming its superior fault sample classification capability across scenarios involving different faulty equipment.

5. Conclusions

To address the problems of significant domain distribution differences between source domain data and target domain data, as well as the scarcity of labeled samples in the target domain in industrial scenarios, this paper proposes a cross-domain few-shot intelligent fault diagnosis method based on Mixup data augmentation. Firstly, the method introduces the Mixup data augmentation approach, which linearly combines data from the source domain query set and the target domain query set in a specific proportion to generate a mixed query set. This enables the model to learn correlations and features between different domain data, enhancing its ability to handle domain distribution differences and improving generalization across domains, thereby boosting performance on the target domain dataset. Secondly, a self-attention-based feature decoupling module is employed. With the supervision signal from the domain discriminator, it decomposes sample features into domain-independent and domain-related features. This allows the model to acquire cross-domain transferable feature representations while retaining domain-specific knowledge, further alleviating the problem of large domain distribution differences in cross-domain few-shot learning. Then, a multi-task learning mechanism consisting of sample classification and domain classification tasks is utilized to optimize the network model parameters. Finally, experimental verification on multiple sets of bearing datasets, including laboratory and public sets, through comparative analysis with various diagnostic methods, demonstrates that the proposed method significantly enhances the fault recognition ability of the diagnostic model under conditions of large domain distribution differences and scarce labeled samples in the target domain.

Author Contributions

Conceptualization, K.Y.; methodology, K.Y. and Q.Z.; software, Y.L. and Q.Z.; validation, K.Y., Y.L. and Q.Z.; formal analysis, Q.Z.; investigation, K.Y.; resources, K.Y.; data curation, Y.L.; writing—original draft preparation, Q.Z.; writing—review and editing, K.Y., Y.L. and Y.Z.; visualization, B.X.; supervision, B.X.; project administration, Y.Z.; funding acquisition, K.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation, grant number 62206298, 62403113. This work was also supported by the Natural Science Foundation of Jiangsu Province (No. BK20221111), the China Postdoctoral Science Foundation (No. 2022M710542, 2024M750374), and the Chongqing Postdoctoral Science Foundation (No. 2022NSCQ-BHX3987).

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We would like to express our deepest appreciation for the valuable comments and suggestions provided by the reviewers.

Conflicts of Interest

Author Bin Xing was employed by the company Chongqing Innovation Center of Industrial Big-Data Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Wang, T.; Han, Q.; Chu, F.; Feng, Z. Vibration based condition monitoring and fault diagnosis of wind turbine planetary gearbox: A review. Mech. Syst. Signal Process. 2019, 126, 662–685. [Google Scholar] [CrossRef]
  2. Li, J.; Luo, W.; Bai, M. Review of research on signal decomposition and fault diagnosis of rolling bearing based on vibration signal. Meas. Sci. Technol. 2024, 35, 092001. [Google Scholar] [CrossRef]
  3. Liu, D.; Cui, L.; Cheng, W. A review on deep learning in planetary gearbox health state recognition: Methods, applications, and dataset publication. Meas. Sci. Technol. 2023, 35, 012002. [Google Scholar] [CrossRef]
  4. Dašić, M.; Almog, R.; Agmon, L.; Yehezkel, S.; Halfin, T.; Jopp, J.; Ya’akobovitz, A.; Berkovich, R.; Stankovic, I. Role of trapped molecules at sliding contacts in lattice-resolved friction. ACS Appl. Mater. Interfaces 2024, 16, 44249–44260. [Google Scholar] [CrossRef]
  5. Dašić, M.; Ponomarev, I.; Polcar, T.; Nicolini, P. Tribological properties of vanadium oxides investigated with reactive molecular dynamics. Tribol. Int. 2022, 175, 107795. [Google Scholar] [CrossRef]
  6. Gkagkas, K.; Ponnuchamy, V.; Dašić, M.; Stanković, I. Molecular dynamics investigation of a model ionic liquid lubricant for automotive applications. Tribol. Int. 2017, 113, 83–91. [Google Scholar] [CrossRef]
  7. Dašić, M.; Stanković, I.; Gkagkas, K. Molecular dynamics investigation of the influence of the shape of the cation on the structure and lubrication properties of ionic liquids. Phys. Chem. Chem. Phys. 2019, 21, 4375–4386. [Google Scholar] [CrossRef]
  8. Yan, C.; Zhao, M.; Lin, J. Fault signature enhancement and skidding evaluation of rolling bearing based on estimating the phase of the impulse envelope signal. J. Sound Vib. 2020, 485, 115529. [Google Scholar] [CrossRef]
  9. Zonta, T.; Da, C.; Rosa, R.; Lima, M.; Da, T.; Li, G. Predictive maintenance in the industry 4.0: A systematic literature review. Comput. Ind. Eng. 2020, 150, 106889. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Zhou, X.; Gao, C.; Lin, J.; Ren, Z.; Feng, K. Contrastive learning-enabled digital twin framework for fault diagnosis of rolling bearing. Meas. Sci. Technol. 2024, 36, 015026. [Google Scholar] [CrossRef]
  11. Vashishtha, G.; Chauhan, S.; Sehri, M.; Hebda-Sobkowicz, J.; Zimroz, R.; Dumond, P.; Kumar, R. Advancing machine fault diagnosis: A detailed examination of convolutional neural networks. Meas. Sci. Technol. 2024, 36, 022001. [Google Scholar] [CrossRef]
  12. Chen, B.; Shen, C.; Shi, J.; Kong, L.; Tan, L.; Wang, D.; Zhu, Z. Continual learning fault diagnosis: A dual-branch adaptive aggregation residual network for fault diagnosis with machine increments. Chin. J. Aeronaut. 2023, 36, 361–377. [Google Scholar] [CrossRef]
  13. Feng, K.; Ji, J.; Ni, Q.; Beer, M. A review of vibration-based gear wear monitoring and prediction techniques. Mech. Syst. Signal Process. 2023, 182, 109605. [Google Scholar] [CrossRef]
  14. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  15. Cen, J.; Yang, Z.; Liu, X.; Xiong, J.; Chen, H. A review of data-driven machinery fault diagnosis using machine learning algorithms. J. Vib. Eng. Technol. 2022, 10, 2481–2507. [Google Scholar] [CrossRef]
  16. Xiang, L.; Yang, X.; Hu, A.; Su, H.; Wang, P. Condition monitoring and anomaly detection of wind turbine based on cascaded and bidirectional deep learning networks. Appl. Energy 2022, 305, 117925. [Google Scholar] [CrossRef]
  17. Ding, Y.; Jia, M.; Miao, Q.; Cao, Y. A novel time-frequency transformer based on self-attention mechanism and its application in fault diagnosis of rolling bearings. Mech. Syst. Signal Process. 2022, 168, 108616. [Google Scholar] [CrossRef]
  18. Li, X.; Ding, Q.; Sun, J. Remaining useful life estimation in prognostics using deep convolution neural networks. Reliab. Eng. Syst. Saf. 2018, 172, 1–11. [Google Scholar] [CrossRef]
  19. Xu, Z.; Li, C.; Yang, Y. Fault diagnosis of rolling bearings using an improved multi-scale convolutional neural network with feature attention mechanism. ISA Trans. 2021, 110, 379–393. [Google Scholar] [CrossRef]
  20. Wang, X.; Mao, D.; Li, X. Bearing fault diagnosis based on vibro-acoustic data fusion and 1D-CNN network. Measurement 2021, 173, 108518. [Google Scholar] [CrossRef]
  21. Niu, G.; Liu, E.; Wang, X.; Ziehl, P.; Zhang, B. Enhanced discriminate feature learning deep residual CNN for multitask bearing fault diagnosis with information fusion. IEEE Trans. Ind. Inform. 2022, 19, 762–770. [Google Scholar] [CrossRef]
  22. Cao, H.; Shao, H.; Zhong, X. Unsupervised domain-share CNN for machine fault transfer diagnosis from steady speeds to time-varying speeds. J. Manuf. Syst. 2022, 62, 186–198. [Google Scholar] [CrossRef]
  23. Liu, R.; Yang, B.; Zio, E.; Chen, X. Artificial intelligence for fault diagnosis of rotating machinery: A review. Mech. Syst. Signal Process. 2018, 108, 33–47. [Google Scholar] [CrossRef]
  24. Meng, Z.; Zhan, X.; Li, J.; Pan, Z. An enhancement denoising autoencoder for rolling bearing fault diagnosis. Measurement 2018, 130, 448–454. [Google Scholar] [CrossRef]
  25. Wang, P.; Li, J.; Wang, S.; Zhang, F.; Shi, J.; Shen, C. A new meta-transfer learning method with freezing operation for few-shot bearing fault diagnosis. Meas. Sci. Technol. 2023, 34, 074005. [Google Scholar] [CrossRef]
  26. Zhu, Z.; Lei, Y.; Qi, G.; Chai, Y.; Mazur, N.; An, Y.; Huang, X. A review of the application of deep learning in intelligent fault diagnosis of rotating machinery. Measurement 2023, 206, 112346. [Google Scholar] [CrossRef]
  27. Li, C.; Li, S.; Zhang, A.; He, Q.; Liao, Z.; Hu, J. Meta-learning for few-shot bearing fault diagnosis under complex working conditions. Neurocomputing 2021, 439, 197–211. [Google Scholar] [CrossRef]
  28. Long, J.; Chen, Y.; Huang, H.; Yang, Z.; Huang, Y.; Li, C. Multidomain variance-learnable prototypical network for few-shot diagnosis of novel faults. J. Intell. Manuf. 2024, 35, 1455–1467. [Google Scholar] [CrossRef]
  29. Zhang, X.; Su, Z.; Hu, X.; Han, Y.; Wang, S. Semisupervised momentum prototype network for gearbox fault diagnosis under limited labeled samples. IEEE Trans. Ind. Inform. 2022, 18, 6203–6213. [Google Scholar] [CrossRef]
  30. Chai, Z.; Zhao, C. Fault-prototypical adapted network for cross-domain industrial intelligent diagnosis. IEEE Trans. Autom. Sci. Eng. 2021, 19, 3649–3658. [Google Scholar] [CrossRef]
  31. Zhang, Y.; Han, D.; Tian, J.; Shi, P. Domain adaptation meta-learning network with discard-supplement module for few-shot cross-domain rotating machinery fault diagnosis. Knowl.-Based Syst. 2023, 268, 110484. [Google Scholar] [CrossRef]
  32. Wang, Y.; Yan, J.; Ye, X.; Jing, Q.; Wang, J.; Geng, Y. Few-shot transfer learning with attention mechanism for high-voltage circuit breaker fault diagnosis. IEEE Trans. Ind. Appl. 2022, 58, 3353–3360. [Google Scholar] [CrossRef]
  33. Tian, Y.; Wang, Y.; Peng, X.; Zhang, W. A fault diagnosis method for few-shot industrial processes based on semantic segmentation and hybrid domain transfer learning. Appl. Intell. 2023, 53, 28268–28290. [Google Scholar] [CrossRef]
  34. Chen, W.; Liu, Y.; Kira, Z.; Wang, Y.; Huang, J. A closer look at few-shot classification. arXiv 2019, arXiv:1904.04232. [Google Scholar] [CrossRef]
  35. Campbell, J.; Dawson, M.; Zisserman, A.; Xie, W.; Nellåker, C. Deep facial phenotyping with Mixup augmentation. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis, Aberdeen, UK, 19–21 July 2023; Springer Nature: Cham, Switzerland, 2023; pp. 133–144. [Google Scholar] [CrossRef]
  36. Zhang, Y.; Gong, M.; Li, J.; Feng, K.; Zhang, M. Few-shot learning with enhancements to data augmentation and feature extraction. IEEE Trans. Neural Netw. Learn. Syst. 2024, 36, 6655–6668. [Google Scholar] [CrossRef]
  37. Zhang, Z.; Chang, D.; Zhu, R.; Li, X.; Ma, Z.; Xue, J. Query-aware cross-Mixup and cross-reconstruction for few-shot fine-grained image classification. IEEE Trans. Circuits Syst. Video Technol. 2024, 35, 1276–1286. [Google Scholar] [CrossRef]
  38. Xiao, K.; Wang, Z.; Li, J. Semantic-guided robustness tuning for few-shot transfer across extreme domain shift. In Proceedings of the European Conference on Computer Vision, Milano, Italy, 29 September–4 October 2024; Springer Nature: Cham, Switzerland, 2024; pp. 303–320. [Google Scholar] [CrossRef]
  39. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar] [CrossRef]
  40. Cao, K.; Peng, J.; Chen, J.; Hou, X.; Ma, A. Adversarial style Mixup and improved temporal alignment for cross-domain few-shot action recognition. Comput. Vis. Image Underst. 2025, 255, 104341. [Google Scholar] [CrossRef]
  41. Zhuo, L.; Fu, Y.; Chen, J.; Cao, Y.; Jiang, Y. Tgdm: Target guided dynamic Mixup for cross-domain few-shot learning. In Proceedings of the 30th ACM International Conference on Multimedia, Lisboa, Portugal, 10–14 October 2022; pp. 6368–6376. [Google Scholar] [CrossRef]
  42. Zhang, H.; Cisse, M.; Dauphin, Y.; Lopez-Paz, D. Mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar] [CrossRef]
  43. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar] [CrossRef]
  44. Yu, K.; Wang, X.; Cheng, Y.; Feng, K.; Zhang, Y.; Xing, B. Dual structural consistent partial domain adaptation network for intelligent machinery fault diagnosis. IEEE Trans. Instrum. Meas. 2024, 73, 3520413. [Google Scholar] [CrossRef]
  45. Diederik, P.; Jimmy, B. Adam: A method for stochastic optimization. arXiv 2017, arXiv:1412.6980. [Google Scholar] [CrossRef]
  46. Li, K.; Ping, X.; Wang, H.; Chen, P.; Cao, Y. Sequential fuzzy diagnosis method for motor roller bearing in variable operating conditions based on vibration analysis. Sensors 2013, 13, 8013–8041. [Google Scholar] [CrossRef]
  47. Case Western Reserve University Bearing Data Center. Available online: https://engineering.case.edu/bearingdatacenter/download-data-file (accessed on 10 October 2024).
  48. Smith, W.; Randall, R. Rolling element bearing diagnostics using the Case Western Reserve University data: A benchmark study. Mech. Syst. Signal Process. 2015, 64, 100–131. [Google Scholar] [CrossRef]
  49. Zhao, C.; Zio, E.; Shen, W. Domain generalization for cross-domain fault diagnosis: An application-oriented perspective and a benchmark study. Reliab. Eng. Syst. Saf. 2024, 245, 109964. [Google Scholar] [CrossRef]
Figure 1. An overall framework of the diagnostic method based on cross-domain few-shot learning.
Figure 1. An overall framework of the diagnostic method based on cross-domain few-shot learning.
Machines 13 00807 g001
Figure 2. The framework of feature decoupling module based on self-attention for cross-domain few-shot intelligent fault diagnosis.
Figure 2. The framework of feature decoupling module based on self-attention for cross-domain few-shot intelligent fault diagnosis.
Machines 13 00807 g002
Figure 3. Laboratory bearing fault test bench.
Figure 3. Laboratory bearing fault test bench.
Machines 13 00807 g003
Figure 4. T-SNE clustering feature distributions of the proposed method in test case 1: (a) feature distribution of source domain training samples; (b) feature distribution of target domain training samples; (c) feature distribution of testing samples.
Figure 4. T-SNE clustering feature distributions of the proposed method in test case 1: (a) feature distribution of source domain training samples; (b) feature distribution of target domain training samples; (c) feature distribution of testing samples.
Machines 13 00807 g004
Figure 5. T-SNE clustering feature distributions for comparison methods in test case 1: (a) feature distribution of testing samples for comparison method 1; (b) feature distribution of testing samples for comparison method 2; (c) feature distribution of testing samples for comparison method 3.
Figure 5. T-SNE clustering feature distributions for comparison methods in test case 1: (a) feature distribution of testing samples for comparison method 1; (b) feature distribution of testing samples for comparison method 2; (c) feature distribution of testing samples for comparison method 3.
Machines 13 00807 g005
Table 1. Test case 1 working condition settings during the training process.
Table 1. Test case 1 working condition settings during the training process.
DomainSource DomainTarget Domain
DatasetLaboratory bearing datasetJNU bearing dataset
Working
condition
SpeedNumber of samplesSpeedNumber of samples
Task 11500 r/min1001000 r/min5/10
Task 2900 r/min1001000 r/min5/10
Task 31500 r/min100600 r/min5/10
Task 4900 r/min100600 r/min5/10
Table 2. The recognition results in test case 1, where five labeled samples are available for each fault condition in the target domain.
Table 2. The recognition results in test case 1, where five labeled samples are available for each fault condition in the target domain.
MethodTask 1Task 2Task 3Task 4
Comparison method 1Average accuracy (%)76.6375.0576.3274.84
Standard deviation2.94172.78702.47923.0078
Comparison method 2Average accuracy (%)85.2483.4082.0680.53
Standard deviation1.64831.82271.76211.8890
Comparison method 3Average accuracy (%)85.9585.7987.5887.26
Standard deviation1.07291.36791.04921.8878
Proposed methodAverage accuracy (%)87.3286.2189.6991.21
Standard deviation0.58240.48780.56251.0325
Table 3. The recognition results in test case 1, where 10 labeled samples are available for each fault condition in the target domain.
Table 3. The recognition results in test case 1, where 10 labeled samples are available for each fault condition in the target domain.
MethodTask 1Task 2Task 3Task 4
Comparison method 1Average accuracy (%)83.0681.9982.3182.16
Standard deviation2.47251.83542.18872.0171
Comparison method 2Average accuracy (%)88.3990.3490.1889.17
Standard deviation1.25951.38651.32981.5010
Comparison method 3Average accuracy (%)92.6194.3992.2892.78
Standard deviation1.49481.29641.44021.2025
Proposed methodAverage accuracy (%)98.3398.6295.4496.28
Standard deviation0.39170.39310.41530.5707
Table 4. Test case 2 working condition settings during the training process.
Table 4. Test case 2 working condition settings during the training process.
DomainSource DomainTarget Domain
DatasetCWRU bearing datasetHUST bearing dataset
Working conditionLoadNumber of samplesSpeedNumber of samples
Task 11 HP10040 Hz5/10
Task 23 HP10040 Hz5/10
Task 31 HP10060 Hz5/10
Task 43 HP10060 Hz5/10
Table 5. The recognition results in test case 2, where five labeled samples are available for each fault condition in the target domain.
Table 5. The recognition results in test case 2, where five labeled samples are available for each fault condition in the target domain.
MethodTask 1Task 2Task 3Task 4
Comparison method 1Average accuracy (%)81.7981.2680.3182.04
Standard deviation2.49321.80342.41831.5228
Comparison method 2Average accuracy (%)87.4687.7988.1688.42
Standard deviation1.50141.71031.67271.5701
Comparison method 3Average accuracy (%)91.1191.8493.1691.05
Standard deviation1.07431.30051.03391.3831
Proposed methodAverage accuracy (%)95.0593.2695.4292.21
Standard deviation0.60800.39640.48630.5426
Table 6. The recognition results in test case 2, where 10 labeled samples are available for each fault condition in the target domain.
Table 6. The recognition results in test case 2, where 10 labeled samples are available for each fault condition in the target domain.
MethodTask 1Task 2Task 3Task 4
Comparison method 1Average accuracy (%)84.5084.7483.9185.01
Standard deviation1.62351.59641.58332.1690
Comparison method 2Average accuracy (%)89.9490.5092.1793.28
Standard deviation1.14401.36151.46361.5947
Comparison method 3Average accuracy (%)96.1195.6196.8996.88
Standard deviation1.06881.08740.91961.0940
Proposed methodAverage accuracy (%)98.8398.9599.5599.28
Standard deviation0.32650.20950.28280.2812
Table 7. Precision rates of various diagnostic methods on test case 2.
Table 7. Precision rates of various diagnostic methods on test case 2.
MethodNIFRFOF
Comparative method 10.73570.88531.00000.8850
Comparative method 20.82930.92430.95640.8972
Comparative method 30.95730.98191.00000.9782
Proposed method0.98900.99711.00000.9890
Table 8. Recall rates of various diagnostic methods on test case 2.
Table 8. Recall rates of various diagnostic methods on test case 2.
MethodNIFRFOF
Comparative method 10.94450.75560.90740.8519
Comparative method 20.91110.81120.96670.9278
Comparative method 30.98150.93891.00000.9889
Proposed method0.99720.97781.00001.0000
Table 9. F-Scores rates of various diagnostic methods on test case 2.
Table 9. F-Scores rates of various diagnostic methods on test case 2.
MethodNIFRFOF
Comparative method 10.82080.74760.92030.8368
Comparative method 20.80790.85340.96680.9122
Comparative method 30.95690.95481.00000.9780
Proposed method0.99170.98881.00000.9890
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yu, K.; Li, Y.; Zhan, Q.; Zhang, Y.; Xing, B. Intelligent Fault Diagnosis for Cross-Domain Few-Shot Learning of Rotating Equipment Based on Mixup Data Augmentation. Machines 2025, 13, 807. https://doi.org/10.3390/machines13090807

AMA Style

Yu K, Li Y, Zhan Q, Zhang Y, Xing B. Intelligent Fault Diagnosis for Cross-Domain Few-Shot Learning of Rotating Equipment Based on Mixup Data Augmentation. Machines. 2025; 13(9):807. https://doi.org/10.3390/machines13090807

Chicago/Turabian Style

Yu, Kun, Yan Li, Qiran Zhan, Yongchao Zhang, and Bin Xing. 2025. "Intelligent Fault Diagnosis for Cross-Domain Few-Shot Learning of Rotating Equipment Based on Mixup Data Augmentation" Machines 13, no. 9: 807. https://doi.org/10.3390/machines13090807

APA Style

Yu, K., Li, Y., Zhan, Q., Zhang, Y., & Xing, B. (2025). Intelligent Fault Diagnosis for Cross-Domain Few-Shot Learning of Rotating Equipment Based on Mixup Data Augmentation. Machines, 13(9), 807. https://doi.org/10.3390/machines13090807

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop