Next Article in Journal
An Integrated Design Method for Elderly-Friendly Game Products Based on Online Review Mining and the BTM–AHP–AD–TOPSIS Framework
Previous Article in Journal
Effect of Water Content and Cementation on the Shear Characteristics of Remolded Fault Gouge
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

KACFormer: A Novel Domain Generalization Model for Cross-Individual Bearing Fault Diagnosis

School of Mechanical & Electrical Engineering, China University of Mining and Technology-Beijing, Beijing 100083, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(14), 7932; https://doi.org/10.3390/app15147932
Submission received: 17 June 2025 / Revised: 9 July 2025 / Accepted: 14 July 2025 / Published: 16 July 2025

Abstract

Featured Application

The proposed cross-individual diagnosis framework enables bearing fault detection across different bearing types without requiring target domain data, offering a plug-and-play solution for predictive maintenance in renewable energy systems that is particularly suitable for wind turbine maintenance and other industrial applications where labeled fault samples are scarce.

Abstract

Fault diagnosis methods based on deep learning have been widely applied to bearing fault diagnosis. However, current methods usually diagnose on the same individual device, which cannot guarantee reliability in real industrial scenarios, especially for new individual devices. This article explores a practical cross-individual scenario and proposes a Kolmogorov–Arnold enhanced convolutional transformer (KACFormer) model to improve both general feature representation and cross-individual capabilities. Specifically, the Kolmogorov–Arnold representation theorem is embedded into convolution and multi-head attention mechanisms to develop novel Kolmogorov–Arnold enhanced convolution (KAConv) and Kolmogorov–Arnold enhanced attention (KAA). The adaptive activation function enhances its nonlinear modeling ability. Comprehensive experiments are performed on two public datasets, demonstrating the superior generalization of the proposed KACFormer model with a higher accuracy of 95.73% and 91.58% compared to existing advanced models.

1. Introduction

Bearing fault diagnosis is critical for the smooth operation of rotating machinery, which plays an essential role in various industries, from power generation to manufacturing. Bearings are fundamental components that facilitate rotation and support mechanical systems. In recent years, many scholars have researched how to extract information from vibration signals for fault diagnosis [1,2], proposing numerous advanced methods [3], among which intelligent fault diagnosis based on deep learning has garnered particular attention in recent years [4,5].
Zhang et al. [6] first designed a deep Convolutional Neural Network (CNN) with wide first-layer kernels (WDCNN) for fault diagnosis under the source domain setting, where both training and testing data were collected from the same set of bearing individuals. This architecture was tailored to enhance the network’s ability to capture long-range dependencies and extract detailed features from the vibration signals. Further, they proposed a convolutional neural network with training interference (TICNN) [7] to enhance performance under noisy environmental conditions and different working loads, which better meets the need of practical industrial scenarios. Lin et al. [8] proposed a novel meta-learning framework called generalized MAML to analyze the acceleration and acoustic signals obtained from bearings under different working conditions. Xu et al. [9] introduced a zero-shot fault semantics learning model for diagnosing compound faults. Additionally, Chen et al. [10] developed a dual adversarial learning-based multi-domain adaptation network for collaborative fault diagnosis across both bearings and gearboxes, with comparative results confirming its superior performance and effectiveness over other methods. Most current intelligent fault diagnosis models are trained and tested within the same individual. However, in practical industrial scenarios, the test objects are usually new components with potential individual differences [11], including manufacturing differences, production inconsistencies, assembly issues, and variations in fault manifestation. Consequently, differences in data distribution between the test and training individuals may hinder the model’s ability to accurately perform fault classification. Recently, scholars have begun to focus on more practical cross-domain issues beyond cross-condition scenarios, such as cross-machine problems [12,13], which are more closely aligned with real-world industrial applications. Cross-individual scenarios are practical, yet they remain insufficiently explored. This presents a challenging task that requires further attention [14,15].
To demonstrate the challenge, the widely cited Case Western Reserve University (CWRU) bearing dataset serves as an illustrative example. Vibration signals corresponding to three distinct bearing fault types, Ball, Inner Race (IR), and Outer Race (OR) from the drive end (DE), were segmented using a sliding window approach to construct the source domain dataset. Respectively, Figure 1a–c and Figure 2a–c present the t-SNE [16] (t-Distributed Stochastic Neighbor Embedding) visualization of the training and testing data distribution for the in-domain task and cross-individual task, where the blue markers represent the dimensionally reduced training set samples and the red markers denote the test set samples in the embedded space.
The t-SNE (t-Distributed Stochastic Neighbor Embedding) visualization reveals the following: (1) compact clustering of same-individual samples (intra-domain consistency), and (2) pronounced distribution shifts between the source (training) and target (testing) individuals, empirically demonstrating the cross-individual generalization challenge in bearing fault diagnosis.
Inspired by the Kolmogorov–Arnold representation theorem (KART) [17], Liu et al. proposed Kolmogorov–Arnold networks (KANs) as promising alternatives to multi-layer perceptrons (MLPs). With a fixed activation function and linear weights replaced by a univariate function parametrized as a spline, KANs tend to show a better nonlinearity expressive capability. This groundbreaking architecture offers promising potential for enhancing model generalization and accuracy. S. Ni et al. [18] applied KANs to sEMG signal classification and achieved exceptionally high accuracy on three datasets, while Cheon et al. [19] employed KANs in remote sensing to improve both efficiency and performance in RS applications. Building on these advancements, KANs hold promise for bearing fault diagnosis based on vibration signals. Considering that not all segments of the spectrum contribute equally to the fault characteristics [20] and the inherent periodicity of vibration signal data, convolution operations and attention mechanisms each offer distinct advantages. The translational invariance of convolution enables the detection of similar local features at different positions, which is crucial for capturing periodic patterns in vibration signals. On the other hand, the attention mechanism, by weighing based on contextual information, captures long-range dependencies and important global information. So far, the architecture of combining CNN and Transformer has achieved remarkable progress in the field of fault diagnosis to enhance diagnostic accuracy and robustness [21,22]. Therefore, we consider integrating KART into classic CNN and Self-attention modules, proposing two novel modules, KAConv and KAA. This aims to improve the model’s capacity to learn fault types by boosting its ability to capture nonlinear features. The main contributions of this paper are as follows:
(1)
Cross-individual scenarios are explored to provide a reasonable evaluation framework. Under this benchmark, models’ cross-individual generalization capability in real-world scenarios is effectively evaluated.
(2)
A novel Kolmogorov–Arnold enhanced convolutional transformer (KACFormer) model is proposed, which is inspired by KART. KAConv and KAA are designed to improve both general feature representation and cross-individual capabilities.
(3)
Comprehensive comparative experiments are conducted on two public datasets, and experiment results demonstrate the effectiveness and superiority of the proposed method.
The structure of this paper is as follows: Section 2 introduces the method proposed in this paper. Section 3 carries on the experiment and analysis from different aspects. Section 4 gives the main conclusion.

2. Proposed Methods

2.1. Problem Definition

This study focuses on addressing the challenge of cross-individual bearing fault diagnosis by developing a deep learning model capable of recognizing variable fault classes. The primary objective is to train the model using labelled vibration signal data from a diverse set of bearing instances. To be specific, the bearing set used for training  S t r a i n i n g  comprises bearing instances  n 1 , n 2 , n t 1 , whereas the test bearing set    S t e s t  consists of bearing instances  m 1 , m 2 , m t 2 . It is ensured that the intersection of these two sets is an empty set, i.e., S t r a i n i n g     S t e s t = . Simultaneously, the sets of fault classes in both the training and test bearing datasets fully overlap and are drawn from the same fault set  C f = 1 , 2 , , n f a u l t , where  n f a u l t  represents the total number of fault categories. From the two sets of bearing instances described above, the training set and the test set are constructed as follows:
D t r a i n = x i t r a i n , y i t r a i n i = 1 , 2 , , N t r a i n } ,
D t e s t = x i t e s t , y i t e s t i = 1 , 2 , , N t e s t } ,
where  y i t r a i n , y i t e s t C f . Each  x i t r a i n  and  x i t e s t  is an  l s e q -dimensional vibration signal vector, i.e., x i t r a i n , x i t e s t R l s e q . Let the model be a function  F x ; Θ : R l s e q C f , where  Θ  denotes the set of trainable parameters. The key objective is to determine the parameter set  Θ  that maximizes the classification accuracy on  D t e s t . The accuracy on the test set can be defined as
A D t e s t ; Θ = 1 N t e s t j = 1 N t e s t I ( F x ; Θ = y i t e s t ) ,
where  I ·  is the indicator function, which is defined as follows:
I F x ; Θ = y i t e s t = 1 ,     i f   F x ; Θ = y i t e s t 0 ,     i f   F x ; Θ y i t e s t
The optimization problem could be formulated as follows:
Θ * = arg max Θ A D t e s t ; Θ .
This formulation ensures the model achieves reliable fault detection and diagnosis for the bearing individuals in the target set.

2.2. Overall Framework

As shown in Figure 3, the overall framework of our work is divided into three parts: data processing and dataset creation, model training and training process monitoring, and model testing and performance evaluation.
(1)
The task sets are constructed following the methodology in Section 2.1. The validation set is obtained by sampling a specific proportion from the source domain data, ensuring no overlap with the training set.
(2)
The KACFormer model is trained for  n n u m _ e p o c h  on the training set. At the end of each epoch, the model is evaluated with the validation set. Notably, target individuals remain unseen by the model throughout the training process. Additionally, only the model that achieves the highest accuracy on the validation set will be saved.
(3)
After training for  n n u m _ e p o c h  epochs, the best-performing model is tested on the test set made from distinct individuals. Cross-individual accuracy and confusion metrics are utilized to assess the model’s performance.

2.3. Proposed Modules

2.3.1. 1D-KAConv

Convolutional computation provides translational invariance, enabling the detection of similar patterns regardless of their position in the sequence [23]. Additionally, its local receptive fields facilitate the effective extraction of short-term patterns, peaks, and transient events. However, despite these strengths, conventional convolutional filters are fundamentally linear operators, limiting their ability to capture the intricate nonlinear relationships that are inherent in complex time-series data. This shortfall can hamper the model’s overall representational power when dealing with signals exhibiting nonstationary characteristics. The success of prior works motivates our integration of nonlinear mappings within the Kanlinear module [24]. The incorporation of the Kanlinear computational module effectively addresses the limitations of linear convolutional layers in modeling nonlinear relationships, thereby enhancing the model’s suitability for complex cross-individual tasks. The KAConv kernel is the essence of the KAConv module. Figure 4 shows the computation process.
Considering the optimizability of KANs, an integrated neuron connected by residual activation, the Kanlinear unit, is designed. Given an input x, the output of the Kanlinear unit is the weighted sum of a trainable spline function and a fixed activation function,
ϕ x = w b b x + w s s p l i n e x ,
where  w b  and  w s  are weighs of corresponding residual branch. To be specific, the  b x  here is a  S i L U x  function:
b x = S i L U x = x 1 + e x .
This unique kernel processes relationships within the receptive field and progressively maps the initial signals to a lower dimensional KA feature space. The outputs of step  t  of a KAConv kernel with a size of  K  can be formulated as
y t = i = 0 K ( w t + i b S i L U x t + i + w t + i s s p l i n e x t + i )
where  x t + i  is the input signal at time  ( t + i ) w t + i b  is the weight for the  S i L U x  part at position  i  of the kernel, and  w t + i s  is the weight for the  s p l i n e x t + i  part at position  i  of the kernel. Based on this, we define the 1D multi-channel KAConv operation as
K A C o n v ( X ) : R n 1 × l 1 R n 2 × l 2
where  n 1  and  l 1  are the number of channels and the sequence length of the input data, respectively, while  n 2  and  l 2  correspond to the number of channels and the sequence length of the output feature.

2.3.2. Channel-Wise Multihead KAA

Multi-Head Self-attention mechanism is a key defining characteristic of transformer models [25]. The Self-attention mechanism captures global dependencies within data, enabling each element in a sequence to interact with all others, which ensures the comprehensive understanding and integration of context across the entire sequence. By mapping input signals into a space with the same dimension, the KA-Attention (KAA) mechanism enhances the model’s nonlinear expressive power, empowering models to uncover and leverage complex patterns, which is shown in Figure 5.
Given an input sequence represented by a matrix  X R n × d , where  n  is the sequence length and  d  is the dimensionality of each input vector, the KAA mechanism computes three distinct matrixes: KAQueries ( Q K A ), KAKeys ( K K A ), and KAValues ( V K A ). These matrixes are derived through learned linear transformations of the input:
Q K A = X W Q K A ,
K K A = X W K K A ,
V K A = X W V K A ,
where  W Q K A R d × d k W K K A R d × d k , and  W V K A R d × d V  are learnable Kanlinear matrixes for  Q K A K K A , and  V K A , respectively. To be specific, they can be formulated as matrixes:
W Q K A = ϕ 11 q ϕ 1 d k q ϕ d 1 q ϕ d d k q ,
W K K A = ϕ 11 k ϕ 1 d k k ϕ d 1 k ϕ d d k k ,
W V K A = ϕ 11 v ϕ 1 d k v ϕ d 1 v ϕ d d k v ,
where each  ϕ  is a learnable Kanlinear unit. Once  Q K A K K A , and  V K A  are computed, the KAA mechanism calculates the attention scores:
K A A Q K A , K K A , V K A = s o f t m a x Q K A K K A T d k V K A ,
where the softmax function ensures that the attention weights are normalized, facilitating a weighted aggregation of the values  V K A    based on their relevance to each query  Q K A .
Through the feature extraction operations of the preceding KAConv layers, the input signal is transformed into multi-channel data. To capture and optimize the distinct characters of each channel, the channel-wise multi-head KAA (CMKAA) is designed. Signals from each channel are assigned to different KAA heads. The outputs of each head are concatenated, which is shown in Figure 6. Formally, this process can be formulated as
C M K A A X = C o n c a t K A A 1 ( x 1 ) , K A A h ( x h ) ,
where  x i  is the output sequence of the  i th channel and  h  is the number of heads.

2.4. KACFormer Model

The overall structure of the KACFormer model is illustrated in Figure 7. Initially, the input signal is encoded into multi-channel features with the KAConv Encoder. The encoder consists of two KAConv Blocks. Each block is equipped with a KAConv layer, a SiLU activation layer, and a pooling layer. Note that the normalizing operation is dropped intentionally. According to the relevant literature, large vibration amplitudes are indicative of bearing faults [19]. Therefore, a maxpooling layer is employed in the first KAConv Block to accentuate peak characteristics. To reduce the overfitting risk and the impact of irrelevant high activations caused by noise, an avgpooling layer is employed in the second KAConv Block. The ith KAConv Block computation can be formally written as below:
B l o c k K A C o n v i X = P o o l i n g S i L U K A C o n v X .
Once the row signal is encoded, it is sent into the transformer block. To better align with the KATR, adjustments are made based on the classic transformer architecture by Vaswani, Ashish, et al. [25]. Considering the volume of data, the transformer architecture is designed to be shallow with only one layer to avoid the risks of overfitting, improving its model generalization capacity. Meanwhile, a KA feed-forward (KAFF) layer with SiLU activation is designed, mapping its input to a vector with the same dimension:
K A F F ( x ) = S i L U w b S i L U x + w s s p l i n e x .
The residual structures of the classic transformer are retained to keep the optimizability of the model. Given an input matrix  X n c h a n n e l × l s e q , let
R e s = L a y e r n o r m ( C M K A A X + X ) ,
The output of the CMKAA transformer can be written as follows:
O u t p u t = L a y e r n o r m K A F F R e s + R e s .

3. Experiments

To validate the effectiveness and the necessity of the proposed modules in the task, comprehensive experiments on two open datasets are conducted, including ablation study and contrast experiments. The code is written in Python 3.10.14 with Pytorch 2.4.1+cu124 on the computer, manufactured by Lenovo in Beijing, China, with an NVIDIA GeForce RTX 4060 Laptop GPU.

3.1. Task Setups

3.1.1. Task 1 on PU Dataset

The dataset for the first cross-individual task is sourced from the PU dataset from the Paderborn University Konstruktions und Antriebstechnik (KAt) data center [26]. Vibration signals from the bearing housing are collected using piezoelectric accelerometers at a sampling frequency of 64 kHz. The experimental test platform, including a drive motor, load motor, flywheel, torque sensor, and module for testing bearing vibration data, is shown in Figure 8.
A total of 32 different bearings are included in the PU dataset. These bearings vary in terms of their manufacturer, type, and severity of damage, aligning with the objectives of our cross-individual study. The bearing faults in the dataset are generated using two primary methods: electrical discharge machining (EDM) and life acceleration treatment. To ensure the experiments closely resemble real-world industrial scenarios, only the data obtained through life acceleration treatment is utilized. The fault positions considered in this study include Healthy (N), Outer Race (OR), and Inner Race (IR). Additionally, to avoid an imbalanced dataset, only a subset of the provided bearings is used. The chosen bearings for the training and validation sets are K001 (N), K002 (N), K003 (N), KA04 (OR), KA15 (OR), KA16 (OR), KI04 (IR), KI14 (IR), and KI16 (IR). For the testing sets, we selected K006 (N), KA22 (OR), and KI21 (IR).
All data is collected under the same working conditions, with a rotational speed of 900 rpm, load torque of 0.7 Nm, and radial force of 1000 N. The official working condition code is “N09_M07_F10”. Detailed information about the chosen bearings is provided on Table 1.

3.1.2. Task 2 on CWRU Dataset

The dataset for Task 2 is derived from the highly cited CWRU dataset [27]. The test bench consists of a 2 HP motor, a torque transducer, and a power dynamometer. The accelerometers are mounted on the drive side and the fan side of the housing to collect vibration signals, as shown in Figure 9.
The CWRU dataset provides two distinct types of bearings for cross-individual task research, namely the drive end (DE) bearing (6205-2RS JEM SKF; deep groove ball bearing) and the fan end (FE) bearing (6203-2RS JEM SKF; deep groove ball bearing). According to official records, the parameters for the two different bearings are listed in Table 2.
The CWRU dataset provides a relatively smaller scale of data for cross-individual fault diagnosis research. According to the relevant literature [28], models tend to face greater challenges in distinguishing fault classes when subjected to higher loads. Regarding the difficulty caused by a relatively insufficient data volume, Task 2 was set on the working condition with 0 HP load. In Task 2, we utilize data collected from the DE bearings to train the model, while data from the FE bearings is used to test the model. Task 2 is also a three-class classification task, including Healthy, OR fault and IR fault.

3.1.3. Other Settings

The sliding window has a length of 1024 and an overlapping rate of 0.5. All models are optimized using the Adam optimizer [29] with the learning rate set to 0.001. Cross-entropy is used as the loss function to complete backpropagation and parameter updates for the models. To prevent model overfitting and achieve optimal performance on cross-individual tasks, multiple training runs are performed with different numbers of epochs, and the epoch number that yields the best accuracy on the target bearings is selected. Through experiments, the optimal number of epochs for different models ranges from 5 to 20. Additionally, to avoid the randomness of a single experiment, each model is run 15 times on the same task, and the average performance is taken as the result. The distribution of the training, validation, and testing data for both tasks is presented in Table 3. As shown in Figure 10, the statistical analysis using Shapiro–Wilk normality testing and confidence interval estimation demonstrates that 15 experimental repetitions yield sufficiently reliable results, with a narrow 95% CI [94.69%, 96.77%] and low variability (CV = 1.90%).

3.2. Model Selection

The hyperparameters of the model are listed in Table 4.

3.3. Ablation Experiment

To evaluate the efficacy of the proposed modules, the ablation experiment is carried out. The models participating in the ablation experiment are as follows:
(A)
CNN+ KAConv. This model has the same structure as the classic WDCNN model. However, all conventional convolution modules are replaced by KAConv modules.
(B)
KACFormer–Transformer. The output of KAConv Encoder is sent directly into the MLP classifier.
(C)
KACFormer–KAA. This model is equipped with the WD-KAConv Encoder, followed by a traditional transformer layer.
(D)
KACFormer–KAConv. Replace the WD-KAConv Encoder with the same structured convolution encoder.
(E)
Non-channel-wise KACFormer (N-KACFormer). Assign more channels to each KAA head.
(F)
KACFormer. The proposed model.

3.4. Contrast Experiment

To validate the advancement of our proposed KACFormer model, we conducted a comprehensive comparative experiment involving both traditional classic models and modern advanced models, ranging from CNN-based methods to contemporary Transformers. The compared models include classic CNN architecture as well as frameworks that share a similar CNN-Transformer structure to our approach. Below are detailed descriptions of the contrasted models:
(A)
WDCNN. Pioneering work in the field of fault diagnosis, serving as the baseline model [6].
(B)
CNN-LSTM. A fault diagnosis model utilizing the CNN module and the LSTM module with 100 neurons set for calculation [30].
(C)
Convformer. A network designed to extract robust features by integrating both global and local information, with the goal of enhancing the end-to-end fault diagnostic performance of gearboxes under high levels of noise [21]. To better adapt to the mission, the small version of the Conformer is utilized.
(D)
Liconvformer. A novel advanced lightweight fault diagnosis model with separable multiscale convolution and broadcast Self-attention [22].
(E)
Clformer. A transformer based on convolutional embedding and Self-Attention mechanisms [31].
(F)
MobileNetV2. This neural network enhances the state-of-the-art performance of mobile models across various tasks, benchmarks, and a wide range of model sizes [32].
(G)
ResNet18. ResNet18 is a convolutional neural network architecture known for its use of residual connections, enabling efficient training of deeper models while maintaining strong performance in image classification and other computer vision tasks [33].
(H)
KACFormer. The proposed model.

3.5. Results and Analysis

3.5.1. Ablation Results

The ablation results are listed in Table 5 and Figure 11. The corresponding confusion matrixes are listed in Figure 12a–c and Figure 13a–c.

3.5.2. Contrast Results

The results of the experiment are listed in Table 6 and Figure 14. The confusion matrixes are listed in Figure 15a–d and Figure 16a–d.

3.5.3. Analysis

Other Evaluations
While fault diagnosis accuracy remains the primary industrial concern, a comprehensive model evaluation requires multi-dimensional metrics. In addition to accuracy and confusion matrix metrics, we have incorporated other domain-generalization-specific evaluations: Precision, Recall, and F1-score (all macro-averaged). The relevant data is shown in Table 7.
Learning Dynamics
To obtain a deeper understanding of the learning dynamics of KACFormer in cross-individual learning tasks, the following experiments are conducted: the model is tested on the testing set to observe phenomena such as convergence and potential overfitting after each epoch. The number of epochs is set to be 100.
As illustrated in Figure 17, on Task 1, the KACFormer network achieves its best accuracy on the target domain when it begins to converge on the source domain, typically around 5 to 20 epochs. After the 20th epoch, as training continues, the model’s performance on the target domain begins to degrade. Such a phenomenon is not as pronounced on Task 2. As illustrated in Figure 18, the model converges quickly and demonstrates relatively strong stability on the target dataset. However, the rising validation loss may indicate a potential over-fitting risk. Such a difference may be due to the volume of data and the inherent differences between the individuals in target and training datasets. Moreover, as shown in the confusion matrixes listed in Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16, the features of healthy bearings in the PU and CWRU dataset are more distinct compared to the other two categories, and models participating in the experiment demonstrate strong cross-domain performance. However, there is a significant degree of confusion between Inner Race (IR) and Outer Race (OR) faults.
Why SiLU
In the model, the SiLU activation function is employed in the encoder based on experimental validation. This implementation is consistent with the original KAN authors’ use of SiLU activation in their work [16], further validating our design choice. As demonstrated in Table 8, using ReLU activation resulted in accuracy degradation in both tasks compared to SiLU. This performance gap may be attributed to SiLU’s superior properties: smoothness, non-monotonicity, and boundedness below (unbounded above), which make it particularly suitable for approximating smooth functions, and this may be a latent characteristic exhibited in vibration signal processing.
Working Condition
Working condition variability is a widely discussed challenge in fault diagnosis, with numerous scholars dedicated to cross-condition diagnosis [34,35,36]. Since model performance may vary under different operating conditions, considering data availability, experiments are conducted on Task 1 based on the PU dataset—evaluating the same bearing samples under different conditions. Table 9 presents KACFormer’s accuracy for Task 1 across these three distinct working conditions.
From the experimental data, it can be concluded that when the source domain and target domain individuals work under the same working conditions, the robustness of the algorithm will not be affected.

4. Conclusions

In this work, a cross-individual bearing fault diagnosis scenario is explored. By integrating KART into traditional convolution and attention mechanisms, a novel model, KACFormer, is proposed to explore the potential of leveraging the nonlinear modeling capabilities and individual generalization of KANs. KACFormer enhances feature representation while preserving the translation invariance of convolution and the context-aware semantic modeling ability of attention mechanisms. Extensive experiments on two public datasets demonstrate the superior generalization performance of KACFormer, achieving state-of-the-art accuracy rates of 95.73% and 91.58% in cross-individual scenarios. These results highlight the potential effectiveness of KACFormer in improving fault diagnosis under varying operational conditions and individual differences. Beyond bearing fault diagnosis, the proposed KACFormer framework exhibits potential for generalization to other similar industrial scenarios involving vibration signal processing and domain adaptation, such as health monitoring for gearboxes and wind turbine condition monitoring [37].
Despite the advancement achieved, limitations remain. (1) Compared to MLP-based models, KANs-embedded models take a relatively longer time to train. Further study will be carried out to reduce the complexity of the model while maintaining the cross-individual accuracy. (2) Furthermore, the model should prioritize detecting smaller, incipient faults to better align with industrial requirements, as early-stage fault diagnosis is crucial for predictive maintenance and to minimize machine downtime. However, due to limitations in the current datasets, experiments regarding the model’s minimum detectable fault size were not conducted. This constitutes an important direction for future research.

Author Contributions

Conceptualization, S.S.; methodology, S.S.; software, S.S.; validation, M.X. and P.L.; investigation, P.Y.; resources, J.Y.; writing—original draft preparation, S.S.; writing—review and editing, S.S.; visualization, P.Y. and T.W. supervision, S.S.; project administration, J.Y.; funding acquisition, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

The work is supported by the National Natural Science Foundation of China, General Project, 52374167.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CNNConvolutional Neural Network
WDCNNCNNs with Wide First-Layer Kernels
TICNNWDCNN with Training Interference
MLPMulti-Layer Perceptron
KARTKolmogorov–Arnold Representation Theorem
KANKolmogorov–Arnold Network
CWRUCase Western Reserve University
PUPaderborn University
FEFan End
DEDrive End
t-SNE t-Distributed Stochastic Neighbor Embedding

References

  1. Gupta, A.; Onumanyi, A.J.; Ahlawat, S.; Prasad, Y.; Singh, V.; Abu-Mahfouz, A.M. DAT: A robust Discriminant Analysis-based Test of unimodality for unknown input distributions. Pattern Recognit. Lett. 2024, 182, 125–132. [Google Scholar] [CrossRef]
  2. Martínez-Rego, D.; Fontenla-Romero, O.; Alonso-Betanzos, A.; Principe, J.C. Fault detection via recurrence time statistics and one-class classification. Pattern Recognit. Lett. 2016, 84, 8–14. [Google Scholar] [CrossRef]
  3. Chen, C.; Pan, H.; Tong, J.; Wang, Y.; Cheng, J.; Zheng, J. Multi-task Twin Pinball Kernel Matrix Classifier for Pattern Classification. IEEE Trans. Instrum. Meas. 2025, 74, 3524109. [Google Scholar]
  4. Chen, Z.; Liu, J.; Du, Z.; Fan, X.; Luo, H. A Noise-Resilient Fault Diagnosis Method Based on Optimized Residual Networks. IEEE Trans. Instrum. Meas. 2025, 74, 3532410. [Google Scholar] [CrossRef]
  5. Jiao, J.; Zhao, M.; Lin, J.; Liang, K. A comprehensive review on convolutional neural network in machine fault diagnosis. Neurocomputing 2020, 417, 36–63. [Google Scholar] [CrossRef]
  6. Zhang, W.; Peng, G.; Li, C.; Chen, Y.; Zhang, Z. A new deep learning model for fault diagnosis with good anti-noise and domain adaptation ability on raw vibration signals. Sensors 2017, 17, 425. [Google Scholar] [CrossRef] [PubMed]
  7. Zhang, W.; Li, C.; Peng, G.; Chen, Y.; Zhang, Z. A deep convolutional neural network with new training methods for bearing fault diagnosis under noisy environment and different working load. Mech. Syst. Signal Process. 2018, 100, 439–453. [Google Scholar] [CrossRef]
  8. Lin, J.; Shao, H.; Zhou, X.; Cai, B.; Liu, B. Generalized MAML for few-shot cross-domain fault diagnosis of bearing driven by heterogeneous signals. Expert Syst. Appl. 2023, 230, 120696. [Google Scholar] [CrossRef]
  9. Xu, J.; Liang, S.; Ding, X.; Yan, R. A zero-shot fault semantics learning model for compound fault diagnosis. Expert Syst. Appl. 2023, 221, 119642. [Google Scholar] [CrossRef]
  10. Chen, X.; Shao, H.; Xiao, Y.; Yan, S.; Cai, B.; Liu, B. Collaborative fault diagnosis of rotating machinery via dual adversarial guided unsupervised multi-domain adaptation network. Mech. Syst. Signal Process. 2023, 198, 110427. [Google Scholar] [CrossRef]
  11. Fan, J.; Qi, Y.; Liu, L.; Gao, X.; Li, Y. Application of an information fusion scheme for rolling element bearing fault diagnosis. Meas. Sci. Technol. 2021, 32, 075013. [Google Scholar] [CrossRef]
  12. Zhou, Y.; Dong, Y.; Zhou, H.; Tang, G. Deep dynamic adaptive transfer network for rolling bearing fault diagnosis with considering cross-machine instance. IEEE Trans. Instrum. Meas. 2021, 70, 3525211. [Google Scholar] [CrossRef]
  13. Qian, Q.; Qin, Y.; Luo, J.; Wang, Y.; Wu, F. Deep discriminative transfer learning network for cross-machine fault diagnosis. Mech. Syst. Signal Process. 2023, 186, 109884. [Google Scholar] [CrossRef]
  14. He, Y.; Shen, W. Msit: A cross-machine fault diagnosis model for machine-level cnc spindle motors. IEEE Trans. Reliab. 2023, 73, 792–802. [Google Scholar] [CrossRef]
  15. He, Y.; Shen, W. MSRCN: A cross-machine diagnosis method for the CNC spindle motors with compound faults. Expert Syst. Appl. 2023, 233, 120957. [Google Scholar] [CrossRef]
  16. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
  17. Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Hou, T.Y.; Tegmark, M. Kan: Kolmogorov-arnold networks. arXiv 2024, arXiv:2404.19756. [Google Scholar] [PubMed]
  18. Al-Qaness, M.A.A.; Ni, S. TCNN-KAN: Optimized CNN by Kolmogorov-Arnold Network and Pruning Techniques for sEMG Gesture Recognition. IEEE J. Biomed. Health Inform. 2024, 29, 188–197. [Google Scholar] [CrossRef] [PubMed]
  19. Cheon, M.; Mun, C. Combining KAN with CNN: KonvNeXt’s Performance in Remote Sensing and Patent Insights. Remote Sens. 2024, 16, 3417. [Google Scholar] [CrossRef]
  20. Li, X.; Zhang, W.; Ding, Q. Understanding and improving deep learning-based rolling bearing fault diagnosis with attention mechanism. Signal Process. 2019, 161, 136–154. [Google Scholar] [CrossRef]
  21. Han, S.; Shao, H.; Cheng, J.; Yang, X.; Cai, B. Convformer-NSE: A novel end-to-end gearbox fault diagnosis framework under heavy noise using joint global and local information. IEEE/ASME Trans. Mechatron. 2022, 28, 340–349. [Google Scholar] [CrossRef]
  22. Yan, S.; Shao, H.; Wang, J.; Zheng, X.; Liu, B. LiConvFormer: A lightweight fault diagnosis framework using separable multiscale convolution and broadcast self-attention. Expert Syst. Appl. 2024, 237, 121338. [Google Scholar] [CrossRef]
  23. Huang, S.Y.; An, W.J.; Zhang, D.S.; Zhou, N.R. Image classification and adversarial robustness analysis based on hybrid quantum–classical convolutional neural network. Opt. Commun. 2023, 533, 129287. [Google Scholar] [CrossRef]
  24. Majidiyan, H.; Enshaei, H.; Howe, D.; Wang, Y. An integrated framework for real-time sea-state estimation of stationary marine units using wave buoy analogy. J. Mar. Sci. Eng. 2024, 12, 2312. [Google Scholar] [CrossRef]
  25. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30. Available online: https://proceedings.neurips.cc/paper/2017/hash/3f5ee243547dee91fbd053c1c4a845aa-Abstract.html (accessed on 13 July 2025).
  26. Fu, Z.; Liu, Z.; Ping, S.; Li, W.; Liu, J. ACGAN: A motor bearing fault diagnosis model based on an auxiliary classifier generative adversarial network and transformer network. ISA Trans. 2024, 149, 381–393. [Google Scholar] [CrossRef] [PubMed]
  27. Loparo, K.A. Case Western Reserve University Bearing Data Center. Bearings Vibration Data Sets, Case Western Reserve University (2012) [EB/OL]. Available online: https://engineering.case.edu/bearingdatacenter (accessed on 13 July 2025).
  28. Zheng, X.; Yue, C.; Wei, J.; Xue, A.; Ge, M.; Kong, Y. Few-shot intelligent fault diagnosis based on an improved meta-relation network. Appl. Intell. 2023, 53, 30080–30096. [Google Scholar] [CrossRef]
  29. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  30. Zhi, Z.; Liu, L.; Liu, D.; Hu, C. Fault detection of the harmonic reducer based on CNN-LSTM with a novel denoising algorithm. IEEE Sens. J. 2021, 22, 2572–2581. [Google Scholar] [CrossRef]
  31. Fang, H.; Deng, J.; Bai, Y.; Feng, B.; Li, S.; Shao, S.; Chen, D. CLFormer: A lightweight transformer based on convolutional embedding and linear self-attention with strong robustness for bearing fault diagnosis under limited sample conditions. IEEE Trans. Instrum. Meas. 2021, 71, 1–8. [Google Scholar] [CrossRef]
  32. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  33. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  34. Cui, Z.; Cao, H.; Ai, Z.; Wang, J. A multi-adversarial joint distribution adaptation method for bearing fault diagnosis under variable working conditions. Appl. Sci. 2023, 13, 10606. [Google Scholar] [CrossRef]
  35. Xu, Y.; Cai, W.; Xie, T. Fault diagnosis of subway traction motor bearing based on information fusion under variable working conditions. Shock Vib. 2021, 2021, 5522887. [Google Scholar] [CrossRef]
  36. Yu, Z.; Zhang, C.; Deng, C. An improved GNN using dynamic graph embedding mechanism: A novel end-to-end framework for rolling bearing fault diagnosis under variable working conditions. Mech. Syst. Signal Process. 2023, 200, 110534. [Google Scholar] [CrossRef]
  37. Tiboni, M.; Remino, C.; Bussola, R.; Amici, C. A review on vibration-based condition monitoring of rotating machinery. Appl. Sci. 2022, 12, 972. [Google Scholar] [CrossRef]
Figure 1. (a) t-SNE visualizations on the same individual with ball fault; (b) t-SNE visualizations on the same individual with OR fault; (c) SNE visualizations on the same individual with IR fault.
Figure 1. (a) t-SNE visualizations on the same individual with ball fault; (b) t-SNE visualizations on the same individual with OR fault; (c) SNE visualizations on the same individual with IR fault.
Applsci 15 07932 g001aApplsci 15 07932 g001b
Figure 2. (a) t-SNE visualizations on cross-individual task with ball fault; (b) t-SNE visualizations on cross-individual task with OR fault; (c) t-SNE visualizations on cross-individual task with IR fault.
Figure 2. (a) t-SNE visualizations on cross-individual task with ball fault; (b) t-SNE visualizations on cross-individual task with OR fault; (c) t-SNE visualizations on cross-individual task with IR fault.
Applsci 15 07932 g002aApplsci 15 07932 g002b
Figure 3. Overall framework of cross-individual diagnosis.
Figure 3. Overall framework of cross-individual diagnosis.
Applsci 15 07932 g003
Figure 4. KAConv computation process.
Figure 4. KAConv computation process.
Applsci 15 07932 g004
Figure 5. KA-Attention computation process.
Figure 5. KA-Attention computation process.
Applsci 15 07932 g005
Figure 6. Channel-wise multi-head KAA.
Figure 6. Channel-wise multi-head KAA.
Applsci 15 07932 g006
Figure 7. Overall structure of KACFormer model.
Figure 7. Overall structure of KACFormer model.
Applsci 15 07932 g007
Figure 8. Test platform for PU dataset.
Figure 8. Test platform for PU dataset.
Applsci 15 07932 g008
Figure 9. Test platform on CWRU dataset.
Figure 9. Test platform on CWRU dataset.
Applsci 15 07932 g009
Figure 10. Stability validation (accuracy of KACFormer in 15 repeated experiments on PU dataset).
Figure 10. Stability validation (accuracy of KACFormer in 15 repeated experiments on PU dataset).
Applsci 15 07932 g010
Figure 11. Ablation results.
Figure 11. Ablation results.
Applsci 15 07932 g011
Figure 12. (a) Confusion matrixes of ablation experiment on PU dataset (Group 1). (b) Confusion matrixes of ablation experiment on PU dataset (Group 2). (c) Confusion matrixes of ablation experiment on PU dataset (Group 3).
Figure 12. (a) Confusion matrixes of ablation experiment on PU dataset (Group 1). (b) Confusion matrixes of ablation experiment on PU dataset (Group 2). (c) Confusion matrixes of ablation experiment on PU dataset (Group 3).
Applsci 15 07932 g012aApplsci 15 07932 g012b
Figure 13. (a) Confusion matrixes of ablation experiment on CWRU dataset (Group 1). (b) Confusion matrixes of ablation experiment on CWRU dataset (Group 2). (c) Confusion matrixes of ablation experiment on CWRU dataset (Group 3).
Figure 13. (a) Confusion matrixes of ablation experiment on CWRU dataset (Group 1). (b) Confusion matrixes of ablation experiment on CWRU dataset (Group 2). (c) Confusion matrixes of ablation experiment on CWRU dataset (Group 3).
Applsci 15 07932 g013aApplsci 15 07932 g013b
Figure 14. Contrast results.
Figure 14. Contrast results.
Applsci 15 07932 g014
Figure 15. (a) Confusion matrixes of contrast experiment on PU dataset. (Group 1). (b) Confusion matrixes of contrast experiment on PU dataset. (Group 2). (c) Confusion matrixes of contrast experiment on PU dataset. (Group 3). (d) Confusion matrixes of contrast experiment on PU dataset. (Group 4).
Figure 15. (a) Confusion matrixes of contrast experiment on PU dataset. (Group 1). (b) Confusion matrixes of contrast experiment on PU dataset. (Group 2). (c) Confusion matrixes of contrast experiment on PU dataset. (Group 3). (d) Confusion matrixes of contrast experiment on PU dataset. (Group 4).
Applsci 15 07932 g015aApplsci 15 07932 g015b
Figure 16. (a) Confusion matrixes of contrast experiment on CWRU dataset (Group 1). (b) Confusion matrixes of contrast experiment on CWRU dataset (Group 2). (c) Confusion matrixes of contrast experiment on CWRU dataset (Group 3). (d) Confusion matrixes of contrast experiment on CWRU dataset (Group 4).
Figure 16. (a) Confusion matrixes of contrast experiment on CWRU dataset (Group 1). (b) Confusion matrixes of contrast experiment on CWRU dataset (Group 2). (c) Confusion matrixes of contrast experiment on CWRU dataset (Group 3). (d) Confusion matrixes of contrast experiment on CWRU dataset (Group 4).
Applsci 15 07932 g016aApplsci 15 07932 g016b
Figure 17. Training process on PU dataset.
Figure 17. Training process on PU dataset.
Applsci 15 07932 g017
Figure 18. Training process on CWRU dataset.
Figure 18. Training process on CWRU dataset.
Applsci 15 07932 g018
Table 1. Detailed information on the chosen bearings from the PU dataset.
Table 1. Detailed information on the chosen bearings from the PU dataset.
Data SetsBearing CodeManufacturerDamageCombinationArrangementExtent of DamageFault TypeFault
Code
Training setsK001IBU----N1
K002IBU----N1
K003IBU----N1
KA04FAGfatigue: pittingSno repetition1OR2
KA15FAGPlastic deforms:
Indentations
Sno repetition1OR2
KA16MTKfatigue: pittingRrandom2OR2
KI04MTKfatigue: pittingMno repetition1IR3
KI14MTKfatigue: pittingMno repetition1IR3
KI16FAGfatigue: pittingSno repetition3IR3
Testing setsK006IBU----N1
KA22IBU/IBBfatigue: pittingSno repetition1OR2
KI21FAGfatigue: pittingSno repetition1IR3
Table 2. Detailed information on the chosen bearings from the CWRU dataset.
Table 2. Detailed information on the chosen bearings from the CWRU dataset.
BearingInside DiameterOutside DiameterThicknessBall DiameterPitch Diameter
DE0.98432.04720.59060.31261.537
FE0.66931.57480.47240.26561.122
Table 3. Data distribution for both tasks.
Table 3. Data distribution for both tasks.
Training Validation Testing
Task1N1996N998N998
IR2003IR1001IR998
OR2004OR1002OR998
Task2N159N79N235
IR158IR79IR235
OR157OR78OR236
Table 4. Hyperparameters of the selected model.
Table 4. Hyperparameters of the selected model.
BlocksLayersParametersOUTPUT SHAPE
Input signal--1 × 2048
KAConv Block1KAConvchannel = 16, k = 64, s = 16, p = 116 × 61
ActivationSiLU16 × 61
MaxPool1Dchannel = 16, k = 2, s = 2, p = 016 × 30
KAConv Block2KAConvchannel = 32, k = 2, s = 2, p = 132 × 16
ActivationSiLU32 × 16
MaxPool1Dchannel = 32, k = 2, s = 2, p = 032 × 8
CMKAA TransformerCMKAAh = 32, embed dim = 8, dropout = 0.132 × 8
Flatten-1 × 256
KFFn = 32, activation = SiLU1 × 256
Layer norm-1 × 256
Drop outp = 0.1 1 × 256
MLP classifierFc layer1n = 100, activation = ReLU1 × 100
Fc layer2 n = 3, activation = ReLU1 × 3
Softmax -1 × 3
Table 5. Ablation results.
Table 5. Ablation results.
TaskCNN+ KAConvKACFormer-TransformerKACFormer-KAAKACFormer-KAConv.N-KACFormerKAC-Former
PU 89.14%64.66%95.25%68.09%93.75%95.73%
CWRU90.50%91.34%89.99%82.73%90.74%91.58%
Table 6. Contrast results.
Table 6. Contrast results.
TaskWDCNNCNN-LSTMConvformerLiconvformerClformerMobileNetV2ResNet18KAC-
Former
PU80.26%93.85%73.48%89.23%91.76%89.11%81.24%95.73%
CWRU84.49%68.74%79.55%71.11%72.92%84.74%88.38%91.58%
Table 7. Precision, Recall, and F1-score (macro-averaged) on both tasks.
Table 7. Precision, Recall, and F1-score (macro-averaged) on both tasks.
PrecisionRecallF1-Score
Task 10.96860.96710.9670
Task 20.93170.91670.9159
Table 8. Model accuracy with different activation functions on different tasks.
Table 8. Model accuracy with different activation functions on different tasks.
SiLUReLU
Task 1 on PU95.73%88.74%
Task 2 on CWRU91.58%89.98%
Table 9. Model accuracy under different working conditions.
Table 9. Model accuracy under different working conditions.
N09_M07_F10N15_M01_F10N15_M07_F10
Accuracy95.73%95.31%95.36%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shu, S.; Xu, M.; Liu, P.; Yang, P.; Wu, T.; Yang, J. KACFormer: A Novel Domain Generalization Model for Cross-Individual Bearing Fault Diagnosis. Appl. Sci. 2025, 15, 7932. https://doi.org/10.3390/app15147932

AMA Style

Shu S, Xu M, Liu P, Yang P, Wu T, Yang J. KACFormer: A Novel Domain Generalization Model for Cross-Individual Bearing Fault Diagnosis. Applied Sciences. 2025; 15(14):7932. https://doi.org/10.3390/app15147932

Chicago/Turabian Style

Shu, Shimin, Muchen Xu, Peifeng Liu, Peize Yang, Tianyi Wu, and Jie Yang. 2025. "KACFormer: A Novel Domain Generalization Model for Cross-Individual Bearing Fault Diagnosis" Applied Sciences 15, no. 14: 7932. https://doi.org/10.3390/app15147932

APA Style

Shu, S., Xu, M., Liu, P., Yang, P., Wu, T., & Yang, J. (2025). KACFormer: A Novel Domain Generalization Model for Cross-Individual Bearing Fault Diagnosis. Applied Sciences, 15(14), 7932. https://doi.org/10.3390/app15147932

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop