Next Article in Journal
Correction: Mølvig et al. Terahertz Cross-Correlation Spectroscopy and Imaging of Large-Area Graphene. Sensors 2023, 23, 3297
Previous Article in Journal
Fluxgate Magnetometers Based on New Physical Principles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Rolling Bearing Fault Diagnosis via Temporal-Graph Convolutional Fusion

1
School of Mechatronics Engineering, Henan University of Science and Technology, Luoyang 471003, China
2
Collaborative Innovation Center of Henan Province for High-End Bearing, Luoyang 471003, China
3
Luoyang Bearing Research Institute Technology Co., Ltd., Luoyang 471033, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(13), 3894; https://doi.org/10.3390/s25133894
Submission received: 6 May 2025 / Revised: 13 June 2025 / Accepted: 18 June 2025 / Published: 23 June 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
To address the challenge of incomplete fault feature extraction in rolling bearing fault diagnosis under small-sample conditions, this paper proposes a Temporal-Graph Convolutional Fusion Network (T-GCFN). The method enhances diagnostic robustness through collaborative extraction and dynamic fusion of features from time-domain and frequency-domain branches. First, Variational Mode Decomposition (VMD) was employed to extract time-domain Intrinsic Mode Functions (IMFs). These were then input into a Temporal Convolutional Network (TCN) to capture multi-scale temporal dependencies. Simultaneously, frequency-domain features obtained via Fast Fourier Transform (FFT) were used to construct a K-Nearest Neighbors (KNN) graph, which was processed by a Graph Convolutional Network (GCN) to identify spatial correlations. Subsequently, a channel attention fusion layer was designed. This layer utilized global max pooling and average pooling to compress spatio-temporal features. A shared Multi-Layer Perceptron (MLP) then established inter-channel dependencies to generate attention weights, enhancing critical features for more complete fault information extraction. Finally, a SoftMax classifier performed end-to-end fault recognition. Experiments demonstrated that the proposed method significantly improved fault recognition accuracy under small-sample scenarios. These results validate the strong adaptability of the T-GCFN mechanism.

1. Introduction

Rolling bearings are ubiquitous and critical components in diverse mechanical systems. However, their operational wear and eventual failure represent significant industrial challenges, posing substantial safety hazards and incurring considerable economic losses. Notably, bearing faults are implicated in over 30% of rotating machinery failures, underscoring the urgency for effective diagnostic strategies [1,2,3]. The early detection and pre-emptive replacement of deteriorating bearings are paramount to avert catastrophic system breakdowns. Consequently, the development and implementation of robust real-time monitoring and timely maintenance protocols for rolling bearings, particularly during initial operational phases, are imperative for industrial applications.
In recent decades, vibration-based fault diagnosis for rolling bearings has advanced significantly, primarily branching into two main paradigms: signal processing techniques and artificial intelligence (AI) approaches [4]. Traditional signal processing methods typically rely on the extraction and analysis of fault-sensitive features from vibration signatures. For instance, Reference [5] introduced the Harmonic Interference Resistant Fast Kurtogram (HR-Fast-Kurtogram), enhancing the accuracy of fault feature extraction and diagnosis within short, second-scale time windows. Similarly, Reference [6] demonstrated improved diagnostic precision by integrating Variational Mode Decomposition (VMD) with wavelet thresholding for enhanced signal denoising. Despite their utility in specific contexts, these conventional methods often suffer from limitations, including a heavy dependence on domain expertise for feature engineering, susceptibility to ambient noise, and poor generalization capabilities, which restrict their applicability in complex and variable industrial environments.
Concurrently, propelled by advancements in AI and computational power, deep learning (DL) has emerged as a dominant methodology in fault diagnosis. Architectures such as Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), and Graph Convolutional Networks (GCNs) have been extensively applied to bearing defect identification. Exemplifying this trend, Reference [7] presented a lightweight CNN model, achieving significant diagnostic efficiency without compromising accuracy. Reference [8] developed a parallel architecture combining CNNs and Long Short-Term Memory (LSTM) networks, validated experimentally to yield high diagnostic accuracy. Furthermore, Reference [9] proposed a sophisticated model employing a GCN-based self-attention LSTM autoencoder, demonstrating superior accuracy on benchmark bearing datasets. Nevertheless, while these DL approaches facilitate end-to-end diagnosis, their efficacy is often contingent upon large volumes of labeled training data, and they can exhibit limited robustness against noise interference and varying operational conditions, thereby constraining their practical deployment.
Addressing the prevalent challenge of inadequate fault feature extraction from limited vibration data—a common limitation of existing methods—this paper introduces a novel approach for rolling bearing fault diagnosis based on Temporal-Graph Convolutional Fusion (TGCF). Through rigorous experimental validation and comparative analysis, we demonstrate that the proposed TGCF method exhibits superior diagnostic performance and enhanced generalization capability compared to state-of-the-art techniques.

2. Theoretical Basis

Accurate and robust fault diagnosis of rolling bearings is paramount for maintaining industrial safety and operational efficiency. However, the inherent complexity of vibration signals, often compounded by challenges such as limited fault samples and the need for comprehensive feature extraction, poses significant hurdles for effective diagnosis. Traditional methods often struggle to capture the intricate patterns embedded within these signals. In response, advanced deep learning architectures have emerged as powerful tools due to their capability for automatic and hierarchical feature learning [10]. Specifically, Temporal Convolutional Networks excel at extracting multi-scale temporal dependencies from sequential data, while Graph Convolutional Networks are highly effective in modeling complex spatial relationships and non-Euclidean data structures. These sophisticated models, along with mechanisms for integrating diverse feature representations, form the core theoretical foundation for developing robust and intelligent fault diagnosis systems. This section provides a concise overview of these essential deep learning concepts, laying the groundwork for understanding our proposed T-GCFN.

2.1. Temporal Convolutional Network

The Temporal Convolutional Network (TCN) represents a specialized deep learning framework adept at processing sequential data, particularly, one-dimensional vibration signals [11]. While fundamentally derived from Convolutional Neural Networks (CNNs), TCNs incorporate crucial architectural enhancements: causal convolutions and dilated convolutions, typically implemented within residual blocks.
Causal convolutions enforce a strict temporal constraint, ensuring that the output at any given time step t is convolved only with inputs from time step t and earlier, thereby preventing any information leakage from future time steps [12]. This property is essential for modeling time-series data where future information is inherently unavailable.
Dilated convolutions are employed to exponentially increase the receptive field without a concomitant explosion in computational cost or network parameters. Unlike standard convolutions, depicted in Figure 1, dilated convolutions introduce a ‘dilation factor’, d, which defines a fixed spacing between the kernel taps applied to the input sequence. Figure 2 illustrates this dilation for d = 1; standard convolution corresponds to d = 0 in this specific representation, though often defined with d = 1 in the broader literature. Critically, d is typically increased exponentially with network depth (e.g., d = 2l for layer l). This strategy allows the network to access a very large history (receptive field) with relatively few layers, efficiently capturing long-range temporal dependencies while mitigating the parameter burden associated with achieving similar receptive fields using standard convolutional approaches [13].
The architecture of a dilated causal convolution layer, structurally illustrated in Figure 3, synergistically combines the attributes of both dilation and causality. By incorporating a dilation factor d into the convolutional kernel, it expands the receptive field, enabling the capture of wider contextual dependencies without inflating the parameter count or computational overhead [14].
Given a one-dimensional input sequence X = ( x 0 , , x T ) R T × d f e a t , where T is the sequence length (time steps) and dfeat represents the feature dimensionality per time step, and a filter f of size n (indexed i 0 , , n 1 ), the dilated causal convolution output yt at time step t and the overall TCN receptive field w are formulated as follows:
F ( x t ) = i = 0 n 1 f ( i ) x t d i
w = ( n 1 ) ( b m 1 ) / ( b 1 ) + 1
where n denotes the kernel size (filter length); the term t d · i represents the index of the historical input element accessed by the i-th filter tap for the convolution centered at time t, incorporating the dilation factor d; m signifies the total number of layers in the TCN stack; and b is the base used for the exponential increase in the dilation factor across layers (commonly b =2, such that dilation d = bl at layer l).
A TCN model is typically constructed by stacking multiple residual blocks. Each residual block comprises several layers, including dilated causal convolutions and weight normalization, among other potential components. A defining characteristic is the residual connection, which adds the input to the block (x) to the output derived from the block’s internal transformations (F(x)). The detailed architecture of a typical residual block is illustrated in Figure 4. This residual learning framework allows the network to learn modifications (the residual F(x)) relative to the identity mapping of the block’s input. The adoption of such residual connections is crucial for effectively training deeper network architectures, alleviating potential gradient vanishing issues, and ultimately enhancing the model’s representational capacity and classification performance [15].

2.2. Graph Convolutional Network

Graph Convolutional Networks (GCNs) represent a class of neural networks specifically designed for processing data structured as graphs. The core principle involves generalizing the convolution operation from regular grid structures (e.g., images) to irregular graph domains. This is achieved by iteratively aggregating information from a node’s local neighborhood (including itself and its adjacent nodes) and subsequently transforming the aggregated information, thereby learning powerful node representations [16].
GCNs operate on graph data, formally defined as G = (V, E, A). Here, V R n × d constitutes the node feature matrix, where n is the total number of nodes, and d is the dimensionality of the features associated with each node. E represents the set of edges connecting the nodes, and A is the adjacency matrix, encoding the graph’s connectivity structure.
In this study, the input graph structure is derived using the k-Nearest Neighbors (KNN) algorithm. Within this KNN graph framework, each node is connected to its k most similar neighbors based on their feature vectors. An illustrative example of such a KNN graph, configured with k = 5, is presented in Figure 5. The specific formulation for identifying the nearest neighbors for a given node xi is provided below:
min ( x i ) = K N N ( k , x i , β )
where k specifies the number of nearest neighbors to retrieve, and β denotes the sample subset (or dataset) from which neighbors are selected.
Within the constructed KNN graph, the weight eij of the edge connecting nodes xi and xj can be computed using a Gaussian kernel, reflecting their feature similarity:
e i j = exp ( x i x j 2 2 γ 2 ) ,   for   x j min ( x i )
where eij denotes the edge weight, and γ is a hyperparameter controlling the bandwidth of the Gaussian kernel.
GCN approaches are broadly categorized into spectral methods, rooted in graph signal processing, and spatial methods, which define convolutions directly on graph topology [17]. The GCN employed in this work belongs to the spectral domain category. Unlike regular grids, graph nodes exhibit variable neighborhood sizes and lack a canonical ordering, rendering standard convolution operations inapplicable. Spectral GCNs circumvent this by defining convolution in the Fourier domain associated with the graph, which is achieved using the graph Laplacian matrix [18].
For a given graph G = (V, E, A), the graph Laplacian L is defined as:
L = D A
where D is the degree matrix, a diagonal matrix whose i-th diagonal element Dii represents the degree (number of connections) of node i, and A is the adjacency matrix, where Aij = 1 if an edge exists between node i and node j, and Aij = 0 otherwise.
The degree Dii is calculated by summing the connections for node i from the adjacency matrix:
D i i = j = 1 n A i j
Figure 6 provides an illustrative example showcasing the degree matrix, adjacency matrix, and the resulting graph Laplacian for a simple undirected graph.
A commonly used variant is the symmetrically normalized graph Laplacian, Lsym, defined as:
L s y m = I D 1 / 2 A D 1 / 2
where I is the identity matrix, A is the adjacency matrix, and D is the diagonal degree matrix.
As Lsym is a real symmetric and positive semi-definite matrix, the Spectral Theorem guarantees its eigendecomposition:
L s y m = U Λ U T
where U is the matrix whose columns are the orthonormal eigenvectors of Lsym, and Λ is the diagonal matrix containing the corresponding real, non-negative eigenvalues ( 0 λ i 2 ).
The matrix U containing the eigenvectors serves as the basis for the Graph Fourier Transform (GFT). The GFT transforms a graph signal y R n (representing features or values at each node in the spatial/vertex domain) into the spectral domain y ^ via:
y ^ = U T y
Conversely, the Inverse Graph Fourier Transform (IGFT) reconstructs the spatial signal from its spectral representation:
y ^ = U y ^
Conceptually, the GFT decomposes the graph signal into components associated with different graph frequencies (the eigenvalues), analogous to the classical Fourier transform. In application contexts like fault diagnosis, this transformation aims to project sensor signals onto a basis where distinct patterns (e.g., corresponding to different fault types) might be more readily separable by analyzing their spectral coefficients y ^ .
Spectral graph convolution is fundamentally defined by applying a filter g θ directly to the graph eigenvalues in the spectral domain [19]. The convolution of a graph signal x with a spectral filter g θ is determined by:
x g θ = U g θ ( Λ ) U T x
where g θ ( Λ ) represents the filter function g θ , parameterized by learnable weights θ , applied element-wise to the diagonal eigenvalue matrix Λ. However, the explicit computation of the eigendecomposition ( U , Λ ) required for Equation (11) is computationally prohibitive for large graphs.
To overcome this limitation, approximations based on Chebyshev polynomials were developed, which were subsequently simplified by Kipf & Welling into a highly efficient first-order approximation [20]. This approach cleverly avoids the explicit eigendecomposition. It involves introducing self-loops into the adjacency matrix and then applying symmetric normalization. This leads to the widely adopted GCN layer-wise propagation rule:
Z = σ ( A ^ X W )
where X R n × d i n is the input node feature matrix (or the output from the previous layer), with n nodes and din input features per node; W R d i n × d o u t is the layer-specific trainable weight matrix, transforming features from dimension din to dout; and A ^ R d i n × d o u t is the pre-processed adjacency matrix incorporating self-loops and symmetric normalization, effectively performing localized feature aggregation. σ ( · ) denotes a non-linear activation function (e.g., ReLU) applied element-wise, and Z R d i n × d o u t is the resulting output feature matrix, representing the node embeddings after the graph convolution and activation.

3. Proposed Temporal-Graph Convolutional Fusion Network Model

This paper introduces a novel rolling bearing fault diagnosis model based on a Temporal-Graph Convolutional Fusion Network (T-GCFN), designed to automatically extract and fuse salient fault features from rolling bearing vibration data.
The overall end-to-end workflow of the proposed T-GCFN-based fault diagnosis system, from raw signal acquisition to final fault classification, is depicted in Figure 7. Following this high-level overview, the detailed architecture of the T-GCFN model itself, illustrating its internal components and data processing within the network, is presented in Figure 8.
The process begins with the input vibration signals undergoing an overlapping sampling strategy to augment the dataset size, subsequently partitioned into training, validation, and test sets according to predefined ratios. These partitioned data segments are then pre-processed through two distinct branches: Variational Mode Decomposition (VMD) for time-domain components and the Fast Fourier Transform (FFT) for frequency-domain features.
The intrinsic mode functions (IMFs) resulting from VMD serve as the input to the TCN branch of the model. Concurrently, the frequency spectrum obtained via FFT (applied after initial cleaning; see below) is utilized to construct a k-Nearest Neighbors (KNN) graph, which forms the input for the GCN branch. These processed representations are fed simultaneously into the respective TCN and GCN modules to extract complementary temporal dynamics and graph-structural (spatial) features associated with bearing faults. The features extracted from both branches are subsequently integrated within a dedicated feature fusion module, followed by adaptive pooling (e.g., global average pooling) and a final fully connected layer to yield the diagnostic classification.
Recognizing that raw industrial vibration signals are often contaminated by significant noise and irrelevant components, which can impede effective feature extraction, we implement a preliminary denoising and normalization step prior to data partitioning and feature extraction. Specifically, the input time-domain signal segment X = ( x 0 , , x t ) undergoes min–max normalization to produce a cleaned signal, Xclean:
X c l e a n = X X min X max X min
where Xmin and Xmax are the minimum and maximum values within the segment X, respectively, and Xclean is the resulting normalized time-domain signal.
The TCN branch processes the VMD-derived components. It may initially employ a 1D convolution (with a 1 × 128 kernel), followed by a stack of TCN residual blocks. As described in the original text, four such blocks (or parallel paths) are used, potentially with exponentially increasing dilation factors (d = 1, 2, 4, 8) to capture multi-scale temporal dependencies, consistent with the architecture shown in Figure 4. Features extracted from these blocks are then fused (e.g., via concatenation) into a comprehensive temporal representation.
While time-domain analysis is direct, subtle fault signatures can be obscured by noise. Conversely, the frequency spectrum often exhibits more pronounced changes indicative of specific fault conditions. Therefore, we transform the cleaned time-domain signal Xclean into the frequency domain using FFT:
X f f t = F F T ( X c l e a n )
where Xfft represents the resulting frequency spectrum signal.
The GCN branch utilizes this frequency spectrum Xfft as the basis for node features. A KNN graph is constructed from these spectral features (as detailed previously; cf. Figure 5 and Equation (4)), capturing relationships between signal segments based on spectral similarity. This graph is then fed into a series of GCN layers (potentially employing shared weights) for graph-based feature extraction. This process might involve graph pooling operations for dimensionality reduction, followed by standard components like Batch Normalization and ReLU activation functions.
To effectively integrate the distinct fault features extracted by the TCN and GCN branches, a dedicated feature fusion module is employed, structurally detailed in Figure 9. This module incorporates a channel attention mechanism, inspired by techniques like the Squeeze-and-Excitation block [21], designed to adaptively recalibrate channel-wise feature responses. The core idea is to explicitly model inter-channel dependencies, thereby enhancing informative features crucial for diagnosis while suppressing less relevant ones.
Specifically, the mechanism first utilizes Global Average Pooling (GAP) and Global Maximum Pooling (GMP) to independently aggregate spatial information across the input feature map F R H × W × C (assuming 2D features F resulting from the concatenation/fusion of TCN and GCN outputs; adjust dimensions if 1D). This ‘squeeze’ operation produces two distinct channel descriptor vectors: F a v g R 1 × 1 × C and F m a x R 1 × 1 × C , which encapsulate channel-wise statistics.
Subsequently, to capture complex cross-channel dependencies and generate attention weights (the ‘excitation’ step), both Favg and Fmax are fed into a shared Multi-Layer Perceptron (MLP). This MLP typically consists of one or more hidden layers (often incorporating ReLU activation internally), followed by an output layer. The outputs of the MLP processing of Favg and Fmax are then merged using element-wise summation.
The merged vector is then passed through a Sigmoid activation function (σ) to produce the final channel attention weights M c F R 1 × 1 × C . These weights, ranging between 0 and 1, signify the relative importance of each channel.
Finally, these attention weights M c F are used to recalibrate the original input feature map F via channel-wise multiplication (broadcasting the weights along spatial dimensions), yielding the refined output features F o u t : F o u t = M C F F . This process selectively modulates the features based on their channel-wise significance. The generation of the channel attention weights M c F can be summarized by the following equation:
M C F = σ ( M L P ( G A P ( F ) ) + M L P ( G M P ( F ) ) )
where F represents the input feature map to the attention module; C denotes the number of channels in F; GAP and GMP signify the Global Average Pooling and Global Maximum Pooling operations, respectively; MLP indicates the shared Multi-Layer Perceptron; and σ represents the Sigmoid activation function.

4. Experimental Results and Analysis

4.1. Experimental Dataset

The experimental validation employs the widely recognized public dataset from Case Western Reserve University (CWRU), a standard benchmark extensively utilized for research in rolling bearing fault diagnosis [22]. The dataset was acquired from an experimental test rig comprising a 1.47 kW (2 HP) induction motor, a dynamometer (serving as the load device), and the test bearings. Vibration data were captured using accelerometers mounted at various locations on the motor housing (specifically, data from the drive-end bearing sensor were used in this study, consistent with common practice).
The experiments utilized SKF-6205-2RS deep-groove ball bearings. A comprehensive dataset encompassing ten distinct bearing health states was constructed: one Normal (N) condition and nine fault conditions, comprising three fault types—Inner Race (IR), Outer Race (OR), and Ball (B)—each introduced with three different defect diameters: 0.007 inches (0.1778 mm), 0.014 inches (0.3556 mm), and 0.021 inches (0.5334 mm).
Data were collected under four distinct motor load conditions: 0 HP, 1 HP, 2 HP, and 3 HP, with vibration signals sampled at a frequency of 12 kHz. Each data sample consists of 1024 contiguous data points. To augment the dataset, samples were extracted using a sliding window approach with a 75% overlap. For each of the 10 bearing states under a specific load condition, 400 individual samples were generated. The resulting dataset was then divided using stratified random sampling into training, validation, and testing subsets with a ratio of 70%, 15%, and 15%, respectively. Details of the dataset composition are summarized in Table 1.

4.2. Model Training Configuration

The experiments were implemented using the Python3.10 programming language and the PyTorch2.0 deep learning framework. All model training and evaluation were conducted on a workstation equipped with an AMD Ryzen 5 4500U CPU operating at 2.38 GHz and 16 GB of RAM.
Key architectural parameters and hyperparameters for the proposed T-GCFN model are detailed in Table 2.
TCN Input: The number of intrinsic mode functions (IMFs) decomposed by VMD, serving as input to the TCN branch, was set to 4. This value was determined as optimal through preliminary analysis employing the Pearson correlation coefficient method to select the most informative modes.
GCN Input: The input features for the GCN branch were derived from the FFT spectrum of the 1024-point signal segments. Owing to the inherent conjugate symmetry of the FFT for real-valued signals, only the first half of the spectrum (representing non-redundant frequencies), corresponding to 512 data points, was utilized for constructing the graph node features. The graph structure itself was generated using the k-Nearest Neighbors (KNN) algorithm, with the number of neighbors k set to 10. This means each node (representing a signal segment’s spectrum) in the graph is connected to its 10 most similar neighbors based on spectral features.
Key hyperparameters influencing model performance were optimized through empirical validation. A batch size of 64 and an initial learning rate of 0.001 were selected based on these experiments. The Adam optimizer was employed for training, chosen for its efficiency and stability, attributed to its integration of momentum and adaptive learning rate capabilities, which facilitate rapid convergence. Cross-entropy loss served as the objective function, selected for its suitability in multi-class classification tasks and its capacity to provide informative gradients conducive to effective model training. The model was trained for a total of 50 epochs.

4.3. Experimental Results and Analysis

The diagnostic performance of the proposed T-GCFN model was evaluated primarily based on classification accuracy and loss function values. Experiments were conducted using the CWRU dataset subset (0 HP load condition) detailed in Table 1. Representative curves illustrating the training and validation loss, as well as accuracy, over the training epochs for a typical run are presented in Figure 10 and Figure 11, respectively.
As depicted in Figure 10 and Figure 11, the proposed model demonstrates remarkably stable training behavior on the CWRU dataset. Convergence is rapid, with both training and validation accuracy stabilizing at 99.91% and the corresponding loss values converging to 0.0004 within approximately the first 10 epochs. This indicates efficient learning and good generalization from the training set to the validation set.
To provide a detailed visualization of the model’s classification performance on unseen data, the confusion matrix for the test set predictions is presented in Figure 12. In this matrix, each row corresponds to the predicted class labels, while each column represents the true class labels. The diagonal elements quantify the number of correctly classified instances for each bearing state. The confusion matrix clearly indicates that the proposed model achieves exceptionally high accuracy across all 10 rolling bearing health states in the test set, with nearly all instances correctly assigned to their respective classes, demonstrating robust diagnostic capability.

4.4. Experimental Results and Analysis in Small-Sample Scenarios

In practical industrial settings, acquiring sufficient fault data is often challenging, as machinery typically operates under normal conditions for extended periods. This data scarcity necessitates models that perform robustly, even when trained on limited samples.
To rigorously evaluate the proposed T-GCFN model’s performance under such constraints, we simulated small-sample scenarios by creating subsets of the original training data containing 10%, 30%, 50%, 70%, 90%, and 100% of the full training samples (while keeping the validation and test sets complete). In this study, ‘small-sample conditions’ specifically refer to scenarios where the amount of available training data per fault type is significantly limited. For the CWRU dataset used in our experiments, 10% of the original training samples translates to 28 samples per fault class. This quantity is considered a challenging small sample size in rolling bearing fault diagnosis, as it severely restricts the model’s ability to learn robust and generalizable features from abundant data, thereby increasing the risk of overfitting and poor generalization in real-world applications.
Regarding the selection of this specific minimum percentage, while even smaller proportions, such as 5% (corresponding to 14 samples per fault type), could theoretically represent more extreme data scarcity: our preliminary investigations indicated that 10% already presents substantial challenges for deep learning models to achieve reliable convergence and meaningful diagnostic performance. This choice allows for a rigorous assessment of our T-GCFN’s ability to extract discriminative features from limited data, which is a primary focus of this paper, while ensuring that the dataset remains sufficiently representative for model training and comparative analysis.

4.4.1. Ablation Experiment

An ablation study was conducted to validate the contribution and necessity of the key components within the proposed T-GCFN architecture. The following six model variants were compared:
Model 1: The proposed T-GCFN model (full architecture).
Model 2: T-GCFN variant using simple feature concatenation instead of the channel attention fusion module.
Model 3: Baseline TCN model using only the raw 1D vibration signal as input.
Model 4: Multi-scale TCN model processing VMD components (akin to the TCN branch of T-GCFN).
Model 5: Baseline GCN model using a KNN graph constructed from time-domain features.
Model 6: GCN model using a KNN graph constructed from FFT frequency-domain features (akin to the GCN branch of T-GCFN).
To ensure stability and mitigate random variations, each experiment was repeated five times, and the average accuracy across these runs is reported as the final result. The average test accuracies of these models on the CWRU 0 HP dataset, under varying training data proportions, are presented in Figure 13. The x-axis represents the percentage of training data used, and the y-axis represents the average test accuracy.
As illustrated in Figure 13, the proposed T-GCFN (Model 1) consistently outperforms all other ablation variants across all tested data proportions, underscoring the effectiveness and synergistic contribution of its integrated components.
Analysis of Pre-processing and Branch Architectures (Model 3 vs. 4; Model 5 vs. 6): Comparing Model 4 (VMD + Multi-scale TCN) with Model 3 (basic TCN), a significant accuracy improvement (e.g., 5.61% at the 10% data level) is observed. This highlights the benefit of VMD pre-processing and the multi-scale TCN architecture for extracting informative temporal features, especially under data scarcity. Similarly, comparing Model 6 (GCN on FFT features) with Model 5 (GCN on time-domain features) reveals a substantial accuracy gain (approx. 10% at the 10% data level) when using frequency-domain inputs. This confirms the suitability of spectral features for graph-based fault representation via GCN. Overall, Models 4 and 6 consistently outperform their respective baselines (Models 3 and 5), further validating the effectiveness of the chosen VMD, FFT, and multi-scale TCN strategies.
Analysis of Dual-Branch Structure (Model 2 vs. 4 and 6): Comparing the dual-branch model with concatenation (Model 2) against the single-branch models (Model 4 and Model 6), Model 2 generally achieves higher accuracy across different data levels. Notably, as the training data size decreases, the performance degradation of Model 2 is less pronounced compared to Models 4 and 6, maintaining accuracy above 80% even at the 10% level. This suggests that the parallel feature extraction from both temporal (VMD + TCN) and spectral (FFT + GCN) domains enhances robustness in low-data regimes.
Analysis of Fusion Mechanism (Model 1 vs. 2): This comparative analysis highlights the key technical innovation of the T-GCFN model: its adaptive channel attention fusion mechanism. Comparing the full T-GCFN model with channel attention (Model 1) against the variant using simple concatenation (Model 2), Model 1 consistently achieves superior accuracy, particularly evident with the full dataset (99.91%). Crucially, under severe data scarcity (e.g., 10% data), Model 1 maintains accuracy above 90%, whereas Model 2 experiences a more significant drop. This notable performance gap suggests that simple concatenation might lead to information loss or suboptimal feature integration. In contrast, the proposed channel attention mechanism effectively and intelligently fuses the complementary features from the two branches, dynamically weighting their contributions to enhance salient diagnostic information. This adaptive integration capability is fundamental to significantly boosting diagnostic performance, especially in challenging small-sample scenarios.
To further investigate the complexity and computational cost of the models, Table 3 summarizes the average accuracy, parameter quantity, FLOPs, and training time for all models at the 10% dataset level (corresponding to 400 samples).
As shown in Table 3, Model 1 (the proposed T-GCFN) achieves the highest average accuracy of 93.21%. Compared to Model 2 (T-GCFN with simple concatenation), Model 1 improves accuracy by 9.03% (93.21% vs. 84.18%), with a modest increase in parameters (0.99 × 105), FLOPs (0.02 G) and training time (18 s). This demonstrates the effectiveness of the channel attention fusion mechanism in significantly boosting performance despite a slight increase in computational cost. Furthermore, Model 1 significantly outperforms single-branch models (Model 3–6) in accuracy, highlighting the necessity of the dual-branch architecture for robust fault diagnosis.
Despite its enhanced performance, Model 1 remains relatively lightweight with 4.11 × 105 parameters and 0.12 GFLOPs. While its training time (120 s) is longer than simpler models, the substantial gain in diagnostic accuracy, especially under small-sample conditions, justifies this. This balance makes T-GCFN a practical solution for online monitoring, where precision is paramount.

4.4.2. Cross-Load Condition Experiment

To further assess the model’s generalization capability and robustness under varying operating conditions (cross-load)—a critical requirement for real-world deployment—we conducted cross-load diagnostic experiments in the small-sample setting. Specifically, models were trained using data subsets (10% to 100%, as described above) from the 1 HP load condition and subsequently tested on the full dataset from the 2 HP load condition. The performance of the proposed T-GCFN (Model 1) was compared against five other representative deep learning models commonly used in fault diagnosis:
Model 1: The proposed T-GCFN model.
Model 2: Lightweight CNN [7].
Model 3: Transformer [23].
Model 4: CNN-GRU [24].
Model 5: CNN-LSTM [8].
Model 6: GCN-LSTM [9].
Again, average results over five independent runs are reported. The average test accuracies under these cross-load, small-sample conditions are presented in Figure 14, where the x-axis denotes the model type, and the y-axis represents the average accuracy achieved with varying proportions of 1 HP training data.
As shown in Figure 14, while the diagnostic accuracy of all models generally improves as the proportion of training data increases, the proposed T-GCFN model (Model 1) consistently achieves the highest accuracy across all training data proportions compared to the other five baseline models under this cross-load scenario. At the most challenging 10% data level, the Lightweight CNN (Model 2) and Transformer (Model 3) exhibit the lowest accuracies (85.24% and 86.31%, respectively), suggesting difficulty in extracting generalizable features under combined data scarcity and domain shift (load variation). Models 4, 5, and 6 (CNN-GRU, CNN-LSTM, GCN-LSTM) maintain respectable accuracies around 90% at the 10% level, but their performance varies more across different data proportions. In contrast, the proposed T-GCFN model demonstrates superior robustness; its accuracy degradation from 100% data to 10% data is minimal (only a 6.79% drop), consistently staying well above 90%. This highlights its strong generalization capability and resilience to variations in operating conditions, even when trained on limited data. In summary, these results demonstrate that the proposed T-GCFN model exhibits significantly better diagnostic performance and robustness compared to existing methods under challenging cross-load, small-sample scenarios.

4.4.3. High-Noise Condition Experiments

To rigorously evaluate the robustness and superiority of the proposed T-GCFN (Model 1) under noisy conditions, we conducted extensive comparative experiments against the comparative models introduced in Section 4.4.2. These experiments were performed across various Signal-to-Noise Ratio (SNR) levels, mimicking realistic industrial environments with varying noise interference. The diagnostic accuracies of these models across varying SNR levels for the 10% sample size, representing a challenging small-sample scenario, are presented in Figure 15.
As evident from Figure 15, the proposed T-GCFN (Model 1) consistently demonstrates superior diagnostic performance compared to all comparative models across the tested noise levels. For instance, even under the most challenging condition of SNR of 2 dB, T-GCFN (Model 1) achieves an accuracy of 86.48%, significantly outperforming the best comparative model at this SNR, GCN-LSTM (Model 6), which only reached 80.32%. This represents an improvement of 6.16 percentage points in accuracy. As the SNR increases, all models show an improvement in accuracy, but the performance gap between T-GCFN and other models remains substantial, particularly at lower SNR values.
This remarkable robustness of T-GCFN can be attributed to its dual-branch architecture and the adaptive channel attention fusion mechanism, which enable the model to extract more discriminative and noise-resilient features from both temporal and graph-structural domains. These comprehensive comparative results unequivocally prove the superior robustness and generalization capability of the T-GCFN model for intelligent fault diagnosis in challenging noisy and data-scarce industrial environments.

4.5. Generalization Experiment on the Jiangnan University Bearing Dataset

To further evaluate the generalization capability of the proposed T-GCFN model across different types of rolling bearings and operating conditions, experiments were conducted using a publicly available dataset from Jiangnan University (JNU) [25]. This dataset features two distinct bearing types, N205 (cylindrical roller bearing) and NU205 (cylindrical roller bearing), under four health states: Normal (N), Inner Race fault (IR), Rolling Element fault (RE), and Outer Race fault (OR). Faults were artificially introduced using wire cutting. Vibration signals were acquired at a sampling frequency of 50 kHz under three different operating speeds: 600 r/min, 800 r/min, and 1000 r/min.
For each health state under each operating condition, 800 samples were generated, resulting in a total of 3200 samples per condition. The data segment length, pre-processing methods (VMD, FFT, normalization), and T-GCFN model parameters were kept identical to those used for the CWRU dataset experiments. This procedure resulted in three distinct datasets, denoted Dataset A (600 r/min), Dataset B (800 r/min), and Dataset C (1000 r/min), corresponding to the three operating speeds. The composition of these datasets is detailed in Table 4.
Experiments were performed independently for each of the three operating conditions (Datasets A, B, and C). The diagnostic accuracies achieved by the T-GCFN model are presented in Table 5.
As shown in Table 5, the T-GCFN model exhibits a slight decrease in accuracy on the JNU dataset compared to the CWRU dataset. This minor reduction might potentially be attributed to factors such as signal distortion caused by interference during data acquisition or transmission specific to the JNU setup. Nevertheless, the model achieves a high overall average accuracy of 99.59% across the three different operating conditions. This consistently strong performance further substantiates the excellent generalization ability and robustness of the proposed T-GCFN model for rolling bearing fault diagnosis across varying bearing types and operational speeds.

5. Conclusions

Addressing the challenges of incomplete feature information, sensitivity to operating conditions, and noise interference inherent in rolling bearing fault diagnosis, particularly under small-sample scenarios, this paper proposed a novel diagnostic method based on a T-GCFN. Comprehensive experimental validation confirmed the superiority of the proposed approach, leading to the following key conclusions:
(1)
The T-GCFN model introduces a novel dual-branch architecture that synergistically extracts complementary temporal and graph-structural fault features from raw signals. Its core innovation lies in the adaptive fusion of multi-domain features via a channel attention mechanism, which dynamically weights and integrates information from both TCN and GCN branches. This unique design effectively addresses the challenge of incomplete feature extraction by enhancing salient diagnostic information.
(2)
Extensive experiments were conducted on the CWRU benchmark dataset under simulated small-sample conditions, encompassing constant load, cross-load variations, and high-noise environments. Ablation studies systematically validated the rational design and contribution of each component within the T-GCFN model. Under challenging cross-load conditions, the proposed T-GCFN consistently outperformed several state-of-the-art deep learning models, including Lightweight CNN, Transformer, CNN-GRU, CNN-LSTM, and GCN-LSTM, demonstrating superior generalization. Furthermore, the model exhibited strong robustness against high levels of noise. Its generalization capability was further substantiated through successful application to the distinct Jiangnan University (JNU) bearing dataset.
(3)
Furthermore, the T-GCFN model exhibits promising potential for real-time system monitoring. Its efficient inference capabilities, coupled with its end-to-end learning paradigm, make it suitable for rapid and automated fault detection in dynamic industrial settings. The computational efficiency of its core components (TCN, GCN) and pre-processing steps (VMD, FFT) further supports its applicability for online diagnostic tasks.
In summary, the proposed T-GCFN fault diagnosis method demonstrates enhanced and comprehensive feature extraction capabilities, achieves higher diagnostic accuracy, and shows strong potential for real-time deployment compared to existing approaches, especially under challenging conditions like data scarcity, varying loads, and noise interference. However, the current study primarily focused on single fault types. The applicability and performance of the T-GCFN model for diagnosing compound faults (e.g., simultaneous inner race cracking and roller spalling) warrant further investigation in future work.

Author Contributions

Conceptualization, F.L. and Y.L.; methodology, Y.L. and D.W.; software, Y.L.; validation, F.L.; formal analysis, F.L.; investigation, Y.L.; resources, D.W.; data curation, F.L. and Y.L.; writing—original draft preparation, F.L.; writing—review and editing, Y.L. and D.W.; visualization, F.L.; supervision, D.W.; project administration, D.W.; funding acquisition, Y.L. and D.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Henan Provincial Major Science and Technology Special Project (Grant No. 221100220100).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in the following repositories: The CWRU dataset is openly available at the Case Western Reserve University Bearing Data Center website at https://engineering.case.edu/bearingdatacenter (accessed on 2 May 2025). The Jiangnan University dataset is available at the CSDN blog at https://blog.csdn.net/qq_41731978/article/details/121733115 (accessed on 2 May 2025).

Conflicts of Interest

Author Dongfeng Wang was employed by the company Luoyang Bearing Research Institute Technology Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Yan, G.X.; Chen, J.; Bai, Y.; Yu, C. A survey on fault diagnosis approaches for rolling bearings of railway vehicles. Processes 2022, 10, 724. [Google Scholar] [CrossRef]
  2. Li, J.N.; Luo, W.G.; Bai, M.S. Review of research on signal decomposition and fault diagnosis of rolling bearing based on vibration signal. Meas. Sci. Technol. 2024, 35, 092001. [Google Scholar] [CrossRef]
  3. Raj, K.K.; Kumar, S.; Kumar, R.R. Systematic review of bearing component failure: Strategies for diagnosis and prognosis in rotating machinery. Arab. J. Sci. Eng. 2024, 50, 5353–5375. [Google Scholar] [CrossRef]
  4. Li, X.; Ma, Z.Q.; Yuan, Z.H.; Mu, T.M.; Du, G.X.; Liang, Y.; Liu, J.W. A review on convolutional neural network in rolling bearing fault diagnosis. Meas. Sci. Technol. 2024, 35, 072002. [Google Scholar] [CrossRef]
  5. Chen, Q.; Qian, X. Fault diagnosis of wind turbine main bearings based on harmonic elimination and spectral kurtosis. Acta Energiae Solaris Sin. 2025, 46, 1–8. [Google Scholar]
  6. Hamdaoui, H.; Ngiejungbwen, L.A.; Gu, J.N.; Tang, S.X. Improved signal processing for bearing fault diagnosis in noisy environments using signal denoising, time–frequency transform, and deep learning. J. Braz. Soc. Mech. Sci. Eng. 2023, 45, 576. [Google Scholar] [CrossRef]
  7. Jiang, L.I.; Shi, C.Z.; Sheng, H.S.; Li, X.J.; Yang, T.G. Lightweight CNN architecture design for rolling bearing fault diagnosis. Meas. Sci. Technol. 2024, 35, 126142. [Google Scholar] [CrossRef]
  8. Fu, G.H.; Wei, Q.J.; Yang, Y.S. Bearing fault diagnosis with parallel CNN and LSTM. Math. Biosci. Eng. MBE 2024, 21, 2385–2406. [Google Scholar] [CrossRef]
  9. Lee, D.; Choo, H.; Jeong, J. Gcn-based lstm autoencoder with self-attention for bearing fault diagnosis. Sensors 2024, 24, 4855. [Google Scholar] [CrossRef]
  10. Huang, G.Z.; Lei, W.P.; Dong, X.M.; Zou, D.L.; Chen, S.J.; Dong, X. Stage-based remaining useful life prediction for bearings using GNN and correlation-driven feature extraction. Machines 2025, 13, 43. [Google Scholar] [CrossRef]
  11. Ding, L.; Li, Q. Fault diagnosis of rotating machinery using novel self-attention mechanism TCN with soft thresholding method. Meas. Sci. Technol. 2024, 35, 047001. [Google Scholar] [CrossRef]
  12. Shang, Z.W.; Liu, H.; Zhang, B.; Feng, Z.H.; Li, W.X. Multi-view feature fusion fault diagnosis method based on an improved temporal convolutional network. Insight-Non-Destr. Test. Cond. Monit. 2023, 65, 559–569. [Google Scholar] [CrossRef]
  13. Wu, Y.Q.; Dai, J.Y.; Yang, X.Q.; Shao, F.M.; Gong, J.C.; Zhang, P.; Liu, S.D. The fault diagnosis of rolling bearings based on FFT-SE-TCN-SVM. Actuators 2025, 14, 152. [Google Scholar] [CrossRef]
  14. Yu, M.Y.; Li, Y.P.; Meng, G.L.; Wang, Y.B. An intelligent fault diagnosis for rotating machine under strong noise based on cross-attention-driven spatial-temporal feature fusion and duplexing time sequence convolution optimization. Eng. Appl. Artif. Intell. 2025, 153, 110897. [Google Scholar] [CrossRef]
  15. Shi, H.T.; Shang, Y.J.; Zhang, X.C.; Tang, Y.H. Research on the Initial Fault Prediction Method of Rolling Bearings Based on DCAE-TCN Transfer Learning. Shock Vib. 2021, 2021, 5587756. [Google Scholar] [CrossRef]
  16. Li, T.F.; Zhou, Z.; Li, S.N.; Sun, C.; Yan, R.Q.; Chen, X.F. The emerging graph neural networks for intelligent fault diagnostics and prognostics: A guideline and a benchmark study. Mech. Syst. Signal Process. 2022, 168, 108653. [Google Scholar] [CrossRef]
  17. Chen, Z.X.; Ji, J.C.; Ni, Q.; Ye, B.Y.; Ding, X.X.; Yu, W.N. Bi-structural spatial–temporal network for few-shot fault diagnosis of rotating machinery. Mech. Syst. Signal Process. 2025, 227, 112378. [Google Scholar] [CrossRef]
  18. Yuan, Z.H.; Ma, Z.Q.; Li, X.; Li, J.J. A multichannel MN-GCN for wheelset-bearing system fault diagnosis. IEEE Sens. J. 2022, 23, 2481–2494. [Google Scholar] [CrossRef]
  19. Gong, X.Y.; Feng, K.P.; Zhi, Z.H.; Gao, Y.Y.; Du, W.L. Multiple fault diagnosis for rolling bearings method employing CEEMD-GCN based on horizontal visibility graph. Meas. Sci. Technol. 2022, 34, 035022. [Google Scholar]
  20. Ma, J.C.; Fu, Y.; Cheng, T.L.; He, D.Q.; Cao, H.R.; Yu, B. SCADA data-driven spatio-temporal graph convolutional neural network for wind turbine fault diagnosis. IEEE Trans. Instrum. Meas. 2025, 74, 3521910. [Google Scholar] [CrossRef]
  21. Gao, T.Y.; Yang, J.L.; Zhang, B.Q.; Li, Y.L.; Zhang, H.Y. A fault diagnosis method based on feature-level fusion of multi-sensor information for rotating machinery. Meas. Sci. Technol. 2023, 35, 036109. [Google Scholar] [CrossRef]
  22. Case Western Reserve University Bearing Data Center. Available online: https://engineering.case.edu/bearingdatacenter/download-data-file (accessed on 1 May 2025.).
  23. Deng, F.Y.; Zheng, S.X.; Hao, R.J. A lightweight scale-aware modulated swin transformer model for axle box bearing fault diagnosis. J. Xi’an Jiaotong Univ. 2024, 58, 83–93. [Google Scholar]
  24. Zhang, L.; Zhen, C.Z.; Yi, J.Y. Dual-channel feature fusion CNN-GRU for gearbox fault diagnosis. J. Vib. Shock 2021, 40, 239–245+294. [Google Scholar]
  25. Tan, Q.Y.; Ma, P.; Zhang, H.L. Rolling bearing fault diagnosis based on graph convolutional networks. Noise Vib. Control 2023, 43, 101–108+116. [Google Scholar]
Figure 1. Standard convolution.
Figure 1. Standard convolution.
Sensors 25 03894 g001
Figure 2. Dilated convolution.
Figure 2. Dilated convolution.
Sensors 25 03894 g002
Figure 3. Structure of dilated causal convolution.
Figure 3. Structure of dilated causal convolution.
Sensors 25 03894 g003
Figure 4. Structure of the temporal convolutional residual block.
Figure 4. Structure of the temporal convolutional residual block.
Sensors 25 03894 g004
Figure 5. Structure of the KNN graph.
Figure 5. Structure of the KNN graph.
Sensors 25 03894 g005
Figure 6. Undirected graph.
Figure 6. Undirected graph.
Sensors 25 03894 g006
Figure 7. Overall end-to-end workflow of the proposed T-GCFN fault diagnosis method.
Figure 7. Overall end-to-end workflow of the proposed T-GCFN fault diagnosis method.
Sensors 25 03894 g007
Figure 8. T-GCFN fault diagnosis model architecture.
Figure 8. T-GCFN fault diagnosis model architecture.
Sensors 25 03894 g008
Figure 9. Channel attention feature fusion layer.
Figure 9. Channel attention feature fusion layer.
Sensors 25 03894 g009
Figure 10. Loss curve.
Figure 10. Loss curve.
Sensors 25 03894 g010
Figure 11. Accuracy curve.
Figure 11. Accuracy curve.
Sensors 25 03894 g011
Figure 12. Confusion matrix.
Figure 12. Confusion matrix.
Sensors 25 03894 g012
Figure 13. Ablation study results.
Figure 13. Ablation study results.
Sensors 25 03894 g013
Figure 14. Variable condition experimental results.
Figure 14. Variable condition experimental results.
Sensors 25 03894 g014
Figure 15. Diagnostic accuracy of different models under various noise levels (10% samples).
Figure 15. Diagnostic accuracy of different models under various noise levels (10% samples).
Sensors 25 03894 g015
Table 1. Composition of the CWRU rolling bearing experimental dataset subset (per load condition).
Table 1. Composition of the CWRU rolling bearing experimental dataset subset (per load condition).
No.ConditionDefect Size (mm)Training SamplesValidation SamplesTest SamplesLabel
1Normal028060600
2Inner Race Fault0.177828060601
3Inner Race Fault0.355628060602
4Inner Race Fault0.533428060603
5Outer Race Fault0.177828060604
6Outer Race Fault0.355628060605
7Outer Race Fault0.533428060606
8Ball Fault0.177828060607
9Ball Fault0.355628060608
10Ball Fault0.533428060609
Table 2. Architectural parameters of the proposed T-GCFN model.
Table 2. Architectural parameters of the proposed T-GCFN model.
Layer TypeKernel SizeInput ChannelsOutput Shape
TCN Branch InputN/A44 × 1024
GCN Branch InputN/A1010 × 512
Initial Conv Layer (TCN)1 × 12888 × 128
TCN Residual Block(s)1 × 28/16/3232 × 512
GCN Layers512 × 5121010 × 512
Feature Fusion ModuleN/A6010 × 512
Global Average Pooling N/A601 × 512
Output LayerN/A101 × 10
Table 3. The complexity and computational cost of each model.
Table 3. The complexity and computational cost of each model.
ModelAverage Accuracy (%)Parameter Quantity (105)FLOPs (G)Training Time (s)
Model 193.214.110.12120
Model 284.183.120.10102
Model 370.11.210.0653
Model 475.611.860.0764
Model 560.252.140.0778
Model 670.252.310.0884
Table 4. Composition of the Jiangnan University (JNU) rolling bearing experimental dataset (per operating condition).
Table 4. Composition of the Jiangnan University (JNU) rolling bearing experimental dataset (per operating condition).
DatasetConditionTraining SamplesValidation SamplesTest SamplesLabel
A/B/CNormal5601201200
Inner Race Fault (IR)5601201201
Rolling Element Fault (RE)5601201202
Outer Race Fault (OR)5601201203
Table 5. Diagnostic results on the Jiangnan University (JNU) dataset.
Table 5. Diagnostic results on the Jiangnan University (JNU) dataset.
ModelAccuracy (%)
ABCAverage
T-GCFN99.5899.5699.6299.59
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, F.; Li, Y.; Wang, D. Rolling Bearing Fault Diagnosis via Temporal-Graph Convolutional Fusion. Sensors 2025, 25, 3894. https://doi.org/10.3390/s25133894

AMA Style

Li F, Li Y, Wang D. Rolling Bearing Fault Diagnosis via Temporal-Graph Convolutional Fusion. Sensors. 2025; 25(13):3894. https://doi.org/10.3390/s25133894

Chicago/Turabian Style

Li, Fan, Yunfeng Li, and Dongfeng Wang. 2025. "Rolling Bearing Fault Diagnosis via Temporal-Graph Convolutional Fusion" Sensors 25, no. 13: 3894. https://doi.org/10.3390/s25133894

APA Style

Li, F., Li, Y., & Wang, D. (2025). Rolling Bearing Fault Diagnosis via Temporal-Graph Convolutional Fusion. Sensors, 25(13), 3894. https://doi.org/10.3390/s25133894

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop