Next Article in Journal
Phosphite as a Sustainable and Versatile Alternative for Biostimulation, Biocontrol, and Weed Management in Modern Agriculture
Next Article in Special Issue
A Review of Synchronous Fixed-Frequency Microgrid Droop Control Systems Based on Global Positioning System
Previous Article in Journal
Determination of Pressure Jump Dependence and Time Constants of Hydraulic Pumps with Constant Pressure and Variable Flow
Previous Article in Special Issue
Q-Learning-Incorporated Robust Relevance Vector Machine for Remaining Useful Life Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Temporal Fusion Channel Network with Multi-Channel Hybrid Attention for the Remaining Useful Life Prediction of Rolling Bearings

1
Institute of Intelligent Manufacturing, Nanjing Tech University, Nanjing 210009, China
2
China Academy of Space Technology, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Processes 2024, 12(12), 2762; https://doi.org/10.3390/pr12122762
Submission received: 6 November 2024 / Revised: 27 November 2024 / Accepted: 3 December 2024 / Published: 5 December 2024

Abstract

:
The remaining useful life (RUL) prediction of rolling bearings is crucial for optimizing maintenance schedules, reducing downtime, and extending machinery lifespan. However, existing multi-channel feature fusion methods do not fully capture the correlations between channels and time points in multi-dimensional sensor data. To address the above problems, this paper proposes a multi-channel feature fusion algorithm based on a hybrid attention mechanism and temporal convolutional networks (TCNs), called MCHA-TFCN. The model employs a dual-channel hybrid attention mechanism, integrating self-attention and channel attention to extract spatiotemporal features from multi-channel inputs. It uses causal dilated convolutions in TCNs to capture long-term dependencies and incorporates enhanced residual structures for global feature fusion, effectively extracting high-level spatiotemporal degradation information. The experimental results on the PHM2012 dataset show that MCHA-TFCN achieves excellent performance, with an average Root-Mean-Square Error (RMSE) of 0.091, significantly outperforming existing methods like the DANN and CNN-LSTM.

1. Introduction

Rolling bearings constitute fundamental mechanical components that are extensively utilized across diverse industrial applications [1]. The performance and service life of these bearings directly impact the reliability of entire systems. Traditional bearing life prediction methods, primarily based on statistical approaches and empirical formulas, have not adequately considered the influence of material characteristics on bearing performance degradation. Psuj et al. [2] revealed that the complex interplay of mechanical properties, microstructural characteristics, and surface integrity of bearing materials significantly influences their operational behavior. Through multi-sensor fusion approaches, they demonstrated precise characterization of material defects. Similarly, Alzhanov et al. [3] established that dynamic material properties directly correlate with component longevity under operational conditions. In industrial applications, bearings are subjected to complex environmental conditions and sustained high-intensity operational loads, which inevitably lead to material degradation and performance deterioration. Such degradation mechanisms can precipitate catastrophic system failures, resulting in substantial equipment damage, elevated maintenance expenditure, and significant safety hazards. Empirical studies have demonstrated that bearing-related failures constitute approximately 45% to 55% of rotating machinery malfunctions [4], emphasizing the paramount importance of systematic bearing health monitoring and management. To address these critical engineering challenges, extensive research initiatives have been directed toward the development of methodologies for predicting the remaining useful life (RUL) of bearings. Contemporary RUL prediction methodologies can be systematically categorized into model-based approaches [5] and data-driven paradigms [6,7]. Data-driven methods are currently the main research focus due to their ease of implementation and ability to avoid the high costs and complexity of model-based methods.
Data-driven methods rely on monitoring data, significantly reducing dependence on equipment degradation mechanisms and expert experience. By analyzing and processing the multi-dimensional feature information of rotating parts, potential health status or degradation features can be extracted from the data. Data-driven methods can be roughly divided into three categories: statistical methods [8,9], machine learning methods [10], and deep learning methods [11]. Deep learning methods have become mainstream in data-driven research because they reduce reliance on expert experience, empirical formulas, and physical modeling knowledge. They can directly extract degradation patterns from raw data. With the improvement in industrial data collection technology, a large amount of operational process data has been collected. Deep learning is widely used in RUL prediction due to its powerful data processing capabilities. Standard deep learning models include convolutional neural networks (CNNs) and recurrent neural networks (RNNs). Several scholars use convolutional network variants to construct mappings from vibration data to the remaining life of bearings. For example, Yang et al. [12] proposed an RUL prediction method based on a dual CNN model architecture combined with a mapping algorithm. The early fault points and RUL prediction results identified by the two CNN models were mapped using intermediate confidence variables to achieve accurate RUL prediction. Wu et al. [13] proposed a data-driven deep learning fusion model based on an attention mechanism and a parallel-branch CNN-LSTM model. The model uses a CNN and an LSTM with an attention mechanism to extract spatial and temporal features in parallel, thereby improving prediction accuracy. Du et al. [14] proposed a CNN prediction model based on the global attention mechanism (GAM), solving the problems of insufficient feature extraction and loss of critical features and achieving accurate RUL prediction for bearings.
To explore time series models, many scholars research CNN models to retain CNNs’ good spatial perception ability while incorporating time series features. Bai et al. [15] proposed a dilated causal convolution (DCC) method. It uses dilated DCC combined with causal and dilated convolution and stacks residual blocks to expand the receptive field, addressing CNNs’ limitations in capturing long-term temporal dependencies compared to RNNs. Song et al. [16] combined a temporal convolutional network (TCN) with the attention mechanism and proposed a deep learning RUL prediction method based on weighted sequence data representation. Different industrial sensors and time steps are weighted through a distributed attention mechanism. The time series is then feature-extracted using a time domain convolution module with shared weights. Zeng et al. [17] proposed a rotating machinery RUL prediction method based on a dynamic graph and spatiotemporal network (STNet), using a graph convolution network and Bi-LSTM to capture spatial and temporal correlations, improving prediction performance in single-sensor monitoring scenarios.
Additionally, with the recent development of sensor and data storage technology, a large amount of production process data has been generated, and existing methods still need to perceive adequate information in the data. Fully utilizing these data can significantly improve the reliability of equipment monitoring systems. Liang et al. [18] proposed a bearing RUL prediction method based on multi-sensor data fusion and a bidirectional temporal attention convolution network (Bi-TACN). They merged multi-sensor data into multi-channel data and combined it with channel weighting to improve the model’s feature extraction capability. Lei et al. [19] proposed an effective information extraction method for multi-dimensional data using a multi-head attention network. By adaptively evaluating and selecting features, feature information can be mined to reduce information loss. Ren [20] proposed a new industrial health indicator prediction network (MCTAN) based on multi-channel temporal attention. It uses channel attention and local attention mechanisms to effectively extract long-term temporal dependencies, improving the prediction accuracy of the multi-input model and reducing delayed predictions. In summary, the multi-input model can effectively improve the model’s data extraction capabilities.
However, despite the success of multi-branch, multi-input deep learning methods in RUL prediction for rolling bearings, some problems remain:
(1) Traditional single-channel networks can only extract features independently from each channel, which leads to a failure in fully capturing the potential correlations between different channels. The degradation process of rolling bearings is a complex and multi-dimensional phenomenon, with signals collected by different sensors often carrying complementary health condition information. Single-channel networks fail to fully utilize this information, resulting in feature extraction that is partial and not comprehensive enough.
(2) Data from different sensor channels may contain distinct fault features. If the correlations between channels are not fully explored, key information may be lost during feature fusion, leading to redundancy or missing important features. Moreover, current multi-channel networks often use static fusion methods that cannot adaptively adjust channel weights based on dynamic input, making it difficult to identify the most relevant features for RUL prediction.
This paper proposes a temporal feature fusion network based on multi-channel hybrid attention and a TCN to address these issues. It is designed to effectively fuse information from various channels and capture detailed patterns and correlations in complex datasets. This network model integrates channel attention with an enhanced TCN framework to mine spatial and temporal dependencies, achieving comprehensive feature extraction. The fusion process employs attention mechanisms and weighted aggregation, enabling the network to emphasize salient features dynamically. The contributions of this paper are as follows:
(1) A novel dual-channel fusion block is designed: The MCHA-TFCN block differs from the TCN block. The MCHA-TFCN block changes the single-channel input of the original TCN block to dual-channel and uses a Squeeze-and-Excitation Network (SENet) to extract channel weights, fusing the channel input into a single-channel output. Therefore, the MCHA-TFCN block can achieve the adaptive weighted fusion of two feature sets and reduce the number of channels. End-to-end fusion of multi-channel features is achieved by stacking MCHA-TFCN blocks.
(2) Features fused by channel attention are combined with the self-attention mechanism to improve the model’s perception of multi-dimensional degradation information, forming a spatiotemporal-level degradation perception.
(3) An MCHA-TFCN model for the RUL prediction of rolling bearings is proposed. This model consists of multiple MCHA-TFCN blocks, which integrate statistical features and deep feature mapping into the remaining service life of bearings, achieving better prediction results than advanced methods in experiments. The model’s success guides further research on multi-channel feature fusion and temporal feature learning in RUL prediction.
The rest of this paper is organized as follows: Section 2 briefly introduces the background knowledge, Section 3 details the proposed method, Section 4 verifies its performance on the bearing dataset, and Section 5 provides our conclusions.

2. Methodology

2.1. Temporal Convolutional Network

The TCN is a variant of the traditional CNN. It uses DCC and residual block stacking to expand the receptive field, addressing the limitation of the CNN in capturing long-term temporal dependencies from time series data compared to the RNN. It has recently been proven to be more accurate, simple, and precise than standard recurrent networks such as long short-term memory networks (LSTM) and CNNs on sequence data.

2.1.1. Dilated Convolution

Dilated convolution, also known as hole convolution, is a particular application of the convolution operation proposed by Oord et al. [21] to address the issue of the small receptive field in causal convolution. It is designed to increase the receptive field of the convolution layer, thus increasing the range of input data that each output unit of the network model can perceive. Dilated convolution expands the receptive field by inserting holes in the standard convolution kernel, thereby obtaining a more comprehensive range of input information without increasing the parameters or the computational load. As illustrated in Figure 1, a four-layer cubic convolution network is shown, where the white circles represent padding, the blue circles represent the input data, the yellow circles represent the hidden layer, and the green circles represent the output layer. The solid and dashed arrows represent the neurons directly involved in the convolution calculation. By comparing Figure 1 and Figure 2, it is evident that the receptive field corresponding to a single neuron in the output layer differs significantly. This difference increases exponentially with the number of convolution layers. The calculation formula for dilated convolution is as follows:
y ( t ) = k = 0 K 1 x ( t + d · k ) · w ( k )
where y ( t ) is the t-th element of the output sequence, x ( t ) is the t-th element of the input sequence, w ( k ) is the k-th element of the convolution kernel, d is the dilation factor, and K is the size of the convolution kernel.

2.1.2. Causal Convolution

Causal convolution was proposed by Long et al. [22] and later utilized in TCNs by Bai et al. [15]. The core concept of causal convolution is that the information at the current and preceding moments serves as the cause, while the predicted result serves as the effect. The expected outcome is solely influenced by the information at the current and preceding moments. Unlike the traditional RNN, which uses future and current data for prediction, causal convolution aligns more closely with the logic of RUL prediction. Additionally, compared to the serial structure of the RNN, causal convolution allows for parallel data processing, enhancing the efficiency of network computations. A structural diagram of dilated causal convolution combined with dilated convolution is shown in Figure 3. As illustrated in Figure 3, combining dilated causal convolution with dilated convolution increases the receptive field, enabling the model to understand better and predict dynamic changes in time series.
The calculation formula for dilated causal convolution is as follows:
y ( t ) = K 1 k = 0 x ( t k ) · w ( k )

2.1.3. Residual Connection

The residual structure was first proposed by He et al. [23] to address the issues of gradient vanishing and network degradation due to increased network layers. The core idea of residual connection is to introduce a direct connection method that skips one or more layers, allowing the network input to be directly passed to the next layer. The calculation formula for residual connection is as follows:
y = F x + x
where F x represents the processing of the input time series signal x = { x 0 , x 1 , x 2 , , x T } by convolution, activation function, etc.
A 1 × 1 convolution block designs the residual connection to form an identity mapping. When the network gradient decreases sharply with increased network layers, the derivative can be regarded as “short-circuited”. The identity mapping replaces the original, maintaining network performance or preventing degradation. Thus, the depth of the network model can be increased effectively.
The TCN comprises several connected and stacked residual blocks, each primarily consisting of a dilated causal convolution layer, a weight normalization layer, a ReLU activation function, and a dropout layer [24]. Dropout is a commonly used regularization technique aimed at preventing neural network overfitting. The structure of the TCN residual block is illustrated in Figure 4.
The spatial topology of the TCN residual block with a dilation factor of 1 and a convolution kernel of 3 is illustrated in Figure 5.
For the multi-channel input characteristics in this study, the TCN can perform efficient parallel calculations and more stable gradient propagation than the RNN, avoiding gradient vanishing or explosion issues. Additionally, the predicted output of the TCN is not influenced by future data, aligning better with actual bearing operation data collection.

2.2. Attention Mechanism

The origins of attention mechanisms can be traced back to their successful applications in natural language processing. Inspired by the human visual system, Bahdanau et al. [25] designed an attention mechanism combined with a sequence-to-sequence (Seq2Seq) model, which enables the model to adaptively focus on the relevant part of the input sequence when generating each target word, thereby improving the translation quality. The attention mechanism has gradually diversified in recent years, and many variants and extensions have emerged. For example, the channel attention mechanism focuses on enhancing the importance of different channels in the feature map; self-attention [26] allows the model to consider the information of all positions in the sequence simultaneously when processing the sequence; and multi-head attention [27] captures the information of different subspaces in the sequence by using multiple self-attention as “heads” in parallel.

2.2.1. Channel Attention

The Squeeze-and-Excitation Network, also known as channel attention, was proposed by Hu et al. [28]. As shown in Figure 6, the channel attention mechanism mainly introduces nonlinearity by pulling up, compressing, and superimposing activation functions on the input multiple channel data and learns the weight of each channel to give high weights to channels with high influence and low weights to channels with low importance. By adjusting the contribution of multi-channel features through weighted channels, greater attention can be paid to channels that are more important to specific tasks.
The channel attention mechanism is mainly implemented through squeeze, excitation, and scale.
(1) The squeeze calculation formula is as follows:
z c = F s q X = 1 W W i = 1 x c i
where F s q ( · ) is the squeeze function that compresses the feature matrix X = [ x 1 , x 2 , , x c ] R W × C with channel count C and size 1 × W to a weight vector z C of size 1 × 1 × C .
(2) The excitation calculation process is as follows:
s c = F e x z c , W = σ g z c , W
where F e x ( ) is the excitation function that takes the C-dimensional vector z C , obtained from global pooling in the squeeze step, and performs a fully connected operation to compress the dimensions to a C / r dimensional vector. This vector is then activated by ReLU to increase nonlinearity, followed by another fully connected operation to restore it to a C-dimensional vector. Finally, sigmoid activation is applied (scaling the values between 0 and 1), resulting in the weight vector s c .
(3) The scale calculation formula is as follows:
x ˜ c = F scale x c , s c = s c · x c
where x c is a feature of channel C in the feature space X. The corresponding channel excitation value s c is applied to obtain the weighted feature x ˜ c , resulting in the final weighted feature matrix
X ˜ = [ x ˜ 1 , x ˜ 2 , , x ˜ c ]

2.2.2. Self-Attention Mechanism

Self-attention calculates the attention of each element relative to all other elements, allowing the model to capture the contextual relationship within the time series. Because the TCN has a vast receptive field compared to ordinary CNNs, combining self-attention can help it better learn the time series information of input features. At the same time, like the TCN, self-attention can parallelize computing and process feature information in parallel with the TCN to improve network operation efficiency. For the time series feature X = x 1 , x 2 , x i , , x n , x i R d , the query Q matrix, fundamental K matrix, value V matrix, and attention calculation formula can be expressed as
Q = X f W Q
K = X f W K
V = X f W V
Attention ( Q , K , V ) = softmax Q K T d k V
where W Q , W K , and  W V are the weight matrices of the Q, K, and V matrices; d k represents the dimensions of Q, K, and V; softmax ( · ) is the function of the normalized attention weights; and  Attention Q , K , V represents the weighted output.

3. RUL Prediction Method Based on MCHA-TFCN

3.1. Key Ideas

During the operation of mechanical equipment, sensors at different positions will generate multiple monitoring signals, which can be regarded as other characteristics of bearings. The existing life prediction methods are mostly single-channel information input, which cannot fully detect potential degradation information for feature information, and the serial processing efficiency of data is low. Therefore, it is necessary to build a multi-channel parallel computing prediction network.
This paper proposes a bearing life prediction method based on multi-channel feature fusion. By analyzing and processing different features and learning channel weight coefficients and feature weight coefficients through mixed attention, the valuable information of multi-dimensional features is retained to the maximum extent. To enhance the readability, Figure 7 outlines the main work of this paper. The process is mainly divided into three steps: (1) data processing, (2) prediction model construction, and (3) model training prediction. The details of the process are as follows.

3.2. Data Processing

The original vibration signal is usually high-dimensional and contains redundant information, making establishing an accurate degradation model difficult. Data processing can effectively reduce the data dimension and eliminate noise, which can help better capture the degradation information of the bearing. This paper processes the data to improve the prediction model’s reliability and stability. It uses the filtered time domain features combined with deep multi-scale features as the model input. The data processing process mainly includes data standardization and feature engineering. The details of the process are as follows:
(1) Data standardization:
x n o r m i t = x i t x m e a n i t x m a x i t x m i n i t
where x i t and x n o r m i t are the i-th values of the original and standardized feature sequences at time t, and x m e a n i t , x m a x i t , and x m i n i t are the average, maximum, and minimum values of the i-th feature sequence values at all times, respectively. Because the life length of the bearing is inconsistent, to ensure the generalization performance of the model, the label is also standardized, and the formula is as follows:
L a b e l t = 1 R U L t R U L T
where R U L t and R U L T represent the running time of the current sampling point time t and the complete failure time T, respectively.
(2) Feature engineering:
To construct a statistical feature pool, this paper uses a sliding window data processing method to extract time domain features such as the mean, variance, and root mean square from vibration data. It also uses feature screening indicators to screen out suitable time domain features. The application of feature screening enables the identification of degradation-relevant features, effectively reducing feature space dimensionality while optimizing predictive accuracy. This experiment uses correlation and monotonicity to form a comprehensive feature evaluation index to evaluate the features. The construction of feature screening indicators needs to fit the equipment degradation process, and suitable features should meet specific conditions:
(1) Correlation: the high correlation between the feature sequence and the performance change trend of the equipment.
(2) Monotonicity: degradation is an irreversible process.
Therefore, this paper uses comprehensive feature evaluation indicators to evaluate the features. The formula for comprehensive composite evaluation indicators J X is as follows:
J ( X ) = ω 1 Cror 1 : k + ω 2 Mon 1 : k
s . t ω i > 0 i ω i = 1 , i = 1 , 2 , 3
where ω 1 and ω 2 are the weight coefficients of the correlation coefficient index C o r r ( X ) and the monotonicity index M o n ( X ) , which are set to 0.6 and 0.4, respectively.
C o r r ( X ) = K i x i t i i x i i t i K i x i 2 i x i 2 K i t i 2 i t i 2
M o n ( X ) = 1 K 1 i ε ( x i x i 1 ) i ε ( x i 1 x i )
where K represents the number of degradation features, X = x 1 , x 2 , , x k represents the obtained degradation feature sequence, x i represents the trend sequence of the obtained condition monitoring indicators, ε ( · ) represents the step function, x ¯ i represents the smoothed degradation feature, and  T = t 1 , t 2 , t 3 , t k represents the corresponding time sequence.
To improve the quality of input features, enhance the performance and robustness of the model, and optimize the feature fusion effect, the filtered time domain features are combined with deep multi-scale convolution features as the input of the model. The original data are convolved with multiple convolution kernels of different sizes, and the ReLU function is introduced to provide nonlinearity and generate multi-scale deep features with different receptive fields.
The multi-scale convolution calculation formula is as follows:
y ( i ) = m = 0 k 1 x ( i + m ) × w j ( m )
where, for a given one-dimensional input signal x and a one-dimensional convolution kernel of scale k, the output y of the convolution operation is the value at position i.

3.3. Construction of Prediction Model MCHA-TFCN

3.3.1. Design of Feature Fusion Module

To solve the problem of the incomplete and unbalanced feature extraction of different channel input degradation, this study proposes a novel dual-channel hybrid attention (DCHA) module by combining self-attention and channel attention mechanisms. This module aims to utilize these features better and enable the model to divide input data’s importance adaptively.
The design of the DCHA module is shown in Figure 8. In the dual-channel fusion module at the same level, one-channel attention is shared. When the dual-channel features are input, they are adaptively weighted with the input features at the same level, and the weighted features of the two are added together and then input into the self-attention module for feature fusion. This design aims to enable the high-level features obtained after fusion to consider global information better, thereby improving the model’s ability to understand and integrate information from different channels at the entire level.
A i = σ ( W 2 δ ( W 1 g i ) )
S i = Attention ( F i W i Q , F i W i K , F i W i V )
where A i represents the channel attention weight, δ represents the ReLU activation function, σ represents the Sigmoid activation function, g i is the feature vector after pooling, W i Q , W i K , and W i V are the weight matrices of the Q, K, and V matrices, respectively, and  S i represents the corresponding self-attention feature matrix.
In the specific implementation process, self-attention captures the correlation of different positions in the input features. At the same time, channel attention helps to explore the information interaction between different channels. The DCHA module can more accurately capture the correlation between different channels by sharing channel attention weights, thereby improving the model’s feature fusion ability. This module reduces the number of model parameters while improving computational efficiency, providing an effective solution for improving the performance of the model’s adaptive partitioning of input data importance.

3.3.2. Feature Fusion Network Construction

To achieve the fusion of multiple channel features, this paper adds a DCHA module to the convolution block of the TCN and designs a dual-channel fusion module MCHA-TFCN structure based on hybrid attention. A structure diagram is shown in Figure 4, and the adaptive weighted fusion of 2 N features is achieved by stacking N layers of modules. Each residual block contains a multi-input dilated causal convolution layer, a dropout layer, and a batch normalization layer and uses a residual method to connect the input and output of the residual blocks of different layers; the output of the two residual blocks at each level is the input of the residual block of the next layer. The feature fusion calculation process of the MCHA-TFCN module for the input channel feature X i is shown in Formulas (20) and (21).
F fusion = A i S i + A i + 1 S i + 1
o = A c t i v a t i o n X i + X i + 1 + F fusion
where A i and A i + 1 represent the channel attention weight and the corresponding feature matrix after self-attention processing, F f u s i o n represents the fused features, ⊙ means the element-by-element multiplication, X i and X i + 1 are the features to be fused, and o is the output.
Multi-channel feature fusion is achieved by continuously stacking the MCHA-TFCN dual-channel fusion module to form the MCHA-TFCN framework. The overall structure of the model is shown in Figure 9. Figure 9 describes the specific structure of the developed MCHA-TFCN, which consists of three modules, namely the feature extraction module DCHA that combines multi-head attention and channel attention, the feature fusion module of the MCHA-TFCN module for stacking, and the dense layer that maps the features to the remaining lifespan. The dense layer (Dense) consists of a fully connected layer, an activation function (PReLU), and a dropout layer, as shown in Table 1. The dropout layer plays a role in regularization to reduce overfitting, and the PReLU function improves the gradient disappearance problem.
According to the above network structure, the RUL prediction algorithm based on MCHA-TFCN can be simplified into a pseudo-code, as shown in Algorithm 1.
Algorithm 1 RUL prediction method based on MCHA-TFCN
Input:Training set data for a single working condition: X train , and corresponding labels Y.
Output:MCHA-TFCN model trained under a single working condition.
  
Training:
1Initialize the network parameters of the feature acquisition module and MCHA-TFCN.
2for epochs = 1 , 2 , 3 , , max do:
3    Input training data X train in batch order into the model to be trained.
4    Extract time domain and multi-scale depth features, and output a multi-dimensional feature sequence x ˜ i to be fused.
5    During the network forward propagation process, Formulas (18) and (19) are used to weight the input features, and Formulas (20) and (21) are used to perform feature fusion.
6    Calculate the loss of the predicted result and the actual label (MSELoss).
7    Use the Adam method to back-propagate and update the model parameters.
8end for
9Save the trained network structure.
  
Test:
10Input the test data X test into the model to obtain the prediction results Y ^ .
11Evaluate the prediction effect of the model based on the evaluation index (RMSE).

4. Verification

4.1. Dataset Description

The performance evaluation of our proposed methodology was conducted on the IEEE PHM2012 challenge dataset, with particular emphasis on feature fusion effectiveness and prediction accuracy for rolling bearing RUL estimation. Figure 10 depicts the experimental configuration.
The experimental equipment mainly consists of three functional modules: the rotation module, radial force loading module, and measurement module. The platform accelerates the degradation of the simulated bearing by applying different radial forces and using different speeds on the tested bearing. It is categorized into three working conditions based on the magnitude of the force and the speed. The division of the dataset adopts the division scheme set by the competition. The specific information is shown in Table 2.
The measurement module mainly consists of two vibration sensors to collect the horizontal and vertical vibration signals of the tested bearing, and the axial and radial vibration signals are sampled on the bearing installation, with a sampling frequency of 25.6 kHz. The data collector collects and records 0.1 s samples every 10 s, that is, 2560 data points. The actual value of the bearing life is the time of the entire operation process of the bearing from the initial state to failure. The experimental end condition set by the accelerated experimental device is that the amplitude of the vibration signal exceeds 20 g.

4.2. Implementation Details

The input and output data structure of the proposed method and the corresponding module are shown in Table 3. The RMSE is used in this paper as the model’s loss function, and the Adam optimizer is used to update the model parameters. The grid search method is used to obtain the hyperparameters of MCHA-TFCN. The setting of the hyperparameters is shown in Table 4. The dynamic learning rate adjustment method changes the learning rate size when the model reaches the milestone number and the model parameters have been fine-tuned. The time domain feature input of the model is the top four time domain features selected according to the screening index, namely kurtosis, peak-to-peak value, root mean square, and mean.

4.3. Effect Verification

There are two primary purposes of the experimental part. The first is to verify the effectiveness of the proposed MCHA-TFCN method for multi-channel feature fusion. The second is to verify the accuracy of the model for RUL prediction.
(1) Effectiveness Verification
Given the effectiveness of feature fusion, the model is subjected to ablation experiments to verify the effectiveness of the proposed feature fusion module. In the ablation experiment, all test networks use the same network parameter configuration as MCHA-TFCN. Specifically, this experiment systematically compares the following three models: the MCHA-TFCN model, the TCN combined with the self-attention mechanism (i.e., TCN + SelfAT), and the essential TCN input.
To avoid randomness in the experimental results, the experiment repeated the prediction results of MCHA-TFCN five times and took the average as the final result. At the same time, for the TCN and TCN + SelfAT with single-channel input, this experiment input the filtered statistical features and multi-scale deep features into the network and, in turn, took the average of the output results. This operation was intended to eliminate the interference of prediction results caused by the difference in input data between the MCHA-TFCN and TCN. The experimental results are shown in Table 5 and Table 6.
In Table 5 and Table 6, TCN(1) represents the traditional TCN whose input data are the deep features of a single channel, and TCN(2) represents the average result obtained when the input data are the time domain feature kurtosis, peak-to-peak value, root mean square, and mean value input into the traditional TCN in turn.
According to the data in the table, by comparing the results of taking the average of the deep features of the single-channel input of the TCN and the multiple time domain feature inputs with the results of the proposed method, it can be seen that MCHA-TFCN can effectively fuse multi-dimensional features, has a good feature enhancement effect, and can obtain more information related to degradation.
(2) Prediction accuracy comparison experiment
To prove the superiority of the proposed MCHA-TFCN, the MCHA-TFCN is compared with four advanced feature fusion models (i.e., CNN-LSTM, Bi-TACN [26], DANN [29], and TCN-SA [30]) in the experiment. These algorithm models are compared on the bearing test sets of working conditions 1 and 2. Comparison tables of the evaluation index RMSE results are shown in Table 7 and Table 8.
From the RMSE comparison results, the RMSE of the proposed method is worse than that of the other methods. However, some individual data performance could be better than that of different methods; overall, it is better than the other methods. From the comparison of the average RMSE under working condition 1 and working condition 2, it can be seen that MCHA-TFCN can predict the degradation trend of all tested bearings, indicating that MCHA-TFCN is effective in bearing RUL prediction. Therefore, the proposed MCHA-TACN bearing RUL prediction method can fuse the bearing degradation information with time and space perception from the multi-channel feature data under a single working condition and maximize the valuable information of multi-dimensional features while improving the prediction effect of the model under a single working condition.

4.4. RUL Estimation Results and Analysis

After processing the bearing data of working condition 1 and working condition 2 in the test set, we input them into the prediction model of the corresponding working condition that was trained, and obtained the prediction result diagram shown below.
As shown in Figure 11a–e, warping occurs in the last rapid degradation stage of bearings 1–3, 1–5, and 1–6, but is slightly lacking. Still, the overall prediction result of working condition 1 has a good fit with the real-life label. It can be seen from Figure 12a–e that the bearing prediction curves all fluctuate slightly near the actual value of RUL, and the curve fitting degree is high, which shows good performance.
The proposed method can accurately reflect the degradation trend for model training under a single working condition. This shows that the high-level features fused by the MCHA-TFCN prediction model can effectively reflect the degree of bearing degradation under a single working condition.
The above experimental results show that the proposed MCHA-TFCN remaining service life prediction method not only has an excellent weighted fusion effect for multi-channel features, but also has a good ability to extract spatiotemporal features and has high accuracy in RUL prediction for a single working condition.

5. Conclusions

The RUL prediction of rolling bearings is an essential component of health management systems for rotating parts and is of great significance for reducing or even avoiding equipment failures. The proposed architecture addresses several critical challenges in RUL prediction through methodological advances. A dual-channel fusion module with Squeeze-and-Excitation Networks (SENet) enables the adaptive weighted fusion of sensor channel features, while a hybrid attention mechanism combining self-attention and channel attention enhances the model’s ability to interpret multi-dimensional degradation features. The improved temporal convolutional network (TCN) framework with attention mechanisms achieves more effective modeling of long-term temporal dependencies in bearing degradation processes. The prediction results demonstrate that the MCHA-TFCN model achieves lower RMSE and evaluation scores superior to those of several existing prediction methods.
While these preliminary results are promising, several methodological limitations require careful consideration. The multi-channel hybrid attention mechanism introduces significant computational demands, requiring substantial resources and extended training periods. The framework demonstrates sensitivity to data quality and sensor integrity, where feature extraction may be compromised by signal noise. The current validation remains limited to specific operational conditions, suggesting the need for broader evaluation across different bearing types and operating scenarios. The complex architecture also presents challenges for real-time applications, requiring a careful balance between prediction accuracy and computational efficiency in industrial settings.
Future research could address these limitations through several key approaches. The development of computationally efficient attention mechanisms and robust feature extraction methods may help maintain predictive performance while reducing computational demands. The framework’s generalization capability might be improved through transfer learning strategies and validation across diverse operational conditions. Engineering implementation could benefit from distributed computing frameworks and adaptive parameter optimization. Data quality enhancement could be pursued through advanced signal processing, while fusion optimization might be achieved through novel feature fusion strategies and multi-modal information integration.

Author Contributions

Conceptualization, C.W.; Software, H.Q.; Validation, J.J.; Resources, X.H.; Writing—Original Draft, H.Q.; Writing—Review and Editing, D.Z.; Supervision, D.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported partly by the National Natural Science Foundation of China under Grants 62203213 and 62333010, and partly by the Natural Science Foundation of Jiangsu Province under Grant BK20220332.

Data Availability Statement

The source code and datasets necessary for generating the PHM2010 dataset, along with detailed instructions to reproduce this study, are available at https://www.kaggle.com/datasets/rabahba/phm-data-challenge-2010 (accessed on 20 January 2024).

Conflicts of Interest

The authors declare no competing interests.

References

  1. Nieto, P.J.G.; García-Gonzalo, E.; Lasheras, F.S.; de Cos Juez, F.J. Hybrid PSO–SVM-based method for forecasting of the remaining useful life for aircraft engines and evaluation of its reliability. Reliab. Eng. Syst. Saf. 2015, 138, 219–231. [Google Scholar] [CrossRef]
  2. Psuj, G. Multi-sensor data integration using deep learning for characterization of defects in steel elements. Sensors 2018, 18, 292. [Google Scholar] [CrossRef] [PubMed]
  3. Alzhanov, N.; Tariq, H.; Amrin, A.; Zhang, D.; Spitas, C. Modelling and simulation of a novel nitinol-aluminium composite beam to achieve high damping capacity. Mater. Today Commun. 2023, 35, 105679. [Google Scholar] [CrossRef]
  4. Wang, Y.; Zhang, S.; Cao, R.; Xu, D.; Fan, Y. A rolling bearing fault diagnosis method based on the WOA-VMD and the GAT. Entropy 2023, 25, 889. [Google Scholar] [CrossRef]
  5. Pecht, M.; Gu, J. Physics-of-failure-based prognostics for electronic products. Trans. Inst. Meas. Control 2009, 31, 309–322. [Google Scholar] [CrossRef]
  6. Yu, W.; Tu, W.; Kim, I.Y.; Mechefske, C. A nonlinear-drift-driven Wiener process model for remaining useful life estimation considering three sources of variability. IEEE Trans. Instrum. Meas. 2021, 212, 107631. [Google Scholar] [CrossRef]
  7. Soualhi, A.; Medjaher, K.; Zerhouni, N. Bearing health monitoring based on Hilbert–Huang transform, support vector machine, and regression. Trans. Inst. Meas. Control 2014, 64, 52–62. [Google Scholar] [CrossRef]
  8. Cui, L.; Wang, X.; Wang, H.; Ma, J. Research on remaining useful life prediction of rolling element bearings based on time-varying Kalman filter. IEEE Trans. Instrum. Meas. 2019, 69, 2858–2867. [Google Scholar] [CrossRef]
  9. Aggab, T.; Vrignat, P.; Avila, M.; Kratz, F. Remaining useful life estimation based on the joint use of an observer and a hidden Markov model. J. Risk Reliab. 2022, 236, 676–695. [Google Scholar] [CrossRef]
  10. Gao, D.; Huang, M. Prediction of remaining useful life of lithium-ion battery based on multi-kernel support vector machine with particle swarm optimization. J. Power Electron. 2017, 17, 1288–1297. [Google Scholar]
  11. Wang, C.; Lu, N.; Cheng, Y.; Jiang, B. A data-driven aero-engine degradation prognostic strategy. IEEE Trans. Cybern. 2019, 51, 1531–1541. [Google Scholar] [CrossRef] [PubMed]
  12. Yang, B.; Liu, R.; Zio, E. Remaining useful life prediction based on a double-convolutional neural network architecture. IEEE Trans. Ind. Electron. 2019, 66, 9521–9530. [Google Scholar] [CrossRef]
  13. Wu, M.; Ye, Q.; Mu, J.; Fu, Z.; Han, Y. Remaining Useful Life Prediction via a data-driven deep learning fusion model–CALAP. IEEE Access 2023, 11, 112085–112096. [Google Scholar] [CrossRef]
  14. Du, X.; Jia, W.; Yu, P.; Shi, Y.; Gong, B. RUL prediction based on GAM–CNN for rotating machinery. J. Braz. Soc. Mech. Sci. Eng. 2023, 45, 142. [Google Scholar] [CrossRef]
  15. Bai, S.; Kolter, J.Z.; Koltun, V. An empirical evaluation of generic convolutional and recurrent networks for sequence modeling. arXiv 2018, arXiv:1803.01271. [Google Scholar]
  16. Song, Y.; Gao, S.; Li, Y.; Jia, L.; Li, Q.; Pang, F. Distributed attention-based temporal convolutional network for remaining useful life prediction. IEEE Internet Things J. 2020, 8, 9594–9602. [Google Scholar] [CrossRef]
  17. Zeng, X.; Yang, C.; Liu, J.; Zhou, K.; Li, D.; Wei, S.; Liu, Y. Remaining useful life prediction for rotating machinery based on dynamic graph and spatial–temporal network. Meas. Sci. Technol. 2022, 34, 035102. [Google Scholar] [CrossRef]
  18. Liang, H.; Cao, J.; Zhao, X. Multi-sensor data fusion and bidirectional-temporal attention convolutional network for remaining useful life prediction of rolling bearing. Meas. Sci. Technol. 2023, 34, 105126. [Google Scholar] [CrossRef]
  19. Nie, L.; Xu, S.; Zhang, L. Multi-Head Attention Network with Adaptive Feature Selection for RUL Predictions of Gradually Degrading Equipment. Actuators 2023, 12, 158. [Google Scholar] [CrossRef]
  20. Ren, L.; Liu, Y.; Huang, D.; Huang, K.; Yang, C. MCTAN: A novel multichannel temporal attention-based network for industrial health indicator prediction. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 6456–6467. [Google Scholar] [CrossRef]
  21. Van Den Oord, A.; Dieleman, S.; Zen, H.; Simonyan, K.; Vinyals, O.; Graves, A.; Kalchbrenner, N.; Senior, A.; Kavukcuoglu, K. Wavenet: A generative model for raw audio. arXiv 2016, arXiv:1609.03499. [Google Scholar]
  22. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7 September 2015; pp. 3431–3440. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 5 July 2016; pp. 770–778. [Google Scholar]
  24. Cao, Y.; Ding, Y.; Jia, M.; Tian, R. A novel temporal convolutional network with residual self-attention mechanism for remaining useful life prediction of rolling bearings. Reliab. Eng. Syst. Saf. 2021, 215, 107813. [Google Scholar] [CrossRef]
  25. Bahdanau, D. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  26. Vaswani, A. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–7 December 2017. [Google Scholar]
  27. Wang, X.; Li, Y.; Xu, Y.; Liu, X.; Zheng, T.; Zheng, B. Remaining useful life prediction for aero-engines using a time-enhanced multi-head self-attention model. Aerospace 2023, 10, 80. [Google Scholar] [CrossRef]
  28. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 1 July 2018; pp. 7132–7141. [Google Scholar]
  29. Li, X.; Zhang, W.; Ma, H.; Luo, Z.; Li, X. Data alignments in machinery remaining useful life prediction using deep adversarial neural networks. Knowl.-Based Syst. 2020, 197, 105843. [Google Scholar] [CrossRef]
  30. Wang, Y.; Deng, L.; Zheng, L.; Gao, R.X. Temporal convolutional network with soft thresholding and attention mechanism for machinery prognostics. J. Manuf. Syst. 2021, 60, 512–526. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of traditional convolution calculation.
Figure 1. Schematic diagram of traditional convolution calculation.
Processes 12 02762 g001
Figure 2. Schematic diagram of dilated convolution calculation.
Figure 2. Schematic diagram of dilated convolution calculation.
Processes 12 02762 g002
Figure 3. Schematic diagram of dilated causal convolution calculation.
Figure 3. Schematic diagram of dilated causal convolution calculation.
Processes 12 02762 g003
Figure 4. Structural diagram of TCN residual block.
Figure 4. Structural diagram of TCN residual block.
Processes 12 02762 g004
Figure 5. Schematic diagram of dilated convolution with residual connection.
Figure 5. Schematic diagram of dilated convolution with residual connection.
Processes 12 02762 g005
Figure 6. Schematic diagram of channel attention mechanism.
Figure 6. Schematic diagram of channel attention mechanism.
Processes 12 02762 g006
Figure 7. The RUL prediction process of the proposed method.
Figure 7. The RUL prediction process of the proposed method.
Processes 12 02762 g007
Figure 8. Diagram of DCHA structure.
Figure 8. Diagram of DCHA structure.
Processes 12 02762 g008
Figure 9. MCHA-TFCN structure diagram.
Figure 9. MCHA-TFCN structure diagram.
Processes 12 02762 g009
Figure 10. Rolling bearing test bench.
Figure 10. Rolling bearing test bench.
Processes 12 02762 g010
Figure 11. Display of bearing prediction effect under working condition 1.
Figure 11. Display of bearing prediction effect under working condition 1.
Processes 12 02762 g011
Figure 12. Display of bearing prediction effect under working condition 2.
Figure 12. Display of bearing prediction effect under working condition 2.
Processes 12 02762 g012
Table 1. Parameters of dense mapping layer.
Table 1. Parameters of dense mapping layer.
No.LayerSetting Value
1Linear2560, 256
2BatchNorm1d256
3Tanh/
4Dropout0.2
5Linear256, 1
6PReLU/
Table 2. PHM 2012 dataset partitioning.
Table 2. PHM 2012 dataset partitioning.
TaskConditionsTrainTest
Task-ILoad (N): 4000;
Speed (r/min): 1800
B1_1-B1_2B1_3-B1_7
Task-IILoad (N): 4200;
Speed (r/min): 1650
B2_1-B2_2B2_3-B2_7
Task-IIILoad (N): 5000;
Speed (r/min): 1500
B3_1-B3_2B3_3
Table 3. Input and output data structure of MCHA-TFCN.
Table 3. Input and output data structure of MCHA-TFCN.
ModuleInput-SizeOutput-Size
Feature input32, 2, 256016, 32, 1, 2560
MCHA-TFCN-I16, 32, 1, 25608, 32, 32, 2560
MCHA-TFCN-II8, 32, 32, 25604, 32, 128, 2560
MCHA-TFCN-III4, 32, 128, 25602, 32, 64, 2560
MCHA-TFCN-IV2, 32, 64, 256032, 32, 2560
AdaptiveAvgPool1d32, 32, 256032, 1, 2560
Table 4. Hyperparameter settings.
Table 4. Hyperparameter settings.
Hyper-ParameterInput-Size
Batch size32
Max epochs300
Initial learning rate0.1
Gamma0.1
Milestones10, 100, 150, 200
Number of features to be fused 2 n , n = 4
Kernel sizes of multiscale conv1d layers 1 × 1 , 3 × 1 , 13 × 1 , 31 × 1
Strides of multiscale conv1d layers1, 1, 2, 2
Hidden channel list of MCHA-TFCN[32, 128, 64, 32]
Self-attention a k = 2560 , 2560 , 2560
Dropout rate0.2
Table 5. Ablation experiment (condition 1).
Table 5. Ablation experiment (condition 1).
TestRMSE
TCN(1)TCN(2)TCN-SAMCHA-TFCN
1–30.3550.1940.1570.109
1–40.3030.1150.1050.044
1–50.2090.1880.1300.118
1–60.2200.1550.1390.126
1–70.0970.1500.1090.057
Avg0.2370.1510.1280.091
Table 6. Ablation experiment (condition 2).
Table 6. Ablation experiment (condition 2).
TestRMSE
TCN(1)TCN(2)TCN-SAMCHA-TFCN
2–30.4090.2270.2300.102
2–40.3460.1060.1640.117
2–50.2160.1870.1760.099
2–60.2570.2180.1420.104
2–70.2930.1830.2480.108
Avg0.3040.1840.1900.106
Table 7. Comparison of RMSE results (condition 1).
Table 7. Comparison of RMSE results (condition 1).
TestCNN-LSTMDANNTCN-SABi-TACNMCHA-TFCN
1–30.1260.3350.1170.0900.109
1–40.1070.2510.0850.1130.044
1–50.1520.2160.0860.0900.118
1–60.1540.2090.1010.0160.126
1–70.1290.1920.1290.1490.057
Avg0.1270.2220.1040.1140.091
Table 8. Comparison of RMSE results (condition 2).
Table 8. Comparison of RMSE results (condition 2).
TestCNN-LSTMDANNTCN-SABi-TACNMCHA-TFCN
2–30.1740.1680.2300.2030.102
2–40.1360.1060.0640.1370.117
2–50.1500.2280.1500.1760.099
2–60.1670.2370.1540.1420.104
2–70.1830.1920.2630.2480.108
Avg0.1620.1860.1720.1810.106
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, C.; Jiang, J.; Qi, H.; Zhang, D.; Han, X. A Novel Temporal Fusion Channel Network with Multi-Channel Hybrid Attention for the Remaining Useful Life Prediction of Rolling Bearings. Processes 2024, 12, 2762. https://doi.org/10.3390/pr12122762

AMA Style

Wang C, Jiang J, Qi H, Zhang D, Han X. A Novel Temporal Fusion Channel Network with Multi-Channel Hybrid Attention for the Remaining Useful Life Prediction of Rolling Bearings. Processes. 2024; 12(12):2762. https://doi.org/10.3390/pr12122762

Chicago/Turabian Style

Wang, Cunsong, Junjie Jiang, Heng Qi, Dengfeng Zhang, and Xiaodong Han. 2024. "A Novel Temporal Fusion Channel Network with Multi-Channel Hybrid Attention for the Remaining Useful Life Prediction of Rolling Bearings" Processes 12, no. 12: 2762. https://doi.org/10.3390/pr12122762

APA Style

Wang, C., Jiang, J., Qi, H., Zhang, D., & Han, X. (2024). A Novel Temporal Fusion Channel Network with Multi-Channel Hybrid Attention for the Remaining Useful Life Prediction of Rolling Bearings. Processes, 12(12), 2762. https://doi.org/10.3390/pr12122762

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop