Next Article in Journal
Using Inertial Measurement Units and Machine Learning to Classify Body Positions of Adults in a Hospital Bed
Previous Article in Journal
Tracking Boats on Amazon Rivers—A Case Study with the LoRa/LoRaWAN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Deep-Learning Method for Remaining Useful Life Prediction of Power Machinery via Dual-Attention Mechanism

by
Fan Wang
1,
Aihua Liu
1,2,*,
Chunyang Qu
1,
Ruolan Xiong
1 and
Lu Chen
1
1
School of Naval Architecture, Ocean and Energy Power Engineering, Wuhan University of Technology, Wuhan 430070, China
2
Sanya Science and Education Innovation Park, Wuhan University of Technology, Sanya 572024, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(2), 497; https://doi.org/10.3390/s25020497
Submission received: 8 December 2024 / Revised: 7 January 2025 / Accepted: 15 January 2025 / Published: 16 January 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
Remaining useful life (RUL) prediction is a cornerstone of Prognostic and Health Management (PHM) for power machinery, playing a crucial role in ensuring the reliability and safety of these critical systems. In recent years, deep learning techniques have shown great promise in RUL prediction, providing more reliable and accurate outcomes. However, existing models often struggle with comprehensive feature extraction, especially in capturing the complex behavior of power machinery, where non-linear degradation patterns arise under varying operational conditions. To tackle this limitation, we propose a multi-feature fusion model leveraging a dual-attention mechanism. Initially, convolutional neural networks (CNNs) and channel attention mechanisms are employed to preliminarily extract spatial features. Subsequently, a layer combining a Gate Recurrent Unit (GRU) and self-attention mechanisms is used to further extract and integrate temporal features. Finally, RUL values are predicted via regression. The effectiveness of the proposed method was validated on C-MAPSS datasets, and its superior performance in RUL prediction was demonstrated through comparative analysis with other methods.

1. Introduction

With the rapid advancement of technologies such as the Industrial Internet, the Internet of Things, and 5G, the performance and reliability of power machinery, particularly turbofan engines, have become increasingly crucial. Turbofan engines, as the core power source for modern aircraft, operate under extreme conditions and are subject to high demands for reliability to ensure safe and efficient flight. As a key component of PHM systems, the prediction of RUL plays a critical role in the operation of power machinery such as aircraft engines and their components [1,2]. The RUL prediction method is divided into a physical-based method and data-driven methods. By using existing machine learning methods and theories, data-driven methods learn information about the degradation process from the original condition-aware data and establish a mapping relationship between the degradation data and the RUL of the equipment. The data-driven method, which only relies on the learning capability of established machine learning models with little prior knowledge or detailed understanding of the equipment’s operation mechanism, has garnered widespread attention in both industry and academia, as it offers an alternative to the physical model-based methods. With the explosive growth of artificial intelligence, many machine learning algorithms have been used to predict the RUL of the equipment and have achieved notable results, such as support vector machines (SVMs) [3,4] and support vector regression (SVR) models [5].
As an end-to-end learning method, deep learning has attracted intense attention in RUL prediction. Unlike traditional machine learning methods, deep learning methods, which do not require manual feature selection, as they can automatically learn representative features from raw sensory data [6,7], have shown great promise in the RUL prediction of power machinery. Ma et al. [8] applied stacked sparse autoencoders to extract degradation features from multi-sensor monitoring data of aircraft engines, followed by logistic regression to predict the remaining useful life of the engines. Huang et al. [9] proposed two independent bidirectional LSTM models for predicting the RUL of turbofan engines. The model uses operating conditions as auxiliary input information to process degradation signals from multiple sensors. Zhai et al. [10] fused sensor data with a CNN, encoded it using a transformer, and extracted features with Long Short-Term Memory (LSTM). A contrastive learning-based tendency retention module was introduced to preserve degradation information. Validated on the NASA C-MAPSS dataset, the method showed superior prediction accuracy compared to existing methods. Zeng et al. [11] integrated attention mechanisms, Graph Attention Networks (GATs), and transformers (AGATT). The method is validated on the C-MAPSS dataset, and the results show that AGATT outperforms state-of-the-art methods in three of the four prediction tasks and achieves comparable results in the remaining one. ZHANG et al. [12] improved the traditional deep belief network and proposed a multi-target deep belief network collection method. Each depth believes that the network output occupies a certain weight on the output of the network collection. At the same time, the test demonstrates the decisive superiority of the method for analyzing NASA’s aviation engine data. Xiao et al. [13] ingeniously utilized noise to enhance the RUL prediction performance of LSTM for aircraft engines. However, two drawbacks limit the widespread application of this RUL prediction method. Kui Hu et al. [14] constructed multiple deep bidirectional recurrent neural networks (DBRNNs) with different neuron structures to extract degradation features in both forward and backward directions. These degradation features were then fed into multiple regression decision tree models for RUL prediction. DBRNN demonstrated superior performance compared to contrasting methods.
As widely studied deep learning methods, CNN and GRU are often used to capture the spatial and temporal dependencies of data in RUL prediction [15,16]. Xiao et al. [17] selected the extracted features in the time domain, frequency domain, time–frequency domain to construct the trend features, and predicted the bearing RUL by the GRU model. Shang et al. [18] directly applied convolution and pooling operations to the raw data to extract features, which were then fed into bidirectional gated recurrent units for temporal degradation information extraction. Finally, a fully connected layer was employed for RUL prediction, significantly enhancing computational efficiency. Li et al. [19] established a multi-scale CNN with a robust feature extraction ability to reflect the probability distribution of the RUL. Experimental results showed that their model performed better than other advanced models. Wang et al. [20] proposed a method combining a transformer encoder and a temporal convolutional neural network (TCNN) for RUL prediction. The transformer captures long-range dependencies, while the TCNN addresses local feature sensitivity. The model, tested on the C-MAPSS dataset, shows strong performance, particularly under complex conditions. The results demonstrated that this combined method is more accurate than using a single network for prediction. Although the aforementioned deep learning methods have achieved competitive performance, the RUL prediction method based on GRU and CNN still has some limitations: the conventional GRU only uses learned features at the last step for regression or classification. The learned features at other time steps may also have some contribution. Therefore, it is proved to be an effective method to assign weight to features reasonably. Nevertheless, the feature extraction capability of the CNN network is excellent, but it does not weigh different features, which leads to the extracted features being treated as equally important.
The self-attention mechanism (SAM) was first proposed by Bahdanau [21]. In general, the self-attention mechanism provides direct correlations between the different states of a sequence in a predictive model [22]. As the name implies, the self-attention mechanism can assign different attention weights to different parts, making the model more focused on the crucial part of the training process [23]. Fan et al. [24] proposed an end-to-end model called the Trend Attention Fully Convolutional Network to apply the attention mechanism to RUL prediction. Besides these traditional DL models, hybrid and revised DL models for RUL prediction have also been studied. Que et al. [25] connected GRU with the (Dynamic Time Distortion Warping, DTW) attention mechanism, which focuses on the weights of different time steps to improve information integrity. Yao et al. [26] introduced the attention mechanism into the GRU model and effectively predicted the remaining life of different types of rolling bearings. In conclusion, the self-attention mechanism can effectively solve the limitation of the GRU, and the combination of the GRU and self-attention mechanism can better extract features in the time dimension. Nevertheless, GRU-SAM ignores the importance of spatial dimension features. Therefore, this paper proposes an advanced RUL prediction method that leverages a dual-attention mechanism, combining CNN with the channel attention mechanism (CNN-CAM) and GRU with the self-attention mechanism (GRU-SAM). This model is designed to address the limitations of existing methods by dynamically assigning different weights to spatial and temporal features, thus improving the prediction accuracy for power machinery like turbofan engines.
Based on previous studies, this paper introduces the framework of the dual-attention mechanism. CNN-CAM can assign different weights to channels and extract features from spatial dimensions. Moreover, the GRU-SAM can further determine the importance of features and sequentially apply higher weights in the time dimension. To verify the validity of the proposed RUL prediction method, the C-MAPSS dataset was evaluated and compared with some advanced methods using the same dataset. The main contributions of this paper are summarized below.
(1)
This paper proposes an RUL prediction method based on a dual-attention framework, combining CNN with GRU. The CNN-CAM assigns different weights to the channels, enhancing the focus on key features, while the GRU-SAM assigns higher weights to more crucial temporal features, thereby improving feature extraction and prediction accuracy.
(2)
By combining features from both spatial and temporal dimensions, the proposed feature extraction mechanism effectively captures essential degradation information from the raw condition-monitoring data. This method fully leverages state-aware features, significantly enhancing the RUL prediction accuracy, especially for power machinery like turbofan engines.
(3)
The performance of the proposed method is evaluated using the C-MAPSS dataset. The experimental results demonstrate that the proposed method significantly improves RUL prediction accuracy compared to existing advanced methods, providing an effective solution for health management and fault prediction in power machinery.
The rest of this article is organized as follows: Section 2 describes the established architecture. Section 3 discusses the experimental details and results analysis. Finally, Section 4 concludes this paper and provides some future insights.

2. Methodology

Figure 1 shows the architecture of the proposed method, which consists of three parts: data preparation, model construction, RUL prediction. First, the raw sensory data are normalized and divided into training and datasets. Then, the training set data are input into the model for training. Finally, the model is verified, and the RUL prediction is completed with the test set.
In this paper, a double-attention-based model is developed to tackle the RUL prognostics. Figure 2 shows the structure of the proposed method. In this study, we propose a method comprising two key modules: CNN-CAM and GRU-SAM, each leveraging specific attention mechanisms to enhance feature extraction capabilities. Initially, the input data are processed through the CNN-CAM network. The CNN is responsible for extracting spatial features from the input data. Subsequently, these spatial features are processed through CAM, which dynamically emphasizes the most informative channels. This step ensures that the model effectively captures critical spatial features necessary for accurate predictions. Next, the features processed by the CNN-CAM module are passed to the GRU-SAM module. The GRU handles the sequential nature of the data, preserving temporal dependencies. To further optimize the extraction of temporal features, an SAM is applied. This mechanism allows the model to dynamically focus on the most relevant time steps, thereby enhancing the representation of temporal features and ensuring more accurate predictions. Finally, the fused spatiotemporal features are passed through a regression layer to achieve RUL prediction. Details of the various parts of the method are described below.

2.1. CNN-CAM

As a classical feedforward neural network, CNN was first proposed by LeCun [27] to solve the image processing problem. It is mainly composed of several convolution layers and pooling layers. By constructing multiple filters, the features extracted by CNN will become more and more abstract with the deepening of the network hierarchy [28]. For CNN, the original input data are convolved by multiple local filters through the convolution layer. The subsequent pooling layer extracts the essential features with fixed lengths [29]. Due to the excellent feature extraction capability of CNN-CAM, this paper proposes a framework that integrates CNN and the channel attention mechanism, as shown in Figure 3.
First, the original time series data are preprocessed to form the input data. The input data features can be expressed as follows:
X = x 1 , x 2 , x 3 , x k , where X R T × K
where x k denotes the K th channel in the feature graph X; while T and k represent the time step and the amount of sensor data.
The convolution calculation is shown as follows:
C i = φ W i X i 1 + b i
where X i 1 represents the input of the i th convolution layer, W i denotes the i th convolution kernel, and φ and b i represent the activation function and the function’s offset term.
Next, CAM is used to process the output data of the convolution layer. Firstly, the feature graph X output by the convolution layer is squeezed. Then, the spatial information in the data is integrated through the global average pooling layer and global maximum pooling layer. After extrusion, the spatial dimension of the given feature map changes from h × w × c to 1 × 1 × c. The generated channel descriptor corresponding to the feature graph X of each channel can be described as follows:
P k = S X k = 1 h × w i = 1 h j = 1 w x k ( i , j )
where x k ( i , j ) represents the element in row i, column j of the feature graph X.
The channel descriptor P k is obtained from the feature graph after extrusion operation. The global distribution of channel feature responses is embedded in the descriptor, which helps the network utilize the information from the global receptive field at a lower level. Subsequently, P k is excited by the two fully connected layers and two tanh activation functions. During the excitation process, extraction descriptor L is generated. L is expressed as follows:
L = E ( P , F C ) = tanh F C 2 tanh F C 1 P
where F C 1 is a dimension reduction layer, F C 2 is a dimension increase layer, and tanh is an activation function. Two fully connected layers reduce the model complexity through dimensionality reduction, significantly simplifying the calculation process and minimizing the number of parameters. And the tanh function sets the output weight between [−1, 1], which avoids the problem of center symmetry. Then, weighted processing is used to re-calibrate the original feature in CAM.
Finally, in order to integrate the data on the two branches of the global average pooling layer and global maximum pooling layer, the element summation is assigned to compose the output feature graph. Then, the weights are normalized by using the softmax function. Then, a scaling operation is used to assign normalized weights to the features of each channel. The final output of this module is transformed directly into the next network layer.
This paper uses CNN-CAM to extract the feature values in the spatial dimension. In order to extract the feature information more comprehensively, it is necessary to extract the feature values in the time dimension of the data. Therefore, GRU-SAM is implemented to perform feature extraction in the time dimension. The specific process is described below.

2.2. GRU-SAM

GRU is a type of neural network derived from recurrent neural networks (RNNs), first proposed by Cho et al. [30] Compared to traditional CNNs, GRU not only extracts temporal sequence information but also effectively addresses the vanishing gradient and exploding gradient problems inherent in conventional RNNs. The core idea of GRU is the introduction of gating mechanisms, as illustrated in Figure 4. It comprises two gating units: the update gate and the reset gate. These gating units control the flow of information within the network, allowing the selective forgetting or updating of certain information. This mechanism enables GRUs to better handle long-term dependencies while reducing the number of parameters that need to be trained. The computational formulas for the GRU model are as follows:
z t = σ W z h t 1 , x t
r t = σ W r h t 1 , x t
h ˜ t = tanh W h r t h t 1 , x t
h t = 1 z t h t 1 + z t h ˜ t
where z t is the output vector at time step t of the update gate; W z denotes its weight parameters; h t 1 represents the hidden state of the previous moment; x t is the input of the time step t; σ indicates the sigmoid activation function; r t is the output vector at time step t of the reset gate; W r denotes its weight parameters; h t is the hidden vector at time step t; W h denotes its weight parameters; tanh is the hyperbolic tangent activation function; and h t signifies the memorized state vector.
While the GRU network is capable of processing time series data, it can only output a fixed-length sequence and does not differentiate between features in terms of their importance. Simultaneously, for excessively long sequences, the capacity of the units to store information may need to be increased, which may result in reduced prediction accuracy. In order to solve these two problems, this paper proposes a method combining the self-attention mechanism and GRU.
Originating from the differences in the attention of the human visual system [31,32], the attention mechanism is used to allocate different feature information resources [33]. As technology evolves, self-attentive mechanisms are combined with deep learning methods and have achieved notable success. In the prediction process, the attention model will autonomously find the features that are more important to the prediction result and assign higher weights to them so as to improve the accuracy of the prediction.
In this paper, the learning feature output by the GRU network is expressed as H = h 1 , h 2 , h n T , where T represents the transpose operation, and h i = x 1 , x 2 , x n T is the self-attention mechanism’s input, and represents the data’s time step. The specific formula is as follows:
u i = softmax W T h i + b
a i = exp u i i = 1 n exp u i
where W T and b are the weight matrix and bias terms in the attention network, a i is the influence of the i th eigenvalue on the target sequence data, and all are added to 1. Then, the feature sequence is weighted to obtain the output sequence, which can be expressed as follows:
A H = i = 1 n a i u i
The integration of GRU with SAM allows the model to dynamically focus on more relevant features, enhancing the feature representation and significantly improving the model’s performance.

3. Experimental Study

In this section, we discuss the validation of the RUL prediction method presented in this paper and compare it with some advanced methods using generic datasets. Section 3.1 describes the details of this dataset; Section 3.2 introduces the related technologies and details of data preprocessing in detail; and Section 3.3 proposes two indexes to evaluate RUL prediction. In Section 3.4, the practical steps and determining relevant hyperparameters are elaborated. Section 3.5 discusses the effects of some hyperparameters on the results and uses ablation experiments to determine the validity of the methods studied in this paper. Finally, the results obtained in the experiments are compared with those obtained by some current advanced methods.

3.1. Dataset Description

To investigate the degradation modeling and life prediction of aero-engines, NASA used the Commercial Modular Aerospace Propulsion Simulation System (C-MAPSS) to simulate a series of performance degradation processes of major engine components (as shown in Figure 5) and made the corresponding engine performance degradation dataset publicly available [34].
The components of generate power include fans, gas chambers, low-pressure turbines, high-pressure compressors, and nozzles. The engine is designed to perform simulated operations at a sea level of 40,000 feet with 90,000 pounds of thrust, Mach numbers of 0 to 0.90, and temperatures of −60 to 103 °F. The C-MAPSS dataset includes four sub-datasets, namely FD001, FD002, FD003, and FD004. Each sub-dataset is further divided into training and test datasets. The training and test datasets are composed of the sensor data of a specific engine during each run cycle, helping to predict the true RUL of each engine in the test file with the given sensor measurements. The datasets are arranged in the N × 26 matrix. Each data sample includes 26 variables:
  • Engine unit number of each engine;
  • Degradation cycle of each turbofan engine;
  • Operation setting parameters of three turbofan engines;
  • Sampling data of 21 sensors on the turbofan engine in each operation cycle.
N is the number of signals recorded for each engine. Table 1 shows the specific information of the dataset [35].
The C-MAPSS dataset has detailed multi-sensor measurements, different operating conditions, and benchmark states in the RUL prediction field. Due to its comprehensive representation of turbofan engine degradation and widespread use in academic research, the C-MAPSS dataset was chosen in this paper to validate the proposed method.

3.2. Data Preprocessing

3.2.1. Data Normalization

Collected from multiple sensors, the dataset has different value ranges. Figure 6a shows the degradation data of “Sensor 2” for 100 engines in the FD001 dataset, which illustrates much noise. Figure 6b shows the lifetime distribution of 100 engines in the FD001 dataset. Since the uneven life distribution increases the computational complexity of the model, it is necessary to normalize the data to speed up the convergence and improve the prediction accuracy. In this paper, the Min–Max method is adopted to normalize the data of the training set and the test set. The formula is as follows:
x i = x i x i min x i max x i min
where x i represents the value of the current time point; x i max  and  x i min refer to the maximum and minimum values of the current sensor at all time points, respectively; and x i denotes the data obtained after normalized calculation. The normalized data will be within the range of [0, 1].
It is essential to mention here that the sensors labeled as 1, 5, 6, 10, 16, 18, and 19 show relatively smooth behavior in the degradation experiment, which means that these data do not show regression characteristics, i.e., they do not provide significant deterioration information. Therefore, the data from these seven sensors should be excluded from this paper, and the data from the other 14 sensors will be applied to train the model proposed in this paper.

3.2.2. Sliding Window Processing

The sliding window method is commonly employed to partition multivariate time series data, enabling the model to capture temporal dependencies across various time steps. An illustrative example of sliding window processing is depicted in Figure 7. Selecting an appropriate window length is crucial, as a window that is too large can increase the complexity of the model and reduce its practical utility, while a window that is too small may fail to adequately capture the underlying relationships in the time series. The normalized raw data, after being processed by a sliding window, become X = x 1 , x 2 , x k .

3.2.3. RUL Label Settings

In the C-MAPSS dataset, sensor output data from the turbofan engine remain relatively stable during the initial phase due to the absence of faults or degradation. Hence, early prediction of remaining useful life (RUL) is not only futile but also computationally inefficient. Prior studies [36,37] indicate that a piecewise linear model can confine the maximum RUL within a specific range, thereby preventing excessive RUL predictions. Consequently, this study employs a piecewise linear model to process engine RUL, as illustrated in Figure 8. The maximum RUL is set to 125; any value exceeding 125 is uniformly capped at 125. To validate this maximum RUL setting, comparative analyses were conducted using four different values: 150, 140, 130, and 125, across four datasets. The results, depicted in Figure 9, demonstrate that a maximum RUL value of 125 yields the most accurate predictions.

3.3. Evaluation Metrics

In order to verify the validity and accuracy of the method, two evaluation metrics were utilized: root means square error (RMSE [38]) and score function (Score [39]). RMSE can measure the extent to which the predicted RUL value deviates from the real RUL value. At the same time, Score is an index proposed by PHM08 data competition to evaluate the predicted performance. The smaller the value of the two indexes is, the better the prediction performance will be. These two evaluation indexes will be used comprehensively to evaluate the model’s prediction performance. The formula of RMSE is as follows:
RMSE = 1 N i = 1 N y ^ i y i 2
where N represents the total number of samples, and y ^ i and y i are predicted from the RUL value and the actual RUL, respectively.
The scoring function is one of NASA’s evaluation metrics for this open question. In the RUL forecast, if the predicted value is less than the actual value, it is considered an early forecast, and the subsequent decision will be more conservative based on the result. The situation where the predicted value is greater than the true value is called a late forecast, which may lead to an accident or incident that cannot be reflected by a single RMSE.
In order to improve the evaluation system, this paper introduces the scoring function, which expressed as follows:
Score = i = 1 N e y ^ i y i 13 1 , y ^ i < y i i = 1 N e y i y i 10 1 , y ^ i y i
where N represents the total number of samples, and y ^ i and y i represent the predicted RUL value and the actual RUL values, respectively.
The relationship between Score and RMSE is shown in Figure 9. With the increase in the error between the predicted and actual values, the RMSE increases linearly, and the Score curve increases exponentially. Early prediction (left side of Figure 10) is better than late prediction (right side of Figure 10) because the goal is to prevent engine failure. The late prediction will be dominated by a single outlier, thus hiding the actual overall accuracy of the model. Therefore, in practice, RMSE and Score should be integrated to evaluate the model’s performance comprehensively.

3.4. Experimental Setup

To validate the accuracy and effectiveness of the model, the training data of FD001 to FD004 were tested first in this paper. The original perception data were extracted by one-dimensional convolution and put into the channel attention mechanism for adaptive weight allocation. The features output from CNN-CAM were entered as input into the network combined with GRU-SAM for training and weight allocation again. The experiment used the training set to train the model and then the test set to verify the model. The algorithm iterated over each dataset ten times (32 epochs were included in each iteration) and recorded the best RMSE and Score once in each iteration. Applying a dropout rate of 0.2 (i.e., 20% of randomly selected hidden layer neurons were ignored) helped to minimize the overfitting that tends to occur during training [40]. Finally, the best data for the iteration were saved. The hyperparameters of the proposed method are shown in Table 2.

3.5. Result Analysis

3.5.1. Impact of Window Size

Window size is one of the most critical parameters in the model, directly influencing its performance outcomes. The impact of window size on the proposed performance prediction model was investigated using the FD001-FD004 datasets. Five window sizes—25, 35, 45, 55, and 60—were evaluated, with results shown in Figure 11. Overall, RMSE varied with window size, and trends differed across datasets. In the FD001 dataset, performance improved as window size increased from 25 to 35, likely due to more informative data being included. However, further increases in window size resulted in decreased predictive performance, possibly due to overfitting. As shown in Table 1, while the FD003 dataset is broadly similar to FD001, it has two operational modes, necessitating a larger window size. Therefore, more complex data may require additional information to enhance RUL prediction accuracy. This is corroborated by the FD002 and FD004 datasets, where model performance improved with increasing window size. Hence, different datasets require tailored window sizes to optimize RUL prediction accuracy. The optimal window sizes for the four datasets are summarized in Table 3.

3.5.2. Ablation Study of the Proposed Architecture

To determine the validity of the method proposed in this paper, ablation experiments were performed for validation. This paper used the model with (or without) CAM to explore the capability of CAM and the model with (or without) SAM to analyze the role of SAM in the feature weight assignment. Specifically, three methods, including No CAM, No SAM, and the model used in this paper, were selected for the comparative study, and the experimental results are shown in Table 4 and Figure 12.
Figure 12 presents a comparative analysis of the predictive performance of various model configurations through ablation experiments. The introduction of the SAM and the CAM significantly enhanced prediction performance. This improvement is more intuitively illustrated in Table 4, where it is shown that the removal of CAM resulted in a marked decline in model performance, underscoring its crucial role in capturing key features. Although CAM had a greater impact on overall error reduction, the SAM also significantly improved model performance by capturing long-term dependencies and dynamically weighting inputs. In summary, both SAM and CAM are pivotal in enhancing predictive model performance. This study proposes that integrating these two mechanisms enables the model to perform well across all datasets, thereby validating the effectiveness and necessity of the dual-attention mechanism.

3.5.3. Prognostic Results Analysis

The predicted results and the actual RUL values are shown in Figure 13. As can be seen from the figure, the predicted RULs of the four datasets closely match the actual RUL values. The fitted curves on FD001 and FD003 are close to the actual values, showing that the model can effectively predict the RUL. On the other hand, most of the predicted results in the FD002 and FD004 datasets are smaller than the actual results, which indicates that the model is capable of significantly mitigating the risks associated with late predictions.

3.5.4. Comparison with the State-of-the-Art Methods

Table 5 and Table 6 present the experimental results of the proposed method compared to several state-of-the-art methods across four datasets. Results highlighted in bold indicate the best performance among all methods.
As shown in Table 5, the proposed method achieved the lowest RMSE on three of the four datasets (FD002, FD003, and FD004), demonstrating its superior prediction accuracy. On the FD001 dataset, although the RMSE of the proposed method was slightly higher than other advanced methods such as the double-attention-based architecture [47], the overall average RMSE across all datasets (14.47) was the best among the compared methods. This indicates that the proposed method performs well in both simple and complex scenarios. In particular, the proposed method showed a significant improvement in the complex datasets FD002 and FD004, where the reduction in RMSE compared to other methods, such as MODBNE [12], was particularly pronounced. This highlights the proposed method’s robustness and adaptability in handling complex operational conditions.
Furthermore, as seen in the Score metrics of Table 6, the proposed method consistently achieved better scores across three datasets (FD002, FD003, and FD004). While the Score on FD001 was slightly worse than BLSTM-CNN [48], the proposed method still maintained a balance between prediction accuracy and early failure detection. Early predictions can help reduce maintenance costs and avoid potential damages in practical scenarios, as demonstrated by the superior performance in FD004.
In summary, compared with advanced methods like the double-attention-based architecture [47] and BiGRU-TSAM [46], the proposed method surpassed them in RMSE and Score on average, showing its ability to achieve accurate and reliable RUL predictions while maintaining a practical advantage in handling complex conditions.

4. Conclusions

This study proposes a feature fusion framework for RUL prediction, integrating CNN and GRU with a dual-attention mechanism. Initially, spatial features were extracted using CNN-CAM, while temporal features were subsequently extracted using GRU-SAM, enabling the fusion of spatio-temporal features to enhance feature extraction capabilities. Finally, the trained neural network model was applied for RUL prediction. In this study, we utilized the C-MAPSS datasets and evaluated our proposed method using RMSE and score function metrics, comparing it with other state-of-the-art methods. Additionally, ablation experiments were conducted to validate the necessity and effectiveness of our approach. Experimental results demonstrate that our proposed method outperformed other advanced methods, effectively integrating information on power machinery characteristics through the dual-attention mechanism. Future research will focus on two key aspects: first, exploring more advanced feature fusion techniques and incorporating additional domain knowledge to further enhance model performance; second, optimizing computational efficiency by exploring lightweight model architectures or hardware-accelerated solutions, enabling the proposed method to achieve faster training and inference times while maintaining accuracy.

Author Contributions

Investigation, L.C. and R.X.; data curation, C.Q.; writing—original draft, F.W.; writing—review and editing, A.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Sanya Science and Education Innovation Park of Wuhan University of Technology (Grant No. 2022KF0019).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fu, S.; Zhang, Y.; Lin, L.; Zhao, M.; Zhong, S.-s. Deep residual LSTM with domain-invariance for remaining useful life prediction across domains. Reliab. Eng. Syst. Saf. 2021, 216, 108012. [Google Scholar] [CrossRef]
  2. Zhu, Y.; Wu, J.; Wu, J.; Liu, S. Dimensionality reduce-based for remaining useful life prediction of machining tools with multisensor fusion. Reliab. Eng. Syst. Saf. 2022, 218, 108179. [Google Scholar] [CrossRef]
  3. Nieto, P.G.; García-Gonzalo, E.; Lasheras, F.S.; de Cos Juez, F.J. Hybrid PSO–SVM-based method for forecasting of the remaining useful life for aircraft engines and evaluation of its reliability. Reliab. Eng. Syst. Saf. 2015, 138, 219–231. [Google Scholar] [CrossRef]
  4. Ordóñez, C.; Lasheras, F.S.; Roca-Pardiñas, J.; de Cos Juez, F.J. A hybrid ARIMA–SVM model for the study of the remaining useful life of aircraft engines. J. Comput. Appl. Math. 2019, 346, 184–191. [Google Scholar] [CrossRef]
  5. Khelif, R.; Chebel-Morello, B.; Malinowski, S.; Laajili, E.; Fnaiech, F.; Zerhouni, N. Direct remaining useful life estimation based on support vector regression. IEEE Trans. Ind. Electron. 2016, 64, 2276–2285. [Google Scholar] [CrossRef]
  6. Liao, L.; Jin, W.; Pavel, R. Enhanced restricted Boltzmann machine with prognosability regularization for prognostics and health assessment. IEEE Trans. Ind. Electron. 2016, 63, 7076–7083. [Google Scholar] [CrossRef]
  7. Zhao, R.; Yan, R.; Chen, Z.; Mao, K.; Wang, P.; Gao, R.X. Deep learning and its applications to machine health monitoring. Mech. Syst. Signal Process. 2019, 115, 213–237. [Google Scholar] [CrossRef]
  8. Ma, J.; Su, H.; Zhao, W.l.; Liu, B. Predicting the remaining useful life of an aircraft engine using a stacked sparse autoencoder with multilayer self-learning. Complexity 2018, 2018, 3813029. [Google Scholar] [CrossRef]
  9. Huang, C.G.; Huang, H.Z.; Li, Y.F. A Bidirectional LSTM Prognostics Method Under Multiple Operational Conditions. Ph.D. Thesis, University of Electronic Science and Technology of China, Chengdu, China, 2019. [Google Scholar]
  10. Zhai, Z.; Wen, J.; Wang, F.; Zhao, Z.; Guo, Y.; Chen, X. Remaining Useful Life Prediction of Aero-Engine Based on Transformer with Tendency Retainment. In Proceedings of the 2022 International Conference on Sensing, Measurement & Data Analytics in the Era of Artificial Intelligence (ICSMD), Harbin, China, 30 November–2 December 2022; pp. 1–6. [Google Scholar]
  11. Zeng, M.; Wu, F.; Cheng, Y. Remaining Useful Life Prediction via Spatio-Temporal Channels and Transformer. IEEE Sens. J. 2023, 23, 29176–29185. [Google Scholar] [CrossRef]
  12. Zhang, C.; Lim, P.; Qin, A.K.; Tan, K.C. Multiobjective deep belief networks ensemble for remaining useful life estimation in prognostics. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2306–2318. [Google Scholar] [CrossRef]
  13. Xiao, L.; Tang, J.; Zhang, X.; Bechhoefer, E.; Ding, S. Remaining useful life prediction based on intentional noise injection and feature reconstruction. Reliab. Eng. Syst. Saf. 2021, 215, 107871. [Google Scholar] [CrossRef]
  14. Hu, K.; Cheng, Y.; Wu, J.; Zhu, H.; Shao, X. Deep bidirectional recurrent neural networks ensemble for remaining useful life prediction of aircraft engine. IEEE Trans. Cybern. 2021, 53, 2531–2543. [Google Scholar] [CrossRef]
  15. Yu, W.; Kim, I.Y.; Mechefske, C. Remaining useful life estimation using a bidirectional recurrent neural network based autoencoder scheme. Mech. Syst. Signal Process. 2019, 129, 764–780. [Google Scholar] [CrossRef]
  16. Song, X.; Chen, K.; Li, X.; Sun, J.; Hou, B.; Cui, Y.; Zhang, B.; Xiong, G.; Wang, Z. Pedestrian trajectory prediction based on deep convolutional LSTM network. IEEE Trans. Intell. Transp. Syst. 2020, 22, 3285–3302. [Google Scholar] [CrossRef]
  17. Li, X.; Teng, W.; Peng, D.; Ma, T.; Wu, X.; Liu, Y. Feature fusion model based health indicator construction and self-constraint state-space estimator for remaining useful life prediction of bearings in wind turbines. Reliab. Eng. Syst. Saf. 2023, 233, 109124. [Google Scholar] [CrossRef]
  18. Shang, Y.; Tang, X.; Zhao, G.; Jiang, P.; Lin, T.R. A remaining life prediction of rolling element bearings based on a bidirectional gate recurrent unit and convolution neural network. Measurement 2022, 202, 111893. [Google Scholar] [CrossRef]
  19. Li, H.; Zhao, W.; Zhang, Y.; Zio, E. Remaining useful life prediction using multi-scale deep convolutional neural network. Appl. Soft Comput. 2020, 89, 106113. [Google Scholar] [CrossRef]
  20. Wang, H.K.; Cheng, Y.; Song, K. Remaining useful life estimation of aircraft engines using a joint deep learning model based on TCNN and transformer. Comput. Intell. Neurosci. 2021, 2021, 5185938. [Google Scholar] [CrossRef]
  21. Bahdanau, D. Neural machine translation by jointly learning to align and translate. arXiv 2014, arXiv:1409.0473. [Google Scholar]
  22. Raffel, C.; Ellis, D.P. Feed-forward networks with attention can solve some long-term memory problems. arXiv 2015, arXiv:1512.08756. [Google Scholar]
  23. Lin, Z.; Feng, M.; Santos, C.N.d.; Yu, M.; Xiang, B.; Zhou, B.; Bengio, Y. A structured self-attentive sentence embedding. arXiv 2017, arXiv:1703.03130. [Google Scholar]
  24. Fan, L.; Chai, Y.; Chen, X. Trend attention fully convolutional network for remaining useful life estimation. Reliab. Eng. Syst. Saf. 2022, 225, 108590. [Google Scholar] [CrossRef]
  25. Que, Z.; Jin, X.; Xu, Z. Remaining useful life prediction for bearings based on a gated recurrent unit. IEEE Trans. Instrum. Meas. 2021, 70, 1–11. [Google Scholar] [CrossRef]
  26. Yao, D.; Li, B.; Liu, H.; Yao, J.; Pi, Y. Residual life prediction of rolling bearing based on attention GRU algorithm. J. Vib. Shock 2021, 40, 116–123. [Google Scholar]
  27. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  28. Lawrence, S.; Giles, C.L.; Tsoi, A.C.; Back, A.D. Face recognition: A convolutional neural-network approach. IEEE Trans. Neural Netw. 1997, 8, 98–113. [Google Scholar] [CrossRef]
  29. Sun, M.; Song, Z.; Jiang, X.; Pan, J.; Pang, Y. Learning pooling for convolutional neural network. Neurocomputing 2017, 224, 96–104. [Google Scholar] [CrossRef]
  30. Cho, K. On the properties of neural machine translation: Encoder-decoder approaches. arXiv 2014, arXiv:1409.1259. [Google Scholar]
  31. Xu, K. Show, attend and tell: Neural image caption generation with visual attention. arXiv 2015, arXiv:1502.03044. [Google Scholar]
  32. Du, W.; Wang, Y.; Qiao, Y. Recurrent spatial-temporal attention network for action recognition in videos. IEEE Trans. Image Process. 2017, 27, 1347–1360. [Google Scholar] [CrossRef]
  33. Carrasco, M.; Barbot, A. Spatial attention alters visual appearance. Curr. Opin. Psychol. 2019, 29, 56–64. [Google Scholar] [CrossRef] [PubMed]
  34. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage propagation modeling for aircraft engine run-to-failure simulation. In Proceedings of the 2008 International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008; pp. 1–9. [Google Scholar]
  35. Zhang, Y.; Xin, Y.; Liu, Z.w.; Chi, M.; Ma, G. Health status assessment and remaining useful life prediction of aero-engine based on BiGRU and MMoE. Reliab. Eng. Syst. Saf. 2022, 220, 108263. [Google Scholar] [CrossRef]
  36. Zhao, Z.; Liang, B.; Wang, X.; Lu, W. Remaining useful life prediction of aircraft engine based on degradation pattern learning. Reliab. Eng. Syst. Saf. 2017, 164, 74–83. [Google Scholar] [CrossRef]
  37. Hsu, C.S.; Jiang, J.R. Remaining useful life estimation using long short-term memory deep learning. In Proceedings of the 2018 IEEE International Conference on Applied System Invention (ICASI), Chiba, Japan, 13–17 April 2018; pp. 58–61. [Google Scholar]
  38. Yu, W.; Kim, I.Y.; Mechefske, C. An improved similarity-based prognostic algorithm for RUL estimation using an RNN autoencoder scheme. Reliab. Eng. Syst. Saf. 2020, 199, 106926. [Google Scholar] [CrossRef]
  39. Xiang, S.; Qin, Y.; Luo, J.; Pu, H.; Tang, B. Multicellular LSTM-based deep learning model for aero-engine remaining useful life prediction. Reliab. Eng. Syst. Saf. 2021, 216, 107927. [Google Scholar] [CrossRef]
  40. Moradi, R.; Cofre-Martel, S.; Droguett, E.L.; Modarres, M.; Groth, K.M. Integration of deep learning and Bayesian networks for condition and operation risk monitoring of complex engineering systems. Reliab. Eng. Syst. Saf. 2022, 222, 108433. [Google Scholar] [CrossRef]
  41. Sateesh Babu, G.; Zhao, P.; Li, X.L. Deep convolutional neural network based regression approach for estimation of remaining useful life. In Proceedings of the Database Systems for Advanced Applications: 21st International Conference, DASFAA 2016, Dallas, TX, USA, 16–19 April 2016; Proceedings, Part I 21. Springer: Berlin/Heidelberg, Germany, 2016; pp. 214–228. [Google Scholar]
  42. Su, C.; Li, L.; Wen, Z. Remaining useful life prediction via a variational autoencoder and a time-window-based sequence neural network. Qual. Reliab. Eng. Int. 2020, 36, 1639–1656. [Google Scholar] [CrossRef]
  43. Ellefsen, A.L.; Bjørlykhaug, E.; Æsøy, V.; Ushakov, S.; Zhang, H. Remaining useful life predictions for turbofan engine degradation using semi-supervised deep architecture. Reliab. Eng. Syst. Saf. 2019, 183, 240–251. [Google Scholar] [CrossRef]
  44. Shang, Z.; Zhang, B.; Li, W.; Qian, S.; Zhang, J. Machine remaining life prediction based on multi-layer self-attention and temporal convolution network. Complex Intell. Syst. 2022, 8, 1409–1424. [Google Scholar] [CrossRef]
  45. Lin, R.; Wang, H.; Xiong, M.; Hou, Z.; Che, C. Attention-based Gate Recurrent Unit for remaining useful life prediction in prognostics. Appl. Soft Comput. 2023, 143, 110419. [Google Scholar] [CrossRef]
  46. Zhang, J.; Jiang, Y.; Wu, S.; Li, X.; Luo, H.; Yin, S. Prediction of remaining useful life based on bidirectional gated recurrent unit with temporal self-attention mechanism. Reliab. Eng. Syst. Saf. 2022, 221, 108297. [Google Scholar] [CrossRef]
  47. Liu, L.; Song, X.; Zhou, Z. Aircraft engine remaining useful life estimation via a double attention-based data-driven architecture. Reliab. Eng. Syst. Saf. 2022, 221, 108330. [Google Scholar] [CrossRef]
  48. Guo, X.; Xu, X.; Guo, J. Self-attentive optimization algorithm for aero-engine life prediction based on improved GRU. J. Aerosp. Power 2024, 39, 20220984. [Google Scholar]
Figure 1. Flow chart of the proposed method.
Figure 1. Flow chart of the proposed method.
Sensors 25 00497 g001
Figure 2. Structure of the proposed method.
Figure 2. Structure of the proposed method.
Sensors 25 00497 g002
Figure 3. Framework structure of CNN combined with channel attention mechanism.
Figure 3. Framework structure of CNN combined with channel attention mechanism.
Sensors 25 00497 g003
Figure 4. GRU network cell structure.
Figure 4. GRU network cell structure.
Sensors 25 00497 g004
Figure 5. Simplified diagram of the turbofan engine.
Figure 5. Simplified diagram of the turbofan engine.
Sensors 25 00497 g005
Figure 6. (a) # 2 Sensor measurement for 100 engines. (b) Life distribution of 100 engines.
Figure 6. (a) # 2 Sensor measurement for 100 engines. (b) Life distribution of 100 engines.
Sensors 25 00497 g006aSensors 25 00497 g006b
Figure 7. Sliding window processing.
Figure 7. Sliding window processing.
Sensors 25 00497 g007
Figure 8. Piecewise linear model.
Figure 8. Piecewise linear model.
Sensors 25 00497 g008
Figure 9. The proposed method analyzes the experimental results in different RUL values.
Figure 9. The proposed method analyzes the experimental results in different RUL values.
Sensors 25 00497 g009
Figure 10. Piecewise linear model.
Figure 10. Piecewise linear model.
Sensors 25 00497 g010
Figure 11. The proposed method has experimental results of different window sizes on four datasets. (a) FD001: RMSE. (b) FD002: RMSE (c) FD003: RMSE. (d) FD004: RMSE.
Figure 11. The proposed method has experimental results of different window sizes on four datasets. (a) FD001: RMSE. (b) FD002: RMSE (c) FD003: RMSE. (d) FD004: RMSE.
Sensors 25 00497 g011
Figure 12. Experimental results of the ablation study.
Figure 12. Experimental results of the ablation study.
Sensors 25 00497 g012
Figure 13. The prediction results of the method in this paper on four datasets and the real RUL fitting graph. (a) FD001: (b) FD002 (c) FD003 (d) FD004.
Figure 13. The prediction results of the method in this paper on four datasets and the real RUL fitting graph. (a) FD001: (b) FD002 (c) FD003 (d) FD004.
Sensors 25 00497 g013aSensors 25 00497 g013b
Table 1. Description of the C-MAPSS dataset.
Table 1. Description of the C-MAPSS dataset.
DatasetsFD001FD002FD003FD004
Engine units for training100260100249
Engine units for testing100259100248
Sensor measurements21212121
Maximum cycle362378525543
Minimum cycle128128145128
Operation conditions1616
Fault conditions1122
Table 2. Hyperparameters of the proposed method.
Table 2. Hyperparameters of the proposed method.
HyperparametersValue
Number of convolution kernels for Conv4
Activation function in CNNHyperbolic tangent (tanh)
GRU number of hidden units in the network50
Number of neurons in FC layer behind GRU[50, 10]
Window sizes for FD001 to FD00435/60/45/60
Learning rate0.001
Dropout0.2
Kernel sizes for Conv1 to Conv51 × 3
Activation function in GRU networkReLU
Table 3. Window size information of the four datasets.
Table 3. Window size information of the four datasets.
DatasetsFD001FD002FD003FD004
Time window size35605560
No. of training time windows18,23148,81921,82057,763
No. of testing time windows10,69629,07013,69637,742
Table 4. Results of RMSE in the ablation study on four datasets.
Table 4. Results of RMSE in the ablation study on four datasets.
MethodsFD001FD002FD003FD004
CNN + GRU-SAM(No CAM)15.8117.3714.4821.79
CNN-CAM + GRU (No SAM)14.1816.5413.1220.51
CNN-CAM + GRU-SAM (proposed)13.0315.4112.2116.43
Note: Bold numbers represent the best performance of the list.
Table 5. Results of the proposed method and other methods on RMSE.
Table 5. Results of the proposed method and other methods on RMSE.
RMSE
MethodsFD001FD002FD003FD004Average
CNN [41]18.4530.2919.8229.1624.43
MODBNE [12]15.0425.0512.5128.6620.315
VAE+LSTM [42]15.8825.7814.2923.9319.97
GA+RBM+LSTM [43]12.5622.7312.122.6617.5125
Multi-attention+TCN [44]13.2519.5713.4321.6916.985
ABGRU [45]12.8317.9713.2321.5516.4
BiGRU-TSAM [46]12.5618.9412.4520.4716.105
Doubleattention-based architecture [47]12.2517.0813.3919.8615.645
BLSTM-CNN [48]12.1316.0111.9618.114.55
Proposed13.0315.4112.2116.4314.47
Note: Bold numbers represent the best performance of the list.
Table 6. Results of the proposed method and other methods on Score.
Table 6. Results of the proposed method and other methods on Score.
Score
MethodsFD001FD002FD003FD004
CNN [41]12861357015967886
MODBNE [12]33455854216557
VAE+LSTM [42]32249903094720
GA+RBM+LSTM [43]23133662512840
Multi-attention+TCN [44]23516552392415
ABGRU [45]22120722793625
BiGRU-TSAM [46]21322642323610
Doubleattention-basedarchitecture [47]19815752901741
BLSTM-CNN [48]17412302421513
Proposed2177961891029
Note: Bold numbers represent the best performance of the list.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, F.; Liu, A.; Qu, C.; Xiong, R.; Chen, L. A Deep-Learning Method for Remaining Useful Life Prediction of Power Machinery via Dual-Attention Mechanism. Sensors 2025, 25, 497. https://doi.org/10.3390/s25020497

AMA Style

Wang F, Liu A, Qu C, Xiong R, Chen L. A Deep-Learning Method for Remaining Useful Life Prediction of Power Machinery via Dual-Attention Mechanism. Sensors. 2025; 25(2):497. https://doi.org/10.3390/s25020497

Chicago/Turabian Style

Wang, Fan, Aihua Liu, Chunyang Qu, Ruolan Xiong, and Lu Chen. 2025. "A Deep-Learning Method for Remaining Useful Life Prediction of Power Machinery via Dual-Attention Mechanism" Sensors 25, no. 2: 497. https://doi.org/10.3390/s25020497

APA Style

Wang, F., Liu, A., Qu, C., Xiong, R., & Chen, L. (2025). A Deep-Learning Method for Remaining Useful Life Prediction of Power Machinery via Dual-Attention Mechanism. Sensors, 25(2), 497. https://doi.org/10.3390/s25020497

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop