Next Article in Journal
UniROS: ROS-Based Reinforcement Learning Across Simulated and Real-World Robotics
Previous Article in Journal
A Performance Study of Deep Neural Network Representations of Interpretable ML on Edge Devices with AI Accelerators
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid Model Based on a Dual-Attention Mechanism for the Prediction of Remaining Useful Life of Aircraft Engines

1
Key Laboratory of Metallurgical Equipment and Control Technology of Ministry of Education, Wuhan University of Science and Technology, Wuhan 430081, China
2
Hubei Key Laboratory of Mechanical Transmission and Manufacturing Engineering, Wuhan University of Science and Technology, Wuhan 430081, China
3
Precision Manufacturing Institute, Wuhan University of Science and Technology, Wuhan 430081, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(18), 5682; https://doi.org/10.3390/s25185682
Submission received: 6 August 2025 / Revised: 3 September 2025 / Accepted: 9 September 2025 / Published: 11 September 2025
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

Estimating the Remaining Useful Life (RUL) of aircraft engines plays a vital role in the field of prognostics and health management. In multi-dimensional time series regression tasks, accurately capturing both time series features and sensor features, as well as integrating these two types of features, poses a significant challenge for RUL prediction. The sensor features represent the weights of each sensor on the RUL prediction results. To overcome this challenge, we introduce a hybrid model based on a dual-attention mechanism. Initially, a temporal feature extraction block is applied to map the time-step dimension into a hidden representation space, facilitating the capture of complex temporal dynamics. These patterns are then refined using a multi-head self-attention mechanism. Subsequently, a sensor feature extraction block is applied to capture sensor-specific characteristics. Each sensor sequence is treated as a separate channel, compressed to derive sensor weights, and integrated to form global features that fuse temporal and sensor-level representations. Finally, RUL is estimated via a regression layer. The proposed method is demonstrated to be effective on the Commercial Modular Aero-Propulsion System Simulation (C-MAPSS) dataset. Compared with the state-of-the-art CTNet model, the proposed method achieves 7% and 9% gains in RMSE and Score, respectively, on the FD001 dataset.

1. Introduction

Prognostics and health management (PHM) is an advanced, integrated framework that encompasses fault diagnosis, RUL prediction, predictive maintenance, and related functionalities [1,2]. Among these, RUL prediction is a critical component, serving as a key indicator of the real-time condition of mechanical equipment [3]. It entails predicting how much longer machinery will function effectively by analyzing data collected from sensors. By analyzing this data, the current degradation state of the equipment can be inferred, allowing for reliable forecasts of its future operational capacity. Such predictions enable the implementation of predictive maintenance strategies, which not only substantially reduce production costs but also mitigate the risk of equipment failures and associated safety risks [4,5].
At present, methods for predicting the remaining useful life (RUL) of mechanical equipment are generally divided into three primary categories: model-based methods, data-driven methods, and hybrid approaches that integrate both paradigms [6,7]. The model-based approach describes the degradation process of a device by constructing physical models or mathematical equations and combining them with relevant empirical knowledge for RUL prediction. These methods are highly interpretable due to their transparency and clear physical underpinnings. However, their applicability to complex systems is often limited by the difficulty of accurately modeling such systems [8]. Data-driven approaches utilize historical operational data, such as sensor data, fault data, etc., to train models and learn device degradation patterns [9]. These approaches eliminate the need for intricate physical modeling and exhibit strong adaptability, but they typically need a large amount of high-quality data and considerable computational resources. Hybrid approaches aim to integrate the transparency of model-based techniques with the adaptability and learning strength of data-driven models. While they offer a promising balance, they are often more complex to design and require careful parameter tuning.
Driven by ongoing advancements in industrial big data and technological innovation, the volume of historical data generated by machinery and equipment has grown substantially. This progress has significantly accelerated the development of data-driven approaches and has attracted growing interest from researchers [10]. A variety of deep learning architectures—including convolutional neural networks (CNNs) [11,12], recurrent neural networks (RNNs) [13,14,15], graph neural networks (GNNs) [16,17], and Transformers [18,19] have been extensively utilized in the field of RUL prediction. These models offer powerful nonlinear mapping capabilities and strong representational learning capacity, providing distinct advantages over traditional model-based approaches.
However, individual models, such as CNNs, RNNs, and GNNs, typically focus on extracting a single type of feature and often fail to simultaneously model both temporal dependencies and inter-sensor correlations [13,20,21]. To overcome this limitation, researchers have designed various model variants that are capable of extracting features across both temporal and spatial dimensions. Nevertheless, despite these advancements, the effectiveness of jointly extracting features from both domains remains constrained [22,23].
The introduction of attention mechanisms (AMs) led researchers to integrate them into CNN and RNN architectures, resulting in moderate improvements in model performance [24,25]. However, these enhancements did not fully address the inherent limitations of CNNs and RNNs. Transformer models have brought notable advancements and have shown strong performance in predicting RUL [26,27]. Unlike CNNs, Transformer architectures substantially enhance the utility of attention mechanisms by enabling more effective global feature extraction [28]. The ability of Transformers to capture extended temporal dependencies makes them well-suited for time-series-based RUL estimation.
Despite their advantages, Transformer models also have limitations. Although they are effective at extracting global features, they often struggle to capture local patterns [29]. This poses a challenge in RUL prediction, where equipment degradation frequently manifests as subtle, short-term changes between adjacent time steps—patterns that Transformers may overlook [29,30]. Integrating customized attention mechanisms allows the model to prioritize adjacent time steps and lessen the impact of more distant ones, enhancing its sensitivity to localized degradation patterns.
AMs have been widely adopted across various domains, including machine translation, time series forecasting, and image recognition [31]. As a typical multidimensional time series regression task, RUL prediction has also benefited from the integration of AMs. Studies utilizing attention for RUL estimation from temporal and feature viewpoints typically fall into two major categories.
The first method focuses on extracting global features. In this strategy, CNNs are employed to capture global representations, followed by attention mechanisms that highlight important features while suppressing less relevant ones [32]. However, this method inherits the limitations of CNNs: although local features are well captured, the method does not sufficiently represent the device’s overall degradation characteristics.
The second approach emphasizes temporal dynamics. Models such as Transformers [33] utilize attention mechanisms to identify subtle variations between time points and learn degradation patterns over time. While effective in modeling temporal dependencies, the standard Transformer architecture does not inherently capture inter-sensor relationships. It tends to focus on the temporal behavior of individual sensors, whereas in RUL prediction, capturing the interactions and correlations among multiple sensors is equally crucial.
Additionally, hybrid models have been widely explored in recent studies. For example, Xiang et al. [34] proposed a single-gated RNN with a differential weighted information storage mechanism, Xiang et al. [35] developed a dynamic self-learning neural network, and Li et al. [36] introduced a channel-independent bidirectionally gated Mamba with an interactive recurrent mechanism. All of these approaches achieved outstanding results in their respective domains. Such models preserve the strengths of their baseline architectures while providing additional advantages. In contrast, the model proposed in this paper integrates two baseline models in series rather than in a hierarchical configuration.
Based on earlier findings, this work presents a dual-attention-based hybrid framework for RUL prediction. The model employs a multi-head full attention mechanism to learn temporal dependencies from sensor data over time, along with an improved channel attention module to assign importance to individual sensors. By combining both attention mechanisms, the model effectively learns temporal patterns as well as inter-sensor relationships. Despite the model’s simple structural design, it achieves high prediction accuracy when evaluated using NASA’s C-MAPSS dataset.
This study offers the following key contributions:
(1)
To overcome the limitations of traditional models in feature extraction, the proposed approach introduces distinct attention mechanisms to separately capture temporal and sensor-specific features, thereby enhancing the richness of the learned representations.
(2)
For temporal feature extraction, a multi-head full attention mechanism is employed. Specifically, an inverted module from the iTransformer architecture is adopted to allow the model to focus on the temporal behavior of individual sensor sequences while disregarding inter-sensor interference.
(3)
For sensor feature extraction, a channel attention mechanism is utilized to learn sensor-specific weights. This study is, to our knowledge, the earliest to implement a channel attention strategy tailored for sensor-wise feature learning in the context of RUL estimation.
The paper will proceed as follows: Section 2 elaborates on the hybrid model framework; Section 3 presents the experiments carried out on the C-MAPSS dataset; and Section 4 concludes the paper with key observations.

2. Proposed Methodology

This section outlines the architecture and design of the proposed Hybrid Model with Dual-Attention Mechanism (HMDAM), developed for predicting RUL from multi-sensor time-series data. As shown in Figure 1, the architecture includes two core components: temporal attention and sensor-specific feature attention mechanisms.
The Temporal Attention Mechanism (TAM) is constructed to uncover hidden temporal relationships within the sensor data. It operates independently on each sensor’s time series, focusing exclusively on intra-sensor temporal patterns without incorporating inter-sensor interactions. This module allows the model to track and learn degradation patterns as they develop over time.
To enhance feature relevance, the feature attention mechanism is used to process the temporal attention block’s output. It captures distinct characteristics from each sensor, allowing the model to assess their individual contributions to system degradation. With a lightweight and efficient structure, the feature attention module integrates sensor-level information while preserving the temporal context learned in the preceding stage.
Overall, the HMDAM consists of three integral parts: a temporal feature extraction block (TFEB), a sensor feature extraction block (SFEB), and a regression component for RUL prediction. The TFEB is designed to capture sequential degradation patterns over time, while the SFEB focuses on identifying sensor-specific contributions. The combined operation of these two modules results in a robust and well-rounded feature representation. The regression module then predicts RUL based on the integrated features. The full structural design of the model is depicted in Figure 2. A comprehensive overview of the model’s architecture and its constituent modules is presented in the remaining part of this section.

2.1. Time Feature Extraction Block

The TFEB consists of three main components: an embedding layer, the TAM, and a projection layer. The primary objective is to transform raw multi-sensor time-series data into meaningful temporal representations that capture the underlying degradation behavior of mechanical systems.
This module is tailored to detect nuanced variations over time, allowing the model to better grasp patterns associated with system degradation. The process begins with the embedding layer, which performs a dimensional transformation by expanding the time dimension of the input data. This transformation facilitates the extraction of fine-grained temporal features in subsequent stages, while deliberately ignoring inter-sensor correlations at this point to ensure a focus on individual temporal dynamics.
Subsequently, a TAM is employed on the expanded time sequences, allowing the model to focus on the most relevant temporal segments for degradation assessment. This enhances its responsiveness to temporal fluctuations and strengthens its capacity to identify early degradation indicators.
Finally, a projection layer is applied to return the extracted temporal features to their initial dimensional space. This operation maintains consistency in feature representation prior to forwarding the output to the subsequent module in the model.
The architecture and functionality of each individual subcomponent are elaborated in the subsections that follow.

2.1.1. Embedding Layer

The embedding layer performs two essential functions. First, it transposes the input sensor data to facilitate independent time-series feature extraction for each sensor. Second, it increases the dimensionality of the time series, projecting the original temporal input into a higher-dimensional space to enable the extraction of more nuanced temporal patterns. This transposition operation, originally introduced in the iTransformer model, has been shown to significantly enhance the effectiveness of temporal feature extraction [37].
Given the original sensor input data X = { x 1 , , , , , x t } R T × N , where   T is the number of time steps and N is the number of sensor channels, the embedding layer first transposes the input to obtain X T R N × T . A linear transformation is then applied to project the temporal dimension T into a higher-dimensional hidden space of size d m o d e l . This process is defined as
H = L i n e a r ( X T ) R N × d m o d e l
In Equation (1), H = { h 1 , , , , , h N } R N × d m o d e l , where N is the number of sensors and d m o d e l is the dimensionality of the hidden space. L i n e a r denotes a fully connected layer that performs a linear mapping of the transposed time sequences into the hidden representation space.

2.1.2. Temporal Attention Mechanism

The TAM captures time-dependent patterns in the sensor data and converts them into rich, continuous feature representations. TAM consists of a multi-head full attention layer, two-layer normalization modules, and a feed-forward neural network (FFN). It processes the output H from the embedding layer and enhances it by learning temporal correlations within each individual sensor’s time series.
TAM takes the input matrix H R N × d m o d e l and passes it through a multi-head full attention layer to capture temporal relationships. Specifically, H is linearly projected into query ( Q ), key ( K ), and value ( V ) matrices as follows:
Q = H W Q
K = H W K
V = H W V
In these equations, Q ,   K R N × d k and V R N × d v , where d k and d v represent the dimensions of the projected key/query and value vectors, respectively. Typically, d k = d v . The matrices W Q ,   W K R d m o d e l × d k , and W V R d m o d e l × d v are learnable parameters.
To improve training stability and speed up convergence, layer normalization is used after the attention step. The FFN is then used to further refine the extracted temporal features, followed by an additional normalization step to maintain consistency in the feature distribution.
Next, the query, key and value matrices are used to compute the scaled dot-product attention. This operation captures time dependency by calculating the attention weights of each sensor at different times. The attention weights are calculated using the following formula:
A t t e n t i o n Q , K , V = s o f t m a x ( Q K T d k ) V
In Equation (5), the resulting matrix A t t e n t i o n R N × d v represents the learned temporal attention features. The scaling factor d k is introduced to keep the dot product values within a manageable range, helping to maintain gradient stability during training.
Since TAM employs a multi-head full attention mechanism, it can capture multiple types of temporal dependencies within the time series. Every attention head applies distinct trainable weight matrices to transform the query, key, and value inputs individually. As a result, each head computes a distinct representation of temporal relationships, Allowing the model to capture different facets of the degradation behavior. Each sub-attention module produces an output H e a d n R N × d v , where n denotes the index of the attention head.
The combined outputs from all attention heads are transformed via a linear network layer to generate the unified representation of the multi-head attention mechanism. The computations are defined as follows:
H e a d n = A t t e n t i o n ( Q W Q n ,   K W K n , V W V n )
M u l t i H e a d Q , K , V = C o n c a t ( H e a d 1 , H e a d n ) W O
In these expressions, W Q n ,   W K n R d m o d e l × d k , and W V n R d m o d e l × d v represent the trainable projection matrices associated with the n -th attention head. The output projection matrix W O R ( n d v ) × d m o d e l is the output projection matrix used to map the concatenated multi-head outputs back into the original feature space.
By attending to multiple representational subspaces across time, the model gains improved awareness of temporal dynamics and generates more expressive feature representations.
After the multi-head attention step, the resulting output undergoes layer normalization to promote training stability and accelerate model convergence. This process is represented as
Y = L a y e r N o r m ( M u l t i H e a d )
In Equation (8), Y R N × d m o d e l r denotes the normalized output and M u l t i H e a d R N × d m o d e l denotes the result produced by the multi-head attention module.
Next, the normalized output Y is then input into a position-wise FFN, enabling nonlinear mapping to strengthen the model’s feature extraction capabilities. This component enhances the model’s capacity to identify intricate degradation behaviors. The FFN comprises two fully connected layers and is defined as
Y 1 = L i n e a r ( Y ) R N × d f f
Y 2 = L i n e a r ( Y 1 ) R N × d m o d e l
In these expressions, Y 1 R N × d f f is the intermediate representation and Y 2 R N × d m o d e l represents the ultimate result produced by the feed-forward network. The parameter d f f is the hidden dimensionality of the FFN, typically set larger than d m o d e l to allow for richer representations.
Finally, the output Y 2 is combined with the original input Y via a residual pathway, and the result is subsequently normalized through an additional layer normalization process. This operation preserves the original features while enhancing training stability:
Y 3 = L a y e r N o r m ( Y 2 + Y )
In Equation (11), Y 3 R N × d m o d e l represents the output produced by the TAM module, which represents the extracted temporal features of the sensor data.

2.1.3. Projection Layer

The projection layer serves to transform the temporal features obtained from the TAM into the same dimensional space as the original input. This is achieved through a linear transformation followed by a transposition operation. The transformation ensures that the learned temporal representations can be seamlessly integrated with subsequent components of the model. The projection operation can be expressed as
Y T T = L i n e a r ( Y 3 ) R N × T
Y T = ( Y T T ) T R T × N
In these expressions, Y T R T × N denotes the concluding output of the TFEB. It retains the same shape as the original input sensor data while embedding rich temporal information extracted through the dual-attention processing. This output serves as the input to the subsequent SFEB.

2.2. Sensor Feature Extraction Block

The SFEB is designed to enhance the representation of sensor-specific information by integrating it with the temporal features obtained from the TFEB. This module is derived from the SENet model proposed by Hu et al. [38] in 2018. SENet assigns weight coefficients to each channel through lightweight feature extraction. In this work, the SENet-based module is employed to compute weight coefficients for individual sensors. The SFEB consists of multiple layers, incorporating pooling layers, dense (fully connected) layers, and nonlinear activation functions. Its primary objective is to ensure that the final feature representation retains both temporal dynamics and sensor-specific characteristics, which significantly contribute to the reliability of RUL forecasting.
The process begins with a permutation operation applied to the output of the TFEB, denoted as Y T , in order to reorganize the data dimensions. Specifically, the single-channel temporal output is reshaped into N distinct channels—each corresponding to a different sensor. This transformation prepares the data for channel-wise processing in the subsequent layers, enabling the model to learn sensor-dependent degradation features more effectively.
Next, the squeeze-and-excitation (SE) operations are applied to each sensor channel to capture sensor-specific importance. This mechanism consists of two main steps.
(1)
Squeeze: The temporal data in each sensor channel is compressed into a global feature vector, summarizing the overall temporal behavior of each sensor.
(2)
Excitation: A nonlinear transformation is applied to the global feature vector, and attention weights are generated using a sigmoid activation function. These weights are then applied to adjust the original input, thereby enhancing the contribution of important sensor features and suppressing less informative ones.
The process begins with a permutation operation on the output of the TFEB, which rearranges the data for channel-wise processing:
Y 4 = P e r m u t e ( Y T ) R N × T × 1
In this expression, Y 4 R N × T × 1 represents the permuted data, where each sensor is now treated as an independent channel, with its temporal data aligned along the second dimension.
Next, a squeeze operation is applied to each channel to obtain a compact global representation. This is accomplished using global average pooling, which computes the mean value of each sensor’s time series, effectively summarizing its overall contribution:
Y 5 = F s q ( Y 4 ) R N × 1 × 1
In Equation (15), Y 5 R N × 1 × 1 is the output of the squeeze operation. Each element in Y 5 represents the average temporal value for one sensor, forming a compact representation of that sensor’s overall behavior across time.
Following the squeeze operation, an excitation process is applied to Y 5 to model the interdependencies among different sensor channels. This step allows the network to adaptively highlight informative sensor features, enhancing the expressiveness of the extracted features. The excitation process consists of three main components: a full connection layer, a ReLU activation function, and another fully connected layer. The computations are as follows:
Y 6 = L i n e a r ( Y 5 ) R N r × 1 × 1
Y 7 = R e L U ( Y 6 ) R N r × 1 × 1
Y 8 = L i n e a r ( Y 7 ) R N × 1 × 1
In these expressions, r denotes a reduction ratio that controls the dimensionality of the intermediate layer, effectively reducing computational complexity while retaining representational capacity. The resulting excitation vector Y 8 R N × 1 × 1 contains learned weights that indicate the weight or priority given to different sensor channels.
To obtain the final sensor-wise attention weights, a sigmoid activation function is used to the excitation output Y 8 , producing normalized weight representations for each sensor channel. These weights are then used to rescale the input temporal features Y 4 through element-wise multiplication, effectively integrating both temporal and sensor-specific characteristics. A final permutation operation restores the data to its original dimensional format. The process is defined by the following equations:
Y 9 = S i g m o i d ( Y 8 ) R N × 1 × 1
Y 10 = S c a l e ( Y 9 Y 4 ) R N × T × 1
Y O = P e r m u t e ( Y 10 ) R 1 × T × N
In these expressions, denotes element-wise multiplication; Y 9 contains the learned sensor attention weights; Y 10 represents the weighted combination of temporal and sensor features; and Y O R 1 × T × N is the final output of the SFEB, encoding a fused representation of temporal and sensor-specific characteristics in a format consistent with the original input dimensions.

2.3. Regressor

The regressor is responsible for mapping the global features extracted by the TFEB and the SFEB to the RUL prediction. The projection process within the regressor consists of two main steps.
First, the two-dimensional output vector Y O R 1 × T × N is reshaped into a one-dimensional vector Y O 1 R ( T N ) × 1 . This reshaping operation allows the subsequent layers to process the global features as a flattened vector.
Next, Y O 1 is subsequently processed through three fully connected layers that progressively reduce dimensionality while improving the model’s capacity to learn robust, generalized feature representations. These layers help capture complex nonlinear relationships in the fused feature space, ultimately enhancing both the precision and reliability of the RUL estimation.
The corresponding mathematical operations are defined as follows:
Y O 1 = R e s h a p e ( Y O ) R ( T N ) × 1
Y O 2 = L i n e a r ( Y O 1 ) R n 1 × 1
Y O 3 = L i n e a r ( Y O 2 ) R n 2 × 1
R U L = L i n e a r ( Y O 3 ) R 1 × 1
In these expressions, Y O 1 , Y O 2 , and Y O 3 are intermediate representations; { n 1 , n 2 } are the hidden dimensions used during the dimensionality reduction process; and R U L R 1 × 1 represents the final scalar prediction of the RUL.
This composition enables the model to capture a rich mapping from the fused temporal–sensor feature space to the target output, enhancing prediction reliability and the model’s robustness to varying operational patterns.

3. Experimental Study

This section presents a detailed evaluation of the developed hybrid model. First, the dataset used in the experiments is introduced, highlighting its characteristics and relevance to the RUL prediction task. Next, the experimental setup, including data preprocessing procedures, model configuration, and evaluation metrics, is described in detail. Finally, we discuss the ablation analysis and comparison of the model against leading methods to confirm its strength in terms of effectiveness, resilience, and adaptability.

3.1. Dataset

This study utilized the C-MAPSS (Commercial Modular Aero-Propulsion System Simulation) dataset, a benchmark dataset released by NASA [39]. As shown in Table 1, C-MAPSS is segmented into four sub-datasets—FD001 to FD004—distinguished by varying operational contexts and failure modes. Specifically, FD001 and FD003 each operate under a single condition, with FD001 containing one fault mode and FD003 involving multiple fault types. In contrast, FD002 and FD004 are more complex, featuring several operating conditions and up to six distinct fault modes, which increases the difficulty of accurate RUL prediction.
Each dataset subset includes both a training and testing partitions. The training portion consists of full engine life-cycle data, capturing the complete progression from normal functioning to eventual failure. All engines are equipped with 21 sensors, which continuously monitor various operational parameters such as temperature, pressure, and fan speed.
In this study, the test set was constructed by randomly selecting partial life-cycle sequences from each engine in the training data. These truncated sequences simulate real-world scenarios where the engine has not yet failed, thereby allowing the model to predict the RUL. This configuration allows for a practical assessment of the model’s prediction capabilities across different operating scenarios and patterns of degradation.

3.2. Experimental Setting

3.2.1. Data Preprocessing

Prior to experimentation, the raw C-MAPSS dataset requires preprocessing and conversion into a structured format that is appropriate for model training and assessment. The preprocessing pipeline consists of four primary steps: data filtering, data normalization, sliding window configuration, and sample-label construction. The overall preprocessing workflow is illustrated in Figure 3.
(1)
Data filtering
The C-MAPSS dataset includes sensor readings from 21 different sensors installed on each engine. These sensors continuously monitor a range of operational and environmental conditions throughout the engine’s lifecycle. However, further analysis indicated that certain sensor readings showed minimal change or remained nearly constant throughout an engine’s operational lifespan. These constant or near-constant signals typically do not contribute meaningful information about the degradation process and may introduce redundancy or noise into the model.
To ensure more informative and high-quality input data, these uninformative sensors were removed. Using the FD001 subset as a representative case, Figure 4 shows the time-series trends of all 21 sensors. It is evident that sensors 1, 5, 6, 10, 16, 18, and 19 displayed constant or near-constant behavior across all operational cycles. Consequently, these sensors were removed from the feature set utilized in later stages of model development. These seven sensors measured the following physical quantities: total inlet temperature, inlet pressure, total duct pressure, engine pressure ratio, burner air-fuel ratio, set fan speed, and set core engine equivalent speed.
This filtering step ensures that the model focuses only on sensors that carry diagnostic information relevant to the RUL prediction, thereby reducing input dimensionality and potential overfitting.
(2)
Data normalization
The sensor readings in the C-MAPSS dataset exhibit considerable variation in both scale and fluctuation range across different sensors. Directly using these raw, unnormalized values can obscure subtle but important patterns—especially for sensors whose numerical values fall within narrow ranges. Such disparities can bias the model toward features with larger magnitudes and reduce its ability to learn meaningful degradation patterns from less prominent signals.
To ensure consistency and eliminate scale-related bias, each sensor’s data was scaled using Min–Max normalization. This approach transforms all sensor readings into a common scale within the range [0, 1], thereby preserving the relative variation patterns while standardizing feature magnitudes.
The normalization method was mathematically formulated and is presented in Equation (26):
x n o r m i , j = x i , l x m i n j x m a x j x m i n j [ 0 ,   1 ]
In this expression, x i , l represents the data value recorded by the j -th sensor at the i -th time step; x n o r m i , j indicates the normalized counterpart of the original value; and x m a x j and x m i n j correspond to the highest and lowest recorded values of the j -th sensor, respectively.
Aligning feature scales through this normalization enhances both the stability and efficiency of the training process.
(3)
Sliding window settings
To boost the training sample size and better learn short-term temporal patterns, the original time series is divided into overlapping fixed-length segments using a sliding window method. This method is particularly effective in modeling the gradual evolution of system degradation over time.
In this method, a fixed window size is defined, and the window slides across each time series with a stride of 1. At each step, a window of sensor readings is extracted, resulting in a large set of overlapping samples. This dense sampling strategy ensures that subtle degradation patterns are captured from various temporal contexts, improving the model’s ability to learn robust short-term and transitional features.
The benefits of this approach are twofold: it augments the training set by generating multiple samples from each time series, thus mitigating the issue of data scarcity; in addition, this approach maintains the sequential order of data within each segment, enabling the model to capture short-term degradation patterns effectively.
Once the segmentation is complete, all windowed sequences are aggregated to construct the final training dataset.
(4)
Training set and label creation
To generate the RUL labels, the original values are replaced using a piecewise linear strategy that better reflects real-world degradation patterns. As illustrated in Figure 5a, the RUL of an engine naturally declines over time throughout its operational life. However, instead of modeling this decline as a strictly linear decay from the initial time step, a segmented degradation rule is applied.
In particular, the RUL is assumed to remain constant during the early stage of engine operation and begins to decrease linearly only after reaching a predefined time threshold, as depicted in Figure 5b. This approach reflects the practical observation that engines typically operate without significant degradation during the initial phase of their life cycle. In this study, the threshold was set to 125 cycles, consistent with prior work [40].
After preprocessing, the FD001 dataset is transformed into multiple sets of n × t matrices paired with 1 × 1 labels, where n denotes the window length and t represents the number of selected sensors. These sensors primarily monitor physical quantities such as temperature, pressure, rotational speed, and airflow rate. Finally, the training set is further split for validation purposes. The training data is split by randomly allocating 80% for model training and reserving the remaining 20% for validation. The test set stays fixed to allow an unbiased comparison of model results.

3.2.2. Evaluation Metrics

To evaluate the performance of the proposed approach, three widely used metrics were employed: Root Mean Square Error (RMSE), Mean Absolute Error (MAE), and the Score function. RMSE is a standard metric in regression analysis that quantifies the average magnitude of the prediction errors. MAE is a statistical measure of the average difference between predicted and actual values. They are defined as
R M S E = 1 N i = 1 N ( Y i ^ Y i ) 2
M A E = 1 N i = 1 N Y i ^ Y i
where Y i is the true RUL of the i -th test sample, Y i ^ is the corresponding predicted RUL value, and N represents the total number of samples in the test set.
Accurate and timely RUL predictions are crucial in industrial contexts, as delays can compromise safety and disrupt processes. Therefore, the Score function is designed to be more sensitive to late predictions, making it more aligned with the needs of industrial production environments [41]. The Score function is computed using the following formula:
S c o r e = i = 1 n e Y i ^ Y i 13 1 ,   f o r   Y i ^ < Y i i = 1 n e Y i ^ Y i 10 1 ,   f o r   Y i ^ > Y i
In Equation (29), Y i ^ and Y i denote the estimated and actual RULs, respectively, and n indicates the total number of samples used for testing. In engine degradation scenarios, early predictions are generally more valuable than late ones. Accordingly, the asymmetric structure defined by parameter settings “10” and “13” assigns a higher penalty to late predictions [39].

3.2.3. Training Parameter Settings

To maximize model effectiveness, a comprehensive grid search is conducted to refine the training and structural parameters. This exhaustive search helps identify the best configuration for maximizing prediction accuracy and model stability. The final training hyperparameter configurations are listed in Table 2 with the corresponding structural parameters provided in Table 3.

3.3. Analysis of HMDAM

3.3.1. Setting of Sliding Window Size

The window length controls the extent of historical information available to the model along the temporal dimension. If the window is too short, the model cannot capture long-term degradation patterns, resulting in performance loss. Conversely, an overly long window may introduce early-stage noise unrelated to degradation and increase information complexity, which also degrades performance. Therefore, selecting an appropriate window length is critical for achieving reliable predictions. Given the heterogeneity in the C-MAPSS sub-datasets, which was characterized by distinct fault patterns and operating environments, a tailored analysis of sliding window lengths was needed to optimize the HMDAM performance for each dataset.
This study investigated the impact of sliding window sizes ranging from 20 to 50, in increments of 2, with all other parameters held constant during the experiments. Model performance was evaluated using two metrics: RMSE and the S-score. To mitigate the influence of randomness during training, each configuration was run five times for 100 epochs, and the best performance among the five runs was reported as the final result.
As shown in Figure 6, subfigures (a), (b), and (d) follow the regression line, whereas subfigure (c) exhibits pronounced fluctuations. This behavior may stem from the fact that FD003 contains multiple failure modes, leading to more complex degradation dynamics. In addition, FD003 has fewer samples than FD002 and FD004, which may increase the statistical variance during hyperparameter tuning and, in turn, cause the fitted curve to fluctuate more strongly. As depicted in Figure 6, for the FD001 dataset, both the RMSE and Score reached their lowest values when the sliding window length was 30, making it the optimal choice. In contrast, for FD002, FD003, and FD004, the points with the minimum RMSE and Score values did not coincide. In such cases, the window size associated with the lowest RMSE was selected to ensure better predictive accuracy. Accordingly, the optimal window lengths determined for FD002, FD003, and FD004 were 26, 46, and 36, respectively. To summarize, the best-performing sliding window lengths for FD001 through FD004 were 30, 26, 46, and 36.

3.3.2. Model Parameter Settings

To identify the optimal hyperparameters for the proposed model, a combination of control variable methodology and grid search was employed. Specifically, four key hyperparameters were selected, and four candidate values were assigned to each hyperparameter for the experimental evaluation. The range of values considered for each hyperparameter is presented in Table 4. Model performance was evaluated using two metrics: RMSE and S-score.
As illustrated in Figure 7, Figure 8, Figure 9 and Figure 10, the optimal hyperparameters for the model varied between the different sub-datasets, reflecting the differences in operational conditions and fault modes. The RMSE and Score continued to serve as the primary evaluation metrics. In cases where the RMSE and Score metrics yielded conflicting results, the configuration with the best RMSE value was selected as the optimal setting given its stronger correlation with predictive accuracy.
For the FD001 sub-dataset, the optimal hyperparameters were as follows: three encoding layers, four attention heads, a batch size of 128, and a model dimension of 32. The optimal hyperparameter configurations for the remaining sub-datasets (FD002, FD003, and FD004) are summarized in Table 5.

3.3.3. HMDAM Ablation Experiment

An ablation analysis was conducted to assess the contribution of each individual component in the HMDAM architecture. Three model variants were developed for comparison:
(1)
The baseline Transformer model without any enhancements;
(2)
The Transformer model equipped with only the TFEB, referred to as iTransformer;
(3)
The Transformer model integrated solely with the SFEB, referred to as ST.
Each model was tested under the same experimental setup, with their performance measured using RMSE and the Score function. Following model training, the results—summarized in Table 6—demonstrate the effectiveness of the TFEB and SFEB modules and the overall improvement provided by their integration into the HMDAM.
As shown in Table 6, the Transformer model integrated with the TFEB module (iTransformer) consistently outperformed the baseline Transformer across all four sub-datasets. Similarly, the Transformer model enhanced with the SFEB module also demonstrated improved performance compared to the original Transformer model. Notably, the proposed HMDAM model, which combines both the TFEB and SFEB modules, achieved the best overall performance. The findings demonstrate that both the TFEB and SFEB modules play a vital role in improving RUL prediction accuracy, and their integration leads to complementary enhancements in model performance.

3.3.4. Analysis of Model Results

After determining the optimal hyperparameters, as outlined in the previous section, the best-performing model configuration was applied to each sub-dataset to predict the RUL using the respective test sets. Figure 11 presents the predicted results, demonstrating a close match between the estimated and actual RUL values for all four sub-datasets. The strong alignment between the predicted and actual values suggests that the HMDAM model delivers high predictive accuracy and demonstrates strong generalization capabilities in estimating the RUL of aero-engine systems.
To further verify the efficacy of the model, several engine units were randomly sampled from each of the four sub-datasets for visual analysis. As shown in Figure 12, the RUL estimates generated by the HMDAM model align more closely with the actual RUL curves than those from the other three baseline models, underscoring its enhanced predictive performance.
Notably, the model maintained strong predictive accuracy during both the initial and final phases of engine degradation. However, some limitations were observed during the mid-stage of degradation, particularly in the FD003 dataset. As shown in Figure 12c, the actual RUL began to decline during this transitional phase, while the model’s predictions lagged behind. This discrepancy is attributed to the complex nature of the mid-stage, which contains overlapping signals from both healthy and deteriorating conditions. The subtle degradation indicators during this period pose a greater challenge for accurate modeling, leading to a temporary reduction in predictive performance.

3.3.5. Model Efficiency Analysis

Using the FD001 dataset as an example, this study compared the four models in terms of FLOPs and parameter counts, with the results summarized in Table 7. the number of FLOPs of the proposed model was 30% lower than that of the ST model and only 1.36% higher than that of the Transformer model. By contrast, the iTransformer achieved markedly lower number of FLOPs, representing just 53% of that of the HMDAM. In terms of parameter counts, the difference between the Transformer and HMDAM was only 2%, while iTransformer’s parameter count was less than 1% higher. Conversely, the ST model required 12% more parameters. These results indicate that the proposed model improves prediction accuracy without substantially increasing the computational overhead, thereby offering a practical solution for real-world RUL prediction tasks under cost constraints.

3.3.6. Comparison with Other Methods

This section presents a performance comparison between the HMDAM model and several state-of-the-art approaches in this domain. The performance outcomes are detailed in Table 8 and Table 9, where the highest scores for each metric are emphasized in bold. These comparisons are intended to highlight the HMDAM model’s effectiveness and competitive edge in predicting RUL under varying operational scenarios.
As shown in Table 7 and Table 8, the Transformer model outperformed the CNN model, while the BGT model surpassed the DCNN model in all metrics. This can be attributed to the fact that CNNs are well suited for extracting local features but have limited ability to capture global dependencies. In contrast, Transformers can effectively model both local and global features when sufficient data and network depths are available [42].
On the FD001 sub-dataset, the HMDAM delivered superior performance, surpassing all the baseline models in terms of both RMSE and Score. Specifically, it surpassed the CTNet model [43], with improvements of approximately 7% in RMSE and 9% in Score, demonstrating its effectiveness in scenarios with simpler operating conditions and failure modes.
Although the HMDAM did not achieve the top rank for all four sub-datasets, it consistently performed at a high level. For the more complex FD002, FD003, and FD004 sub-datasets, the HMDAM model consistently secured second place in the performance rankings, closely following the BGT model [44] and CTNet. These results affirm that the HMDAM maintains strong predictive performance across varying levels of dataset complexity, confirming its robustness and generalizability.
Regarding the Score metric, although the HMDAM model performed well on the FD001 dataset, it lagged behind the BGT and CTNet models on the other datasets. This suggests that the HMDAM has limitations in mitigating RUL overestimation. The CTNet model combines the multi-head self-attention mechanism of Transformers with the cross-channel information propagation of graph neural networks, which increases its complexity [43]. Such complexity enhances its ability to fit complex data, leading to superior performance. By contrast, the HMDAM does not markedly increase FLOPs or the parameter count relative to Transformer models, reflecting its relatively low complexity. This may account for its weaker performance in the Score metric compared with CTNet.
Table 8. RMSE comparison between HMDAM and other advanced methods.
Table 8. RMSE comparison between HMDAM and other advanced methods.
MethodFD001FD002FD003FD004Average
BiLSTM ([45])13.6523.1813.7424.8618.86
DCNN ([46])12.6122.3612.6423.3117.73
CatBoos ([47])15.821.416.022.418.90
CDLSTM ([48])13.9917.5312.1520.9116.15
HMC ([49])13.8420.7414.4122.7317.93
BiGRU-AS ([50])13.6820.8115.5327.3119.33
DSAN ([30])13.422.0615.1221.0317.90
DAA ([51])12.2517.0813.3919.8615.65
IMDSSN ([52])12.1417.4012.3519.7815.42
BGT ([44])12.0911.4610.1613.8911.9
CTNet ([43])11.6413.6711.2814.6212.80
HMDAM10.8215.3311.2117.4813.71
Table 9. Score comparison between HMDAM and other advanced methods.
Table 9. Score comparison between HMDAM and other advanced methods.
MethodFD001FD002FD003FD004Average
BiLSTM ([45])295413031754302543
DCNN ([46])273.71041.2284.112,4665858.9
CatBoos ([47])398.73493.2584.23203.41919.9
CDLSTM ([48])320175822126331233
HMC ([49])42719,400297710,3748295
BiGRU-AS ([50])284245442847081968.5
DSAN ([30]336194625136711571.3
DAA ([51])19815752901741951
IMDSSN ([52])206.111775.15229.542852.811265.9
BGT ([44])262.67550.52196.94963.36493.37
CTNet ([43])187809187844506.75
HMDAM170.071030.42239.471738.19794.53

4. Conclusions and Future Research

This study proposes the HMDAM model, which extracts temporal and sensor-level features and integrates them to estimate the RUL of aircraft engines. The architecture consists of two primary modules: the TFEB, which models the evolution of time-dependent features and uncovers hidden patterns in different time scales, and the SFEB, which enhances sensor-level representation by learning channel-wise attention weights. To obtain the final RUL output, the fused features are processed through a regression block composed of dense layers and a Sigmoid function.
To assess the proposed model’s efficacy, detailed experiments were executed. A thorough investigation of sliding window configurations and model hyperparameters across the C-MAPSS sub-datasets ensured optimal performance. Ablation studies confirmed the contribution of each model component, while comparative evaluations against several state-of-the-art methods demonstrated that the HMDAM consistently achieved superior or competitive performance—particularly on simpler datasets such as FD001. In the efficiency analysis, the proposed model showed improved prediction accuracy without a substantially increased computational cost. This makes it a practical solution for RUL prediction tasks under limited computational resources.
Despite its promising results, the HMDAM model has certain limitations. Since the computational resources required by the proposed model do not markedly exceed those of the baseline, its computational accuracy may be suboptimal in complex scenarios. In addition, the model shows a tendency to overestimate RUL. For example, on the FD004 dataset, it underperformed compared with other advanced models. Furthermore, it lacks transferability, as models trained on FD001 cannot be directly applied to other sub-datasets.
In future work, we plan to enhance the robustness and generalization capability of the HMDAM by incorporating multi-scale feature fusion techniques to better capture degradation trends across diverse and complex engine states, thus, enhancing the model’s relevance and utility in practical prognostic applications.

Author Contributions

Writing—original draft: C.H.; Validation: C.H.; Software: C.H.; Conceptualization: Z.L.; Methodology: Z.L.; Supervision: Z.L.; Writing—review & editing: C.Z.; Writing—review & editing: Z.Z.; Writing—review & editing: L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This project is supported by the National Natural Science Foundation of China under grant 62173260, the National Natural Science Foundation of China under grant 62303358, and the Hubei Provincial Natural Science Foundation of China under grant 2024AFB641.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lei, Y.G.; Li, N.P.; Guo, L.; Li, N.B.; Yan, T.; Lin, J. Machinery health prognostics: A systematic review from data acquisition to RUL prediction. Mech. Syst. Signal Proc. 2018, 104, 799–834. [Google Scholar] [CrossRef]
  2. Hu, Y.; Miao, X.W.; Si, Y.; Pan, E.S.; Zio, E. Prognostics and health management: A review from the perspectives of design, development and decision. Reliab. Eng. Syst. Saf. 2022, 217, 108063. [Google Scholar] [CrossRef]
  3. Xia, M.; Li, T.; Shu, T.X.; Wan, J.F.; de Silva, C.W.; Wang, Z.R. A Two-Stage Approach for the Remaining Useful Life Prediction of Bearings Using Deep Neural Networks. IEEE Trans. Ind. Inform. 2019, 15, 3703–3711. [Google Scholar] [CrossRef]
  4. She, D.M.; Jia, M.P. A BiGRU method for remaining useful life prediction of machinery. Measurement 2021, 167, 108277. [Google Scholar] [CrossRef]
  5. Zhou, J.H.; Qin, Y.; Luo, J.; Wang, S.L.; Zhu, T. Dual-Thread Gated Recurrent Unit for Gear Remaining Useful Life Prediction. IEEE Trans. Ind. Inform. 2023, 19, 8307–8318. [Google Scholar] [CrossRef]
  6. Pan, Y.; Kang, S.J.; Kong, L.G.; Wu, J.J.; Yang, Y.H.; Zuo, H.F. Remaining useful life prediction methods of equipment components based on deep learning for sustainable manufacturing: A literature review. AI EDAM-Artif. Intell. Eng. Des. Anal. Manuf. 2025, 39, e4. [Google Scholar] [CrossRef]
  7. Song, L.Y.; Lin, T.J.; Jin, Y.; Zhao, S.K.; Li, Y.; Wang, H.Q. Advancements in bearing remaining useful life prediction methods: A comprehensive review. Meas. Sci. Technol. 2024, 35, 092003. [Google Scholar] [CrossRef]
  8. Ge, M.F.; Liu, Y.B.; Jiang, X.X.; Liu, J. A review on state of health estimations and remaining useful life prognostics of lithium-ion batteries. Measurement 2021, 174, 109057. [Google Scholar] [CrossRef]
  9. Ferreira, C.; Goncalves, G. Remaining Useful Life prediction and challenges: A literature review on the use of Machine Learning Methods. J. Manuf. Syst. 2022, 63, 550–562. [Google Scholar] [CrossRef]
  10. Kong, Z.M.; Cui, Y.; Xia, Z.; Lv, H. Convolution and Long Short-Term Memory Hybrid Deep Neural Networks for Remaining Useful Life Prognostics. Appl. Sci. 2019, 9, 4156. [Google Scholar] [CrossRef]
  11. Long, W.; Yan, D.; Liang, G. A new ensemble residual convolutional neural network for remaining useful life estimation. Math. Biosci. Eng. 2019, 16, 862–880. [Google Scholar] [CrossRef]
  12. Bhatt, D.; Patel, C.; Talsania, H.; Patel, J.; Vaghela, R.; Pandya, S.; Modi, K.; Ghayvat, H. CNN Variants for Computer Vision: History, Architecture, Application, Challenges and Future Scope. Electronics 2021, 10, 2470. [Google Scholar] [CrossRef]
  13. Guo, L.; Li, N.P.; Jia, F.; Lei, Y.G.; Lin, J. A recurrent neural network based health indicator for remaining useful life prediction of bearings. Neurocomputing 2017, 240, 98–109. [Google Scholar] [CrossRef]
  14. Chemali, E.; Kollmeyer, P.J.; Preindl, M.; Ahmed, R.; Emadi, A. Long Short-Term Memory Networks for Accurate State-of-Charge Estimation of Li-ion Batteries. IEEE Trans. Ind. Electron. 2018, 65, 6730–6739. [Google Scholar] [CrossRef]
  15. Sherstinsky, A. Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) network. Phys. D Nonlinear Phenom. 2020, 404, 132306. [Google Scholar] [CrossRef]
  16. Wu, S.W.; Sun, F.; Zhang, W.T.; Xie, X.; Cui, B. Graph Neural Networks in Recommender Systems: A Survey. ACM Comput. Surv. 2023, 55, 1–37. [Google Scholar] [CrossRef]
  17. Yang, X.Y.; Zheng, Y.; Zhang, Y.; Wong, D.S.H.; Yang, W.D. Bearing Remaining Useful Life Prediction Based on Regression Shapalet and Graph Neural Network. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  18. Han, K.; Wang, Y.H.; Chen, H.T.; Chen, X.H.; Guo, J.Y.; Liu, Z.H.; Tang, Y.H.; Xiao, A.; Xu, C.J.; Xu, Y.X.; et al. A Survey on Vision Transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 87–110. [Google Scholar] [CrossRef] [PubMed]
  19. Ding, Y.F.; Jia, M.P.; Miao, Q.H.; Cao, Y.D. A novel time-frequency Transformer based on self-attention mechanism and its application in fault diagnosis of rolling bearings. Mech. Syst. Signal Proc. 2022, 168, 108616. [Google Scholar] [CrossRef]
  20. Wang, L.; Cao, H.R.; Xu, H.; Liu, H.C. A gated graph convolutional network with multi-sensor signals for remaining useful life prediction. Knowl.-Based Syst. 2022, 252, 109340. [Google Scholar] [CrossRef]
  21. Huang, C.G.; Huang, H.Z.; Li, Y.F.; Peng, W.W. A novel deep convolutional neural network-bootstrap integrated method for RUL prediction of rolling bearing. J. Manuf. Syst. 2021, 61, 757–772. [Google Scholar] [CrossRef]
  22. Liu, Y.W.; Sun, J.; Shang, Y.L.; Zhang, X.D.; Ren, S.; Wang, D.T. A novel remaining useful life prediction method for lithium-ion battery based on long short-term memory network optimized by improved sparrow search algorithm. J. Energy Storage 2023, 61, 106645. [Google Scholar] [CrossRef]
  23. Ma, P.; Li, G.F.; Zhang, H.L.; Wang, C.; Li, X.K. Prediction of Remaining Useful Life of Rolling Bearings Based on Multiscale Efficient Channel Attention CNN and Bidirectional GRU. IEEE Trans. Instrum. Meas. 2024, 73, 1–13. [Google Scholar] [CrossRef]
  24. Xiang, L.; Wang, P.H.; Yang, X.; Hu, A.J.; Su, H. Fault detection of wind turbine based on SCADA data analysis using CNN and LSTM with attention mechanism. Measurement 2021, 175, 109094. [Google Scholar] [CrossRef]
  25. Li, M.W.; Xu, D.Y.; Geng, J.; Hong, W.C. A hybrid approach for forecasting ship motion using CNN-GRU-AM and GCWOA. Appl. Soft Comput. 2022, 114, 108084. [Google Scholar] [CrossRef]
  26. Chen, D.Q.; Hong, W.C.; Zhou, X.Z. Transformer Network for Remaining Useful Life Prediction of Lithium-Ion Batteries. IEEE Access 2022, 10, 19621–19628. [Google Scholar] [CrossRef]
  27. Zhang, Z.Z.; Song, W.; Li, Q.Q. Dual-Aspect Self-Attention Based on Transformer for Remaining Useful Life Prediction. IEEE Trans. Instrum. Meas. 2022, 71, 1–11. [Google Scholar] [CrossRef]
  28. Brauwers, G.; Frasincar, F. A General Survey on Attention Mechanisms in Deep Learning. IEEE Trans. Knowl. Data Eng. 2023, 35, 3279–3298. [Google Scholar] [CrossRef]
  29. Chen, Z.H.; Wu, M.; Zhao, R.; Guretno, F.; Yan, R.Q.; Li, X.L. Machine Remaining Useful Life Prediction via an Attention-Based Deep Learning Approach. IEEE Trans. Ind. Electron. 2021, 68, 2521–2531. [Google Scholar] [CrossRef]
  30. Xia, J.; Feng, Y.W.; Teng, D.; Chen, J.Y.; Song, Z.C. Distance self-attention network method for remaining useful life estimation of aeroengine with parallel computing. Reliab. Eng. Syst. Saf. 2022, 225, 108636. [Google Scholar] [CrossRef]
  31. Niu, Z.Y.; Zhong, G.Q.; Yu, H. A review on the attention mechanism of deep learning. Neurocomputing 2021, 452, 48–62. [Google Scholar] [CrossRef]
  32. Zhang, Q.; Ye, Z.J.; Shao, S.Y.; Niu, T.L.; Zhao, Y.W. Remaining useful life prediction of rolling bearings based on convolutional recurrent attention network. Assem. Autom. 2022, 42, 372–387. [Google Scholar] [CrossRef]
  33. Fan, Z.Y.; Li, W.R.; Chang, K.C. A Two-Stage Attention-Based Hierarchical Transformer for Turbofan Engine Remaining Useful Life Prediction. Sensors 2024, 24, 824. [Google Scholar] [CrossRef] [PubMed]
  34. Xiang, S.; Li, P.H.; Huang, Y.; Luo, J.; Qin, Y. Single gated RNN with differential weighted information storage mechanism and its application to machine RUL prediction. Reliab. Eng. Syst. Saf. 2024, 242, 109741. [Google Scholar] [CrossRef]
  35. Xiang, S.; Zheng, X.Y.; Miao, J.G.; Qin, Y.; Li, P.H.; Hou, J.; Ilolov, M. Dynamic Self-Learning Neural Network and Its Application for Rotating Equipment RUL Prediction. IEEE Internet Things J. 2025, 12, 12257–12266. [Google Scholar] [CrossRef]
  36. Li, P.H.; Zheng, X.Y.; Xiang, S.; Hou, J.; Qin, Y.; Kurboniyon, M.S.; Ren, W. Channel Independence Bidirectional Gated Mamba With Interactive Recurrent Mechanism for Time Series Forecasting. In Proceedings of the IEEE Transactions on Industrial Electronics, Piscataway, NJ, USA, 10 July 2025. [Google Scholar]
  37. Liu, Y.; Hu, T.; Zhang, H.; Wu, H.; Wang, S.; Ma, L.; Long, M. iTransformer: Inverted Transformers Are Effective for Time Series Forecasting. arXiv 2024, arXiv:2310.06625. [Google Scholar] [CrossRef]
  38. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef]
  39. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage propagation modeling for aircraft engine run-to-failure simulation. In Proceedings of the 2008 International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008; pp. 1–9. [Google Scholar]
  40. Zheng, S.; Ristovski, K.; Farahat, A.; Gupta, C. Long Short-Term Memory Network for Remaining Useful Life Estimation. In Proceedings of the 2017 IEEE International Conference on Prognostics and Health Management (ICPHM), Dallas, TX, USA, 19–21 June 2017; pp. 88–95. [Google Scholar]
  41. Zhang, Y.; Xin, Y.; Liu, Z.-W.; Chi, M.; Ma, G. Health status assessment and remaining useful life prediction of aero-engine based on BiGRU and MMoE. Reliab. Eng. Syst. Saf. 2022, 220, 108263. [Google Scholar] [CrossRef]
  42. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  43. Liu, H.; Sun, Y.C.; Wang, H.; Zhang, H.Y. CTNet: Improving the non-stationary predictive ability of remaining useful life of aero-engine under multiple time-varying operating conditions. Measurement 2025, 243, 116345. [Google Scholar] [CrossRef]
  44. Xiang, F.F.; Zhang, Y.M.; Zhang, S.Y.; Wang, Z.L.; Qiu, L.M.; Choi, J.H. Bayesian gated-transformer model for risk-aware prediction of aero-engine remaining useful life. Expert Syst. Appl. 2024, 238, 121859. [Google Scholar] [CrossRef]
  45. Wang, J.; Wen, G.; Yang, S.; Liu, Y. Remaining Useful Life Estimation in Prognostics Using Deep Bidirectional LSTM Neural Network. In Proceedings of the 2018 Prognostics and System Health Management Conference (PHM-Chongqing), Chongqing, China, 26–28 October 2018; pp. 1037–1042. [Google Scholar]
  46. Li, H.; Zhao, W.; Zhang, Y.; Zio, E. Remaining useful life prediction using multi-scale deep convolutional neural network. Appl. Soft Comput. 2020, 89, 106113. [Google Scholar] [CrossRef]
  47. Deng, K.; Zhang, X.; Cheng, Y.; Zheng, Z.; Jiang, F.; Liu, W.; Peng, J. A remaining useful life prediction method with long-short term feature processing for aircraft engines. Appl. Soft Comput. 2020, 93, 106344. [Google Scholar] [CrossRef]
  48. Sayah, M.; Guebli, D.; Zerhouni, N.; Masry, Z.A. Towards Distribution Clustering-Based Deep LSTM Models for RUL Prediction. In Proceedings of the 2020 Prognostics and Health Management Conference (PHM-Besançon), Besançon, France, 4–7 May 2020; pp. 253–256. [Google Scholar]
  49. Benker, M.; Furtner, L.; Semm, T.; Zaeh, M.F. Utilizing uncertainty information in remaining useful life estimation via Bayesian neural networks and Hamiltonian Monte Carlo. J. Manuf. Syst. 2021, 61, 799–807. [Google Scholar] [CrossRef]
  50. Duan, Y.; Li, H.; He, M.; Zhao, D. A BiGRU Autoencoder Remaining Useful Life Prediction Scheme with Attention Mechanism and Skip Connection. IEEE Sens. J. 2021, 21, 10905–10914. [Google Scholar] [CrossRef]
  51. Liu, L.; Song, X.; Zhou, Z. Aircraft engine remaining useful life estimation via a double attention-based data-driven architecture. Reliab. Eng. Syst. Saf. 2022, 221, 108330. [Google Scholar] [CrossRef]
  52. Zhang, J.; Li, X.; Tian, J.; Luo, H.; Yin, S. An integrated multi-head dual sparse self-attention network for remaining useful life prediction. Reliab. Eng. Syst. Saf. 2023, 233, 109096. [Google Scholar] [CrossRef]
Figure 1. Overall architecture of the HMDAM model.
Figure 1. Overall architecture of the HMDAM model.
Sensors 25 05682 g001
Figure 2. Data processing procedure in the HMDAM framework.
Figure 2. Data processing procedure in the HMDAM framework.
Sensors 25 05682 g002
Figure 3. Overview of the data preprocessing procedure.
Figure 3. Overview of the data preprocessing procedure.
Sensors 25 05682 g003
Figure 4. Sensor data distribution in the FD001 dataset.
Figure 4. Sensor data distribution in the FD001 dataset.
Sensors 25 05682 g004
Figure 5. Two degradation patterns of Engine 1 in FD001: (a) linear degradation model; (b) piecewise linear degradation model.
Figure 5. Two degradation patterns of Engine 1 in FD001: (a) linear degradation model; (b) piecewise linear degradation model.
Sensors 25 05682 g005
Figure 6. Performance using different time window lengths: (a) FD001; (b) FD002; (c) FD003; (d) FD004.
Figure 6. Performance using different time window lengths: (a) FD001; (b) FD002; (c) FD003; (d) FD004.
Sensors 25 05682 g006
Figure 7. Performance comparison under varying hyperparameter settings for FD001: (a) number of encoder layers; (b) number of heads; (c) batch size; (d) model dimension.
Figure 7. Performance comparison under varying hyperparameter settings for FD001: (a) number of encoder layers; (b) number of heads; (c) batch size; (d) model dimension.
Sensors 25 05682 g007
Figure 8. Performance comparison under varying hyperparameter settings for FD002: (a) number of encoder layers; (b) number of heads; (c) batch size; (d) model dimension.
Figure 8. Performance comparison under varying hyperparameter settings for FD002: (a) number of encoder layers; (b) number of heads; (c) batch size; (d) model dimension.
Sensors 25 05682 g008
Figure 9. Performance comparison under varying hyperparameter settings for FD003: (a) number of encoder layers; (b) number of heads; (c) batch size; (d) model dimension.
Figure 9. Performance comparison under varying hyperparameter settings for FD003: (a) number of encoder layers; (b) number of heads; (c) batch size; (d) model dimension.
Sensors 25 05682 g009
Figure 10. Performance comparison under varying hyperparameter settings for FD004: (a) number of encoder layers; (b) number of heads; (c) batch size; (d) model dimension.
Figure 10. Performance comparison under varying hyperparameter settings for FD004: (a) number of encoder layers; (b) number of heads; (c) batch size; (d) model dimension.
Sensors 25 05682 g010
Figure 11. RUL prediction results using the HMDAM model for (a) FD001; (b) FD002; (c) FD003; (d) FD004.
Figure 11. RUL prediction results using the HMDAM model for (a) FD001; (b) FD002; (c) FD003; (d) FD004.
Sensors 25 05682 g011
Figure 12. RUL prediction performance for selected engine units across the four sub-dataset: (a) the 34th engine in test set FD001; (b) the 138th engine in test set FD002; (c) the 46th engine in test set FD003; (d) the 12th engine in test set FD004.
Figure 12. RUL prediction performance for selected engine units across the four sub-dataset: (a) the 34th engine in test set FD001; (b) the 138th engine in test set FD002; (c) the 46th engine in test set FD003; (d) the 12th engine in test set FD004.
Sensors 25 05682 g012
Table 1. Summary of the C-MAPSS data.
Table 1. Summary of the C-MAPSS data.
DatasetC-MAPSS
FD001FD002FD003FD004
Training engines100260100249
Test engines100256100248
Operating Conditions1616
Fault modes1122
Training set size20,63153,75924,72045,918
Test set size100259100218
Table 2. Configuration of training parameters.
Table 2. Configuration of training parameters.
HyperparameterDescriptionOption
Batch sizeThe number of samples for each backpropagation32
OptimizerAlgorithm for minimizing lossAdam
Training epochsThe number of backpropagations for each sample100
Learning rate (lr)Initial learning rate of training0.001–0.0001
Dropout rateProportion of samples discarded0.2
Table 3. Structural configuration of the proposed model.
Table 3. Structural configuration of the proposed model.
ComponentsLayersParametersOption
TFEBEncoder layerNumber of Conv1d layers2
Kernel size of Conv1d layer1
Number of norm layers2
Number of hidden dimensions32
Number of extended dimensions128
Number of heads12
ActivationReLU
Number of encoder layers2
SFEBLinear network layerNumber of hidden dimensions32
Reduction4
ActivationReLU
Number of linear network layers2
Projection layerActivationSigmoid
RegressorLinear network layerActivationReLU
Number of linear network layers3
Prediction layerActivationSigmoid
Table 4. Configuration of hyperparameter values.
Table 4. Configuration of hyperparameter values.
ParameterABCD
Number of encoder layers1234
Number of heads481216
Batch size163264128
Model dimension3264128256
Table 5. Optimal hyperparameter settings for each sub-dataset.
Table 5. Optimal hyperparameter settings for each sub-dataset.
ParameterFD001FD002FD003FD004
Number of encoder layers3334
Number of heads416816
Batch size128326432
Model dimension32323232
Table 6. Performance comparison of different methods.
Table 6. Performance comparison of different methods.
MethodsFD001FD002FD003FD004
RMSES-ScoreMAERMSES-ScoreMAERMSES-ScoreMAERMSES-ScoreMAE
Transformer14.23379.5610.3218.623751.6112.7412.86326.5510.2622.879862.5915.87
iTransformer11.47190.379.1015.9249914.0812.32298.088.9720.09246915.6
ST12.58273.669.5117.352587.5511.9312.53291.1710.122.825936.0615.89
HMDAM10.82170.079.0215.331130.4210.7211.21239.478.9617.481738.1911.88
Table 7. Comparison of FLOPs and parameter counts of the models.
Table 7. Comparison of FLOPs and parameter counts of the models.
MethodNumber of FLOPsNumber of Parameters
Transformer810.82 K87.39 K
iTransformer441.92 K89.66 K
ST1.185 M100.50 K
HMDAM821.25 K89.02 K
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

He, C.; Li, Z.; Zheng, C.; Zhang, Z.; Zhang, L. A Hybrid Model Based on a Dual-Attention Mechanism for the Prediction of Remaining Useful Life of Aircraft Engines. Sensors 2025, 25, 5682. https://doi.org/10.3390/s25185682

AMA Style

He C, Li Z, Zheng C, Zhang Z, Zhang L. A Hybrid Model Based on a Dual-Attention Mechanism for the Prediction of Remaining Useful Life of Aircraft Engines. Sensors. 2025; 25(18):5682. https://doi.org/10.3390/s25185682

Chicago/Turabian Style

He, Chenwen, Zixiang Li, Chenyu Zheng, Zikai Zhang, and Liping Zhang. 2025. "A Hybrid Model Based on a Dual-Attention Mechanism for the Prediction of Remaining Useful Life of Aircraft Engines" Sensors 25, no. 18: 5682. https://doi.org/10.3390/s25185682

APA Style

He, C., Li, Z., Zheng, C., Zhang, Z., & Zhang, L. (2025). A Hybrid Model Based on a Dual-Attention Mechanism for the Prediction of Remaining Useful Life of Aircraft Engines. Sensors, 25(18), 5682. https://doi.org/10.3390/s25185682

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop