Next Article in Journal
The Angular Momentum Unloading of the Asymmetric GEO Satellite by Using Electric Propulsion with a Mechanical Arm
Previous Article in Journal
Effects of Pitch Stabilization Buffer on the Dynamic Performance of Frame-Type Landing Gear
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Transformer Encoder Enhanced by an Adaptive Graph Convolutional Neural Network for Prediction of Aero-Engines’ Remaining Useful Life

1
School of Mechanical Engineering, Xi’an Jiaotong University, Xi’an 710049, China
2
National Key Lab of Aerospace Power System and Plasma Technology, Xi’an Jiaotong University, Xi’an 710049, China
3
School of Future Technology, Xi’an Jiaotong University, Xi’an 710049, China
*
Author to whom correspondence should be addressed.
Aerospace 2024, 11(4), 289; https://doi.org/10.3390/aerospace11040289
Submission received: 25 February 2024 / Revised: 2 April 2024 / Accepted: 6 April 2024 / Published: 9 April 2024

Abstract

:
Accurate prediction of remaining useful life (RUL) plays a significant role in ensuring the safe flight of aircraft. With the recent rapid development of deep learning, there has been a growing trend towards more precise RUL prediction. However, while many current deep learning methods are capable of extracting spatial features—those along the sensor dimension—through convolutional kernels or fully connected layers, their extraction capacity is often limited due to the small scale of kernels and the high uncertainty associated with linear weights. Graph neural networks (GNNs), emerging as effective approaches for processing graph-structured data, explicitly consider the relationships between sensors. This is akin to imposing a constraint on the training process, thereby allowing the learned results to better approximate real-world situations. In order to address the challenge of GNNs in extracting temporal features, we augment our proposed framework for RUL prediction with a Transformer encoder, resulting in the adaptive graph convolutional transformer encoder (AGCTE). A case study using the C-MAPSS dataset is conducted to validate the effectiveness of our proposed model.

1. Introduction

Aero-engines are the core power component of aircraft and whether they operate normally will have a crucial effect on the safety of the flight. However, aero-engines work under harsh environments [1] and have numerous components, meaning their reliability limited, which both make them prone to failure. But when the failure can be anticipated, carrying out maintenance at unreasonable times leads to the waste of time and cost. Thus, how to make an accurate prediction of the occurrence time of failure, that is, the remaining useful life (RUL) of current components, is a significant issue that needs to be researched.
The methods of RUL prediction can be roughly divided into two categories, model-based and data-based [2]. Model-based methods [3,4] predict the RUL through analyzing the degradation processes, such as cracks, fatigue, and deformation [5], which requires sufficient research on the mechanism of the degradation processes. For aero-engine complex systems, it is not possible to fully analyze the degradation mechanism of every component or the whole system, so we can only resort to data-driven methods. With the fast development of sensor technologies, a great number of data from aero-engines can be acquired [6], making it possible to implement data-driven methods in industrial applications. And this may be another of the reasons why data-driven methods have been a hot topic in the RUL prediction field in recent years [1] in addition to the rapid advances in artificial intelligence.
Data driven methods can be further classified into machine learning methods and deep learning methods. Conventional machine learning methods, such as Bayesian belief networks [7], support vector regression [8], and the support vector machine [9], have achieved positive results but demand extensive efforts for manual pre-feature engineering [1]. On the contrary, deep learning (DL) methods could map the monitoring data with not much preprocessing, such as max–min normalization, directly into the RUL values, possessing the strengths of convenience, low application threshold, and generality to a certain extent. Common DL methods include convolutional neural network (CNN)-based, recurrent neural network (RNN)-based, auto-encoder (AE)-based, Transformer-based methods, and their combination. For example, Li et al. [10] proposed a deep CNN (DCNN) framework for the RUL prediction. Sayah et al. [11] combined deep long short-term memory networks (LSTM) with Gaussian mixture models (GMMs) and applied it to the RUL prediction of complex industrial system components. Chen et al. [12] proposed a quadratic function-based deep convolutional AE (DCAE) to construct health indicators (HIs) and found that the extracted indicators have more predictive power than those bulit from traditional methods. Kamei et al. [13] developed two predictive models, LSTM and the Transformer, to predict the RUL, combined with two decentralized federated learning (FL) algorithms and verified the performance of FL.
Although these methods achieved great success, they each have their own drawbacks. For the CNN-based and AE-based methods, they could extract spatial features through convolutional kernels and fully connected layers along the dimension of the sensor, as shown in Figure 1, but they lack the capacity to model time series changes. For RNN-based and transformer-based methods, they all have fully connected layers along the dimension of the sensor, thus possessing the ability to extract spatial features to a certain extent. Moreover, different from CNN-based and AE-based methods, they could model time series changes so better results may be achieved. Furthermore, compared with RNN-based methods, Transformer-based methods have the capacity to capture long time dependencies, so they may perform better than RNN-based methods when dealing with long time series data [13]. However, for extracting spatial features using fully connected layers, a problem exists, i.e., only the loss function is used to constrain the values of linear weights and the learned weights may not reflects the true relationships between sensors, leading to suboptimal results. Thus, to make our trained model better extract true relationships between sensors, some other methods need to be introduced.
Graph neural networks (GNNs) [14], emerging as a kind of effective deep learning method, are developed to deal with graph-structured data. As the sensors in aero-engines intercorrelate with each other, it forms a kind of natural graph-structured data, with the sensors as nodes and relationships as edges. GNNs take the relationships between nodes into account explicitly, equivalent to adding a constraint compared to a linear weight whose elements could take any value. So the learned results may have a better approximation to the true situations. But general GNNs could not model time series changes, so some researchers combined GNNs with RNN-based methods to perform RUL prediction. For example, Li et al. [15] proposed a directed acyclic graph (DAG) network that combines LSTM and a CNN to predict the RUL. Kong et al. [16] combined a graph attention network (GAT) and LSTM to develop a graph neural network (GNN)-based spatio-temporal fusion attention (STFA) approach. But these methods may not be suitable for extracting long time dependencies, which may lead to suboptimal results. Moreover, little research combined GNN with the Transformer, whose performance is better than RNN when we deal with long time series data, to make RUL prediction. Based on this research result, we propose a new method called the adaptive graph convolutional transformer encoder (AGCTE) using the graph convolutional network (GCN) [17] and the Transformer encoder.
In AGCTE, the GCN part is used to model the sensor relationships and helps capture spatial features. The Transformer part is used to model the long time series changes. As a result, both spatial and temporal features could be extracted.
The remainder of this paper is organized as follows. Section 2 presents the AGCTE model used in this paper. Section 3 gives the descriptions of the experiments. Results and some analysis are shown in Section 4. This paper is finally concluded in Section 5.

2. Methodology

In this section, we will elaborate on our proposed architecture, as illustrated in Figure 2. It mainly contains two parts, adaptive GCN layers and the Transformer encoder.

2.1. Adaptive GCN Layers

To use GCN, first, the adjacency matrix that describes the connection relationships between nodes should be defined. In some of the literature [6,17], a pre-defined graph is used, which is calculated as follows:
cos θ = C o v ( X i , X j ) σ X i σ X j = E X i X j E X i E X j σ X i σ X j ,
A i j = cos θ , cos θ > 0 0 , cos θ < 0 ,
where X i and X j are two 1-dimensional time series data; E · means mathematical expectation function; σ X i and σ X j are the standard deviation of X i and X j , respectively; and cos θ is the cosine similarity between X i and X j . A problem exists in such a method: a predefined graph may not be best suitable for the GCN’s message passing process [18]. So in this paper, we use an adaptive generation approach. Following [19], we add a learnable embedding vector for every node, representing each sensor’s own characteristics:
V = [ v 1 T , , v N T ] T R N × d ,
where d represents the dimension of the embedding vector and N is the number of nodes. Then, following [20], we could acquire a learnable adjacency matrix using the following equation:
A = Softmax LeakyReLU ( V · V T ) ,
where Softmax is used to normalize the learned matrix and LeakyReLU is a non-linear activation function to enhance the learning capacity. Compared to the predefined graph, the adaptive graph is acquired during the process of training under the constraint of minimizing the loss function values. Thus, this graph may better serve the GCN’s message passing process. In classical GCN [21], the expression of GCN layer is as follows:
Y = σ A ^ X W ,
where Y is the processed signal and X is the original signal; A ^ = D ˜ 1 / 2 A ˜ D ˜ 1 / 2 , A ˜ = A 0 + I , and A 0 is the adjacency matrix, I is the identity matrix, D ˜ is a diagonal matrix and its element D ˜ i i = j = 1 N A ˜ i j ; W is the weight matrix; and σ · is an activation function. In this paper, as Softmax function has already normalized the adjacency matrix, we use A to perform the message passing process and the adaptive GCN layer is as follows:
Y = σ [ A X W ] .

2.2. Transformer Encoder

The Transformer is first proposed for natural language processing (NLP) in [22]. It consists of two parts, the encoder and decoder. Following the work in [23], we only use the encoder part to capture the short-term and long-term time series information. The encoder part is composed of input embedding, positional embedding, multi-head attention, skip connection, layer-normalization, and a feed-forward part. Their detailed descriptions are as follows.

2.2.1. Input Embedding

One of the functions of input embedding is to transform the dimension of the input to be suitable for the further calculation of multi-head attention. But the choice of the input embedding approaches, such as fully connected layer or CNN, could have a non-neligible effect on the final results. According to the experiments in [23], a fully connected layer could have a better processing effect than CNN, so in this paper, we take the same strategy and its expression is
Y = Y W i + b i ,
where W i is a learnable weight matrix and b i is a learnable bias.

2.2.2. Positional Embedding

When multi-head attention is performed, unlike RNN-based methods, the sequential order of the input data loses. As a result, in order to keep such part of information, a positional embedding is added into the input after input embedding.
For a signal Y R T × D , where T is the time dimension and D is the feature dimension, first, a scaled factor D is used to enlarge the signal value to avoid the added positional embedding dominating the input. That is,
Y = Y · D .
Then, the following positional embedding is added
p t , 2 d = sin t / 10,000 2 d / D ,
p t , 2 d + 1 = cos t / 10,000 2 d / D ,
Y = Y + p .

2.2.3. Multi-Head Attention

Multi-head attention is composed of several parallel self-attention modules [24]. A self-attention function can be described as mapping a query and a set of key-value pairs to an output [22], where query, key, and value are the linear transformation of input, as
Q i = Y W i q ,
K i = Y W i k ,
V i = Y W i v ,
where W i q , W i k , a n d   W i v R D × D k are learnable weight matrixes and Q i , K i ,   a n d   V i are query, key, and value, respectively. Then, the attention value can be calculated as follows:
Attention Q i , K i , V i = Softmax Q i K i T D k V i .
For multi-head attention, we need to repeat the above process h times, where h is the number of heads. To be specific,
MultiHead Q , K , V = Concat head 1 , , head h W O ,
where head i = Attention ( Q i , K i , V i ) , W O R h D k × D model , D model = D k h .

2.2.4. Feed-Forward Function

After the multi-head attention module, there is a feed-forward function consisting of two linear maps in the Transformer encoder. In this paper, the activation function is chosen as the GeLU function and the expression of this part is
Y = GeLU Y W 1 + b 1 W 2 + b 2 ,
where W 1 R D model × D 1 , W 2 R D 1 × D 2 , b 1 R D 1 , a n d b 2 R D 2 . The most important reason why we use the GeLU function is that it performs better in the Transformer than the ReLU function in many tasks and it follows some of the same work [23,25].

2.2.5. Skip Connection and Layernormalization

Skip connection and layernormalization are added twice in a transformer encoder layer and their positions are shown in Figure 2. Skip connection proves to be quite significant in ensuring the output does not converge to rank-1 matrix, keeping the performance of the Transformer [26].

3. Experiment

In this section, we give the description of the validation dataset we use and the basic experiment setting. The main contents of this section are dataset description, data preprocessing, evaluation metrics, and main implementation details.

3.1. Dataset Description

The C-MAPSS (Commercial Modular Aero-Propulsion System Simulation) dataset [27] is utilized to validate the effectiveness of our proposed model, AGCTE. This datatset is an aero-engine operation datasets, composed of four subdatasets, FD001 to FD004. Each subdataset contains numerous aero-engines and has three files for training, whose data are of rul-to-failure kind, testing and corresponeding RUL values of testing set. The engines in the C-MAPSS dataset may work under different working conditions and have different failure modes. And there are 21 sensors recording the different operating parameters such as pressure and temperature. The detailed descriptions of the engines in C-MAPSS dataset and the sensors are list in Table 1 and Table 2, respectively.

3.2. Data Preprocessing

There are 21 sensors available for our calculation. However, the values of seven of them do not change along with the time, which means they could not reflect the degradation process of the engines and are of no use for predicting the RUL. As a result, they are removed from the sensor list and the final chosen sensors are sensor 2, 3, 4, 7, 8, 9, 11, 12, 13, 14, 15, 17, 20, and 21.
After the sensor selection, the max–min normalization is performed to normalize each sensor data to [0, 1]. As the work conditions are multiple in FD002 and FD004 subdatasets, first K-means clustering is performed and then in each cluster class, the max–min normalization is utilized as follows:
y norm i , c = y i , c y min i , c y max i , c y min i , c
where y norm i , c is the normalized sensor data; y i , c is the original sensor data; and y max i , c   a n d   y min i , c are the maximum and minimum of the i-th sensor under the c-th working condition, respectively. This calculation expression is also suitable for subdatasets FD001 and FD003.
Before training, in order to acquire more data samples, the sliding window method [23] is utilized to split the training set. Each slide forms a data sample and the goal is to predict the RUL of the next time after the end of this slide. Moreover, for RUL values used in model training, the maximum is set to 125 following previous research shown in Figure 3 [24] as at the initial operation stage of an engine, the degradation process is negligible so the RUL value could be considered as constant when we train our model for RUL prediction.

3.3. Evaluation Metrics

In this paper, two same evaluation metrics, root mean square error (RMSE) and scoring function (SF), are utilized to estimate the efficacy of the proposed model. RMSE and SF are defined as follows:
δ i = RUL i ^ RUL i ,
RMSE = 1 | D test | i = 1 | D test | δ i 2 ,
SF = i = 1 | D test | SF i , with SF i = e δ i / 13 1 , δ i < 0 e δ i / 10 1 , δ i > 0 ,
where RUL i ^ and RUL i are the predicted and ground-truth RUL of the i-th engine; | D test | is the engine number of the testing set. The curves of RMSE and SF are plotted in Figure 4. The Score Function (SF) indicator was initially introduced in the PHM08 data competition and has since been widely utilized in remaining useful life (RUL) prediction to evaluate the performance of RUL models. Its significance can be elucidated as follows: when the predicted results deviate significantly from the actual RUL values, the resultant loss is substantial and the greater the disparity, the more pronounced the loss. Therefore, to mitigate such scenarios, the SF is formulated as an exponential function of the deviation, enabling the selection of models capable of minimizing substantial losses. Moreover, it is notable that there is a greater penalty imposed on predictions where the estimated RUL exceeds the actual RUL. This is because if the predicted RUL falls short of the actual RUL, it may not lead to catastrophic consequences, as appropriate maintenance measures could potentially extend the RUL. Conversely, if the predicted RUL surpasses the actual RUL, the situation is reversed. Consequently, predictions with an overestimation incur heavier penalties, reflecting the increased severity of such discrepancies.

3.4. Implementation Details

For the python library, our method is implemented in Pytorch. We use an Adam optimizer with 1 × 10 3 learning rate and set the number of adaptive GCN layers, Transformer encoder layers, and multi-head attention heads all to two. The feed-forward dimension in the Transformer encoder is set to 10. For FD001 and FD003, the model dimension of the Transformer encoder is set to 16, no dropout is used, and the activation function of adaptive GCN layers is tanh. While for FD002 and FD004, the model dimension of transformer encoder is set to 14, no activation function of adaptive GCN layers is used and dropout equals 0.2.

4. Results and Analysis

We compare our proposed model with 12 baselines, they are CNN-based methods [10,28], RNN-based methods [29,30], autoencoder-based methods [31], Transformer-based methods [13,24,32], and GNN-based methods [1,15,16,33]. The comparison results are listed in Table 3 and Table 4.
From the results, it can be seen that our proposed model performs better than other methods in FD002 and FD004 subdatasets, especially for RMSE indicators. And in terms of both RMSE and SF indicators, the average performance of our model ranks first, demonstrating the effectiveness of our proposed model.
Moreover, four examples of the predicted RUL are depicted in Figure 5 and the comparison of the predicted RUL and the ground-truth RUL of the testing set of each subdataset are shown in Figure 6. From Figure 5, we can see that the error between the predicted RUL and the ground-truth RUL is relatively small in much of the time. In this phenomenon, our model could learn the linear degradation trend of aero-engines relatively well. From Figure 6, we can see that the run-to-failure prediction results are quite close to the ground-truth values, which is consistent with the results shown in Table 3 and Table 4 and demonstrates the effectiveness of our proposed model again.
In order to further verify the function of combining the GCN and Transformer encoder together, we perform an ablation study on our model, removing the GCN and Transformer encoder parts to form AGCTE-w/o GCN and AGCTE-w/o Transformer models, respectively. The RMSE of the experiments are listed in Table 5.
From Table 5, it can be observed that average performances of the ablation models both have some decreases at different degrees, illustrating the effectiveness of combining GCN and the Transformer encoder together. Moreover, the performance of the model AGCTE-w/o Transformer decreases greatly, which demonstrates the significance of the Transformer encoder in RUL prediction. In addition, as depicted in Figure 1, the Transformer has fully connected layers along sensor dimension, thus possessing the capacity of extracting spatial features to a certain extent. But as we can see in Table 4, after adding the GCN part, the performance has certain improvements. This experiment result reveals that GCN can better extract spatial features that fully connect layers, as GCN takes the relationships between sensor explicitly. The learned adjacency matrixes are shown in Figure 7.
As for why our model performs better than other models on FD002 and FD004 subdatasets, we think the reasons can be two-fold. On one hand, we employ the normalization method tailored to different operating conditions, which will help reduce the effect that the variation in sensor values under different operating conditions has on the extraction of features related to the remaining using life (RUL) indicator. On the other hand, there has been one related work [23] purely using the Transformer encoder to perform RUL prediction on the same dataset. And we can see the results in [23]: the performance on FD002 and FD004 is better. Moreover, in ablation experiments, we can see AGCTE-w/o GCN, that is, the model only using the Transformer encoder, performs even better than our baselines on FD002 and FD004. Both results may demonstrate a conclusion that Transformer can better handle multiple operating condition data in RUL prediction. This may be attributed to the strong capacity of the Transformer to extract temporal features with no need to process time series data in sequence, as the data in multiple operating conditions could vary greatly point by point. And this may be another reason why our model performs better than other baselines. For the Transformer-1 and Transformer-2 method, there may be some slight differences from our model structure and setting. In the end, we found that the score value calculation equation in [23] is different from the popular setting, which is the reverse from that in [23]. So we think it may not be fair to compare our results with those in [23]. We foresee some small errors in their preprocessing code, which may also have unpredictable effects on the results.

5. Conclusions

In this paper, we propose a new model called AGCTE, which combines adaptive GCN and the transformer encoder together. The adaptive GCN part and the transformer encoder part are mainly utilized to extract spatial features and temporal features, respectively. The case study on the C-MAPSS dataset is carried out to validate the effectiveness of our proposed method. The results show the superiority of AGCTE compared with CNN-based, RNN-based, transformer-based, and GNN-based methods. In addition, the ablation study shows that combining GCN and transformer is reasonable and effective.
For further potential direction, as the current method extract spatial features and temporal features relatively independently using two separate modules, this may not be the most natural way to extract spatial–temporal features. As a result, developing a method that could extract spatial and temporal features at the same time may be the potential direction in the future.

Author Contributions

Conceptualization, M.M. and Z.W.; methodology, M.M. and Z.W.; software, Z.W.; validation, Z.W.; writing—original draft preparation, Z.W.; writing—review and editing, Z.W. and Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China (No. 52205124), the Postdoctoral Science Fund (No. 2021M702634), the Basic Research Program of China (No. 2022-JCJQ-JJ-0643), and the Basic Research Fund of Xi’an Jiaotong University (No. xxj032022011).

Data Availability Statement

The dataset used in this paper is an open dataset simulutad by NASA and the download address is https://www.nasa.gov/intelligent-systems-division/discovery-and-systems-health/pcoe/pcoe-data-set-repository/ (accessed on 24 February 2024).

Conflicts of Interest

There are no conflicts of interest.

References

  1. Wang, H.; Zhang, Z.; Li, X.; Deng, X.; Jiang, W. Comprehensive Dynamic Structure Graph Neural Network for Aero-engine Remaining Useful Life Prediction. IEEE Trans. Instrum. Meas. 2023, 72, 3533816. [Google Scholar] [CrossRef]
  2. Chen, J.; Shen, C.; Jing, Z.; Wu, Y.; Chen, R. Remaining useful life prediction of aircraft flap control system with mode transition. AIAA J. 2022, 60, 1104–1115. [Google Scholar] [CrossRef]
  3. Wang, H.; Ma, X.; Zhao, Y. An improved Wiener process model with adaptive drift and diffusion for online remaining useful life prediction. Mech. Syst. Signal Proc. 2019, 127, 370–387. [Google Scholar] [CrossRef]
  4. Chiachío, J.; Jalón, M.L.; Chiachío, M.; Kolios, A. A Markov chains prognostics framework for complex degradation processes. Reliab. Eng. Syst. Saf. 2020, 195, 106621. [Google Scholar] [CrossRef]
  5. Li, Y.; Li, J.; Wang, H.; Liu, C.; Tan, J. Knowledge enhanced ensemble method for remaining useful life prediction under variable working conditions. Reliab. Eng. Syst. Saf. 2024, 242, 109748. [Google Scholar] [CrossRef]
  6. Li, T.; Zhao, Z.; Sun, C.; Yan, R.; Chen, X. Hierarchical attention graph convolutional network to fuse multi-sensor signals for remaining useful life prediction. Reliab. Eng. Syst. Saf. 2021, 215, 107878. [Google Scholar] [CrossRef]
  7. Zhang, X.; Kang, J.; Jin, T. Degradation modeling and maintenance decisions based on Bayesian belief networks. IEEE Trans. Reliab. 2014, 63, 620–633. [Google Scholar] [CrossRef]
  8. Benkedjouh, T.; Medjaher, K.; Zerhouni, N.; Rechak, S. Health assessment and life prediction of cutting tools based on support vector regression. J. Intell. Manuf. 2015, 26, 213–223. [Google Scholar] [CrossRef]
  9. Bezerra Souto Maior, C.; das Chagas Moura, M.; Didier Lins, I.; Lopez Droguett, E.; Henrique Lima Diniz, H. Remaining useful life estimation by empirical mode decomposition and support vector machine. IEEE Latin Am. Trans. 2016, 14, 4603–4610. [Google Scholar] [CrossRef]
  10. Li, X.; Ding, Q.; Sun, J.-Q. Remaining useful life estimation in prognostics using deep convolution neural networks. Reliab. Eng. Syst. Saf. 2018, 172, 1–11. [Google Scholar] [CrossRef]
  11. Sayah, M.; Guebli, D.; Noureddine, Z.; Al Masry, Z. Deep LSTM enhancement for RUL prediction using Gaussian mixture models. Autom. Control Comp. Sci. 2021, 55, 15–25. [Google Scholar] [CrossRef]
  12. Chen, D.; Qin, Y.; Wang, Y.; Zhou, J. Health indicator construction by quadratic function-based deep convolutional auto-encoder and its application into bearing RUL prediction. ISA Trans. 2021, 114, 44–56. [Google Scholar] [CrossRef] [PubMed]
  13. Kamei, S.; Taghipour, S. A comparison study of centralized and decentralized federated learning approaches utilizing the transformer architecture for estimating remaining useful life. Reliab. Eng. Syst. Saf. 2023, 233, 109130. [Google Scholar] [CrossRef]
  14. Wu, Z.; Pan, S.; Chen, F.; Long, G.; Zhang, C.; Philip, S.Y. A comprehensive survey on graph neural networks. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4–24. [Google Scholar] [CrossRef] [PubMed]
  15. Li, J.; Li, X.; He, D. A directed acyclic graph network combined with CNN and LSTM for remaining useful life prediction. IEEE Access 2019, 7, 75464–75475. [Google Scholar] [CrossRef]
  16. Kong, Z.; Jin, X.; Xu, Z.; Zhang, B. Spatio-temporal fusion attention: A novel approach for remaining useful life prediction based on graph neural network. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  17. Liu, N.; Guo, J.; Chen, S.; Zhang, X. Aero-Engines Remaining Useful Life Prognostics Based on Multi-Hierarchical Gated Recurrent Graph Convolutional Network. In Proceedings of the 2023 International Conference on Cyber-Physical Social Intelligence (ICCSI), Xi’an, China, 20–23 October 2023; pp. 642–647. [Google Scholar]
  18. Gilmer, J.; Schoenholz, S.S.; Riley, P.F.; Vinyals, O.; Dahl, G.E. Neural message passing for quantum chemistry. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 1263–1272. [Google Scholar]
  19. Deng, A.; Hooi, B. Graph neural network-based anomaly detection in multivariate time series. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 2–9 February 2021; pp. 4027–4035. [Google Scholar]
  20. Zhang, X.; Leng, Z.; Zhao, Z.; Li, M.; Yu, D.; Chen, X. Spatial-temporal dual-channel adaptive graph convolutional network for remaining useful life prediction with multi-sensor information fusion. Adv. Eng. Inform. 2023, 57, 102120. [Google Scholar] [CrossRef]
  21. Kipf, T.N.; Welling, M. Semi-supervised classification with graph convolutional networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  23. Ogunfowora, O.; Najjaran, H. A Transformer-based Framework For Multi-variate Time Series: A Remaining Useful Life Prediction Use Case. arXiv 2023, arXiv:2308.09884. [Google Scholar]
  24. Mo, Y.; Wu, Q.; Li, X.; Huang, B. Remaining useful life estimation via transformer encoder enhanced by a gated convolutional unit. J. Intell. Manuf. 2021, 32, 1997–2006. [Google Scholar] [CrossRef]
  25. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  26. Dong, Y.; Cordonnier, J.-B.; Loukas, A. Attention is not all you need: Pure attention loses rank doubly exponentially with depth. In Proceedings of the International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 2793–2803. [Google Scholar]
  27. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage propagation modeling for aircraft engine run-to-failure simulation. In Proceedings of the 2008 International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008; pp. 1–9. [Google Scholar]
  28. Sateesh Babu, G.; Zhao, P.; Li, X.-L. Deep convolutional neural network based regression approach for estimation of remaining useful life. In Proceedings of the Database Systems for Advanced Applications: 21st International Conference, DASFAA 2016, Dallas, TX, USA, 16–19 April 2016; Proceedings, Part I 21. pp. 214–228. [Google Scholar]
  29. Zheng, S.; Ristovski, K.; Farahat, A.; Gupta, C. Long short-term memory network for remaining useful life estimation. In Proceedings of the 2017 IEEE international conference on prognostics and health management (ICPHM), Dallas, TX, USA, 19–21 June 2017; pp. 88–95. [Google Scholar]
  30. Ellefsen, A.L.; Bjørlykhaug, E.; Æsøy, V.; Ushakov, S.; Zhang, H. Remaining useful life predictions for turbofan engine degradation using semi-supervised deep architecture. Reliab. Eng. Syst. Saf. 2019, 183, 240–251. [Google Scholar] [CrossRef]
  31. Liu, X.; Xiong, L.; Zhang, Y.; Luo, C. Remaining Useful Life Prediction for Turbofan Engine Using SAE-TCN Model. Aerospace 2023, 10, 715. [Google Scholar] [CrossRef]
  32. Mo, H.; Iacca, G. Evolutionary neural architecture search on transformers for RUL prediction. Mater. Manuf. Process. 2023, 38, 1881–1898. [Google Scholar] [CrossRef]
  33. Wang, L.; Cao, H.; Xu, H.; Liu, H. A gated graph convolutional network with multi-sensor signals for remaining useful life prediction. Knowl.-Based Syst. 2022, 252, 109340. [Google Scholar] [CrossRef]
Figure 1. The illustration of CNN and fully connected layer extracting spatial features.
Figure 1. The illustration of CNN and fully connected layer extracting spatial features.
Aerospace 11 00289 g001
Figure 2. Overall architecture of the proposed model.
Figure 2. Overall architecture of the proposed model.
Aerospace 11 00289 g002
Figure 3. Piece-wise RUL of the C-MAPSS dataset.
Figure 3. Piece-wise RUL of the C-MAPSS dataset.
Aerospace 11 00289 g003
Figure 4. The curve of RMSE and SF.
Figure 4. The curve of RMSE and SF.
Aerospace 11 00289 g004
Figure 5. Four examples of predicted RUL in the testing set.
Figure 5. Four examples of predicted RUL in the testing set.
Aerospace 11 00289 g005
Figure 6. Comparison of the predicted RUL and the ground-truth RUL of the testing set.
Figure 6. Comparison of the predicted RUL and the ground-truth RUL of the testing set.
Aerospace 11 00289 g006
Figure 7. The learned adjacency matrixes in four subdatasets.
Figure 7. The learned adjacency matrixes in four subdatasets.
Aerospace 11 00289 g007
Table 1. The details of the engines in the dataset.
Table 1. The details of the engines in the dataset.
DatasetFD001FD002FD003FD004
Training EU100260100249
Testing EU100259100248
Working Conditions1616
Fault types1122
Table 2. The details of sensor measurement.
Table 2. The details of sensor measurement.
NumberSymbolDescriptionUnits
1T2Total temperature at fan inlet°R
2T24Total temperature at LPC outlet°R
3T30Total temperature at HPC outlet°R
4T50Total temperature at LPT outlet°R
5P2Pressure at fan inletpsia
6P15Total pressure in bypass-ductpsia
7P30Total pressure at HPC outletpsia
8Nfphysical fan speedrpm
9Ncphysical core speedrpm
10eprEngine pressure ratio (P50/P2)-
11Ps30Static pressure at HPC outletpsia
12phiRatio of fuel flow to Ps30pps/psi
13NRfCorrected fan speedrpm
14NRcCorrected core speedrpm
15BPRBypass ratio-
16farBBurner fuel-air ratio-
17htBleedBleed Enthalpy-
18Nf_dmdDemanded fan speedrpm
19PCNfR_dmdDemanded corrected fan speedrpm
20W31HPT coolant bleedlbm/s
21W32LPT coolant bleedlbm/s
Table 3. The RMSE experimental results of the C-MAPSS testing dataset.
Table 3. The RMSE experimental results of the C-MAPSS testing dataset.
MethodsFD001FD002FD003FD004
CNN [28]18.4530.2919.8229.1624.43
DCNN [10]12.6122.3612.6423.3117.73
LSTM-FNN [29]16.1424.4916.1828.1721.25
RBM-LSTM-FNN [30]12.5622.7312.1022.6617.51
DSAE-TCN [31]18.01----
GCU-Transformer [24]11.2722.8111.4224.8617.59
Transformer-1 [13]13.5216.1117.1019.7716.63
Transformer-2 [32]11.5016.1411.3520.0014.75
DAG [15]11.9620.3412.4622.4316.80
STFA [16]11.3519.1711.6421.4115.89
CDSG [1]11.2618.1312.0319.7315.29
GGCN [33]11.8217.2412.2117.3614.66
AGCTE (this paper)12.4613.712.9515.8313.74
Table 4. The SF experimental results of the C-MAPSS testing dataset.
Table 4. The SF experimental results of the C-MAPSS testing dataset.
MethodsFD001FD002FD003FD004Avg.
CNN [28]128713,570159678866084.75
DCNN [10]27410,41228412,4665859
LSTM-FNN [29]33844585255502795.5
RBM-LSTM-FNN [30]231336625128401672
DSAE-TCN [31]161----
GCU-Transformer [24]-----
Transformer-1 [13]287.071436.74263.642784.621193.02
Transformer-2 [32]20211312272298964.5
DAG [15]229273053533701716
STFA [16]194.442493.09224.532760.131418.05
CDSG [1]188174021823321119.5
GGCN [33]186.61493.7245.191371.5824.25
AGCTE (this paper)259.37833.41372.441520.05746.32
Table 5. The RMSE of ablation study experiments.
Table 5. The RMSE of ablation study experiments.
MethodsFD001FD002FD003FD004Avg.
AGCTE12.4613.7012.9515.8313.74
AGCTE-w/o GCN13.614.1612.8616.3614.26
difference1.140.46−0.090.530.51
AGCTE-w/o Transformer13.2722.8312.7335.9421.19
difference0.819.13−0.2220.117.76
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, M.; Wang, Z.; Zhong, Z. Transformer Encoder Enhanced by an Adaptive Graph Convolutional Neural Network for Prediction of Aero-Engines’ Remaining Useful Life. Aerospace 2024, 11, 289. https://doi.org/10.3390/aerospace11040289

AMA Style

Ma M, Wang Z, Zhong Z. Transformer Encoder Enhanced by an Adaptive Graph Convolutional Neural Network for Prediction of Aero-Engines’ Remaining Useful Life. Aerospace. 2024; 11(4):289. https://doi.org/10.3390/aerospace11040289

Chicago/Turabian Style

Ma, Meng, Zhizhen Wang, and Zhirong Zhong. 2024. "Transformer Encoder Enhanced by an Adaptive Graph Convolutional Neural Network for Prediction of Aero-Engines’ Remaining Useful Life" Aerospace 11, no. 4: 289. https://doi.org/10.3390/aerospace11040289

APA Style

Ma, M., Wang, Z., & Zhong, Z. (2024). Transformer Encoder Enhanced by an Adaptive Graph Convolutional Neural Network for Prediction of Aero-Engines’ Remaining Useful Life. Aerospace, 11(4), 289. https://doi.org/10.3390/aerospace11040289

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop