Next Article in Journal
Dim and Small Target Detection with a Combined New Norm and Self-Attention Mechanism of Low-Rank Sparse Inversion
Previous Article in Journal
Telerehabilitation with ARC Intellicare to Cope with Motor and Respiratory Disabilities: Results about the Process, Usability, and Clinical Effect of the “Ricominciare” Pilot Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Unsupervised Machine Learning Approach for Monitoring Data Fusion and Health Indicator Construction

1
Ship Comprehensive Test and Training Base, Naval University of Engineering, Wuhan 430033, China
2
School of Electrical Engineering, Naval University of Engineering, Wuhan 430033, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(16), 7239; https://doi.org/10.3390/s23167239
Submission received: 5 July 2023 / Revised: 3 August 2023 / Accepted: 13 August 2023 / Published: 18 August 2023
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
The prediction of system degradation is very important as it serves as an important basis for the formulation of condition-based maintenance strategies. An effective health indicator (HI) plays a key role in the prediction of system degradation as it enables vital information for critical tasks ranging from fault diagnosis to remaining useful life prediction. To address this issue, a method for monitoring data fusion and health indicator construction based on an autoencoder (AE) and a long short-term memory (LSTM) network is proposed in this study to improve the predictability and effectiveness of health indicators. Firstly, an unsupervised method and overall framework for HI construction is built based on a deep autoencoder and an LSTM neural network. The neural network is trained fully based on the normal operating monitoring data and then the construction error of the AE model is adopted as the health indicator of the system. Secondly, we propose related machine learning techniques for monitoring data processing to overcome the issue of data fusion, such as mutual information for sensor selection and t-distributed stochastic neighbor embedding (T-SNE) for operating condition identification. Thirdly, in order to verify the performance of the proposed method, experiments are conducted based on the CMAPSS dataset and results are compared with algorithms of principal component analysis (PCA) and a vanilla autoencoder model. Result shows that the LSTM-AE model outperforms the PCA and Vanilla-AE model in the metrics of monotonicity, trendability, prognosability, and fitness. Fourthly, in order to analyze the impact of the time step of the LSMT-AE model on HI construction, we construct and analyze the system HI curve under different time steps of 5, 10, 15, 20, and 25 cycles. Finally, the results demonstrate that the proposed method for HI construction can effectively characterize the health state of a system, which is helpful for the development of further failure prognostics and converting the scheduled maintenance into condition-based maintenance.

1. Introduction

In the field of Prognostics and Health Management (PHM) [1], real-time prediction of equipment degradation is crucial as it serves as an important basis for developing Condition-Based Maintenance (CBM) strategies [2]. However, in most cases, the data measured by the monitoring system cannot track the degradation status of the system. Therefore, a model that can predict system degradation or a quantifiable indicator is needed to evaluate the health status of the system, such as the system health indicator (HI). HI reflects the current health status of the system and is an important basis for predicting the trend of system fault development and performance degradation, which can be used to estimate the time when the system reaches failure [3]. Therefore, constructing a reasonable and accurate system HI is an important foundation for developing CBM strategies. For example, HI plays an important role in anomaly detection [4], remaining useful life (RUL) prediction [5], and health state prediction [6], and the effectiveness of the implementation in these fields heavily depends on HI construction [7].
The construction of HI can be divided into two types: model-based methods and data-driven methods, or directly using parameters with clear physical meanings in the system, such as the crack length of mechanical devices and the vibration amplitude of rotating machinery. Among these, model-based methods require in-depth analysis of system characteristics and the establishment of physical failure models based on this, which usually requires rich domain knowledge [8]. Currently, with the development of data collection and storage technology, data-driven methods have become the mainstream method for constructing HI. The core idea is to fuse multiple measurement data to obtain the comprehensive HI of the system.
The evaluation of system health status can be summarized into two questions: what data to use, and how to process and analyze the data. Usually, the most easily obtained data for a system are its monitoring data, which are typically time series data. The monitoring data of the system actually reflect many characteristics of the system, and by deeply mining the system monitoring data, a large amount of relevant information about the system’s health status can be obtained. For example, in [9], several data-driven models have been proposed and applied for diagnostics and prognostics purposes in complex systems based on monitoring data. In [10], a PHM model for the prediction of component failures and the system lifetime is proposed by combining monitoring data, time-to-failure data, and background engineering knowledge of the systems. Likewise, in [11], the degradation index of a rolling bearing is constructed based on the monitoring data.
Regarding the issue of HI construction, a certain number of researchers have adopted traditional machine learning methods such as principal component analysis [12], quantitative programming [13], support vector machines [14], hidden Markov model [15], and Bayesian linear regression [16]. In recent years, with the massive accumulation of data and the development of computer hardware, as an important branch of machine learning, deep learning technology has been rapidly developed. Currently, deep learning techniques have been applied to HI construction in relevant literature; these techniques include vanilla neural networks [17], recurrent neural networks (RNNs) [18], long short-term memory networks [19,20], convolutional neural networks [21,22], and generative adversarial networks [23].
Among the current numerous deep learning methods, an autoencoder is a kind of unsupervised learning method, which is essentially a custom-built neural network architecture. The architecture of the autoencoder consists of a decoder and an encoder. It was first proposed by Rumelhart et al. and is mainly used for data dimension reduction [24]. A deep autoencoder is a multi-layer feedforward neural network with narrow middle and wide ends. It compresses the input data into a low dimensional projection through the encoder, and then reconstructs the output data through the decoder, trying to make them similar enough. Thus, the dimension of input data can be reduced in a hierarchical manner, achieving high-quality data reconstruction. With the development of autoencoder methods, derivative methods such as the denoising autoencoder [25], sparse autoencoder [26], and stacked autoencoder [27] have been invented. In the field of PHM research, the autoencoder has also been gradually applied to anomaly detection [28,29], RUL prediction [30], fault diagnosis [31,32], and health status assessment [33].
For HI construction, we use a basic assumption that when the system’s health status deteriorates, the monitoring data of the system will gradually deviate from the range of normal data, and the more the system performance degrades, the greater the degree of deviation. Therefore, the autoencoder can successfully reconstruct normal data samples after training, but for the abnormal data of the system and the monitoring data after system degradation, an autoencoder model is expected to obtain a large reconstruction error, and the more serious the system degradation, the larger the reconstruction error should be. Based on this, it is reasonable to use autoencoder to build a system HI based on system monitoring data. However, as mentioned earlier, the monitoring data of the system are typically time series data, and methods mentioned above do not fully consider the time series characteristics of monitoring data and the correlation between variables. The LSTM neural network is a kind of recurrent neural network that is specially designed to solve the vanishing gradient problem existing in general RNNs. It can learn the relationship between variables in time series, capture the time series characteristics of monitoring data, and better realize the encoding and decoding of time series data.
By comparison, for time series analysis, sensor selection, also known as feature selection, is one of the important factors that determine the final effect, which plays a vital role in the quantitative analysis of data. At present, for the feature analysis of the time series of the HI, in most cases, the selection of variables is mainly undertaken manually by observing the trend, amplitude, noise, and other characteristics of the variables. This approach has problems such as strong subjectivity, inability to conduct quantitative analysis, and inaccuracy [34]. To solve these problems, a sensor selection method based on mutual information (MI) is proposed to determine the input data of the autoencoder. Mutual information is a measure of the interdependence between two variables, and indicates how much information is shared between two variables [35]. The greater the mutual information between two variables, the stronger their correlation [36]. It has become one of the important methods in feature selection for calculating and analyzing the mutual information of features, judging the correlation between features, and then selecting the features with strong correlations with target variables [37]. However, for feature extraction and analysis of time series, one of the main problems of mutual information-based methods is identifying suitable target variables; that is, the key to mutual information-based methods is to identify appropriate target data. In most cases, there is no intuitive and directly available target information and the existing literature and research are mainly focused on the improvement in the feature extraction algorithm based on mutual information itself. To solve the above problems, we propose a target information extraction method based on PCA dimensionality reduction. On this basis, a mutual information method coupled with target variables is constructed to distinguish and extract applicable monitoring sensor data.
In view of the above situation, we propose an HI construction method based on a deep autoencoder, which is used to assess system health status based on high-dimensional monitoring time series data. The model is composed of a neural network composed of an encoder and a decoder, and LSTM is used to optimize the autoencoder network structure. While extracting the information of monitoring time series data, the method can complete the bidirectional mapping of data between the high-dimensional feature space and the low-dimensional latent space. The autoencoder compresses the time series into a low-dimensional space, and then uses the decoder to reconstruct the compressed data, minimize the reconstruction error, and construct the HI of the system by the degree of the reconstruction error.

1.1. Highlight and Contribution

To summarize, we propose an HI construction method based on LSTM and autoencoder neural networks for the health assessment of complex systems. This method is a data-driven approach that uses system health operating data to train neural networks, and uses the reconstruction error of the decoder as the basis for HI construction. To verify the effectiveness of the method, we conduct HI construction experiments based on the CMAPSS dataset and compared with the most widely used PCA dimensionality reduction method for extracting HI from multiple metrics. In summary, the contributions of this research are as follows:
(1) We propose a construction method for a system HI based on an autoencoder and LSTM neural network. As the system gradually degrades and reaches failure, the monitoring data will gradually deviate from the normal range. Based on this assumption, the LSTM neural network is used to obtain the data series characteristics of the system monitoring data, and the system health index is constructed according to the reconstruction error of the autoencoder.
(2) We design a specific technical process for implementing HI construction based on the above theoretical methods. In this process, we focus on data processing of sensor selection and operating condition identification, and propose a sensor selection method based on mutual information and operating condition identification method based on t-distributed stochastic neighbor embedding.
(3) The effectiveness of the proposed method is verified based on a publicly available aviation engine operation dataset. The experimental results are compared with other methods to verify the feasibility of the proposed method. Through qualitative and quantitative analysis, it is shown that the proposed method does not require the establishment of an accurate mathematical model for the system, and can effectively achieve the health status assessment of the system.

1.2. Organization

The remainder of this paper is organized as follows. In Section 2, we describe the architecture of the proposed system and introduce the theoretical foundation of the methodology. Results of the experiments are presented and discussed in Section 3. Finally, the conclusions and future work are set out in Section 4.

2. Methodology

2.1. Overall Framework

The key to conducting system health assessment is to establish a system health assessment model, that is, to establish a mapping relationship between multidimensional monitoring variables and the system HI. In the actual process, the system will operate alternately under different working conditions, and due to the different operating states of the system, there may be significant differences in the monitoring data of the system. Furthermore, under different operating conditions, the degradation rate and failure threshold of the system also vary, which will increase the difficulty of evaluating the health status. Based on the idea of divide and conquer, a system HI construction method and process, as shown in Figure 1, are proposed for the problem of system health assessment in multiple working conditions. This process includes the following main steps: Step 1: select sensor data for health assessment that contain rich information and can reflect system degradation; Step 2: divide the working conditions based on sensor data and classify the data samples according to different working conditions; Step 3: establish health assessment models based on the autoencoder for different working conditions and calculate the comprehensive HI; Step 4: integrate the comprehensive HI of all monitoring data in chronological order to obtain the system degradation curve in the form of HI sequences.

2.2. Deep Autoencoder

An autoencoder is an unsupervised neural network with training in two stages. It can simplify high-dimensional data into low-dimensional data, and then reconstruct low-dimensional data into high-dimensional input vectors. In the process, the reconstructed high-dimensional vectors are made as close to high-dimensional input vectors as possible. On this basis, a deep autoencoder builds a deep model through superposition. The hidden state of the upper layer is taken as the input of the lower layer. The model is trained with greedy learning layer by layer. Its basic structure is shown in Figure 2. In this case, the encoder is used to transform the high-dimensional data in the input layer into the low-dimensional data in the hidden layer. Meanwhile, the decoder reconstructs the low-dimensional data in the hidden layer into the high-dimensional data in the output layer. Through constant training, the autoencoder structure can deliver reconstructed output that is approximate to the original input. This is mainly intended to denoise the data and extract its truly useful features.
When the data X are input into the encoder, they are transformed by dimension reduction layer by layer until the bottleneck layer is reach, because the number of nodes in the encoder’s neural network decreases layer by layer. Next, the decoder is used to expand such data into X . The difference between the input X and the output X is taken as the optimization objective function to adjust the parameters of the neural network, so as to guarantee their approximation as much as possible. To be specific, the input data are assumed to be X, so that the output of the encoder is:
h ( X ) = σ W ( E N ) X + b ( E N )
where σ is the encoder activation function, and W(EN) and b(EN) are the encoder’s neural network weight matrix and bias vector, respectively. The output of the decoder is:
X = σ W ( D E ) h ( X ) + b ( D E )
where W(DE) and b(DE) are the decoder’s neural network weight matrix and bias vector, respectively. To the autoencoder as a whole, the reconstruction error to be minimized can be expressed as:
D ( X , X ) = X X 2
While searching for the parameters needed to minimize the reconstruction error, the autoencoder can gradually adjust the parameters of the network to achieve the permissible convergence value of the reconstruction error, that is:
W ( E N ) , b ( E N ) , W ( D E ) , b ( D E ) = arg min X X 2

2.3. Long Short-Term Neural Network

Long short-term memory (LSTM) is a special type of a recurrent neural network. With the series data as its input, an RNN recurs in the evolution direction of a series, and has all nodes linked. Normally, it is trained by a back propagation through time (BPTT) algorithm. Nevertheless, the inference rules of a BPTT algorithm make it very easy for an error gradient to cause nonlinear behaviors such as vanishing or exploding gradients after the back propagation of multiple time steps. The LSTM neural network relies on gate control to combine short- and long-term memories, which resolves the problem of a vanishing gradient to some extent [38]. The internal structure of an LSTM neural network unit is given in Figure 3.
An LSTM unit consists of three gates, that is, input, forget, and output. The input and output gates are used to control the input and output activation of the memory unit, respectively. The forget gate can update the state of the unit. The module formed by these gates guarantees that the error travels in the form of a constant instead of a product in the network. This can resolve the problem of a vanishing or exploding gradient. The parameters of the unit are updated by the following equations:
f t = sgmoid W x f · x t + W h f · h t 1 + b f
i t = sgmoid W x i · x t + W h i · h t 1 + b i
C t = tanh W x a · x t + W h a · h t 1 + b a
o t = sgmoid W x o · x t + W h o · h t 1 + b o
C t = f t C t 1 + i t C t
h t = o t tanh C t
where ht−1 and Ct−1 are the output and the state of the unit at the previous time; xt is the input at the current time; f, i, and o are the forget, input, and output gates, respectively; ft is the forget control signal for the state of the unit at the previous time; ft Ct−1 is the information kept at the previous time; C t is the state of the candidate unit at the current time; it is the control signal of C t ; ht−1 is the final output; ot is the output control signal; Wx is the link matrix of input and hidden layers; Wh is the link matrix of output and hidden layers; and is the dot product between matrices.

2.4. Sensor Selection Based on Mutual Information

Mutual information is a measure of the interdependence between two variables, which mainly represents the correlation between two variables. Assuming that the joint probability distribution of two variables X and Y is p x , y , and their marginal probability distributions are p x and p y , respectively, then mutual information M I X , Y is the relative entropy of the joint probability distribution p x , y and the marginal probability distribution p x and p y , and is defined as follows:
M I ( X , Y ) = x y f x , y ( x , y ) log f x , y ( x , y ) f x ( x ) f y ( y ) d x d y
When variables X and Y are completely independent or mutually independent, their mutual information is 0. The greater the mutual information between two variables, the stronger the correlation between them. In data processing, sensors with large mutual information with the target variable should be selected to improve the prediction of the algorithm. At the same time, features with small mutual information should be eliminated to reduce data redundancy, which is the so-called feature selection based on maximum correlation and minimum redundancy.
In the application of the mutual information method, the probability density estimation function can be used to approximate the marginal probability distribution to simplify the above formula:
M I ( X , Y ) = 1 N i = 1 N log f ^ x , y x i , y i f ^ x x i f ^ y y i
where f ^ x x i and f ^ y y i are the joint probability distributions of X and Y , respectively, f ^ x , y x i , y i is the joint probability density estimation function, and N is the number of data samples.

2.5. Operating Condition Identification Based on T-SNE

The idea of T-SNE is to convert the Euclidean distance between data points into conditional probability, and then use the conditional probability to measure the similarity between data points. The greater the similarity, the closer the distance between data points. The advantage of measuring the distance between samples in the form of probability is that in the process of mapping from high-dimensional space to low-dimensional space, although the distance between sample points may vary due to spatial differences, it can still maintain the relative distance between sample points, so that the originally distant sample points still maintain a relatively far distance. Specifically, for the sample point x i , in the original high-dimensional space, its similarity to other sample points such as x j can be expressed by the conditional probabilities p j / i and p i / j of the Gaussian distribution as:
p j / i = exp x i x j 2 / 2 σ i 2 k i exp x i x k 2 / 2 σ i 2
p i / j = exp x j x i 2 / 2 σ j 2 k j exp x j x k 2 / 2 σ j 2
where k is the number of nearest neighbors of each point, σ i and σ j are the Gaussian variances of x i and x j , respectively. Then, the symmetric joint probability distribution of data points in the original space is:
p i j = p j / i + p i / j 2 n
where N is the size of the data. The variance σ in Formulas (13) and (14) is calculated by the binary search of perplexity P e r p P i , which is a manually defined parameter. Its function is to measure the number of points around each point in the effective domain that can form a smooth epidemic, and is defined as follows:
P e r p P i = 2 H P i
where H P i is the Shannon entropy of P i and is defined as follows:
H P i = j P j / i log 2 P j / i
In the low-dimensional space after dimensionality reduction, the mapping points of sample points x i and x j are t i and t j , respectively. Assuming that the symmetric joint probability distribution between the two is represented by q i j , it can be expressed as:
q i j = 1 + t i t j 2 1 k l 1 + t k t l 2 1
where k and l respectively represent any two different points in a low dimensional space, defined by a t-distribution with a degree freedom of 1, which makes the points within the same cluster aggregate more tightly and the points in different clusters become more distant.

2.6. Reconstruction Error as Health Indicator

In the training process of LSTM-AE, we only use monitoring data of the initial state of the system, that is, the health status data of the system. Figure 4 shows the workflow of HI construction based on LSTM-AE, where L is the length of the LSTM time window, which is one of the hyperparameters of the model. For the monitoring data of the system, the time window with a length of L will gradually move to the right and intercept the time series with a length of L (the dimension of the time series depends on the number of sensors used). Then, the time series X is calculated by the decoder and encoder to obtain the reconstructed time series X .
As only normal data are used in the process of training the autoencoder, when the system state gradually deteriorates, the autoencoder model has poor reconstruction ability for the deviated monitoring data as it has not seen any data after state deviation. Therefore, the reconstruction error for the monitoring data of the system will gradually increase [39]. In the method proposed in this paper, the reconstruction error of system monitoring data using LSTM-AE will be used as the HI and is defined as follows:
ε = X X , X = x 1 x 2 x L    X = x 1 x 2 x L
where L is the length of the LSTM time window, X is the input data, and X is the data reconstructed by the autoencoder.

3. Experiments

To verify the effectiveness of the method proposed in this paper, the LSTM-AE-based HI construction method is validated using the C-MAPSS aircraft engine performance degradation simulation dataset published by NASA. The methods of sensor selection based on MI and T-SNE-based operating condition identification are applied. The effectiveness of the method is compared and analyzed with algorithms such as PCA and Vanilla-AE. All simulation experiments were carried out in a PyCharm 2021 (Community Edition) environment on a PC with an AMD Ryzen 7 3700X 8-core CPU and 16 GB memory.

3.1. C-MAPPS Datasets

As shown in Figure 5, a structure diagram of the aero-engine based on C-MAPSS contains several components such as a fan, combustion chamber, high/low pressure booster, turbine, and nozzle [40].
The C-MAPSS dataset consists of four sub-datasets (FD001 to FD004), each corresponding to different working conditions and fault modes. As shown in Table 1, FD001 and FD003 contain one operating condition, while FD002 and FD004 both contain six operating conditions. FD001 and FD002 have one fault mode, while FD003 and FD004 both have two fault modes. The number of engines in the datasets of FD001 to FD004 is 100, 260, 100, and 249, respectively. In these datasets, a total of 21 different sensors are used to measure the working condition, including physical information collected from different locations during the operating cycle of the aircraft engine (such as temperature, pressure, and speed).
During data processing, as shown in Figure 6, the data taken from the first 50 cycles of each engine were regarded as normal data, and only the normal data of each engine were used in the training of an autoencoder.

3.2. Data Preprocessing

Due to the different measurement intervals of different sensors, the data need to be normalized to make all sensor data ranges consistent, so as to achieve a better prediction effect. We adopt Z-score standardization to normalize the input data:
X = X X ¯ σ
where X is the time series data, X ¯ and σ are the average and variance of X , respectively, and X is the normalized monitoring data.
Figure 7 shows the normalized change trend of the original data of FD001, FD002, FD003, and FD004 in the C-MAPSS dataset. From Figure 4, it can be seen that the sensor data of FD001 and FD003 show a clear upward or downward trend, while FD002 and FD004 do not show see a clear pattern. The reason for this problem is that, as mentioned earlier, there are six different operating conditions in the FD002 and FD004 datasets. During the entire life cycle of the engine, the engine operates alternately between different operating conditions. For the situations of FD002 and FD004, the T-SNE method proposed in Section 2.5 should be used to prepare for the identification and division of such alternating operating conditions.

3.3. Sensor Selection

For the feature selection problem of time series based on mutual information, one of the difficulties is to find appropriate target variables [41]. For monitoring data, in most cases, the monotonicity and predictability of sensors cannot be guaranteed, nor can they fully reflect the health and degradation of the system. Therefore, it is not possible to randomly select a sensor signal as the target variable. To address this issue, we propose a time series target variable construction method based on PCA. Firstly, the PCA method is applied to reduce the dimensionality of multidimensional time series variables, and then the obtained principal components are smoothed. After that, the smoothed principal component is taken as the target variable of mutual information. In this way, the obtained target variable can represent the overall performance change of the system.
The mutual information correlation between different sensors and target variables is analyzed by taking engine 1 in the FD001 dataset as an example. It should be noted that some sensor signals in the monitoring data are constant values, which do not change with engine operation, such as s1, s5, s6, s10, s16, s18, and s19. After removing these data, a total of 14 sensor signals are obtained in the C-MAPSS dataset. Further, in order to reduce the redundancy between data, based on the feature selection idea of maximum correlation and minimum redundancy [42], before feature selection, some variables with high redundancy among themselves are removed according to the value of mutual information between variables. After the above selection, Table 2 shows the mutual information values of nine different sensors with the target variable.
It can be seen from Table 3 that the five sensors with high mutual information with the target variable are s7, s9, s11, s12, and s13, corresponding to total pressure at the high-pressure compressor outlet, physical core speed, static pressure at the high-pressure compressor outlet, ratio of fuel flow to Ps30, and corrected fan speed, respectively, as shown in Table 3.

3.4. Operating Condition Identification

The C-MAPSS datasets FD002 and FD004 contain six different operating conditions, which continuously alternate between each operating condition during the working process. In order to build a system HI, real-time and accurate identification of system operating conditions is necessary. We propose and apply the T-SNE method for real-time operating condition recognition of the monitoring data for datasets FD002 and FD004.
We conduct operating condition identification on the FD002 and FD004 datasets based on the five sensor signals selected in Section 3.3. Figure 8 shows the effects of FD002 and FD004 using the T-SNE method for condition identification. From the graph, it can be seen that the data points of different operating conditions are effectively separated into different regions. The T-SNE method can accurately identify the engine operating conditions and preserve the general structure of the monitoring data. Thus, we can divide and normalize the data of FD002 and FD004 in Section 3.2, and eliminate the impact of variable operating conditions.

3.5. Performance Metrics

For HI construction, a set of metrics that can describe the quality of HI is needed, including monotonicity, trendability, and prognosability [43,44]. Among these, monotonicity represents the potential positive or negative trend of features. Assuming that the system does not undergo maintenance, it should be monotonic on the timeline for degraded features. The definition of monotonicity is as follows:
m o n o t o n i c i t y = 1 M j = 1 M k = 1 N j 1 sgn x j ( k + 1 ) x j ( k ) N j 1
where M is the number of HIs, N j is the length of the jth HI, x j ( k + 1 ) and x j ( k ) are the values of the jth HI at time of k + 1 and k , respectively. In PHM, the value of monotonicity is between 0 and 1, and the closer it is to 1, the more monotonic the predictor, which is more beneficial for prediction.
Trendability measures the similarity among the trajectories of the HI in multiple run-to-failure experiments. Generally, Pearson correlation is used to describe whether a set of HI has the same basic trend, and its calculation methods are as follows:
t r e n d a b i l i t y = min j , k corr x j , x k ,     j , k = 1 , , M
where M is the number of HIs, x j and x k are the values of the jth and kth HI respectively, corr is the Pearson correlation function. When the lengths of x j and x k are different, interpolation processing is required. The value of trendability is between 0 and 1. The closer it approaches the value of 1, the better its predictability.
Prognosability reflects the dispersion of HI between faults and normal states, specifically:
p r o g n o s a b i l i t y = exp std j x j N j mean j x j ( 1 ) x j N j ,     j = 1 , 2 , M
where M is the number of HIs, N j is the length of the jth HI, x j is the value of the jth HI, std j is the standard deviation function, mean j is the average function. The value of prognosability is between 0 and 1. The closer it is to 1, the better its predictability.
The fitness is the weighted sum of the three predictive indicators mentioned above:
f i t n e s s = w m × M o n o t o n i c i t y + w t × T r e n d a b i l i t y + w p × P r o g n o s a b i l i t y
where w m , w t , w p are the weight values, which are all equal to 1 in this paper.

3.6. Experimental Setting

To test the performance of the proposed model, in the CMAPSS dataset, normal operating data from FD001, FD002, FD003, and FD004 were selected, with 80% samples from each dataset being used for training, and the remaining 20% being used as testing data. Firstly, the number of neurons in the input layer of the model is determined by the input data. Due to the selection of five sensor signals in Section 3.3, the number of neurons in the input and output layers is five. The LSTM-AE model and Vanilla-AE neural network structure ultimately adopted in this article include an input layer, three encoding layers, an intermediate layer, three decoding layers, and an output layer. It should also be noted that LSTM-AE has an additional preprocessing step compared to Vanilla-AE. The purpose of this preprocessing is to reshape the input data into (samples, time steps, features). The network parameter settings are determined by the grid search method, where the shared bottleneck layer is a LSTM layer with eight nodes. A dropout layer with a dropout rate of 0.5 is employed after the shared bottom layer to prevent the model overfitting. The weights in the network are initialized randomly, and the model is trained using an Adam optimizer with a learning rate of 0.001. The final network structure, input size, and output size of each layer of LSTM-AE and Vanilla-AE, as well as the hyperparameters, are tabulated in Table 4, Table 5 and Table 6.

3.7. Health Indicator Evaluation

To verify the effectiveness of the proposed method, experiments were conducted on the four sets of data in C-MAPSS, FD001, FD002, FD003, and FD004. Among these, FD002 and FD004 contain six different operating conditions. Comparative analysis is conducted with PCA and Vanilla-AE, two widely used HI construction methods. Table 4 shows the performance of the three methods on the four datasets.
Table 7 shows the performance of the three methods on four datasets. We post the best results for each metric and dataset in bold font in Table 7. Firstly, for the two datasets FD002 and FD004, LSTM-AE outperforms PCA and Vanilla-AE in terms of monotonicity, trendability, prognosability, and fitness. For the FD001 and FD003 datasets, LSTM-AE outperforms PCA and Vanilla-AE in terms of monotonicity, but its trendability and prognosability are similar to, but slightly lower than, those of the PCA method, and both are higher than those of the Vanilla-AE method.
From Table 7, it can be seen that the LSTM-AE method has achieved better results than PCA and Vanilla-AE in terms of monotonicity, trendability, prognosability, and fitness. The exceptions are the two metrics of trendability and prognosability, which are slightly lower than those of the PCA method on the FD001 and FD003 datasets. In addition, we also notice that the PCA method performed well on the FD001 and FD003 datasets, but on the FD002 and FD004 datasets, the prognosability metric is almost equal to 0, which indicates that the PCA method has weak data processing ability for variable operating conditions.
In order to more intuitively demonstrate the predictive effects of the three methods, we visualize the HI curves of the engines obtained from the three methods in four datasets, as shown in Figure 9. From Figure 9, we can see that the PCA method has a certain upward trend in HI during the initial operation phase of the system during the long-life cycle, while Vanilla-AE and LSTM-AE are relatively stable in the initial stage. From the perspective of system degradation, the system has a good state in the initial stage, and HI should theoretically remain relatively stable at this time. Therefore, we believe that compared to PCA methods, AE-based methods can better reflect the degradation status of the system. On the other hand, comparing the HI curves of Vanilla-AE and LSTM-AE, it can be seen that due to the time series characteristics of monitoring data learned by LSTM neural networks, their HI curves are more stable and smoother compared to those of Vanilla-AE, thus having better predictive performance.

3.8. The Effect of LSTM Time Step on Predictive Performance

For the LSTM-AE model, time step is one of the most important hyperparameters and directly affects the prediction performance of HI construction. To analyze the impact of different time steps on the HI performance of the system, we compare the performance of the LSTM-AE model under different time steps. Table 8 shows the comparison of time steps of 5, 10, 15, 20, 25, and 30 cycles.
From Table 8, we can see that increasing the time step size can obviously improve the predictive performance of HI. When the time step size is 20 cycles, the proposed method achieves the highest number of optimal indicators on the four datasets FD001, FD002, FD003, and FD004, with item counts of 3, 4, 2, and 1, respectively. However, it should be noted that when using the proposed method for HI construction, the effect may deteriorate if the time step is too long. For example, when the step reaches 25 cycles, the predictive performance of various indicators decreases. The reason for this problem is that when the time step size is too large, the time series information of system degradation contained within a single step range is too much, which means the degradation information learned by the model becomes too vague. This is because a single time step may contain information such as the system operating normally, in addition to degradation or failure. Therefore, special attention should be paid to the hyperparameter of time step.
Figure 10 shows the sensor signals of the LSTM-AE model before and after AE reconstruction at a time step of 20 cycles. The reconstructed data are the s7, s9, s11, s12, and s13 sensor signals of the first engine in the FD001, FD002, FD003, and FD004 datasets. In the figure, the blue curve represents the original signal of the sensor, while the yellow curve represents the reconstructed signal. It can be seen from the figure that the amplitude of the reconstructed signal is basically smaller than that of the original signal.
In an ideal situation, the sensor signal reconstructed by the LSTM-AE model will basically overlap with the original signal in the initial stage. As the number of engine cycles increases, the engine will gradually deviate from its normal operating state, and the reconstructed signal will gradually deviate from the original signal. At this point, the reconstruction error, i.e., HI, will also gradually increase. Next, we analyze the specific situation of data reconstruction. Firstly, taking FD001 as an example, it can be seen from the signal fitting that when the engine is running at around 150 cycles, the reconstructed signal begins to generate a certain deviation, indicating that the engine begins to exhibit abnormalities at around 150 cycles and the engine’s performance begins to deteriorate. In addition, during the operation of the engine, due to different types of fault inputs in the simulation dataset, the system performance may deteriorate, causing some sensors’ signals to still be in normal working conditions, while others may experience significant deviations. For example, the s12 signal of FD002 has always had a good fitting effect during the process. We also notice that there is a certain deviation between the reconstructed signal of AE and the original signal for sensor signal s9. The reason for this phenomenon may be that S9 represents the physical core speed of the system, which fluctuates greatly and frequently during the engine operating cycle, and, at this time, the model’s reconstruction ability is limited.

4. Conclusions

In this paper, we propose an HI construction method based on an autoencoder and a long short-term memory neural network. The proposed method, LSTM-AE, is an unsupervised method as the model is fully trained based on the system’s normal operating monitoring data, and the abnormal data are not needed. The construction error of the AE model is adopted as the system HI. We built the overall framework for the monitoring data fusion and HI construction based on LSTM-AE and the data processing workflow, including techniques such as mutual information for sensor selection and t-distributed stochastic neighbor embedding for operating condition identification. Experiments were conducted based on the CMAPSS dataset to test the utility of the LSTM-AE on the construction of system HI. The LSTM-AE method showed better algorithm performance compared to PCA and Vanilla-AE models in the metrics of monotonicity, trendability, prognosability, and fitness. LSTM-AE proved capable of constructing an effective HI that can characterize the health state of the system. This ensures that the constructed HI is conducive to future analysis of failure prognostics, as well as RUL prediction, etc.
Regarding future work, we plan to apply our proposal in hybrid autoencoder architectures, such as deep adversarial autoencoders (AAEs), as AAEs can learn better representations and extract more robust features. Another direction of future work is to develop new indicators for prognostics; for example, instead of directly constructing an HI for prognostics, exploring indicators of feature importance that are interrelated for prognostics features might be promising.

Author Contributions

Methodology, L.H.; Software, X.P.; Formal analysis, X.P.; Investigation, Y.L.; Resources, Y.L.; Writing—original draft, L.H.; Writing—review & editing, L.H.; Supervision, L.G.; Funding acquisition, L.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data sharing is not applicable to this article as no new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Hu, Y.; Miao, X.; Si, Y.; Pan, E.; Zio, E. Prognostics and health management: A review from the perspectives of design, development and decision. Reliab. Eng. Syst. Saf. 2022, 217, 108063. [Google Scholar] [CrossRef]
  2. Quatrini, E.; Costantino, F.; Di Gravio, G.; Patriarca, R. Condition-Based Maintenance—An Extensive Literature Review. Machines 2020, 8, 31. [Google Scholar] [CrossRef]
  3. Tsui, K.L.; Chen, N.; Zhou, Q.; Hai, Y.; Wang, W.; Wu, S. Prognostics and Health Management: A Review on Data Driven Approaches. Math. Probl. Eng. 2015, 2015, 793161. [Google Scholar] [CrossRef]
  4. Bataineh, A.A.; Mairaj, A.; Kaur, D. Autoencoder based Semi-Supervised Anomaly Detection in Turbofan Engines. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 41–47. [Google Scholar] [CrossRef]
  5. Nguyen, K.T.P.; Medjaher, K. An automated health indicator construction methodology for prognostics based on multi-criteria optimization. ISA Trans. 2021, 113, 81–96. [Google Scholar] [CrossRef]
  6. Wen, P.; Zhao, S.; Chen, S.; Li, Y. A generalized remaining useful life prediction method for complex systems based on composite health indicator. Reliab. Eng. Syst. Saf. 2021, 205, 107241. [Google Scholar] [CrossRef]
  7. Xiao, L.; Tang, J.; Zhang, X.; Bechhoefer, E.; Ding, S. Remaining useful life prediction based on intentional noise injection and feature reconstruction. Reliab. Eng. Syst. Saf. 2021, 215, 107871. [Google Scholar] [CrossRef]
  8. Ding, D.; Han, Q.L.; Wang, Z.; Ge, X. A Survey on Model-Based Distributed Control and Filtering for Industrial Cyber-Physical Systems. IEEE Trans. Ind. Inform. 2019, 15, 2483–2499. [Google Scholar] [CrossRef]
  9. Cofre-Martel, S.; Droguett, E.L.; Modarres, M. Big Machinery Data Preprocessing Methodology for Data-Driven Models in Prognostics and Health Management. Sensors 2021, 21, 6841. [Google Scholar] [CrossRef]
  10. Rocchetta, R.; Gao, Q.; Mavroeidis, D.; Petkovic, M. A robust model selection framework for fault detection and system health monitoring with limited failure examples: Heterogeneous data fusion and formal sensitivity bounds. Eng. Appl. Artif. Intell. 2022, 114, 105140. [Google Scholar] [CrossRef]
  11. Zhu, W.; Ni, G.; Cao, Y.; Wang, H. Research on a rolling bearing health monitoring algorithm oriented to industrial big data. Measurement 2021, 185, 110044. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Xin, Y.; Liu, Z.; Chi, M.; Ma, G. Health status assessment and remaining useful life prediction of aero-engine based on BiGRU and MMoE. Reliab. Eng. Syst. Saf. 2022, 220, 108263. [Google Scholar] [CrossRef]
  13. Guo, L.; Yu, Y.; Duan, A.; Gao, H.; Zhang, J. An unsupervised feature learning based health indicator construction method for performance assessment of machines. Mech. Syst. Signal Process. 2022, 167, 108573. [Google Scholar] [CrossRef]
  14. Islam, M.M.M.; Prosvirin, A.E.; Kim, J. Data-driven prognostic scheme for rolling-element bearings using a new health index and variants of least-square support vector machines. Mech. Syst. Signal Process. 2021, 160, 107853. [Google Scholar] [CrossRef]
  15. Wang, S.; Chen, J.; Wang, H.; Zhang, D. Degradation evaluation of slewing bearing using HMM and improved GRU. Measurement 2019, 146, 385–395. [Google Scholar] [CrossRef]
  16. Lyu, Z.; Wang, G.; Tan, C. A novel Bayesian multivariate linear regression model for online state-of-health estimation of Lithium-ion battery using multiple health indicators. Microelectron. Reliab. 2022, 131, 114500. [Google Scholar] [CrossRef]
  17. Wu, Y.; Yuan, M.; Dong, S.; Lin, L.; Liu, Y. Remaining useful life estimation of engineered systems using vanilla LSTM neural networks. Neurocomputing 2018, 275, 167–179. [Google Scholar] [CrossRef]
  18. Chen, L.; Xu, G.; Zhang, S.; Yan, W.; Wu, Q. Health indicator construction of machinery based on end-to-end trainable convolution recurrent neural networks. J. Manuf. Syst. 2020, 54, 1–11. [Google Scholar] [CrossRef]
  19. Cheng, Y.; Hu, K.; Wu, J.; Zhu, H.; Shao, X. A convolutional neural network based degradation indicator construction and health prognosis using bidirectional long short-term memory network for rolling bearings. Adv. Eng. Inform. 2021, 48, 101247. [Google Scholar] [CrossRef]
  20. Ye, Z.; Yu, J. Health condition monitoring of machines based on long short-term memory convolutional autoencoder. Appl. Soft Comput. 2021, 107, 107379. [Google Scholar] [CrossRef]
  21. Li, P.; Zhang, Z.; Grosu, R.; Deng, Z.; Hou, J.; Rong, Y.; Wu, R. An end-to-end neural network framework for state-of-health estimation and remaining useful life prediction of electric vehicle lithium batteries. Renew. Sustain. Energy Rev. 2022, 156, 111843. [Google Scholar] [CrossRef]
  22. Marei, M.; Zaatari, S.E.; Li, W. Transfer learning enabled convolutional neural networks for estimating health state of cutting tools. Robot. Comput.-Integr. Manuf. 2021, 71, 102145. [Google Scholar] [CrossRef]
  23. Behera, S.; Misra, R. Generative adversarial networks based remaining useful life estimation for IIoT. Comput. Electr. Eng. 2021, 92, 107195. [Google Scholar] [CrossRef]
  24. Li, P.; Pei, Y.; Li, J. A comprehensive survey on design and application of autoencoder in deep learning. Appl. Soft Comput. 2023, 138, 110176. [Google Scholar] [CrossRef]
  25. Vincent, P.; Larochelle, H.; Bengio, Y.; Manzagol, P.A. Extracting and Composing Robust Features with Denoising Autoencoders. In Proceedings of the Twenty-Fifth International Conference (ICML 2008) on Machine Learning, Helsinki, Finland, 5–9 June 2008. [Google Scholar]
  26. She, D.; Jia, M.; Pecht, M.G. Sparse auto-encoder with regularization method for health indicator construction and remaining useful life prediction of rolling bearing. Meas. Sci. Technol. 2020, 31, 105005. [Google Scholar] [CrossRef]
  27. Xu, F.; Huang, Z.; Yang, F.; Wang, D.; Tsui, K.L. Constructing a health indicator for roller bearings by using a stacked auto-encoder with an exponential function to eliminate concussion. Appl. Soft Comput. 2020, 89, 106119. [Google Scholar] [CrossRef]
  28. Jakubowski, J.; Stanisz, P.; Bobek, S.; Nalepa, G.J. Anomaly Detection in Asset Degradation Process Using Variational Autoencoder and Explanations. Sensors 2022, 22, 291. [Google Scholar] [CrossRef] [PubMed]
  29. Nguyen, H.D.; Tran, K.P.; Thomassey, S.; Hamad, M. Forecasting and Anomaly Detection approaches using LSTM and LSTM Autoencoder techniques with the applications in supply chain management. Int. J. Inf. Manag. 2021, 57, 102282. [Google Scholar] [CrossRef]
  30. Han, T.; Pang, J.; Tan, A.C.C. Remaining useful life prediction of bearing based on stacked autoencoder and recurrent neural network. J. Manuf. Syst. 2021, 61, 576–591. [Google Scholar] [CrossRef]
  31. Yang, Z.; Xu, B.; Luo, W.; Chen, F. Autoencoder-based representation learning and its application in intelligent fault diagnosis: A review. Measurement 2022, 189, 110460. [Google Scholar] [CrossRef]
  32. Wu, X.; Zhang, Y.; Cheng, C.; Peng, Z. A hybrid classification autoencoder for semi-supervised fault diagnosis in rotating machinery. Mech. Syst. Signal Process. 2021, 149, 107327. [Google Scholar] [CrossRef]
  33. Lee, H.; Lim, H.J.; Chattopadhyay, A. Data-driven system health monitoring technique using autoencoder for the safety management of commercial aircraft. Neural Comput. Appl. 2021, 33, 3235–3250. [Google Scholar] [CrossRef]
  34. Li, J.; Ren, W.; Han, M. Mutual Information Variational Autoencoders and Its Application to Feature Extraction of Multivariate Time Series. Int. J. Pattern Recognit. Artif. Intell. 2022, 36, 2255005. [Google Scholar] [CrossRef]
  35. Stögbauer, H.; Grassberger, P.; Kraskov, A. Estimating mutual information. Phys. Rev. E 2004, 69, 066138. [Google Scholar]
  36. Pascoal, C.; Oliveira, M.R.; Pacheco, A.; Valadas, R. Theoretical evaluation of feature selection methods based on mutual information. Neurocomputing 2017, 226, 168–181. [Google Scholar] [CrossRef]
  37. Cheng, J.; Sun, J.; Yao, K.; Xu, M.; Cao, Y. A variable selection method based on mutual information and variance inflation factor. Spectrochim. Acta Part A Mol. Biomol. Spectrosc. 2022, 268, 120652. [Google Scholar] [CrossRef] [PubMed]
  38. Zaremba, W.; Sutskever, I.; Vinyals, O. Recurrent neural network regularization. arXiv 2014, arXiv:1409.2329. [Google Scholar]
  39. de Pater, I.; Mitici, M. Developing health indicators and RUL prognostics for systems with few failure instances and varying operating conditions using a LSTM autoencoder. Eng. Appl. Artif. Intell. 2023, 117, 105582. [Google Scholar] [CrossRef]
  40. Saxena, A.; Goebel, K.; Simon, D.; Eklund, N. Damage propagation modeling for aircraft engine run-to-failure simulation. In Proceedings of the 2008 International Conference on Prognostics and Health Management, Denver, CO, USA, 6–9 October 2008; IEEE: Piscataway, NJ, USA, 2008; pp. 1–9. [Google Scholar]
  41. Khaire, U.M.; Dhanalakshmi, R. Stability of feature selection algorithm: A review. J. King Saud Univ.-Comput. Inf. Sci. 2022, 34, 1060–1073. [Google Scholar] [CrossRef]
  42. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  43. Liu, C.; Sun, J.; Liu, H.; Lei, S.; Hu, X. Complex engineered system health indexes extraction using low frequency raw time-series data based on deep learning methods. Measurement 2020, 161, 107890. [Google Scholar] [CrossRef]
  44. Baptista, M.L.; Goebel, K.; Henriques, E.M.P. Relation between prognostics predictor evaluation metrics and local interpretability SHAP values. Artif. Intell. 2022, 306, 103667. [Google Scholar] [CrossRef]
Figure 1. Autoencoder-based health status assessment process (The different colors mean different working conditions, being separated by the algorithm).
Figure 1. Autoencoder-based health status assessment process (The different colors mean different working conditions, being separated by the algorithm).
Sensors 23 07239 g001
Figure 2. Deep autoencoder.
Figure 2. Deep autoencoder.
Sensors 23 07239 g002
Figure 3. Structure of a long short-term memory network unit.
Figure 3. Structure of a long short-term memory network unit.
Sensors 23 07239 g003
Figure 4. Workflow of HI construction based on LSTM-AE.
Figure 4. Workflow of HI construction based on LSTM-AE.
Sensors 23 07239 g004
Figure 5. Structure diagram of a commercial modular aero-propulsion system.
Figure 5. Structure diagram of a commercial modular aero-propulsion system.
Sensors 23 07239 g005
Figure 6. Data partition in the full lifecycle of engine.
Figure 6. Data partition in the full lifecycle of engine.
Sensors 23 07239 g006
Figure 7. Normalized C-MAPSS sensor monitoring data from FD001 to FD004.
Figure 7. Normalized C-MAPSS sensor monitoring data from FD001 to FD004.
Sensors 23 07239 g007
Figure 8. Clustering results of six operating conditions, with a sample size of 5000 cycles. Left-hand plot is FD002, and the right-hand plot is FD004.
Figure 8. Clustering results of six operating conditions, with a sample size of 5000 cycles. Left-hand plot is FD002, and the right-hand plot is FD004.
Sensors 23 07239 g008
Figure 9. HIs built on PCA, Vanilla-AE, and LSTM-AE methods. (a) The first row is the PCA method applied to FD001, FD002, FD003, and FD004 respectively. (b) The second row is the Vanilla-AE method applied to FD001, FD002, FD003, and FD004, respectively. (c) The third row is the LSTM -AE method applied to FD001, FD002, FD003, and FD004, respectively.
Figure 9. HIs built on PCA, Vanilla-AE, and LSTM-AE methods. (a) The first row is the PCA method applied to FD001, FD002, FD003, and FD004 respectively. (b) The second row is the Vanilla-AE method applied to FD001, FD002, FD003, and FD004, respectively. (c) The third row is the LSTM -AE method applied to FD001, FD002, FD003, and FD004, respectively.
Sensors 23 07239 g009
Figure 10. Raw and reconstructed sensor signals with LSTM-AE—s7, s9, s11, s12, and s13 of the CMAPSS dataset. The first row is the LSTM-AE method applied to FD001. The second row is the LSTM-AE method applied to FD002. The third row is the LSTM-AE method applied to FD003. The fourth row is the LSTM-AE method applied to FD004.
Figure 10. Raw and reconstructed sensor signals with LSTM-AE—s7, s9, s11, s12, and s13 of the CMAPSS dataset. The first row is the LSTM-AE method applied to FD001. The second row is the LSTM-AE method applied to FD002. The third row is the LSTM-AE method applied to FD003. The fourth row is the LSTM-AE method applied to FD004.
Sensors 23 07239 g010
Table 1. Basic information of C-MAPSS dataset.
Table 1. Basic information of C-MAPSS dataset.
DatasetsFD001FD002FD003FD004
Operating condition1616
Fault mode1122
Number of engines100260100249
Table 2. Mutual information between sensor signal and target variable.
Table 2. Mutual information between sensor signal and target variable.
Sensor NumberValue of MIUnits
s20.570bit
s70.697bit
s80.597bit
s91.000bit
s110.951bit
s120.705bit
s130.682bit
s170.485bit
s200.506bit
Table 3. Selected engine sensors for HI construction.
Table 3. Selected engine sensors for HI construction.
Sensor NumberDescriptionUnits
s7Total pressure at High-Pressure Compressor outletkpa
s9Physical core speedrpm
s11Static pressure at High-Pressure Compressor outletkpa
s12Ratio of fuel flow to Ps30
s13Corrected fan speedrpm
Table 4. Neural network structure of LSTM-AE.
Table 4. Neural network structure of LSTM-AE.
ModelNeural NetworkNeuron Number
Input layerLSTM5
EncoderLSTM128
EncoderLSTM64
EncoderLSTM32
Middle layerLSTM8
DecoderLSTM32
DecoderLSTM64
DecoderLSTM128
Output layerDense5
Table 5. Neural network structure of Vanilla-AE.
Table 5. Neural network structure of Vanilla-AE.
ModelNeural NetworkNeuron Number
Input layerDense5
EncoderDense128
EncoderDense64
EncoderDense32
Middle layerDense8
DecoderDense32
DecoderDense64
DecoderDense128
Output layerDense5
Table 6. Hyperparameters of neural networks.
Table 6. Hyperparameters of neural networks.
HyperparametersValue
Batch size128
Learning rate0.001
OptimizerAdam
LossMae
Epochs200
Dropout rate0.5
Table 7. Performance results of PCA, Vanilla-AE, and LSTM-AE for FD001, FD002, FD003, and FD004 (best results for each metric and dataset are presented in bold font).
Table 7. Performance results of PCA, Vanilla-AE, and LSTM-AE for FD001, FD002, FD003, and FD004 (best results for each metric and dataset are presented in bold font).
PredictorFD001FD002
MonoTrenProgFitnessMonoTrenProgFitness
PCA0.3350.8920.8552.0820.1940.00040.7510.9454
Vanilla-AE0.2460.7960.8181.8600.1420.4190.6171.178
LSTM-AE0.4490.890 0.8412.1800.4460.8610.8222.129
PredictorFD003FD004
MonoTrenProgFitnessMonoTrenProgFitness
PCA0.3300.8050.7011.8360.1690.00020.5810.7502
Vanilla-AE0.2150.640 0.4961.3510.150.3830.5291.062
LSTM-AE0.4190.7180.6051.7420.2720.3870.6411.300
monotonicity—mono, trendability—tren–, prognosability—prog, and fitness.
Table 8. Results of the performance of Vanilla-AE for time step of 5, 10, 15, 20, and 25 cycles (best results for each metric and dataset are presented in bold font).
Table 8. Results of the performance of Vanilla-AE for time step of 5, 10, 15, 20, and 25 cycles (best results for each metric and dataset are presented in bold font).
Time step (cycle)FD001FD002
MonoTrenProgFitnessMonoTrenProgFitness
50.298 0.880 0.8522.030 0.184 0.356 0.537 1.077
100.313 0.801 0.739 1.853 0.296 0.776 0.657 1.729
150.406 0.709 0.795 1.910 0.358 0.791 0.648 1.797
200.4490.8900.841 2.1800.4460.8610.8222.129
250.4480.8860.8172.151 0.3580.7720.7291.859
Time step (cycle)FD003FD004
MonoTrenProgFitnessMonoTrenProgFitness
50.329 0.845 0.561 1.735 0.216 0.371 0.549 1.136
100.383 0.860 0.330 1.573 0.284 0.454 0.610 1.348
150.418 0.707 0.545 1.670 0.316 0.4060.614 1.336
200.419 0.718 0.6051.7420.272 0.387 0.6411.300
250.4480.8200.4411.709 0.3230.1880.4530.964
monotonicity—mono–, trendability—tren–, prognosability—prog, and fitness.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, L.; Pan, X.; Liu, Y.; Gong, L. An Unsupervised Machine Learning Approach for Monitoring Data Fusion and Health Indicator Construction. Sensors 2023, 23, 7239. https://doi.org/10.3390/s23167239

AMA Style

Huang L, Pan X, Liu Y, Gong L. An Unsupervised Machine Learning Approach for Monitoring Data Fusion and Health Indicator Construction. Sensors. 2023; 23(16):7239. https://doi.org/10.3390/s23167239

Chicago/Turabian Style

Huang, Lin, Xin Pan, Yajie Liu, and Li Gong. 2023. "An Unsupervised Machine Learning Approach for Monitoring Data Fusion and Health Indicator Construction" Sensors 23, no. 16: 7239. https://doi.org/10.3390/s23167239

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop