1. Introduction
In the field of Prognostics and Health Management (PHM) [
1], real-time prediction of equipment degradation is crucial as it serves as an important basis for developing Condition Based Maintenance (CBM) strategies [
2]. However, in most cases, the data measured by the monitoring system cannot track the degradation status of the system. Therefore, a model that can predict system degradation or a quantifiable indicator is needed to evaluate the health status of the system, such as the system health indicator (HI). HI reflects the current health status of the system and is an important basis for predicting the trend of system fault development and performance degradation, which can be used to estimate the time when the system reaches failure [
3]. Therefore, constructing a reasonable and accurate system HI is an important foundation for developing CBM strategies. For example, HI plays an important role in anomaly detection [
4], remaining useful life (RUL) prediction [
5], and health state prediction [
6], and the effectiveness of the implementation in these fields heavily depends on HI construction [
7].
The construction of HI can be divided into two types: model-based methods and data-driven methods, or directly using parameters with clear physical meanings in the system, such as the crack length of mechanical devices and the vibration amplitude of rotating machinery. Among them, model-based methods require in-depth analysis of system characteristics and the establishment of physical failure models based on this, which usually requires rich domain knowledge[
8]. Currently, with the development of data collection and storage technology, data-driven methods have become the mainstream method for constructing HI. The core idea is to fuse multiple measurement data to obtain the comprehensive HI of the system.
The evaluation of system health status can be summarized into two questions: what data to use, and how to process and analyze the data. Usually, the most easily obtained data for a system are its monitoring data, which are time series data typically. The monitoring data of the system actually reflect many characteristics of the system, and by deeply mining the system monitoring data, a large amount of relevant information about the system’s health status can be obtained. For example, in [
9], several data-driven models have been proposed and applied for diagnostics and prognostics purposes in complex systems based on monitoring data. In [
10], a PHM model for the prediction components failures and system lifetime is proposed by combining monitoring data, time-to-failure data and background engineering knowledge of the systems. Likewise, in [
11], the degradation index of rolling bearing is constructed based on the monitoring data.
Regarding the issue of HI construction, a certain number of researchers have adopted traditional machine learning methods such as principal component analysis [
12], quantitative programming [
13], support vector machines [
14], hidden Markov model [
15], and Bayesian linear regression [
16]. In recent years, with the massive accumulation of data and the development of computer hardware, as an important branch of machine learning, deep learning technology has been rapidly developed. Currently, deep learning techniques have been applied to HI construction in relevant literature, including: vanilla neural network [
17], recurrent neural network (RNN) [
18], long short term memory network [19, 20], convolutional neural network [21, 22], and generative adversarial network [
23] et al.
Among the current numerous deep learning methods, autoencoder is a kind of unsupervised learning method, which is essentially a custom-built neural network architecture. The architecture of autoencoder consists of decoder and encoder. It was first proposed by Rumelhart et al. and is mainly used for data dimension reduction [
24]. Deep autoencoder is a multi-layer feedforward neural network with narrow middle and wide ends. It compresses the input data into a low dimensional projection through the encoder, and then reconstructs the output data through the decoder, trying to make them similar enough. Thus, the dimension of input data can be reduced in a hierarchical manner, achieving high-quality data reconstruction. With the development of autoencoder methods, derivative methods such as denoising autoencoder [
25], sparse autoencoder [
26] and stacked autoencoder [
27], et al. have been invented. In the field of PHM research, autoencoder has also been gradually applied to anomaly detection [28, 29], RUL prediction [
30] , fault diagnosis [31, 32] and health status assessment [
33].
For HI construction, we have a basic assumption that when the system’s health status deteriorates, the monitoring data of the system will gradually deviate from the range of normal data, and the more system performance degrades, the greater the degree of deviation. Therefore, the autoencoder can successfully reconstruct normal data samples after training, but for the abnormal data of the system and the monitoring data after system degradation, an autoencoder model is expected to get a large reconstruction error, and the more serious the system degradation, the larger the reconstruction error should be. Based on this, it is reasonable to use autoencoder to build system HI based on system monitoring data. However, as mentioned earlier, the monitoring data of the system are typical time series data, and methods mentioned above do not fully consider the time series characteristics of monitoring data and the correlation between variables. LSTM neural network is a kind of recurrent neural network, which is specially designed to solve the vanishing gradient problem existing in general RNN. It can learn the relationship between variables in time series, capture the time series characteristics of monitoring data, and better realize the encoding and decoding of time series data.
On the other hand, for time series analysis, sensor selection, also known as feature selection, is one of the important factors that determine the final effect, which plays a vital role in the quantitative analysis of data. At present, for the feature analysis of the time series of HI, in most cases, the selection of variables is mainly done manually by observing the trend, amplitude, noise and other characteristics of the variables, which has problems such as strong subjectivity, inability to quantitatively analyze, and inaccuracy[
34]. To solve this problem, a sensor selection method based on mutual information (MI) is proposed to determine the input data of Autoencoder. Mutual information is a measure of the interdependence between two variables, which indicates how much information is shared between two variables[
35]. The greater the mutual information between two variables, the stronger the correlation they are [
36]. It has become one of the important methods for feature selection to calculate and analyze the mutual information of features, judge the correlation between features, and then select the features with strong correlation with target variables[
37]. However, for feature extraction and analysis of time series, one of the main problems of mutual information-based methods is to seek for suitable target variables, that is, the key of mutual information- based methods is to identify an appropriate target data. In most cases, there is no intuitive and directly available target information and the existing literature and researches are mainly focus on the improvement of feature extraction algorithm based on mutual information itself. To solve the above problems, we propose a target information extraction method based on PCA dimensionality reduction. On this basis, a mutual information method coupled with target variables is constructed to distinguish and extract applicable monitoring sensor data.
In view of above situation, we propose an HI construction method based on deep autoencoder, which is used to assess system health status based on high-dimensional monitoring time series data. The model is composed of a neural network composed of an encoder and a decoder, and LSTM is used to optimize the autoencoder network structure. While extracting the information of monitoring time series data, the method can complete the bidirectional mapping of data between high dimensional feature space and low dimensional latent space. The autoencoder compresses the time series into a low dimensional space, and then uses the decoder to reconstruct the compressed data, minimize the reconstruction error, and construct the HI of the system by the degree of the reconstruction error.
To sum up, we propose an HI construction method based on LSTM and autoencoder neural networks for the health assessment of complex system. This method is a data-driven approach that uses system health operating data to train neural networks, and uses the reconstruction error of the decoder as the basis for HI construction. To verify the effectiveness of the method, we conduct HI construction experiments based on the CMAPSS dataset and compared with the most widely used PCA dimensionality reduction method for extracting HI from multiple metrics. In summary, the contributions of this research are as follows:
(1) We propose a construction method of system HI based on Autoencoder and LSTM neural network. As the system gradually degrades and reaches failure, the monitoring data will gradually deviate from the normal range. Based on this assumption, the LSTM neural network is used to obtain the data series characteristics of the system monitoring data, and the system health index is constructed according to the reconstruction error of the Autoencoder.
(2) We design a specific technical process for implementing HI construction based on the above theoretical methods. In this process, we focus on data processing of sensor selection and operating condition identification, and propose a sensor selection method based on mutual information and operating condition identification method based on t-distributed stochastic neighbor embedding.
(3) The effectiveness of the proposed method is verified based on a publicly available aviation engine operation dataset. The experimental results are compared with other methods to verify the feasibility of the proposed method. Through qualitative and quantitative analysis, it is shown that the proposed method does not require the establishment of an accurate mathematical model for the system, and can effectively achieve the health status assessment of the system.
The remainder of this paper is organized as follows. In
Section 2, we describe the architecture of the proposed system and introduce the theoretical foundation of the methodology. Results of the experiments are presented and discussed in
Section 3. Finally, the conclusions and future work are set out in
Section 4.
Figure 1.
Autoencoder based health status assessment process.
Figure 1.
Autoencoder based health status assessment process.
Figure 2.
Deep autoencoder.
Figure 2.
Deep autoencoder.
Figure 3.
Structure of a long short-term memory network unit.
Figure 3.
Structure of a long short-term memory network unit.
Figure 5.
Structure diagram of a commercial modular aero-propulsion system.
Figure 5.
Structure diagram of a commercial modular aero-propulsion system.
Figure 6.
Data partition in the full lifecycle of engine.
Figure 6.
Data partition in the full lifecycle of engine.
Figure 7.
Normalized C-MAPSS sensor monitoring data from FD001 to FD004.
Figure 7.
Normalized C-MAPSS sensor monitoring data from FD001 to FD004.
Figure 8.
Trends in principal component of FD001 and FD003.
Figure 8.
Trends in principal component of FD001 and FD003.
Figure 9.
Clustering results of 6 operating conditions, with a sample size of 5000 cycles. Left-hand plot is FD002, and the right-hand plot is FD004.
Figure 9.
Clustering results of 6 operating conditions, with a sample size of 5000 cycles. Left-hand plot is FD002, and the right-hand plot is FD004.
Figure 10.
FD002 and FD004 monitoring data after condition identification and normalization.
Figure 10.
FD002 and FD004 monitoring data after condition identification and normalization.
Figure 11.
HIs built on (a) PCA, (b) Vanilla-AE and (c) LSTM-AE residuals for four test degradation profiles belonging to dataset FD001, FD002, FD003 and FD004.
Figure 11.
HIs built on (a) PCA, (b) Vanilla-AE and (c) LSTM-AE residuals for four test degradation profiles belonging to dataset FD001, FD002, FD003 and FD004.
Figure 12.
Actual and reconstructed sensor measurements with the LSTM-AE —s7、s9、s11、s12 and s13 of FD001, FD002, FD003 and FD004.
Figure 12.
Actual and reconstructed sensor measurements with the LSTM-AE —s7、s9、s11、s12 and s13 of FD001, FD002, FD003 and FD004.
Table 1.
Basic information of C-MAPSS dataset.
Table 1.
Basic information of C-MAPSS dataset.
Datasets |
FD001 |
FD002 |
FD003 |
FD004 |
Operating condition |
1 |
6 |
1 |
6 |
Fault mode |
1 |
1 |
2 |
2 |
Number of engines |
100 |
260 |
100 |
249 |
Table 2.
Mutual information between sensor signal and target variable.
Table 2.
Mutual information between sensor signal and target variable.
Sensor number |
Value of MI |
|
s2 |
0.570 |
|
s7 |
0.697 |
|
s8 |
0.597 |
|
s9 |
1.000 |
|
s11 |
0.951 |
|
s12 |
0.705 |
|
s13 |
0.682 |
|
s17 |
0.485 |
|
s20 |
0.506 |
|
Table 3.
Selected engine sensor for HI construction.
Table 3.
Selected engine sensor for HI construction.
Sensor number |
Description |
Units |
s7
|
Total pressure at High-Pressure Compressor outlet |
kpa |
s9
|
Physical core speed |
rpm |
s11
|
Static pressure at High-Pressure Compressor outlet |
kpa |
s12
|
Ratio of fuel flow to Ps30 |
— |
s13
|
Corrected fan speed |
rpm |
Table 4.
Results of performance of PCA, Vanilla-AE and LSTM-AE for FD001, FD002, FD003 and FD004.
Table 4.
Results of performance of PCA, Vanilla-AE and LSTM-AE for FD001, FD002, FD003 and FD004.
Predictor |
FD001 |
|
FD002 |
Mono |
Tren |
Prog |
Fitness |
|
Mono |
Tren |
Prog |
Fitness |
PCA |
0.335 |
0.892 |
0.855 |
2.082 |
|
0.194 |
0.0004 |
0.751 |
0.9454 |
Vanilla-AE |
0.246 |
0.796 |
0.818 |
1.860 |
|
0.142 |
0.419 |
0.617 |
1.178 |
LSTM-AE |
0.449 |
0.890 |
0.841 |
2.180 |
|
0.446 |
0.861 |
0.822 |
2.129 |
Predictor |
FD003 |
|
FD004 |
Mono |
Tren |
Prog |
Fitness |
|
Mono |
Tren |
Prog |
Fitness |
PCA |
0.330 |
0.805 |
0.701 |
1.836 |
|
0.169 |
0.0002 |
0.581 |
0.7502 |
Vanilla-AE |
0.215 |
0.640 |
0.496 |
1.351 |
|
0.15 |
0.383 |
0.529 |
1.062 |
LSTM-AE |
0.419 |
0.718 |
0.605 |
1.742 |
|
0.272 |
0.387 |
0.641 |
1.300 |
Table 7.
Results of performance of Vanilla-AE for time step 5, 10, 15,20 and 25cycles.
Table 7.
Results of performance of Vanilla-AE for time step 5, 10, 15,20 and 25cycles.
Time step (cycle) |
FD001 |
|
FD002 |
Mono |
Tren |
Prog |
Fitness |
|
Mono |
Tren |
Prog |
Fitness |
5 |
0.298 |
0.880 |
0.852 |
2.030 |
|
0.184 |
0.356 |
0.537 |
1.077 |
10 |
0.313 |
0.801 |
0.739 |
1.853 |
|
0.296 |
0.776 |
0.657 |
1.729 |
15 |
0.406 |
0.709 |
0.795 |
1.910 |
|
0.358 |
0.791 |
0.648 |
1.797 |
20 |
0.449 |
0.890 |
0.841 |
2.180 |
|
0.446 |
0.861 |
0.822 |
2.129 |
25 |
0.448 |
0.886 |
0.817 |
2.151 |
|
0.358 |
0.772 |
0.729 |
1.859 |
Time step (cycle) |
FD003 |
|
FD004 |
Mono |
Tren |
Prog |
Fitness |
|
Mono |
Tren |
Prog |
Fitness |
5 |
0.329 |
0.845 |
0.561 |
1.735 |
|
0.216 |
0.371 |
0.549 |
1.136 |
10 |
0.383 |
0.860 |
0.330 |
1.573 |
|
0.284 |
0.454 |
0.610 |
1.348 |
15 |
0.418 |
0.707 |
0.545 |
1.670 |
|
0.316 |
0.406 |
0.614 |
1.336 |
20 |
0.419 |
0.718 |
0.605 |
1.742 |
|
0.272 |
0.387 |
0.641 |
1.300 |
25 |
0.448 |
0.820 |
0.441 |
1.709 |
|
0.323 |
0.188 |
0.453 |
0.964 |