Next Article in Journal
Hybrid Pipeline Hardware Architecture Based on Error Detection and Correction for AES
Previous Article in Journal
5G Standalone and 4G Multi-Carrier Network-in-a-Box Using a Software Defined Radio Framework
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Feature Selection Method Based on Random Forest Algorithm for Wind Turbine Condition Monitoring

1
School of Modern Post, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
School of Artificial and Intelligence, Beijing University of Posts and Telecommunications, Beijing 100876, China
3
School of Data Science and Media Intelligence, Communication University of China, Beijing 100024, China
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(16), 5654; https://doi.org/10.3390/s21165654
Submission received: 24 July 2021 / Revised: 10 August 2021 / Accepted: 19 August 2021 / Published: 22 August 2021
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Feature selection and dimensionality reduction are important for the performance of wind turbine condition monitoring models using supervisory control and data acquisition (SCADA) data. In this paper, an improved random forest algorithm, namely Feature Simplification Random Forest (FS_RF), is proposed, which is capable of identifying features closely correlated with wind turbine working conditions. The Euclidian distances are employed to distinguish the weight of the same feature among different samples, and its importance is measured by means of the random forest algorithm. The selected features are finally verified by a two-layer gated recurrent unit (GRU) neural network facilitating condition monitoring. The experimental results demonstrate the capacity and effectiveness of the proposed method for wind turbine condition monitoring.

1. Introduction

Compared with traditional energy sources, wind energy is clean and renewable; thus, wind power has spread worldwide [1,2,3]. However, wind turbines often suffer from frequent malfunctions and failures, which might cause long downtime and significant maintenance costs [4]. For instance, the rotor blades are one of the main components of a wind turbine, and these often fail as the age of the wind farm grows [5]. To prevent high financial losses, condition monitoring and fault prognosis for wind turbines attract a great deal of attention.
In previous studies, condition monitoring methods for wind turbines were mainly carried out with signals collected by sensors. This motivated the research into data-driven wind turbine condition monitoring methods that are capable of estimating working conditions and detecting faults. Nizwan et al. [6] used a Discrete Wavelet Transform (DWT) to analyze vibrational signals in order to achieve fault detection for bearings, where the DWT was employed to decompose signals in different frequency ranges. Sun et al. [7] proposed a method to detect weak features in early faults of rolling bearings in wind turbines. They combined the multiwavelet denoising technique with the threshold of the data-driven block and separated features from noises. Zhang et al. [8] successfully localized the fault planet gear in wind turbine gearboxes using the acoustic emission technique.
Compared with the above monitoring methods, as a comprehensive tool, the supervisory control and data acquisition (SCADA) system has been configured in each wind turbine for working condition supervision. It can provide a large number of parameters that can provide information on the turbine operating condition; therefore, a large number of SCADA data mining methods have been developed. SCADA data were analyzed to construct a model for predicting or detecting the bearing faults of a wind turbine in [9], and faults were predicted by the model 1.5 h before their occurrence. In [10], the authors presented a virtual model to predict two parameters using SCADA data in wind turbines, and the results indicated that the accuracy of the model depended to a large extent on the selected input parameters. The Intelligent System for Predictive Maintenance was applied to monitor the gearbox conditions of wind turbines in [11]. However, among the numerous SCADA parameters, only a few of them are prominent in fault diagnosis and condition monitoring models [12]. Unfortunately, most of the traditional methods [9,10,11,13] have more or less ignored the effects of the interrelation among the SCADA parameters on the model output, to some extent by choosing input parameters according to field experience.
In recent years, deep learning techniques have provided powerful mathematical tools for the fault prognosis and condition monitoring of wind turbines. A large amount of works that have used deep learning techniques were reported in [14]. Jiang et al. [15] employed a denoising autoencoder (DAE) model with time series information from SCADA data to achieve the detection of faults in wind turbines. The optimized long short-term memory (LSTM) neural network, which uses cosine loss, was proposed for the fault diagnosis of a wind turbine gearbox in [16]. Bangalore et al. [17] employed an artificial neural network (ANN) as a condition monitoring method using SCADA data, and the final results proved the effectiveness of the method. All of these approaches [15,16,17] used SCADA data, and input features that are almost manually selected or hand-crafted may restrict the performance of the models. Dimensionality reduction by selecting the most closely related features is a prerequisite to ensure the accuracy of condition estimation models. In [18], the authors developed a prediction model and a diagnosis model using SCADA data for wind turbine generators, where the prediction model was used to predict the remaining useful life of the wind turbine generators. Besides, in [18], the authors also proposed a data preprocessing procedure including data cleaning, feature selection, feature reduction, and data set balancing. In [19], the authors constructed a normal behavior model using support vector regression with a Gaussian kernel to diagnose the faults of wind turbine generators, and the dimensionality of features was reduced by using principal component analysis. Kong et al. [20] introduced a feature selection method with Pearson correlation coefficients in their fault detection model to diagnose the gearbox failures of a wind turbine. Ferreira et al. [21] presented an approach which used decision trees for feature selection and the condition monitoring of tool wear. Wei et al. [22] used a random forest algorithm to select feature parameters and feed them into a constituted deep neural network to detect whether the permanent magnets in a wind turbine had fallen off or not. In [21,22], a method was used that included decision trees to achieve the purpose of condition monitoring. Feature selection was used for variables that can reflect a special component condition in [18,19,20]. However, in this study, the obtained SCADA datasets have only one fault of blade breakages, and no SCADA variables can directly indicate the conditions of blades. Due to this fact, we have to perform feature selection for all SCADA variables.
According to the results of literature studies, better features can simplify the complexity of models and improve the accuracy of condition monitoring models, but the influences of the quality of the selected features on the model performance still remain unclear to date. In this paper, an improved random forest algorithm, the Feature Simplification Random Forest (FS_RF) algorithm, is proposed for the feature selection of SCADA data, in which features that most significantly show the wind turbine’s state are chosen. The gated recurrent unit (GRU) method with the selected features is dedicated to achieving the monitoring of the wind turbine condition. To validate the performance of the proposed method, a comparison is carried out with some other feature selection algorithms. The final results indicate the effectiveness of the proposed method.
This paper is organized as follows. Section 2 presents a brief review of existing feature selection algorithms. The FS_RF algorithm is expounded in Section 3. In Section 4, the experiment setup is depicted and the effectiveness of the proposed feature dimensionality reduction algorithm is evaluated. Finally, the conclusions are given in Section 5.

2. Related Feature Selection Algorithms

Generally, there are mainly three families of feature selection algorithms [23,24,25]: filters, wrappers and embedding methods. The differences between these three basic families are in how the learning algorithm is incorporated to evaluate and select features.
In the filter methods [23,26,27,28,29,30], the selected features are evaluated only by the intrinsic properties of the data without running a learning algorithm. These methods neither rely on any machine learning methods nor require cross-validation. For example, the Pearson correlation coefficient method, as one of the filter methods, was introduced by Kong et al. [20] for the feature selection of SCADA data for wind turbine gearbox condition diagnosis. The Pearson correlation coefficient is used to detect the degree of linear correlation between two continuous variables. This method is suitable for solving regression problems but is not appropriate for classification problems. Another filter method is the variance threshold method, which removes only those features whose variance does not satisfy a certain threshold.
The wrapper methods [24,25] select features by “wrapping” the exploration in a learning algorithm and then estimate feature subsets according to the property of the classifier on each candidate feature subset. An obvious drawback of these methods is the high computational cost of the wrapper methods, since the classifier has to be trained and tested for each candidate feature subset. In practice, it is found that using the wrapper methods requires a large amount of computation resources and time when facing high-dimensional SCADA data. Therefore, these methods are not used in this paper.
The embedded methods [31,32] integrate the feature search and the learning algorithms into a simple optimization formula, employing the advantages of both the wrapper methods combined with machine learning algorithms and the high computational efficiency of the filter methods. For instance, Wei et al. [22] constructed a deep neural network in order to detect the falling off of permanent magnets from wind turbines and used the random forest algorithm as a feature selection method. The random forest algorithm [33] is an embedded method; it is a combination of tree predictors in which each tree depends on the values of independently sampled random vectors that are identically distributed trees in the forest. The final result is obtained by casting a vote for the most popular class using all the decision trees (DTs). The procedure of the random forest algorithm is summarized below and shown in Figure 1.
Step 1: P 1 , P 2 , P 3 . . . P n are sampled randomly from the total training dataset D as n subsets, and the bootstrap sampling method is used in this process;
Step 2: A DT is constructed for each of the n subsets, thus obtaining n classification results;
Step 3: Each DT votes for the most popular class, and this can determine the optimal result.
In the DTs, the minimum Gini value is employed as the splitting criterion of the nodes, and the corresponding features are considered to be excellent features. The impurity degree of each node is described by the Gini value [33], calculated using Formula (1):
Gini ( t ) = 1 j = 1 n [ p ( j t ) ] 2
where p ( j t ) denotes the probability of risk class j at node t. Once the value of Gini ( t ) is 0, the sample data at node t are recognized as the same risk class. The greater the value of Gini ( t ) , the less available the gained information.
In addition, the L1-SVM algorithm [31], one of the main embedded methods, is an embedded sparse method that uses L1 regularization for linear SVM formulas instead of the standard L2 Margin for selecting features. The L1-SVM algorithm and the random forest random algorithm are employed in this research for comparison.
As depicted above, all the existing algorithms have their own obvious drawbacks while working with SCADA data. Nevertheless, the random forest algorithm behaves better compared with other feature selection methods. However, even in the random forest algorithm, the classification results of the DTs are heavily influenced by the redundant features of the sampled SCADA data. Therefore, this paper proposes the FS_RF algorithm to improve the performance of the random forest algorithm.

3. Proposed Algorithm

In [20], to detect whether the permanent magnets in the wind turbines were dislodged, the random forest algorithm was employed for feature selection for construct a condition monitoring model. However, in this paper, in order to diagnose whether the wind turbine blades are broken or not, the feature selection of all SCADA variables is necessary. When facing high-dimensional data, the feature simplification (FS) algorithm can reduce the impact of redundant features on the random forest algorithm.
The FS algorithm removes features that have little impact on the operating state of the wind turbines. The purpose of this process is to decrease the interference of some useless features in the calculation of Gini values in the decision trees of the random forest algorithm. The final results in Section 4.4 indicate that the features that are selected by the FS_RF algorithm are different from features selected using the random forest algorithm.
The feature simplification (FS) algorithm mainly calculates the correlation of a feature with the samples from positive and negative classes and assigns corresponding weights to each of the features, where the positive class denotes the SCADA data of the normal working conditions and the negative class the fault. Figure 2 shows the flowchart of the FS algorithm. The detailed algorithm is depicted step by step as follows:
Step 1: Sample R is randomly drawn from the SCADA dataset, sample S of the same class is adjacent to sample R, and sample D is drawn randomly from a different class;
Step 2: A feature (denoted as A i ) is selected from each of the three samples (R, S and D) in turn;
Step 3: The Euclidean distances between feature A i of sample R and that of sample S are calculated, denoted as d ( R A i , S A i ) , and those of feature A i between sample R and sample D, denoted as d ( R A i , D A i ) ;
Step 4: Steps 1 to 3 are repeated, and the weight of each feature A i is computed. The weight computation formula is shown as Formula (2):
W A i = W A i d R A i , S A i m d R A i , D A i m ( i = 1 , 2 , 3 , , n )
where W ( A i ) is the weight of feature A i and the initial value of each feature weight is assigned to 0; m denotes the repetitions; and n is the total number of features in the SCADA data. In Figure 2, in order to save computation cost, k is assigned to 0.6 × P, where P is the total numbers of samples in a SCADA dataset.
In step 3, if the two Euclidean distances (i.e., d ( R A i , R S i ) , d ( R A i , D S i ) ) are significantly different, this means that the feature A i has a greater ability to distinguish the positive or the negative samples among R, S, and D, and then the weight of feature A i should be increased. Otherwise, its weight should be decreased.

4. A Case Study

In this section, the proposed method is evaluated with 10 min SCADA data on two SCADA datasets collected from two Aeolon58 wind turbines, where the two SCADA datasets contain 418,078 and 404,933 items, respectively. These datasets include normal operating conditions of the wind turbines and failure conditions after the breakage of a blade. A GRU model is constructed to estimate the condition of the wind turbines. The general flow is as follows:
Step 1: The SCADA data are preprocessed by the proposed feature selection method;
Step 2: The SCADA data are divided into training sets and testing sets, and the splitting ratio is set to 0.6;
Step 3: The GRU model is built;
Step 4: The hyper-parameters are initialized (refer to Section 4.4);
Step 5: The training sets are fed to train the GRU model;
Step 6: The correctness of the feature selection method is evaluated.

4.1. Data Preprocessing

The SCADA data acquired from a real wind farm supervision system frequently have missing entries, which will lead to discontinuities in the time series. Moreover, a SCADA system usually involves a manifold of variables, which may affect the accuracy of the wind turbine condition estimation model if these missing data are not properly processed [22]. It is therefore necessary to preprocess the SCADA data.

4.1.1. Missing Value Processing

Considering the need to maintain the temporal order of the SCADA data, the missing values of the SCADA data should not be neglected directly. In this paper, the local mean replacement method to pad the missing values is employed, which can be expressed as follows:
x m = m k 1 m 1 x i + m + k + 1 m + 1 x i 2 k
where x m is the value of the missing data, and k represents the number of available data values near the missing data, which is set to 3 in this study, in order to smooth the data series curve.

4.1.2. Feature Selection

In the proposed method, the FS algorithm is incorporated into the feature selection process of the random forest. Firstly, the features with zero or very low weights in the SCADA dataset are initially removed using the feature simplification algorithm; secondly, the random forest algorithm is employed to calculate the importance of features (i.e., columns in the SCADA dataset); and finally, feature selection is achieved based on feature importance.
On the basis of the FS_RF algorithm, the importance of the 75 features in the wind turbine SCADA dataset are computed, and the 28 most important output features are retained as the input of the GRU model. The selected features using the FS_RF algorithm are shown in Table 1.

4.2. Constructing the Gated Recurrent Unit Model for Condition Monitoring

When evaluating the conditions of a wind turbine, the data from the SCADA system should be continuously fed into the model in the form of a time series. The common fully-connected neural network is weak in perceiving the change in the conditions of the wind turbine in real time, while the convolutional neural network can only monitor the magnitude of the local change. Obviously, these two methods are both insufficiently sensitive to the input data and cannot reflect the state change of the wind turbine in real time.
A recurrent neural network (RNN) is a type of artificial neural network with directed cycles of dependencies between nodes [34]. This construction allows the network to retain previous state information between successive time steps. The value of every time step is considered, affecting the temporal result. Generally, given a s e q u e n c e X = [ x 1 , x 2 , , x n ] , x t R k is the input of time step t. The process of the RNN preserving states can be defined by the following state transfer function [35,36]:
h t = f W x t + U h t 1 + b
where U R d × d represents the matrix between the adjacent time-step of hidden layers and itself; W R d × k is the matrix, which denotes the ordinary weights between the input layer and hidden layers; b R d is the bias parameter; U, W and b are shared parameters and can learn during model training; h t R d denotes the corresponding hidden state when the input is x t ; and f, which is widely used in RNNs, is the hyperbolic tangent activation function.
Nonetheless, standard RNNs suffer from gradient disappearance and gradient explosion during training, which may make it difficult to obtain satisfactory results. More importantly, standard RNNs are unable to remember long-term data and discard relevant prior states—a problem known as “fading memory” [37].
The gated recurrent unit (GRU) evolved on the basis of the RNN and has become popular due to its better information storage and ability to access prior conditions. In contrast to traditional RNNs, the hidden unit of the GRU is replaced by a gated unit, which does not require a separate storage unit to regulate the flow of information within the unit. The structure of this model is shown in Figure 3.
The calculation of the GRU layer can be described as follows [34]:
z t = σ g W z x t + U z h t 1 + b z
r t = σ g W r x t + U r h t 1 + b r
h ˜ t = tanh W h x t + U t r t h t 1
h t = 1 z t h t 1 + z t h ˜ t
where all W R d × k , U R d × d and b R d , as the included learnable parameters, should be shared by each step, and they are all able to learn in the period of the GRU model training; ⊙ represents the element-wise product; d and k are hyper-parameters representing the dimensions of the input and hidden vector; h t is the output vector, which includes information of the current unit when the input vector is x t ; z t and r t are the vectors for update gate and reset gate; σ g is the sigmoid activation function; and tanh is the hyperbolic tangent activation function.
The GRU model can resolve the drawbacks of the standard RNNs. The update gate vector and the reset gate vector are employed to decide whether the information should be remembered or forgotten and to learn adaptive weights of different features. Moreover, compared with the long short-term memory (LSTM) network, the GRU has better performance with fewer parameters to suppress overfitting [38].

4.3. Loss Function and Evaluation Criteria

The evaluation of the feature selection algorithm is performed according to the accuracy of the model. We use binary cross-entropy as a loss function, and its mathematical expression could be expressed as follows:
l n = y n log y ^ n + 1 y n log 1 y ^ n
loss ( z , y ) = mean l 0 , l 1 , l 2 , , l N 1
where N denotes the total number of samples; y ^ n is the probability that the nth sample is a positive case; and y n the true label of the nth sample.
To measure the influence of the feature selection on the deep learning model, the F1-Score is used as a criterion. Here, “0” denotes a healthy condition of the wind turbine and “1” represents an abnormal condition. The basic parameters in the standard performance metrics are adopted: true positive (TP), false positive (FP), false negative (FN), and true negative (TN).
As a statistical measurement, the F1-Score combines the precision and recall of the classification model, with a maximum value of 1 and a minimum value of 0. The F1-Score can be mathematically expressed as follows:
Accuracy = T P + T N T P + F P + F N + T N
Precision = T P T P + F P
Recall = T P T P + F N
F 1 Score = 2 T P 2 T P + F N + F P = 2 · Precision · Recall Precision + Recall
The aforementioned metrics are specific to each category, and they measure the capacity of the condition monitoring model to distinguish certain circumstances (i.e., fault) from normal circumstances (i.e., health). In addition, accuracy is also employed to evaluate the overall model performance for the positive class and negative class.

4.4. Experiment Results

In this section, a GRU model is constructed and applied to the wind turbine condition estimation. Two SCADA datasets are employed to evaluate the proposed algorithm, in which both healthy and failure operating condition data are recorded. In the proposed method, the selected features, preprocessed using the FS_RF algorithm, are fed into the GRU model. The detailed setup of the GRU model is shown in Table 2.
To further evaluate the performance of the FS_RF algorithm, a comparison is conducted with other feature selection algorithms—i.e., the random forest algorithm, the variance threshold algorithm, and the L1-SVM algorithm—where the threshold of the variance threshold method is set to 0.4. The results of different feature selection methods are described in Table 3 and Table 4, respectively.
The experimental results show that different feature selection methods result in a variance in the number of selected features. In Table 3, the F1-Score of the GRU condition estimation model which uses the proposed method reaches 95.21, surpassing the method in second place by about 0.41. In Table 4, the F1-Score of the best performing model using the FS_RF algorithm is 91.24, which exceeds the model in second place by approximately 0.97.
As shown in Table 3 and Table 4, the FS_RF algorithm performs well on both data samples. Further, 28 features with a higher impact on the metrics are selected using the FS_RF algorithm, while the random forest algorithm selects only 22 features. This means that the final F1-Score of the model using the FS_RF algorithm is higher than that of the random forest algorithm.

4.5. Discussion

In this study, we propose a FS algorithm to optimize the random forest algorithm for SCADA data. The FS_RF algorithm behaves more efficiently when choosing the important features from the data sample than other algorithms. The reason for this may be that the features selected by the FS algorithm reduce the complexity of the later calculation of feature importance using the random forest algorithm. In addition, it is worth noting that none of the condition monitoring models constructed for sample 2 using the four feature selection methods performed as well as those for sample 1. This may have been caused by the degree of the failures of the wind turbine; i.e., minor failures could cause insignificant variations in the parameters of the SCADA data, while severe faults cause remarkable changes, which in turn affects the diagnosis ability of the GRU models.

5. Conclusions

Feature selection and dimensionality reduction on the SCADA data of wind turbines remain a perplexing problem. Although prior research has provided hope for a better result of condition estimation, there still remains room for improvement.
This paper proposes the FS_RF algorithm for feature selection and dimensionality reduction on SCADA data, and the approach is evaluated using a GRU deep learning model. The proposed algorithm is able to select features from SCADA datasets that better denote the operating state of the wind turbines by means of computing the weight of a feature using Euclidian distances among sample datasets and measuring its importance. The results obtained in the experiments demonstrate the applicability of the proposed method.
In addition, the findings of this research have a number of important implications for future practice by providing a reference for feature selection and dimensionality reduction on massive datasets in other fields.

Author Contributions

Conceptualization, G.L. and C.W.; methodology, G.L.; software, G.L.; validation, G.L.; formal analysis, G.L.; investigation, G.L.; resources, C.W.; data curation, G.L.; writing—original draft preparation, G.L.; writing—review and editing, C.W. and D.Z.; visualization, G.L.; supervision, G.Y. and C.W.; project administration, C.W.and G.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Provincial Key Research and Development Plan (Industry Foresight and Common Key Technologies), Jiangsu Province, China [Grant number: BE2016032] and the Major Scientific and Technological Support and Independent Innovation Project, Jiangsu Province, China [Grant number: BE2010019].

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to thank the support of Beijing University of Posts and Telecommunications.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MDPIMultidisciplinary Digital Publishing Institute
DOAJDirectory of open access journals
TLAThree letter acronym
LDLinear dichroism

References

  1. Qiao, W.; Lu, D. A survey on wind turbine condition monitoring and fault diagnosis—Part I: Components and subsystems. IEEE Trans. Ind. Electron. 2015, 62, 6536–6545. [Google Scholar] [CrossRef]
  2. Liu, Y.; Tan, L. Tip clearance on pressure fluctuation intensity and vortex characteristic of a mixed flow pump as turbine at pump mode. Renew. Energy 2018, 129, 606–615. [Google Scholar] [CrossRef]
  3. Hao, Y.; Tan, L. Symmetrical and unsymmetrical tip clearances on cavitation performance and radial force of a mixed flow pump as turbine at pump mode. Renew. Energy 2018, 127, 368–376. [Google Scholar] [CrossRef]
  4. Gao, Z.; Ding, S.X.; Cecati, C. Real-time fault diagnosis and fault-tolerant control. IEEE Trans. Ind. Electron. 2015, 62, 3752–3756. [Google Scholar] [CrossRef] [Green Version]
  5. Chou, J.-S.; Chiu, C.-K.; Huang, I.-K.; Chi, K.-N. Failure analysis of wind turbine blade under critical wind loads. Eng. Fail. Anal. 2013, 27, 99–118. [Google Scholar] [CrossRef]
  6. Nizwan, C.; Ong, S.; Yusof, M.; Baharom, M. A wavelet decomposition analysis of vibration signal for bearing fault detection. In IOP Conference Series: Materials Science and Engineering; IOP Publishing: Bristol, UK, 2013; p. 012026. [Google Scholar]
  7. Sun, H.; Zi, Y.; He, Z. Wind turbine fault detection using multiwavelet denoising with the data-driven block threshold. Appl. Acoust. 2014, 77, 122–129. [Google Scholar] [CrossRef]
  8. Zhang, Y.; Lu, W.; Chu, F. Planet gear fault localization for wind turbine gearbox using acoustic emission signals. Renew. Energy 2017, 109, 449–460. [Google Scholar] [CrossRef]
  9. Kusiak, A.; Verma, A. Analyzing bearing faults in wind turbines: A data-mining approach. Renew. Energy 2012, 48, 110–116. [Google Scholar] [CrossRef]
  10. Kusiak, A.; Li, W. Virtual models for prediction of wind turbine parameters. IEEE Trans. Energy Convers. 2009, 25, 245–252. [Google Scholar] [CrossRef]
  11. Garcia, M.C.; Sanz-Bobi, M.A.; Del Pico, J. SIMAP: Intelligent System for Predictive Maintenance: Application to the health condition monitoring of a windturbine gearbox. Comput. Ind. 2006, 57, 552–568. [Google Scholar] [CrossRef]
  12. Wilkinson, M.; Darnell, B.; Van Delft, T.; Harman, K. Comparison of methods for wind turbine condition monitoring with SCADA data. IET Renew. Power Gener. 2014, 8, 390–397. [Google Scholar] [CrossRef]
  13. Zaher, A.; McArthur, S.; Infield, D.; Patel, Y. Online wind turbine fault detection through automated SCADA data analysis. Wind Energy Int. J. Prog. Appl. Wind Power Convers. Technol. 2009, 12, 574–593. [Google Scholar] [CrossRef]
  14. Helbing, G.; Ritter, M. Deep Learning for fault detection in wind turbines. Renew. Sustain. Energy Rev. 2018, 98, 189–198. [Google Scholar] [CrossRef]
  15. Jiang, G.; Xie, P.; He, H.; Yan, J. Wind turbine fault detection using a denoising autoencoder with temporal information. IEEE/Asme Trans. Mechatron. 2017, 23, 89–100. [Google Scholar] [CrossRef]
  16. Yin, A.; Yan, Y.; Zhang, Z.; Li, C.; Sanchez, R.V. Fault Diagnosis of Wind Turbine Gearbox Based on the Optimized LSTM Neural Network with Cosine Loss. Sensors 2020, 20, 2339. [Google Scholar] [CrossRef] [Green Version]
  17. Bangalore, P.; Tjernberg, L.B. An artificial neural network approach for early fault detection of gearbox bearings. IEEE Trans. Smart Grid 2015, 6, 980–987. [Google Scholar] [CrossRef]
  18. Zhao, Y.; Li, D.; Dong, A.; Kang, D.; Lv, Q.; Shang, L. Fault prediction and diagnosis of wind turbine generators using SCADA data. Energies 2017, 10, 1210. [Google Scholar] [CrossRef] [Green Version]
  19. Castellani, F.; Astolfi, D.; Natili, F. SCADA Data Analysis Methods for Diagnosis of Electrical Faults to Wind Turbine Generators. Appl. Sci. 2021, 11, 3307. [Google Scholar] [CrossRef]
  20. Kong, Z.; Tang, B.; Deng, L.; Liu, W.; Han, Y. Condition monitoring of wind turbines based on spatio-temporal fusion of SCADA data by convolutional neural networks and gated recurrent units. Renew. Energy 2020, 146, 760–768. [Google Scholar] [CrossRef]
  21. Ferreira, B.; Silva, R.; Pereira, V. Feature selection using non-binary decision trees applied to condition monitoring. In Proceedings of the 2017 22nd IEEE International Conference on Emerging Technologies and Factory Automation (ETFA), Limassol, Cyprus, 12–15 September 2017; pp. 1–7. [Google Scholar]
  22. Teng, W.; Cheng, H.; Ding, X.; Liu, Y.; Ma, Z.; Mu, H. DNN-based approach for fault detection in a direct drive wind turbine. IET Renew. Power Gener. 2018, 12, 1164–1171. [Google Scholar] [CrossRef]
  23. Yu, L.; Liu, H. Efficient feature selection via analysis of relevance and redundancy. J. Mach. Learn. Res. 2004, 5, 1205–1224. [Google Scholar]
  24. Kohavi, R.; John, G.H. Wrappers for feature subset selection. Artif. Intell. 1997, 97, 273–324. [Google Scholar] [CrossRef] [Green Version]
  25. Guyon, I.; Weston, J.; Barnhill, S.; Vapnik, V. Gene selection for cancer classification using support vector machines. Mach. Learn. 2002, 46, 389–422. [Google Scholar] [CrossRef]
  26. Song, L.; Smola, A.; Gretton, A.; Borgwardt, K.M.; Bedo, J. Supervised feature selection via dependence estimation. In Proceedings of the 24th International Conference on Machine Learning, Corvalis, OR, USA, 20–24 June 2007; pp. 823–830. [Google Scholar]
  27. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  28. Kira, K.; Rendell, L.A. A practical approach to feature selection. In Machine Learning Proceedings; Elsevier: Amsterdam, The Netherlands, 1992; pp. 249–256. [Google Scholar]
  29. He, X.; Cai, D.; Niyogi, P. Laplacian score for feature selection. Adv. Neural Inf. Process. Syst. 2005, 18, 507–514. [Google Scholar]
  30. Zhao, Z.; Liu, H. Spectral feature selection for supervised and unsupervised learning. In Proceedings of the 24th International Conference on Machine Learning, Corvalis, OR, USA, 20–24 June 2007; pp. 1151–1157. [Google Scholar]
  31. Zhu, J.; Rosset, S.; Tibshirani, R.; Hastie, T.J. 1-norm support vector machines. In Advances in Neural Information Processing Systems; Citeseer: Princeton, NJ, USA, 2003. [Google Scholar]
  32. Vapnik, V.N. An overview of statistical learning theory. IEEE Trans. Neural Netw. 1999, 10, 988–999. [Google Scholar] [CrossRef] [Green Version]
  33. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef] [Green Version]
  34. Schmidhuber, J. Deep learning in neural networks: An overview. Neural Netw. 2015, 61, 85–117. [Google Scholar] [CrossRef] [Green Version]
  35. Zhang, J.; Wang, P.; Yan, R.; Gao, R.X. Long short-term memory for machine remaining life prediction. J. Manuf. Syst. 2018, 48, 78–86. [Google Scholar] [CrossRef]
  36. Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:14061078. [Google Scholar]
  37. Zhao, R.; Wang, D.; Yan, R.; Mao, K.; Shen, F.; Wang, J. Machine health monitoring using local feature-based gated recurrent unit networks. IEEE Trans. Ind. Electron. 2017, 65, 1539–1548. [Google Scholar] [CrossRef]
  38. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:14123555. [Google Scholar]
Figure 1. The procedure of the random forest algorithm.
Figure 1. The procedure of the random forest algorithm.
Sensors 21 05654 g001
Figure 2. The flowchart of the feature simplification algorithm.
Figure 2. The flowchart of the feature simplification algorithm.
Sensors 21 05654 g002
Figure 3. The structure of GRU.
Figure 3. The structure of GRU.
Sensors 21 05654 g003
Table 1. Selected SCADA features.
Table 1. Selected SCADA features.
NoFeatureNoFeature
1Nacelle temperature2Hub temperature
3Reactive power control status4Generator active power
5Converter grid side voltage6Blade 1 battery box temperature
7Blade 2 battery box temperature8Blade 3 battery box temperature
9Blade 1 converter box temperature10Blade 2 converter box temperature
11Blade 3 converter box temperature12Blade 1 angle
13Blade 2 angle14Blade 3 angle
15Main bearing temperature 116Main bearing temperature 2
17Engine room control cabinet temperature18Hub control cabinet temperature
19Generator stator temperature 120Generator stator temperature 2
21Generator stator temperature 322Generator stator temperature 4
23Generator stator temperature 524Generator stator temperature 6
25Wind measurement tower temperature26Converter inlet pressure
27Converter outlet pressure28Absolute wind direction
Table 2. Structure and hyper-parameter setup of the GRU model.
Table 2. Structure and hyper-parameter setup of the GRU model.
Structure and Hyper-Parameter SettingValues
StructureInput size28
The number of neurons in the GRU layer50
The number of neurons in the fully connected layer10
The number of neurons in the output layer2
Training settingsBatch size64
Learning rate0.01
The number of GRU layers2
Dropout rate in the GRU layers0.2
The total number of epochs50
Table 3. Results on the sample 1.
Table 3. Results on the sample 1.
MethodNum of FeaturesAccuracyPrecisionRecallF1-Score
VarianceThrehold-GPU5790.3191.5789.3191.79
L1-SVM-GPU6193.1894.3292.3494.17
RandomForest-GRU2294.0094.7193.3794.80
FS_RF-GRU(Proposed)2894.5095.1393.9395.21
Table 4. Results on the sample 2.
Table 4. Results on the sample 2.
MethodNum of FeaturesAccuracyPrecisionRecallF1-Score
VarianceThrehold-GPU5781.6581.5981.7582.44
L1-SVM-GPU6189.8389.8189.9290.27
RandomForest-GRU2288.6488.6088.7888.94
FS_RF-GRU(Proposed)2890.7790.7190.8391.24
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, G.; Wang, C.; Zhang, D.; Yang, G. An Improved Feature Selection Method Based on Random Forest Algorithm for Wind Turbine Condition Monitoring. Sensors 2021, 21, 5654. https://doi.org/10.3390/s21165654

AMA Style

Li G, Wang C, Zhang D, Yang G. An Improved Feature Selection Method Based on Random Forest Algorithm for Wind Turbine Condition Monitoring. Sensors. 2021; 21(16):5654. https://doi.org/10.3390/s21165654

Chicago/Turabian Style

Li, Guo, Chensheng Wang, Di Zhang, and Guang Yang. 2021. "An Improved Feature Selection Method Based on Random Forest Algorithm for Wind Turbine Condition Monitoring" Sensors 21, no. 16: 5654. https://doi.org/10.3390/s21165654

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop