You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

16 January 2023

Rolling Bearing Fault Feature Selection Method Based on a Clustering Hybrid Binary Cuckoo Search

,
,
and
1
Key Laboratory of Grain Information Processing and Control, Ministry of Education, Henan University of Technology, Zhengzhou 450001, China
2
Zhengzhou Key Laboratory of Machine Perception and Intelligent System, Henan University of Technology, Zhengzhou 450001, China
3
School of Electrical Engineering, Henan University of Technology, Zhengzhou 450001, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Fault Detection, Diagnosis, and Prognosis Techniques towards System Reliability Enhancement

Abstract

In order to solve the low accuracy in rolling bearing fault diagnosis caused by irrelevant and redundant features, a feature selection method based on a clustering hybrid binary cuckoo search is proposed. First, the measured motor signal is processed by Hilbert–Huang transform technology to extract fault features. Second, a clustering hybrid initialization technique is given for feature selection, combining the Louvain algorithm and the feature number. Third, a mutation strategy based on Levy flight is proposed, which effectively utilizes high-quality information to guide subsequent searches. In addition, a dynamic abandonment probability is proposed based on population sorting, which can effectively retain high-quality solutions and accelerate the convergence of the algorithm. Experimental results from nine UCI datasets show the effectiveness of the proposed improvement strategy. The open-source bearing dataset is used to compare the fault diagnosis accuracy of different algorithms. The experimental results show that the diagnostic error rate of this method is only 1.13%, which significantly improves classification accuracy and effectively realizes feature dimension reduction in fault datasets. Compared to similar methods, the proposed method has better comprehensive performance.

1. Introduction

With the development of modern industry, motors are ubiquitous in manufacturing applications [1]. Rolling bearings are the core components and vulnerable parts of machinery. Their health directly affects the performance, efficiency, stability, and life of the machine. According to statistics from the American Electric Power Research Institute, 41% of motor faults are caused by bearing damage, which is the primary cause of motor faults [2]. In order to maintain the safe operation of equipment, early identification of rolling bearing defects is essential [3].
Traditional bearing fault diagnosis is mainly based on model analysis and signal processing [4]. However, these two methods often rely on the accumulation of experience. Due to the limitations of these two methods, they cannot meet the reliability requirements of modern production equipment. With the development of artificial intelligence technology, fault diagnosis methods based on machine learning show good performance in terms of motor health detection. Machine learning-based bearing fault diagnosis can be divided into three steps. The first step is to extract features, the second step is feature selection, and the third step is classifier recognition [5].
Extraction of the running data of the bearings is an important step in the realization of rolling bearing condition monitoring. Feature extraction is the process of extracting attributes from motor signals [6]. So far, researchers have studied various information analysis techniques, such as the fast Fourier transform (FFT), envelope analysis (EA), wavelet transform (WT), and Hilbert–Huang transform (HHT) [7]. The HHT is one of the most widely used signal analysis tools in nonlinear and nonstationary signal analysis. Therefore, this paper uses HHT technology to extract fault features. Not all the time–frequency domain features extracted by HHT technology are conducive to fault diagnosis. Irrelevant and redundant features not only reduce the efficiency of model operation, but also lead to a decline in model recognition performance [8]. Therefore, in order to avoid the interference of redundant and irrelevant features, it is necessary to use effective methods that select the optimal feature subset from multi-dimensional features.
Feature selection is an essential data preprocessing method, which can effectively reduce data dimensionality and improve classification performance by removing redundant or irrelevant features [9]. Currently, two widely used strategies in feature selection methods are the filter and wrapper methods [10]. The filter method obtains the intrinsic correlation of features through univariate statistics. This method is independent of the learning algorithm, so filter calculation is cheaper than calculation using the wrapper method [11]. The wrapper method uses the classifier as part of the evaluation function, so it usually works better than filters. Recently, swarm intelligence algorithms such as the genetic algorithm (GA) [12], particle swarm optimization (PSO) [13], grey wolf optimization (GWO) [14], and crow search algorithm (CSA) [15] have been widely used in the search process of wrapper-based methods. Furthermore, the cuckoo search algorithm (CS) [16] is inspired by cuckoo production and breeding behavior. It is a promising metaheuristic algorithm due to having fewer adjustment parameters and good search ability. The CS algorithm has a more effective search ability than the GA and PSO [17]. However, the CS algorithm still faces the problems of slow convergence speed and random initialization when solving feature selection problem.
Based on the problems mentioned above, a feature selection method based on a clustering hybrid binary cuckoo search (CHBCS) is proposed. The main contributions of this study are briefly presented as follows:
  • Propose a strategy to extract the time–frequency domain features of motor signals based on the Hilbert–Huang transform.
  • In order to reduce redundant features in a population, a clustering hybrid initialization method is presented. The method uses the Louvain algorithm to cluster features and initializes the population according to the clustering information and the number of features, which can effectively remove redundant features.
  • A mutation strategy based on Levy flights is proposed to improve the update formula. This strategy can effectively utilize the high-quality information of the population by guiding the subsequent search with several high-quality individuals.
  • The proposed dynamic (Pa) probability strategy adaptively adjusts the (Pa) probability based on population rankings to preserve the high-quality solution of the current population.
This article is organized as follows. The related work is discussed in Section 2. Section 3 introduces the feature extraction method. Section 4 describes the proposed algorithm in detail. In Section 5, the effectiveness of the proposed CHBCS algorithm is verified by experiments. Finally, Section 6 provides the conclusion.

3. Feature Extraction

In this paper, the Hilbert–Huang transform is used to extract the features of bearing signals. This technique first decomposes a column of time series data using the empirical mode decomposition (EMD) algorithm. It then obtains the operating characteristics of the time series data using the Hilbert transform [34].
The vibration signal obtained from rotating machinery is usually nonstationary, complex, and nonlinear, which does not meet the preconditions of the Hilbert transform [35]. Therefore, it is necessary to use EMD to decompose the nonlinear stationary signal into a stationary signal. The Hilbert transform is then performed on the intrinsic mode function (IMF) obtained by decomposition, and the complex signal is obtained for further analysis. For the given signal x(t), the signal decomposition process in the EMD method is shown in Figure 1.
Figure 1. Signal decomposition flow chart of the EMD method.
According to the IMF and the residual component r(t), the original signal x(t) can be reconstructed, as shown in Equation (1).
x ( t ) = i = 1 n c i ( t ) + r n ( t ) .
Each IMF component ci(t) obtained via EMD of the signal is subjected to the Hilbert transform to generate H[ci(t)]. The equation is as follows:
H [ c i ( t ) ] = 1 π c i ( t ) t τ d τ ,
where H[ci(t)] and ci(t) are conjugate complex pairs. The analysis signal zi(t) is shown in Equation (3).
z i ( t ) = a i ( t ) e j θ i ( t ) ,
where ai(t) and θi(t) are time functions, as shown in Equations (4) and (5).
a i ( t ) = c i 2 ( t ) + H 2 [ c i ( t ) ] ,
θ i ( t ) = arctan H [ c i ( t ) ] c i ( t ) .
The Hilbert time–frequency spectrum matrix of the motor operation signal is obtained by the above calculation. The features of the signal of the time–frequency domain are extracted according to the information of the Hilbert time–frequency spectrum matrix [27]. In the time domain, this paper obtained five characteristic curves by calculating the maximum (T-max), mean (T-mean), mean square error (T-mse), root mean square (T-rms), and standard deviation (T-std) of each column of the Hilbert time spectrum matrix. For each characteristic curve, the maximum value, mean value, mean square error, root mean square value, and standard deviation can be calculated. Thus, a total of 25 statistical features were obtained from five characteristic curves, recorded as F1–F25. Using the same program in the frequency domain, the five characteristic curves of F-max, F-mean, F-mse, F-rms, and F-std are obtained. A total of 25 statistical features from the frequency domain were extracted and recorded as F26–F50. Figure 2 shows the process of the Hilbert–Huang transform establishing a feature set containing 50 statistical features.
Figure 2. Feature extraction process diagram.

5. Experiments and Analysis

This section verifies the feasibility of the proposed method through two experiments. First, the performance of feature selection method was verified using the famous UCI benchmark dataset. To confirm the efficacy of the improved method, the suggested CHBCS algorithm is contrasted with other feature selection algorithms based on the cuckoo search. Second, fault diagnosis tasks were performed on the bearing fault dataset. In bearing fault diagnosis, the CHBCS algorithm is compared to the classical optimization algorithm.

5.1. Performance Analysis of the CHBCS Algorithm

5.1.1. UCI Dataset

This experiment tested the performance of the suggested method using nine well-known benchmark datasets from the UCI database [40]. These datasets are widely used for performance comparisons of feature selection. The number of features, samples, and classifications in the datasets are shown in Table 1.
Table 1. Dataset information.
The range difference between the original data will affect the classification results of the classifier. The feature sample χij is normalized by the maximum and minimum normalization method to obtain χij.
χ i j = χ i j χ i min χ i max χ i min ,
where χimax and χimin represent the maximum values and minimum values of the i-th feature in the sample.

5.1.2. Experiment Setting

This section of the experiment is divided into two steps to study the impact of the proposed improvement strategy on the performance of the algorithm. First, the clustering hybrid initialization method is compared to other initialization methods. In the second stage, the CHBCS algorithm is compared to other BCS algorithms to verify the effectiveness of the improved strategy in improving classification performance and convergence speed.
A 10-fold cross-validation [41] was used to generate a training set and a test set from the normalized dataset. The feature subset was evaluated using a KNN classifier (in this article k = 5). The maximum number of iterations for all experiments in this paper was 100, the population size was 50, and the probability Pa was set to 0.25. All algorithms were implemented in Python, and all experiments were conducted on a PC with an Intel (R) Core (TM) i5-8500T CPU @ 2.10 GHz and 8.0 GB of RAM.

5.1.3. Effect of Initialization Strategy on Feature Selection

To validate the validity of the clustering hybrid initialization strategy proposed in this paper, random (Rand) initialization and initialization based on clustering (BC) [42] are compared to the clustering hybrid (CH) initialization method proposed in this paper. These three methods use the standard BCS algorithm for a subsequent search.
Table 2 shows the average classification error rate of BCS algorithms with different initialization strategies. Results show that CH initialization achieves the lowest error rate for the nine datasets. Its classification performance is improved compared to other initialization methods.
Table 2. Classification error rate of three initialization strategies.
Figure 7 intuitively reflects the quantitative relationship between the number of features of the three initialization strategies in the nine datasets. Table 2 and Figure 7 show that the CH initialization method not only reduces the number of features by a significant amount, but also achieves a lower classification error rate. In all datasets, the BCS (CH) algorithm obtains fewer features than the other two algorithms. The CH initialization method can effectively reduce redundant features by using clustering information and a hybrid initialization strategy. The advantages are evident in datasets with a high number of features, such as Hillvalley, Libras, and Musk1.
Figure 7. Average histogram of the number of features for the three algorithms.
The clustering hybrid initialization method considers the similarity between features, which can effectively avoid the influence of redundant features and reduce the number of features. Thus, while maintaining the classification error rate, clustering hybrid initialization can effectively minimize the number of features.

5.1.4. Comparison with Other Cuckoo Search Algorithms

This section verifies the impact of the mutation strategy based on Levy flight and the dynamic Pa probability strategy on algorithm performance. The CHBCS algorithm is compared to the binary cuckoo search algorithm (BCS) [28] and the enhanced binary cuckoo search (EBCS) [29] under unified use of the CH initialization strategy.
Experimental results for the three algorithms after being run independently 20 times in the nine datasets are shown in Table 3. The results show that the CHBCS algorithm achieves the lowest average classification error rate for all datasets and obtains the lowest average number of features for eight datasets. In the Musk1 dataset, the average error rate of the CHBCS algorithm is 29.51% and 38.17% lower than that of the BCS algorithm and the EBCS algorithm, respectively, and the average number of features is the lowest. In the Exactly dataset, the CHBCS algorithm achieves 100% accuracy and the lowest average number of features in 20 experiments, with classification performance and stability also being better than other algorithms. The two proposed improvement strategies can effectively improve the classification performance of the CS algorithm.
Table 3. Classification results of three algorithms.
Furthermore, Figure 8 displays the convergence curves of the three algorithms for each of the nine datasets. In the iterative process, the CHBCS algorithm can search for a better solution than the other two algorithms. On datasets such as Hillvalley, Libras, and Musk1, the convergence curve drops rapidly and performance is significantly improved. In the Exactly dataset, the convergence curve of the CHBCS algorithm converges faster than the curves for the BCS algorithm and the EBCS algorithm, and the curve is smoother. In the HeartEW dataset, the fitness value of the BCS algorithm stops decreasing after 30 iterations, and its classification performance is far from that of the CHBCS algorithm.
Figure 8. Iterative fitness curves of the three algorithms.
The CHBCS algorithm can effectively search for the optimal solution, mainly because the mutation strategy based on Levy flight improves the exploration ability of the BCS algorithm and DPa can effectively retain high-quality solutions and accelerate the algorithm’s convergence. Therefore, the above results prove that the CHBCS algorithm can effectively enhance the search ability of cuckoos and improve the convergence performance of the BCS algorithm.

5.2. Bearing Fault Diagnosis Experiment

5.2.1. Data Sources

The experimental data for rolling bearings used in this paper come from the Bearing Data Center of Case Western Reserve University (CWRU) [5]. The CWRU benchmark is a common dataset that is widely used in bearing fault research. The experimental bearing adopts artificial single point electric spark damage. The damage diameter is divided into three groups (0.007 inches, 0.014 inches, and 0.021 inches). Fault location can be divided into rolling ball (B), inner ring (IR), and outer ring (OR) groups. There are ten bearing states, including the normal state and different fault states. The vibration signal data are obtained from the acceleration sensor of the drive end under the no-load condition of the motor, with a sampling frequency of 12 kHz. The measurement data in the dataset are provided as a single file, with each file containing about 120,000 data points that correspond to each bearing state. The length of each signal sample was set to 2000 sample points. A total of 50 test samples were obtained for each health state. This information for the CWRU dataset is shown in Table 4.
Table 4. CWRU bearing data.
Figure 9 shows the time domain waveforms of the motor vibration signals under 10 bearing conditions. It can be seen from the diagram that it is difficult to distinguish fault types using only vibration signals. Therefore, the signal was analyzed using the feature extraction method in Section 3, and a50-dimensional fault feature set was extracted.
Figure 9. Vibration signal of ten bearing states.

5.2.2. Experimental Design

To prove the effectiveness of the CHBCS feature selection method in rolling bearing fault diagnosis, the CHBCS was compared to the BCS [28], EBCS [29], BPSO [43], GA [44], HHO [45], WOA [46], SSA [47], GWO [48], and a method using all features (ALL) using the CWRU dataset. The parameter settings of the nine algorithms are shown in Table 5.
Table 5. Parameter settings.
The overall fault diagnosis model is shown in Figure 10. During feature selection training, the KNN classifier (in this article k = 5) and a 10-fold cross-validation were used to assess the classification performance of the feature subset.
Figure 10. Diagram of the fault diagnosis model.

5.2.3. Experimental Results and Analysis

The t-distributed stochastic neighbor embedding (t-SNE) manifold learning algorithm [49] transforms the feature set before and after feature selection into a two-dimensional distribution, which can more intuitively show the quality of the feature subset. The resulting distribution map is shown in Figure 11. When feature selection is not performed, fault types two, three, four, six, and nine are tightly distributed and overlap. After feature selection based on the CHBCS algorithm, irrelevant interference features are eliminated and all fault states are effectively separated. Therefore, the method proposed in this paper can effectively remove the redundant features that affect classification and improve the performance of the classifier.
Figure 11. Two-dimensional distribution map of t-SNE.
Table 6 shows the diagnostic performance obtained using the nine algorithms and all feature (ALL) methods.
Table 6. Classification results of the CHBCS algorithm and other algorithms.
From the perspective of diagnostic performance, the nine algorithms shown in Table 6 achieve better diagnostic accuracy than when using all features. The average diagnostic error rate of the CHBCS algorithm was only 1.13%, which is 81.17% lower than the ALL method and the lowest diagnostic error rate among the nine algorithms. In the CWRU dataset, 50 features were extracted. The nine algorithms reduced dimensionality to varying degrees. The average number of selected features with the CHBCS algorithm was 4.75, and the feature selection rate was only 9.5%. The Compared to the ALL, BCS, EBCS, BPSO, GA, HHO, WOA, SSA, and GWO algorithms, the fitness value of the CHBCS algorithm was 87.20%, 55.02%, 50.62%, 61.87%, 40.84%, 50.25%, 42.23%, 68.23%, and 40.30% lower, respectively. The experimental results show that the CHBCS algorithm can significantly improve diagnostic performance.
Convergence performance analysis of the algorithm was also performed. The fitness convergence curves obtained by the nine algorithms after being run 20 times independently are shown in Figure 12. The average fitness of the SSA is 0.0620, which is the worst among the nine algorithms. It converges to the global optimum before the 10th iteration, indicating that the SSA algorithm easily falls toward the local optimum. The quality of the feature subset searched by the CHBCS algorithm is significantly better than the other eight algorithms. By comparing the iterative curves of the CHBCS, BCS, and EBCS, it can be seen that the CHBCS algorithm can effectively improve the convergence speed of the CS algorithm. In addition, the initial value of the convergence curve of the CHBCS algorithm is less than 0.07, and the quality of the initial population is improved. The results prove the effectiveness of the clustering hybrid initialization method.
Figure 12. Fitness convergence curves for nine kinds of algorithm.
The optimal fitness values obtained by the nine algorithms after being run 20 times independently are shown in Figure 13. The fitness value of the CHBCS algorithm is the lowest among the nine broken lines, mainly falling between 0.03 and 0.01. Compared to the other eight algorithms, the fluctuations in the line chart of the CHBCS algorithm are the smallest. The above analysis shows that the CHBCS algorithm has better stability than the BCS, EBCS, BPSO, GA, HHO, WOA, SSA, and GWO algorithms.
Figure 13. The optimal fitness value of nine algorithms after being run 20 times.
In rolling bearing fault diagnosis, the CHBCS method can effectively identify various bearing fault types. In terms of diagnostic accuracy, feature number, convergence speed, and stability, it has outperformed other traditional optimization algorithms. Therefore, the CHBCS feature selection method can effectively solve the problem of bearing fault diagnosis.

6. Conclusions

This paper proposes a feature selection method based on the clustering hybrid binary cuckoo search to address the problem of low accuracy in rolling bearing fault diagnosis that is caused by irrelevant and redundant features. Hilbert–Huang transform technology effectively realizes the strategy of extracting features from the signal time–frequency domain. The clustering hybrid initialization method proposed by the CHBCS algorithm uses the similarity between features to cluster, effectively reducing redundant features and controlling the number of features through a hybrid initialization method. In addition, the proposed mutation strategy based on Levy flight and the dynamic Pa probability strategy can effectively promote the algorithm’s convergence. Finally, an experiment was carried out on the CWRU rolling bearing data. The results show that the diagnostic error rate of the CHBCS algorithm was only 1.13% and that a low-dimensional feature subset that is more sensitive to the fault state could be obtained, which thus provides an effective solution for rolling bearing fault diagnosis. The method will be further improved in future studies and field measured data will be used for experiments to eliminate simulation errors.

Author Contributions

Conceptualization, L.S. and Y.X.; methodology, Y.X.; software, T.C.; validation, L.S., Y.X. and T.C.; formal analysis, Y.X.; resources, L.S. and T.C.; data curation, B.F.; writing—original draft preparation, Y.X.; writing—review and editing, L.S. and T.C.; visualization, B.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant nos. 62173127 and 61973104); the Scientific and Technological Innovation Leaders in Central Plains (No. 224200510008); the Henan Excellent Young Scientists Fund (No. 212300410036); the Program for Science and Technology Innovation Talents in Universities of Henan Province (Grant no. 21HASTIT029); the Training Program for Young Backbone Teachers in Universities of Henan Province (Grant no. 2019GGJS089); the Innovative Funds Plan of Henan University of Technology (Grant no. 2020ZKCJ06); the Zhengzhou Science and Technology Collaborative Innovation Project (No. 21ZZXTCX06); and the Cultivation Program of Young Backbone Teachers in Henan University of Technology (Grant chentianfei).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors thank Case Western Reserve University for providing free access to the bearing vibration experimental data on their website.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Neupane, D.; Seok, J. Bearing Fault Detection and Diagnosis Using Case Western Reserve University Dataset With Deep Learning Approaches: A Review. IEEE Access 2020, 8, 93155–93178. [Google Scholar] [CrossRef]
  2. Nyanteh, Y.; Edrington, C.; Srivastava, S.; Cartes, D. Application of artificial intelligence to real-time fault detection in permanent-magnet synchronous machines. IEEE Trans. Ind. Appl. 2013, 49, 1205–1214. [Google Scholar] [CrossRef]
  3. Lin, S.L. Application of Machine Learning to a Medium Gaussian Support Vector Machine in the Diagnosis of Motor Bearing Faults. Electronics 2021, 10, 2266. [Google Scholar] [CrossRef]
  4. Hou, J.B.; Wu, Y.; Ahmad, A.S.; Gong, H.; Liu, L. A Novel Rolling Bearing Fault Diagnosis Method Based on Adaptive Feature Selectionand Clustering. IEEE Access 2021, 9, 99756–99767. [Google Scholar] [CrossRef]
  5. Zhang, X.; Zhao, B.Y.; Lin, Y. Machine Learning Based Bearing Fault Diagnosis Using the Case Western Reserve University Data: A Review. IEEE Access 2021, 9, 155598–155608. [Google Scholar] [CrossRef]
  6. Dhamande, L.S.; Chaudhari, M.B. Compound gear-bearing fault feature extraction using statistical features based on time-frequency method. Measurement 2018, 125, 63–77. [Google Scholar] [CrossRef]
  7. Tian, J.; Morillo, C.; Azarian, M.H.; Pecht, M. Motor Bearing Fault Detection Using Spectral Kurtosis-Based Feature Extraction Coupled With K -Nearest Neighbor Distance Analysis. IEEE Trans. Ind. Electron. 2016, 63, 1793–1803. [Google Scholar] [CrossRef]
  8. Guan, X.Y.; Chen, G. Sharing pattern feature selection using multiple improved genetic algorithms and its application in bearing fault diagnosis. J. Mech. Sci. Technol. 2019, 33, 129–138. [Google Scholar] [CrossRef]
  9. Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A survey on evolutionary computation approaches to feature selection. IEEE Trans. Evol. Comput. 2016, 20, 606–626. [Google Scholar] [CrossRef]
  10. Mao, W.T.; Wang, L.Y.; Feng, N. A New Fault Diagnosis Method of Bearings Based on Structural Feature Selection. Electronics 2020, 8, 1406. [Google Scholar] [CrossRef]
  11. Liu, H.; Yu, L. Toward integrating feature selection algorithms for classification and clustering. IEEE Trans. Knowl. Data Eng. 2005, 17, 491–502. [Google Scholar]
  12. Tam, W.; Cheng, L.; Wang, T.; Xia, W.; Chen, L. An improved genetic algorithm based robot path planning method without collision in confined workspace. Int. J. Model. Identif. Control 2019, 33, 120–129. [Google Scholar] [CrossRef]
  13. Shafqat, W.; Malik, S.; Lee, K.T.; Kim, D.H. PSO Based Optimized Ensemble Learning and Feature Selection Approach for Efficient Energy Forecast. Electronics 2021, 10, 2188. [Google Scholar] [CrossRef]
  14. Al-tashi, Q.; Kadir, S.J.A.; Rais, H.M.; Mirjalili, S.; Alhussian, H. Binary Optimization Using Hybrid Grey Wolf Optimization for Feature Selection. IEEE Access 2019, 7, 39496–39508. [Google Scholar] [CrossRef]
  15. Anter, A.M.; Muntaz, A. Feature selection strategy based on hybrid crow search optimization algorithm integrated with chaos theory and fuzzy c-means algorithm for medical diagnosis problems. Soft Comput. 2020, 24, 1565–1584. [Google Scholar] [CrossRef]
  16. Yang, X.S.; Deb, S. Cuckoo Search via Lévy flights. In Proceedings of the 2009 World Congress on Nature & Biologically Inspired Computing, Coimbatore, India, 9–11 December 2009. [Google Scholar]
  17. Gandomi, A.H.; Yang, X.S.; Alavi, A.H. Cuckoo search algorithm: A metaheuristic approach to solve structural optimization problems. Eng. Comput. 2013, 29, 17–35. [Google Scholar] [CrossRef]
  18. Xue, B.; Zhang, M.J.; BrowneSchool, W.N. Particle swarm optimisation for feature selection in classification: Novel initialisation and updating mechanisms. Appl. Soft Comput. 2014, 18, 261–276. [Google Scholar] [CrossRef]
  19. Cui, B.; Pan, H.; Wang, Z. Fault diagnosis of roller bearings base on the local wave and approximate entropy. J. North Univ. China 2012, 33, 552–558. [Google Scholar]
  20. Zheng, J.; Cheng, J.; Yang, Y.; Luo, S. A rolling bearing fault diagnosis method based on multi-scale fuzzy entropy and variable predictive model-based class discrimination. Mech. Mach. Theory 2014, 78, 187–200. [Google Scholar] [CrossRef]
  21. Tang, X.H.; Wang, J.C.; Lu, J.G.; Liu, G.K.; Chen, J.D. Improving Bearing Fault Diagnosis Using Maximum Information Coefficient Based Feature Selection. Appl. Sci. 2018, 8, 2143. [Google Scholar] [CrossRef]
  22. Tang, X.H.; He, Q.; Gu, X.; Li, C.J.; Zhang, H.; Lu, J.G. A novel bearing fault diagnosis method based on GL-mRMR-SVM. Processes 2020, 8, 784. [Google Scholar] [CrossRef]
  23. Lu, L.; Yan, J.H.; de Silva, C.W. Dominant feature selection for the fault diagnosis of rotary machines using modified genetic algorithm and empirical mode decomposition. J. Sound Vib. 2015, 344, 464–483. [Google Scholar] [CrossRef]
  24. Rauber, T.W.; De, A.B.F.; Varejao, F.M. Heterogeneous Feature Models and Feature Selection Applied to Bearing Fault Diagnosis. IEEE Trans. Ind. Electron. 2015, 62, 637–646. [Google Scholar] [CrossRef]
  25. Shan, Y.H.; Zhou, J.Z.; Jiang, W.; Liu, J.; Xu, Y.H.; Zhao, Y.J. A fault diagnosis method for rotating machinery based on improved variational mode decomposition and a hybrid artificial sheep algorithm. Meas. Sci. Technol. 2019, 30, 055002. [Google Scholar] [CrossRef]
  26. Nayana, B.R.; Geethanjali, P. Improved Identification of Various Conditions of Induction Motor Bearing Faults. IEEE Trans. Instrum. Meas. 2020, 69, 1908–1919. [Google Scholar] [CrossRef]
  27. Lee, C.Y.; Le, T.A. An Enhanced Binary Particle Swarm Optimization for Optimal Feature Selection in Bearing Fault Diagnosis of Electrical Machines. IEEE Access 2021, 9, 102671–102686. [Google Scholar] [CrossRef]
  28. Rodrigues, D.; Pereira, L.A.M.; Almeida, T.N.S.; Papa, J.P.; Souza, A.N.; Ramos, C.C.O.; Yang, X.S. BCS: A Binary Cuckoo Search algorithm for feature selection. In Proceedings of the 2013 IEEE International Symposium on Circuits and Systems, Beijing, China, 19–23 May 2013. [Google Scholar]
  29. Salesi, S.; Cosma, G. A novel extended binary cuckoo search algorithm for feature selection. In Proceedings of the 2017 2nd International Conference on Knowledge Engineering and Applications 2017, London, UK, 21–23 October 2017. [Google Scholar]
  30. Pandey, A.C.; Rajipoot, D.S.; Saraswat, M. Feature selection method based on hybrid data transformation and binary binomial cuckoo search. J. Ambient. Intell. Humaniz. Comput. 2019, 11, 719–738. [Google Scholar] [CrossRef]
  31. Abd El Aziz, M.; Hassanien, A.E. Modified cuckoo search algorithm with rough sets for feature selection. Neural. Comput. Appl. 2018, 29, 925–934. [Google Scholar] [CrossRef]
  32. Kelidari, M.; Hamidzadeh, J. Feature selection by using chaotic cuckoo optimization algorithm with levy flight, opposition-based learning and disruption operator. Soft Comput. 2021, 25, 2911–2933. [Google Scholar] [CrossRef]
  33. Alia, A.; Taweel, A. Enhanced Binary Cuckoo Search with Frequent Values and Rough Set Theory for Feature Selection. IEEE Access 2021, 9, 119430–119453. [Google Scholar] [CrossRef]
  34. Kabla, A.; Mokrani, K. Bearing fault diagnosis using Hilbert-Huang transform (HHT) and support vector machine (SVM). Mech. Ind. 2016, 17, 3. [Google Scholar] [CrossRef]
  35. Chegini, S.N.; Bagheri, A.; Najafi, F. A new intelligent fault diagnosis method for bearing in different speeds based on the FDAF-score algorithm, binary particle swarm optimization, and support vector machine. Soft Comput. 2019, 24, 10005–10023. [Google Scholar] [CrossRef]
  36. Sosa-Cabrera, G.; García-Torres, M.; Gómez-Guerrero, S.; Schaerer, C.E.; Divina, F. A Multivariate approach to the Symmetrical Uncertainty Measure: Application to Feature Selection Problem. Inf. Sci. 2019, 2019, 494. [Google Scholar] [CrossRef]
  37. Chen, Q.; Zhao, L.; Lu, J.; Kuang, G.; Wang, N.; Jiang, Y. Modified two-dimensional Otsu image segmentation algorithm and fast realisation. IET Image Process. 2012, 6, 426–433. [Google Scholar] [CrossRef]
  38. Traag, V.A. Faster unfolding of communities: Speeding up the Louvain algorithm. Phys. Rev. E 2015, 92, 032801. [Google Scholar] [CrossRef] [PubMed]
  39. Wang, Q.F.; Wang, S.; Wei, B.K.; Chen, W.W.; Zhang, Y.F. Weighted K-NN Classification Method of Bearings Fault Diagnosis with Multi-Dimensional Sensitive Features. IEEE Access 2021, 9, 45428–45440. [Google Scholar] [CrossRef]
  40. Ouadfel, S.; Elaziz, M.A. Enhanced Crow Search Algorithm for Feature Selection. Expert Syst. Appl. 2020, 159, 113572. [Google Scholar] [CrossRef]
  41. Fushiki, T. Estimation of prediction error by using K-fold cross-validation. Stat. Comput. 2011, 21, 137–146. [Google Scholar] [CrossRef]
  42. Li, W.; Chao, X.Q. Improved particle swarm optimization method for feature selection. J. Front. Comput. Sci. Technol. 2019, 13, 990–1004. [Google Scholar]
  43. Mirjalili, S.; Lewis, A. S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol. Comput. 2013, 9, 1–14. [Google Scholar] [CrossRef]
  44. Tan, F.; Fu, X.Z.; Zhang, Y.P. A genetic algorithm-based method for feature subset selection. Soft Comput. 2008, 12, 111–120. [Google Scholar] [CrossRef]
  45. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H.L. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Sys. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  46. Mirjalili, S.; Lewis, A. The whale optimization algorithm. Adv. Eng. Softw. 2016, 95, 51–67. [Google Scholar] [CrossRef]
  47. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp swarm algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  48. Mirjalili, S.; Mirjalili, S.M.; Lewis, A. Grey wolf optimizer. Adv. Eng. Softw. 2014, 69, 46–61. [Google Scholar] [CrossRef]
  49. Jiang, W.; Xu, Y.H.; Chen, Z.; Zhang, N.; Zhou, J.Z. Fault diagnosis for rolling bearing using a hybrid hierarchical method based on scale-variable dispersion entropy and parametric t-SNE algorithm. Measurement 2022, 191, 110843. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.