Multi-Manifold Learning Fault Diagnosis Method Based on Adaptive Domain Selection and Maximum Manifold Edge
Abstract
1. Introduction
- (1)
- Most of the fault diagnosis methods for rolling bearings focus on the mean value of all samples or the mean value of a certain class of samples, while the description of local information is ignored. Therefore, based on modeling the multi-manifold subspace and using within-class and between-class differences, this paper innovatively applies the multi-manifold learning dimensionality reduction algorithm to realize the efficient classification of fault modes of rolling bearings.
- (2)
- To solve the problem of sensitive parameter selection in the field of multi-manifold learning, the relationship between local sample density and neighborhood parameter values was tried, and the local neighborhood parameter size of each sample point was adjusted adaptively by defining the density scaling factor, so as to build the within- and between-manifold graphs adaptively. This helps to recognize boundaries between multiple manifolds, making the data samples more distinguishable.
- (3)
- Two different types of datasets are selected for experimental analysis, and multiple nonlinear dimension reduction algorithms (KPCA, LLE, LLRMM, and UMAP) are compared to verify the proposed algorithm’s advantages in dimension reduction effect and recognition rate. The results show that using a manifold space for feature fusion reduction can lead to better classification.
2. Manifold Learning
- (1)
- Search for nearest neighbor points: The first step in the LLE algorithm is determining the nearest neighbors of each data point. Given a dataset , where is the original dimensionality of the data and is the number of samples. For each sample , the neighbors are determined by either choosing the closest Euclidean distance points or by selecting points within a fixed radius around the sample. These neighbors are then used to form the neighborhood index set ensuring that the local relationships between data points are preserved.
- (2)
- Calculate the optimal reconstruction weight matrix: each sample point is linearly reconstructed from its nearest neighbor data points , and the reconstruction error of all data points can be expressed in the following mathematical form:
- (3)
- Represent low-dimensional embedding information: The spatial position relationship between all sample points and the nearest neighbor set is kept unchanged in the mapping from high to low dimensions. Namely is kept unchanged after linear reconstruction, and the low-dimensional reconstruction error is minimized.
3. ALLRMM Model
3.1. Optimization of Local Neighborhood Selection Method
- (1)
- Sensitivity analysis of neighborhood selection: Generally, an optimal domain size should be taken. If the domain size is too small, the continuous topological space will be constructed into multiple separated subgraphs, resulting in the loss of connections between certain points and the failure to reflect the global characteristics. If it is too large, the neighborhood of certain sample data points will come from other folding planes.
- (2)
- Sample density analysis: As shown in Figure 4, owing to the unevenly distributed manifold structure, overlapped neighborhoods are conducive to information transmission and the protection of the local linearity hypothesis. Therefore, to maximize overlap, four places with high local density of sample points and more sample points in the neighborhood are selected to make the neighborhood larger, resulting in neighborhood overlap.
- (3)
- Adaptive neighborhood selection: The density scaling factor of each sample point was calculated according to the density scaling factor algorithm proposed in the previous section, and the initial neighborhood value was adjusted adaptively according to the size of the density scaling factor. Consequently, the neighborhood value parameter of the high-density sample point increased while that of the low-density sample point decreased. Considering that the extreme value of the initial neighborhood value significantly affects the adaptive results, the initial extreme neighborhood value is preliminarily adjusted, followed by the adaptive iteration to get the ideal neighborhood value. The initial neighborhood value is based on the features of the datasets. The k value adaptation is performed using the density scaling factor algorithm. The density scaling factor is computed based on the local density of sample points in the high-dimensional space. As the local density of a sample point increases, becomes larger; conversely, as the inter-sample density decreases, becomes smaller. The adaptive iteration process is determined as follows:
3.2. Improved Edge Margin Algorithm for Local Linear Embedded Manifolds
- (1)
- The construction of manifold graphs: The within-manifold graphs reflect the local relations in the manifold comprising similar data. For any sample point (the i-th column of dataset ) in the within-manifold graph, sample points with the same category and minimum distance are selected to form the local neighborhood of sample point . Therefore, in this neighborhood, similar to that in LLE, the minimum linear error is expressed by the closest neighbor points of the same class of sample point .
- (2)
- Manifold margin definition: To describe the degree of dispersion between manifolds with different class labels, a new manifold margin is defined. The manifold margin is the distance between points on the manifold and other manifolds minus the within-manifold graph distance of the manifold . The within-manifold graph distance of the manifold can be described as the degree of convergence and dispersion inside the manifold and can be expressed by the within-manifold graph divergence matrix of the manifold. Therefore, the definition of manifold margin is as shown below:
- (3)
- Optimization solution: According to the within-manifold and between-manifold graph divergence matrices, and the previously defined manifold edge distance, the purpose of the proposed algorithm is to find a subspace on which data of different manifolds can be more easily distinguished. That is, the edge distance of manifolds should be maximized in the low-dimensional subspace, and the within-class difference should be minimized.
Algorithm 1 Algorithm for ALLRMM |
Input: original data , data category , initial nearest neighbor number , low-dimensional space dimension ; Output: Transform matrix and data after dimensionality reduction. Step 1: Determine the domain size of each data sample point adaptively through density scaling factor ; Step 2: Construct the within-manifold and between-manifold graphs by adaptive selection of . Step 3: Calculate the weight matrix of the corresponding between-manifold and within-manifold graphs using the formula and calculate the corresponding divergence matrix ; Step 4: Calculate the manifold edge distance according to the manifold divergence matrix; Step 5: Solve the feature decomposition equation . The eigenvectors corresponding to the first maximum eigenvalues are selected to form the transformation matrix and obtain the data after dimensionality reduction by solving . |
3.3. Analysis of the Initial Dimensionality Reduction Effect of Multi-Manifolds
3.4. Complexity
4. Experimental Design
4.1. Dataset
4.2. Data Preprocessing
5. Analysis of Experimental Result
5.1. Analysis of the Dimension Reduction Effect of Multi-Manifold
5.2. Analysis of Algorithm Recognition Accuracy
6. Conclusions
Author Contributions
Funding
Conflicts of Interest
References
- Huang, D.; Hua, X.; Mi, B.; Liu, Y.; Zhang, Z. Incipient fault diagnosis on active disturbance rejection control. Sci. China Inf. Sci. 2021, 65, 199202. [Google Scholar] [CrossRef]
- Wang, C.; Wang, H.; Liu, M. A dual-attention feature fusion network for imbalanced fault diagnosis with two-stream hybrid generated data. J. Intell. Manuf. 2024, 35, 1707–1719. [Google Scholar] [CrossRef]
- Ding, X.; Li, Q.; Lin, L.; He, Q.; Shao, Y. Fast time-frequency manifold learning and its reconstruction for transient feature extraction in rotating machinery fault diagnosis. Measurement 2019, 141, 380–395. [Google Scholar] [CrossRef]
- Huang, Z.; Shao, J.; Guo, W.; Li, W.; Zhu, J.; He, Q.; Fang, D. Tool Wear Prediction Based on Multi-Information Fusion and Genetic Algorithm-Optimized Gaussian Process Regression in Milling. IEEE Trans. Instrum. Meas. 2023, 72, 2516716. [Google Scholar] [CrossRef]
- Qin, Y.; Yang, J.; Zhou, J.; Pu, H.; Mao, Y. A new supervised multi-head self-attention autoencoder for health indicator construction and similarity-based machinery RUL prediction. Adv. Eng. Inform. 2023, 56, 101973. [Google Scholar] [CrossRef]
- Tenenbaum, J.B.; de Silva, V.; Langford, J.C. A Global Geometric Framework for Nonlinear Dimensionality Reduction. Science 2000, 290, 2319–2323. [Google Scholar] [CrossRef] [PubMed]
- Roweis, S.T.; Saul, L.K. Nonlinear Dimensionality Reduction by Locally Linear Embedding. Science 2000, 290, 2323–2326. [Google Scholar] [CrossRef]
- Belkin, M.; Niyogi, P. Laplacian eigenmaps for dimensionality reduction and data representation. Neural Comput. 2003, 15, 1373–1396. [Google Scholar] [CrossRef]
- Peng, K.; Guo, Y. Fault detection and quantitative assessment method for process industry based on feature fusion. Measurement 2022, 197, 111267. [Google Scholar] [CrossRef]
- Su, Z.; Tang, B.; Zhao, M. Rotating machinery fault diagnosis based on multiple fault manifolds. J. Vib. Eng. 2015, 28, 309–315. [Google Scholar]
- Chen, G.; Liu, F.; Huang, W. Sparse discriminant manifold projections for bearing fault diagnosis. J. Sound Vib. 2017, 399, 330–344. [Google Scholar] [CrossRef]
- He, Y.-L.; Li, K.; Zhang, N.; Xu, Y.; Zhu, Q.-X. Fault Diagnosis Using Improved Discrimination Locality Preserving Projections Integrated With Sparse Autoencoder. IEEE Trans. Instrum. Meas. 2021, 70, 3527108. [Google Scholar] [CrossRef]
- Su, Z.; Xu, H.; Luo, J.; Zheng, K.; Zhang, Y. Fault Diagnosis Method Based on a New Manifold Learning Framework. J. Intell. Fuzzy Syst. 2018, 34, 3413–3427. [Google Scholar] [CrossRef]
- Zhang, N.; Tian, Y.; Wang, X.-W.; Xu, Y.; Zhu, Q.-X.; He, Y.-L. Novel Bootstrap-Based Discriminant NPE Integrated With Orthogonal LPP for Fault Diagnosis. IEEE Trans. Instrum. Meas. 2023, 72, 3506309. [Google Scholar] [CrossRef]
- Ma, Y.; Wu, X. Local-Dictionary Sparsity Discriminant Preserving Projections for Rotating Machinery Fault Diagnosis Based on Pre-Selected Multi-Domain Features. IEEE Sens. J. 2022, 22, 8781–8794. [Google Scholar] [CrossRef]
- Yang, C.; Ma, S.; Han, Q. Unified discriminant manifold learning for rotating machinery fault diagnosis. J. Intell. Manuf. 2022, 34, 3483–3494. [Google Scholar] [CrossRef]
- Xu, Z.M.; Hu, R.M.; Chen, J.; Chen, C.; Li, J.F.; Li, H.Y. Semisupervised Discriminant Multi-manifold Analysis for Action Recognition. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2951–2962. [Google Scholar] [CrossRef]
- Carlini, L.P.; Junior, G.F.M.; Giraldi, G.A.; Thomaz, C.E. A new method of selecting safe neighbors for the Riemannian Manifold Learning algorithm. IEEE Lat. Am. Trans. 2021, 19, 89–97. [Google Scholar] [CrossRef]
- Yu, W.; Bian, J.; Nie, F.; Wang, R.; Li, X. Unsupervised Subspace Learning With Flexible Neighboring. IEEE Trans. Neural Netw. Learn. Syst. 2023, 34, 2043–2056. [Google Scholar] [CrossRef]
- Wan, Z.; Yang, R.; Huang, M.; Liu, W.; Zeng, N. EEG fading data classification based on improved manifold learning with adaptive neighborhood selection. Neurocomputing 2022, 482, 186–196. [Google Scholar] [CrossRef]
- Zhao, L.; Huang, D.R.; Song, J. Fault Diagnosis Method Based on Fractal Theory and Its Application in Wind Power Systems. Acta Armamentarii 2012, 8, 167–173. [Google Scholar]
- Huang, H.; Baddour, N. Bearing vibration data collected under time-varying rotational speed conditions. Data Brief 2018, 21, 1745–1749. [Google Scholar] [CrossRef]
- Li, B.; Guo, W.; Zhang, X.-L. A global manifold margin learning method for data feature extraction and classification. Eng. Appl. Artif. Intell. 2018, 75, 94–101. [Google Scholar] [CrossRef]
Bearing Health Conditions | Speed Varying Conditions | |||
---|---|---|---|---|
Increasing Speed | Decreasing Speed | Increasing Then Decreasing Speed | Decreasing then Increasing Speed | |
Healthy | H-A-1 | H-B-1 | H-C-1 | H-D-1 |
H-A-2 | H-B-2 | H-C-2 | H-D-2 | |
H-A-3 | H-B-3 | H-C-3 | H-D-3 | |
Faulty(inner race fault) | I-A-1 | I-B-1 | I-C-1 | I-D-1 |
I-A-2 | I-B-2 | I-C-2 | I-D-2 | |
I-A-3 | I-B-3 | I-C-3 | I-D-3 | |
Faulty(outer race fault) | O-A-1 | O-B-1 | O-C-1 | O-D-1 |
O-A-2 | O-B-2 | O-C-2 | O-D-2 | |
O-A-3 | O-B-3 | O-C-3 | O-D-3 |
Feature Name | Mathematical Formula | Feature Description |
---|---|---|
Mean value | Arithmetic average of signal amplitudes | |
Maximum amplitude | Highest instantaneous value in the signal | |
Minimum amplitude | Lowest instantaneous value in the signal | |
Peak-to-Peak value | Dynamic range of signal amplitudes | |
Root mean square | Quadratic mean representing signal energy | |
Variance | Measure of signal dispersion | |
Raw skewness | Unnormalized third moment for asymmetry | |
Raw kurtosis | Unnormalized fourth moment for tail heaviness | |
Smoothness value | Sensitivity to small amplitude variations | |
Normalized kurtosis | Energy-compensated impulse detection | |
Normalized skewness | Variance-normalized distribution asymmetry | |
Waveform factor | Ratio of RMS to mean absolute value | |
Crest factor | Peak-to-RMS ratio for impulse detection | |
Peak-to-Mean ratio | Peak amplitude relative to DC component | |
Smooth Peak ratio | Peak normalized by smoothness value | |
Absolute mean | Magnitude of DC component |
Feature Name | Mathematical Formula | Feature Description |
---|---|---|
Spectral centroid | Energy-weighted mean frequency, denotes frequency weight of the i-th component | |
Mean square frequency | Second moment of spectral distribution | |
Root mean square frequency | Effective bandwidth measure | |
Frequency variance | Dispersion around spectral centroid | |
Spectral standard deviation | Standard deviation of frequency distribution |
Feature Name | Mathematical Formula | Feature Description |
---|---|---|
Minimum singularity exponent | Exponent for most singular regions | |
Maximum singularity exponent | Exponent for smoothest regions | |
Multifractal spectrum width | Range of singularity strengths | |
Maximum fractal dimension | Peak dimension in multifractal spectrum | |
Spectrum asymmetry | Asymmetry measure of singularity spectrum | |
Generalized Hurst exponent | Scaling exponent function | |
Mean Hurst exponent | Average scaling behavior | |
Hurst exponent range | Multifractality strength indicator | |
Singularity exponent range | Width of singularity support |
Initial Neighborhood Value | 2 | 5 | 10 | 15 | 20 | 25 | |
---|---|---|---|---|---|---|---|
Rolling bearing fault dataset adaptive value | LLE algorithm | 2 | 5 | 10 | 15 | 20 | 25 |
LLRMM algorithm | 2 | 5 | 10 | 15 | 20 | 25 | |
ALLRMM algorithm | 2 | 3 | 2 | 2 | 3 | 2 |
Method | Average Recognition Rate |
---|---|
Total features | 91.67% |
LLE | 90.00% |
KPCA | 92.00% |
UMAP | 93.30% |
LLRMM | 93.33% |
ALLRMM | 96.67% |
Method | Average Recognition Rate |
---|---|
Total features | 86.67% |
KPCA | 71.33% |
UMAP | 76.65% |
LLE | 81.00% |
LLRMM | 89.33% |
ALLRMM | 92.33% |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhao, L.; Ding, J.; Li, P.; Chi, X. Multi-Manifold Learning Fault Diagnosis Method Based on Adaptive Domain Selection and Maximum Manifold Edge. Sensors 2025, 25, 5384. https://doi.org/10.3390/s25175384
Zhao L, Ding J, Li P, Chi X. Multi-Manifold Learning Fault Diagnosis Method Based on Adaptive Domain Selection and Maximum Manifold Edge. Sensors. 2025; 25(17):5384. https://doi.org/10.3390/s25175384
Chicago/Turabian StyleZhao, Ling, Jiawei Ding, Pan Li, and Xin Chi. 2025. "Multi-Manifold Learning Fault Diagnosis Method Based on Adaptive Domain Selection and Maximum Manifold Edge" Sensors 25, no. 17: 5384. https://doi.org/10.3390/s25175384
APA StyleZhao, L., Ding, J., Li, P., & Chi, X. (2025). Multi-Manifold Learning Fault Diagnosis Method Based on Adaptive Domain Selection and Maximum Manifold Edge. Sensors, 25(17), 5384. https://doi.org/10.3390/s25175384