Next Article in Journal
Lower-Limb Motor Imagery Recognition Prototype Based on EEG Acquisition, Filtering, and Machine Learning-Based Pattern Detection
Next Article in Special Issue
Fault Diagnosis of Rolling Bearings Using Denoising Multi-Channel Mixture of CNN and Mamba-Enhanced Adaptive Self-Attention LSTM
Previous Article in Journal
A Highly Sensitive Silicon Nanowire Array Field Effect Transistor Biosensor for Detecting HBV-DNA and AFP
Previous Article in Special Issue
Structural Component Identification and Damage Localization of Civil Infrastructure Using Semantic Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Diagnostic Framework for Interpretable Bearing Fault Classification Using CNN and Dual-Stage Feature Selection

by
Mohamed Elhachemi Saouli
1,2,*,
Mostefa Mohamed Touba
2 and
Adel Boudiaf
3
1
LESIA Laboratory of Research, University of Mohamed Khider Biskra, Biskra 07000, Algeria
2
Li3CUB Laboratory of Research, University of Mohamed Khider Biskra, Biskra 07000, Algeria
3
Research Center in Industrial Technologies—CRTI, Cheraga, P.O. Box 64, Algiers 16014, Algeria
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(20), 6386; https://doi.org/10.3390/s25206386
Submission received: 18 August 2025 / Revised: 2 October 2025 / Accepted: 9 October 2025 / Published: 16 October 2025
(This article belongs to the Special Issue AI-Assisted Condition Monitoring and Fault Diagnosis)

Abstract

Timely and accurate fault diagnosis in rotary machinery is essential for ensuring system reliability and minimizing unplanned downtime. While deep learning approaches, particularly Convolutional Neural Networks (CNNs), have demonstrated strong performance in vibration-based fault classification, their limited interpretability poses challenges for adoption in safety-critical environments. To address this, the present study introduces a hybrid diagnostic framework that integrates CNN-based transfer learning with interpretable supervised classification, aiming to enhance both predictive accuracy and model transparency. A key innovation of this work lies in the dual-stage feature selection process, combining Analysis of Variance (ANOVA) and Permutation Feature Importance (PFI) to refine deep features extracted from a pre-trained VGG19 network. This strategy improves both dimensionality reduction and classification performance in a statistically grounded, model-agnostic manner. Furthermore, SHapley Additive exPlanations (SHAP) are employed to interpret the predictions, offering insight into the most influential features driving the classification decisions. Experimental evaluation on the Case Western Reserve University (CWRU) bearing dataset confirms the effectiveness of the proposed approach, achieving 100% classification accuracy using ten-fold cross-validation. By uniting high performance with transparent decision-making, the framework demonstrates strong potential for explainable and reliable fault diagnosis in industrial settings.

1. Introduction

The accurate and timely diagnosis of failures in rotary machinery remains a critical objective in industrial maintenance and operational efficiency. Among various diagnostic techniques, vibration analysis has emerged as one of the most effective due to its high sensitivity to structural and mechanical anomalies [1,2]. The Fourth Industrial Revolution has driven the integration of vibration analysis with artificial intelligence (AI) to support predictive maintenance and improve system reliability and performance [3]. However, this integration also introduces several new challenges, including data noise [4], complex feature extraction [5], high dimensionality [6], limited model interpretability [7], and the necessity for real-time processing [8]—factors that collectively complicate the development of accurate and dependable bearing fault diagnosis systems.
Early efforts predominantly employed machine learning (ML) models that relied on handcrafted feature extraction from vibration signals. Common approaches included time-domain descriptors such as root mean square (RMS), mean, kurtosis, and skewness [9]; frequency-domain methods such as Fourier Transform (FT) [10], Envelope Analysis (EA) [11]; and time–frequency analyses such as Short-Time Fourier Transform (STFT) [12], Hilbert–Huang Transform (HHT) [13], Wavelet Analysis (WA) [14], and Variational Mode Decomposition (VMD) [15]. These features were fed into classifiers to detect specific faults. While effective, such methods rely heavily on domain expertise, extensive preprocessing, and manual feature engineering [16]. Additionally, ML classifiers may underperform when faced with high-dimensional or noisy input features that are not sufficiently informative [17,18].
Beyond traditional approaches, recent developments have introduced data-driven methods such as phase diagram analysis, which extract features directly from the geometric structure of signals in reconstructed phase space, without requiring prior knowledge or extensive preprocessing [19]. These techniques have shown strong potential in capturing transient and nonlinear dynamics, thereby enhancing feature extraction for machinery fault diagnosis. However, their effectiveness can be constrained by noise sensitivity, computational complexity, and challenges in parameter selection.
To address these limitations, deep learning (DL) particularly using Convolutional Neural Networks (CNNs) has emerged as a powerful alternative. CNNs automatically learn hierarchical features from raw data, enabling end-to-end diagnostics. Transfer learning (TL) with pre-trained CNN models enhances efficiency by leveraging knowledge from large-scale datasets, thereby reducing the need for extensive labeled data and computational resources [20].
Despite their strong performance, CNNs are often criticized for their opaque internal representations. This limits their deployment in safety-critical settings, where decision transparency is essential. While feature visualization techniques provide some insight into CNN behavior, practical interpretability often remains elusive especially when decisions are based on complex feature hierarchies [21].
In parallel, feature selection has emerged as an important strategy to reduce dimensionality, improve learning efficiency, and enhance model transparency. Methods such as filter [22], wrapper [23], and embedded [24] approaches aim to identify the most informative features, thus facilitating better generalization and interpretability. However, when applied in isolation, these methods are still limited by the quality and nature of the input features [23,25].
To overcome these limitations, hybrid frameworks that combine deep feature extraction with traditional ML classification have been proposed. These approaches integrate various combinations of signal processing, CNN-based feature learning, and conventional classifiers. For example, a compressive sensing (CS)-based filter feature selection approach was proposed in [22], applied to features extracted using Fast Fourier Transform (FFT) and wavelet transform, with classification performed using Artificial Neural Networks (ANN) and Support Vector Machines (SVM). Another study introduced an ensemble learning framework based on transfer learning with a LeNet-5 model, where features extracted from wavelet-transformed signals were classified using a Random Forest (RF) algorithm [26]. Similarly, a three-stage model in [27] utilized time–frequency-domain feature extraction, followed by Neighborhood Component Analysis (NCA) for selection and classification via ML algorithms. Additional frameworks have explored combinations such as CNN-SVM [28], VMD-ANN [29], and Wavelet-CNN [30]. While many of these methods improve diagnostic accuracy, relatively few explicitly address the interpretability of the diagnostic process particularly at the interface between deep learning feature extraction and decision-level classification.
Most existing fault diagnosis models focus primarily on improving accuracy, with limited emphasis on understanding or explaining model decisions. As a result, the integration of deep feature learning with interpretable supervised classification remains an underexplored yet critical area for developing trustworthy diagnostic systems.
In light of these considerations, this study proposes a hybrid interpretable diagnostic framework that balances predictive performance with interpretability. The proposed approach combines CNN-based VGG19 transfer learning for feature extraction with a dual-stage feature selection mechanism, leveraging Analysis of Variance (ANOVA) and Permutation Feature Importance (PFI) to identify the most informative features. These refined features are then used to train multiple supervised learning algorithms, including Random Forest (RF), Extreme Gradient Boosting (XGBoost), Logistic Regression (LR), K-Nearest Neighbors (KNN), Artificial Neural Networks (ANN), and Support Vector Machines (SVM). This classification is followed by models’ explanation via SHAP (SHapley Additive exPlanations) analysis. This combination contributes to building a high performance and interpretable intelligent bearing fault diagnosis.
While the internal representations learned by CNNs remain partially opaque, this study improves interpretability by analyzing the CNNs’ feature outputs. Through systematic selection and analysis of these deep features, and their use in interpretable supervised models, the framework enables a clearer understanding of which characteristics contribute most significantly to fault classification. This enhances diagnostic transparency without sacrificing accuracy.
The proposed framework is validated using the Case Western Reserve University (CWRU) bearing dataset. The results demonstrate high classification performance results, achieving 100% accuracy. The comparative results show the enhancement in accuracy, adopting the dual-stage feature selection approach. Comparative analysis with existing techniques confirms the effectiveness and reliability of the proposed method. This paper mainly contributes to the following:
  • Proposes a dual-stage approach for dimensionality reduction and feature selection based on ANOVA and PFI, designed for general applicability across various supervised machine learning models.
  • Investigates the interpretability of the CNN black-box by analyzing the extracted deep features.
  • Provides ML model explanation.
  • Improves both classification performance and interpretability of intelligent bearing fault diagnosis framework.
The remainder of this paper is structured as follows: Section 1 is an introduction. Section 2 presents the bearing fault and vibration analysis in rotary machinery. Section 3 illustrates a review of the dataset used in this study. Section 4 introduces the feature extraction technique. Section 5 describes the proposed dual-stage feature selection approach. Section 6 outlines the overall methodology. Section 7 presents the experimental results, including a comparative study. Section 8 discusses the model interpretability using SHAP analysis, and finally, the paper conclusion in Section 9.

2. Bearing Faults and Vibration Analysis in Rotary Machines

Electric motors, as a central class of rotary machinery, are responsible for nearly half of the world’s total electricity usage and account for around 70% of industrial energy demand [31]. This extensive reliance highlights their critical importance in maintaining the continuity and efficiency of modern industrial processes. Electrical motors, in particular, are expected to operate continuously over their intended service life while avoiding unexpected shutdowns or severe malfunctions. However, despite their structural robustness and proven reliability, studies have shown that these machines are still prone to progressive wear and failure. Statistical investigations conducted by the Institute of Electrical and Electronics Engineers (IEEE) [32] reveal that bearing faults are implicated in approximately 40–50% of motor failures (Table 1), thereby establishing bearings as the single most common and severe source of malfunction in rotary machinery.
Several condition monitoring techniques have been developed to reduce failure risks and enable early fault detection.
The methods include the following:
  • Infrared Thermography (IT);
  • Ultrasonic Testing (UT);
  • Electrical Signature Analysis (ESA);
  • Acoustic sound-based condition monitoring (ASCM).
These methods provide useful diagnostic information by identifying thermal anomalies, high-frequency acoustic emissions, current irregularities, or abnormal noise patterns, respectively [33]. However, their effectiveness in detecting incipient mechanical faults is often limited, and their reliability may be affected by operational and environmental conditions.
By contrast, Vibration Analysis (VA) has become the most reliable and widely applied approach for bearing fault diagnosis. Localized defects in the inner race, outer race, or rolling elements generate periodic impact responses that manifest as distinct vibration signatures, enabling accurate fault localization and classification. The method offers comprehensive diagnostic information across the time, frequency, and time-frequency domains, while remaining non-intrusive, cost-effective, and easily deployable using standard accelerometers, thus supporting predictive maintenance and large-scale industrial monitoring.

3. Dataset Review

This study uses the vibration signals obtained from the Case Western Reserve University (CWRU) Bearing Data Centre [34]. The public accessibility of this dataset facilitates fair and consistent comparisons of various methodologies proposed for bearing fault diagnosis, thereby serving as both a benchmark and a source of motivation for this research. The CWRU experimental benchmark for the data collection shown in Figure 1 consists of an electric induction motor on one side, a torque transducer in the middle, and a dynamometer on the other end to simulate the load. The experimental data was obtained based on the accelerations generated by an electric motor under various conditions, including altering the nature of failure, severity, and the shaft rotational speed.
The bearing used in this study is the SKF 6205-2RSL JEM single-row deep groove ball bearing, which consists of four main components: the inner race, outer race, ball, and cage. These components are illustrated in Figure 2, while the detailed technical specifications of the bearing are provided in Table 2.
The bearing defects can be caused by different reasons, including insufficient lubrication, mechanical friction, rotor misalignment [2]. Accordingly, four types of bearing faults are defined: (the inner race fault (fIRF), the outer race fault (fORF), the ball fault (fBF), and the cage fault (fCF)). Each of them has a distinctive signature, known as characteristic frequency. The associated frequencies for each failure in function of rotation frequency (fr) of bearing are calculated using the following equations [35]:
f I R F = N 2 f r 1 + B D C D cos β
f O R F = N 2 f r 1 B D C D cos β
f B F = C D B D f r 1 B D 2 C D 2 cos 2 β
f C F = f r 2 1 B D C D cos β

4. Feature Extraction

Effective classification of bearing faults essentially depends on the ability to identify and leverage the most relevant features. Therefore, feature extraction is considered the most critical and sensitive step in the classification task, as it directly influences the accurate fault recognition and distinction. In this work, a CNN is employed to extract features from two-dimensional 2D image representation of vibration signal, based on transfer learning using the pre-trained model VGG19. A brief description of the pre-trained model is provided in the next section:

4.1. Signal Conversion

The concept behind signal-to-image conversion, as illustrated in Figure 3, involves transforming raw time-domain signals into images, offering an alternative preprocessing approach. To create an M × M image, a segmented signal of length M 2 is derived from the raw vibration signal. In this study, 64 × 64 grayscale images are generated, with the segmented signal having a length of 4096. The segmented signal values are represented as L(i), where i = 1, …, M2, while the pixel strength of the resulting image is denoted by P (j, k), where j, k = 1, …, M. The converted grayscale images are given by the following formula:
P j , k = r o u n d   L j 1 × M + K min L max L min L   × 255
Here, round {⋅} represents the rounding function, and the transformation maps values to the interval [0, 255], producing a grayscale image [36]. This straightforward mapping converts 1D raw signals into 2D textured representations, where each operational condition whether normal or indicative of a bearing defect corresponds directly to a distinct class of similarly textured images. Notably, this approach does not rely on any predefined parameters, making it a flexible and effective preprocessing technique. The choice of this specific 1D to 2D conversion method over other alternatives, such as wavelet scalograms, was guided by its simplicity and computational efficiency. Unlike scalograms, which require careful selection of mother wavelets and scale parameters potentially introducing additional complexity and computational overhead the proposed method is parameter-free and preserves the temporal texture of the signal directly. This enables effective feature representation while allowing for direct application of image-based deep learning techniques, such as convolutional neural networks, without prior transformation design.
Furthermore, the selection of a 64 × 64 image size, corresponding to a segment length of 4096, was based on practical considerations. This configuration strikes a balance between capturing sufficient signal detail and maintaining computational efficiency. A segment length of 4096 provides enough data to preserve fault-related characteristics, while the 64 × 64 image format ensures that the resulting images are neither too large to hinder model training nor too small to lose critical information. This resolution is also consistent with prior literature, where similar image sizes have proven effective in signal-based condition monitoring tasks.

4.2. Feature Extraction Using VGG19

VGG19, considered one of the most frequently used paradigms in image classification, is a deep learning architecture designed on a convolutional neural network, with nineteen layers, sixteen of which are convolutional, while the remaining three are fully connected layers. Its design stands on 3 by 3 filters that allow the model to catch the most relevant details. VGG19 uses the ImageNet dataset. This model is designed to classify images into one of a thousand distinct object categories [37]. As depicted in the architecture shown in Figure 4a, the first 16 layers consist exclusively of convolutional operations used for feature extraction. These convolutional layers are organized into five distinct blocks, each followed by a max-pooling layer to progressively reduce spatial dimensions while preserving essential features. The remaining layers constitute the classification component of the network. The model input consists of images resized to 224 × 224 resolution with three color channels, and the output corresponds to image labels. In this study, the original 64 × 64 grayscale images derived from vibration signals are resized to 224 × 224 and replicated across three channels prior to input, ensuring compatibility with VGG19 without compromising signal information.
VGG19 was selected over other pre-trained CNN architectures primarily due to its straightforward and well-established structure, which facilitates the visualization and understanding learned features, an important consideration in condition monitoring applications. While deeper networks like ResNet offer improved accuracy via residual connections, their increased complexity can hinder transparent analysis of feature extraction. The simplicity of VGG19 thus enables effective feature representation alongside easier interpretability.
In this study, we use the pre-trained VGG19 model to extract features [37]. However, several machine-learning techniques are applied for classification. This deep feature extraction process generates a huge number of features, some of which are impactful, while others are not. This issue results in increased computational time due to an oversized feature vector, along with misclassification problems caused by irrelevant features and high dimensionality. Therefore, there is a need for feature selection methods to reduce the number of features by choosing only the most important ones as mentioned in Figure 4b. The feature selection in our case is performed using combined feature selection methods.

5. Feature Selection

In the bearing fault diagnosis, the extraction and interpretation of meaningful features from complex models is of paramount importance [38]. The proposed work leverages deep features extracted via VGG19, a pre-trained CNN, to enhance the diagnostic capability of ML algorithms. While deep learning models such as the VGG19 have demonstrated superior performance in feature extraction of vibration signal [39,40], they inherently function as ‘’black box’’ models [7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41]. This characteristic impedes interpretability, which is crucial in critical applications such as fault detection in rotating machinery. Consequently, understanding the relevance and impact of deep features becomes essential, not only to improve ML model performance but also to foster trust, transparency, and insight into model behavior.
To address this interpretability challenge and simultaneously enhance model efficiency, a hybrid feature selection methodology is proposed, combining Analysis of Variance (ANOVA) F-statistics with Permutation Feature Importance (PFI). This dual-stage approach capitalizes on the complementary strengths of both techniques to reduce feature dimensionality, eliminate irrelevant or redundant features, and provide insights into the relationships between features and model predictions.

5.1. ANOVA-Based Pre-Filtering

The first stage of the proposed method utilizes the ANOVA F-statistic as a preliminary filter to assess the statistical significance of each extracted feature with respect to the target fault classes. ANOVA, a well-established parametric statistical test, evaluates the variance between group means relative to the variance within the groups, thereby identifying features that contribute significantly to class separation. Mathematically, the F-statistic is defined as follows [42]:
F = V a r i a n c e   b e t w e e n   g r o u p V a r i a n c e   w i t h i n   g r o u p = i = 1 k n i x ¯ i x ¯ 2 / ( K 1 ) i = 1 k j = 1 n i n i x i j x ¯ i 2 / ( N k )
where x ¯ i is the mean of group i ,   x ¯ is the overall mean, x i j denotes the j-th observation in group i , n i is the number of samples in group i , k is the number of groups, and N is the total number of samples. Features with low F-values, indicating minimal discriminatory power, are discarded, thereby reducing dimensionality and improving computational efficiency.
The rationale for selecting ANOVA lies in its simplicity, scalability, and effectiveness in identifying linearly separable features [43]. Importantly, it helps to mitigate the adverse effect of multicollinearity, by removing highly correlated or statistically insignificant features in advance [44].

5.2. Permutation Feature Importance: Model-Agnostic Interpretability

Following the ANOVA filtering, the retained subset of features undergoes evaluation using Permutation Feature Importance. PFI is a model-agnostic technique that quantifies the importance of each feature by measuring the change in the model’s prediction error upon randomly shuffling its values [45]. The greater the degradation in performance, the more important the feature is deemed to be.
The key advantages of PFI include the following [45,46,47]:
  • Applicability to any machine learning model, including tree-based, kernel-based, and neural models.
  • Effectiveness in detecting nonlinear relationships and complex feature interactions.
  • Enhanced interpretability by directly linking features to model output changes.

5.3. Implementation Steps

ANOVA is selected for its simplicity, scalability, and effectiveness in identifying linearly separable features, while also addressing multicollinearity, an inherent limitation of Permutation Feature Importance (PFI). As PFI is sensitive to highly correlated features, which may distort the interpretation of feature relevance, the preliminary application of ANOVA effectively reduces redundancy and filters out statistically insignificant features. This integration enhances the stability and reliability of the importance rankings generated by PFI, ultimately supporting a more interpretable and robust feature selection process.
The combined feature selection approach involves the following steps as illustrated in workflow in Figure 5:
  • Feature extraction: Extract deep features from the penultimate layer of the VGG19 model for each input sample.
  • Preform ANOVA F-test filtering:
    • Compute the F-statistic for each feature across different fault classes.
    • Select features exceeding a predefined significance threshold.
  • Preform Permutation Feature Importance:
    • Apply the selected features to train a machine learning model.
    • Iteratively permute each feature and measure the impact on model accuracy.
    • Rank features based on the observed degradation in performance.
  • Final selected feature subset: Retain top-ranked features for subsequent model training and evaluation.

5.4. Benefits of the Combined Approach

The integration of ANOVA and PFI offers a robust framework for both dimensionality reduction and interpretability enhancement. From a practical standpoint, this combined methodology contributes to:
Improved interpretability: By ranking and analyzing the contribution of individual deep features, it provides insights into the inner workings of the VGG19-based feature extractor, helping demystify its ‘‘black box’’ nature.
Enhanced model performance: By eliminating irrelevant and redundant features, the resulting ML models demonstrate increased accuracy and reduced generalization error.
Reduced execution time: The pruning of high-dimensional feature sets streamlines training and inference, making the approach suitable for real-time or embedded diagnostic applications.
Model behavior understanding: Facilitates a deeper understanding of how specific features influence the classifier’s decision boundaries, thereby aligning ML predictions with domain knowledge in fault diagnostics.
In summary, the proposed dual-stage feature selection framework serves as a critical enabler for both the performance and interpretability of machine learning models in bearing fault diagnosis. By bridging the gap between deep feature representations and transparent decision-making, this approach enhances the reliability and applicability of intelligent the bearing diagnostic systems in industrial settings.

6. Proposed Methodology

The workflow of the proposed methodology, given in Figure 6, intends to improve the performance of different machine learning algorithms in classifying rolling bearing faults. The conducted experiment using the (CWRU) dataset analyzed the effectiveness of the bearing fault diagnosis system. The proposed strategy is based on integration between deeply extracted features obtained from deep CNN (VGG19) transfer learning, dual-stage feature selection approach using the ANOVA algorithm and PFI and SHAP analysis for supervised classification model explanation. The proposed system works in five main steps as follows:
  • Converting the vibration signal to a 64 by 64 grayscale image.
  • Extracting the deep features using the VGG19 pre-trained model.
  • Selecting the most important features using dual stage feature selection.
  • Classifying by different machine learning classifiers.
  • Explain models using SHAP analysis.
Within the first step, the vibration signal is converted to a 64 by 64 grayscale image. At first, the collected vibration signals are subjected to a data preprocessing process to prepare them for conversion. The process consists of removing errors from the raw data, eliminating duplicate data, and transferring the data type. The second step is about feature extraction. VGG19 extracts all deep features, generating 4096 features from the output of the first fully connected layer (fc1) of the pre-trained VGG19 model. The next move is more sensitive, the third step in our proposed strategy. It concerns feature selection. In this step, features are selected, employing the dual stage approach, to use them in a fourth move, involving classification utilizing various machine learning classifiers. The last step in our methodology workflow is dedicated to explaining the outcomes of the fourth step.

7. Results and Discussion

The vibration data were captured by an accelerometer fixed to the bearing housing at the drive end of the induction motor and collected under various conditions, including health operation, inner race fault, outer race fault, and ball fault, at different motor loads of 0, 1, 2, and 3 hp and for fault diameters of 0.007, 0.014 and 0.021 inches. The data are recorded at a sampling rate of 12,000 samples per second and saved in MATLAB format files. Samples of the vibration signals under the different operating conditions are shown in Figure 7.
Table 3 describes the ten different faulted bearing classes used within this experimental dataset. In this experiment, we construct two datasets, each one containing different classes of rolling bearing faults. Dataset A has four classes corresponding to the first four labels of Table 3, and dataset B has ten classes covering all bearing labels in Table 3. The final construction of datasets is presented in Table 4. Figure 8 presents samples of reconstructed images representing the used classes. The datasets are balanced, and each category has the same number of images. Figure 9 represents a violin plot of both dataset A and B, it shows a detailed view of the dataset’s distribution.
Following the extraction of deep features using the pre-trained VGG19, a two-stage feature selection strategy is employed to identify the most relevant and informative features for each classifier from deep extracted features. In the first stage, ANOVA is utilized to statistically rank all extracted features based on their ability to discriminate between fault classes. The top-ranked features are then selected to reduce dimensionality while retaining the most diagnostically significant information.
In the second stage, PFI is applied independently for each supervised learning classifier. This step quantifies the contribution of each selected feature to the model’s predictive performance by measuring the decrease in accuracy when a feature’s values are randomly shuffled. This classifier-specific refinement ensures optimal feature relevance tailored to each model’s structure and learning dynamics.
Figure 10 illustrates examples of the most influential features identified for each classification model on Dataset B, based on the corresponding decrease in accuracy. This dual-stage selection process not only reduces computational complexity and enhances model generalization but also significantly contributes to provide global interpretability. By systematically linking the CNN-derived features to their impact on classification outcomes, the proposed method transforms abstract deep features into transparent, tractable inputs for supervised models. This approach mitigates the inherent opacity of CNN models by enhancing the traceability of their outputs, thereby promoting greater explainability, reliability, and accountability in bearing fault diagnosis framework.
This study employs the tenfold cross-validation (CV) method [48] to evaluate the effectiveness of the proposed rolling bearing fault recognition strategy and to obtain an unbiased performance estimate of the various machine learning algorithms adopted in this work. In each fold, 90% of the data is used for training and 10% for testing, ensuring that every sample appears in the test set exactly once across all iterations. Each partition also maintains a balanced distribution of class samples to preserve fault type representativeness throughout the validation process. The overall performance is then computed as the average accuracy across the ten folds.
To rigorously prevent test–training data leakage, the data are first partitioned into folds, and for each iteration, all preprocessing, feature engineering, and model training are performed solely on the training subset of that fold. The test data remain completely unseen and isolated until the final evaluation stage in each fold. This strict separation ensures that the model does not gain access directly or indirectly to any information from the test set during training or feature selection, thereby eliminating leakage and preserving the validity of the results. Furthermore, the following metrics measurements as described in [49,50] are used to evaluate the performance of each classification algorithm.
  • Classification accuracy (CA) is computed by dividing the number of correct assessments for a particular class by the total number of assessments for that class. Accuracy represents the efficiency of classification.
  • Precision describes which rolling bearing fault is recognized as belonging to a particular class actually belongs to this class.
  • Recall determines which rolling bearing fault is recognized correctly.
  • F1-score functions as a single assessment value for classification models by computing the harmonic mean between precision and recall.
  • AUC (area under the receiver-operating curve) is an interpolation between the true positive and false positive rates. It is used to analyze the performance of classification models. An AUC greater than 90% indicates an excellent classification result.
  • Execution time in machine learning execution time is cumulated time in seconds used in training and testing the model.
Our experiments were carried out on a computer running Microsoft Windows 10 operating system and an Intel® Core (TM) i5-11400H CPU with 16 GB memories, using Python 3.9.12. The proposed system was evaluated on two datasets: one with 4 classes and another with 10 classes, covering various bearing fault types and severity levels. To achieve the optimal results for each classifier, we designed models according to the kernel hyperparameters that illustrated in Table 5.
A detailed comparison among various classifiers including: FR, XGBoost, LR, KNN, ANN, SVM on all the performance parameters is given in Table 6. The comprehensive analysis indicates a real improvement of all the machine learning models in classifying rolling bearing faults, adopting the workflows proposed in this study. Particularly, in terms of diagnostic accuracy and computational efficiency, the Support Vector Machine (SVM) algorithm exhibited superior performance achieving classification accuracies of 100% and 99.66%, with notably reduced computational times of 0.1 s and 6.8 s on Datasets A and B, respectively.
A comparative analysis was conducted to assess the effectiveness of the proposed system with and without the dual-stage feature selection approach. As shown in Table 7, feature selection consistently improved both classification accuracy and computational efficiency across Datasets A and B. For instance, the SVM model achieved an average accuracy gain of 0.41% on Dataset B, while its computational time decreased significantly from 34 s to 6.8 s, demonstrating the ability of the dual-stage method to eliminate redundant features and enhance classifier performance. The confusion matrices (Figure 11 and Figure 12) provide a detailed view of class-wise predictions, while the ROC curves (Figure 13 and Figure 14) confirm the models’ strong discriminatory capability through high true positive rates and low false positive rates.
Importantly, these improvements extend beyond algorithmic performance to directly enhance defect identification in rolling bearings. The higher accuracies correspond to more reliable differentiation between normal conditions and specific fault types (inner race, outer race, and rolling element defects), while reduced computational time supports faster and more practical deployment in industrial settings. Thus, the proposed hybrid diagnostic framework not only surpasses traditional classifiers in numerical performance but also strengthens the quality of fault detection and enables timely, reliable diagnosis for predictive maintenance applications.
Table 8 summarizes a comparative analysis between the proposed system and a selection of recent methodologies applied to the CWRU dataset. The results demonstrate that the proposed approach significantly outperforms several existing techniques, exhibiting superior performance in both classification accuracy and computational efficiency.
Overall, these findings substantiate the effectiveness of the proposed strategy in achieving robust diagnostic performance while maintaining a lightweight and computationally efficient design.
Although the proposed framework demonstrates strong performance under controlled experimental conditions, we acknowledge that the framework relies on labeled data for supervised learning, which may limit its generalizability to unlabeled or real-world scenarios. The framework has been primarily validated on relatively well-defined fault types within a specific dataset, which may not fully capture the complexity and variability of faults occurring in diverse, real-world machinery and operational conditions. Its effectiveness may be challenged in noisy industrial environments where vibration signals are often contaminated with significant background noise and interference. Such noise can degrade the quality of the extracted features and complicate accurate fault classification. Complex mechanical systems often exhibit overlapping fault signatures and evolving degradation patterns that could reduce the robustness and generalizability of the model.

8. Model’s Explanation via SHAP Analysis

Building upon the global interpretability insights provided by PFI, this section introduces SHapley Additive exPlanations (SHAP) to achieve local interpretability explaining individual predictions and elucidating how specific deep features influence model outputs [51]. While PFI quantifies the overall importance of each feature by measuring its impact on model performance when permuted, SHAP offers a more granular perspective by attributing a precise contribution of each feature to a single prediction [52]. This localized interpretability is essential in high-stakes domains such as bearing fault diagnosis, where model transparency, trust, and robustness are critical for reliable decision-making [53].
SHAP is grounded in cooperative game theory and computes feature attributions based on Shapley values, the average marginal contribution of each feature across all possible feature combinations [54]. In addition to PFI which provides global rankings, SHAP explains why a model outputs a specific prediction for a given input, identifying whether each feature increases or decreases the likelihood of a particular class [51,55]. Thus, SHAP not only complements PFI but enhances interpretability by revealing the direction and magnitude of each feature’s influence at the instance level.
Figure 15 illustrates the SHAP values obtained for the different fault classes with dataset B, using the Support Vector Machine (SVM) model as a representative example. The plots display the initial features selected through the proposed dual-stage feature selection approach arranged along the y-axis according to their impact on the model’s output. These features are ranked by their mean SHAP values, with the x-axis representing the average contribution to the prediction. A positive SHAP value for a feature indicates that its presence increases the likelihood of the corresponding fault class, while a negative SHAP value implies a shift away from it.
This visualization facilitates a clear understanding of the contribution of individual features to the model’s decision-making process. For example, a low value of feature n283 positively influences the prediction of fault classes IR_007, IR_014, and IR_021, while negatively impacting the classification of OR_007, OR_014, and OR_021. Conversely, a high value of feature n416 enhances the likelihood of predicting fault classes OR_007, OR_014, and OR_021, while exerting a negative effect on the prediction of IR_007, IR_014, and IR_021.
In this way, it becomes evident how individual features differentially affect class-specific outcomes, offering a nuanced understanding of the model’s internal logic.
Such insight reveals that certain deep features such as n283 and n416 consistently correspond to characteristic vibration behaviors of specific fault types. Features strongly linked to inner race faults capture transient, high-frequency components associated with the impulsive impacts caused by inner race defects. Conversely, features driving outer race fault predictions exhibit periodic patterns reflecting the stationary nature of outer race damage.
Thus, SHAP serves as a bridge between abstract deep features and physically interpretable fault mechanisms, offering insight into how the model internally differentiates between fault types based on the signal characteristics they produce.
This analysis can be systematically extended to all machine learning models employed in this study, allowing consistent evaluation of feature contributions across classifiers. The integration of SHAP within the proposed hybrid diagnostic framework thus provides a critical layer of interpretability, facilitating transparent, trustworthy, and actionable insights in the context of real time intelligent bearing fault diagnosis.

9. Conclusions

This study presented a hybrid diagnostic framework that effectively bridges the gap between high-accuracy fault detection and model interpretability in vibration-based condition monitoring of rotary machinery. By integrating deep feature extraction through VGG19-based transfer learning with a dual-stage feature selection mechanism, comprising ANOVA and PFI, the framework enables both superior diagnostic accuracy and enhanced model transparency. This combination facilitates the identification of the most discriminative CNN-derived features, providing a systematic understanding of the representations driving fault classification.
A central innovation of the proposed framework lies in its dual-stage feature selection strategy, which leverages both statistical relevance (ANOVA) and model-based importance (PFI) to systematically refine deep features for classification. This approach, to the best of our knowledge, has not been extensively explored in vibration-based fault diagnosis. Combined with SHAP-based interpretability analysis, the framework not only achieves high diagnostic accuracy but also ensures transparency in the classification process, an essential requirement for practical industrial applications.
Furthermore, SHAP analysis is employed to rigorously interpret the predictions of the supervised models, offering quantitative insight into individual feature contributions. This deepens understanding of the decision-making process and substantially improves the transparency, trustworthiness, and accountability of the overall system.
Experimental validation using the CWRU bearing dataset confirms the effectiveness of the proposed approach, achieving up to 100% classification accuracy while maintaining computational efficiency suitable for real-time applications. Overall, the framework advances the field by transforming complex deep learning outputs into interpretable, reliable insights, thereby supporting trustworthy, transparent decision-making in safety-critical industrial environments. Consequently, further investigation is necessary to adapt and optimize the methodology for practical applications involving multi-source noise and more intricate fault scenarios typical of industrial environments, where signal contamination and overlapping fault signatures may pose additional challenges. In addition, future work will aim to extend the proposed framework by incorporating multi-sensor data fusion, integrating heterogeneous sources such as acoustic emissions and thermal measurements to enhance the robustness and reliability of fault diagnosis. Moreover, the development and evaluation of computationally efficient model architectures and feature selection techniques specifically optimized for deployment on edge computing platforms will be prioritized, facilitating real-time condition monitoring within resource-constrained environments. Furthermore, efforts will be directed towards validating the generalizability of the framework across a wider variety of machinery types and fault modalities. In addition, the exploration of advanced model interpretability methods beyond SHAP, including but not limited to counterfactual explanations and causal inference techniques, is anticipated to yield deeper insights into model decision-making processes and further advance transparency in industrial prognostic systems.

Author Contributions

Conceptualization, M.E.S. and M.M.T.; methodology, M.E.S.; software, M.E.S.; validation, M.E.S. and A.B.; formal analysis, M.M.T.; investigation, M.E.S.; resources, A.B.; data curation, A.B.; writing—original draft preparation, M.E.S.; writing—review and editing, M.E.S.; visualization, A.B.; supervision, M.M.T.; project administration, M.M.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This study uses the publicly available CRWU dataset https://csegroups.case.edu/bearingdatacenter, accessed on 15 February 2024.

Acknowledgments

The authors gratefully acknowledge the support of the University Mohamed Khider of Biskra and the Algerian Ministry of Higher Education and Scientific Research.

Conflicts of Interest

The authors declare that they have no conflicts of interest, financial or otherwise, that could have influenced the research, authorship, or publication of this article.

Nomenclature

ANOVAAnalyze of Variance
PFIPermutation Feature Importance
SHAPSHapley Additive exPlanations
CVCross Validation

References

  1. Yang, Y.; Zhang, Y.; Tan, X. Review on vibration-based structural health monitoring techniques and technical codes. Symmetry 2021, 13, 1998. [Google Scholar] [CrossRef]
  2. Xu, F.; Ding, N.; Li, N.; Liu, L.; Hou, N.; Xu, N.; Chen, X. A review of bearing failure modes, mechanisms and causes. Eng. Fail. Anal. 2023, 152, 107518. [Google Scholar] [CrossRef]
  3. Zonta, T.; Da Costa, C.A.; da Rosa Righi, R.; de Lima, M.J.; Da Trindade, E.S.; Li, G.P. Predictive maintenance in the industry 4.0: A systematic literature review. Comput. Ind. Eng. 2020, 150, 106889. [Google Scholar] [CrossRef]
  4. Chegini, S.N.; Bagheri, A.; Najafi, F. Application of a new EWT-based denoising technique in bearing fault diagnosis. Measurement 2019, 144, 275–297. [Google Scholar] [CrossRef]
  5. Caesarendra, W.; Tjahjowidodo, T. A review of feature extraction methods in vibration-based condition monitoring and its application for degradation trend estimation of low-speed slew bearing. Machines 2017, 5, 21. [Google Scholar] [CrossRef]
  6. Buchaiah, S.; Shakya, P. Bearing fault diagnosis and prognosis using data fusion-based feature extraction and feature selection. Measurement 2022, 188, 110506. [Google Scholar] [CrossRef]
  7. Yang, H.; Li, X.; Zhang, W. Interpretability of deep convolutional neural networks on rolling bearing fault diagnosis. Meas. Sci. Technol. 2022, 33, 055005. [Google Scholar] [CrossRef]
  8. Zhang, X.; Rane, K.P.; Kakaravada, I.; Shabaz, M. Research on vibration monitoring and fault diagnosis of rotating machinery based on internet of things technology. Nonlinear Eng. 2021, 10, 245–254. [Google Scholar] [CrossRef]
  9. Yadav, O.P.; Pahuja, G.L. Bearing health assessment using time domain analysis of vibration signal. Int. J. Image Graph. Signal Process. 2020, 10, 27. [Google Scholar] [CrossRef]
  10. Luo, X.; Wang, H.; Han, T.; Zhang, Y. FFT-trans: Enhancing robustness in mechanical fault diagnosis with Fourier transform-based transformer under noisy conditions. IEEE Trans. Instrum. Meas. 2024, 73, 2515112. [Google Scholar] [CrossRef]
  11. Alonso-González, M.; Díaz, V.G.; Pérez, B.L.; G-Bustelo, B.C.P.; Anzola, J.P. Bearing fault diagnosis with envelope analysis and machine learning approaches using CWRU dataset. IEEE Access 2023, 11, 57796–57805. [Google Scholar] [CrossRef]
  12. Liu, D.; Cheng, W.; Wen, W. Rolling bearing fault diagnosis via STFT and improved instantaneous frequency estimation method. Procedia Manuf. 2020, 49, 166–172. [Google Scholar] [CrossRef]
  13. Liu, Z.; Peng, D.; Zuo, M.J.; Xia, J.; Qin, Y. Improved Hilbert–Huang transform with soft sifting stopping criterion and its application to fault diagnosis of wheelset bearings. ISA Trans. 2022, 125, 426–444. [Google Scholar] [CrossRef]
  14. Qin, C.; Wang, D.; Xu, Z.; Tang, G. Improved empirical wavelet transform for compound weak bearing fault diagnosis with acoustic signals. Appl. Sci. 2020, 10, 682. [Google Scholar] [CrossRef]
  15. Ni, Q.; Ji, J.C.; Feng, K.; Halkon, B. A fault information-guided variational mode decomposition (FIVMD) method for rolling element bearings diagnosis. Mech. Syst. Signal Process. 2022, 164, 108216. [Google Scholar] [CrossRef]
  16. Sánchez, R.V.; Macancela, J.C.; Ortega, L.R.; Cabrera, D.; García Márquez, F.P.; Cerrada, M. Evaluation of hand-crafted feature extraction for fault diagnosis in rotating machinery: A survey. Sensors 2024, 24, 5400. [Google Scholar] [CrossRef]
  17. Gougam, F.; Rahmoune, C.; Benazzouz, D.; Afia, A.; Zair, M. Bearing faults classification under various operation modes using time domain features, singular value decomposition, and fuzzy logic system. Adv. Mech. Eng. 2020, 12, 1687814020967874. [Google Scholar] [CrossRef]
  18. Soomro, A.A.; Muhammad, M.B.; Mokhtar, A.A.; Saad, M.H.M.; Lashari, N.; Hussain, M.; Palli, A.S. Insights into modern machine learning approaches for bearing fault classification: A systematic literature review. Results Eng. 2024, 102, 700. [Google Scholar]
  19. Mohamad, T.H.; Nazari, F.; Nataraj, C. A review of phase space topology methods for vibration-based fault diagnostics in nonlinear systems. J. Vib. Eng. Technol. 2020, 8, 393–401. [Google Scholar] [CrossRef]
  20. Chen, X.; Yang, R.; Xue, Y.; Huang, M.; Ferrero, R.; Wang, Z. Deep transfer learning for bearing fault diagnosis: A systematic review since 2016. IEEE Trans. Instrum. Meas. 2023, 72, 3508221. [Google Scholar] [CrossRef]
  21. Chen, G.; Yuan, J.; Zhang, Y.; Zhu, H.; Huang, R.; Wang, F.; Li, W. Enhancing reliability through interpretability: A comprehensive survey of interpretable intelligent fault diagnosis in rotating machinery. IEEE Access 2024, 12, 103348–103379. [Google Scholar] [CrossRef]
  22. Ahmed, H.; Nandi, A.K. Compressive sampling and feature ranking framework for bearing fault classification with vibration signals. IEEE Access 2018, 6, 44731–44746. [Google Scholar] [CrossRef]
  23. Hui, K.H.; Ooi, C.S.; Lim, M.H.; Leong, M.S.; Al-Obaidi, S.M. An improved wrapper-based feature selection method for machinery fault diagnosis. PLoS ONE 2017, 12, e0189143. [Google Scholar] [CrossRef] [PubMed]
  24. Imane, M.; Rahmoune, C.; Benazzouz, D. Rolling bearing fault feature selection based on standard deviation and random forest classifier using vibration signals. Adv. Mech. Eng. 2023, 15, 16878132231168503. [Google Scholar] [CrossRef]
  25. Hopf, K.; Reifenrath, S. Filter methods for feature selection in supervised machine learning applications—Review and benchmark. arXiv 2021, arXiv:2111.12140. [Google Scholar]
  26. Xu, G.; Liu, M.; Jiang, Z.; Söffker, D.; Shen, W. Bearing fault diagnosis method based on deep convolutional neural network and random forest ensemble learning. Sensors 2019, 19, 1088. [Google Scholar] [CrossRef]
  27. Cui, B.; Weng, Y.; Zhang, N. A feature extraction and machine learning framework for bearing fault diagnosis. Renew. Energy 2022, 191, 987–997. [Google Scholar] [CrossRef]
  28. Yuan, L.; Lian, D.; Kang, X.; Chen, Y.; Zhai, K. Rolling bearing fault diagnosis based on convolutional neural network and support vector machine. IEEE Access 2020, 8, 137395–137406. [Google Scholar] [CrossRef]
  29. Liang, X.; Yao, J.; Zhang, W.; Wang, Y. A novel fault diagnosis of a rolling bearing method based on variational mode decomposition and an artificial neural network. Appl. Sci. 2023, 13, 3413. [Google Scholar] [CrossRef]
  30. Yu, X.; Liang, Z.; Wang, Y.; Yin, H.; Liu, X.; Yu, W.; Huang, Y.A. wavelet packet transform-based deep feature transfer learning method for bearing fault diagnosis under different conditions. Measurement 2022, 201, 111597. [Google Scholar] [CrossRef]
  31. Fong, J.; de Wachter, B.; Sangiorgio, I.; Ntaras, N.; Zarkadoula, M.; Almeida, A.T. Policy recommendations to accelerate the replacement of inef-ficient electric motors in the EU. Energy Effic. 2025, 18, 53. [Google Scholar] [CrossRef]
  32. Bazzi, A.M.; Krein, P.T. Review of methods for real-time loss minimization in induction machines. IEEE Trans. Ind. Appl. 2010, 46, 2319–2328. [Google Scholar] [CrossRef]
  33. Bagri, I.; Tahiry, K.; Hraiba, A.; Touil, A.; Mousrij, A. Vibration signal analysis for intelligent rotating machinery diagnosis and prognosis: A comprehensive systematic literature review. Vibration 2024, 7, 1013–1062. [Google Scholar] [CrossRef]
  34. Case Western Reserve University Bearing Data Center. Available online: https://csegroups.case.edu/bearingdatacenter (accessed on 15 February 2024).
  35. Boudiaf, A.; Moussaoui, A.; Dahane, A.; Atoui, I. A comparative study of various methods of bearing faults diagnosis using the Case Western Reserve University data. J. Fail. Anal. Prev. 2016, 16, 271–284. [Google Scholar] [CrossRef]
  36. Hoang, D.T.; Kang, H.J. Rolling element bearing fault diagnosis using convolutional neural network and vibration image. Cogn. Syst. Res. 2019, 53, 42–50. [Google Scholar] [CrossRef]
  37. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  38. Zhang, S.; Zhang, S.; Wang, B.; Habetler, T.G. Deep learning algorithms for bearing fault diagnostics—A comprehensive review. IEEE Access 2020, 8, 29857–29881. [Google Scholar] [CrossRef]
  39. Neupane, D.; Seok, J. Bearing fault detection and diagnosis using Case Western Reserve University dataset with deep learning approaches: A review. IEEE Access 2020, 8, 93155–93178. [Google Scholar] [CrossRef]
  40. Wen, L.; Li, X.; Li, X.; Gao, L. A new transfer learning based on VGG-19 network for fault diagnosis. In Proceedings of the 2019 IEEE 23rd International Conference on Computer Supported Cooperative Work in Design (CSCWD), Porto, Portugal, 6–8 May 2019; pp. 205–209. [Google Scholar]
  41. Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 2018, 51, 1–42. [Google Scholar] [CrossRef]
  42. Kim, T.K. Understanding one-way ANOVA using conceptual figures. Korean J. Anesthesiol. 2017, 70, 22. [Google Scholar] [CrossRef]
  43. Kamalov, F.; Sulieman, H.; Alzaatreh, A.; Emarly, M.; Chamlal, H.; Safaraliev, M. Mathematical Methods in Feature Selection: A Review. Mathematics 2025, 13, 996. [Google Scholar] [CrossRef]
  44. Chan, J.Y.L.; Leow, S.M.H.; Bea, K.T.; Cheng, W.K.; Phoong, S.W.; Hong, Z.W.; Chen, Y.L. Mitigating the multicollinearity problem and its machine learning approach: A review. Mathematics 2022, 10, 1283. [Google Scholar] [CrossRef]
  45. Altmann, A.; Toloşi, L.; Sander, O.; Lengauer, T. Permutation importance: A corrected feature importance measure. Bioinformatics 2010, 26, 1340–1347. [Google Scholar] [CrossRef]
  46. Molnar, C.; König, G.; Bischl, B.; Casalicchio, G. Model-agnostic feature importance and effects with dependent features: A conditional subgroup approach. Data Min. Knowl. Discov. 2024, 38, 2903–2941. [Google Scholar] [CrossRef]
  47. Orlenko, A.; Moore, J.H. A comparison of methods for interpreting random forest models of genetic association in the presence of non-additive interactions. BioData Min. 2021, 14, 1. [Google Scholar] [CrossRef] [PubMed]
  48. Nti, I.K.; Nyarko-Boateng, O.; Aning, J. Performance of machine learning algorithms with different K values in K-fold cross-validation. Int. J. Inf. Technol. Comput. Sci. 2021, 13, 61–71. [Google Scholar]
  49. Tharwat, A. Classification assessment methods. Appl. Comput. Inform. 2021, 17, 168–192. [Google Scholar] [CrossRef]
  50. Luque, A.; Carrasco, A.; Martín, A.; de Las Heras, A. The impact of class imbalance in classification performance metrics based on the binary confusion matrix. Pattern Recognit. 2019, 91, 216–231. [Google Scholar] [CrossRef]
  51. Lundberg, S.M.; Lee, S.I. A unified approach to interpreting model predictions. Adv. Neural Inf. Process. Syst. 2017, 30, 1–10. [Google Scholar]
  52. Wang, H.; Liang, Q.; Hancock, J.T.; Khoshgoftaar, T.M. Feature selection strategies: A comparative analysis of SHAP-value and importance-based methods. J. Big Data 2024, 11, 44. [Google Scholar] [CrossRef]
  53. Rudin, C.; Chen, C.; Chen, Z.; Huang, H.; Semenova, L.; Zhong, C. Interpretable machine learning: Fundamental principles and 10 grand challenges. Stat. Surv. 2022, 16, 1–85. [Google Scholar] [CrossRef]
  54. Li, M.; Sun, H.; Huang, Y.; Chen, H. Shapley value: From cooperative game to explainable artificial intelligence. Auton. Intell. Syst. 2024, 4, 2. [Google Scholar] [CrossRef]
  55. Molnar, C. Interpretable Machine Learning; Lulu.com: Morrisville, NC, USA, 2020. [Google Scholar]
Figure 1. The CRWU benchmark.
Figure 1. The CRWU benchmark.
Sensors 25 06386 g001
Figure 2. Bearing’s components.
Figure 2. Bearing’s components.
Sensors 25 06386 g002
Figure 3. Converting 1D vibration signal to 2D image methodology.
Figure 3. Converting 1D vibration signal to 2D image methodology.
Sensors 25 06386 g003
Figure 4. (a) VGG19 pre-trained model. (b) The proposed ensemble of VGG19 deep feature extraction, dual-stage feature selection and machine learning classification.
Figure 4. (a) VGG19 pre-trained model. (b) The proposed ensemble of VGG19 deep feature extraction, dual-stage feature selection and machine learning classification.
Sensors 25 06386 g004
Figure 5. Workflow of the adopted dual-stage feature selection method.
Figure 5. Workflow of the adopted dual-stage feature selection method.
Sensors 25 06386 g005
Figure 6. Workflow for the proposed strategy.
Figure 6. Workflow for the proposed strategy.
Sensors 25 06386 g006
Figure 7. Samples of raw vibration signals under different operating conditions form CRWU dataset.
Figure 7. Samples of raw vibration signals under different operating conditions form CRWU dataset.
Sensors 25 06386 g007
Figure 8. Samples of the converted greyscale image signals under different operating conditions.
Figure 8. Samples of the converted greyscale image signals under different operating conditions.
Sensors 25 06386 g008
Figure 9. Violin plot: (a)—distribution of Dataset A, (b)—distribution of Dataset B.
Figure 9. Violin plot: (a)—distribution of Dataset A, (b)—distribution of Dataset B.
Sensors 25 06386 g009
Figure 10. Top feature per model: (a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM.
Figure 10. Top feature per model: (a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM.
Sensors 25 06386 g010
Figure 11. The confusion matrix of proposed models ((a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM) with the dataset A.
Figure 11. The confusion matrix of proposed models ((a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM) with the dataset A.
Sensors 25 06386 g011
Figure 12. The confusion matrix of proposed models ((a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM) with the dataset B.
Figure 12. The confusion matrix of proposed models ((a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM) with the dataset B.
Sensors 25 06386 g012
Figure 13. ROC curves of proposed models ((a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM) with the dataset A.
Figure 13. ROC curves of proposed models ((a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM) with the dataset A.
Sensors 25 06386 g013
Figure 14. ROC curves of proposed models ((a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM) with the dataset B.
Figure 14. ROC curves of proposed models ((a) RF, (b) XGBoost, (c) LR, (d) KNN, (e) ANN, (f) SVM) with the dataset B.
Sensors 25 06386 g014
Figure 15. SVM Shapley values for each fault class: ((a) IR_007, (b) IR_0014, (c) IR_021, (d) OR_007 (e) OR_014, (f) OR_021 (g) BA_007, (h) BA_0014, (i) BA_021, (j) Normal) in dataset B.
Figure 15. SVM Shapley values for each fault class: ((a) IR_007, (b) IR_0014, (c) IR_021, (d) OR_007 (e) OR_014, (f) OR_021 (g) BA_007, (h) BA_0014, (i) BA_021, (j) Normal) in dataset B.
Sensors 25 06386 g015aSensors 25 06386 g015b
Table 1. IEEE distribution of common faults in electric motors.
Table 1. IEEE distribution of common faults in electric motors.
Fault CategoryApproximate Contribution (%)
Bearings faults40–50%
Stator faults30–40%
Rotor faults~10%
Other faults<10%
Table 2. Bearing’s technical parameters.
Table 2. Bearing’s technical parameters.
Technical ParametersValue
Inner race diameter d25.001 (mm)
Outer race diameter D51.998 (mm)
Ball diameter BD7.940 (mm)
Pitch diameter CD39.04 (mm)
Ball number N9 (mm)
Overall width B15 (mm)
Contact angle β0 (rad)
Table 3. The fault classes used from the CWRU dataset.
Table 3. The fault classes used from the CWRU dataset.
Faut ClassFault Diameter (Inch)Fault Class
Healthy status/Normal
Inner race0.007IR_007
Outer race (centered at 6:00)0.007OR_007
Ball0.007BA_007
Inner race0.014IR_014
Outer race (centered at 6:00)0.014OR_014
Ball0.014BA_014
Inner race0.021IR_021
Outer race (centered at 6:00)0.021OR_021
Ball0.021BA_021
Table 4. The constructed dataset.
Table 4. The constructed dataset.
DatasetNB of ClassesLoad (HP)NB of Images
A40, 1, 2 and 3464
B10 1160
Table 5. Hyperparameters used for classification models.
Table 5. Hyperparameters used for classification models.
Classification ModelHyperparameters DefinitionValues
Random ForestNumber of trees100
(RF)Number of attributes at each split10–15
Extreme Gradient BoostingBoostergptree
(XGBoost)Learning rate0.01–0.3
Limit depth7
Number of trees100
Subsample0.5
Logistic RegressionLasso regressionLasso (L1)
(LR)Strength C1
K-Nearest NeighborsNumber of neighbors5–10
(KNN)FunctionEuclidean
Artificial Neural NetworksNumber of neurons in hidden layers100
(ANN)Activation functionReLu
OptimizerAdam
Learning rate0.001
Support Vector MachineKernel functionPolynomial
(SVM)Kernel degree3
Cost C1
Gamma0.1
Table 6. Comparative analysis of model performance metrics on Datasets A and B.
Table 6. Comparative analysis of model performance metrics on Datasets A and B.
Metric PerformanceDatasetRFXGBoostLRKNNANNSVM
Average accuracy 99.55%99.76%99.78%100%100%100%
Precision 99.60%99.80%99.80%100%100%100%
RecallA99.55%99.76%99.8%100%100%100%
F1-score 99.50%99.65%99.70%100%100%100%
AUC 99.65%99.80%99.85%100%100%100%
Average execution time (S) 1.70.60.50.012.40.1
Average accuracy 98.79%99.22%99.56%98.51%99.33%99.66%
PrecisionB98.50%99.20%99.60%98/50%90.40%99.70%
Recall 98.79%99.22%99.56%98.51%99.33%99.66%
F1-score 98.22%99.02%99.49%98.15%99.13%99.36%
AUC 99.11%99.45%99.55%99.65%99.81%100%
Average execution time (S) 7.75.79.13.1196.8
Table 7. Classification performances with and without dual-stage feature selection.
Table 7. Classification performances with and without dual-stage feature selection.
ModelDatasetWithout Feature Selection ApproachWith Feature Selection Approach
AccuracyExecution TimeAccuracyExecution Time
RF 99.00%4.199.55%1.7
XGBoost 99.01%2.899.76%0.6
LRA99.11%2.599.78%0.5
KNN 99.78%1.3100%0.01
ANN 99.70%44100%2.4
SVM 100%5.5100%0.1
RF 98.40%15.398.79%7.7
XGBoost 98.80%17.199.22%5.7
LR 99.12%23.299.56%9.1
KNNB98.22%13.298.51%3.1
ANN 99.10%68.199.33%19
SVM 99.25%3499.66%6.8
Table 8. Comparative results with existing recent works on CRWU dataset.
Table 8. Comparative results with existing recent works on CRWU dataset.
Reference/YearMethods UsedExecution Time
(S)
Average Accuracy
[26]/(2019)CWT-CNN-RF99.08114.57
[27]/(2022)Three stage learning-
SVMN/A86.11%
KNNN/A92.59%
ANNN/A95.37%
[29]/(2023)VMD-ANN2.499.3%
[11]/(2023)AE-Fine KNN12.26100%
Our modelsHybrid framework-
RF1.799.55%
XGBoost0.699.76%
LR0.599.78%
KNN0.01100%
ANN2.4100%
SVM0.1100%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Saouli, M.E.; Touba, M.M.; Boudiaf, A. Hybrid Diagnostic Framework for Interpretable Bearing Fault Classification Using CNN and Dual-Stage Feature Selection. Sensors 2025, 25, 6386. https://doi.org/10.3390/s25206386

AMA Style

Saouli ME, Touba MM, Boudiaf A. Hybrid Diagnostic Framework for Interpretable Bearing Fault Classification Using CNN and Dual-Stage Feature Selection. Sensors. 2025; 25(20):6386. https://doi.org/10.3390/s25206386

Chicago/Turabian Style

Saouli, Mohamed Elhachemi, Mostefa Mohamed Touba, and Adel Boudiaf. 2025. "Hybrid Diagnostic Framework for Interpretable Bearing Fault Classification Using CNN and Dual-Stage Feature Selection" Sensors 25, no. 20: 6386. https://doi.org/10.3390/s25206386

APA Style

Saouli, M. E., Touba, M. M., & Boudiaf, A. (2025). Hybrid Diagnostic Framework for Interpretable Bearing Fault Classification Using CNN and Dual-Stage Feature Selection. Sensors, 25(20), 6386. https://doi.org/10.3390/s25206386

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop