Next Article in Journal
Determination of the Spring Constant for the Horizontal Movement of Rigid Rectangular Foundations Resting on Elastic Subsoil
Previous Article in Journal
A Partitioned Rigid Element–Interface Element–Finite Element Method (PRE-IE-FE) for the Slope Stability Analysis of Soil–Rock Binary Structures
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HFDF-EffNetV2: A Lightweight, Noise-Robust Model for Fault Diagnosis in Rolling Bearings

1
Key Laboratory of Advanced Manufacturing and Automation Technology (Guilin University of Technology), Education Department of Guangxi Zhuang Autonomous Region, Guilin 541006, China
2
Guangxi Engineering Research Center of Intelligent Rubber Equipment, Guilin University of Technology, Guilin 541006, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(9), 4902; https://doi.org/10.3390/app15094902
Submission received: 27 March 2025 / Revised: 24 April 2025 / Accepted: 25 April 2025 / Published: 28 April 2025
(This article belongs to the Section Mechanical Engineering)

Abstract

:
In rolling bearing intelligent fault diagnosis (FD), lightweight models are constrained by issues such as noise interference and the scarcity of fault data, making it challenging to achieve real-time, high-accuracy diagnosis on resource-limited devices. To address these challenges, this study proposes a lightweight model that combines the hierarchical fine-grained decision fusion (HFDF) strategy with an improved EfficientNetV2 architecture (HFDF-EffNetV2). The model optimizes depth and width multiplicity factors to enhance parameter utilization efficiency. It uses pyramidal convolution (PyConv) combined with Fused-MBConv (Fused-MBPyConv) to obtain multi-scale time-frequency information. Additionally, an enhanced MBConv, termed BSMB-Conv-MLCA, integrates subspace blueprint separable convolution (BSConv-S) with mixed local channel attention (MLCA) extract deep-grained fault features. The HFDF strategy outputs confidence in stages and updates weights to prevent the model from falling into local overfitting when handling confusable samples. Experimental results on Case 1 and Case 2 show that HFDF-EffNetV2 achieved 100% accuracy with diagnostic times of 18.67 millisecond (ms) and 17.56 ms, respectively, and 1.85 million (M) parameters. Under noisy conditions, average accuracies reached 98.19% and 85.68%, respectively. Additionally, the model performed well with small samples, yielding accuracies of 98.69% and 97.51%. These results highlight its superior robustness to noise and lightweight performance compared with other advanced models.

1. Introduction

Rolling bearings, which are crucial components of mechanical equipment, are prone to various types of damage due to their long running time and harsh working environment [1]. If damaged components cannot be detected and replaced promptly, outcomes include disruption of the normal operation of the mechanical equipment, lower production efficiency, and even serious consequences such as substantial economic losses and casualties [2]. Accurate, real-time detection and diagnosis of the operating condition of rolling bearings is therefore essential [3].
In fault diagnosis (FD), vibration signals are widely used as effective diagnostic signals due to their ability to quickly and accurately reflect equipment operating conditions [4,5,6]. Traditional FD methods mainly rely on signal-decomposition techniques such as feature mode decomposition [7], all time-scale decomposition [8] and Blaschke mode decomposition [9]. These methods identify fault types by analyzing the characteristics of the vibration signal, but they heavily rely on empirical knowledge and are difficult to interpret for people without relevant expertise [10]. Compared to signal-decomposition methods, deep learning techniques have been more widely researched and applied in industry because they do not require a priori knowledge and have superior feature-extraction capability [11,12]. For instance, Chang et al. [13] proposed an FD model combining a one-dimensional convolutional neural network (CNN) with a two-dimensional improved ResNet-50, which utilizes multi-modal information to enhance the robustness of the model. Han et al. [14] proposed the MT-ConvFormer FD model, which utilizes multi-task learning to extract fault information. Song et al. [15] proposed an FD model based on multimodal data fusion and the subtraction-average-based optimizer to address the challenge of poor model generalization with small samples by combining CNN and support vector machines. Although these approaches effectively improve FD results by introducing various mechanisms, they also require extensive parameter calculations, leading to increased consumption of storage and computational resources.
In recent years, the rapid development of embedded components and digital twin technology have provided new solutions for FD in rolling bearings in special operating scenarios [16]. Due to the advantages of edge computing, such as fast transmission and low latency, the lightweight FD methods for rolling bearings have attracted significant attention in recent research [17]. To enable real-time diagnosis, Fan et al. [18] proposed a lightweight, multi-scale, multi-attention feature fusion network that significantly enhances fault feature identification. Wang et al. [19] proposed an AUTO-CNN architectural design method, aiming to achieve a balance between accuracy and use of computational resources. Teta et al. [20] designed a lightweight CNN and fine-tuned it using the energy-valley optimizer. Wang et al. [21] proposed a lightweight FD model that integrates frequency-slice wavelet transform with an improved neural-transformer that emphasizes the global capture of important fine-grained information. Xie et al. [22] designed a lightweight pyramid attention residual network for fault diagnosis under changing operating conditions. However, while the above methods have the advantages of their light weights and the promising results they achieve through different technical approaches, they are over-reliant on global information and neglect to utilize the information extracted at different stages for a more comprehensive diagnosis. Although lightweight models offer advantages such as low computational cost and high diagnostic speed, they face several challenges in practical applications. For instance, issues such as limited fault data and noise interference in real-world scenarios significantly hinder the practical deployment of lightweight models [23,24,25,26].
Aiming to address the challenges of achieving light weight, high accuracy, strong robustness to noise, and real-time responsiveness in practical industrial applications, we propose HFDF-EffNetV2, a lightweight FD model, with the following main contributions:
  • The improved EfficientNetV2 was constructed based on multiplicity factors combined with the Mish activation function. Additionally, Fused-MBPyConv and BSMBConv-MLCA were designed to combine multi-scale shallow features with multi-dimensional deep features for efficient extraction of representative fault features from time-frequency feature maps (FMs).
  • The hierarchical fine-grained decision fusion (HFDF) strategy is proposed to perform weight updating in phases to alleviate the tendency of the lightweight model to fall into local overfitting when dealing with confusable fault types and small samples.
The proposed model was evaluated under noisy and small-sample conditions and compared with six other models. The results indicate that the proposed model outperforms these others in terms of robustness to noise and stability while remaining lightweight.
The remainder of this study is structured as follows. Section 2 presents the materials and methods; Section 3 describes the results, providing detailed architectural parameters; Section 4 discusses the robustness to noise and stability of HFDF-EffNetV2 in two cases; Section 5 concludes this study.

2. Materials and Methods

2.1. Continuous Wavelet Transform

In a rolling-bearing FD task, a fault signal from monitored equipment is easily lost in noise due to vibrations from the equipment and environmental factors; this noisiness increases the difficulty of later feature extraction [27]. The continuous wavelet transform (CWT) provides rich time-frequency information, which more effectively highlights the fault characteristics of the target components [28]. Considering the limited computational resources of edge devices, this study employed the CWT to convert the signals into time-frequency FMs with a resolution of 128 × 128 × 3 ; the FMs were then used as fault samples. The CWT is calculated as shown in Equation (1):
CWT ( a , b ) = 1 a x ( t ) ψ * ( t b a ) dt
where x ( t ) is the input signal, ψ * ( t ) is the complex conjugate of the scaled and shifted wavelet function ψ ( t ) , a is the scale factor, and b is the translation factor.

2.2. Pyramidal Convolution

Due to the limited receptive field of a single convolutional kernel, when the feature information in the time-frequency FM is located at the edges, small convolutional kernels are not sufficiently sensitive to edge information, resulting in the loss of edge details. Although large convolutional kernels have a larger receptive field, they require a substantial increase in parameters. As shown in Figure 1, pyramidal convolution (PyConv) comprises convolutional kernel of varying sizes, with the kernel sizes gradually increasing from bottom to top. In addition, PyConv introduces the grouped convolution. The input FMs are then processed through grouped convolution and subsequently fused into a complete output FM along the channel dimension. Compared to atrous spatial pyramid pooling, which utilizes the dilated convolution technique, PyConv offers superior feature capture capabilities and demonstrates better computational efficiency [29].

2.3. Subspace Blueprint Separable Convolution

Haase et al. [30] visualized the convolutional kernels of VGG-19, InceptionV2, and ResNet-50 and found that many FMs were linearly transformed by a special convolution kernel, referred to as the ‘Blueprint’, demonstrating intra-kernel correlation. Building on this concept, the authors propose the subspace blueprint separable convolution (BSConv-S), which effectively suppresses irrelevant features in the convolution operation through a subspace separation mechanism. In BSConv-S, the convolution process consists of three sub-processes, as follows:
  • The input tensor X is mapped to an M -dimensional subspace by M pointwise convolutional kernels of size 1 × 1   ( ω 1 B , ω 2 B , , ω m B , , ω M B ) , compressing the number of channels.
  • Another N pointwise convolutional kernels of size 1 × 1   ( ω 1 A , ω 2 A , , ω n A , , ω N A ) are applied to the result of step 1.
  • Finally, N depthwise separable convolutional kernels of size k × k   ( B 1 , B 2 , , B n , , B N ) are applied to the results of step 2.
The computational relationship between the input tensor X and the output tensor Y is shown in Equation (2) as follows:
Y n , : , : = X × ω M B × ω n A × B n , n = 1 , , N

2.4. Mixed Local Channel Attention Mechanism

Although attention mechanisms like squeeze-and-excitation (SE) and the convolutional block attention module have shown good performance in many tasks, the mixed local channel attention (MLCA) mechanism excels at capturing high-level feature representations. It offers better adaptability and enables more detailed channelwise feature learning across multiple levels, particularly when handling complex features [31]. It applies two steps of average pooling (AP) to transform the input feature vectors into a shape of ( 1 × C × ks × ks ) , with the first branch capturing global information and the second branch capturing local spatial information. After one-dimensional convolution, the two one-dimensional vectors are restored to their original resolution through unaverage pooling (UNAP) and the information is then fused to achieve hybrid attention. As shown in Figure 2, Conv1d denotes one-dimensional convolution, and the kernel size k of the one-dimensional convolution in MLCA is proportional to the channel C , indicating that only the relationship between each channel and its k neighboring channels are considered when capturing local cross-channel interaction information. The value of k is calculated as in Equation (3).
k = Φ ( C )   = lo g 2 ( C ) γ + d γ   odd
where C is the number of channels; k is convolutional kernel size; γ and d are hyperparameters, with the default value set to 2; the function odd ensures that k is odd by adding 1 if k is even.

2.5. EfficientNetV2

In the field of FD, accuracy is a crucial metric; therefore, while pursuing a lightweight design, it is essential to ensure high diagnostic accuracy, especially in the context of complex tasks [32,33,34].
EfficientNetV2 is a lightweight model introduced by Tan et al. [35] in 2021 that incorporates the Fused-MBConv and MBConv layers and is available in several versions, including B0~B3, S, and M. The model employs a progressive learning strategy that adaptively adjusts the regularization technique and the input image size, accelerating the training process while maintaining high accuracy. Through neural-architecture search, EfficientNetV2 optimizes three key parameters: input-image resolution, network depth, and channel width. Among the different versions, the M version, with depth and width scaling factors of 1.8 and 1.4, respectively, is suited for high-resolution scenarios, achieving a classification accuracy of 85.1% on the ImageNet dataset. In contrast, the S version, with depth and width scaling factors of 1.2 and 1.1, respectively, focuses more on balancing the number of parameters with accuracy, making it more suitable for lower-resolution applications.
EfficientNetV2-S is structured into seven stages; this is one MBConv stage fewer than the M version, resulting in a significant reduction in the number of parameters. As shown in the Table 1, stage zero consists of a standard convolutional layer. As illustrated in Figure 3a, stages one to three employ the Fused-MBConv layer, which utilizes 3 × 3 convolution for channel upscaling, followed by 1 × 1 convolution for downscaling. In contrast, stages four to six utilize the MBConv layer, as shown in Figure 3b. It begins by introducing features into a higher-dimensional space through 1 × 1 convolution, and this step is followed by feature extraction with a 3 × 3 depthwise convolution. Subsequently, the SE Attention mechanism is applied to capture significant channelwise features. Finally, a 1 × 1 convolution maps the processed features back to the input space. The final stage includes a 1 × 1 convolutional layer, an AP layer, and a fully connected (FC) layer. This streamlined architecture contributes to its reduced parameter count while maintaining efficient performance.

3. Results

The FD process of HFDF-EffNetV2 is shown in Figure 4. First, the vibration signals collected by the sensors during the machinery operation are transmitted to the digital device. Subsequently, data preprocessing using the CWT is applied to obtain time-frequency FM, which are then divided into training and test sets. The model-training process outputs the predicted probabilities in stages, computes the loss using the dynamic adaptive weighted cross-entropy (DAWCE) function during forward propagation, and updates the weights via backpropagation. Finally, the test set is fed into the trained model for diagnosis, and the diagnostic results are visualized and analyzed from multiple perspectives.

3.1. HFDF-EffNetV2

HFDF-EffNetV2 is constructed by optimizing the depth and width multiplicity factors of EfficientNetV2-S according to the resolution of the input time-frequency FM. In addition, the Mish [36] activation function is employed to enhance gradient flow, thereby improving model training stability. The Mish activation function is calculated as in Equation (4), below:
Mish ( δ ) = δ × tanh   ( ln   ( 1 + e δ ) )
where δ is the input value and tanh   is the hyperbolic tangent function that maps the result to the range ( 1 ,   1 ) .

3.1.1. Improving Fused-MBConv with PyConv

EfficientNetV2 introduces Fused-MBConv due to the limited hardware acceleration support for depthwise separable convolutions. Replacing all layers with Fused-MBConv increases the computational overhead of the network; therefore, it is used only in the shallow stages of the model.
To enlarge the receptive field in the shallow stages and capture multiscale feature information, this study constructed Fused-MBPyConv based on combining Fused-MBConv with PyConv, as shown in Figure 5. When the expansion rate is not 1, Fused-MBPyConv first expands the number of input channels using PyConv. A convolutional kernel of size 1 is then applied to map the expanded features to the desired number of output channels. Finally, the residual connection is introduced to enhance feature representation, prevent information loss, and mitigate the vanishing gradient problem. The specific implementation is given by Equation (5). The input FMs are assumed to be X R C × H × W , where C , H , and W denote the channel dimension, height, and width of the input FM, respectively.
F i = C o n v ( ks i × ks i , G i ) ( X i ) R C 4 × H × W F = [ F 0 ; F 1 ; F 2 ; F 3 ] R ( E × C ) × H × W Y = Conv ( F ) R C × H × W F o = X + Y R C × H × W
where ks i is the convolutional kernel size; G i is the number of groups; and E is the channel expansion factor.

3.1.2. Construction of BSMBConv-MLCA

The MBConv is an inverted linear bottleneck module that utilizes a depthwise separable convolution. In this study, we propose BSMBConv-MLCA, a novel architecture that synergistically integrates BSConv-S with the MLCA mechanism within the MBConv, as shown in Figure 6, to facilitate enhanced extraction of discriminative fault information while suppressing noise interference. BSConv-S effectively reduces the number of parameters by introducing the strategy of subspace separation and has stronger feature extraction. which further enhances the efficiency of parameter utilization. To mitigate information loss during channel compression while augmenting global and local feature interdependencies, we strategically replaced the SE attention module in MBConv with MLCA, deploying it across residual connections to excavate fault information through cross-scale attention propagation. The BSMBConv-MLCA enables the model to focus on both local and global features through multi-branch convolution and pooling operations, efficiently fusing fault information from different channels.

3.1.3. HFDF Strategy

The HFDF strategy is proposed in this study to alleviate the issue of low confidence and susceptibility to local overfitting due to insufficient learning of fault features in lightweight models when facing confusable fault types and small samples. It is based on the concept of multi-granularity image classification [37]. This strategy employs global average pooling (GAP), the FC layer, and the SoftMax activation function at different stages of the model; these are followed by a weighted summation to output the classification probability. The goal is to fully utilize fault information at different levels to enhance the model’s diagnostic confidence and, consequently, improve diagnostic accuracy. This is implemented as shown in Equations (6)–(9), with the probabilities of the two branch outputs denoted as p i 1 and p i 2 to distinguish between them.
p i 1 = e ( x i 1 ) j = 1 n e ( x i 1 ) p i 2 = e ( x i 2 ) j = 1 n e ( x i 2 )
ω 1 = max ( p i 1 ) ω 2 = max ( p i 2 )
α 1 = ω 1 ω 1 + ω 2 α 2 = ω 2 ω 1 + ω 2
P redict = argmax x X   ( α 1 p i 1 + α 2 p i 2 )
where p i is the probability of class i ; x i is the i -th element of the input vectors.
Aiming to resolve the issue that traditional One-Hot encoding tends to make the model overconfident, the proposed method incorporates a label-smoothing [38] strategy into the CE loss. The probability distribution after label smoothing is shown in Equation (10), as follows:
y i = ( 1 ϵ ) , if   ( i = y ) ϵ K 1 , if   ( i y )
where ϵ is the label-smoothing factor and K is the total number of categories.
To enable layerwise importance quantification, we designed a DAWCE loss function based on focal loss (FL) [39]. The FL function introduces the weight factor α to adjust the contribution of samples, as well as a modulation factor ( 1 p t ) σ to guide the model to focus on the confusable samples, where α takes the value ( 0 , 1 ) and σ takes the value [ 0 , + ) . This study introduces ω 1 and ω 2 to construct adaptive adjustment coefficients, which help regulate the feature-learning process between different layers, thereby enabling the model to adaptively change weights when facing different categories of samples. Where σ is ( 2 ω ) , when the model has high confidence in a category, σ decreases, thereby suppressing the gradient update strength for such samples; conversely, when the model has low confidence in a category, σ increases, thereby enhancing the loss and accelerating the gradient update for such samples. Additionally, based on Equation (8), this study designed the dynamic weight factors α 1 and α 2 to dynamically adjust the contributions of the two branch losses to the overall loss. These weights reflect the model’s confidence in each branch. During training, α 1 and α 2 are adjusted based on the confidence of their respective branches, ensuring that the model avoids overfitting to categories with excessively high confidence within a branch. This adjustment helps prevent the over-tuning of the branch’s parameters. Dynamic adjustment is introduced to mitigate the risk of overfitting on easily confusable samples, to enhance the robustness of cross-level feature representation, and to maintain the stability of decision boundaries across categories. The DAWCE loss function is derived as in Equation (12), below:
L ( y i , p i 1 ) = α 1 ( 1 p t 1 ) ( 2 ω 1 ) i = 1 K y i log ( p i 1 ) L ( y i , p i 2 ) = α 2 ( 1 p t 2 ) ( 2 ω 2 ) i = 1 K y i log ( p i 2 )
where p t 1 and p t 2 are the probabilities that the model will correctly predict a sample.
L DAWCE = L ( y I , p i 1 ) + L ( I , p i 2 )

3.2. Determination of the Channel Expansion Factors at Different Stages of HFDF-EffNetV2

In the design of the CNN model architecture, the channel dimension, as a key adjustment parameter of model capacity, has an expansion ratio that is positively correlated with the model’s representational ability and directly affects the degree of overparameterization. Therefore, selecting an appropriate expansion factor is crucial in model design. Rolling bearings in actual operation are subject to interference from factors such as vibrations between equipment and environmental noise. To validate the diagnostic performance of HFDF-EffNetV2 using various expansion factors at different stages, this study simulated real working conditions by adding Gaussian white noise (GWN) with signal-to-noise ratios (SNR) of 2 dB, 4 dB, 6 dB, and 8 dB to the signal samples from the test set in Table 2. The SNR is calculated as in Equation (13), as follows:
SNR dB = 10 log 10 P signal P nosie
where SNR dB is the ratio of signal to noise (the lower the ratio, the higher the noise intensity); P signal is the power spectral density of the original signal; and P nosie is the power spectral density of the noise signal.
The experimental results are shown in Figure 7. When the noise intensity is low, the accuracy gap is not significant; therefore, the 2 dB experimental results under strong noise interference were selected as the basis for choosing the channel expansion factors. Under 2 dB of noise interference, the model achieves an accuracy of 92.60% when the channel expansion factors for different stages are (1, 2, 4, 6). The model reaches a maximum accuracy of 95.88% when the expansion factors are (2, 4, 4, 6), and this value decreases to 95.03% when the expansion factors are increased to (4, 4, 6, 8). The results indicate that an increase in the channel expansion factors helps the model fuse information from different channels, enabling it to extract diverse features more effectively. However, an excessively high expansion factor can cause the model to converge to a local optimum. Therefore, proper control of the channel expansion factors can help avoid interference from redundant information. In this study, the channel expansion factors(2, 4, 4, 6), which yielded the best diagnostic performance, were selected, and the specific architectural parameters of HFDF-EffNetV2 are shown in Table 3.

4. Discussion

To systematically evaluate applicability in engineering of HFDF-EffNetV2 and the characteristics related to its light weight, this study conduced a series of comparative experiments using the CWRU bearing dataset [40] and a self-constructed bearing experimental dataset, aiming to validate the model’s performance in terms of light weight, robustness to noise, and stability with small samples. Six advanced lightweight models were selected for comparison, including EfficientNetV2-S [35], ResNet-18 [41], RepViT [42], SqueezeNet [43], MPNet [44], and MobileNetV3 [45]. Additionally, ablation experiments were conducted on the CWRU bearing dataset to evaluate the effectiveness of the three improved methods. To ensure the reliability of the results, all experiments were repeated five times, and the average value was taken as the final result.

4.1. Experimental Hyperparameter Settings and Evaluation Metrics

Data preprocessing and model training were performed using a deep learning framework on PyTorch 3.11.1 based on the PyCharm platform. The system configuration was as follows: Windows 11, AMD Ryzen 7945HX CPU @ 2.5 GHz, NVIDIA GeForce RTX 4060 GPU (NVIDIA, Santa Clara, CA, USA). The experimental hyperparameters were set as in Table 4.
“Parameters” represents the total number of parameters that the model needs to train, and “time” refers to the time taken for diagnosis with a batch size of 16. These metrics were used to assess the model’s complexity and diagnostic speed. In addition, accuracy, precision (P), recall (R), and F1 score were used to evaluate the diagnostic performance of the model, as calculated in Equations (14)–(17), below:
Accuracy = TP + TN TP + FN + TN + FP
P = TP TP + FP
R = TP TP + FN
F 1 = 2 × P × R P + R
where TP , TN , FP , and FN are the number of true positives, true negatives, false positives, and false negatives, respectively.

4.2. Case 1: CWRU Bearing Dataset

In the experiment, data from the drive-end part of the CWRU bearing dataset with a motor load of 0 horsepower (HP), a speed of 1797 r/min, and a sampling frequency of 12 kilohertz (kHz) were used. The data were divided into four states: normal, inner-ring faults, outer-ring faults, and rolling-ball faults, with single-point fault diameters corresponding to 0.1778 mm (mm), 0.3556 mm, and 0.5334 mm. To ensure the capture of critical fault features, the sample length was set to 1024 data points, with an overlap window of 512, covering approximately 2.5 turns of bearing rotation. The data points were calculated as shown in Equation (18), below:
N P = 60 f R
where N P is the number of data points collected per revolution; f is the sampling frequency; and R is the bearing speed under the corresponding operating condition.
The raw vibration signals were standardized, and then the CWT was applied to generate a total of 2000 samples. Each class contained 130 training samples and 70 test samples, with the detailed composition of the dataset shown in Table 2.

4.2.1. Model Training

The training curves of HFDF-EffNetV2 are shown in Figure 8, where it can be observed that, after a few iterations, the training accuracy reached 100% and the training loss rapidly decreased to around 0.1. Although the loss curve fluctuated several times during the subsequent training, it ultimately converged steadily to 0.011. This indicates that the DAWCE loss function helped the model learn effective fault features, gradually reducing the error and causing it to converge towards optimization.

4.2.2. Comparison and Analysis of Model Diagnostic Performance

Table 5 presents the experimental results for the weight and diagnostic evaluation metrics of different models. The proposed model and EfficientNetV2-S both reached 100% accuracy, P, R, and F1, making them the best diagnostic performers among all compared models. Moreover, with the same optimal diagnostic performance as EfficientNetV2-S, the proposed model was significantly lighter-weight, reducing the number of parameters and the diagnostic time by 91% and 35%, respectively. While the proposed model’s parameters and diagnostic time are not optimal, the model’s performance was good. For example, the model has slightly more parameters than SqueezeNet (by 1.12 million (M)), but the time is 23% shorter. The time is slightly longer than that of ResNet-18 (by 1.14 milliseconds (ms)), but the number of parameters is significantly lower, by 83%. In summary, HFDF-EffNetV2 achieved optimal diagnostic performance while still maintaining an excellent lightweight profile, an outcome that validates the effectiveness and advancement of the model design.

4.2.3. Model Comparison Experiments and Analysis Under Noise Interference

According to Equation (13), GWN with varying SNRs was added to the test set signals to evaluate the robustness of the model to noise. The experimental results are shown in Figure 9.
After noise was added, the proposed model achieved an average accuracy of 98.19%, which is 18.39%, 11.23%, 5.07%, 2.38%, 2.37%, and 1.51% higher than the six comparison models. Furthermore, the proposed model performed optimally under the same noise intensity compared to the other models. As the noise intensity increased, the accuracy decreased by only 3.72%, which is 1.05% less than the decrease observed in the next-best model, EfficientNetV2-S. At 2 dB, the accuracies of SqueezeNet and MobileNetV3 dropped dramatically, to 61.77% and 76.14%, respectively; these values were 34.11% and 19.74% lower than that of the proposed model, showing a significant gap. ResNet-18, RepViT, and MPNet performed well at 6 dB and 8 dB but poorly at 2 dB, with their accuracies lower than that of the proposed model by 16.02%, 6.05%, and 4.51%, respectively. EfficientNetV2-S, which had the greatest number of parameters, performs=ed well, though its accuracy was slightly lower than that of the proposed model, by 2.25%, at 2 dB. The results show that there is a relationship between the model’s noise resistance and its parameter count. More importantly, optimizing the model architecture to enhance its feature extraction capabilities can significantly improve the robustness of the model under noise interference.

4.2.4. Visualization and Analysis Under Noise Interference

To compare the diagnostic performance of the models under noise interference for different fault types with a simulated SNR of 2 dB, this experiment visualized the confusion matrices of EfficientNetV2-S, RepViT, and MPNet, which performed better in the noise experiments, alongside HFDF-EffNetV2, as shown in Figure 10.
All models exhibited varying levels of misclassification in diagnosing fault label 2, with the number of misclassified samples for EfficientNetV2-S, RepViT, and MPNet being 30, 29, and 47, respectively, while the proposed model misclassified only 17 samples. Additionally, RepViT and MPNet exhibited poor diagnostic performance (P) for fault label 9, achieving only 70% and 84%, respectively. Although the performance of the proposed model was not optimal for all fault types, its diagnostic P reached 100% for fault labels 0, 1, 3, 4, and 6, and exceeded 94% for fault labels 5, 7, 8, and 9. The results show that HFDF-EffNetV2 significantly improves diagnostic P for samples with confusable fault types.

4.2.5. Comparative Experiments and Analysis with Small Sample Sizes

Due to the difficulty of obtaining large amounts of fault data in practical engineering applications, this experiment evaluated the diagnostic performance of HFDF-EffNetV2 with small sample sets by dividing the samples based on the total length of the sampled vibration signals and creating three datasets (A, B, and C) of different sizes. The number of training samples in datasets A, B, and C was 100, 200, and 300, respectively, while the number of test samples was fixed at 700.
The experimental results are shown in Table 6. ResNet-18 performed well, with an average accuracy of 98.55%, which is only 0.14% lower than that of the proposed model. EfficientNetV2-S, RepViT, and MPNet, which exhibited relatively good performance in the noise-interference experiments, showed instability when the training data were scarce, with average accuracies lower than that of the proposed model by 2.68%, 6.06%, and 2.72%, respectively. In contrast, the accuracies of SqueezeNet and MobileNetV3 were significantly lower than that of the proposed model, at 10.89% and 7.74%, respectively. On dataset A, the accuracies of all six comparison models decreased significantly, demonstrating that the models failed to extract effective fault features. Although the accuracy of the proposed model decreased compared to those of the other two training datasets on dataset A, it still remained above 98%, which was higher than those of the other six models by 6.35%, 0.09%, 12.69%, 16.12%, 6.09%, and 14.49%, respectively. The results indicate that HFDF-EffNetV2 is less influenced than the other models by the number of training samples and that its performance remains stable across varying sample sizes.

4.2.6. Visualization and Analysis with Small Sample Sizes

To further analyze the feature extraction performance of the model with small samples sizes, this experiment utilized the t-distributed stochastic neighbor embedding (t-SNE) technique to visualize the FD results of ResNet-18, EfficientNetV2-S, MPNet, and the proposed model on dataset A. This approach also allowed for the visualization of the clustering and separation of the extracted fault features after dimensionality reduction.
The visualization results are presented in Figure 11. The separation of fault features extracted by MPNet and EfficientNetV2-S for samples with fault labels 2, 4, 5, 8, and 9 was insufficiently clear, which could lead to misclassification of fault samples. In contrast, ResNet-18 demonstrated better clustering and separation of features than the previous two models, but the clustering centers for fault labels 4 and 5 were relatively close. Although the proposed model yielded a small number of misclassifications, the clustering and separation of fault features across different samples were satisfactory, highlighting its excellent feature extraction.

4.3. Case 2: Self-Constructed Bearing Experimental Dataset

The self-constructed bearing experimental platform primarily consisted of 6206 rolling bearings, a 0.55 kW three-phase asynchronous motor, 1C330ET three-axis vibration and temperature integrated sensors, and couplings, as illustrated in Figure 12. Based on the common faults of rolling bearings in real industrial scenarios, cage faults, inner-ring faults, outer-ring faults, rolling-ball faults, and rust faults were simulated through wire cutting and saltwater immersion, as shown in Figure 13. At a speed of 800 r/min and a sampling frequency of 25.6 kHz, approximately 250,000 datapoints were collected for each fault type. To ensure that a single sample would contain fault information for at least one full revolution, the overlap window was set to 1024 and the sample length was set to 2048, as calculated using Equation (18). The time-frequency FM, obtained after standardizing the original signal and performing the CWT, is shown in Figure 14. The dataset samples were configured to remain the same as in Case 1, as presented in Table 7.

4.3.1. Model Training

The training curves of HFDF-EffNetV2 are shown in Figure 15. The model started at 71.5% accuracy at the first iteration and then improved significantly, reaching 99.23% at the 7-th iteration and achieving 100% training accuracy after 100 iterations. Although there were significant fluctuations in the loss curves at the beginning of the model’s training, the loss values began to converge steadily to 0.01 after 20 iterations, indicating that the model’s weight updates tended to stabilize.

4.3.2. Model Diagnostic Performance Comparison and Analysis

Table 8 presents the experimental results for the lightweight and diagnostic evaluation metrics of different models. The proposed model achieved 100% in four diagnostic evaluation metrics, accuracy, P, R, and F1, demonstrating the best performance among all the compared models. Although the proposed model was not optimal in terms of parameters and diagnostic time, it improved all of the diagnostic evaluation metrics by 2.33% and reduced the time by 29% compared to the best small model of SqueezeNet. Additionally, there was only a 0.04 ms time difference between the proposed model and ResNet-18, but the number of parameters was only 17% of the number in ResNet-18.

4.3.3. Model-Comparison Experiments and Analysis Under Noise Interference

Consistent with the noise-interference experiment in Case 1, GWN with different SNRs was added to the test set signals. The experimental results are shown in Figure 16.
After noise was added, the proposed model still maintained the highest average accuracy, surpassing the other six models by 25.71%, 21.25%, 5.5%, 3.55%, 2.45%, and 1.3%, respectively. As the level of noise interference increased, the accuracies of all models decreased significantly. At 4 dB, SqueezeNet and MobileNetV3 exhibited the largest accuracy gaps with the proposed model, at 38.14% and 31.58%, respectively. Comparatively, ResNet-18, MPNet, and RepViT exhibited slightly better noise resistance than SqueezeNet and MobileNetV3, with deficits of 10.48%, 8.43%, and 6.34% relative to the proposed model, respectively. Although the accuracy of the proposed model was suboptimal at 2 dB, 1.76% lower than the accuracy of EfficientNetV2-S, it showed a notable improvement at 4 dB, surpassing that of EfficientNetV2-S by 4.2%. In summary, these findings suggest that models with larger parameter counts are more resilient to severe noise interference but that an optimal design can also improve the model’s robustness to noise.

4.3.4. Visualization and Analysis Under Noise Interference

In order to show the diagnostic results of the models for different fault types with a simulated SNR of 4 dB, this experiment visualized the confusion matrices of EfficientNetV2-S, RepViT, MPNet, and the proposed model.
The visualization results are shown in Figure 17, where the diagnostic P of the proposed model is 100% for fault labels 3 and 5 and 89% for fault labels 0, 1, and 2. Compared to EfficientNetV2-S, MPNet, and RepViT, which correctly diagnosed all samples with fault label 3, the diagnostic P for fault label 5 differed significantly from that of the proposed model, with differences of 22%, 28%, and 16%, respectively. In addition, all models performed poorly in the diagnosis of fault label 4. Despite the proposed model having eight more misclassified samples compared to the model with the fewest (EfficientNetV2-S), HFDF-EffNetV2 performed better overall. The result demonstrates that the HFDF strategy effectively mitigates the issue of the lightweight model falling into local overfitting.

4.3.5. Comparative Experiments and Analysis Under Small Samples Conditions

Consistent with the experimental setup in Case 1, this experiment was conducted with three different dataset sizes (A, B, and C), where the numbers of training samples were 100, 200, and 300, respectively, and the number of test set samples was fixed at 700.
The experimental results are presented in Table 9 and demonstrate that the proposed model achieves optimal diagnostic accuracy across different datasets. ResNet-18 performed excellently, with an average accuracy only 0.5% lower than that of the proposed model. EfficientNetV2-S and MPNet performed well on datasets B and C, but exhibited poor performance on dataset A, resulting in average accuracies 2.76% and 3.26% lower than those of the proposed model, respectively. RepViT, SqueezeNet, and MobileNetV3 exhibited poor performance overall, with average accuracies 5.96%, 8.95%, and 10.69% lower than those of the proposed model, respectively. As the number of training samples decreased, the accuracy of the proposed model decreased slightly by 2.59%, for a value 0.49% higher than that of ResNet18, the model with the smallest decrease. However, it remained more stable compared to the other five models, which showed decreases of 5.1%, 6.02%, 12.3%, 5.28%, and 13.78%. The results indicate that HFDF-EffNetV2 improves the ability to perceive the relationships between different channels in the input FM and that it continues to capture key information from important channels even with small sample sizes.

4.3.6. Visualization and Analysis with Small Sample Sizes

In this experiment, the t-SNE technique was employed to visualize the FD results of ResNet-18, EfficientNetV2-S, MPNet, and the proposed model in a small-sample experiment using dataset A.
The visualization results are shown in Figure 18. While MPNet and EfficientNetV2-S were effective in clustering and separating the features of fault labels 0, 3, 4, and 5; the clustering and separation of the features for fault labels 1 and 2 were not fully formed, indicating insufficient perception of the differences between fault types, which could result in the misclassification of fault samples. Compared to ResNet-18, the proposed model exhibited more distinct separation for fault labels 1 and 2, with misclassification occurring for only a few samples. The results indicate that HFDF-EffNetV2 effectively establishes a clear classification boundary, demonstrating stronger feature clustering and discriminative ability.

4.4. Analysis of Experimental Results in Two Cases

The results of the comparison experiments conducted demonstrate that the model architecture design significantly influences diagnostic performance. The lightweight architectures of SqueezeNet and MobileNetV3 constrain their capacity to capture salient local fault features, leading to poor robustness. ResNet-18 employs a simple architecture that achieves the fastest diagnostic time, but its limited depth compromises robustness to noise. While the multi-branch architectures of RepViT and MPNet enable effective capture of fault features with enhanced robustness to noise, such architectures inevitably reduce diagnostic speed. EfficientNetV2-S possesses a large model capacity, enabling it to capture deeper fault information and exhibit strong robustness to noise. However, this complexity also makes it sensitive to the size of the training dataset and results in the slowest diagnostic performance. In contrast, although HFDF-EffNetV2 is slightly slower than ResNet-18 and has a marginally higher parameters than SqueezeNet, it utilizes a lightweight multi-branch structure combined with a fine-grained feature extraction and decision fusion approach, significantly enhancing diagnostic accuracy.

4.5. Ablation Experiment and Analysis

Since the construction of HFDF-EffNetV2 involved multiple methods, ablation experiments were conducted on the CWRU bearing dataset to more intuitively visualize the impact of each method on the model’s diagnostic performance. As shown in Table 10, The BASE refers to the model constructed based on EfficientNetV2-S. The BASE + A and BASE + B models incorporated the methods of Fused-MBPyConv and BSMBConv-MLCA, respectively. The BASE + C model employed the HFDF strategy, and the BASE + A + B + C represents the proposed model.
The experimental results are presented in Figure 19 and demonstrate that the application of the three improvement methods significantly enhanced the diagnostic accuracies under strong noise interference. At 2 dB, the accuracy of the BASE + A + B + C was improved by 7.79% compared to the BASE. At 4 dB, BASE + C with the HFDF strategy applied showed improvements of 1.4%, 0.66%, and 0.09% compared to BASE, BASE + A, and BASE + B, respectively. At 6 dB and 8 dB, all models showed improved performance, making the relative improvement of the three enhanced methods compared to BASE non-significant. However, the average accuracies of the three enhanced methods were increased by 1.22%, 1.71%, and 1.46%, respectively, compared to BASE. Overall, all three improvement methods were effective in enhancing the model’s performance under noise interference, with the most significant contributions coming from the application of BSMBConv-MLCA.

5. Conclusions

This study addresses the challenge of achieving real-time and high-accuracy FD on platforms with limited computational resources by proposing a lightweight model called HFDF-EffNetV2.
In noise-interference experiments on the CWRU bearing dataset and a self-constructed bearing dataset, the average accuracy of the proposed model increased by 1.51%, 5.07%, 2.38%, 18.39%, 2.37%, and 11.23% and by 1.3%, 5.5%, 2.45%, 21.25%, 3.55%, and 25.71%, compared to EfficientNetV2-S, ResNet-18, RepViT, SqueezeNet, MPNet, and MobileNetV3, respectively. With small sample sizes, the proposed model outperformed these six models in average accuracy by 2.68%, 0.14%, 6.06%, 10.89%, 2.72%, and 7.74% and by 2.76%, 0.5%, 5.96%, 8.95%, 3.26%, and 10.69%, respectively. These results show that the Fused-MBPyConv and BSMBConv-MLCA proposed in this study achieved deeply coupled resolution of cross-scale fault features while maintaining light weight through feature fusion via staged channel-compression transformations. The HFDF strategy enables global integrated FD with cross-level fault feature relationships through multi-stage information fusion.
Although the proposed model achieved excellent diagnostic results in preliminary experiments and has 1.85 M parameters and faster diagnosis time compared with other lightweight models, which verifies its potential for real-time and highly accurate deployment on edge devices, the model may still be limited by the hardware configurations on resource-constrained edge devices. Indeed, in actual equipment operation, rolling bearings may experience fault types never before encountered by the models. Therefore, future research will combine signal processing with threshold-independent open-set identification strategies to enhance the diagnostic capabilities of the proposed model for unknown faults and evaluate its performance across different edge devices and real-world environments.

Author Contributions

Conceptualization, J.P. and D.Z.; methodology, D.Z.; software, D.Z.; validation, D.Z. and J.P.; formal analysis, D.Z. and T.H.; investigation, J.P.; resources, J.P.; data curation, F.H.; writing—original draft preparation, D.Z.; writing—review and editing, J.P., T.H. and J.N.; visualization, D.Z.; supervision, J.P.; project administration, F.H.; funding acquisition, F.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Guangxi Key Research and Development Program Project Grant No. AB24010202.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The CWRU bearing datasets are available in the references; the Self-constructed Bearing Experimental Dataset will be made available on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

FDFault diagnosis
HFDFHierarchical Fine-grained Decision Fusion
HFDF-EffNetv2HFDF strategy with an improved EfficientNetV2 architecture
PyConvPyramidal convolution
Fused-MBPyConvFused-MBConv with PyConv
BSConv-SSubspace blueprint separable convolution
MLCAMixed local channel attention
BSMBConv-MLCADesigned based on MBConv and applying MLCA and BSConv-S
CWRUCase Western Reserve University
CNNConvolutional neural network
FMFeature maps
CWTContinuous wavelet transform
DAWCEDynamic adaptive weighted cross-entropy
FLFocal loss
GWNGaussian white noise
SNRSignal-to-noise ratio
t-SNEt-distributed stochastic neighbor embedding

References

  1. Chen, X.; Yang, R.; Xue, Y.; Huang, M.; Ferrero, R.; Wang, Z. Deep Transfer Learning for Bearing Fault Diagnosis: A Systematic Review Since 2016. IEEE Trans. Instrum. Meas. 2023, 72, 3508221. [Google Scholar] [CrossRef]
  2. Tao, H.; Qiu, J.; Chen, Y.; Stojanovic, V.; Cheng, L. Unsupervised cross-domain rolling bearing fault diagnosis based on time-frequency information fusion. J. Franklin Inst. 2023, 360, 1454–1477. [Google Scholar] [CrossRef]
  3. Liao, J.-X.; Wei, S.-L.; Xie, C.-L.; Zeng, T.; Sun, J.; Zhang, S.; Zhang, X.; Fan, F.-L. BearingPGA-Net: A Lightweight and Deployable Bearing Fault Diagnosis Network via Decoupled Knowledge Distillation and FPGA Acceleration. IEEE Trans. Instrum. Meas. 2024, 73, 3506414. [Google Scholar] [CrossRef]
  4. Deng, J.; Liu, H.; Fang, H.; Shao, S.; Wang, D.; Hou, Y.; Chen, D.; Tang, M. MgNet: A fault diagnosis approach for multi-bearing system based on auxiliary bearing and multi-granularity information fusion. Mech. Syst. Signal Process. 2023, 193, 110253. [Google Scholar] [CrossRef]
  5. Chen, B.; Zhang, W.; Gu, J.X.; Song, D.; Cheng, Y.; Zhou, Z.; Gu, F.; Ball, A.D. Product envelope spectrum optimization-gram: An enhanced envelope analysis for rolling bearing fault diagnosis. Mech. Syst. Signal Process. 2023, 193, 110270. [Google Scholar] [CrossRef]
  6. Li, J.; Luo, W.; Bai, M. Review of research on signal decomposition and fault diagnosis of rolling bearing based on vibration signal. Meas. Sci. Technol. 2024, 35, 092001. [Google Scholar] [CrossRef]
  7. Miao, Y.; Zhang, B.; Li, C.; Lin, J.; Zhang, D. Feature Mode Decomposition: New Decomposition Theory for Rotating Machinery Fault Diagnosis. IEEE Trans. Ind. Electron. 2023, 70, 1949–1960. [Google Scholar] [CrossRef]
  8. Cheng, Z.; Yang, Y.; Hu, N.; Cheng, Z.; Cheng, J. All time-scale decomposition method and its application in gear fault diagnosis. Struct. Health Monit. 2024, 0, 14759217241289873. [Google Scholar] [CrossRef]
  9. Zheng, X.; Cheng, Z.; Cheng, J.; Yang, Y.; Yang, X. Blaschke mode decomposition: Algorithm and application. Struct. Health Monit. 2025, 14759217241310576. [Google Scholar] [CrossRef]
  10. Wang, Z.; Luo, Q.; Chen, H.; Zhao, J.; Yao, L.; Zhang, J.; Chu, F. A high-accuracy intelligent fault diagnosis method for aero-engine bearings with limited samples. Comput. Ind. 2024, 159, 104099. [Google Scholar] [CrossRef]
  11. Zhu, X.; Zhao, X.; Yao, J.; Deng, W.; Shao, H.; Liu, Z. Adaptive Multiscale Convolution Manifold Embedding Networks for Intelligent Fault Diagnosis of Servo Motor-Cylindrical Rolling Bearing Under Variable Working Conditions. IEEE/ASME Trans. Mechatron. 2024, 29, 2230–2240. [Google Scholar] [CrossRef]
  12. Zhao, X.; Zhu, X.; Liu, J.; Hu, Y.; Gao, T.; Zhao, L.; Yao, J.; Liu, Z. Model-Assisted Multi-source Fusion Hypergraph Convolutional Neural Networks for intelligent few-shot fault diagnosis to Electro-Hydrostatic Actuator. Inf. Fusion 2024, 104, 102186. [Google Scholar] [CrossRef]
  13. Chang, H.; Zhang, X.; Long, Y.; Zhang, Y.; Zhang, K.; Ding, C.; Wang, J.; Li, Y. WCNN-RSN: A novel fault diagnosis method for rolling bearing using multimodal feature fusion. Meas. Sci. Technol. 2024, 35, 126145. [Google Scholar] [CrossRef]
  14. Han, Y.; Zhang, F.; Li, Z.; Wang, Q.; Li, C.; Lai, P.; Li, T.; Teng, F.; Jin, Z. MT-ConvFormer: A Multitask Bearing Fault Diagnosis Method Using a Combination of CNN and Transformer. IEEE Trans. Instrum. 2025, 74, 3501816. [Google Scholar] [CrossRef]
  15. Song, Q.; Wang, J.; Song, Q.; Li, K.; Hao, W.; Jiang, H. Fault diagnosis of HVCB via the subtraction average based optimizer algorithm optimized multi channel CNN-SABO-SVM network. Sci. Rep. 2024, 14, 29507. [Google Scholar] [CrossRef] [PubMed]
  16. Hu, Q.; Fu, X.; Sun, D.; Xu, D.; Guan, Y. A Lightweight Rolling Bearing Fault Diagnosis Method Based on Multiscale Depth-Wise Separable Convolutions and Network Pruning. IEEE Access 2024, 12, 186131–186144. [Google Scholar] [CrossRef]
  17. Liu, L.; Cheng, Y.; Song, D.; Zhang, W.; Tang, G.; Luo, Y. A Lightweight Network With Adaptive Input and Adaptive Channel Pruning Strategy for Bearing Fault Diagnosis. IEEE Trans. Instrum. 2024, 73, 3510911. [Google Scholar] [CrossRef]
  18. Fan, Z.; Xu, X.; Wang, R.; Wang, H. Fan Fault Diagnosis Based on Lightweight Multiscale Multiattention Feature Fusion Network. IEEE Trans. Instrum. 2022, 18, 4542–4554. [Google Scholar] [CrossRef]
  19. Wang, Z.; Lu, H.; Shi, Y.; Wang, X. Lightweight CNN Architecture Design Based on Spatial–Temporal Tensor and Its Application in Bearing Fault Diagnosis. IEEE Trans. Instrum. Meas. 2024, 73, 3504112. [Google Scholar] [CrossRef]
  20. Teta, A.; Korich, B.; Bakria, D.; Hadroug, N.; Rabehi, A.; Alsharef, M.; Bajaj, M.; Zaitsev, I.; Ghoneim, S.S.M. Fault detection and diagnosis of grid-connected photovoltaic systems using energy valley optimizer based lightweight CNN and wavelet transform. Sci. Rep. 2024, 14, 18907. [Google Scholar] [CrossRef]
  21. Wang, C.; Tian, B.; Yang, J.; Jie, H.; Chang, Y.; Zhao, Z. Neural-transformer: A brain-inspired lightweight mechanical fault diagnosis method under noise. Reliab. Eng. Syst. Saf. 2024, 251, 110409. [Google Scholar] [CrossRef]
  22. Xie, Z.; Chen, J.; Shi, Z.; Liu, S.; He, S. Lightweight pyramid attention residual network for intelligent fault diagnosis of machine under sharp speed variation. Mech. Syst. Signal Process. 2025, 223, 111824. [Google Scholar] [CrossRef]
  23. Shin, J.; Lee, S. Robust and lightweight deep learning model for industrial fault diagnosis in low-quality and noisy data. Electronics 2023, 12, 409. [Google Scholar] [CrossRef]
  24. Lee, S.; Kim, T. FRFconv-TDSNet: Lightweight, Noise-Robust Convolutional Neural Network Leveraging Full-Receptive-Field Convolution and Time-Domain Statistics for Intelligent Machine Fault Diagnosis. IEEE Trans. Instrum. Meas. 2024, 73, 3528213. [Google Scholar] [CrossRef]
  25. Wang, C.; Xu, M.; Zhang, Q.; Zhang, D. A Novel Lightweight Rotating Mechanical Fault Diagnosis Framework With Adaptive Residual Enhancement and Multigroup Coordinate Attention. IEEE Trans. Instrum. Meas. 2025, 74, 3514517. [Google Scholar] [CrossRef]
  26. Zhang, J.; Zhao, Z.; Jiao, Y.; Zhao, R.; Hu, X.; Che, R. DPCCNN: A new lightweight fault diagnosis model for small samples and high noise problem. Neurocomputing 2025, 626, 129526. [Google Scholar] [CrossRef]
  27. Li, F.; Zhao, X. A novel approach for bearings multiclass fault diagnosis fusing multiscale deep convolution and hybrid attention networks. Meas. Sci. Technol. 2024, 35, 045017. [Google Scholar] [CrossRef]
  28. Yan, R.; Shang, Z.; Xu, H.; Wen, J.; Zhao, Z.; Chen, X.; Gao, R.X. Wavelet transform for rotary machine fault diagnosis: 10 years revisited. Mech. Syst. Signal Process. 2023, 200, 110545. [Google Scholar] [CrossRef]
  29. Duta, I.C.; Liu, L.; Zhu, F.; Shao, L. Pyramidal Convolution: Rethinking Convolutional Neural Networks for Visual Recognition. arXiv 2020. [Google Scholar] [CrossRef]
  30. Haase, D.; Amthor, M. Rethinking Depthwise Separable Convolutions: How Intra-Kernel Correlations Lead to Improved MobileNets. arXiv 2020. [Google Scholar] [CrossRef]
  31. Wan, D.; Lu, R.; Shen, S.; Xu, T.; Lang, X.; Ren, Z. Mixed local channel attention for object detection. Eng. Appl. Artif. Intell. 2023, 123, 106442. [Google Scholar] [CrossRef]
  32. Gao, S.; He, J.; Pan, H.; Gong, T. A multi-scale and lightweight bearing fault diagnosis model with small samples. Symmetry 2022, 14, 909. [Google Scholar] [CrossRef]
  33. Xie, F.; Li, G.; Song, C.; Song, M. The Early Diagnosis of Rolling Bearings’ Faults Using Fractional Fourier Transform Information Fusion and a Lightweight Neural Network. Fractal and Fractional 2023, 7, 875. [Google Scholar] [CrossRef]
  34. Wang, S.; Feng, Z. Multi-sensor fusion rolling bearing intelligent fault diagnosis based on VMD and ultra-lightweight GoogLeNet in industrial environments. Digital Signal Process. 2024, 145, 104306. [Google Scholar] [CrossRef]
  35. Tan, M.; Le, Q. Efficientnetv2: Smaller Models and Faster Training. arXiv 2021. [Google Scholar] [CrossRef]
  36. Misra, D. Mish: A Self Regularized Non-Monotonic Activation Function. arXiv 2019. [Google Scholar] [CrossRef]
  37. Chen, J.; Wang, P.; Liu, J.; Qian, Y. Label Relation Graphs Enhanced Hierarchical Residual Network for Hierarchical Multi-Granularity Classification. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 4858–4867. [Google Scholar] [CrossRef]
  38. Müller, R.; Kornblith, S.; Hinton, G.E. When Does Label Smoothing Help? arXiv 2019. [Google Scholar] [CrossRef]
  39. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollr, P. Focal Loss for Dense Object Detection. arXiv 2017. [Google Scholar] [CrossRef]
  40. Smith, W.A.; Randall, R.B. Rolling element bearing diagnostics using the Case Western Reserve University data: A benchmark study. Mech. Syst. Signal Process. 2015, 64, 100–131. [Google Scholar] [CrossRef]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  42. Wang, A.; Chen, H.; Lin, Z.; Han, J.; Ding, G. Repvit: Revisiting Mobile CNN From ViT Perspective. In Proceedings of the 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 16–22 June 2024; pp. 15909–15920. [Google Scholar] [CrossRef]
  43. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016. [Google Scholar] [CrossRef]
  44. Liu, Y.; Chen, Y.; Li, X.; Zhou, X.; Wu, D. MPNet: A lightweight fault diagnosis network for rotating machinery. Measurement 2025, 239, 115498. [Google Scholar] [CrossRef]
  45. Howard, A.; Sandler, M.; Chu, G.; Chen, L.-C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for Mobilenetv3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar] [CrossRef]
Figure 1. PyConv structure.
Figure 1. PyConv structure.
Applsci 15 04902 g001
Figure 2. MLCA mechanism structure.
Figure 2. MLCA mechanism structure.
Applsci 15 04902 g002
Figure 3. (a) The Fused-MBConv structure and (b) The MBConv structure.
Figure 3. (a) The Fused-MBConv structure and (b) The MBConv structure.
Applsci 15 04902 g003
Figure 4. FD process based on HFDF-EffNetV2.
Figure 4. FD process based on HFDF-EffNetV2.
Applsci 15 04902 g004
Figure 5. Fused-MBPyConv structure.
Figure 5. Fused-MBPyConv structure.
Applsci 15 04902 g005
Figure 6. BSMBConv-MLCA structure.
Figure 6. BSMBConv-MLCA structure.
Applsci 15 04902 g006
Figure 7. Comparison of the numbers of parameters and diagnostic accuracies of HFDF-EffNetV2 at different expansion factors and at different SNRs on the CWRU bearing dataset.
Figure 7. Comparison of the numbers of parameters and diagnostic accuracies of HFDF-EffNetV2 at different expansion factors and at different SNRs on the CWRU bearing dataset.
Applsci 15 04902 g007
Figure 8. Accuracy and loss curves of HFDF-EfftNetV2 during training in Case 1.
Figure 8. Accuracy and loss curves of HFDF-EfftNetV2 during training in Case 1.
Applsci 15 04902 g008
Figure 9. Comparison of experimental results for different models under different SNRs in case 1.
Figure 9. Comparison of experimental results for different models under different SNRs in case 1.
Applsci 15 04902 g009
Figure 10. Confusion matrix visualization for different models at SNR (2 dB) in Case 1.
Figure 10. Confusion matrix visualization for different models at SNR (2 dB) in Case 1.
Applsci 15 04902 g010
Figure 11. The t-SNE visualization results for different models on dataset A (with a training sample size of 10) in Case 1.
Figure 11. The t-SNE visualization results for different models on dataset A (with a training sample size of 10) in Case 1.
Applsci 15 04902 g011
Figure 12. The self-constructed experimental platform for bearings.
Figure 12. The self-constructed experimental platform for bearings.
Applsci 15 04902 g012
Figure 13. Condition of rolling bearings.
Figure 13. Condition of rolling bearings.
Applsci 15 04902 g013
Figure 14. Time-frequency FM of different fault types obtained from the original signals collected by the self-constructed bearing experimental platform after standardization and CWT.
Figure 14. Time-frequency FM of different fault types obtained from the original signals collected by the self-constructed bearing experimental platform after standardization and CWT.
Applsci 15 04902 g014
Figure 15. Accuracy and loss curves of HFDF-EfftNetV2 during the training process in Case 2.
Figure 15. Accuracy and loss curves of HFDF-EfftNetV2 during the training process in Case 2.
Applsci 15 04902 g015
Figure 16. Comparison of experimental results of different models under different SNRs in Case 2.
Figure 16. Comparison of experimental results of different models under different SNRs in Case 2.
Applsci 15 04902 g016
Figure 17. Confusion matrix visualization for different models at SNR (4 dB) in Case 2.
Figure 17. Confusion matrix visualization for different models at SNR (4 dB) in Case 2.
Applsci 15 04902 g017aApplsci 15 04902 g017b
Figure 18. The t-SNE visualization results for different models on dataset A (with a training sample size of 10) in Case 2.
Figure 18. The t-SNE visualization results for different models on dataset A (with a training sample size of 10) in Case 2.
Applsci 15 04902 g018
Figure 19. Experimental results for different ablation models under different SNRs in Case 1.
Figure 19. Experimental results for different ablation models under different SNRs in Case 1.
Applsci 15 04902 g019
Table 1. EfficientNetV2-S architecture.
Table 1. EfficientNetV2-S architecture.
StageOperatorStrideChannelsLayers
0Conv 3 × 3 2241
1Fused-MBConv 1   1 ,   k 3 × 3 1242
2Fused-MBConv 4 ,   k 3 × 3 2484
3Fused-MBConv 4 ,   k 3 × 3 2644
4MBConv4, k 3 × 3 , SE0.2521286
5MBConv6, k 3 × 3 , SE0.2511609
6MBConv6, k 3 × 3 , SE0.25225615
7Conv 1 × 1 & Pooling & FC -12801
1 The digit following “MBConv” denotes the expansion factor of the input channel.
Table 2. Case Western Reserve University (CWRU) bearing dataset composition.
Table 2. Case Western Reserve University (CWRU) bearing dataset composition.
Fault SizesFault TypeCodeFault Label 1TrainTest
NoneNormalN013070
Inner RingIR1113070
0.1778 mmRolling BallRB1213070
Outer RingOR1313070
Inner RingIR2413070
0.3556 mmRolling BallRB2513070
Outer RingOR2613070
Inner RingIR3713070
0.5334 mmRolling BallRB3813070
Outer RingOR3913070
1 The fault labels are utilized to categorize samples based on various fault types, with each label representing a specific type of fault.
Table 3. The specific architectural parameters of HFDF-EffNetV2.
Table 3. The specific architectural parameters of HFDF-EffNetV2.
StageOperatorStrideExpansion FactorInputOutputLayers
-Conv 3 × 3 2-3201
1Fused-MBPyConv[2, 1, 1]220403
2Fused-MBPyConv[2, 1, 1]440803
3BSMBConv-MLCA[2, 1, 1]4801203
-Conv1x1&GAP&FC--960-1
4BSMBConv-MLCA[2, 1, 1]61201603
-Conv1x1&GAP&FC--1280-1
Table 4. The experimental hyperparameter settings.
Table 4. The experimental hyperparameter settings.
EpochBSLR 1Step LR 1OptimizerWD 2Seed
100160.000520Adam0.0142
1 Step LR indicates the learning rate is halved every twenty rounds of iteration. 2 WD is the weight decay factor, a widely used regularization technique designed to reduce overfitting in the model.
Table 5. Comparison of different models’ diagnostic performance in Case 1.
Table 5. Comparison of different models’ diagnostic performance in Case 1.
ModelAccuracy/%P/%R/%F1/%Parameters/MTime/ms
EfficientNetV2-S100 ± 0.00100 ± 0.00100 ± 0.00100 ± 0.0020.1028.73 ± 0.84
ResNet-1899.71 ± 0.0099.71 ± 0.0099.71 ± 0.0099.71 ± 0.0011.1817.53 ± 0.32
RepViT99.83 ± 0.1499.83 ± 0.1499.83 ± 0.1499.83 ± 0.144.7222.69 ± 0.39
SqueezeNet99.23 ± 0.1299.23 ± 0.1199.23 ± 0.1299.23 ± 0.110.7324.40 ± 0.43
MPNet99.63 ± 0.0799.63 ± 0.0799.63 ± 0.0799.63 ± 0.072.5021.45 ± 0.58
MobileNetV398.40 ± 0.2598.40 ± 0.2498.40 ± 0.2498.40 ± 0.243.9119.34 ± 0.31
HFDF-EffNetV2 (Ours)100 ± 0.00100 ± 0.00100 ± 0.00100 ± 0.001.8518.67 ± 0.31
Table 6. Comparison of experimental results for different models with different training-sample sizes in Case 1.
Table 6. Comparison of experimental results for different models with different training-sample sizes in Case 1.
ModelAccuracy/%
A/(Train10/Test70)B/(Train20/Test70)C/(Train30/Test70)Average
EfficientNetV2-S91.74 ± 1.1497.74 ± 0.4398.54 ± 0.2596.01
ResNet-1898.00 ± 0.4298.74 ± 0.1698.91 ± 0.1598.55
RepViT85.40 ± 1.3294.63 ± 0.8997.86 ± 0.4592.63
SqueezeNet81.97 ± 1.5588.71 ± 0.8292.71 ± 0.5587.80
MPNet92.00 ± 0.9297.34 ± 0.4898.57 ± 0.1895.97
MobileNetV383.60 ± 1.593.86 ± 0.7695.40 ± 0.4490.95
HFDF-EffNetV2 (Ours)98.09 ± 0.5298.91 ± 0.1799.09 ± 0.1298.69
Table 7. Self-Constructed bearing experimental dataset composition.
Table 7. Self-Constructed bearing experimental dataset composition.
Fault TypeFault LabelTrainTest
Normal013070
Cage Fault113070
Inner Ring Fault213070
Outer Ring Fault313070
Rolling Ball Fault413070
Rust Fault513070
Table 8. Comparison of different models’ diagnostic performances in Case 2.
Table 8. Comparison of different models’ diagnostic performances in Case 2.
ModelAccuracy/%P/%R/%F1/%Parameters/MTime/ms
EfficientNetV2-S99.67 ± 0.1999.67 ± 0.1999.67 ± 0.1999.67 ± 0.1920.1028.05 ± 0.72
ResNet-1899.95 ± 0.1099.95 ± 0.0999.95 ± 0.1099.95 ± 0.1011.1817.52 ± 0.62
RepViT99.14 ± 0.4499.16 ± 0.4499.14 ± 0.4499.14 ± 0.444.7223.17 ± 0.77
SqueezeNet97.67 ± 0.1897.67 ± 0.1797.67 ± 0.1897.67 ± 0.180.7324.57 ± 0.40
MPNet99.63 ± 0.2499.63 ± 0.2499.63 ± 0.2499.63 ± 0.242.5021.00 ± 0.51
MobileNetV395.95 ± 0.4895.95 ± 0.4895.95 ± 0.4895.95 ± 0.483.9118.69 ± 0.35
HFDF-EffNetV2 (Ours)100 ± 0.00100 ± 0.00100 ± 0.00100 ± 0.001.8517.56 ± 0.41
Table 9. Comparison of experimental results for different models with different sizes of training samples in Case 2.
Table 9. Comparison of experimental results for different models with different sizes of training samples in Case 2.
ModelAccuracy/%
A/(Train10/Test70)B/(Train20/Test70)C/(Train30/Test70)Average
EfficientNetV2-S91.29 ± 0.7896.58 ± 0.8996.39 ± 0.8494.75
ResNet-1895.71 ± 0.6697.52 ± 0.6197.81 ± 0.2897.01
RepViT88.17 ± 1.1092.28 ± 0.9694.19 ± 0.7791.55
SqueezeNet80.57 ± 1.6192.24 ± 0.5892.87 ± 0.7288.56
MPNet90.81 ± 0.7695.86 ± 0.5296.09 ± 0.3694.25
MobileNetV378.04 ± 1.3590.59 ± 0.7691.82 ± 0.6286.82
HFDF-EffNetV2 (Ours)95.88 ± 0.9198.19 ± 0.4298.47 ± 0.4197.51
Table 10. The number of parameters and diagnostic time for ablation models.
Table 10. The number of parameters and diagnostic time for ablation models.
ModelBASEBASE + ABASE + BBASE + CBASE + A + B + C
Parameters/M2.312.371.672.441.85
Time/ms15.24 ± 0.2917.43 ± 0.3117.18 ± 0.3015.80 ± 0.2918.67 ± 0.31
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, D.; Pan, J.; Huang, T.; Niu, J.; Huang, F. HFDF-EffNetV2: A Lightweight, Noise-Robust Model for Fault Diagnosis in Rolling Bearings. Appl. Sci. 2025, 15, 4902. https://doi.org/10.3390/app15094902

AMA Style

Zhang D, Pan J, Huang T, Niu J, Huang F. HFDF-EffNetV2: A Lightweight, Noise-Robust Model for Fault Diagnosis in Rolling Bearings. Applied Sciences. 2025; 15(9):4902. https://doi.org/10.3390/app15094902

Chicago/Turabian Style

Zhang, Donglei, Jiafang Pan, Tianping Huang, Junlin Niu, and Faguo Huang. 2025. "HFDF-EffNetV2: A Lightweight, Noise-Robust Model for Fault Diagnosis in Rolling Bearings" Applied Sciences 15, no. 9: 4902. https://doi.org/10.3390/app15094902

APA Style

Zhang, D., Pan, J., Huang, T., Niu, J., & Huang, F. (2025). HFDF-EffNetV2: A Lightweight, Noise-Robust Model for Fault Diagnosis in Rolling Bearings. Applied Sciences, 15(9), 4902. https://doi.org/10.3390/app15094902

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop