Next Article in Journal
RFID Sensor with Integrated Energy Harvesting for Wireless Measurement of dc Magnetic Fields
Previous Article in Journal
Detection and Recognition of Visual Geons Based on Specific Object-of-Interest Imaging Technology
Previous Article in Special Issue
Efficient Gearbox Fault Diagnosis Based on Improved Multi-Scale CNN with Lightweight Convolutional Attention
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on Open-Set Recognition Methods for Rolling Bearing Fault Diagnosis

1
School of Electronic Information, Sichuan University, Chengdu 610065, China
2
National Key Laboratory of Science and Technology on Reactor System Design Technology, Nuclear Power Institute of China, Chengdu 610213, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(10), 3019; https://doi.org/10.3390/s25103019
Submission received: 20 March 2025 / Revised: 29 April 2025 / Accepted: 8 May 2025 / Published: 10 May 2025
(This article belongs to the Special Issue Deep Learning Based Intelligent Fault Diagnosis)

Abstract

In rolling bearing fault diagnosis, when an unknown fault is present, the Closed-Set Recognition (CSR) method tends to misclassify it as a known fault. To address this issue, an Open-Set Recognition (OSR) framework is proposed for rolling bearing fault diagnosis in this study. The framework is built upon a serial multi-scale convolutional prototype learning (SMCPL) network, enhanced with an efficient channel attention (ECA) mechanism to extract the most critical fault features. The extracted features are fed into the Density Peak Clustering (DPC) module, which identifies known and unknown classes based on the computed local densities and relative distances. Finally, validation is performed on the Case Western Reserve University (CWRU) dataset, the Xi’an Jiaotong University rolling bearing accelerated life test dataset (XJTU-SY), and the Paderborn University bearing dataset (PU), Germany, and the framework is comprehensively evaluated in terms of several evaluation metrics, such as normalization accuracy and feature visualization. The experimental results show that SMCPL-ECA-DPC outperforms the comparative methods of SMCPL, CPL, ANEDL, CNN, and OpenMax and has high diagnostic performance in the identification of unknown faults.

1. Introduction

Rolling bearings are critical components in mechanical systems, with high demands for reliability and performance. Therefore, the timely and accurate identification and diagnosis of bearing faults are crucial [1,2]. In the past decade, data-driven fault diagnosis methods have rapidly advanced, with machine learning and deep learning being widely applied in intelligent fault diagnosis [3,4,5,6,7,8,9,10,11]. For instance, Xu et al. [5] proposed an improved multi-scale convolutional neural network model combined with a feature attention mechanism (IMS-FACNN) for diagnosing bearing faults in real wind turbines. Ruan et al. [8], based on the physical characteristics of bearing acceleration signals, guided the design of CNN hyperparameters for fault diagnosis. However, these studies [5,6,7,8] were conducted under the assumption that both the training and testing datasets come from the same distribution, resulting in effective fault diagnosis in a Closed-Set Recognition (CSR) context. In practical applications, however, due to experimental constraints, the training set may not cover all fault types and operating conditions. The studies above are prone to misclassification when encountering untrained data. Therefore, this study of Open-Set Recognition (OSR) is of great value, both for the correct classification of known fault types and for the identification of unknown fault types [12].
In traditional fault diagnosis research, CSR typically employs feature extractors, often neural networks, to obtain distinguishable features between known and unknown faults, relying on the Softmax operation for classification. Softmax uses the cross-entropy loss function (CE) for optimization, focusing on maximizing the probability distribution of known classes while neglecting the specific feature relationships between samples and the distance information between classes. This approach has inherent limitations when applied to OSR tasks.
To better solve the OSR problem, researchers have used different approaches. Bendale and Boult [13] proposed a new model layer, OpenMax, to compute the likelihood and estimate the probability that a given input belongs to an unknown class. However, this method relies on a large number of known class samples to estimate the activation vectors and distributions for each class during the training process, which places high demands on the number of known class samples, as well as computational resources. In particular, when certain category samples are few or even scarce, the performance of OpenMax may be significantly affected, as the differentiation between categories may be insufficient, resulting in less clear classification boundaries, which reduces the accuracy of the classification results. In addition, the method is more computationally intensive when the number of samples and the number of categories is large, which can significantly increase the time overhead of training and inference. Rudd et al. [14] proposed the Extreme Value Machine (EVM) based on the statistical Extreme Value Theory (EVT), which handles unknown query classes by selecting the points and distributions with the least redundancy between each category and constructing a compact probabilistic representation of the decision boundaries. However, the EVM approach has some limitations, especially when dealing with complex data distributions. EVMs usually rely on simple model assumptions, such as that the data distribution is linearly differentiable. However, in reality, data distributions are often more complex, and EVMs may lead to an oversimplification of the decision boundary, especially when faced with failure data that are highly nonlinear or multimodal. Such simplifications may not effectively capture the diversity of the data, and the performance of EVM may be degraded, especially when the data are non-uniformly distributed or have large differences between classes. EDL [15] is a learning method for self-training and inference that performs well in uncertainty quantification. However, Yu et al. [16] empirically found that it has low classification accuracy in classification tasks. In order to enhance the separation between outliers and outliers, Yu et al. proposed self-adaptive negative optimization to explicitly compress the evidential value of outliers in unlabeled data for application in open-set semi-supervised learning. However, the method relies on the high quality and quantity of labeled data and requires multiple iterations of optimization with high computational overhead.
In the diagnosis of rolling bearings, on the other hand, the fault data are often high-frequency vibration signals, characterized by different scales and frequencies, and fewer monitoring variables, usually vibration signals such as acceleration, velocity and displacement. The treatment of OSR problems in the field of rolling bearings also belongs to a type of unbalance fault diagnosis. In 2025, among the unbalanced data fault diagnosis research, Cai [17] et al. proposed the Multi-scale Dynamic Graph Mutual Information Network (MDGMIN), which includes graph generation and a multi-scale dynamic edge graph convolutional layer and mutual information learning, to solve the planetary bearing health-monitoring problem under unbalanced data by weighting and fusing specific features extracted from multi-scale dynamic edge graph convolutional layer through the attentional mechanism. Zhao [18] constructed a multi-scale Chebyshev graph convolutional layer with multi-perceptual learning as the backbone network of the multi-perceptual graph convolutional tree embedded network to capture special and shared information for the severe imbalance data of aero-engine bearings. It can be seen that for the extraction of key information from data, the capture of multi-scale information and the focusing of attention mechanisms are highly promising research tools for the present and the future.
In addition to the above methods, due to the powerful feature extraction capability and classification performance demonstrated by CNNs in the field of fault diagnosis [5,6,7,8], Yang [19] et al. proposed the Convolutional Prototype Learning (CPL) framework based on CNNs by using prototype loss (PL) as a regularization to train the data for learning. More discriminative feature representations are produced, and OSR problem solving is greatly aided by CPL’s significant improvements in intra-class compactness and inter-class separability when compared to classic CNN [20,21,22]. But it is difficult for CPL to comprehensively capture the multi-scale information and critical failure information of the samples because it relies on the convolution operation at fixed scales. To overcome the above challenges, a rolling bearing OSR framework is proposed in this paper. The framework is based on the network structure of Serial Multi-scale Convolutional Prototype Learning (SMCPL) and adds the Efficient Channel Attention (ECA) mechanism to extract the feature representations and then uses Density Peak Clustering (DPC) for recognition, which can efficiently solve the problems of insufficient multi-scale feature extraction and insufficient information interaction between channels.
The contributions of this paper are as follows:
(1)
This framework is based on the serial multi-scale convolutional prototype learning network framework for feature extraction while adding an efficient channel attention mechanism so that the SMCPL-ECA network can pay more attention to the features that are important for fault identification.
(2)
The Density Peak Clustering (DPC) method is used to cluster the extracted features for identification.
(3)
The experiments were carried out using the Case Western Reserve University dataset (CWRU), the Xi’an Jiaotong University rolling bearing accelerated life test dataset (XJTU-SY), and the Paderborn University (PU) bearing dataset. For CWRU, 18 experiments with different known and unknown fault types were set up considering different fault diameters, and for XJTU-SY and PU, a total of 12 experiments were set up considering three different types of damage sites under the same operating conditions for testing in order to comprehensively evaluate the feasibility and superiority of the proposed framework.
The rest of this paper is organized as follows: Section 2 introduces the relevant principles of OSR, ECA, SMCPL, and DPC. Section 3 discusses the proposed framework and presents the assessment metrics. Section 4 presents the experimental data, OSR experimental setup, and model parameters. Section 5 presents the experimental results and analysis. Section 6 summarizes the conclusions of this study and future research plans.

2. Related Theory

2.1. Open-Set Recognition (OSR)

Research for OSR has been applied in many fields, including security monitoring [23,24], face recognition [25,26], text categorization [27,28], medical diagnosis [29], etc. However, there are still fewer experiments related to the diagnosis of industrial rolling bearing faults [20]; so, OSR based on rolling bearing faults is a worthwhile research direction. The decision boundary and the treatment of unknown samples for CSR and OSR are illustrated in Figure 1. The CSR method divides all the samples into the feature space of the known classes, and its decision boundary is unbounded, which leads to the possibility that the unknown samples may be incorrectly assigned, as shown in Figure 1b. In contrast, the OSR method defines a bounded region for each known class, and if a sample is not in the region of any known class, it is rejected as unknown.

2.2. Efficient Channel Attention (ECA)

This work decides to process the multi-scale features retrieved by SMCPL using the Efficient Channel Attention (ECA) mechanism [30] in order to focus more on the fault-related portion of the information among the features.
Figure 2 displays the ECA module’s schematic diagram. First, global average pooling (GAP) is used to obtain each channel’s characteristics. Then, a fast 1D convolution with a convolution kernel size of k generates the channel weights and performs local channel interaction. Given the channel dimension C , the convolution kernel size k can be adaptively determined as (1), where the operation x o d d represents the nearest odd integer to x .
k = ψ ( C ) = log 2 C γ + b γ odd
γ is a scaling factor that controls the weight assignment to each channel in the convolution operation. b is a bias term that is used to adjust the results of the convolution operation. In order to simplify the experimental setup, avoid adding unnecessary model complexity, and make the comparison of experimental results more concise, γ and b = 1 are set in the experiment. Finally, the sigmoid activation function is applied to generate the channel weights, which are then multiplied with the original feature map to reweight the features.

2.3. Serial Multi-Scale Convolutional Prototype Learning (SMCPL)

The serial multi-scale approach has lower memory and computational resource requirements compared to parallel multi-scale. Its step-by-step processing makes the model easier to control and debug, avoids the synchronization and management overheads in parallel computing, and provides more stable results when dealing with complex problems. Therefore, this paper chooses the serial multi-scale method to extract information. For an input sample x i , the features extracted by the φ layer of Serial Multi-scale Convolutional Prototype Learning (SMCPL) are given as follows:
f φ ( x i ) = ω φ g θ ( x i ) + b φ
where ω φ and b φ denote the weights and biases of the neurons, respectively, g θ ( x ) refers to the feature extractor that maps the raw data into the feature space. The SMCPL framework extracts multi-scale information from the data through a serial convolutional prototype learning network, which is then fused using the ECA mechanism. The final fused feature f x i is obtained by stacking along the channel dimension. This approach, in comparison to parallel network architectures, reduces computational resource consumption. The equation is as follows:
f ( x i ) = f ˜ a ( x i ) + f ˜ b ( x i ) + f ˜ c ( x i )
where 0 < a < b < c N , with N representing the total number of layers in the network. During training, a class prototype C k is assigned to each known class. C k is derived by computing the mean of the feature vectors for each class in the training set:
C k = 1 T k ( x i , y i ) T k f ( x i )
where f x i denotes the feature of the sample extracted by the feature extraction network, and ( x i , y i ) represents the feature vector and label of the sample. | T k | refers to the number of samples in the training set T k , and k indicates the number of classes. The probability that a class sample x i belongs to a given class in SMCPL-ECA is determined by the Euclidean distance between the sample’s mapped feature and the class prototype, as shown in (5). Here, d ( · ) represents the Euclidean distance, with its calculation provided in (6).
p ( y = k x ) = exp ( d ( f ( x i ) , C k ) ) k = 1 K exp ( d ( f ( x i ) , C k ) )
d ( f ( x i ) , C k ) = f ( x i ) C k 2 2
Based on the probability p ( y = k | x ) , the loss function of the network is defined as the distance cross-entropy (DCE) function, as shown below:
l D C E = log p ( y = k )
To address the potential overfitting issue that may arise when minimizing the DCE loss during training, SMCPL-ECA incorporates the Euclidean distance between the sample and its corresponding class prototype, referred to as the prototype loss l P L , as a regularization term, as described in (8). The final loss function for SMCPL-ECA is defined as (9), where λ is the hyperparameter that controls the weight of the prototype loss [13]. By minimizing the objective loss function l S M C P L E C A , the network ultimately reduces the intra-class feature distance while maximizing the inter-class feature distance. In the experiments, λ was set to 0.001 based on the analysis of the hyperparameters in Figure 3.
l P L = 1 T k i = 1 T k g θ ( x i ) C k 2 2
l S M C P L E C A = l D C E + λ l P L

2.4. Density Peak Clustering (DPC)

The Density Peak Clustering (DPC) algorithm was proposed by Rodriguez and Laio in 2014 [31]. Its core idea is to generate a decision graph with local density ( ρ ) on the x-axis and relative distance ( δ ) on the y-axis. The clustering centers are found through the decision map, and then the rest of the data are assigned to the closest clustering centers, respectively. The formulas for calculating local density ρ and relative distance δ are given in (10) and (12), respectively. For the data point x i with the maximum local density, the formula for its δ is provided in (13).
ρ i = j = 1 , j i n e ( d i j d c ) 2
d i j = x i x j 2
δ i = min j , ρ j > ρ i ( d i j )
δ i = max j ( d i j )
where d i j denotes the distance between the i -th and j -th data points, and d c represents the truncation distance. d c is used to limit the range of the neighborhood considered when calculating the local density. The value of d c will be adjusted according to the specific distribution of the dataset in the experiments of this paper. Based on the distribution characteristics of the distance between data points within each class, the 98% quantile of its distance value is selected as the specific parameter of d c . The input extracted sample features that will each be aggregated after DPC clustering, and the unknown class data features are necessarily different from the known class data features; clustering will become one or more outlier points or outlier clusters as a way to distinguish between the known class and the unknown class.

3. Methodology

3.1. The OSR Fault Diagnosis Framework Based on SMCPL-ECA-DPC

In this section, the proposed rolling bearing OSR fault diagnosis framework is presented. As shown in Figure 4, it consists of three modules: data preprocessing, SMCPL-ECA feature extraction, and DPC fault diagnosis.
(1) The data-preprocessing module will Z-score-normalize the raw acceleration and vibration signal data with a segment length of 1024 to eliminate the negative impact of the scale differences in the monitored variables on the model training, as shown in (14). In addition, the one-dimensional vibration signals were converted into two-dimensional images using an image conversion method [32]. The normalized 1D vectors are reconstructed into 32 × 32 2D feature maps as input.
Z = X μ σ
where Z represents the normalized value, X is the original data, μ denotes the mean of the original data, and σ is the standard deviation of the original data.
(2) The framework’s vital component is the SMCPL-ECA feature extraction module, which uses the data-preprocessing module’s normalized data as input and applies the multi-scale convolutional pooling operation of ECA during training to produce the known fault feature representation. For testing, mixed samples of known and unknown faults are fed into the SMCPL-ECA network architecture to obtain the extracted features of both, which are then fed into the DPC fault diagnosis module.
(3) During training, the DPC fault diagnosis module clusters the features to obtain the maximum relative distance m a x ( δ i ) of the known classes. λ D P C is a threshold value in the range of (0, 1], which is used to control the proportion of training samples misclassified by DPC. In the testing phase, DPC first calculates the relative distance δ x i of the test sample x i to the center of the clustering class of the known classes and then determines whether δ x i is in the range of the known classes so as to assign the sample, as shown in (15). If it falls within the known class range, the sample is classified as a known fault; otherwise, it is identified as an unknown fault. After extensive experiments to optimize and facilitate an intuitive comparison between methods, λ D P C was set to 1.
i f δ x i λ D P C × max ( δ i ) , k n o w n _ c l a s s i f δ x i > λ D P C × max ( δ i ) , u n k n o w n _ c l a s s

3.2. Evaluation Metric

In the performance evaluation of the model, both CSR and OSR performance should be considered comprehensively. This study employs three evaluation metrics: Normalized Accuracy (NA), Normalized F1 Score (NF1), and Youden’s Index (J) [12].
Normalized Accuracy (NA): A u is defined as the proportion of correctly categorized unknown samples to the total number of unknown samples, as expressed in (16). A k is defined as the proportion of correctly categorized known samples to the total number of known samples, as shown in (17). NA provides a comprehensive evaluation of both Close-Set and Open-Set Recognition performance, calculated as a weighted sum of A k and A u , as shown in (18).
A u = T u T u + F u
A k = i = 1 N ( T P i + T N i ) i = 1 N ( T P i + F P i + T N i + F N i )
N A = α A k + ( 1 α ) A u
In this case, TP, TN, FP, and FN stand for the true positives, true negatives, false positives, and false negatives of known samples, respectively. α is the weighting coefficient for A k and A u , with a range of (0, 1). Considering that the closed-set classification ability and the unknown class detection ability in the Open-Set Recognition task are of equal importance, this paper experimentally selects α = 0.5 to balance the contribution of both in the comprehensive evaluation.
Normalized F1 Score (NF1): The precision (P) and recall (R) harmonic means are called NF1. Unknown classes are treated as an extra simple class when used to assess the OSR task. It can be calculated using (21).
P = 1 N i = 1 N T P i T P i + F P i
R = 1 N i = 1 N T P i T P i + F N i
N F 1 = 2 P × R P + R
Better categorization performance is indicated by higher NF1 values, which range from 0 to 1. Notably, NF1 is unaffected by changes in TN, although TN remains an important factor influencing OSR performance. To address this, Scherreik and Rigling proposed the use of the Youden’s index (J) for a more comprehensive evaluation of the unknown class, which is calculated from specificity (S) and recall (R) [33]. The definitions of S and J are given in (22) and (23), respectively. J denotes the ability of the algorithm to avoid failure, bounded by (−1, 1), with a higher value indicating a higher resistance to failure.
S = T N F P + T N
J = S + R 1

4. Experiment

4.1. CWRU Experimental Data and Task Description

The CWRU experimental setup [34] consists of a 1.5 KW (2HP) motor, a torque transducer and decoder, and a power meter, as shown in Figure 5. Due to the extremely large amount of data in the CWRU dataset, after consideration, this paper selects the acceleration dataset with a sampling frequency of 12 kHz, a load of 1 HP, and a drive end (DE), in which three damage diameters of the inner ring failure data, the outer ring failure data, and the rolling element failure data are selected, making a total of nine types of failure data and one type of normal data. Compared with the types of failures that may occur in actual rolling bearings, the 10 types selected in this paper, although limited in number, are sufficient to verify the feasibility and effectiveness of the OSR framework based on SMCPL-ECA-DPC proposed in this paper.
Through overlapping sampling, this paper sets the number of samples for each class to 1000, with each sample presenting a length of 1024. For known classes, when partitioning the dataset, this paper divides it into a training set and a testing set in a 7/3 ratio, consisting of 700 training samples and 300 testing samples, respectively. For unknown classes, only 300 samples are selected to participate in testing alongside known class test samples, without involving them in the training phase. A detailed description of the experimental dataset is provided in Table 1.
Because motors are always in normal operation in real life, normal operating condition 9 is set as the known class when setting up the OSR tasks. In the OSR experiments with the same number of unknown classes, the faults similar to the known classes are targeted to be selected as the unknown classes on the premise that the known classes contain all fault types, and three tasks with different types of known and unknown faults are selected, and the number of unknown classes is 1~6, as shown in Table 2. For example, AE7 (with unknown classes B007(0), IR007(3), and OR007(6)) sets the class with a fault diameter of 7 mil as an unknown class, whereas AE16, AE17, and AE18 all set the class with the same fault diameter as a known class and set the classes with the other two fault diameters as unknown classes. The purpose of this setting is to deeply analyze the performance of the model under new conditions, i.e., its ability to identify unseen fault diameters, reflecting its adaptability and robustness.

4.2. XJTU-SY Experimental Data and Task Introduction

The XJTU-SY dataset is a bearing failure dataset provided by Xi’an Jiaotong University, which contains data on different failure modes, such as normal operation, failure, and wear, collected in real industrial environments, and is specially used for bearing life prediction and fault diagnosis [35]. The XJTU-SY test platform consists of an AC motor, motor speed controller, rotating shaft, support bearing, hydraulic loading system, and test bearing, as shown in Figure 6. After consideration, this paper chooses the following parameters for working condition 1: a rotational speed of 2100 r/min and a radial force of 12 kN, in which three types of failure types of data are selected, which are outer ring failure data, cage failure data, and inner ring outer ring failure data at the same time.
The sample parameters, as well as those of processing, are the same as those for the CWRU data-processing flow, but the difference is that the XJTU-SY data contain two-dimensional features representing the acceleration vibration signals in the horizontal and vertical directions, respectively; so, the normalized features will be reconstructed into a 32 × 64 two-dimensional feature map as input. A detailed description of the experimental dataset is shown in Table 3. The OSR experimental setup is shown in Table 4.

4.3. PU Experimental Data and Task Introduction

The PU dataset is a publicly available dataset from the University of Paderborn, Germany, constructed in 2016 to fuel research on the monitoring of bearing damage conditions in electromechanical drive systems based on motor current signals using a data-driven classification method [36]. The experimental bench shown in Figure 7 simulates a real industrial scenario, and the motor signals are susceptible to electromagnetic interference; so, the signals present a high-noise phenomenon. In this paper, the vibration signals of real damage data samples are selected for processing to verify the performance of the proposed algorithm in a complex environment. The information of the selected data samples is shown in Table 5, and the operating conditions of the bearing data samples are N09_M07_F10, i.e., the rotational speed is 900 rpm, the horsepower is 0.7 Nm, and the radial force is 1000 N. The OSR experimental settings are shown in Table 6.

4.4. Experimental Parameter Settings

The SMCPL-ECA-DPC model presented in this paper is implemented using the Python 3.9 programming language, with a framework built on PyTorch 2.1.0. The hyperparameter configurations are detailed in Table 7. This model selects three scale features for fusion; specifically, after obtaining three different scale features pooled by Max_pooling1, Max_pooling2, and Max_pooling3, they are input into the ECA for fusion, and the fused features are input into the Fully connected layer 1. The number of fully connected layers is set as 3. The dimension of the initialized prototype is set to the number of nodes in the output of fully connected layer 3, and this output is the data feature extracted by SMCPL-ECA, which is then input into DPC for clustering recognition. It is worth noting that since the number of samples in the training set is not the same for all tasks, the batch size will be set adaptively based on the experimental comparison of batch sizes in Figure 3. The batch size is chosen to be 64 when the number of training set samples is large and 32 when it is small. In addition, this paper employed the Adam optimizer and set the learning rate to 0.001 based on the results in Figure 3.
In this paper, CPL and SMCPL are selected for ablation experimental validation based on the proposed SMCPL-ECA-DPC model architecture, in addition to the classical CNN method, the OpenMax method, and the Adaptive Negative Evidential Deep Learning (ANEDL, 2023) [16] method for comparison experiments. The structure and parameters of the six models were optimized through extensive experiments to achieve the best performance of the models used.

5. Experimental Results and Analysis

This section presents the performance evaluation and analysis of the SMCPL-ECA-DPC framework, based on metrics such as NA, NF1, J, and feature visualization.

5.1. CWRU Experimental Results

The computational complexity of different models on the CWRU dataset is shown in Table 8. Due to the limitation of the experimental environment, this study conducts training and inference on the CPU platform; so, the average inference time for a single sample is relatively long. If GPU acceleration is used, the inference efficiency is expected to be significantly improved. Although the inference time of the SMCPL-ECA-DPC model is longer compared with simple CNN and CPL, its response time is still suitable for practical system monitoring and fault diagnosis. The sampling frequency of the CWRU dataset is 12 kHz, and each input sample contains 1024 data points, as described in this article. The sample generation time is calculated to be 1024/12,000 ≈ 85.3 ms. The inference time of the SMCPL-ECA-DPC model is 8.1 ms, which is much lower than the sample generation time of 85.3 ms. Therefore, the SMCPL-ECA-DPC model can be used for real-time monitoring of bearing health.
Tests were conducted according to the OSR experiment serial number in Table 2, and the results of the evaluation indexes obtained from the experiments are shown in Table 9. According to the three key performance indicators listed, the SMCPL-ECA-DPC model proposed in this paper exhibits either optimal or suboptimal performance in terms of overall performance for all 10 types of data jointly involved in the OSR experiments. Figure 8 shows the full closed-set fault diagnosis effect of SMCPL-ECA-DPC with an accuracy of 99.57% on the test set, which indicates that the model proposed in this paper is extremely effective for the extraction of critical fault signals and can provide accurate and reliable diagnosis results.
In the specific test scenario of AE2, the SMCPL-ECA-DPC model achieves a 100% recognition rate of unknown classes, which fully proves its excellent ability in feature extraction and classification tasks. For example, AE12 sets B007(0), B021(2), IR021(5), and OR021(8) as unknown classes. In this task, the proposed model is 2.71%, 7.96%, 6.81%, 13.14%, and 7.7% higher than SMCPL, CPL, ANEDL, CNN, and OpenMax on Na, respectively; on NF1, it is 0.0224, 0.068, 0.0592, 0.1048, and 0.093 higher than SMCPL, CPL, ANEDL, CNN, and OpenMax, respectively; on J, it is 0.0253, 0.0747, 0.0762, 0.1225, and 0.1031 higher than SMCPL, CPL, ANEDL, CNN, and OpenMax, respectively. In order to observe and compare the advantages and disadvantages of the six models more intuitively, the output of Fully connected layer 3 (i.e., the features extracted by the framework) is reduced from 256 dimensions to 2 dimensions for display. Figure 9 shows the spatial distribution of features extracted by AE12 using different models, and the t-distributed Stochastic Neighbor Embedding (t-SNE) algorithm is selected for the visualization algorithm [37].
Figure 9d shows that the CNN extracts the known fault feature B014(1) with a low degree of data dispersion and aggregation, whereas the other models for this class of data extraction still have a small number of data points outside the plotted intra-class range, but it can be clearly observed that the data points of the proposed model in this paper for the extraction of B014(1) are all centrally aggregated into a complete, continuous cluster of points with a complete distribution within the intra-class range, which is also significantly reduced according to the intra-class distance visualization in Figure 10. Intra-class distances are calculated for each class by calculating the Euclidean distance from each sample in the class to the class center and then averaging them, with the input being the Z-score-normalized features and thus dimensionless; inter-class distances are the Euclidean distances between different class centers. The number of interclass distances is extremely high, so it is not convenient to display them. Based on the intra-class distance visualization results and statistical metrics observed in Figure 10 and Table 10, SMCPL-ECA-DPC has the lowest mean (9.89), which is smaller than the comparison model’s, indicating that this model has the smallest intra-class distance and the most compact features. It is reduced by 0.64, 1.37, 0.37, 3.62, and 1.28 compared to SMCPL, CPL, ANEDL, CNN, and OpenMax, respectively. In addition to this, SMCPL-ECA-DPC has the smallest coefficient of variation (CV = 0.127), which indicates that the intra-class distances are the least fluctuating and are more robust.
After introducing the unknown class to participate in the test, by observing Figure 9d–f, it is found that among the known class and unknown class features extracted using CNN, OpenMax, and ANEDL, B007(0) and B021(2) as the unknown classes have a lower recognition rate of the unknown class because they are not involved in the training, and some of their features are distributed in the B014(1) space. For this problem, SMCPL-ECA-DPC greatly improved the separation of unknown and known classes, and the recognition performance is also substantially improved, as shown in Figure 9a. The corresponding normalized confusion matrix for AE12 is shown in Figure 11, in which the unknown class fault is labeled as −1. In the normalized confusion matrix, the rows and columns denote the true labels and predicted labels, respectively, and the numerical values in the matrix represent the correct predicted label samples’ number to the total number of true label samples. The confusion matrix classification results also confirm the superiority of the proposed model in OSR experiments.
Anomaly analysis: In some experiments, such as AE2, AE5, and AE6, the diagnostic performance of CPL is lower than that of CNN, specifically because the loss function used by CPL is different from that of CNN, which makes the same class of data closer, and because of this, misclassification may exist. Because the original data belonging to an unknown class and a known class feature distribution are too close, the training instead of the inter-class distance between the two will be so close that it cannot be separated and recognized. This is precisely why the CNN achieves optimal recognition performance in AE7, as our experimental results show that B007(0), serving as an unknown class, has feature distributions extremely similar to those of B021(2), which is also a rolling element fault, making it indistinguishable after intra-class distance reduction.

5.2. XJTU-SY Experimental Results

Comparison tests are conducted according to the OSR experiment serial numbers in Table 4, and the results of the evaluation metrics on the XJTU-SY dataset are obtained as shown in Table 11. Figure 12 shows the full closed-set fault diagnosis results of SMCPL-ECA-DPC on this dataset with 100% test set accuracy.
Table 11 demonstrates the OSR strong diagnostic performance of SMCPL-ECA-DPC on the XJTU-SY dataset, and the classification performances are all higher than those of other models. Taking BE3 (the unknown class is IORF(2)) as an example, the proposed framework in this paper achieves a 100 percent recognition rate of the unknown class with both CPL and SMCPL, and the computed NA, NF1, and J metrics are improved by 1.35%, 0.0126, and 0.0037, respectively, compared to ANEDL; by 25.67%, 0.1832, and 0.2622, respectively, compared to CNN; and by 21.42%, 0.2144, and 0.2834, respectively, compared to OpenMax. Figure 13 demonstrates the distribution of features extracted by experiment BE3 using the model of this paper. According to the intra-class distance visualization results and the statistical metrics observed in Figure 14 and Table 12, SMCPL-ECA-DPC has the lowest mean value (14.01), which is 0.98, 1.89, 0.76, 2.56, and 1.74 lower than that of SMCPL, CPL, ANEDL, CNN, and OpenMax, respectively, which indicates that the present model is better than other models for the aggregation of intra-class distances. Figure 15 represents the normalized confusion matrix of SMCPL-ECA-DPC.
Anomaly analysis: The reason for the low diagnostic performance of all methods in the BE2, BE5, and BE6 experimental results is that the distributions of ORF(0) and CF(1) in the XJTU-SY dataset are extremely similar in the feature space, which would be difficult to differentiate if both of them were not set up as known classes to be trained to increase the inter-class distance. If any one of the two classes is an unknown class, then the original part of the data distribution belonging to the unknown class during testing will be distributed in the same space with the distribution of the other class so that it cannot be separated and identified. The experiment where both are trained as known classes is BE3, analyzed as above. In experiment BE4, both ORF(0) and CF(1) are used as unknown classes, and the recognition performance of all models is very good precisely because the spatial distributions of these two classes are similar, which are very different from those of IORF(2), and thus the recognition rate of unknown classes reaches 100%.

5.3. PU Experimental Results

Comparison tests are conducted according to the OSR experiment serial numbers in Table 6, and the results of the evaluation metrics on the Paderborn dataset are obtained as shown in Table 13, where the classification performance of SMCPL-ECA-DPC is higher than that of other models in all cases, which once again verifies its strong diagnostic capability for OSR on the real bearing dataset PU, which is a much larger dataset with a much higher noise level. Figure 16 shows the full closed-set fault diagnosis effect of SMCPL-ECA-DPC on this dataset, with an accuracy of 99.33% on the test set. Taking CE2 (unknown class as IR (1)) as an example, the NA, NF1, and J metrics computed by the proposed framework in this paper are higher than those of other models. Figure 17 demonstrates the feature distribution of experiment CE2 extracted using the model of this paper. According to the intra-class distance visualization results and statistical metrics observed in Figure 18 and Table 14, SMCPL-ECA-DPC has the lowest mean value (14.78), which is 0.365, 1.035, 0.13, 1.385, and 0.955 lower than that of SMCPL, CPL, ANEDL, CNN, and OpenMax, respectively, which once again verifies that the present model is more effective for intra-class distance approximation. Figure 19 represents the normalized confusion matrix of SMCPL-ECA-DPC.
Overall, according to the analysis of the experimental results of the three rolling bearing datasets, the framework based on SMCPL-ECA-DPC proposed in this paper can effectively improve the separation of known and unknown classes in the field of rolling bearings, and the diagnostic performance of the unknown classes is better than that of the comparison method, which provides support for the research of OSR.

6. Conclusions

To solve the problem of identifying unknown faults in the field of rolling bearings, this paper proposes an OSR fault diagnosis framework based on SMCPL-ECA-DPC. The experiments on the CWRU dataset take the classes of different fault diameters into account, and comparative experiments are also carried out on the XJTU-SY dataset and the PU dataset, and the results show that the proposed framework is higher than the other models’ in NA, NF1, and J on all the tasks, which proves that its diagnostic performance is superior to that of CPL, SMCPL, ANEDL, CNN, and OpenMax. Compared to the cross-entropy loss function (CE) used in CNN and the OpenMax layer, the use of distance cross-entropy (DCE) and prototype loss (PL) enhances the separation between known and unknown classes and outperforms the adaptive negative optimization evidence learning of ANEDL. In addition, the performance of the network on feature extraction is further strengthened by fusing multi-scale information with the introduction of an efficient channel attention mechanism, which is why this framework is able to efficiently capture the diversity and complexity of fault features and thus efficiently identify the unknown classes.
Future research will be devoted to the in-depth validation of the proposed framework on richer and more diverse rolling bearing fault datasets while continuing to optimize the structural design and hyperparameter configuration of the framework.

Author Contributions

Conceptualization, J.X. and X.Z.; methodology, J.X.; formal analysis, R.X.; investigation, H.W.; resources, Y.W.; writing—original draft preparation, J.X.; writing—review and editing, X.Z.; visualization, H.W.; supervision, X.Z.; funding acquisition, Y.W. and R.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Nuclear Grade Instrumentation Project Grant 2024 under Grant TC240AAKP, and the China Nuclear Power Research and Design Institute Joint Fund under Grant SCU&NPIC-LHCX-20.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are publicly available.

Acknowledgments

The authors wish to acknowledge China Nuclear Power Research and Design Institute.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Randall, R.B.; Antoni, J. Rolling element bearing diagnostics—A tutorial. Mech. Syst. Signal Process. 2011, 25, 485–520. [Google Scholar] [CrossRef]
  2. Chen, Z.; Li, W. Multisensor Feature Fusion for Bearing Fault Diagnosis Using Sparse Autoencoder and Deep Belief Network. IEEE Trans. Instrum. Meas. 2017, 66, 1693–1702. [Google Scholar] [CrossRef]
  3. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  4. Tian, H.; Fan, H.; Feng, M.; Cao, R.; Li, D. Fault Diagnosis of Rolling Bearing Based on HPSO Algorithm Optimized CNN-LSTM Neural Network. Sensors 2023, 23, 6508. [Google Scholar] [CrossRef] [PubMed]
  5. Xu, Z.; Li, C.; Yang, Y. Fault diagnosis of rolling bearings using an Improved Multi-Scale Convolutional Neural Network with Feature Attention mechanism. ISA Trans. 2021, 110, 379–393. [Google Scholar] [CrossRef] [PubMed]
  6. Zhou, Z.; Ai, Q.; Lou, P.; Hu, J.; Yan, J. A Novel Method for Rolling Bearing Fault Diagnosis Based on Gramian Angular Field and CNN-ViT. Sensors 2024, 24, 3967. [Google Scholar] [CrossRef] [PubMed]
  7. Shao, H.; Xia, M.; Han, G.; Zhang, Y.; Wan, J. Intelligent Fault Diagnosis of Rotor-Bearing System Under Varying Working Conditions with Modified Transfer Convolutional Neural Network and Thermal Images. IEEE Trans. Ind. Inform. 2021, 17, 3488–3496. [Google Scholar] [CrossRef]
  8. Ruan, D.; Wang, J.; Yan, J.; Guhmann, C. CNN parameter design based on fault signal analysis and its application in bearing fault diagnosis. Adv. Eng. Inform. 2023, 55, 101877. [Google Scholar] [CrossRef]
  9. O’Higgins, C.; Hester, D.; McGetrick, P.; Cross, E.J.; Ao, W.K.; Brownjohn, J. Minimal Information Data-Modelling (MID) and an Easily Implementable Low-Cost SHM System for Use on a Short-Span Bridge. Sensors 2023, 23, 6328. [Google Scholar] [CrossRef]
  10. Ao, W.K.; Hester, D.; O’Higgins, C.; Brownjohn, J. Tracking long-term modal behaviour of a footbridge and identifying potential SHM approaches. J. Civ. Struct. Health Monit. 2024, 14, 1311–1337. [Google Scholar] [CrossRef]
  11. O’Higgins, C.; Hester, D.; Ao, W.K.; McGetrick, P. A method to maximise the information obtained from low signal-to-noise acceleration data by optimising SSI-COV input parameters. J. Sound. Vib. 2024, 571, 118101. [Google Scholar] [CrossRef]
  12. Geng, C.; Huang, S.-J.; Chen, S. Recent Advances in Open Set Recognition: A Survey. ieee Trans. Pattern Anal. Mach. Intell. 2021, 43, 3614–3631. [Google Scholar] [CrossRef] [PubMed]
  13. Bendale, A.; Boult, T.E. Towards open set deep networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 1563–1572. [Google Scholar]
  14. Rudd, E.M.; Jain, L.P.; Scheirer, W.J.; Boult, T.E. The Extreme Value Machine. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 762–768. [Google Scholar] [CrossRef]
  15. Sensoy, M.; Kaplan, L.; Kandemir, M. Evidential Deep Learning to Quantify Classification Uncertainty. arXiv 2018, arXiv:1806.01768. [Google Scholar] [CrossRef]
  16. Yu, Y.; Deng, D.; Liu, F.; Jin, Y.; Dou, Q.; Chen, G.; Heng, P.A. Adaptive Negative Evidential Deep Learning for Open-set Semi-supervised Learning. arXiv 2023, arXiv:2303.12091. [Google Scholar] [CrossRef]
  17. Cai, W.; Zhao, D.; Wang, T. Multi-scale dynamic graph mutual information network for planet bearing health monitoring under imbalanced data. Adv. Eng. Inform. 2025, 64, 103096. [Google Scholar] [CrossRef]
  18. Zhao, D.; Cai, W.; Cui, L. Multi-perception graph convolutional tree-embedded network for aero-engine bearing health monitoring with unbalanced data. Reliab. Eng. Syst. Saf. 2025, 257, 110888. [Google Scholar] [CrossRef]
  19. Yang, H.-M.; Zhang, X.-Y.; Yin, F.; Liu, C.-L. Robust Classification with Convolutional Prototype Learning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  20. Sun, H.; Yang, B.; Lin, S. An Open Set Diagnosis Method for Rolling Bearing Faults Based on Prototype and Reconstructed Integrated Network. IEEE Trans. Instrum. Meas. 2023, 72, 3502010. [Google Scholar] [CrossRef]
  21. Liu, W.; Nie, X.; Zhang, B.; Sun, X. Incremental Learning with Open-Set Recognition for Remote Sensing Image Scene Classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5622916. [Google Scholar] [CrossRef]
  22. Li, J.; Lin, M.; Wang, B.; Tian, R.; Tan, S.; Li, Y.; Chen, J. Open set recognition fault diagnosis framework based on convolutional prototype learning network for nuclear power plants. Energy 2024, 290, 130101. [Google Scholar] [CrossRef]
  23. Kumar, A.; Kumar, A. Adaptive Security for Human Surveillance Using Multimodal Open Set Biometric Recognition. In Proceedings of the 2014 22nd International Conference on Pattern Recognition (ICPR), Stockholm, Sweden, 24–28 August 2014. [Google Scholar]
  24. Hassen, M.; Chan, P.K. Learning a Neural-network-based Representation for Open Set Recognition. In Proceedings of the 2020 SIAM International Conference on Data Mining (SDM), Cincinnati, OH, USA, 7–9 May 2020. [Google Scholar]
  25. Han, Z.; Fang, C.; Ding, X. Discriminative prototype learning in open set face recognition. In Proceedings of the 2010 20th International Conference on Pattern Recognition (ICPR 2010), Istanbul, Turkey, 23–26 August 2010; pp. 2696–2699. [Google Scholar]
  26. Moeini, A.; Faez, K.; Moeini, H.; Safai, A.M. Open-set face recognition across look-alike faces in real-world scenarios. Image Vis. Comput. 2017, 57, 1–14. [Google Scholar] [CrossRef]
  27. Prakhya, S.; Venkataram, V.; Kalita, J. Open Set Text Classification Using CNNs. In Proceedings of the 4th International Conference on Natural Language Processing (ICON-2017), Kolkata, India, 18–21 December 2017. [Google Scholar]
  28. Kim, D.; Koo, J.; Kim, U.M. OSP-Class: Open Set Pseudo-labeling with Noise Robust Training for Text Classification. In Proceedings of the 2022 IEEE International Conference on Big Data (Big Data), Osaka, Japan, 17–20 December 2022; pp. 5520–5529. [Google Scholar]
  29. Liu, M.; Xu, L.; Zhang, J. learning large margin sparse embeddings for open set medical diagnosis. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2023. [Google Scholar]
  30. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR 2020), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  31. Rodriguez, A.; Laio, A. Clustering by fast search and find of density peaks. Science 2014, 344, 1492–1496. [Google Scholar] [CrossRef]
  32. Zhou, H.; Chen, W.; Cheng, L.; Liu, J.; Xia, M. Trustworthy Fault Diagnosis with Uncertainty Estimation Through Evidential Convolutional Neural Networks. IEEE Trans. Ind. Inform. 2023, 19, 10842–10852. [Google Scholar] [CrossRef]
  33. Scherreik, M.D.; Rigling, B.D. Open Set Recognition for Automatic Target Classification with Rejection. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 632–642. [Google Scholar] [CrossRef]
  34. Loparo, K. Case Western Reserve University Bearing Data Centre Website. Available online: https://engineering.case.edu/bearingdatacenter (accessed on 20 September 2017).
  35. Wang, B.; Lei, Y.; Li, N.; Li, N. A Hybrid Prognostics Approach for Estimating Remaining Useful Life of Rolling Element Bearings. IEEE Trans. Reliab. 2018, 69, 401–412. [Google Scholar] [CrossRef]
  36. Lessmeier, C.; Kimotho, J.K.; Zimmer, D.; Sextro, W. Condition Monitoring of Bearing Damage in Electromechanical Drive Systems by Using Motor Current Signals of Electric Motors: A Benchmark Data Set for Data-Driven Classification. In Proceedings of the European Conference of the PHM Society 2016, Bilbao, Spain, 5–8 July 2016. [Google Scholar] [CrossRef]
  37. van der Maaten, L.; Hinton, G. Visualizing Data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. The differences between CSR and OSR: (a) distribution of the known class space, (b) CSR classification, and (c) OSR classification.
Figure 1. The differences between CSR and OSR: (a) distribution of the known class space, (b) CSR classification, and (c) OSR classification.
Sensors 25 03019 g001
Figure 2. Schematic diagram of the ECA module.
Figure 2. Schematic diagram of the ECA module.
Sensors 25 03019 g002
Figure 3. Hyperparametric sensitivity analysis. (a) Impact of batch size on NA. (b) Impact of learning rate on NA. (c) Impact of prototype loss weights on NA.
Figure 3. Hyperparametric sensitivity analysis. (a) Impact of batch size on NA. (b) Impact of learning rate on NA. (c) Impact of prototype loss weights on NA.
Sensors 25 03019 g003
Figure 4. The OSR fault diagnosis framework based on SMCPL-ECA-DPC.
Figure 4. The OSR fault diagnosis framework based on SMCPL-ECA-DPC.
Sensors 25 03019 g004
Figure 5. CWRU experimental equipment.
Figure 5. CWRU experimental equipment.
Sensors 25 03019 g005
Figure 6. XJTU-SY experimental equipment.
Figure 6. XJTU-SY experimental equipment.
Sensors 25 03019 g006
Figure 7. PU experimental equipment.
Figure 7. PU experimental equipment.
Sensors 25 03019 g007
Figure 8. Confusion matrix for fully closed-set experiments with the SMCPL-ECA-DPC model on the CWRU dataset.
Figure 8. Confusion matrix for fully closed-set experiments with the SMCPL-ECA-DPC model on the CWRU dataset.
Sensors 25 03019 g008
Figure 9. T-SNE visualization for features from AE12.
Figure 9. T-SNE visualization for features from AE12.
Sensors 25 03019 g009
Figure 10. Known intra-class distances for experiment AE12.
Figure 10. Known intra-class distances for experiment AE12.
Sensors 25 03019 g010
Figure 11. Normalized confusion matrix for experiment AE12.
Figure 11. Normalized confusion matrix for experiment AE12.
Sensors 25 03019 g011
Figure 12. Confusion matrix for fully closed-set experiments with the SMCPL-ECA-DPC model on the XJTU-SY dataset.
Figure 12. Confusion matrix for fully closed-set experiments with the SMCPL-ECA-DPC model on the XJTU-SY dataset.
Sensors 25 03019 g012
Figure 13. Distribution of BE3 features extracted by SMCPL-ECA-DPC. (a) Known classes’ distribution space for BE3. (b) Mixed classes’ distribution space for BE3.
Figure 13. Distribution of BE3 features extracted by SMCPL-ECA-DPC. (a) Known classes’ distribution space for BE3. (b) Mixed classes’ distribution space for BE3.
Sensors 25 03019 g013
Figure 14. Known intra-class distances for experiment BE3.
Figure 14. Known intra-class distances for experiment BE3.
Sensors 25 03019 g014
Figure 15. Normalized confusion matrix for SMCPL-ECA-DPC on the XJTU-SY dataset.
Figure 15. Normalized confusion matrix for SMCPL-ECA-DPC on the XJTU-SY dataset.
Sensors 25 03019 g015
Figure 16. Confusion matrix for fully closed-set experiments with the SMCPL-ECA-DPC model on the PU dataset.
Figure 16. Confusion matrix for fully closed-set experiments with the SMCPL-ECA-DPC model on the PU dataset.
Sensors 25 03019 g016
Figure 17. Distribution of CE2 features extracted by SMCPL-ECA-DPC. (a) Known classes’ distribution space for CE2. (b) Mixed classes’ distribution space for CE2.
Figure 17. Distribution of CE2 features extracted by SMCPL-ECA-DPC. (a) Known classes’ distribution space for CE2. (b) Mixed classes’ distribution space for CE2.
Sensors 25 03019 g017
Figure 18. Known intra-class distances for experiment CE2.
Figure 18. Known intra-class distances for experiment CE2.
Sensors 25 03019 g018
Figure 19. Normalized confusion matrix for SMCPL-ECA-DPC on the PU dataset.
Figure 19. Normalized confusion matrix for SMCPL-ECA-DPC on the PU dataset.
Sensors 25 03019 g019
Table 1. Description of the CWRU bearing dataset.
Table 1. Description of the CWRU bearing dataset.
Fault DiameterHealth StatusAbbreviationNumberNumber of Samples
7 milBall faultB0070700/300
14 milBall faultB0141700/300
21 milBall faultB0212700/300
7 milInner race faultIR0073700/300
14 milInner race faultIR0144700/300
21 milInner race faultIR0215700/300
7 milOuter race faultOR0076700/300
14 milOuter race faultOR0147700/300
21 milOuter race fault0R0218700/300
-NormalN9700/300
Table 2. OSR experimental setup of CWRU.
Table 2. OSR experimental setup of CWRU.
Number of Unknown ClassesTaskKnown ClassUnknown Class
1AE10, 1, 3, 4, 5, 6, 7, 8, 92
AE20, 1, 2, 3, 5, 6, 7, 8, 94
AE30, 1, 2, 3, 4, 5, 7, 8, 96
2AE41, 3, 4, 5, 6, 7, 8, 90, 2
AE50, 1, 2, 5, 6, 7, 8, 93, 4
AE69, 0, 1, 2, 3, 4, 5, 67, 8
3AE71, 2, 4, 5, 7, 8, 90, 3, 6
AE80, 2, 3, 5, 6, 8, 91, 4, 7
AE90, 1, 3, 4, 6, 7, 92, 5, 8
4AE101, 4, 5, 7, 8, 90, 2, 3, 6
AE110, 2, 3, 5, 6, 91, 4, 7, 8
AE121, 3, 4, 6, 7, 90, 2, 5, 8
5AE130, 3, 5, 8, 91, 2, 4, 6, 7
AE140, 2, 4, 6, 91, 3, 5, 7, 8
AE151, 3, 6, 7, 90, 2, 4, 5, 8
6AE160, 3, 6, 91, 2, 4, 5, 7, 8
AE171, 4, 7, 90, 2, 3, 5, 6, 8
AE182, 5, 8, 90, 1, 3, 4, 6, 7
Table 3. XJTU-SY bearing dataset description.
Table 3. XJTU-SY bearing dataset description.
Bearing LifetimeFault ElementAbbreviationNumberNumber of Samples
2 h 3 minOuter raceORF0700/300
2 h 2 minCageCF1700/300
52 minInner race and
outer race
IORF2700/300
Table 4. OSR experimental setup of XJTU-SY.
Table 4. OSR experimental setup of XJTU-SY.
Number of Unknown ClassesTaskKnown ClassUnknown Class
1BE11, 20
BE20, 21
BE30, 12
2BE420, 1
BE510, 2
BE601, 2
Table 5. PU bearing dataset description.
Table 5. PU bearing dataset description.
Bearing CodeDamageFault ElementAbbreviationNumberNumber of Samples
K001-NormalN0700/300
KA04fatigue:pittingOuter race faultOR1700/300
KI04fatigue:pittingInner race faultIR2700/300
Table 6. OSR experimental setup of PU.
Table 6. OSR experimental setup of PU.
Number of Unknown ClassesTaskKnown ClassUnknown Class
1CE11, 20
CE20, 21
CE30, 12
2CE420, 1
CE510, 2
CE601, 2
Table 7. Hyperparameter settings for SMCPL-ECA-DPC.
Table 7. Hyperparameter settings for SMCPL-ECA-DPC.
No.Layer TypeKernel NumberKernel SizePaddingBatch
Normalization
ActivationAttentionOutput
0Input------(32, 32)
1Convolution-1-1323 × 3(2, 2)BNPReLU-(34, 34, 32)
2Convolution-1-2323 × 3(2, 2)BNPReLU-(36, 36, 32)
3Max-pooling-1-3 × 3---ECA(12, 12, 32)
4Convolution-2-1643 × 3(2, 2)BNPReLU-(14, 14, 64)
5Convolution-2-2643 × 3(2, 2)BNPReLU-(16, 16, 64)
6Max-pooling-2-3 × 3---ECA(5, 5, 64)
7Convolution-3-11283 × 3(2, 2)BNPReLU-(7, 7, 128)
8Convolution-3-21283 × 3(2, 2)BNPReLU-(9, 9, 128)
9Max-pooling-3-3 × 3---ECA(3, 3, 128)
10Fully connected layer 1---BNPReLU-2016
11Fully connected layer 2---BNPReLU-1024
12Fully connected layer 3----PReLU-256
13Classifier---- l S M C P L E C A --
14DPC-------
Table 8. Comparison of computational complexity of different models on CWRU dataset.
Table 8. Comparison of computational complexity of different models on CWRU dataset.
ModelFLOPsParametersAverage Inference Time per Sample
CPL41.66 M0.94 M6.58 ms
SMCPL43.83 M2.62 M8.21 ms
SMCPL-ECA-DPC(ours)43.83 M2.62 M8.1 ms
CNN41.66 M0.94 M6.05 ms
OpenMax41.66 M0.94 M10.77 ms
ANEDL44.42 M2.74 M12.96 ms
Table 9. NA, NF1, and J for all fault diameters in the CWRU dataset.
Table 9. NA, NF1, and J for all fault diameters in the CWRU dataset.
TaskIndicatorsSMCPL-ECA-DPC(Ours)SMCPLCPLANEDLCNNOpenMax
AE1NA(%)96.58880.8392.4580.369.67
NF10.980.95290.87140.94320.86740.9119
J0.99220.97330.95740.93460.95630.9089
AE2NA(%)10091.9369.7494.5890.7881.06
NF110.95110.77030.96420.9420.7675
J10.97780.93110.97650.96670.7041
AE3NA(%)99.8389.6982.3995.4395.0262.76
NF10.99910.93680.88420.96580.96920.781
J0.99960.97440.95690.98130.98040.7411
AE4NA(%)95.088071.0890.282.1776.12
NF10.96850.85560.76870.92470.87310.8108
J0.97790.910.86990.93610.91980.7829
AE5NA(%)99.8396.6373.0495.8495.0888.35
NF10.9990.9780.79140.96950.96720.8179
J0.99920.98310.85230.97620.97340.7937
AE6NA(%)97.6783.9251.5289.6373.2773.48
NF10.98530.88690.49180.93570.79820.8005
J0.98950.92760.77650.95830.85640.769
AE7NA(%)83.3979.6773.8282.0785.9675.03
NF10.87660.84080.78280.86380.89490.7928
J0.88610.86060.81980.87240.90130.7554
AE8NA(%)83.578.4979.7881.3665.4479.73
NF10.87750.83490.84610.8520.69950.7715
J0.88690.85140.85830.86430.75890.7478
AE9NA(%)99.7598.2884.6793.4689.2592.03
NF10.99820.98770.90530.95580.91910.9145
J0.99810.98820.89490.97630.9250.9055
AE10NA(%)95.9695.3370.4292.8189.4277.67
NF10.96810.96300.75810.94120.91680.7929
J0.96230.95640.72390.9170.90120.7542
AE11NA(%)84.2877.7664.4280.1869.7572.54
NF10.87690.83740.69540.84930.74670.7545
J0.85300.79230.66590.83690.71570.7138
AE12NA(%)99.2196.591.2592.486.1791.51
NF10.99350.97110.92550.93430.88870.9005
J0.99260.96730.91790.91640.87010.8895
AE13NA(%)97.9396.991.4396.1890.0788.67
NF10.98010.9730.920.97340.9050.8813
J0.97520.96280.89720.9680.88080.864
AE14NA(%)92.990.165.8791.6380.9769.27
NF10.9320.90620.64930.92570.83210.6909
J0.91480.88120.59040.90540.77160.6312
AE15NA(%)98.2796.694.7394.5683.0374.5
NF10.9830.96790.94750.93860.84770.7593
J0.97920.95920.93680.95240.79640.694
AE16NA(%)95.2893.588.4392.9485.6484.72
NF10.94420.92410.88730.92250.84940.8467
J0.92920.93860.87630.90420.83460.8373
AE17NA(%)92.4391.8687.3790.6984.3986.76
NF10.91340.91580.86730.91240.8380.8532
J0.93420.92060.87940.9350.85780.8649
AE18NA(%)89.5986.3982.8588.1680.4176.57
NF10.90530.8580.83170.87840.78630.7804
J0.91270.87910.82680.8970.7920.7329
Table 10. Statistical indicators of intra-class distances for AE12.
Table 10. Statistical indicators of intra-class distances for AE12.
MethodMean (μ)Standard Deviation (σ)MinimumMaximumCoefficient of Variation (CV = σ/μ)
SMCPL-ECA-DPC9.891.2548.2411.810.127
SMCPL10.531.4328.3812.810.136
CPL11.261.7459.0314.070.155
ANEDL10.261.6038.6113.340.156
CNN13.511.86711.2215.770.138
OpenMax11.171.6239.0114.080.145
Table 11. NA, NF1 and J on the XJTU-SY dataset.
Table 11. NA, NF1 and J on the XJTU-SY dataset.
TaskIndicatorsSMCPL-ECA-DPC(Ours)SMCPLCPLANEDLCNNOpenMax
BE1NA(%)92.6791.9491.1690.6890.1791.97
NF10.94340.94150.93760.93430.93380.942
J0.93140.92460.91240.9250.9010.9214
BE2NA(%)75.575.1973.7273.8772.6773.26
NF10.81860.80830.79460.7910.80310.814
J0.76490.75610.7620.74930.71980.7274
BE3NA(%)10010010098.6574.3378.58
NF11110.98740.81680.7856
J1110.99630.73780.7166
BE4NA(%)10010010010010099.18
NF1111110.9465
J111110.9837
BE5NA(%)75.9275.5375.1874.887572.16
NF10.67880.67450.66940.67160.66670.6439
J0.35780.35310.34570.34840.33330.4632
BE6NA(%)78.9277.9477.3677.8576.5871.43
NF10.71810.70610.69730.7140.68980.6783
J0.43780.41460.39670.42830.37560.4267
Table 12. Statistical indicators of intra-class distances for BE3.
Table 12. Statistical indicators of intra-class distances for BE3.
MethodMean (μ)Standard Deviation (σ)MinimumMaximumCoefficient of Variation (CV = σ/μ)
SMCPL-ECA-DPC14.011.81312.7315.290.1294
SMCPL14.990.29514.7815.200.0197
CPL15.771.24314.916.650.0778
ANEDL14.771.27413.8715.680.0863
CNN16.571.48815.5117.620.0898
OpenMax15.750.27915.5515.950.0177
Table 13. NA, NF1 and J on the PU dataset.
Table 13. NA, NF1 and J on the PU dataset.
TaskIndicatorsSMCPL-ECA-DPC(Ours)SMCPLCPLANEDLCNNOpenMax
BE1NA(%)90.8387.4385.8589.2877.580.36
NF10.92850.87860.85410.9040.80890.8065
J0.90880.8640.8540.89560.7750.792
BE2NA(%)92.6774.7551.2590.4649.3370.41
NF10.94250.78090.4510.92560.43510.6769
J0.92170.74670.5050.91830.46830.5283
BE3NA(%)93.8390.7188.3891.5382.6783.84
NF10.95260.92340.8920.93750.85820.847
J0.93830.9080.890.92130.82670.8364
BE4NA(%)89.9288.4684.5187.6878.8573.95
NF10.860.85230.8390.82610.7160.648
J0.73110.72540.7060.72620.63720.4333
BE5NA(%)88.3387.3578.460.868760.0864.57
NF10.83940.82470.73160.82180.44590.5436
J0.68890.6750.51620.6742−0.06440.4932
BE6NA(%)93.1791.7684.6892.186374.3
NF10.93020.90970.85620.91760.49370.7834
J0.81780.79540.7260.80420.01330.7684
Table 14. Statistical indicators of intra-class distances for CE2.
Table 14. Statistical indicators of intra-class distances for CE2.
MethodMean (μ)Standard Deviation (σ)MinimumMaximumCoefficient of Variation (CV = σ/μ)
SMCPL-ECA-DPC14.781.456613.7515.810.0986
SMCPL15.1451.106414.3615.930.0731
CPL15.8151.445914.7916.840.0914
ANEDL14.911.317413.9815.840.0884
CNN16.1650.459715.8416.490.0284
OpenMax15.7350.572715.3316.140.0364
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, J.; Wang, Y.; Xu, R.; Wang, H.; Zhou, X. Research on Open-Set Recognition Methods for Rolling Bearing Fault Diagnosis. Sensors 2025, 25, 3019. https://doi.org/10.3390/s25103019

AMA Style

Xu J, Wang Y, Xu R, Wang H, Zhou X. Research on Open-Set Recognition Methods for Rolling Bearing Fault Diagnosis. Sensors. 2025; 25(10):3019. https://doi.org/10.3390/s25103019

Chicago/Turabian Style

Xu, Jia, Yan Wang, Renyi Xu, Hailin Wang, and Xinzhi Zhou. 2025. "Research on Open-Set Recognition Methods for Rolling Bearing Fault Diagnosis" Sensors 25, no. 10: 3019. https://doi.org/10.3390/s25103019

APA Style

Xu, J., Wang, Y., Xu, R., Wang, H., & Zhou, X. (2025). Research on Open-Set Recognition Methods for Rolling Bearing Fault Diagnosis. Sensors, 25(10), 3019. https://doi.org/10.3390/s25103019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop