Next Article in Journal
Planetary Gearboxes Fault Diagnosis Based on Markov Transition Fields and SE-ResNet
Previous Article in Journal
Exploring rounD Dataset for Domain Generalization in Autonomous Vehicle Trajectory Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Calibrated Adaptive Teacher for Domain-Adaptive Intelligent Fault Diagnosis

Intelligent Maintenance and Operations Systems, École Polytechnique Fédérale de Lausanne, 1015 Lausanne, Switzerland
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(23), 7539; https://doi.org/10.3390/s24237539
Submission received: 14 October 2024 / Revised: 21 November 2024 / Accepted: 22 November 2024 / Published: 26 November 2024
(This article belongs to the Section Fault Diagnosis & Sensors)

Abstract

:
Intelligent fault diagnosis (IFD) based on deep learning can achieve high accuracy from raw condition monitoring signals. However, models usually perform well on the training distribution only, and experience severe performance drops when applied to a different distribution. This is also observed in fault diagnosis, where assets are often operated in working conditions different from the ones in which the labeled data have been collected. The scenario where labeled data are available in a source domain and only unlabeled data are available in a target domain has been addressed recently by unsupervised domain adaptation (UDA) approaches for IFD. Recent methods have relied on self-training with confident pseudo-labels for the unlabeled target samples. However, the confidence-based selection of pseudo-labels is hindered by poorly calibrated uncertainty estimates in the target domain, primarily due to over-confident predictions, which limits the quality of pseudo-labels and leads to error accumulation. In this paper, we propose a novel method called Calibrated Adaptive Teacher (CAT), where we propose to calibrate the predictions of the teacher network on target samples throughout the self-training process, leveraging post hoc calibration techniques. We evaluate CAT on domain-adaptive IFD and perform extensive experiments on the Paderborn University (PU) benchmark for fault diagnosis of rolling bearings under varying operating conditions, using both time- and frequency-domain inputs. We compare four different calibration techniques within our framework, where temperature scaling is both the most effective and lightweight one. The resulting method—CAT+TempScaling—achieves state-of-the-art performance on most transfer tasks, with on average 7.5% higher accuracy and 4 times lower calibration error compared to domain-adversarial neural networks (DANNs) across the twelve PU transfer tasks.

1. Introduction

Fault diagnosis of industrial equipment is a crucial task in prognostics and health management. Intelligent fault diagnosis (IFD) based on deep learning has proven to be an effective and flexible solution, attracting extensive research [1,2]. Deep neural networks are able to learn rich representations from extensive labeled data, allowing them to tackle various tasks across different applications. In IFD in particular, they can achieve high classification performance from sensor data such as time series or spectrograms in an end-to-end manner, without the need for incorporating extensive domain knowledge. However, deep learning models usually only perform well on the data distribution they have been trained on. When applied to a different distribution, they may experience a severe performance drop. This is also observed in fault diagnosis, where the industrial assets are often operated in working conditions different from those in which the labeled data have been collected. Furthermore, obtaining labeled data is difficult and costly in real-world industrial settings. This challenge has recently been addressed recently by deep transfer learning (DTL) and domain adaptation (DA) approaches for IFD [3,4].

1.1. Unsupervised Domain Adaptation

Domain adaptation [5] is a type of transfer learning approach aiming at adapting a model from a source domain to a different but related target domain. In particular, unsupervised domain adaptation (UDA) addresses the setting where labeled data are available in the source domain and only unlabeled data are available in the target domain. Domain shift refers to the discrepancies between different domains. In the context of fault diagnosis, domain shift can occur across various scenarios, including different operating conditions [6,7], different units of a fleet [8,9,10,11], data collected under laboratory versus real-world conditions [12], and adaptations from synthetic or simulated data to real data [13,14]. Prevailing approaches for UDA focus on reducing the discrepancy between domains and learning domain-invariant features using the maximum mean discrepancy (MMD) [15], maximum classifier discrepancy (MCD) [16], optimal transport [17], or domain-adversarial training [18]. The latter approach is at the core of domain-adversarial neural networks (DANNs) [18], and has been applied extensively for fault diagnosis [10,11,19].

1.2. Self-Training Methods

Self-training [20,21] has emerged as an effective alternative for domain adaptation. First introduced for semi-supervised learning, self-training consists in iteratively generating a set of pseudo-labels [22] on the unlabeled data and retraining the network under the supervision of these pseudo-labels [23,24,25]. However, noisy and inaccurate pseudo-labels hurt the training process. To address this issue, only the most confident predictions are selected for pseudo-labeling, typically using prediction confidence (i.e., the maximum softmax probability) as a proxy for correctness. Curriculum pseudo-labeling (CPL) [26] is a strategy where pseudo-labels are gradually introduced during the learning process in an “easy-to-hard” manner, starting with the most confident target predictions. Once the model is adapted to the target domain, additional samples can be explored. Confident predictions can be selected using a fixed confidence threshold [27] or an adaptive threshold that dynamically adjusts for each class during training to consider the varying difficulties of classes and enable target samples from low-confidence classes to participate early in the training [26,28]. Alternatively, pseudo-labels can be selected by fixing a proportion of the most confident predictions for each class instead of using an explicit threshold value [29,30]. In [31], the variance of predictions between two network sub-branches is used as a replacement for prediction confidence to estimate uncertainty. In their study, French et al. [32] state that the confidence thresholding stabilizes the training and acts as filter to increase the number of correct pseudo-labels.
Two main challenges arise in self-training algorithms [33]: (1) Choosing a trustworthy proxy measure of classification accuracy on unlabeled data; and (2) selecting a threshold on this measure for pseudo-labeling at each training iteration. While solutions have been proposed to tackle the second challenge in the previously discussed literature, in the form of adaptive thresholds capable of handling varying class difficulties during training, the first challenge has remained unaddressed. In this work, we address the first challenge, which is related to uncertainty estimation and model calibration. It is well known that deep neural networks are often badly calibrated and produce over-confident outputs, even for incorrect predictions [34,35,36]. Furthermore, in the presence of a domain shift, the calibration of a model trained on the source domain will degrade even more in the target domain due to the distribution shift [37,38]. In self-training algorithms for domain adaptation, the confidence-based selection of pseudo-labels is hindered by poorly calibrated confidence estimates in the target domain. This limitation restricts the quality of pseudo-labels, leading to error accumulation.

1.3. UDA and Self-Training in IFD

In the recent literature, various applications of UDA methods for deep learning-based intelligent fault diagnosis have been explored [3,4]. The main directions explored include domain-adversarial training [39,40,41,42], MMD [43,44], and MCD [45]. Among all possible methods, DANNs are established as a strong and reliable baseline on most benchmark datasets [3]. Despite their effectiveness in aligning source and target features, these methods face challenges with complex distributions and high variability, as they rely on global alignment metrics that may not capture fine-grained class distinctions.
As already mentioned, in recent years, self-training UDA methods based on pseudo-labeling of target samples have emerged. These approaches, which are the main focus of this study, have also been explored for IFD applications. Prediction-consistency-guided convolutional neural networks (PCG-CNNs) [46] draw direct inspiration from [32] and uses a Mean Teacher with a consistency loss, a fixed confidence threshold of 0.96 , and a class balance loss. The deep transfer learning with improved pseudo-label learning method (DTL-IPLL) [47] combines MK-MMD feature alignment and pseudo-labeling with a class-wise adaptive threshold as well as a “making-decision-twice” strategy, i.e., predicting twice and discarding the predictions if they differ (which is similar in spirit to Monte Carlo dropout [48]). Wang et al. [49] proposed achieving joint distribution alignment by combining marginal alignment using a DANN with Wasserstein distance and conditional alignment using a triplet loss, using pseudo-labels for target samples. Pseudo-labels are also selected using CPL with a dynamic adaptive threshold. The contrastive cluster center (CCC) [50] approach involves a combination of adversarial training, contrastive cluster alignment and pseudo-labeling, where pseudo-labels of target samples far away from cluster centers are filtered out, but this method does not use self-training. In the semi-supervised learning setting, Ref. [51] uses an aggregation of indicators among which is the entropy on unlabeled samples. Pseudo-labels can also be obtained in an unsupervised way through clustering [52]. Differently, Ref. [53] uses a prototypical network and filters the pseudo-labels using a confidence threshold based on Monte Carlo dropout uncertainty. Finally, Ref. [54] proposes gradually enlarging the set of selected pseudo-labels by using the Euclidean distance in feature space as a measure of confidence, and assigns pseudo-labels using a nearest-neighbor classifier instead of directly predicting with the classifier itself.
However, while these works differ in their use of confidence filtering and distributional alignment techniques, none of them has addressed the challenge of the calibration of confidence estimates for selecting pseudo-labels in the target domain, which may lead to limited robustness and unreliable pseudo-labels, especially in cases of substantial domain shifts where confidence is not well correlated with accuracy in the target domain. Moreover, accurate uncertainty estimates are crucial for trust and adoption by engineers for critical applications, which involves ensuring confidence calibration of models before their deployment.

1.4. Contributions

In this paper, we propose a novel UDA method called Calibrated Adaptive Teacher (CAT). The primary novelty of the proposed approach is in improving the calibration in the target domain, with the goal of increasing the accuracy of selected pseudo-labels. CAT consists in a cross-domain teacher–student architecture, where the student network receives confident target pseudo-labels from the teacher network, which is in turn updated by an exponential moving average of the student’s weights. Domain-adversarial feature learning is leveraged to alleviate the domain gap. This architecture is based upon the Adaptive Teacher (AT) [55], recently introduced for object detection in computer vision, and has never been applied to IFD. To address the issue of target-domain calibration, we propose calibrating the predictions of the teacher network on target samples throughout the self-training process, using post hoc calibration techniques such as temperature scaling [34]. Note that previous works on calibration have focused on calibrating already-trained models, which improves the quality of their uncertainty estimates, but has no effect on their performance. However, the novelty of this work relies on the integration of calibration in the self-training procedure in order to improve both the final accuracy and the calibration.
We apply our proposed method to domain-adaptive intelligent fault diagnosis and perform extensive benchmarks and ablation studies on the Paderborn University (PU) bearing dataset, which is characterized by large domain gaps and provides challenging transfer tasks between operating conditions. Experiments are carried out on time-domain and frequency-domain (Fourier transform) inputs, following the benchmark setup by [3]. We demonstrate state-of-the-art performance on most transfer tasks, both with time-domain and frequency-domain inputs. The code is available at https://github.com/EPFL-IMOS/CAT (accessed on 21 November 2024).
The main contributions of this paper are summarized as follows:
  • We propose a novel unsupervised domain adaptation approach, Calibrated Adaptive Teacher (CAT), aiming to improve the calibration of pseudo-labels in the target domain, and thus the overall accuracy on the target data. Our approach consists in introducing post hoc calibration of the teacher predictions during the training.
  • We evaluate our approach on intelligent fault diagnosis and conduct extensive studies on the Paderborn University (PU) bearing dataset, with both time-domain and frequency-domain inputs.
  • CAT significantly outperforms previous approaches in terms of accuracy on most transfer tasks, and effectively reduces the calibration error in the target domain, leading to increased target accuracy.
  • We compare four different post hoc calibration techniques, and demonstrate that temperature scaling [34] and CPCS [37] are the most effective calibration strategies.
The remainder of the paper is organized as follows. In Section 2, we provide the technical background required, before introducing the details of our method in Section 3. Then, we present experimental settings and results in Section 4. Finally, Section 6 concludes this paper.

2. Background

2.1. Notation

We use the following notation throughout the paper. Since fault diagnostics can be defined as classification problems, we consider K-class classification tasks, whereby each class is typically associated with a different fault type. The source (S) and target (T) data are split into training and test sets, with samples and labels, respectively, denoted by X train S , Y train S , X test S , Y test S , and X train T . In the UDA setting, no labels are available in the target domain. The target training labels Y train T and target test set X test T , Y test T are used solely for evaluation purposes.

2.2. Domain-Adversarial Neural Networks (DANNs)

Domain-adversarial neural networks (DANNs) [18] have emerged as one of the most prominent approaches in UDA. Typically, DANNs comprise a feature encoder, a task-specific module (e.g., a classification head, as in our case), and a domain classifier, also referred to as a discriminator. The feature encoder is shared between the classifier and the discriminator. The underlying principle of this methodology is to train the discriminator to classify the input sample features as belonging to either the source or the target domain. While the discriminator is trained to minimize its classification error, the feature encoder tries to generate indistinguishable features to fool the discriminator, hence the term adversarial. This is commonly achieved using a gradient reversal layer (GRL) [18]. As a result, the marginal distributions of source and target features become aligned, ultimately improving the performance of the source-trained classifier. Domain-adversarial training occurs concurrently with the training of the main task.

2.3. Mean and Adaptive Teachers

Mean Teacher [56] is a variant of the temporal ensembling [57] approach originally proposed for semi-supervised learning, where knowledge is distilled from a teacher network into a student network. The student is trained with standard gradient updating, whereas the teacher is gradually updated through an exponential moving average (EMA) of the student weights, resulting in an ensemble of all the previous iterations of the student network, increasing its robustness to inaccurate and noisy predictions on unlabeled data. The knowledge distillation can be achieved via a consistency loss between the teacher’s and student’s predicted probabilities, or via hard pseudo-labeling of the unlabeled samples by the teacher with a confidence threshold to increase the number of correct labels from the teacher used in the student’s learning. Mean Teacher was extended to domain adaptation in [32], using a consistency loss on predicted target probabilities and a confidence threshold, as well as a class balance loss to minimize the binary cross-entropy between the mean target class probabilities and a uniform distribution. The confidence threshold is proposed as a replacement to the Gaussian ramp-up used in [56,57], to stabilize the training and act as filter to increase the number of correct pseudo-labels from the teacher used in the student’s learning.
Adaptive Teacher (AT) [55] was recently proposed to extend the semi-supervised learning Mean Teacher-based method Unbiased Teacher [58] for UDA in object detection. AT basically combines DANNs and Unbiased Teacher into a single architecture (see Figure 1). Concretely, a domain discriminator is added to the student network to perform feature alignment jointly with mutual teacher–student learning. In AT, the teacher provides hard pseudo-labels of target samples that are filtered using a fixed confidence threshold. Then, the cross-domain student network is trained simultaneously on the labeled source data and the pseudo-labeled target data.

2.4. Model Calibration

Classification models typically output conditional probabilities for each class given an input sample x i by applying a softmax activation on the logits: p i = σ ( z i ) . The predicted class is y ^ i = arg max y p i [ y ] , and its corresponding probability is called the prediction confidence:
c i : = max p i = p i [ y ^ i ] .
A classifier is said to be well calibrated if the confidence estimates are equal to the true accuracy of the predictions. For instance, if the confidence is equal to 0.9, the prediction should be correct 90 percent of the time. The calibration of a model can be represented visually by a reliability diagram [34,59], plotting the expected accuracy as a function of average confidence (see, e.g., Figure 3).In order to estimate these quantities from a finite dataset, we divide the [ 0 , 1 ] interval into M equally spaced bins of size 1 / M , and partition the samples into groups B m : = { i ; m 1 M < c i m M , 1 m M } . Then, we define the expected accuracy and average confidence of a bin as follows:
acc ( B m ) : = 1 | B m | i B m 1 ( y ^ i = y i )
conf ( B m ) : = 1 | B m | i B m c i .
where y i is the true label and 1 is the indicator function, equal to 1 when its argument is true and 0 otherwise. In a perfectly calibrated model, these quantities are supposed to be equal. Whenever we have acc ( B m ) < conf ( B m ) , the model is said to be over-confident, and in the opposite case, it is said to be under-confident. Hence, a well-calibrated model should produce a reliability diagram close to the diagonal. To summarize the calibration quality, a commonly used metric is the expected calibration error (ECE) [60]:
ECE : = m = 1 M | B m | N | acc ( B m ) conf ( B m ) |
where N is the total number of samples. Its value is in the [ 0 , 1 ] range, with 0 corresponding to a perfectly calibrated model. We use a number of bins of M = 10 in the ECE computations throughout the paper, following [60].
While a well-calibrated model is desirable, it was shown empirically that neural networks are often badly calibrated in classification [34], regression [35], and anomaly detection [36]. In response, several techniques for network calibration have been developed, which can be broadly classified into post hoc, train time, and through out-of-distribution detection [61]. Additionally, it was also demonstrated that calibration degrades under domain shift [62]. Thus, in the setting of UDA, models are poorly calibrated in the target domain [38].

3. Proposed Method

3.1. Overview

We propose the Calibrated Adaptive Teacher (CAT) framework for unsupervised domain adaptation, an extension of the Adaptive Teacher (AT) [55]. CAT aims to improve the quality of pseudo-labels generated by the teacher network in the target domain. The primary innovation of this architecture lies in introducing post hoc calibration into the self-training process.
The architecture, summarized in Figure 1, comprises two networks: a teacher and a student network, sharing the same architecture but with different weights.
Student (see Figure 1, on green background). In the student network, both source and target inputs first pass through a feature encoder (➀). This encoder consists of a 1D-CNN followed by a 256-dimensional bottleneck, as detailed in [3]. Subsequently, source and target fault classes are predicted by a fully connected (FC) linear layer with softmax activation (➁). The source and target features are also fed into the gradient reversal layer (GRL) (➂) and domain classifier (DC) for domain-adversarial training (➃). The weights of each component in the student network are then updated through gradient descent training, using the loss function detailed in the following paragraphs.
Teacher (see Figure 1, on orange background). On the teacher side, only target inputs go through a feature encoder (➀) and a linear classification head (➁). Teacher predictions are then used to supervise the student through pseudo-labeling. Before selecting the pseudo-labels based on confidence, we introduce post hoc calibration (➄) to calibrate the teacher predictions, which is the primary novelty in this architecture. Essentially, this calibration transforms the target logits (i.e., the outputs before the softmax) to obtain better-calibrated probabilities after applying the softmax. We then select the most confident predictions as pseudo-labels, using a class-wise adaptive threshold instead of the fixed threshold used in [55] (➅). These pseudo-labels serve as supervision for the student network to train on target samples. The teacher weights are frozen and updated through the EMA of the student weights between each training iteration, as presented in Figure 1.
We compare existing UDA approaches for IFD based on self-training with pseudo-labels in Table 1. To the best of our knowledge, none of these works has investigated the aspect of model calibration. The details of each layer in the architecture are provided in Table 2. Following this overview of our proposed approach, we introduce the key components in more detail.

3.2. Calibrated Self-Training

A model is considered well calibrated if its confidence scores align with the accuracy of predictions. Since our objective is to increase the accuracy of the selected pseudo-labels, and confidence serves as a proxy for accuracy, we aim for well-calibrated target outputs. Therefore, we propose calibrating the teacher network’s outputs before selecting confident pseudo-labels for training. To achieve this calibration, we leverage a post hoc calibration technique. Different post hoc calibration techniques are available for multi-class classification, with differences in effectiveness and the number of parameters. In this study, we compare four previously introduced post hoc calibration techniques: temperature scaling, vector scaling, matrix scaling, and calibrated predictions with covariate shift (CPCS) [37]. All these techniques rely on a labeled hold-out test set from the source domain to learn a transformation of the target logits. Temperature scaling, with a single scalar parameter, was identified to be the most effective method in the study by Guo et al. [34]. Additionally, we evaluate vector and matrix scaling, which adjust each class differently, requiring 2 K and K 2 + K parameters, respectively, where K is the number of classes. The first three techniques do not account for distribution shifts. In the context of a domain shift, we also evaluate CPCS, which addresses covariate shift by applying importance weighting with a logistic domain discriminator on domain-invariant features. These features are obtained through the domain classifier in our approach.
Definition 1
(Calibrated self-training). Let x be an unlabeled input and  z be the logits obtained from the teacher network before the softmax activation σ. The calibrated teacher predictions are defined as
p teacher cal ( x ) = σ ( f ( z ) )
where f : R K R K is a calibration function operating on the logits and K is the number of classes.
We propose four variants of CAT, each incorporating different post hoc calibration techniques for multi-class classification.
CAT-TempScaling Temperature scaling is a common approach that involves scaling the output logits using a single scalar parameter T, called temperature. This parameter is tuned to minimize the negative log-likelihood on a hold-out test set in the source domain. The newly calibrated probabilities are expressed as p i = σ ( z i / T ) . The calibration function is defined as follows:
f : z z / T where T = arg min T R 1 | X test S | ( x , y ) k = 1 K 1 ( y = k ) · log σ ( z / T ) [ k ]
where the first summation is over the source test samples and labels X test S × Y test S . CAT-VectorScaling Vector scaling involve transforming the logits using a linear transformation with a matrix W and a bias vector b : p i = σ ( W z i + b ) , where the matrix is restricted to be diagonal, and the parameters W , b are tuned to minimize the negative log-likelihood on a hold-out source test set. The calibration function is defined as follows:
f : z W z + b where W , b = arg min ( W , b ) diag ( K ) × R K 1 | X test S | ( x , y ) k = 1 K 1 ( y = k ) · log σ ( W z + b ) [ k ]
CAT-MatrixScaling Matrix scaling scales the logits using a similar transformation, but where the matrix can be non-diagonal. The corresponding calibration function is defined as follows:
f : z W z + b where W , b = arg min ( W , b ) R K × K × R K 1 | X test S | ( x , y ) k = 1 K 1 ( y = k ) · log σ ( W z + b ) [ k ]
CAT-CPCS Calibrated prediction with covariate shift (CPCS) [37] proposes to correct for covariate shift by combining adversarial feature alignment (as in DANNs [18]), importance weighting using a logistic domain discriminator, and temperature scaling. First, domain-invariant features are learned through domain-adversarial training. Then, a logistic domain discriminator is trained to distinguish between the features extracted from X train S and X train T , and its output probabilities are used to estimate the importance weight w ( x ) for each sample. Specifically, the weight is equal to the ratio between the probability of belonging to the target domain and the probability of belonging to the source domain, allowing for a translation from the source distribution to the target one. Finally, the optimal temperature is found by minimizing the weighted Brier score [63] (i.e., the mean squared error between outputs and one-hot labels) on the hold-out source test set. The calibration function is defined as follows:
f : z z / T where T = arg min T R 1 | X test S | ( x , y ) w ( x ) k = 1 K 1 ( y = k ) σ ( z / T ) [ k ] 2
All optimization problems for finding the optimal calibration parameters are convex and solved using the optimize.fmin function from scipy.
Note that this step adds computational overhead. Nevertheless, as we update the calibration function only once per epoch, the added cost is minimal. Moreover, the computational complexity of the post hoc calibration is typically relatively low. In TempScaling, the complexity is driven by the computation of the log-likelihood, which is linear in both the test set size and the number of classes, and the number of iterations of the temperature optimization, which is very small for tuning a single scalar parameter and considered constant. Vector and MatrixScaling tune, respectively, 2 K and K 2 + K parameters, adding only a slightly higher constant term due to the higher dimension. CPCS requires, in addition, to train a logistic domain discriminator to compute the importance weights, adding up to a cost with linear complexity in the number of samples and iterations (gradient-based training). No additional adversarial feature alignment is required, as CAT already learns aligned features thanks to the domain classifier.

3.3. Adaptive Confidence Threshold

Curriculum pseudo-labeling methods dynamically adjust the confidence threshold for each class during training, based on the accuracy of each class. We adopt the method introduced in [26]. At training iteration t, the dynamic threshold for class k is defined as
T t ( k ) = a t ( k ) · τ
where a t ( k ) represents the accuracy, and  τ is a fixed threshold value. As proposed in [26], the accuracy can be substituted with the learning effect of the class, which is reflected by the number of high-confidence predictions for this class:
σ t ( k ) = x X train T 1 ( max p teacher ( x ) τ ) · 1 ( arg max y p teacher ( x ) [ y ] = k )
Subsequently, this quantity is scaled, undergoes a non-linear mapping M , and is ultimately used to define the dynamic threshold as follows:
β t ( k ) = σ t ( k ) max k σ t ( k ) M ( x ) = x 2 x T t ( k ) = M ( β t ( k ) ) · τ
At training iteration t, we ultimately define the set of selected target pseudo-labels:
X ^ train T = { x ; x X train T , max p teacher ( x ) T t ( arg max y p teacher ( x ) [ y ] ) }
Y ^ train T = { arg max y p teacher ( x ) [ y ] ; x X train T , max p teacher ( x ) T t ( arg max y p teacher ( x ) [ y ] ) }
In our CAT method, the teacher probabilities p teacher ( x ) are replaced with the calibrated teacher probabilities p teacher cal ( x ) .

3.4. Loss Function and Training Procedure

3.4.1. Student Training

We denote the parameters of the student network as θ = { W E , W C } , where W E and W C represent the weights of the student encoder and classification head, respectively. The weights of the discriminator are denoted as W DC . Instead of the standard DANN [18], we adopt the enhanced approach of smooth domain-adversarial training (SDAT) [64]. This involves using the Sharpness Aware Minimization (SAM) optimizer [65] on the task loss (i.e., supervised source loss) and incorporating the minimum class confusion (MCC) loss [66], which stands as a current state-of-the-art method for domain adaptation in computer vision tasks. The MCC loss serves as a non-adversarial regularization term designed to minimize pairwise class confusion in the target domain.
The loss function of CAT, denoted L CAT , comprises four terms: a supervised source loss L C , a target pseudo-labeling loss L PL , the domain classifier term L DC , and the MCC loss L MCC . The supervised loss is calculated as the cross-entropy between source predictions and labels:
L C ( θ ; X train S , Y train S ) = 1 | X train S | ( x , y ) k = 1 K 1 ( y = k ) · log p ( x ) [ k ]
where the first summation is over the source training samples and labels X train S × Y train S . The pseudo-labeling loss is then calculated as the cross-entropy between predictions and the pseudo-labels:
L PL ( θ ; X ^ train T , Y ^ train T ) = 1 | X ^ train T | ( x , y ) k = 1 K 1 ( y = k ) · log p ( x ) [ k ]
where the first summation is over the pseudo-labels X ^ train T × Y ^ train T . The domain classifier loss is then expressed as a binary cross-entropy between the domain classifier predictions and domain labels, which is set to 1 for the source and 0 for the target training samples:
L DC ( W E , W DC ; X train S , X train T ) = 1 | X train S | x X train S log p DC ( x ) 1 | X train T | x X train T log 1 p DC ( x )
Finally, the total loss is an equally weighted sum of the individual loss terms:
L CAT ( θ , W DC ; X train S , Y train S , X ^ train T , Y ^ train T ) = L C L DC + L PL + L MCC ,
omitting the arguments of each loss term for brevity. Source and target cross-entropies are given equal weights, and Ref. [66] found that a weight of 1 for the MCC loss worked across all their experiments. For the domain-adversarial training, we adopt the same strategy as [18], only used in [3], progressively increasing the GRL coefficient from 0 to 1 following the formula 2 1 + exp ( 10 p ) , where p increases linearly from 0 to 1 during training. The total loss is minimized with respect to the student weights, while the domain classifier loss is minimized with respect to the discriminator weights and maximized with respect to the encoder weights in an adversarial way via the gradient reversal layer. Thus, the overall objective is expressed as follows:
min θ max W DC L CAT ( θ , W DC ; X train S , Y train S , X ^ train T , Y ^ train T )

3.4.2. Teacher Updating

The teacher network parameters are initialized with the student weights at the beginning of self-training. Subsequently, they are updated between each training iteration by the EMA of the student weights with an update rate α :
θ teacher α · θ teacher + ( 1 α ) · θ

3.5. Training Procedure of the Entire Pipeline

The CAT training procedure consists of three phases. The first phase involves a source-only training of the student network using only the supervised loss L C . In the second phase, the domain classifier loss is enabled after T DA epochs for domain-adversarial training. The third phase, mutual teacher–student training, begins at T PL , with  T DA T PL . A warm-up phase is necessary to ensure sufficient pseudo-label quality and to avoid compromising the training by fitting noise [30]. Lastly, the calibration is enabled later in the training at T cal , with  T PL T cal .
During the mutual training phase, at the beginning of each given epoch t, we estimate the class-wise adaptive threshold values T t ( k ) following (12), as well as the calibration function following (6) in CAT-TempScaling, (7) in CAT-VectorScaling, (8) in CAT-MatrixScaling, and (9) in CAT-CPCS, depending on the variant of CAT considered. These estimations involve the unlabeled target training set and the labeled source test set. At every training iteration, a batch of inputs and labels is sampled from the source training set, and an unlabeled batch of inputs is sampled from the target training set. We compute and calibrate the logits of the teacher network to obtain calibrated probabilities, which are subsequently filtered using the adaptive threshold following (13,14). We then minimize the objective (19) by taking a gradient descent step. Finally, the teacher weights are updated by EMA following (20).
The complete training procedure is detailed in Algorithm 1.   
Algorithm 1: Calibrated Adaptive Teacher (CAT) training procedure
Sensors 24 07539 i001

4. Experiments

In this section, we discuss the data and transfer learning tasks, and provide details on the experimental settings and hyperparameters.

4.1. Case Study

Experiments were conducted on the bearing fault diagnostics Paderborn University (PU) dataset [67], which comprises challenging transfer learning tasks across various operating conditions and stands as a benchmark in the recent literature [3,68,69,70]. The dataset comprises motor current signals of an electro-mechanical drive system, allowing for bearing diagnostics without the need for additional acceleration sensors for vibration analysis. The domain adaptation tasks involve adapting between four different operating conditions—rotational speed, load torque and radial force—described in Table 3.
From a domain knowledge and signal processing point of view, domain shifts of varying intensities can be expected. Varying the speed, torque, or force has very different impacts on bearing vibration signals characteristics [71]. Thus, we can expect each domain shift to result in various transfer task difficulties. Changing the speed (transfer tasks 0 1 , 1 2 and 1 3 ) significantly impacts both the frequency content—frequencies will be scaled due to rotational speed, and it can introduce harmonics, broadening the spectrum—and the amplitude. Changing the force (transfer tasks 0 3 , 1 3 and 2 3 ) increases the mechanical stress in the system, which also significantly affects the amplitude, as well as the frequency content to a lesser extent. Torque (transfer tasks 0 2 , 1 2 and 2 3 ) has less impact on the signals, resulting in easier tasks, especially when other parameters are left fixed (tasks 0 2 ).
We adopt the same setting as [3] for a fair comparison. The classification problem has 13 classes and consists in classifying different combination of fault types and severities, represented by different bearing codes, described in Table 4. Each sample consists of a 1024-dimensional input in the time-domain, or in the frequency-domain after fast Fourier transform.

4.2. Training Parameters

We train the model for a total of T total = 300 epochs with a batch size of 64, using the Adam optimizer with a learning rate of 0.001, β 1 = 0.9 , β 2 = 0.999 , and an 2 weight decay of 10 5 . The learning rate is reduced by a factor of 10 at epochs 150 and 250. The domain-adversarial loss is introduced at epoch T DA = 50 . Self-training starts at epoch T PL = 50 for the PU time domain, and at epoch T DA = 100 for the PU frequency domain. In both cases, we introduce the calibration from epoch T cal = 150 . The teacher EMA update rate is set to 0.999 . The fixed threshold value in the adaptive threshold is τ = 0.9 . No weak–strong data augmentations were used. The train–test data splitting follows [3], with 80% of total samples for training and 20% for testing.

5. Results

In this section, we present the results of experiments on the PU dataset. First, we demonstrate the impact of our proposed calibration method on the quality of pseudo-labels and model performance in Section 5.1. Following this, Section 5.2 and Section 5.3 present a quantitative benchmark analysis comparing different UDA methods in terms of performance and calibration. We conduct ablation studies in Section 5.4.

5.1. Calibration and Quality of Pseudo-Labels

We begin by illustrating the motivation behind our approach, which is to address the issue of model calibration in the target domain, aiming to increase the accuracy of pseudo-labels.
In Figure 2, we present the reliability diagrams of the teacher predictions of a trained AT model on the source and target test sets. Blue bars under the diagonal represent over-confident predictions and vice-versa. The red area visualizes the gap between actual and ideal calibration. The initial observation is that the model is over-confident in both domains, but the expected calibration error (ECE) is even higher in the target domain (36.90%) than in the source domain (21.27%).
We present the same reliability diagrams after applying the temperature scaling (i.e., CAT-TempScaling) in Figure 3. Calibration drastically reduces the calibration error on the source domain to 5.41%, as expected, since the temperature was tuned on the source test set. However, it also significantly reduces the ECE on the target domain (down to 10.56%), even though temperature scaling does not account for domain shift. Thanks to feature alignment, the source and target distributions are similar enough to allow for a transfer of temperature scaling to the target data.
As a consequence, we can expect the quality of the selected pseudo-labels to improve, as the confidence more closely matches the true accuracy, and we reduce the number of over-confident wrong target predictions. Figure 4 displays the evolution of target pseudo-label accuracy during training for the transfer task 0 1 , comparing AT + MCC + SDAT (denoted by AT*) and our proposed methods CAT*-TempScaling and CAT*-CPCS. At the epoch where calibration is introduced, we observe an absolute increase in pseudo-label accuracy of around 10%, which is maintained during the rest of the training. At the last iteration, pseudo-label accuracy reaches 47.64% for AT*, and 57.37% and 58.45% for CAT*-TempScaling and CAT*-CPCS, respectively.
In addition, Figure 5 illustrates the evolution of the target test accuracy and ECE during training for the different methods. Specifically, we compare the source-only model, DANN, DANN + MCC + SDAT (denoted as DANN*), AT + MCC + SDAT (denoted as AT*), and our proposed methods CAT*-TempScaling and CAT*-CPCS. Domain adaptation and self-training both start at T DA = T PL = 50 epochs, and calibration is enabled at T cal = 150 epochs. All domain adaptation methods drastically improve the performance compared to the source-only model. AT significantly improves over the DANN baseline, but still exhibits a high ECE. After introducing calibration, the ECE drops significantly (see Figure 5, right), and the increased pseudo-label accuracy also translates into an improvement in accuracy on the target validation set (see Figure 5, left).
Although TempScaling and CPCS effectively improve calibration in the target domain, this is not always the case for VectorScaling and MatrixScaling. The reliability diagrams shown in Figure 6 and Figure 7 demonstrate that these two calibration techniques are able to reduce calibration error more effectively than temperature scaling in the source domain. However, in the target domain, calibration error is not reduced. This observation shows that the higher number of parameters (scaling matrix and bias vector) of these calibration techniques prevents them from generalizing to the target domain, as opposed to temperature scaling and CPCS, which only tune a single scalar parameter on the source test set. In other words, the scaling parameters are too specific to the source domain.

5.2. Comparative Analysis of Performance

In this section, we conduct a comparative analysis on the 12 transfer tasks of the PU bearing fault diagnosis benchmark, considering both time-domain and frequency-domain (FFT) inputs. In most related works, only time-domain inputs are considered (see Table 1). The compared methods include the model trained on the source (source-only), DANN, which is the best-performing domain adaptation method overall in the survey [3] and represents our main baseline, and DANN + MCC + SDAT (denoted as DANN*, where the asterisk indicates that the model integrates the MCC loss and SDAT [64]). We also evaluate AT*, our adaptation of AT [55]. Finally, we assess our proposed calibrated self-training methods CAT* with four different post hoc calibration techniques, namely, CAT*-TempScaling, CAT*-CPCS, CAT*-VectorScaling and CAT*-MatrixScaling. In the case of teacher–student models, we always report the results of the teacher network throughout the paper.
We use the same parameters as [3] for the source-only and DANN training. We assess the performance at the last training iteration, as using the test labels for early stopping is unrealistic. Thus, we compare our results to the “Last-Mean” results of [3]. All our experiments are repeated five times with different random seeds. As performance measures, we report the average accuracy per transfer task and the overall average accuracy. Additionally, we report the average rank, which is better suited for comparing multiple methods over multiple tasks of varying difficulties [72].
The results for the PU dataset in the time domain are presented in Figure 8 and Table 5. The domain adaptation baseline DANN achieves an overall accuracy of 46.53% across all tasks, consistent with the results reported by [3], and significantly higher than the source-only model (33.78%). It is worth noting that some transfer tasks are easy, with the source model already performing well ( 0 2 and 2 0 ), while other tasks are challenging, with accuracies below 50%. DANN* (DANN + MCC + SDAT) has a slightly higher average accuracy of 47.02%. The AT* method obtains a significantly higher average accuracy of 53.13%, outperforming DANN* on every task except the easiest one ( 0 2 ). Overall, the best-performing methods are our proposed CAT*-TempScaling and CAT*-CPCS, with average accuracies of 54.50% and 54.42%, and average ranks of 2.28 and 2.40. The calibrated methods outperform the non-calibrated versions in 10 out of 12 tasks. Among the four calibration techniques, we observe that only TempScaling and CPCS are effective, whereas VectorScaling and MatrixScaling degrade the performance in most tasks.
The results for the frequency-domain inputs are presented in Figure 9 and Table 6. The findings are consistent with those in the time domain, and the methods exhibit similar average ranks. The source-only performance is higher than in the time domain, achieving 44.04% accuracy. The improvement of DANN* (59.84%) over DANN (57.89%) is more pronounced. Once again, the AT* method demonstrates a high average accuracy (63.94%) and significantly outperforms DANN* on all the challenging tasks, while showing comparable performance on the easier ones. As in the time domain, our proposed CAT*-TempScaling and CAT*-CPCS achieve the best performance, achieving 64.76% and 64.92% accuracy, with an equal average rank of 2.76 out of the eight compared methods. CAT*-VectorScaling and CAT*-MatrixScaling, however, fail to improve upon AT*.
To assess the statistical significance of the differences between compared methods, we conducted a statistical post hoc analysis using pairwise Wilcoxon signed-rank tests with Holm’s step-down procedure at a significance level of 0.05 [72]. This procedure compares the sums of ranks across the 12 transfer tasks and five runs, testing against the null hypothesis stating that methods have equal performance. In Table 5, Table 6 and Table 7, the average ranks are reported in the rightmost column, with the rank of the best method and statistically equivalent methods in bold. As a result, our proposed CAT*-TempScaling and CAT*-CPCS have higher ranks than AT*, and this difference is significant in both the time and frequency domains. However, both calibration techniques are equivalent, indicating that the importance weighting in CPCS is not necessary in this case. The features are sufficiently well aligned to make temperature scaling directly transferable and effective.
The result of the statistical tests are summarized in critical difference diagrams [72], shown in Figure 10. These diagrams visualize the average ranks of each method in terms of target accuracy, and connect statistically equivalent methods with a horizontal bar (i.e., when the null hypothesis could not be rejected at a significance level of 0.05). In both the time and frequency domains, CAT*-TempScaling and CAT*-CPCS emerge as the top-performing methods, with equivalent performance. The next best methods, AT* and CAT*-VectorScaling, have significantly different ranks with time-series inputs but not with FFT inputs. The CD-diagram (https://github.com/hfawaz/cd-diagram accessed on 13 October 2024) library was used to create these diagrams.

5.3. Comparative Analysis of Calibration Error

In this section, we compare the calibration error, in terms of expected calibration error (ECE), across different methods. The ECE of the target-domain predictions is presented in Figure 11 and detailed in Table 7 for time-domain inputs, with frequency-domain results available in Appendix A. On average, both the source-only model and DANN exhibit very high ECE values, exceeding 50% on most tasks, particularly the challenging ones. While the Adaptive Teacher shows improved calibration, it still maintains a relatively high average ECE of 34.26%. CAT*-VectorScaling and CAT*-MatrixScaling fail to improve calibration in the target domain on most tasks compared with AT*, aligning with their subpar accuracy performance. As observed previously, their higher number of parameters (scaling matrix and bias vector) prevent them from generalizing to the target domain due to this over-parametrization, as opposed to temperature scaling and CPCS, which only tune a single scalar parameter on the source test set. Finally, CAT*-TempScaling and CAT*-CPCS achieve the lowest calibration error, with average ECE values of 12.77% and 11.31%, and average ranks of 1.67 and 1.48, respectively.

5.4. Ablation Studies

In this section, we present the results of ablation studies, providing justification for the design choices in our proposed method and assessing the impact of each component. Initially, we compare the performance of the AT model using a fixed confidence threshold ( τ = 0.9 ) against an adaptive threshold, as introduced in Section 3. The results on the PU dataset in the time domain (see Table 8) indicate that, in most transfer tasks, particularly the most challenging ones, the adaptive threshold outperforms the fixed threshold in terms of target-domain teacher accuracy. The adaptive thresholding strategy achieves an average accuracy of 48.33%, whereas the fixed thresholding strategy averages 46.22% over all tasks.
In a second ablation study, we investigated the influence of incorporating the MCC loss [66] and the SDAT optimization technique [64] into the AT and CAT methods. We conducted experiments for AT, CAT-TempScaling and CAT-CPCS on the PU dataset in time domain, using different combinations of MCC and/or SDAT, as reported in Table 9. For each method, using MCC and SDAT together results in a substantial performance gain, consistent with findings by [64]. The average target-domain accuracy of the teacher network across all tasks is approximately 4% higher when using both MCC and SDAT compared to not using them.

6. Conclusions

In this paper, we tackled the challenge of model calibration for pseudo-labeling in the context of unsupervised domain adaptation. We proposed a novel method called Calibrated Adaptive Teacher (CAT), drawing inspiration from Mean Teacher, self-training with pseudo-labels, and feature alignment through domain-adversarial training. The primary innovation involves calibrating the predictions of the teacher network in the target domain throughout the self-training process. We explored four post hoc calibration techniques, with temperature scaling and CPCS yielding the best results. Interestingly, both techniques performed similarly, despite the fact that the first one does not account for the domain shift. We believe this is due to the presence of already well-aligned features when calibration is introduced, enabling temperature scaling to transfer to the target domain. Experiments on intelligent fault diagnosis demonstrated that our method is able to improve calibration in the target domain, resulting in increased accuracy. On the Paderborn University bearing dataset, our method outperformed previous unsupervised domain adaptation approaches by a significant margin: on average, with time-series inputs, accuracy is 7.5% higher and calibration error is 4 times lower than DANN across the twelve PU transfer tasks. With frequency-domain inputs, the improvement is +5% in accuracy and there is a 2 times lower calibration error.
Our method is effective whenever the model is badly calibrated in the target domain, which is often the case in deep learning. The main limitation of our approach is that if target predictions are already well calibrated, improvement cannot be expected. Additionally, we observed that our approach is more effective in tackling more challenging tasks, specifically when the baseline accuracy is relatively low, without adaptation, and predictions are poorly calibrated. For example, in the transfer tasks 0→2 and 2→0, the source model already shows high accuracy (over 75%, while harder tasks reach under 30%) and relatively low ECE (around 18%, while other tasks are over 50%), and CAT produces no significant performance gain on these tasks compared with DANN.
The application of weak–strong data augmentations, as implemented in [21,55], requires further investigation in the context of intelligent fault diagnosis. The use of weak augmentations for the teacher network and strong augmentations for the student network is highly beneficial in computer vision applications. Exploring suitable augmentations for time-series and spectrum data represents a promising direction for future research.

Author Contributions

Conceptualization, F.F. and O.F.; methodology, F.F. and O.F.; software, F.F.; validation, F.F. and O.F.; writing—original draft preparation, F.F.; writing—review and editing, F.F. and O.F.; visualization, F.F.; supervision, O.F.; funding acquisition, O.F. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the Swiss National Science Foundation (grant number 200021_200461).

Data Availability Statement

Data and code are publicly available.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ATAdaptive Teacher
CATCalibrated Adaptive Teacher
CPCSCalibrated predictions with covariate shift
CPLCurriculum pseudo-labeling
DADomain adaptation
DANNDomain-adversarial neural network
DCDomain classifier
ECEExpected calibration error
EMAExponential moving average
FCFully connected
FFTFast Fourier transform
GRLGradient reversal layer
IFDIntelligent fault diagnosis
MCCMinimum class confusion
PUPaderborn University
SDATSmooth domain-adversarial training
UDAUnsupervised domain adaptation

Appendix A. ECE Results for Frequency-Domain Inputs

Figure A1. Comparison of target-domain expected calibration error (ECE) for different methods on the PU transfer tasks with frequency-domain input. * = with MCC + SDAT.
Figure A1. Comparison of target-domain expected calibration error (ECE) for different methods on the PU transfer tasks with frequency-domain input. * = with MCC + SDAT.
Sensors 24 07539 g0a1
Table A1. Target-domain expected calibration error (ECE) on the different PU transfer tasks with frequency-domain input (ECE in %).
Table A1. Target-domain expected calibration error (ECE) on the different PU transfer tasks with frequency-domain input (ECE in %).
Method 0 1 0 2 0 3 1 0 1 2 1 3 2 0 2 1 2 3 3 0 3 1 3 2 AverageAverage Rank
Source-only74.826.3035.1968.6972.9480.099.8575.1733.7143.6275.8639.8751.347.97
DANN55.484.2618.1065.5557.8670.616.5848.8619.0324.9663.3024.0438.225.95
DANN*56.093.3318.6167.1760.4664.685.6848.6919.4829.1865.3826.4538.776.08
AT*38.623.2615.9249.1140.1146.125.1934.6317.4927.0750.5928.0429.684.63
CAT*-TempScaling31.542.2010.5025.4620.7622.532.6027.8811.3920.4942.2321.9019.961.78
CAT*-CPCS26.272.7813.3825.1123.2115.612.4315.018.8814.3240.6718.1617.151.82
CAT*-VectorScaling39.982.4613.0332.7228.8529.873.5535.0014.5223.7951.8423.5424.933.03
CAT*-MatrixScaling42.494.0916.3339.1132.1035.495.0641.3018.1127.3053.4525.7528.384.73
* = with MCC + SDAT.

References

  1. Zhao, Z.; Li, T.; Wu, J.; Sun, C.; Wang, S.; Yan, R.; Chen, X. Deep learning algorithms for rotating machinery intelligent diagnosis: An open source benchmark study. ISA Trans. 2020, 107, 224–255. [Google Scholar] [CrossRef] [PubMed]
  2. Lei, Y.; Yang, B.; Jiang, X.; Jia, F.; Li, N.; Nandi, A.K. Applications of machine learning to machine fault diagnosis: A review and roadmap. Mech. Syst. Signal Process. 2020, 138, 106587. [Google Scholar] [CrossRef]
  3. Zhao, Z.; Zhang, Q.; Yu, X.; Sun, C.; Wang, S.; Yan, R.; Chen, X. Applications of Unsupervised Deep Transfer Learning to Intelligent Fault Diagnosis: A Survey and Comparative Study. IEEE Trans. Instrum. Meas. 2021, 70, 3525828. [Google Scholar] [CrossRef]
  4. Li, W.; Huang, R.; Li, J.; Liao, Y.; Chen, Z.; He, G.; Yan, R.; Gryllias, K. A perspective survey on deep transfer learning for fault diagnosis in industrial scenarios: Theories, applications and challenges. Mech. Syst. Signal Process. 2022, 167, 108487. [Google Scholar] [CrossRef]
  5. Redko, I.; Habrard, A.; Morvant, E.; Sebban, M.; Bennani, Y. Advances in Domain Adaptation Theory; Elsevier: Amsterdam, The Netherlands, 2019. [Google Scholar]
  6. Wang, Q.; Michau, G.; Fink, O. Domain Adaptive Transfer Learning for Fault Diagnosis. In Proceedings of the 2019 Prognostics and System Health Management Conference (PHM-Paris), Paris, France, 2–5 May 2019; pp. 279–285, ISSN 2166-5656. [Google Scholar] [CrossRef]
  7. Huang, M.; Yin, J.; Yan, S.; Xue, P. A fault diagnosis method of bearings based on deep transfer learning. Simul. Model. Pract. Theory 2023, 122, 102659. [Google Scholar] [CrossRef]
  8. Michau, G.; Fink, O. Domain Adaptation for One-Class Classification: Monitoring the Health of Critical Systems Under Limited Information. Int. J. Progn. Health Manag. 2019, 10, 4. [Google Scholar] [CrossRef]
  9. Shen, F.; Langari, R.; Yan, R. Transfer between multiple machine plants: A modified fast self-organizing feature map and two-order selective ensemble based fault diagnosis strategy. Measurement 2020, 151, 107155. [Google Scholar] [CrossRef]
  10. Zhu, J.; Chen, N.; Shen, C. A New Multiple Source Domain Adaptation Fault Diagnosis Method Between Different Rotating Machines. IEEE Trans. Ind. Informatics 2021, 17, 4788–4797. [Google Scholar] [CrossRef]
  11. Nejjar, I.; Geissmann, F.; Zhao, M.; Taal, C.; Fink, O. Domain adaptation via alignment of operation profile for Remaining Useful Lifetime prediction. Reliab. Eng. Syst. Saf. 2024, 242, 109718. [Google Scholar] [CrossRef]
  12. Yang, B.; Lei, Y.; Jia, F.; Xing, S. An intelligent fault diagnosis approach based on transfer learning from laboratory bearings to locomotive bearings. Mech. Syst. Signal Process. 2019, 122, 692–706. [Google Scholar] [CrossRef]
  13. Wang, Q.; Taal, C.; Fink, O. Integrating Expert Knowledge with Domain Adaptation for Unsupervised Fault Diagnosis. IEEE Trans. Instrum. Meas. 2022, 71, 3500312. [Google Scholar] [CrossRef]
  14. Lou, Y.; Kumar, A.; Xiang, J. Machinery Fault Diagnosis Based on Domain Adaptation to Bridge the Gap Between Simulation and Measured Signals. IEEE Trans. Instrum. Meas. 2022, 71, 3180416. [Google Scholar] [CrossRef]
  15. Long, M.; Cao, Y.; Wang, J.; Jordan, M. Learning Transferable Features with Deep Adaptation Networks. In Proceedings of the 32nd International Conference on Machine Learning, PMLR, Lille, France, 7–9 July 2015; pp. 97–105, ISSN 1938-7228. [Google Scholar]
  16. Saito, K.; Watanabe, K.; Ushiku, Y.; Harada, T. Maximum Classifier Discrepancy for Unsupervised Domain Adaptation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3723–3732. [Google Scholar] [CrossRef]
  17. Courty, N.; Flamary, R.; Habrard, A.; Rakotomamonjy, A. Joint distribution optimal transportation for domain adaptation. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  18. Ganin, Y.; Ustinova, E.; Ajakan, H.; Germain, P.; Larochelle, H.; Laviolette, F.; March, M.; Lempitsky, V. Domain-Adversarial Training of Neural Networks. J. Mach. Learn. Res. 2016, 17, 1–35. [Google Scholar]
  19. Dai, B.; Frusque, G.; Li, T.; Li, Q.; Fink, O. Smart filter aided domain adversarial neural network for fault diagnosis in noisy industrial scenarios. Eng. Appl. Artif. Intell. 2023, 126, 107202. [Google Scholar] [CrossRef]
  20. Amini, M.R.; Feofanov, V.; Pauletto, L.; Hadjadj, L.; Devijver, E.; Maximov, Y. Self-Training: A Survey. arXiv 2023, arXiv:2202.12040. [Google Scholar] [CrossRef]
  21. Hao, Y.; Forest, F.; Fink, O. Simplifying Source-Free Domain Adaptation for Object Detection: Effective Self-Training Strategies and Performance Insights. arXiv 2024, arXiv:2407.07586. [Google Scholar]
  22. Lee, D.H. Pseudo-Label: The Simple and Efficient Semi-Supervised Learning Method for Deep Neural Networks. In Proceedings of the ICML 2013 Workshop: Challenges in Representation Learning (WREPL), Atlanta, GA, USA, 16–21 June 2013. [Google Scholar]
  23. Blum, A.; Mitchell, T. Combining labeled and unlabeled data with co-training. In Proceedings of the Eleventh Annual Conference on Computational Learning Theory, New York, NY, USA, 24–26 July 1998; COLT’ 98. pp. 92–100. [Google Scholar] [CrossRef]
  24. Amini, M.R.; Gallinari, P. Semi-supervised logistic regression. In Proceedings of the 15th European Conference on Artificial Intelligence, NLD, Lyon, France, 21–26 July 2002; ECAI’02. pp. 390–394. [Google Scholar]
  25. Grandvalet, Y.; Bengio, Y. Semi-supervised Learning by Entropy Minimization. In Proceedings of the NIPS, Vancouver, BC, Canada, 5–8 December 2005. [Google Scholar]
  26. Zhang, B.; Wang, Y.; Hou, W.; Wu, H.; Wang, J.; Okumura, M.; Shinozaki, T. FlexMatch: Boosting Semi-Supervised Learning with Curriculum Pseudo Labeling. In Proceedings of the 35th Conference on Neural Information Processing Systems (NeurIPS 2021), Online, 6–14 December 2021. [Google Scholar] [CrossRef]
  27. Tur, G.; Hakkani-Tür, D.; Schapire, R.E. Combining active and semi-supervised learning for spoken language understanding. Speech Commun. 2005, 45, 171–186. [Google Scholar] [CrossRef]
  28. Zhang, W.; Ouyang, W.; Li, W.; Xu, D. Collaborative and Adversarial Network for Unsupervised Domain Adaptation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3801–3809, ISSN 2575-7075. [Google Scholar] [CrossRef]
  29. Zou, Y.; Yu, Z.; Kumar, B.V.K.V.; Wang, J. Domain Adaptation for Semantic Segmentation via Class-Balanced Self-Training. arXiv 2018, arXiv:1810.07911. [Google Scholar] [CrossRef]
  30. Mei, K.; Zhu, C.; Zou, J.; Zhang, S. Instance Adaptive Self-training for Unsupervised Domain Adaptation. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2020; pp. 415–430. [Google Scholar] [CrossRef]
  31. Zheng, Z.; Yang, Y. Rectifying Pseudo Label Learning via Uncertainty Estimation for Domain Adaptive Semantic Segmentation. Int. J. Comput. Vis. 2021, 129, 1106–1120. [Google Scholar] [CrossRef]
  32. French, G.; Mackiewicz, M.; Fisher, M. Self-ensembling for visual domain adaptation. arXiv 2018, arXiv:1706.05208. [Google Scholar] [CrossRef]
  33. Odonnat, A.; Feofanov, V.; Redko, I. Leveraging Ensemble Diversity for Robust Self-Training in the Presence of Sample Selection Bias. arXiv 2023, arXiv:2310.14814. [Google Scholar]
  34. Guo, C.; Pleiss, G.; Sun, Y.; Weinberger, K.Q. On Calibration of Modern Neural Networks. In Proceedings of the 34th International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 1321–1330, ISSN 2640-3498. [Google Scholar]
  35. Kuleshov, V.; Fenner, N.; Ermon, S. Accurate Uncertainties for Deep Learning Using Calibrated Regression. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018; Volume 80, p. 9. [Google Scholar]
  36. Deng, X.; Jiang, X. On confidence computation and calibration of deep support vector data description. Eng. Appl. Artif. Intell. 2023, 125, 106646. [Google Scholar] [CrossRef]
  37. Park, S.; Bastani, O.; Weimer, J.; Lee, I. Calibrated Prediction with Covariate Shift via Unsupervised Domain Adaptation. In Proceedings of the Twenty Third International Conference on Artificial Intelligence and Statistics, PMLR, Online, 26–28 August 2020; pp. 3219–3229, ISSN 2640-3498. [Google Scholar]
  38. Munir, M.A.; Khan, M.H.; Sarfraz, M.S.; Ali, M. SSAL: Synergizing between Self-Training and Adversarial Learning for Domain Adaptive Object Detection. In Proceedings of the NeurIPS 2021, Online, 6–14 December 2021; p. 13. [Google Scholar]
  39. Zhang, B.; Li, W.; Hao, J.; Li, X.L.; Zhang, M. Adversarial adaptive 1-D convolutional neural networks for bearing fault diagnosis under varying working condition. arXiv 2018, arXiv:1805.00778. [Google Scholar] [CrossRef]
  40. She, D.; Peng, N.; Jia, M.; Pecht, M.G. Wasserstein distance based deep multi-feature adversarial transfer diagnosis approach under variable working conditions. J. Instrum. 2020, 15, P06002. [Google Scholar] [CrossRef]
  41. Li, F.; Tang, T.; Tang, B.; He, Q. Deep convolution domain-adversarial transfer learning for fault diagnosis of rolling bearings. Measurement 2021, 169, 108339. [Google Scholar] [CrossRef]
  42. Yao, Q.; Qian, Q.; Qin, Y.; Guo, L.; Wu, F. Adversarial domain adaptation network with pseudo-siamese feature extractors for cross-bearing fault transfer diagnosis. Eng. Appl. Artif. Intell. 2022, 113, 104932. [Google Scholar] [CrossRef]
  43. Wang, X.; Shen, C.; Xia, M.; Wang, D.; Zhu, J.; Zhu, Z. Multi-scale deep intra-class transfer learning for bearing fault diagnosis. Reliab. Eng. Syst. Saf. 2020, 202, 107050. [Google Scholar] [CrossRef]
  44. Qin, Y.; Qian, Q.; Wang, Z.; Mao, Y. Adaptive manifold partial domain adaptation for fault transfer diagnosis of rotating machinery. Eng. Appl. Artif. Intell. 2023, 126, 107082. [Google Scholar] [CrossRef]
  45. Jia, S.; Wang, J.; Han, B.; Zhang, G.; Wang, X.; He, J. A Novel Transfer Learning Method for Fault Diagnosis Using Maximum Classifier Discrepancy with Marginal Probability Distribution Adaptation. IEEE Access 2020, 8, 71475–71485. [Google Scholar] [CrossRef]
  46. Wu, S.; Jing, X.Y.; Zhang, Q.; Wu, F.; Zhao, H.; Dong, Y. Prediction Consistency Guided Convolutional Neural Networks for Cross-Domain Bearing Fault Diagnosis. IEEE Access 2020, 8, 120089–120103. [Google Scholar] [CrossRef]
  47. Zhu, W.; Shi, B.; Feng, Z. A Transfer Learning Method Using High-quality Pseudo Labels for Bearing Fault Diagnosis. IEEE Trans. Instrum. Meas. 2022, 72, 3502311. [Google Scholar] [CrossRef]
  48. Gal, Y.; Ghahramani, Z. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning. In Proceedings of the 33rd International Conference on Machine Learning, PMLR, New York, NY, USA, 20–22 June 2016; pp. 1050–1059, ISSN 1938-7228. [Google Scholar]
  49. Wang, J.; Zeng, Z.; Zhang, H.; Barros, A.; Miao, Q. An hybrid domain adaptation diagnostic network guided by curriculum pseudo labels for electro-mechanical actuator. Reliab. Eng. Syst. Saf. 2022, 228, 108770. [Google Scholar] [CrossRef]
  50. Chen, P.; Zhao, R.; He, T.; Wei, K.; Yuan, J. Unsupervised structure subdomain adaptation based the Contrastive Cluster Center for bearing fault diagnosis. Eng. Appl. Artif. Intell. 2023, 122, 106141. [Google Scholar] [CrossRef]
  51. Hu, Y.; Baraldi, P.; Di Maio, F.; Zio, E. A Systematic Semi-Supervised Self-adaptable Fault Diagnostics approach in an evolving environment. Mech. Syst. Signal Process. 2017, 88, 413–427. [Google Scholar] [CrossRef]
  52. Tao, X.; Ren, C.; Li, Q.; Guo, W.; Liu, R.; He, Q.; Zou, J. Bearing defect diagnosis based on semi-supervised kernel Local Fisher Discriminant Analysis using pseudo labels. ISA Trans. 2021, 110, 394–412. [Google Scholar] [CrossRef]
  53. Zhang, X.; Su, Z.; Hu, X.; Han, Y.; Wang, S. Semisupervised Momentum Prototype Network for Gearbox Fault Diagnosis Under Limited Labeled Samples. IEEE Trans. Ind. Informatics 2022, 18, 6203–6213. [Google Scholar] [CrossRef]
  54. Long, J.; Chen, Y.; Yang, Z.; Huang, Y.; Li, C. A novel self-training semi-supervised deep learning approach for machinery fault diagnosis. Int. J. Prod. Res. 2022, 61, 8238–8251. [Google Scholar] [CrossRef]
  55. Li, Y.J.; Dai, X.; Ma, C.Y.; Liu, Y.C.; Chen, K.; Wu, B.; He, Z.; Kitani, K.; Vajda, P. Cross-Domain Adaptive Teacher for Object Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; p. 10. [Google Scholar]
  56. Tarvainen, A.; Valpola, H. Mean teachers are better role models: Weight-averaged consistency targets improve semi-supervised deep learning results. arXiv 2018, arXiv:1703.01780. [Google Scholar] [CrossRef]
  57. Laine, S.; Aila, T. Temporal Ensembling for Semi-Supervised Learning. arXiv 2017, arXiv:1610.02242. [Google Scholar] [CrossRef]
  58. Liu, Y.C.; Ma, C.Y.; He, Z.; Kuo, C.W.; Chen, K.; Zhang, P.; Wu, B.; Kira, Z.; Vajda, P. Unbiased Teacher for Semi-Supervised Object Detection. arXiv 2021, arXiv:2102.09480. [Google Scholar] [CrossRef]
  59. Niculescu-Mizil, A.; Caruana, R. Predicting good probabilities with supervised learning. In Proceedings of the 22nd International Conference on Machine Learning—ICML’05, Bonn, Germany, 7–11 August 2005; pp. 625–632. [Google Scholar] [CrossRef]
  60. Naeini, M.P.; Cooper, G.F.; Hauskrecht, M. Obtaining Well Calibrated Probabilities Using Bayesian Binning. AAAI Conf. Artif. Intell. 2015, 2015, 2901–2907. [Google Scholar]
  61. Hebbalaguppe, R.; Prakash, J.; Madan, N.; Arora, C. A Stitch in Time Saves Nine: A Train-Time Regularizing Loss for Improved Neural Network Calibration. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 16060–16069, ISSN 2575-7075. [Google Scholar] [CrossRef]
  62. Ovadia, Y.; Fertig, E.; Ren, J.; Nado, Z.; Sculley, D.; Nowozin, S.; Dillon, J.V.; Lakshminarayanan, B.; Snoek, J. Can You Trust Your Model’s Uncertainty? Evaluating Predictive Uncertainty Under Dataset Shift. arXiv 2019, arXiv:1906.02530. [Google Scholar] [CrossRef]
  63. Brier, G.W. Verification of Forecasts Expressed in Terms of Probability. Mon. Weather Rev. 1950, 78, 1–3. [Google Scholar] [CrossRef]
  64. Rangwani, H.; Aithal, S.K.; Mishra, M.; Jain, A.; Babu, R.V. A Closer Look at Smoothness in Domain Adversarial Training. arXiv 2022, arXiv:2206.08213. [Google Scholar] [CrossRef]
  65. Foret, P.; Kleiner, A.; Mobahi, H. Sharpness-Aware Minimization for Efficiently Improving Generalization. In Proceedings of the ICLR, Virtual Event, 3–7 May 2021. [Google Scholar]
  66. Jin, Y.; Wang, X.; Long, M.; Wang, J. Minimum Class Confusion for Versatile Domain Adaptation. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Vedaldi, A., Bischof, H., Brox, T., Frahm, J.M., Eds.; Springer: Cham, Switzerland, 2020. Lecture Notes in Computer Science. pp. 464–480. [Google Scholar] [CrossRef]
  67. Lessmeier, C.; Kimotho, J.K.; Zimmer, D.; Sextro, W. Condition Monitoring of Bearing Damage in Electromechanical Drive Systems by Using Motor Current Signals of Electric Motors: A Benchmark Data Set for Data-Driven Classification. In Proceedings of the European Conference of the Prognostics and Health Management Society, Bilbao, Spain, 5–8 July 2016. [Google Scholar]
  68. Chen, X.; Yang, R.; Xue, Y.; Huang, M.; Ferrero, R.; Wang, Z. Deep Transfer Learning for Bearing Fault Diagnosis: A Systematic Review Since 2016. IEEE Trans. Instrum. Meas. 2023, 72, 3244237. [Google Scholar] [CrossRef]
  69. Ruan, D.; Wang, J.; Yan, J.; Gühmann, C. CNN parameter design based on fault signal analysis and its application in bearing fault diagnosis. Adv. Eng. Inform. 2023, 55, 101877. [Google Scholar] [CrossRef]
  70. Rombach, K.; Michau, G.; Fink, O. Controlled generation of unseen faults for Partial and Open-Partial domain adaptation. Reliab. Eng. Syst. Saf. 2023, 230, 108857. [Google Scholar] [CrossRef]
  71. Randall, R.B.; Antoni, J. Rolling element bearing diagnostics—A tutorial. Mech. Syst. Signal Process. 2011, 25, 485–520. [Google Scholar] [CrossRef]
  72. Demsar, J. Statistical Comparisons of Classifiers over Multiple Data Sets. J. Mach. Learn. Res. 2006, 7, 1–30. [Google Scholar]
Figure 1. Our proposed Calibrated Adaptive Teacher (CAT). The main novelty involves a post hoc calibration of the teacher predictions in the target domain throughout the self-training process, improving the quality of pseudo-labels.
Figure 1. Our proposed Calibrated Adaptive Teacher (CAT). The main novelty involves a post hoc calibration of the teacher predictions in the target domain throughout the self-training process, improving the quality of pseudo-labels.
Sensors 24 07539 g001
Figure 2. Example of reliability diagram before applying calibration (here, AT on the 0 1 task in time domain). The model has higher expected calibration error (ECE) on the target domain than on the source domain. The dotted line represents ideal calibration.
Figure 2. Example of reliability diagram before applying calibration (here, AT on the 0 1 task in time domain). The model has higher expected calibration error (ECE) on the target domain than on the source domain. The dotted line represents ideal calibration.
Sensors 24 07539 g002
Figure 3. Example of of reliability diagram after applying temperature scaling (here, CAT on the 0 1 task in time domain). Even though temperature scaling is based on the source validation set, ECE is also drastically reduced on the target domain, owing to well-aligned features. The dotted line represents ideal calibration.
Figure 3. Example of of reliability diagram after applying temperature scaling (here, CAT on the 0 1 task in time domain). Even though temperature scaling is based on the source validation set, ECE is also drastically reduced on the target domain, owing to well-aligned features. The dotted line represents ideal calibration.
Sensors 24 07539 g003
Figure 4. Evolution of target pseudo-label accuracy produced by the teacher network during training for different methods. A boost in accuracy is observed after introduction of the calibration in our proposed CAT. Mean ± standard deviation over 5 runs with different random seeds. * = with MCC + SDAT.
Figure 4. Evolution of target pseudo-label accuracy produced by the teacher network during training for different methods. A boost in accuracy is observed after introduction of the calibration in our proposed CAT. Mean ± standard deviation over 5 runs with different random seeds. * = with MCC + SDAT.
Sensors 24 07539 g004
Figure 5. Evolution of target accuracy (left) and ECE (right) during training (here, on the 0 1 task in time domain). AT significantly improves over the DANN baseline. In addition, our proposed CAT effectively reduces calibration error, leading to an improved accuracy. Mean ± standard deviation over 5 runs with different random seeds. * = with MCC + SDAT.
Figure 5. Evolution of target accuracy (left) and ECE (right) during training (here, on the 0 1 task in time domain). AT significantly improves over the DANN baseline. In addition, our proposed CAT effectively reduces calibration error, leading to an improved accuracy. Mean ± standard deviation over 5 runs with different random seeds. * = with MCC + SDAT.
Sensors 24 07539 g005
Figure 6. Example of reliability diagram after applying vector scaling calibration. Calibration improves in the source domain but not in the target domain. The dotted line represents ideal calibration.
Figure 6. Example of reliability diagram after applying vector scaling calibration. Calibration improves in the source domain but not in the target domain. The dotted line represents ideal calibration.
Sensors 24 07539 g006
Figure 7. Example of reliability diagram after applying matrix scaling calibration. Calibration improves in the source domain but not in the target domain. The dotted line represents ideal calibration.
Figure 7. Example of reliability diagram after applying matrix scaling calibration. Calibration improves in the source domain but not in the target domain. The dotted line represents ideal calibration.
Sensors 24 07539 g007
Figure 8. Comparison of target accuracy for different methods on the PU transfer tasks with time-domain input. * = with MCC + SDAT.
Figure 8. Comparison of target accuracy for different methods on the PU transfer tasks with time-domain input. * = with MCC + SDAT.
Sensors 24 07539 g008
Figure 9. Comparison of target test accuracy for different methods on the PU transfer tasks with frequency-domain input. * = with MCC + SDAT.
Figure 9. Comparison of target test accuracy for different methods on the PU transfer tasks with frequency-domain input. * = with MCC + SDAT.
Sensors 24 07539 g009
Figure 10. Critical difference diagrams of the performance of each method in terms of average rank of target accuracy on the PU transfer tasks. Methods connected by a horizontal bar are statistically equivalent (using Wilcoxon–Holm post hoc analysis at a 0.05 significance level). * = with MCC + SDAT.
Figure 10. Critical difference diagrams of the performance of each method in terms of average rank of target accuracy on the PU transfer tasks. Methods connected by a horizontal bar are statistically equivalent (using Wilcoxon–Holm post hoc analysis at a 0.05 significance level). * = with MCC + SDAT.
Sensors 24 07539 g010
Figure 11. Comparison of target-domain expected calibration error (ECE) for different methods on the PU transfer tasks with time-domain input (lower is better). * = with MCC + SDAT.
Figure 11. Comparison of target-domain expected calibration error (ECE) for different methods on the PU transfer tasks with time-domain input (lower is better). * = with MCC + SDAT.
Sensors 24 07539 g011
Table 1. Comparison of self-training unsupervised domain adaptation methods applied to intelligent fault diagnosis.
Table 1. Comparison of self-training unsupervised domain adaptation methods applied to intelligent fault diagnosis.
MethodBackboneTimeFrequencyFeature AlignmentSelf-TrainingPseudo-Label FilteringAuxiliary Loss
PCG-CNN [46]1D-CNN -MT with consistency lossFixed thresholdClass-balance loss
Wang et al. [49]1D-CNN DAWPLAdaptive thresholdTriplet loss
DTL-IPLL [47]1D-CNN MK-MMDPLAdaptive threshold + “make decision twice”-
CAT (ours)1D-CNNDANNMT with PLAdaptive threshold + calibration-
Table 2. Parameters of the CAT architecture used in this study.
Table 2. Parameters of the CAT architecture used in this study.
ModuleLayersParametersNotation
1D-CNN backboneConv1D, BatchNorm, ReLUin = 1, out = 16, kernel = 15 W E
Conv1D, BatchNorm, ReLUin = 16, out = 32, kernel = 3
MaxPool1Dkernel = 2, stride = 2
Conv1D, BatchNorm, ReLUin = 32, out = 64, kernel = 3
Conv1D, BatchNorm, ReLUin = 64, out = 128, kernel = 3
AdaptiveMaxPool1Dout = 128 × 4
FC, ReLUin = 128 × 4, out = 256
Dropoutp = 0.5
BottleneckFC, ReLUin = 256, out = 256
Dropoutp = 0.5
Classification headFC, Softmaxin = 256, out = 13 W C
Domain classifierFC, ReLUin = 256, out = 1024 W DC
Dropoutp = 0.5
FC, ReLUin = 1024, out = 1024
Dropoutp = 0.5
FC, Sigmoidin = 1024, out = 1
Table 3. Domains for the Paderborn University (PU) dataset.
Table 3. Domains for the Paderborn University (PU) dataset.
DomainRotational SpeedLoad TorqueRadial Force
01500 rpm0.7 Nm1000 N
1  900 rpm0.7 Nm1000 N
21500 rpm0.1 Nm1000 N
31500 rpm0.7 Nm  400 N
Table 4. Classification task for the Paderborn University (PU) dataset.
Table 4. Classification task for the Paderborn University (PU) dataset.
ClassBearing CodeDamageElementCombinationCharacteristic
0KA04Fatigue: pittingORSSingle point
1KA15Plastic deform: indentationsORSSingle point
2KA16Fatigue: pittingORRSingle point
3KA22Fatigue: pittingORSSingle point
4KA30Plastic deform: indentationsORRDistributed
5KB23Fatigue: pittingIR (+OR)MSingle point
6KB24Fatigue: pittingIR (+OR)MDistributed
7KB27Plastic deform: indentationsOR+IRMDistributed
8KI14Fatigue: pittingIRMSingle point
9KI16Fatigue: pittingIRSSingle point
10KI17Fatigue: pittingIRRSingle point
11KI18Fatigue: pittingIRSSingle point
12KI21Fatigue: pittingIRSSingle point
Table 5. Target-domain test accuracy on the different PU transfer tasks with time-domain input. Average teacher accuracy over 5 runs (values in %).
Table 5. Target-domain test accuracy on the different PU transfer tasks with time-domain input. Average teacher accuracy over 5 runs (values in %).
Method 0 1 0 2 0 3 1 0 1 2 1 3 2 0 2 1 2 3 3 0 3 1 3 2 AverageAverage Rank
Source-only 14.0276.3330.0223.5724.1816.0976.7314.7131.2332.1625.2732.3933.06-
DANN 38.1979.9753.7435.4239.5727.0579.2036.5349.2347.9327.4547.5746.82-
Source-only15.1578.2330.2924.3325.7117.7376.4113.8733.3431.9226.3232.1233.787.90
DANN36.9681.2251.5635.5540.8928.3578.7733.8347.2947.5629.9446.4146.535.49
DANN*36.5383.5454.4635.6138.5325.7581.4436.5349.8647.2528.5046.2047.025.08
AT*49.3682.1756.1647.8350.6634.9281.7249.1151.8050.0829.9153.8653.133.26
CAT*-TempScaling52.6782.4457.0051.2753.3436.3481.7851.4454.4049.2230.3753.7454.502.28
CAT*-CPCS52.4282.2957.2551.2452.7636.4681.6951.6354.2549.4029.8853.7754.422.40
CAT*-VectorScaling34.5182.0850.9543.6645.5334.1082.0340.1848.3547.9935.9246.7549.344.15
CAT*-MatrixScaling33.6582.0848.2039.9744.5832.0480.9535.3746.1444.7331.0143.4846.855.43
= results from [3] (Last-Mean). * = with MCC + SDAT.
Table 6. Target-domain test accuracy on the different PU transfer tasks with frequency-domain input. Average teacher accuracy over 5 runs (values in %).
Table 6. Target-domain test accuracy on the different PU transfer tasks with frequency-domain input. Average teacher accuracy over 5 runs (values in %).
Method 0 1 0 2 0 3 1 0 1 2 1 3 2 0 2 1 2 3 3 0 3 1 3 2 AverageAverage Rank
Source-only 20.9690.8757.1425.1324.1413.7586.4020.5557.1852.5820.9053.6443.60-
DANN 40.3493.7182.1528.4835.2722.8892.5046.0179.5268.7624.5776.1557.53-
Source-only21.1392.1258.3425.6821.9214.9287.4720.8359.7650.8120.7154.7544.047.98
DANN40.0995.2779.5828.4535.8822.1592.0746.4478.0371.9832.0972.6157.895.76
DANN*42.3096.5880.6430.4137.0731.9893.9850.0079.6469.8332.9872.7359.844.63
AT*52.5596.1281.0941.4150.4439.8593.8656.5079.0369.2838.5368.6463.943.80
CAT*-TempScaling52.9896.2781.5744.4550.8743.3693.8956.9979.0670.1138.2269.3464.762.76
CAT*-CPCS53.4796.1881.6043.9650.6044.3393.8956.9379.5870.0538.7169.7764.922.76
CAT*-VectorScaling50.9296.4081.2443.5348.4041.9794.1055.1878.6169.2537.5569.5063.893.64
CAT*-MatrixScaling49.9795.8280.8541.4147.5736.8593.9852.6477.9468.7337.5569.2862.724.67
= results from [3] (Last-Mean). * = with MCC + SDAT.
Table 7. Target-domain expected calibration error (ECE) on the different PU transfer tasks with time-domain input. Average teacher ECE over 5 runs (values in %; lower is better).
Table 7. Target-domain expected calibration error (ECE) on the different PU transfer tasks with time-domain input. Average teacher ECE over 5 runs (values in %; lower is better).
Method 0 1 0 2 0 3 1 0 1 2 1 3 2 0 2 1 2 3 3 0 3 1 3 2 AverageAverage Rank
Source-only78.9017.4162.4466.6066.1174.2718.2479.2058.1556.8662.8657.1658.187.85
DANN55.7014.8041.8055.4149.5562.5116.2957.5845.8143.5559.9845.2545.696.13
DANN*60.1714.8142.8959.5756.7368.9716.8659.7347.2448.8366.3349.8349.336.88
AT*41.5112.6034.3338.9637.0448.4612.9038.0737.2732.3449.3028.2934.264.15
CAT*-TempScaling14.163.9512.306.649.5418.106.2415.4915.8414.3927.758.8712.771.67
CAT*-CPCS21.096.3610.768.346.7719.284.8912.4411.636.9417.1910.0511.311.48
CAT*-VectorScaling49.195.6531.1426.6725.5531.057.4844.3135.4134.6938.8337.7830.653.35
CAT*-MatrixScaling54.388.0336.6227.2525.9732.5511.2948.8141.3039.7847.1642.6034.644.48
* = with MCC + SDAT.
Table 8. Comparison between fixed and adaptive confidence thresholds in Adaptive Teacher (AT). The results on PU dataset with time-domain input. Average teacher accuracy on target test set over 5 runs (values in %).
Table 8. Comparison between fixed and adaptive confidence thresholds in Adaptive Teacher (AT). The results on PU dataset with time-domain input. Average teacher accuracy on target test set over 5 runs (values in %).
MethodThreshold 0 1 0 2 0 3 1 0 1 2 1 3 2 0 2 1 2 3 3 0 3 1 3 2 Average
ATFixed33.4480.9549.5340.9839.3923.0079.7536.6044.2150.2329.2047.3346.22
ATAdaptive38.7780.8952.5041.6344.3426.2978.8642.8845.9049.0328.9951.2148.33
Table 9. Ablation study on MCC and SDAT in Adaptive Teacher (AT) and our proposed Calibrated Adaptive Teacher (CAT) method. The results on PU dataset with time-domain input. Average teacher accuracy on target test set over 5 runs (values in %).
Table 9. Ablation study on MCC and SDAT in Adaptive Teacher (AT) and our proposed Calibrated Adaptive Teacher (CAT) method. The results on PU dataset with time-domain input. Average teacher accuracy on target test set over 5 runs (values in %).
MethodMCCSDAT 0 1 0 2 0 3 1 0 1 2 1 3 2 0 2 1 2 3 3 0 3 1 3 2 Average
AT 38.7780.8952.5041.6344.3426.2978.8642.8845.9049.0328.9951.2148.33
42.7780.9252.8943.5346.7528.3880.8343.8746.4451.3432.7351.4550.04
45.9882.3254.9845.2846.4130.8981.0448.7150.8044.1826.4450.0550.59
49.3682.1756.1647.8350.6634.9281.7249.1151.8050.0829.9153.8653.13
CAT - TempScaling 44.5881.4053.9547.0446.9927.8778.9244.8248.1449.5928.9952.1850.28
46.5681.3455.1648.9150.1432.4481.1746.3848.3550.6033.2253.0752.28
49.1482.4156.5847.5049.9232.4781.0150.0052.3444.7023.1950.1451.62
52.6782.4457.0051.2753.3436.3481.7851.4454.4049.2230.3753.7454.50
CAT - CPCS 43.9081.5353.8646.1846.4427.3579.1744.2948.5349.6230.1552.4050.28
46.8781.3454.8948.8250.3832.9581.1746.7848.5951.5833.2852.8952.46
49.8282.3857.1947.3749.6530.9881.0151.2352.6845.1923.4049.7151.72
52.4282.2957.2551.2452.7636.4681.6951.6354.2549.4029.8853.7754.42
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Forest, F.; Fink, O. Calibrated Adaptive Teacher for Domain-Adaptive Intelligent Fault Diagnosis. Sensors 2024, 24, 7539. https://doi.org/10.3390/s24237539

AMA Style

Forest F, Fink O. Calibrated Adaptive Teacher for Domain-Adaptive Intelligent Fault Diagnosis. Sensors. 2024; 24(23):7539. https://doi.org/10.3390/s24237539

Chicago/Turabian Style

Forest, Florent, and Olga Fink. 2024. "Calibrated Adaptive Teacher for Domain-Adaptive Intelligent Fault Diagnosis" Sensors 24, no. 23: 7539. https://doi.org/10.3390/s24237539

APA Style

Forest, F., & Fink, O. (2024). Calibrated Adaptive Teacher for Domain-Adaptive Intelligent Fault Diagnosis. Sensors, 24(23), 7539. https://doi.org/10.3390/s24237539

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop