Next Article in Journal
Motion Capture Technologies for Athletic Performance Enhancement and Injury Risk Assessment: A Review for Multi-Sport Organizations
Previous Article in Journal
Using Large Languge Models for Processing Sensor Data
Previous Article in Special Issue
A Digital Signal Processing-Based Multi-Channel Acoustic Emission Acquisition System with a Simplified Analog Front-End
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Source-Free Domain Adaptation Framework for Rotary Machine Fault Diagnosis

1
Department of Electrical and Computer Engineering, Inha University, Incheon 22212, Republic of Korea
2
Department of Computer Engineering, Inha University, Incheon 22212, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(14), 4383; https://doi.org/10.3390/s25144383
Submission received: 16 June 2025 / Revised: 1 July 2025 / Accepted: 10 July 2025 / Published: 13 July 2025
(This article belongs to the Special Issue Sensor Data-Driven Fault Diagnosis Techniques)

Abstract

Intelligent fault diagnosis for rotary machinery often suffers performance degradation under domain shifts between training and deployment environments. To address this, we propose a robust fault diagnosis framework incorporating three key components: (1) an order-frequency-based preprocessing method to normalize rotational variations, (2) a U-Net variational autoencoder (U-NetVAE) to enhance adaptation through reconstruction learning, and (3) a test-time training (TTT) strategy enabling unsupervised target domain adaptation without access to source data. Since existing works rarely evaluate under true domain shift conditions, we first construct a unified cross-domain benchmark by integrating four public datasets with consistent class and sensor settings. The experimental results show that our method outperforms conventional machine learning and deep learning models in both F1-score and recall across domains. Notably, our approach maintains an F1-score of 0.47 and recall of 0.51 in the target domain, outperforming others under identical conditions. Ablation studies further confirm the contribution of each component to adaptation performance. This study highlights the effectiveness of combining mechanical priors, self-supervised learning, and lightweight adaptation strategies for robust fault diagnosis in the practical domain.

1. Introduction

Rotating machinery plays a vital role across various industrial sectors such as manufacturing, power generation, and transportation, serving as key equipment for transmitting power and converting energy. Due to their continuous rotational motion, components such as shafts, bearings, and gears are prone to wear, imbalance, and mechanical looseness. If such anomalies are not detected early, they may lead to equipment failure, production downtime, or even safety accidents. Therefore, fault diagnosis technologies capable of the real-time monitoring and early detection of abnormal behavior are essential to ensure the reliability and productivity of industrial sites.
Vibration signals directly reflect changes in the physical condition of internal components and are capable of capturing fault indications through variations in energy distribution across specific frequency bands. For this reason, vibration signals have been established as a standard input for rotary machinery diagnostics, and numerous studies have reported high classification accuracy for various fault types using these signals [1,2].
Traditional vibration-based fault diagnosis typically involves extracting handcrafted features using signal processing techniques such as FFT (Fast Fourier Transform), EMD (Empirical Mode Decomposition), or Order Tracking, followed by classification using conventional machine learning algorithms such as SVM or decision trees [3,4]. While such approaches can achieve high performance under specific conditions, they heavily depend on expertly designed features and are vulnerable to performance degradation when sensor positions or environmental conditions change. To overcome these limitations, recent studies have introduced deep learning-based models such as CNNs and RNNs, which automate feature extraction and improve classification performance through label-based training [5,6,7].
However, most existing deep learning approaches rely on single-input classification architectures, which often result in overfitting to specific training domains without learning sufficiently generalized representations. To address this structural limitation, we propose a self-supervised framework that reconstructs the input vibration signal to learn a generalized representation and then classifies anomalies by comparing the latent representations of the input and reference signals. This framework goes beyond conventional classification by interpreting the structural differences between input and reference data in latent space, enabling more robust fault diagnosis under domain shifts.
Furthermore, to improve reconstruction accuracy, we designed an encoder–decoder architecture based on U-Net [8] instead of a conventional CNN-based autoencoder. U-Net’s skip connections help preserve multi-resolution features, allowing for the finer reconstruction of frequency details and more precise latent representation learning. This architecture supports self-supervised training even in the absence of labels by minimizing reconstruction loss, thereby providing adaptability to target domains without additional annotation. Since, in many industrial settings, access to source domain data is restricted and labeled data from the target domain is often unavailable, we adopt a Source-Free Domain Adaptation (SFDA) strategy and verify its effectiveness experimentally.
The main contributions of this paper are as follows: (1) We propose an order-frequency-based preprocessing method that normalizes rotational variations across machines, enabling unified representation learning in cross-domain vibration diagnostics. (2) We develop a self-supervised anomaly classification framework based on a U-Net-structured variational autoencoder, which reconstructs input signals and compares latent representations with reference signals for robust fault detection. (3) We introduce a test-time training strategy tailored for Source-Free Domain Adaptation, allowing our model to adapt to unlabeled target domains without access to source data. A unified benchmark of four public datasets is constructed to validate performance under real-world domain shift conditions.

2. Related Work

2.1. Rotary Machine Fault Diagnosis

Faults in rotary machinery are directly linked to equipment downtime, product quality degradation, and financial losses. As such, fault diagnosis technologies that enable early detection and response have played a critical role in industrial applications. In particular, vibration signals have been widely utilized as key indicators for diagnosing the condition of rotating equipment, as they quantitatively capture subtle dynamic variations. Since many fault symptoms appear as patterns in the frequency or time–frequency domain, vibration-based diagnostics have proven highly effective.
Traditional fault diagnosis approaches initially relied on physics-based models or statistical process control techniques to assess the health of rotating machinery. For instance, Farrar et al. [9] proposed a method based on statistical pattern recognition, assuming that minor changes in structural dynamic responses reflect the presence and severity of damage. Their method involved training an autoregressive (AR) model on undamaged signals, then monitoring the mean and variance of the residuals using control charts for damage detection. While such approaches provide interpretable and quantitative diagnostic criteria, they are often sensitive to nonlinearities and external factors such as load variations and temperature changes, limiting their generalizability in real-world operating environments.
With the introduction of machine learning techniques, diagnostic pipelines began to incorporate signal processing methods such as FFT and EMD for feature extraction, followed by classifiers like SVM or ANFIS [10]. For example, Lei et al. [11] used EMD to extract intrinsic mode functions, selected meaningful features using an enhanced distance metric, and fed them into multiple ANFIS classifiers for fault prediction. They further improved classification performance under multiple fault conditions by combining the outputs using a genetic algorithm. Although these methods improved automation and accuracy over statistical approaches, they still required manual feature selection and were prone to performance degradation when applied to different operating environments.
More recently, deep learning-based methods have gained prominence due to their ability to extract high dimensional representations directly from raw vibration signals without explicit preprocessing or feature engineering. Zhang et al. [12] proposed the WDCNN model, which uses wide kernels in the initial layers to suppress high-frequency noise and smaller kernels in subsequent layers to capture fault-specific patterns. This end-to-end approach achieved high classification accuracy and demonstrated robust performance even on real industrial datasets.
Vibration signals are inherently non-stationary and time-varying, so raw time-series inputs cannot capture consistent fault features across speed and load changes. Shao et al. [13] showed that STFT-based time–frequency spectra provide stable fault patterns under varying conditions. Aiordăchioaie [14] further demonstrated that an FFT-based power spectrum achieves higher classification accuracy rather than time-domain methods. However, even frequency-domain inputs shift with RPM variation. Wang et al. [15] used order analysis to normalize rotational speed effects and produce order-domain features that outperform both raw and standard frequency inputs under variable speeds. Yu et al. [16] then resolved the spectral blur of conventional FFT-order spectra by using mean envelope amplitudes to yield high-resolution, noise-robust order spectra. In line with these findings, we apply an FFT followed by an order transform to supply our model with speed-invariant, high-resolution frequency representations.

2.2. Self-Supervised Feature Extraction via VAE

Vibration signals measured from rotary machinery vary in distribution depending on external factors such as rotational speed, load conditions, sensor position, and ambient temperature. Even for the same type of fault, the signal may appear differently under different conditions. To perform reliable diagnostics under such complex variations, it is essential to learn a universal representation that effectively captures the inherent structural characteristics of the input data.
In recent studies, self-supervised learning approaches that reconstruct the input signal itself have been actively explored to overcome the limitations of supervised classification models [17,18,19]. In particular, the VAE has been regarded as a highly suitable architecture for this purpose.
The VAE [20] embeds the input into a probabilistic latent space and reconstructs it based on the sampled latent representation. It is known to learn more generalized representations than a standard autoencoder. Bengio et al. [21] experimentally demonstrated that AE/VAE-based reconstruction learning can provide meaningful representations for downstream tasks, and emphasized the mathematical superiority of VAE in learning the overall data distribution rather than merely compressing features. Since the VAE can learn structured representations of input data without supervision, it is advantageous for building models that are robust to various domain conditions.
Figure 1 visually illustrates the difference between a standard autoencoder (AE) and a variational autoencoder (VAE). While an AE compresses and reconstructs the input deterministically, a VAE maps the input into a latent space modeled as a Gaussian distribution and reconstructs it through sampling. This probabilistic structure improves both the diversity and generalizability of the learned representations.
VAEs have been successfully applied not only in image domains but also in sensor-based time-series data. In particular, they have been utilized in temporal representation learning as an effective method to capture essential patterns within input sequences. For example, when applied to one-dimensional time-series data such as vibration signals, VAE-based reconstruction enables the model to learn high-dimensional representations that include regions containing critical fault information. These representations can be effectively used for downstream tasks such as classification and anomaly detection [22,23].
Recent studies have further extended this reconstruction-based representation learning to multi task structures, where reconstruction and classification are trained simultaneously [24,25]. By jointly optimizing both tasks, the model is encouraged to capture the structural characteristics of the input beyond relying solely on label information, thereby improving generalization under domain shift. This learning paradigm is particularly effective in test-time training (TTT) and Source-Free Domain Adaptation (SFDA) scenarios, where adaptation must be performed without labeled data.
In this study, we adopt such a self-supervised representation learning framework. Our model reconstructs vibration frequency data using a VAE, while simultaneously learning generalized latent vectors. Anomalies are then classified by comparing the latent representation of the input with that of a reference (normal) state. The reconstruction loss is optimized without any labels, and the classification based on latent differences enables robust fault diagnosis even under domain shifts. In this study, we adopt a standard Gaussian prior for the VAE latent space. While alternative priors such as mixture models or β -VAE can offer more expressive latent distributions, they typically require access to additional information such as class labels or structural attributes. However, our framework is designed for source-free and unsupervised domain adaptation, where neither source data nor target labels are available during adaptation. In this setting, the use of a standard normal prior is the most practical and stable choice, allowing the model to learn transferable latent features without relying on any domain-specific assumptions.
Our model reconstructs vibration frequency data using a VAE with a standard Gaussian prior in the latent space, which enables it to learn generalized representations without labels. This choice is particularly suitable for our source-free and unsupervised setting, where additional information required by more expressive priors (e.g., β -VAE [26]) is unavailable. Fault conditions are then identified by comparing the latent representation of the input with that of a reference (normal) state, allowing robust diagnosis across domains with different operating conditions.

2.3. Source-Free Domain Adaptation

Domain adaptation is a core technique for ensuring that a model performs reliably even when deployed in environments with data distributions different from those used during training. It is considered essential for real-world industrial applications. However, most existing domain adaptation methods assume simultaneous access to labeled source domain data and unlabeled target domain data. This assumption often fails in practice, where accessing source data is restricted due to security, privacy regulations, intellectual property, or legal constraints.
Source-Free Domain Adaptation (SFDA) has recently gained attention as a method to address this limitation. SFDA enables model adaptation using only a pre-trained source model and unlabeled target domain data, without requiring access to the original source data. This paradigm provides a practical solution that satisfies both data privacy and model portability requirements, making it particularly suitable for industrial deployment scenarios.
Figure 2 illustrates the general structure of Source-Free Domain Adaptation (SFDA). A pre-trained model consisting of a feature extractor and classifier is kept frozen or partially adaptive, and adaptation is performed using only unlabeled data from the target domain. This process aims to either reduce the distributional gap or reorganize the representation space to better align with the target data. Common approaches include entropy minimization with dispersion regularization to sharpen decision boundaries, or iterative refinement using soft pseudo-labels generated via k-means clustering [27]. Other methods introduce self-supervised auxiliary tasks, such as predicting image rotation or position, to guide adaptation to the new distribution [28], or apply contrastive learning to preserve feature distances between augmented samples during domain shifts [29].
While these techniques originated mainly in the image domain, recent studies have begun to leverage reconstruction-based self-supervised learning to align cross-domain representations at test time, even without labels [30]. In addition, student–teacher frameworks have been introduced to improve prediction stability and enable gradual adaptation without relying heavily on pseudo-labels [31]. Collectively, these efforts reflect an evolution of SFDA toward preserving classification performance while removing the dependency on labeled or source data.
However, applying SFDA to time-series data where sequential continuity and temporal dependency are critical presents additional challenges. Unlike images, the temporal structure often varies significantly across domains, making simple alignment strategies insufficient. To address this, some recent work has proposed masking parts of the time series and training the model to reconstruct them, thereby learning the temporal structure of the target domain in a self-supervised manner [32]. Other approaches train autoregressive interpolation models on the source domain and enforce temporal consistency when applied to the target domain [33]. In time-series image-based domain adaptation, methods have also been proposed to balance class separability and information preservation while aligning cross-domain features [34].
In addition to approaches that rely solely on deep learning, there has been a growing interest in hybrid frameworks that incorporate physical signal preprocessing and domain knowledge into the domain adaptation process. For example, ref. [35] explores the fusion of real data acquired from an actual system with synthetic data generated from a physical model, generating synthetic data for every possible condition. Ref. [36] explores using simulation-based training data to overcome the challenges of limited real-world fault data for bearing fault detection, enabling machine learning models to be trained for fault detection even in the absence of sufficient labeled fault data. Furthermore, ref. [37] proposes a zero-fault-shot learning framework that combines physical preprocessing with model-based approaches to achieve robust domain adaptation for bearing spall type classification. This method demonstrates that integrating domain-specific physical priors with machine learning models can significantly improve adaptation performance.
In this study, we extend existing time-series SFDA approaches by constructing a fault diagnosis framework that combines reconstruction-based self-supervised learning with latent space comparison. Specifically, we reconstruct the representation space and classify anomalies based on the latent difference between the input and a reference (normal) signal, enabling effective adaptation to the target domain without labels.

3. Method

3.1. Mechanically Informed Order Spectrum Preprocessing

Vibration signals measured from rotary machinery exhibit distinctive frequency-domain patterns when mechanical faults or structural anomalies occur. In particular, depending on the type of fault, energy tends to concentrate at integer multiples of the fundamental rotational frequency. Since these patterns are dependent on rotational speed, a change in speed results in a shift in the fault-related frequencies, making consistent fault identification across varying operating conditions difficult. This variability becomes a key reason why deep learning models may learn the same fault under different distributions, ultimately hindering generalization.
To address this issue, the mechanical engineering community has traditionally employed a technique known as order analysis. Order analysis normalizes the frequency axis based on the machine’s rotational speed (RPM), aligning fault features at consistent order indices regardless of operating speed. This approach has long been validated in industrial settings and is considered a standard frequency-domain diagnostic method. Compared to conventional FFT-based analysis, order analysis provides superior consistency in diagnosing the same fault under variable speed conditions.
Despite its effectiveness, most recent deep learning-based diagnostic studies have bypassed such physically grounded preprocessing steps, instead relying solely on raw time-domain signals or standard transformations such as FFT [38], Short-Time Fourier Transform (STFT) [39], or wavelet transform (WT) [40] for end-to-end learning. Consequently, models become highly sensitive to variations in rotational speed, resulting in poor generalization when deployed across different domains. To overcome this limitation, this study incorporates an order transform-based preprocessing method grounded in mechanical principles into the deep learning pipeline.
Although order analysis has also been applied in traditional machine learning-based diagnostic systems, it has typically involved extracting the amplitudes of specific harmonic components (e.g., 1×, 2×, 3×) and feeding them into classifiers such as SVMs or Random Forests [41,42,43]. While effective at summarizing energy in key fault-related frequencies, such approaches inherently discard valuable information present in other spectral regions. They are also limited in their ability to capture latent fault patterns and noise variations, thereby constraining model expressiveness.
In contrast, this study proposes using the entire waveform within the order domain rather than only selected components as the input to the deep learning model. By feeding the full order-domain spectrum to the model, we preserve a wide range of frequency-based characteristics critical for diagnosis, including relative energy distribution across orders, trend shifts, and local patterns. This richer representation enables the model to learn more expressive and discriminative features. The input range is configured to include up to 10 times the fundamental rotational order (10×), thereby maximizing the coverage of physically meaningful frequency bands.
The preprocessing pipeline begins by applying FFT to the input time-series signal. Based on the machine’s rotational speed, the fundamental frequency is calculated and used to convert the frequency axis into the order domain. Because this normalized spectrum aligns fault features along a fixed X-axis regardless of speed, it is well suited for use with the heterogeneous multi-machine, multi-condition dataset used in this study. In other words, by normalizing frequency information across datasets collected from different machines, we enable unified representation learning that is highly beneficial for domain adaptation.
Subsequently, log scaling is applied to control the amplitude’s dynamic range, and MinMax normalization scales the data into the [0, 1] range, making it suitable for neural network input. Finally, a reference signal randomly sampled from normal condition data collected from the same machine is paired with the input after undergoing the same preprocessing steps. This paired data is then used for self-supervised classification based on latent representation differences. The full pseudocode of the preprocessing algorithm is provided in Algorithm 1.
Algorithm 1 Signal and reference preprocessing pipeline
Require: Input signal  x ( t ) , Rotational speed r ^ , Machine ID m
Ensure: Normalized order-domain pair:  ( X ˜ input , X ˜ ref )
 1:
 function Preprocess( x ( t ) , r ^ )
 2:
     // Step 1: FFT
 3:
      X ( f ) FFT ( x ( t ) )
 4:
     // Step 2: Order Transform
 5:
      f 0 r ^ / 60                           ▹ Fundamental frequency
 6:
      o f / f 0                         ▹ Convert to order domain
 7:
      X order ( o ) X ( f )
 8:
     // Step 3: Log Scaling
 9:
      X log ( o ) log ( 1 + | X order ( o ) | )
10:
    // Step 4: Min-Max Normalization
11:
     X ˜ ( o ) X log ( o ) min ( X log ) max ( X log ) min ( X log )
12:
    return  X ˜ ( o )
13:
end function
14:
function FullPipeline( x ( t ) , r ^ , m )
15:
     X ˜ input P R E P R O C E S S ( x ( t ) , r ^ )
16:
    // Step 5: Reference Search
17:
    Randomly select x ref ( t ) from healthy data of machine m
18:
     X ˜ ref P R E P R O C E S S ( x ref ( t ) , r ^ )
19:
    return  ( X ˜ input , X ˜ ref )
20:
end function
Figure 3 summarizes the overall flow of the preprocessing pipeline. The input raw signal is first transformed into the frequency domain via FFT, then realigned to the order domain using the rotational speed. After scaling and reference pairing, the signal is converted into a unified input format that allows integrated modeling across different machines and operating conditions.

3.2. Latent Representation Learning via U-NetVAE with Difference-Aware Classification

The deep learning-based diagnostic model proposed in this study does not simply classify input data. Instead, it quantifies the difference between the input and a normal (reference) signal within an interpretable latent space, and determines anomaly status based on this difference. This framework integrates a VAE structure that learns generalized representations through reconstruction and a classification mechanism that operates on the difference between latent vectors, forming a multi-task learning architecture.
The proposed model, illustrated in Figure 4, consists of three main sub-modules forming a unified architecture: (1) a U-NetVAE encoder that maps the input signal into a latent Gaussian space and captures generalized representations, (2) a decoder that reconstructs the original input from the latent vector to reinforce reconstruction learning, and (3) a classifier that receives the difference between the latent vectors of the input and the reference signal to determine the fault class. This overall pipeline enables both representation learning and classification in a self-supervised manner.
As shown in Figure 5, each encoder and decoder block is constructed using 1D-convolutional residual modules, where skip connections preserve both local and global information throughout the network. The residual blocks include a dimension expansion and contraction process to maximize feature capacity while maintaining sequence structure.
Figure 6 shows the full structure of the U-NetVAE, where the encoder compresses the input sequence and the decoder reconstructs it back to the original shape using transposed convolutions. Skip connections between matching levels ensure feature reuse and reconstruction fidelity.
After feature extraction, the encoder samples from a Gaussian distribution to generate a latent representation z, and incorporates a distribution loss based on KL divergence to ensure the latent space maintains a generalized form. This loss is formulated as in Equation (1).
L distribution = KL ( q ( z x input ) p ( z ) ) = q ( z x input ) log q ( z x input ) p ( z ) d z = q ( z f enc ( x input ) ) log q ( z f enc ( x input ) ) p ( z ) d z p ( z ) N ( 0 , I )
The decoder mirrors the encoder structure and reconstructs the signal into the original 1D sequence using 1D transpose convolutions. The reconstruction loss is defined using the Huber loss, which provides a balanced trade-off between the mean-squared error (MSE) and mean absolute error (MAE), offering both training stability and robustness to noise. The reconstruction loss is given by Equation (2).
L recon = 1 2 ( x input x recon ) 2 if | x input x recon | δ δ · | x input x recon | 1 2 δ 2 otherwise
The classifier receives the difference vector between z input and z ref as the input, and performs classification using a Support Vector Machine (SVM). SVMs are known to perform effectively in margin-based classification tasks in high-dimensional feature spaces, making them well suited for comparison-based decision problems. In this work, the structured nature of the latent difference vector enables reliable anomaly detection using this approach. The classifier is trained using focal loss to handle class imbalance and improve training stability. The focal loss is defined as in Equation (3).
L classify = α t ( 1 p pred ) γ log ( p pred ) if y = 1 α t ( p pred ) γ log ( 1 p pred ) if y = 0
The overall loss function is formulated as a multi-task objective that includes reconstruction loss, distribution regularization loss, and classification loss. To account for instability in the early stages of training, a beta annealing strategy is employed, in which the weight on the KL-divergence term gradually increases as training progresses. The total loss function is defined as in Equation (4).
L total = β ( t ) · L distribution + λ recon · L recon + λ cls · L classify β ( t ) = t T · β max if t T β max otherwise

3.3. Self-Supervised Test-Time Training for Domain Adaptation

To enable effective adaptation to the target domain in scenarios where source domain data is unavailable, this study adopts a self-supervised learning strategy based on test-time training (TTT). TTT adapts to unlabeled target data by updating only selected modules of a pre-trained model. In our framework, TTT is specifically designed to enhance the representation learning capability of the VAE encoder. The adaptation process consists of two stages, assuming that no labels are available in the target domain after the model has been trained on the source domain.

3.3.1. Step 1: Self-Supervised Encoder Adaptation

For each target domain sample x, two augmented versions are created by adding Gaussian noise at different intensities: a weakly augmented version x weak and a strongly augmented version x strong . Each of these inputs, along with the original x, is passed through the encoder and decoder to obtain the reconstructed outputs x ^ , x ^ weak , and x ^ strong . Based on these, three reconstruction-based losses are defined as shown in Equations (5)–(7).
  • Reconstruction Loss: Loss between original input x and reconstruction x ^ :
    L recon = 1 n i = 1 n ( x i x ^ i ) 2
  • Augmentation Loss: Loss between reconstructed weak augmentation x weak and original input x:
    L aug = 1 n i = 1 n ( x i x ^ weak , i ) 2
  • Consistency Loss: Loss between reconstructed weak augmentation x weak and reconstructed strong augmentation x strong :
    L consistency = 1 n i = 1 n ( x ^ weak , i x ^ strong , i ) 2
In addition, the KL-divergence loss between the Gaussian latent distributions is computed for each reconstruction and defined as
L KL = D KL origin + D KL weak + D KL strong
The total loss function is then composed as a weighted sum of the three reconstruction losses and the KL-divergence term, as shown in Equation (9).
L total = L recon + L aug + L consistency + λ KL · L KL , λ KL = 0.1
During this adaptation process, the decoder and classifier are kept frozen, and only the encoder is updated. This design preserves the reconstruction baseline learned from the source domain while allowing the encoder to incorporate structural characteristics specific to the target domain. As a result, the encoder learns a new representation z that is specialized for the target domain while remaining structurally comparable to the original source-based representation.

3.3.2. Step 2: Latent Fusion and Classification

The final classification input is obtained by averaging the latent representation z extracted from the encoder before adaptation and the updated representation z obtained after adaptation. This approach effectively integrates the generality of the source domain representation with the specificity of the adapted representation, enabling robust fault diagnosis while mitigating the effects of domain shift in the target domain.
In addition, this averaging can be viewed as a convex combination that explores the intermediate representation in the latent space between the source and target domains, providing an effective feature representation under moderate distribution shifts. This allows the model to retain the general knowledge from the source domain while incorporating instance-specific information from the target domain in a stable and computationally efficient manner.

4. Experiment

4.1. Dataset Integration

Conventional fault diagnosis models for rotary machinery are typically trained and validated using data collected under a single experimental setup or specific operating conditions. However, such a setup fails to reflect the discrepancies between controlled laboratory environments and the complexities of real-world industrial operations. For instance, vibration signals acquired from laboratory-based machinery are collected under precisely controlled rotational speeds, load conditions, and structural configurations. In contrast, equipment operating in industrial environments is subject to various environmental factors and unpredictable fault scenarios. These differences introduce structural domain shifts that can lead to a significant drop in model performance when a model trained in the laboratory is deployed in the field.
To address this issue, we construct an experimental setup that enables the quantitative evaluation of domain adaptation performance by separating the training and testing datasets across different domains. By explicitly distinguishing between datasets used for training and those used for testing, each reflecting different equipment types and operating conditions, we aim to assess the robustness of fault diagnosis models in settings that simulate real-world deployment scenarios.
To this end, we integrate four publicly available fault diagnosis datasets collected under diverse conditions. Specifically, the Vibration and Motor Current Dataset of Rolling Element Bearing Under Varying Speed Conditions for Fault Diagnosis (VAT), the Mechanical Faults in Rotating Machinery Dataset (DXAI), the VBL-VA001 dataset, and the MaFaulDa dataset are selected. These datasets include major fault conditions commonly found in rotary machinery, making them appropriate for our experimental objectives. Since these datasets differ in terms of their sensor configurations, measurement axes, class definitions, rotational speeds, and fault scenarios, we standardize them into a unified input format and class structure. This preprocessing ensures compatibility for cross-domain validation experiments and enables the evaluation of adaptation across heterogeneous environments. Table 1 and Table 2 describe the environments and other information for each dataset.

4.1.1. VAT

The VAT dataset (Figure 7), was collected by the Korea Advanced Institute of Science and Technology (KAIST) and contains multi-sensor data—including vibration, current, acoustic, and temperature signals measured from rotary machinery under various load and speed conditions. The dataset consists of two configurations. The first includes measurements under three different load conditions: 0 Nm, 2 Nm, and 4 Nm. The second includes bearing fault data (inner race, outer race, and ball faults) recorded at varying rotational speeds ranging from 680 RPM to 2460 RPM.

4.1.2. DXAI

The DXAI dataset (Figure 8), is a vibration dataset constructed by the Universidade Federal de São João del-Rei in Brazil. It includes four operating conditions: normal, unbalance, misalignment, and looseness. The data were collected under experimental conditions designed to closely resemble real industrial environments. The test bench consists of components such as a motor, an inverter, bearings, and pulleys. Each fault condition was repeated five times, with a randomized experiment order and reinitialization to ensure data diversity.

4.1.3. VBL-VA001

The VBL-VA001 dataset (Figure 9), was collected by the Sepuluh Nopember Institute of Technology in Indonesia. It contains data measured at a fixed rotational speed of 3000 RPM and includes four fault conditions: normal, misalignment, unbalance (at two severity levels), and BPFO (Ball Pass Frequency Outer race) faults. The dataset consists of four classes, each with 1000 samples. All experiments were conducted at a sampling frequency of 20 kHz.

4.1.4. MaFaulDa

The MaFaulDa dataset (Figure 10), was developed by the Federal University of Rio de Janeiro in Brazil, using the Alignment–Balance–Vibration Trainer (ABVT) to simulate various fault conditions. The dataset includes several classes: normal, misalignment (horizontal and vertical), unbalance, Ball Pass Frequency Outer race (BPFO), Ball Pass Frequency Inner race (BPFI), and bearing cage faults. It also incorporates variations in fault severity and rotational speeds ranging from 737 RPM to 3686 RPM. All samples were recorded at a 50 kHz sampling rate for a duration of 5 s.

4.1.5. Integration Setting and Results

Since the original datasets differ in terms of class definitions and measurement conditions, we refined them into a unified and interpretable class structure. In particular, the MaFaulDa dataset includes various speeds and load levels even within the same fault category (e.g., unbalance), and misalignment and looseness conditions are collected under horizontal/vertical and different fastening levels, respectively. To avoid excessive subclass fragmentation, we integrated the data into five primary fault categories: normal, misalignment, unbalance, and bearing. The bearing class merges BPFO, BPFI, and bearing cage-related faults. The final class definitions are as follows:
  • Normal: healthy operating condition.
  • Misalignment: shaft misalignment.
  • Unbalance: rotor mass imbalance.
  • Bearing: bearing-related faults (including outer race, inner race, and cage faults).
In rotary machinery diagnostics, orthogonal sensor axes (X and Y) are typically used together. In this study, we utilized both X-axis and Y-axis vibration signals measured from sensors mounted on the motor. The final integrated dataset is as shown in Table 1 and Table 2. The total dataset consists of 62,863 samples. To address the known issue of test–train leakage in condition-based maintenance studies, especially those relying on single-source datasets like CWRU, we constructed a benchmark that integrates four diverse public datasets. These datasets were collected under different machines, sensor setups, rotational speeds, and acquisition protocols. This diversity reduces the risk of model overfitting to a narrow fault distribution.
Moreover, our experimental design strictly separates the source and target domains. VAT, MaFaulDA, and DXAI are used exclusively for training and validation, while the VBL-VA001 dataset is reserved as a test-only target domain. No data from the target domain was used during training, which structurally prevented information leakage and ensured a fair evaluation under realistic domain shift conditions. Table 3 summarizes the integrated dataset distribution. The training–validation–test split was configured as follows: VBL-VA001, MaFaulDa, and DXAI were used as source domains, with an 8:2 split for training and validation, respectively. The VAT dataset was designated as the target domain and was excluded from training. It was used exclusively as a test set to evaluate model adaptation under domain shift conditions.

4.2. Benchmark

4.2.1. Benchmark Setting

To evaluate the performance of the proposed model, we conducted benchmark experiments against a variety of conventional models. The comparison included both machine learning-based and deep learning-based methods. For machine learning models, we used the Support Vector Machine (SVM) [48], Logistic Regression [49], Random Forest [50], Gradient Boosting [51], and K-Nearest Neighbors (KNN) [52] methods. For deep learning models, we adopted 1D-Convolutional Neural Network (1D-CNN) [53], Long Short-Term Memor y(LSTM) [54], and Transformer [55] architectures.
For all machine learning models, the hyperparameters were optimized using grid search on the source domain training data. The deep learning models were implemented with the following architectural settings:
  • 1D-CNN: This is composed of three 1D convolutional blocks followed by a fully connected layer. Each block consists of a convolution layer, batch normalization, ReLU activation, and max pooling. The convolution layers use a kernel size of 3, a stride of 1, and padding of 1, with the output channels set to 16, 32, and 64, respectively.
  • LSTM: This is constructed with two LSTM layers, each with a hidden dimension of 128. A dropout rate of 0.3 is applied between layers, and the input size is set to 2.
  • Transformer: The embedding dimension is set to 64, and the number of heads in the multi-head attention is 4. The encoder output is summarized and passed through a fully connected layer to predict fault classes.
All deep learning models were trained using CrossEntropyLoss and the Adam optimizer, with a learning rate of 1 × 10 3 , a dropout rate of 0.3, and a batch size of 512, over 200 epochs. To assess the effectiveness of the proposed order-frequency-based preprocessing technique, we applied five different input data types across all models:
  • Statistical features from the frequency domain: power, max frequency, mean frequency, median frequency, spectral skewness, spectral kurtosis, peak amplitude, band energy, dominant frequency power, spectral entropy, root-mean-square (RMS) frequency, frequency variance.
  • Statistical features from the time domain: mean, standard deviation, max, min, RMS, skewness, kurtosis, peak, peak-to-peak value, crest factor, impulse factor, shape factor.
  • Raw time-series signals.
  • Frequency-domain signals generated through FFT transformation.
  • Order-frequency-domain signals produced by the proposed preprocessing pipeline.
For machine learning models, the first two statistical feature sets, (1) and (2), were used as inputs. For deep learning models, raw signals (3), FFT-based inputs (4), and order-frequency-transformed inputs (5) were used. These results empirically validate the effectiveness of the proposed preprocessing method, and provide a comprehensive evaluation of the effectiveness of the proposed input transformation.

4.2.2. Benchmark Results

To evaluate the effectiveness of the proposed cross-domain fault classification model, we conducted benchmark experiments comparing it with various existing techniques. The models used in the experiments include traditional machine learning approaches and deep learning-based architectures. Each model was trained using either raw vibration signals or transformed inputs, allowing for a comparison of performance across different modeling strategies.
All models were trained using data collected from the source domain. Performance evaluation was carried out on two datasets: a validation set from the source domain (i.e., without domain shift) and a test set from the target domain (i.e., with domain shift). The evaluation metrics used were precision, recall, and F1-score. These three metrics collectively assess the model’s accuracy and sensitivity, and the trade-off between them. The benchmark results are shown in Table 4.
Precision measures the proportion of correctly predicted positive samples among all samples predicted as positive. A higher precision indicates a lower false positive rate.
Precision = T P T P + F P
Recall measures the proportion of true positive samples correctly identified by the model. A higher recall indicates better sensitivity to actual positives.
Recall = T P T P + F N
The F1-score is the harmonic mean of precision and recall, providing a balanced evaluation when both metrics are equally important. It yields a high value only when both precision and recall are sufficiently high.
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
In the analysis of the experimental results, we present visual comparisons of the F1-score and recall across different models and preprocessing methods. The F1-score, as the harmonic mean of precision and recall, provides a comprehensive measure of overall model performance, while recall is particularly sensitive to performance degradation caused by domain shift. Thus, these two metrics are primarily used in this analysis. In the source domain, most models achieved high performance, with precision, recall, and F1-score all exceeding 0.9 . However, under domain shift, performance varied significantly depending on the model type.
As shown in Figure 11, machine learning models experienced a sharp decline in performance. On average, their F1-score dropped from above 0.9 to below 0.1 , and recall decreased dramatically to the 0.05   0.25 range. For example, SVM achieved an F1-score of 0.92 in the source domain but only 0.10 in the target domain, corresponding to an 89% performance drop.
In contrast, as shown in Figure 12, deep learning models demonstrated more stable adaptation performance. While 1D-CNN and LSTM still suffered from degradation, they maintained average F1-scores in the 0.2 0.3 range in the target domain, indicating greater robustness than traditional machine learning models.
The full framework proposed in this study, which integrates both the preprocessing pipeline and model architecture, demonstrated consistent and superior performance. It achieved a precision of 0.89 , recall of 0.84 , and an F1-score of 0.84 in the source domain, and maintained strong target-domain results with a precision of 0.50 , recall of 0.51 , and an F1-score of 0.47 . Importantly, while most baseline models recorded target-domain recall in the range of 0.1 0.3 , our model consistently outperformed them, achieving a recall of 0.51 and demonstrating enhanced anomaly detection sensitivity and domain adaptation capability.

4.3. Model Implementation and Ablation Study

To ensure stable performance under domain shift, this study introduces three key components: (1) A preprocessing method based on order-frequency transformation. (2) Self-supervised reconstruction learning using a variational autoencoder (VAE). (3) A test-time training (TTT) strategy for unsupervised adaptation to the target domain. To quantitatively assess the contribution of each component, we conduct an ablation study.

4.3.1. Model Implementation

The proposed U-NetVAE model is a U-Net-based variational autoencoder designed for 1D time-series data. The encoder–decoder architecture is enhanced with skip connections to improve reconstruction performance. The latent vectors extracted from the encoder are passed to a downstream classifier, which determines the fault class based on the difference between the latent vectors of the input and a normal (reference) signal. Although our framework is conceptually inspired by SVM-based classification in latent space, we implement it using a single fully connected layer that performs the same linear decision function. This enables end-to-end training with the encoder.
The architecture is configured as follows: the U-NetVAE consists of approximately 117 K parameters, and the classifier has 16.4 K parameters, resulting in a total of about 133 K parameters. The main hyperparameters for the U-NetVAE are as follows: 2 residual blocks, each with 8 residual layers; a latent embedding dimension of 32; and an embedding length of 128. The model is trained with a batch size of 512 for 400 epochs. StepLR is used as the learning rate scheduler for the U-NetVAE, while the classifier uses a Cosine Annealing scheduler.
To prevent early overfitting in the classifier, a 50-epoch warm-up strategy is applied. Both the U-NetVAE and classifier are optimized using the Adam optimizer with an initial learning rate of 1 × 10 3 . To ensure reproducibility, all experiments are conducted with the random seed fixed at 42. The overall training structure is summarized in Table 5.
The test-time training (TTT) strategy is implemented in an offline manner, where only the encoder is updated. TTT leverages unlabeled target domain data to refine the latent representations via self-supervised learning. During this phase, the decoder and classifier are kept frozen. The TTT hyperparameters are as follows: a batch size of 512, a learning rate of 1 × 10 5 , 2 training epochs, and optimization using SGD. This structure is designed to provide effective domain adaptation with minimal computational overhead.

4.3.2. Ablation Study

To quantitatively analyze the contribution of each component in the proposed framework, we conducted an ablation study. The experiments were performed using four different model configurations, as summarized in the following Table 6.
The Baseline model consists only of the encoder, and resembles a 1D-CNN architecture with residual connections. Order-frequency preprocessing, reconstruction-based learning, and TTT strategies are all disabled in this configuration. While the model showed stable performance in the source domain, its performance dropped sharply in the target domain, with an F1-score of just 0.23 , indicating poor adaptation to domain shifts.
When order-frequency preprocessing was added to the baseline, performance in the target domain improved significantly: precision improved from 0.26 to 0.48 , recall by 0.24 points, and the F1-score by 0.23 points. This result highlights the effectiveness of the proposed preprocessing method in mitigating cross-domain distributional shifts and extracting robust features. In the next configuration, reconstruction learning was introduced as an auxiliary task. Although no notable performance gains were observed in either domain, the integration of the VAE structure enabled latent adaptation during the TTT phase. Thus, while the numerical metrics remained similar, the model gained the flexibility to adapt in a self-supervised setting, enhancing its extensibility.
Finally, applying TTT to the full proposed framework led to further improvements in the target domain: precision increased by 0.02 points, recall by 0.01 points, and the F1-score by 0.04 points. This demonstrates that even with short adaptation periods, TTT can compensate for residual performance drops due to domain shift. The increase in F1-score reflects overall performance gains, and the stable recall indicates reliable anomaly detection capability. In summary, each component contributed to performance improvement on its own, and the integrated full framework played a critical role in achieving robust adaptation to unseen target domains.
To further assess the impact of the reconstruction module architecture, we conducted a comparative experiment using three variants: a standard VAE, a vector-quantized VAE (VQ-VAE), and the proposed U-NetVAE. All models were implemented with the same number of layers, latent dimensionality, and optimization settings to ensure fair comparison. Test-time training (TTT) was applied to each model before evaluation.
As shown in Table 7, in the source domain, all models performed relatively well. However in the target domain, the U-Net-VAE showed a notable improvement in the F1-score ( 0.50 ) compared to the VQ-VAE ( 0.35 ) and standard VAE ( 0.15 ). These results indicate that the U-Net architecture plays a key role in enhancing generalization under domain shift by enabling more robust latent representation learning.
According to prior studies [56], updating only the encoder during test-time training while jointly performing a self-supervised task has been reported to be effective. Based on these findings, we adopted the same approach in our work, and as a result, achieved stable performance, with a recall of 0.51 , precision of 0.50 , and an F1-score of 0.50 when updating only the encoder. Additionally, when we included the classifier in the test-time training along with the encoder, we obtained a recall of 0.48 , precision of 0.46 , and an F1-score of 0.46 .
Through these comparative experiments, we confirmed that updating only the encoder provides more stable and generalized performance and can be reliably utilized under domain shift conditions. In contrast, including the classifier in the adaptation process can lead to partial performance improvements; however, it was observed that this approach may result in unstable convergence and potential overfitting depending on the distributional differences between domains.
Additionally, to further evaluate the effectiveness of different test-time training (TTT) methods under identical conditions, we conducted comparative experiments using the same dataset with T3A [57] and TAST [58], which are methods used in time-series data [59,60], and the proposed U-Net-VAE-based TTT approach. As summarized in the Table 8, in the target domain, T3A achieved a precision of 0.28 , recall of 0.33 , and an F1-score of 0.30 , while TAST achieved a precision of 0.43 , recall of 0.31 , and an F1-score of 0.36 . In contrast, the proposed U-Net-VAE-based TTT method achieved the highest performance with a precision of 0.50 , recall of 0.51 , and an F1-score of 0.50 . These findings confirm that our U-Net-VAE-based TTT framework not only achieves superior adaptation performance under domain shifts but also maintains a general and practical architecture that is readily applicable to industrial time-series fault diagnosis without requiring additional architectural modifications.

5. Conclusions

In this study, we proposed a robust fault diagnosis framework for rotary machinery under domain shift conditions. Our method integrates order-frequency-domain preprocessing, self-supervised latent representation learning with a U-NetVAE, and a source-free test-time training (TTT) strategy. Together, these components enable effective domain adaptation without access to source data or target labels. Through extensive benchmark experiments across four public datasets, the proposed approach consistently outperformed conventional models in the presence of domain shift. Ablation studies confirmed the individual contributions of each component.
This work demonstrates the effectiveness of combining mechanical priors, self-supervised learning, and lightweight adaptation for practical industrial diagnostics. Future research will focus on extending the model to handle open-set scenarios and unseen fault types, as well as improving its applicability to low-resource or online monitoring environments.

Author Contributions

Conceptualization, H.J. and S.K.; methodology, H.J. and S.K.; validation, H.J., S.K., and D.S.; formal analysis, S.K. and D.S.; investigation, H.J.; resources, J.K.; data curation, D.S.; writing—original draft preparation, H.J. and S.K.; writing—review and editing, S.K.; visualization, D.S.; supervision, J.K.; project administration, H.J.; funding acquisition, J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by a Korea Institute of Energy Technology Evaluation and Planning (KETEP) grant funded by the Korean government (MOTIE) (20224B10100060, Development of Artificial Intelligence Vibration Monitoring System for Rotating Machinery).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conficts of interest.

References

  1. Tang, B.; Liu, W.; Song, T. Wind turbine fault diagnosis based on Morlet wavelet transformation and Wigner-Ville distribution. Renew. Energy 2010, 35, 2862–2866. [Google Scholar] [CrossRef]
  2. Yan, R.; Gao, R.X.; Chen, X. Wavelets for fault diagnosis of rotary machines: A review with applications. Signal Process. 2014, 96, 1–15. [Google Scholar] [CrossRef]
  3. Konar, P.; Chattopadhyay, P. Bearing fault detection of induction motor using wavelet and Support Vector Machines (SVMs). Appl. Soft Comput. 2011, 11, 4203–4211. [Google Scholar] [CrossRef]
  4. Wan, X.; Wang, D.; Peter, W.T.; Xu, G.; Zhang, Q. A critical study of different dimensionality reduction methods for gear crack degradation assessment under different operating conditions. Measurement 2016, 78, 138–150. [Google Scholar] [CrossRef]
  5. Zhang, W.; Peng, G.; Li, C. Bearings fault diagnosis based on convolutional neural networks with 2-D representation of vibration signals as input. In MATEC Web of Conferences; EDP Sciences: Les Ulis, France, 2017; Volume 95, p. 13001. [Google Scholar]
  6. Ince, T.; Kiranyaz, S.; Eren, L.; Askar, M.; Gabbouj, M. Real-time motor fault detection by 1-D convolutional neural networks. IEEE Trans. Ind. Electron. 2016, 63, 7067–7075. [Google Scholar] [CrossRef]
  7. Guo, X.; Chen, L.; Shen, C. Hierarchical adaptive deep convolution neural network and its application to bearing fault diagnosis. Measurement 2016, 93, 490–502. [Google Scholar] [CrossRef]
  8. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  9. Farrar, C.R.; Doebling, S.W.; Nix, D.A. Vibration–based structural damage identification. Philos. Trans. R. Soc. Lond. Ser. A Math. Phys. Eng. Sci. 2001, 359, 131–149. [Google Scholar] [CrossRef]
  10. Saha, D.K.; Hoque, M.E.; Badihi, H. Development of intelligent fault diagnosis technique of rotary machine element bearing: A machine learning approach. Sensors 2022, 22, 1073. [Google Scholar] [CrossRef]
  11. Lei, Y.; He, Z.; Zi, Y.; Hu, Q. Fault diagnosis of rotating machinery based on multiple ANFIS combination with GAs. Mech. Syst. Signal Process. 2007, 21, 2280–2294. [Google Scholar] [CrossRef]
  12. Zhang, W.; Li, C.; Peng, G.; Chen, Y.; Zhang, Z. A deep convolutional neural network with new training methods for bearing fault diagnosis under noisy environment and different working load. Mech. Syst. Signal Process. 2018, 100, 439–453. [Google Scholar] [CrossRef]
  13. Shao, J.; Huang, Z.; Zhu, Y.; Zhu, J.; Fang, D. Rotating machinery fault diagnosis by deep adversarial transfer learning based on subdomain adaptation. Adv. Mech. Eng. 2021, 13, 16878140211040226. [Google Scholar] [CrossRef]
  14. Aiordăchioaie, D. A comparative analysis of fault detection and process diagnosis methods based on a signal processing paradigm. Discov. Appl. Sci. 2025, 7, 10. [Google Scholar] [CrossRef]
  15. Wang, T.; Li, J.; Zhang, P.; Sun, Y. Fault diagnosis of rotating machinery under time-varying speed based on order tracking and deep learning. J. Vibroeng. 2020, 22, 366–382. [Google Scholar] [CrossRef]
  16. Yu, X.; Feng, Z.; Zhang, D. Adaptive high-resolution order spectrum for complex signal analysis of rotating machinery: Principle and applications. Mech. Syst. Signal Process. 2022, 177, 109194. [Google Scholar] [CrossRef]
  17. He, K.; Chen, X.; Xie, S.; Li, Y.; Dollár, P.; Girshick, R. Masked autoencoders are scalable vision learners. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 16000–16009. [Google Scholar]
  18. Maaløe, L.; Sønderby, C.K.; Sønderby, S.K.; Winther, O. Auxiliary deep generative models. In Proceedings of the International Conference on Machine Learning, PMLR, New York, NY, USA, 20–22 June 2016; pp. 1445–1453. [Google Scholar]
  19. Ding, Z.; Xu, Y.; Xu, W.; Parmar, G.; Yang, Y.; Welling, M.; Tu, Z. Guided variational autoencoder for disentanglement learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7920–7929. [Google Scholar]
  20. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114v10. [Google Scholar]
  21. Bengio, Y.; Courville, A.; Vincent, P. Representation learning: A review and new perspectives. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1798–1828. [Google Scholar] [CrossRef]
  22. Li, J.; Ren, W.; Han, M. Mutual information variational autoencoders and its application to feature extraction of multivariate time series. Int. J. Pattern Recognit. Artif. Intell. 2022, 36, 2255005. [Google Scholar] [CrossRef]
  23. Erhan, D.; Courville, A.; Bengio, Y.; Vincent, P. Why does unsupervised pre-training help deep learning? In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics, Sardinia, Italy, 13–15 May 2010; JMLR Workshop and Conference Proceedings. pp. 201–208. [Google Scholar]
  24. Gouda, W.; Tahir, S.; Alanazi, S.; Almufareh, M.; Alwakid, G. Unsupervised outlier detection in IOT using deep VAE. Sensors 2022, 22, 6617. [Google Scholar] [CrossRef]
  25. Huang, T.; Chen, P.; Li, R. A semi-supervised VAE based active anomaly detection framework in multivariate time series for online systems. In Proceedings of the ACM Web Conference 2022, Lyon, France, 25–29 April 2022; pp. 1797–1806. [Google Scholar]
  26. Higgins, I.; Matthey, L.; Pal, A.; Burgess, C.; Glorot, X.; Botvinick, M.; Mohamed, S.; Lerchner, A. beta-vae: Learning basic visual concepts with a constrained variational framework. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  27. Liang, J.; Hu, D.; Feng, J. Do we really need to access the source data? Source hypothesis transfer for unsupervised domain adaptation. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 6028–6039. [Google Scholar]
  28. Sun, Y.; Wang, X.; Liu, Z.; Miller, J.; Efros, A.; Hardt, M. Test-time training with self-supervision for generalization under distribution shifts. In Proceedings of the International Conference on Machine Learning, PMLR, Virtual, 13–18 July 2020; pp. 9229–9248. [Google Scholar]
  29. Chen, D.; Wang, D.; Darrell, T.; Ebrahimi, S. Contrastive test-time adaptation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 21–24 June 2022; pp. 295–305. [Google Scholar]
  30. Darestani, M.Z.; Liu, J.; Heckel, R. Test-time training can close the natural distribution shift performance gap in deep learning based compressed sensing. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 4754–4776. [Google Scholar]
  31. Sinha, S.; Gehler, P.; Locatello, F.; Schiele, B. Test: Test-time self-training under distribution shift. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 2759–2769. [Google Scholar]
  32. Wang, Y.; Gong, P.; Wu, M.; Ott, F.; Li, X.; Xie, L.; Chen, Z. Temporal Source Recovery for Time-Series Source-Free Unsupervised Domain Adaptation. arXiv 2024, arXiv:2409.19635. [Google Scholar]
  33. Ragab, M.; Eldele, E.; Wu, M.; Foo, C.S.; Li, X.; Chen, Z. Source-free domain adaptation with temporal imputation for time series data. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Long Beach, CA, USA, 6–10 August 2023; pp. 1989–1998. [Google Scholar]
  34. Mohammadi, S.; Belgiu, M.; Stein, A. A source-free unsupervised domain adaptation method for cross-regional and cross-time crop mapping from satellite image time series. Remote Sens. Environ. 2024, 314, 114385. [Google Scholar] [CrossRef]
  35. Leturiondo, U.; Mishra, M.; Galar, D.; Salgado, O. Synthetic data generation in hybrid modelling of rolling element bearings. Insight-Non-Destr. Test. Cond. Monit. 2015, 57, 395–400. [Google Scholar] [CrossRef]
  36. Sobie, C.; Freitas, C.; Nicolai, M. Simulation-driven machine learning: Bearing fault classification. Mech. Syst. Signal Process. 2018, 99, 403–419. [Google Scholar] [CrossRef]
  37. Matania, O.; Cohen, R.; Bechhoefer, E.; Bortman, J. Zero-fault-shot learning for bearing spall type classification by hybrid approach. Mech. Syst. Signal Process. 2025, 224, 112117. [Google Scholar] [CrossRef]
  38. Cooley, J.W.; Tukey, J.W. An algorithm for the machine calculation of complex Fourier series. Math. Comput. 1965, 19, 297–301. [Google Scholar] [CrossRef]
  39. Gabor, D. Theory of communication. Part 1: The analysis of information. J. Inst. Electr. Eng.-Part III Radio Commun. Eng. 1946, 93, 429–441. [Google Scholar] [CrossRef]
  40. Mallat, S. A Wavelet Tour of Signal Processing; Elsevier: Amsterdam, The Netherlands, 1999. [Google Scholar]
  41. Xue, H.; Wang, H.; Song, L.; Chen, P. Structural fault diagnosis of rotating machinery based on distinctive frequency components and support vector machines. In Proceedings of the International Conference on Intelligent Computing, Chongqing, China, 8–9 January 2011; Springer: Berlin/Heidelberg, Germany, 2011; pp. 341–348. [Google Scholar]
  42. Wu, J.D.; Bai, M.R.; Su, F.C.; Huang, C.W. An expert system for the diagnosis of faults in rotating machinery using adaptive order-tracking algorithm. Expert Syst. Appl. 2009, 36, 5424–5431. [Google Scholar] [CrossRef]
  43. Bai, M.R.; Jeng, J.; Chen, C. Adaptive order tracking technique using recursive least-square algorithm. J. Vib. Acoust. 2002, 124, 502–511. [Google Scholar] [CrossRef]
  44. Jung, W.; Kim, S.H.; Yun, S.H.; Bae, J.; Park, Y.H. Vibration, acoustic, temperature, and motor current dataset of rotating machine under varying operating conditions for fault diagnosis. Data Brief 2023, 48, 109049. [Google Scholar] [CrossRef] [PubMed]
  45. Atmaja, B.T.; Ihsannur, H.; Suyanto; Arifianto, D. Lab-scale vibration analysis dataset and baseline methods for machinery fault diagnosis with machine learning. J. Vib. Eng. Technol. 2024, 12, 1991–2001. [Google Scholar] [CrossRef]
  46. Marins, M.A.; Ribeiro, F.M.; Netto, S.L.; Da Silva, E.A. Improved similarity-based modeling for the classification of rotating-machine failures. J. Frankl. Inst. 2018, 355, 1913–1930. [Google Scholar] [CrossRef]
  47. Brito, L.C.; Susto, G.A.; Brito, J.N.; Duarte, M.A. Mechanical Faults in Rotating Machinery Dataset (Normal, Unbalance, Misalignment, Looseness). 2022. Available online: https://www.research.unipd.it/handle/11577/3469613 (accessed on 17 April 2025).
  48. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  49. Hosmer, D.W., Jr.; Lemeshow, S.; Sturdivant, R.X. Applied Logistic Regression; John Wiley & Sons: Hoboken, NJ, USA, 2013. [Google Scholar]
  50. Breiman, L. Random forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  51. Friedman, J.H. Greedy function approximation: A gradient boosting machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  52. Cover, T.; Hart, P. Nearest neighbor pattern classification. IEEE Trans. Inf. Theory 1967, 13, 21–27. [Google Scholar] [CrossRef]
  53. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  54. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  55. Ashish, V. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, I. [Google Scholar]
  56. Liu, Y.; Kothari, P.; Van Delft, B.; Bellot-Gurlet, B.; Mordan, T.; Alahi, A. Ttt++: When does self-supervised test-time training fail or thrive? Adv. Neural Inf. Process. Syst. 2021, 34, 21808–21820. [Google Scholar]
  57. Iwasawa, Y.; Matsuo, Y. Test-time classifier adjustment module for model-agnostic domain generalization. Adv. Neural Inf. Process. Syst. 2021, 34, 2427–2440. [Google Scholar]
  58. Jang, M.; Chung, S.Y.; Chung, H.W. Test-time adaptation via self-training with nearest neighbor information. arXiv 2022, arXiv:2207.10792. [Google Scholar]
  59. Jang, G.D.; Han, D.K.; Lee, S.W. Test-Time Adaptation for EEG-Based Driver Drowsiness Classification. In Proceedings of the International Conference on Pattern Recognition and Artificial Intelligence, Xiamen, China, 20–22 September 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 410–424. [Google Scholar]
  60. Ragab, M.; Eldele, E.; Tan, W.L.; Foo, C.S.; Chen, Z.; Wu, M.; Kwoh, C.K.; Li, X. Adatime: A benchmarking suite for domain adaptation on time series data. ACM Trans. Knowl. Discov. Data 2023, 17, 1–18. [Google Scholar] [CrossRef]
Figure 1. Autoencoder (AE) and variational autoencoder (VAE) architecture. (a) In the AE, the encoder f enc maps the input x input to a latent vector z, and the decoder f dec reconstructs x recon from z. (b) In the VAE, the encoder f enc computes the mean vector μ and standard deviation vector σ of the latent-space distribution from the input x input . A sample z sample is then drawn from the normal distribution p ( z ) = N ( μ , σ 2 ) and passed to the decoder f dec to reconstruct x recon .
Figure 1. Autoencoder (AE) and variational autoencoder (VAE) architecture. (a) In the AE, the encoder f enc maps the input x input to a latent vector z, and the decoder f dec reconstructs x recon from z. (b) In the VAE, the encoder f enc computes the mean vector μ and standard deviation vector σ of the latent-space distribution from the input x input . A sample z sample is then drawn from the normal distribution p ( z ) = N ( μ , σ 2 ) and passed to the decoder f dec to reconstruct x recon .
Sensors 25 04383 g001
Figure 2. Source-Free Domain Adaptation Framework. Blue indicates source-domain samples orange indicates target-domain samples. The “Distribution Difference” arrow highlights the discrepancy between the two domains. The classifier is trained on source data, and target samples misclassified in the mixed set are marked with red crosses X. The closed lock icon signifies that source data cannot be used during adaptation, while the open lock icon indicates that only target data is used. The dashed line denotes the decision boundary of the adapted classifier.
Figure 2. Source-Free Domain Adaptation Framework. Blue indicates source-domain samples orange indicates target-domain samples. The “Distribution Difference” arrow highlights the discrepancy between the two domains. The classifier is trained on source data, and target samples misclassified in the mixed set are marked with red crosses X. The closed lock icon signifies that source data cannot be used during adaptation, while the open lock icon indicates that only target data is used. The dashed line denotes the decision boundary of the adapted classifier.
Sensors 25 04383 g002
Figure 3. Data preprocessing results.
Figure 3. Data preprocessing results.
Sensors 25 04383 g003
Figure 4. Proposed method.
Figure 4. Proposed method.
Sensors 25 04383 g004
Figure 5. Model block structure.
Figure 5. Model block structure.
Sensors 25 04383 g005
Figure 6. U-Net-based VAE structure.
Figure 6. U-Net-based VAE structure.
Sensors 25 04383 g006
Figure 7. VAT data collection environment.
Figure 7. VAT data collection environment.
Sensors 25 04383 g007
Figure 8. DXAI data collection environment.
Figure 8. DXAI data collection environment.
Sensors 25 04383 g008
Figure 9. VBL data collection environment.
Figure 9. VBL data collection environment.
Sensors 25 04383 g009
Figure 10. MaFaulDa data collection environment.
Figure 10. MaFaulDa data collection environment.
Sensors 25 04383 g010
Figure 11. Machine learning model results.
Figure 11. Machine learning model results.
Sensors 25 04383 g011
Figure 12. Deep learning model results.
Figure 12. Deep learning model results.
Sensors 25 04383 g012
Table 1. Dataset specifics (Part 1: VAT and VBL-VA001).
Table 1. Dataset specifics (Part 1: VAT and VBL-VA001).
DatasetsVATVBL-VA001
Items
MotorFour-pole AC motor by SIEMENS, Germany (3HP)Panasonic, Japan GP-129JXK (125W, 3000RPM)
Accelerometers4× PCB352C34 at bearings A, B, United StatesLOG-0002-100G on motor (X,Y,Z), United States
Other SensorsAcoustic mic, 4× CT sensor, 2× K-type thermocouples-
Sampling Frequency25.6 kHz (vibration/temp), 51.2 kHz (acoustic)20 kHz
Fault TypesBearing fault, unbalance, misalignmentNormal, misalignment, unbalance (2 levels), BPFO
Operating ConditionsLoad: 0, 2, 4 Nm/RPM: 3010, 680–2460Fixed 3000 RPM
Acquisition Time and RepetitionNormal: 120s, Faulty: 60 s, 7 × 300 s, 1 × 600 s1000 samples/class, 5 s/sample
Data Collection InstitutionKorea Advanced Institute of Science and TechnologySepuluh Nopember Institute of Technology, Indonesia
SourceJung et al. (2023) [44], Mendeley Data, 10.17632/vxkj334rzv.7Atmaja et al. (2024) [45], J. Vibration Eng. Tech. 12(2)
Table 2. Dataset specifics (Part 2: MaFaulDA and DXAI).
Table 2. Dataset specifics (Part 2: MaFaulDA and DXAI).
MaFaulDADXAI
MotorDC motor, 1/4CV, ABVT trainer, United StatesThree-phase induction motor, B56 B4, 1650 RPM, 0.09 kW, 220 V, 0.70 A, Germany
AccelerometersIMI 601A01, IMI 604B31, United States4× PCB352C33 (2 per bearing), United States
Other SensorsShure(United States) SM81 condenser microphone-
Sampling Frequency50 kHz25 kHz
Fault Types8568 labeled faultsNormal, unbalance, misalignment, looseness
Operating Conditions60–3686 RPM (varied)Fixed 1238 RPM
Acquisition Time and Repetition1951 samples (varied), 5 s/sample5 tests/class, 420 samples/test, 1 s/sample
Data Collection InstitutionFederal University of Rio de JaneiroUniv. of São João del-Rei, Universidade Federal de Uberlandia, Universita degli Studi di Padova
SourceMaFaulDa Database (2018) [46], www02.smt.ufrj.br/~offshore/mfs, accessed on 17 April 2025Lucas Brito et al. (2022) [47], Expert Systems with Applications, 10.1016/j.eswa.2023.120860
Table 3. Data integration results.
Table 3. Data integration results.
DatasetVAT-MCDVBL-VA001MaFaulDADXAITotal
Class
normal10778000392210011,569
misalignment215180003984210016,235
unbalance358580002664210016,349
bearing214280008568-18,710
Total895532,00015,608714562,863
Table 4. Experimental results comparison for domain adaptation.
Table 4. Experimental results comparison for domain adaptation.
Model NameData Preprocessing MethodValidation (Source Domain)Test (Target Domain)
SVMTime Featureprecision: 0.92; recall: 0.92; F1: 0.92precision: 0.07; recall: 0.18; F1: 0.09
Frequency Featureprecision: 0.99; recall: 0.99; F1: 0.99precision: 0.06; recall: 0.05; F1: 0.05
Random ForestTime Featureprecision: 0.97; recall: 0.96; F1: 0.97precision: 0.06; recall: 0.25; F1: 0.10
Frequency Featureprecision: 0.99; recall: 0.99; F1: 0.99precision: 0.05; recall: 0.22; F1: 0.08
Logistic RegressionTime Featureprecision: 0.77; recall: 0.77; F1: 0.76precision: 0.07; recall: 0.18; F1: 0.10
Frequency Featureprecision: 0.85 recall: 0.86;; F1: 0.85precision: 0.07; recall: 0.11; F1: 0.08
Gradient BoostingTime Featureprecision: 0.96; recall: 0.95; F1: 0.96precision: 0.06; recall: 0.25; F1: 0.10
Frequency Featureprecision: 0.99; recall: 0.99; F1: 0.99precision: 0.10; recall: 0.18; F1: 0.12
KNNTime Featureprecision: 0.93; recall: 0.93; F1: 0.93precision: 0.32; recall: 0.25; F1: 0.28
Frequency Featureprecision: 0.99; recall: 0.98; F1: 0.99precision: 0.05; recall: 0.25; F1: 0.08
1D-CNNRaw Signalprecision: 0.86; recall: 0.85; F1: 0.85precision: 0.22; recall: 0.27; F1: 0.24
FFTprecision: 0.99; recall: 0.99; F1: 0.99precision: 0.27; recall: 0.25; F1: 0.25
Our Methodprecision: 0.99; recall: 0.99; F1: 0.99precision: 0.34; recall: 0.31; F1: 0.32
Raw Signalprecision: 0.77; recall: 0.70; F1: 0.69precision: 0.32; recall: 0.22; F1: 0.26
LSTMFFTprecision: 0.94; recall: 0.93; F1: 0.93precision: 0.36; recall: 0.19; F1: 0.24
Our Methodprecision: 0.98; recall: 0.98; F1: 0.98precision: 0.08; recall: 0.25; F1: 0.12
Raw Signalprecision: 0.93; recall: 0.92; F1: 0.92precision: 0.40; recall: 0.30; F1: 0.21
TransformerFFTprecision: 0.99; recall: 0.99; F1: 0.99precision: 0.34; recall: 0.14; F1: 0.11
Our Methodprecision: 0.99; recall: 0.99; F1: 0.99precision: 0.57; recall: 0.25; F1: 0.10
Our MethodOur Methodprecision: 0.89; recall: 0.84; F1: 0.86precision: 0.50; recall: 0.51; F1: 0.50
Table 5. Training setting.
Table 5. Training setting.
ModelSchedulerOptimizer (Training)Optimizer (TTT)
U-NetVAEStepLRAdamSGD (Only Encoder)
ClassifierCosine AnnealingAdam-
Table 6. Ablation study results.
Table 6. Ablation study results.
Experiment SettingOrder PreprocessingReconstructionTTTSource Domain TestTarget Domain Test
Baselinexxxprecision: 0.89
recall: 0.83
F1: 0.82
precision: 0.26
recall: 0.26
F1: 0.23
+Order Spectrum
Preprocessing
oxxprecision: 0.89
recall: 0.84
F1: 0.84
precision: 0.48
recall: 0.50
F1: 0.46
+Reconstructionooxprecision: 0.89
recall: 0.84
F1: 0.84
precision: 0.48
recall: 0.50
F1: 0.46
+Oursoooprecision: 0.89
recall: 0.84
F1: 0.86
precision: 0.50
recall: 0.51
F1: 0.50
Legend: o = included; x = not included.
Table 7. Comparison of classification performance using different reconstruction modules after applying test-time training (TTT).
Table 7. Comparison of classification performance using different reconstruction modules after applying test-time training (TTT).
ModelPrecision
(Source)
Recall
(Source)
F1
(Source)
Precision
(Target)
Recall
(Target)
F1
(Target)
VAE0.860.740.800.150.160.15
VQ-VAE0.810.780.790.380.320.35
U-Net-VAE (Ours)0.890.840.860.500.510.50
Table 8. Comparison of classification performance by applying different test-time training (TTT).
Table 8. Comparison of classification performance by applying different test-time training (TTT).
ModelPrecision (Target)Recall (Target)F1 (Target)
T3A0.280.330.30
TAST0.430.310.36
U-Net-VAE (Ours)0.500.510.50
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jeong, H.; Kim, S.; Seo, D.; Kwon, J. Source-Free Domain Adaptation Framework for Rotary Machine Fault Diagnosis. Sensors 2025, 25, 4383. https://doi.org/10.3390/s25144383

AMA Style

Jeong H, Kim S, Seo D, Kwon J. Source-Free Domain Adaptation Framework for Rotary Machine Fault Diagnosis. Sensors. 2025; 25(14):4383. https://doi.org/10.3390/s25144383

Chicago/Turabian Style

Jeong, Hoejun, Seungha Kim, Donghyun Seo, and Jangwoo Kwon. 2025. "Source-Free Domain Adaptation Framework for Rotary Machine Fault Diagnosis" Sensors 25, no. 14: 4383. https://doi.org/10.3390/s25144383

APA Style

Jeong, H., Kim, S., Seo, D., & Kwon, J. (2025). Source-Free Domain Adaptation Framework for Rotary Machine Fault Diagnosis. Sensors, 25(14), 4383. https://doi.org/10.3390/s25144383

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop