You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

7 May 2025

Multi-Variable Transformer-Based Meta-Learning for Few-Shot Fault Diagnosis of Large-Scale Systems

,
and
Department of Automation, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Advances in Sensors for Online Condition Monitoring and Fault Diagnosis

Abstract

Fault diagnosis in large-scale systems presents significant challenges due to the complexity and high dimensionality of data, as well as the scarcity of labeled fault data, which are hard to obtain during the practical operation process. This paper proposes a novel approach, called Multi-Variable Meta-Transformer (MVMT), to tackle these challenges. In order to deal with the multi-variable time series data, we modify the Transformer model, which is the currently most popular model on feature extraction of time series. To enable the Transformer model to simultaneously receive continuous and state inputs, we introduced feature layers before the encoder to better integrate the characteristics of both continuous and state variables. Then, we adopt the modified model as the base model for meta-learning—more specifically, the Model-Agnostic Meta-Learning (MAML) strategy. The proposed method leverages the power of Transformers for handling multi-variable time series data and employs meta-learning to enable few-shot learning capabilities. The case studies conducted on the Tennessee Eastman Process database and a Power-Supply System database demonstrate the exceptional performance of fault diagnosis in few-shot scenarios, whether based on continuous-only data or a combination of continuous and state variables.

1. Introduction

Large-scale systems, such as industrial machinery, power grids, and transportation networks, are fundamental to modern infrastructure. Ensuring their dependability and safety requires precise and timely fault detection and diagnosis (FDD). Traditional knowledge-driven fault diagnosis methods rely heavily on extensive mechanistic knowledge and expert experience. However, as system complexity increases, constructing a diagnostic model solely based on first principles becomes infeasible. Consequently, since the early 21st century, FDD in large-scale systems has transitioned into the data-driven era.
With the continuous advancement of deep learning, its application in FDD has garnered significant attention. Deep neural networks excel at extracting features, allowing for precise modeling of complex nonlinear relationships in large-scale systems. However, as network size grows, so does the demand for vast amounts of labeled data. One of the key challenges in data-driven FDD is the limited availability of fault samples for each specific fault category. To address this issue, meta-learning has emerged as a promising approach. Meta-learning, or “learning to learn” [], enables models to generalize from a few examples, making it particularly suitable for fault diagnosis in large-scale systems where labeled data are scarce. It involves deriving prior knowledge from numerous similar few-shot fault diagnosis tasks and utilizing this prior knowledge to build a model that quickly adapts to new fault scenarios. Depending on the type of prior knowledge, meta-learning can be classified into initialization-based, metric-based, and optimizer-based methods.
Model-Agnostic Meta-Learning (MAML) is one of the most effective initialization-based approaches. MAML learns an optimal set of initial network weights from a large number of source domain samples, which serves as prior knowledge []. This knowledge is then applied to the target domain, enabling rapid adaptation for few-shot fault diagnosis tasks. Initially, MAML was implemented using Convolutional Neural Networks (CNNs), which are well suited for feature extraction from 2D image data. However, when analyzing multivariate time series data in complex systems, CNNs may fall short due to their limited ability to capture temporal dependencies. Therefore, introducing Recurrent Neural Networks (RNNs) or Long Short-Term Memory (LSTM) networks into the MAML strategy as base models seems to be a valuable solution for time series analysis challenges. But the recurrent nature of these architectures requires extensive computations and backpropagation, significantly slowing down the meta-training process. As a result, traditional recurrent architectures, despite their effectiveness in multivariate time series classification, are not well suited for MAML-based FDD.
To overcome these limitations, recent advancements in deep learning have introduced Transformers, which have demonstrated exceptional performance in handling complex, high-dimensional data. Unlike RNNs, Transformers utilize position encoding to retain sequence information without relying on recurrence, making them a compelling choice for meta-learning applications. In real-world fault diagnosis scenarios, multivariate time series data often consist of both continuous analog and discrete state variables. To effectively capture the unique characteristics of these variable types, we propose an enhanced Transformer model. In our approach, analog and state variables are processed through separate embedding layers to extract their respective feature representations before being fed into the Transformer’s encoder for Multi-Head Self-Attention operations. This architecture ensures that both types of variables are effectively merged and utilized for fault diagnosis.
By integrating this enhanced multi-variable Transformer model with MAML, we introduce the Multi-Variable Meta-Transformer (MVMT), a novel approach specifically designed for small-sample fault diagnosis in large-scale systems. Unlike existing methods, MVMT effectively captures the unique characteristics of both continuous analog and discrete state variables through separate embedding layers, ensuring a more comprehensive representation of multivariate time series data. Furthermore, by leveraging the self-attention mechanism of Transformers, our approach overcomes the limitations of recurrent structures, significantly improving both computational efficiency and adaptation speed in meta-learning.
To rigorously evaluate the effectiveness of MVMT, we conducted extensive experiments on the Tennessee Eastman Process (TEP) dataset and a satellite Power-Supply System dataset. Comparative analysis with state-of-the-art neural networks, including LSTM, CNN, and Vision Transformer (ViT) combined with meta-learning, demonstrates that MVMT achieves superior fault diagnosis accuracy while maintaining high training efficiency. This work makes three key contributions: (1) introducing a Transformer-based meta-learning framework tailored for multivariate time series fault diagnosis in the few-shot scenario, (2) proposing a novel embedding mechanism that effectively integrates both analog and state variables, and (3) validating the effectiveness of MVMT through comprehensive experimental comparisons on benchmark and real-world datasets. Our findings underscore the potential of Transformer-based meta-learning in addressing the challenges of few-shot fault diagnosis in complex large-scale systems.
The article will progress as follows: Section 2 will discuss related work, focusing on the use of Transformers in fault diagnosis and advancements in meta-learning for addressing fault diagnosis with limited data. Section 3, Methodology, will detail the MVMT method, outlining its model structure and training approaches. Section 4 will cover the experimental setup, detailing the datasets used and the experimental configurations. Section 5 will showcase the experimental results and offer a detailed analysis. Finally, Section 6 will present conclusions regarding the MVMT method.

3. Methodology

3.1. Overall Framework of the Multi-Variable Meta-Transformer (MVMT)

By seamlessly integrating MAML, the Transformer model, and multi-variable fusion strategies, we introduce the Multi-Variable Meta-Transformer (MVMT). Designed for high efficiency and accuracy, MVMT enables rapid feature extraction and classification, even when only a few samples of both analog and state variables are available. The overall framework of MVMT is depicted in Figure 2.
Figure 2. Overall framework of the Multi-Variable Meta-Transformer.
The MVMT framework is structured into three key stages:
  • Data Collection and Pre-processing: Faults are classified as either source faults or target faults based on the number of available samples. Source faults, being common, have a sufficient number of samples, whereas target faults are rare and have only a limited number of samples. These faults are then allocated to a meta-training task set and a meta-testing task set, respectively. Notably, the faults in the meta-testing set are entirely distinct from those in the meta-training set. Unlike conventional train–test splits, fault types included in the meta-training set do not reappear in the meta-testing set.
  • Meta-Training Phase: In this stage, a MAML strategy is employed to train the multi-variable Transformer model, optimizing its initial parameters. This process yields a well-initialized multi-variable Transformer encoder, which serves as the foundation for the subsequent meta-testing phase.
  • Meta-Testing Phase: Here, the pre-trained multi-variable Transformer model is fine-tuned using the limited samples of rare faults, enabling it to effectively perform the final few-shot fault diagnosis task.
The following sections provide a detailed explanation of the meta-learning techniques, the Transformer model, and the overall architecture and training process of the MVMT model.

3.2. Meta-Learning Framework

To enhance the few-shot learning capability of the base model, we integrate it into a meta-learning framework, specifically leveraging Model-Agnostic Meta-Learning (MAML). MAML optimizes model parameters to facilitate rapid adaptation to new tasks, even with only a few available data samples.
In the context of few-shot learning for fault diagnosis, each task T consists of N fault types (N-way), with K support samples (K-shot) per type, Q query samples, and a specific loss function L T i .
The core idea of MAML is to determine an optimal model initialization θ such that, under the distribution of the entire classification task set, a model initialized with θ can be quickly trained to achieve high performance. In other words, this initialization enables the model to efficiently adapt to various classification tasks within the task set.
The MAML algorithm consists of two key optimization processes: meta-training and meta-testing. During meta-training, the model learns from multiple tasks to enhance its ability to adapt rapidly. The objective is to identify an optimal initial parameter θ from the source domain dataset during meta-training and transfer it as prior knowledge to the target domain. This allows the model to fine-tune efficiently with only a few gradient updates in the target domain, enabling effective adaptation to new tasks.
Figure 3 illustrates the relationship between the meta-training and meta-testing phases.
Figure 3. Relationship between meta-training and meta-testing. The red dot denotes the optimal initial parameters (prior) learned during the meta-training phase.
The overall process of the MAML algorithm is as follows:
  • Sample B tasks from the task distribution T i p ( T ) . Each task consists of N fault types, with K support samples for task-specific training and Q query samples for task-specific validation.
  • For each task T i , use the K × N support samples to compute the updated parameters θ i through S iterations of inner-loop training:
    θ i = θ α θ L T i ( f θ )
    where α is the inner-loop learning rate.
  • Evaluate the updated model using the query sets from all B tasks. The task-specific loss function L T i is used to compute the overall meta-objective:
    min θ T i p ( T ) L T i ( f θ i S )
  • Update the model parameters θ using the meta-gradient:
    θ θ β θ T i p ( T ) L T i ( f θ i S )
    where β is the meta-learning rate (outer-loop learning rate).
Figure 4 illustrates the workflow of the MAML algorithm. Step 2 represents the inner loop of the process, while steps 1 through 4 form the outer loop.
Figure 4. MAML algorithm workflow.
The objective of the outer loop is to optimize the model’s initial parameters through training on multiple tasks. Specifically, the outer loop attempts to find a good initialization parameter θ that allows the model to quickly adapt to new tasks. The outer loop updates the initialization parameters using the meta-loss from multiple tasks, thereby enhancing the model’s generalization ability.
The inner loop is used for short-term optimization on each task, with the goal of adjusting the model parameters through a few gradient updates based on the task’s data. The optimization process in the inner loop typically uses gradient descent to update the model parameters. For each task T i , the model is fine-tuned on the task-specific training data during the inner loop.
While the objectives of the outer and inner loops are different, they are interdependent. The inner loop fine-tunes the parameters on each task, enabling the model to rapidly adapt to the task-specific features. Meanwhile, the outer loop optimizes the initial parameters based on the training results from multiple tasks, allowing the initialization to generalize to different tasks. In MAML, the outer and inner loops alternate, progressively enhancing the model’s adaptability through multiple training iterations.
For more details, please check Algorithm 1.
Algorithm 1 MAML for Classification
Input:         α : learning rate for inner updates
           β : learning rate for meta-update
           p ( T ) : distribution over tasks
Output:     Model parameters θ
1:          Initialize model parameters θ
2:          while not done do
3:             Sample batch of tasks T i p ( T )
4:             for all  T i  do
5:                Sample K datapoints D T i = { ( x k , y k ) }
6:                Evaluate θ L T i ( f θ ) using D T i
7:                Compute adapted parameters with gradient descent:
                 θ i = θ α θ L T i ( f θ )
8:             end for
9:             Sample new datapoints D T i = { ( x k , y k ) }
10:           Compute meta-objective:
             θ θ β θ T i p ( T ) L T i ( f θ i )
11:        end while
During meta-testing, the model uses the learned initial parameters to quickly adapt to new tasks with a few gradient steps, enabling effective few-shot learning.
Meta-SGD is an extension of MAML that not only learns the initial parameters but also learns the learning rates for each parameter during the meta-training process. This provides an additional layer of flexibility and can lead to faster convergence and better performance.
The main modification in Meta-SGD is that each parameter θ i has an associated learning rate α i . During the meta-training phase, both the initial parameters and the learning rates are updated.
During meta-testing, the model uses the learned initial parameters and learning rates to quickly adapt to new tasks with a few gradient steps, enhancing the effectiveness of few-shot learning. For more details, please check Algorithm 2.
In practical fault diagnosis applications, the number of fault categories in tasks from the source domain may not necessarily match those in the target domain. To address this issue, we introduced an adaptive classifier. While traditional MAML uses a fixed classification head, the adaptive classifier generates a specific classifier for each task, enabling dynamic adaptation to the number of categories in different tasks. For more details, please check Algorithm 3.
Algorithm 2 Meta-SGD for Classification
Input:        α : learning rates for inner updates (learned)
          β : learning rate for meta-update
          p ( T ) : distribution over tasks
Output:      Model parameters θ and learning rates α
1:         Initialize model parameters θ and learning rates α
2:         while not done do
3:            Sample batch of tasks T i p ( T )
4:            for all  T i  do
5:               Sample K datapoints D T i = { ( x k , y k ) }
6:               Evaluate θ L T i ( f θ ) using D T i
7:               Compute adapted parameters with learned
                  gradient descent:
                   θ i = θ α θ L T i ( f θ )
8:            end for
9:            Sample new datapoints D T i = { ( x k , y k ) }
10:             Compute meta-objective:
                θ θ β θ T i p ( T ) L T i ( f θ i )
                α α β α T i p ( T ) L T i ( f θ i )
11:       end while
Algorithm 3 MAML with adaptive classifier.
Input          α : Inner-loop learning rate
           β : Meta-learning rate
           p ( T ) : Task distribution
Output     Optimized model parameters θ
1:          Initialize model parameters θ
2:          while not converged do
3:              Sample a batch of tasks T i from task distribution p ( T )
4:              for each task T i  do
5:                  Sample support set D T i support and query set D T i query
6:                  Construct the adaptive classifier g ϕ using D T i support
7:                  Compute task-specific loss:
                    L T i = ( x j , y j ) D T i support ( g ϕ ( f θ ( x j ) ) , y j )
8:                  Compute task-specific parameter updates:
                      θ T i = θ α θ L T i
9:              end for
10:            Compute meta-loss on the query set:
               L meta = T i ( x j , y j ) D T i query ( g ϕ ( f θ T i ( x j ) ) , y j )
11:            Update meta-parameters:
               θ θ β θ L meta
12:       end while

3.3. Multi-Variable Transformer

The Transformer [], a deep network based on self-attention, was initially proposed for natural language-processing tasks and has demonstrated exceptional capability in capturing long-range dependencies. In this study, we extend the Transformer model to multi-variable time series data, which are prevalent in complex systems such as process industries, energy systems, and on-orbit spacecraft.

3.3.1. Overall Structure of Transformer

The overall architecture of the Transformer, as shown in Figure 5, consists of two main components: the encoder and the decoder. The encoder extracts high-level features from the input data, while the decoder utilizes these features to generate predictions.
Figure 5. Transformer structure includes multiple encoders and decoders [].
The encoder consists of N identical layers, each comprising two key components:
  • Multi-Head Self-Attention Mechanism
  • Feed-Forward Network (FFN)
Given an input time series X = ( x 1 , x 2 , , x T ) , it is first processed through an embedding layer and positional encoding before being fed into the Transformer encoder:
H ( 0 ) = Embedding ( X ) + PE ( X )
The encoder computes representations as follows:
H ( l ) = FFN ( MultiHead ( H ( l 1 ) , H ( l 1 ) , H ( l 1 ) ) )
where
-
H ( l ) is the representation at layer l.
-
Multi-head attention allows the model to learn various time dependencies.
-
The Feed-Forward Network (FFN) applies nonlinear transformations to enhance feature representations.
The decoder has a similar structure but includes an additional cross-attention layer to incorporate information from the encoder output. The decoder computation is as follows:
Z ( l ) = FFN ( MultiHead ( Z ( l 1 ) , H , H ) )
where
-
Z ( l ) is the decoder representation at layer l.
-
The cross-attention mechanism allows the decoder to access encoder representations H , improving prediction performance.
The decoder’s final output is passed through a linear layer followed by a Softmax function to obtain the final prediction:
y ^ t = Softmax ( W o Z ( N ) + b o )
where
-
W o and b o are learnable parameters.
-
For classification tasks, Softmax outputs a probability distribution over classes.
-
For regression tasks, the model directly outputs the predicted values.
Multi-Head Self-Attention is a crucial component of the Transformer model, enabling it to dynamically assign varying levels of importance to different variables. The self-attention mechanism is formulated as follows:

3.3.2. Self-Attention Mechanism

The core of the Transformer model is the self-attention mechanism, which dynamically weighs the importance of different time steps and variables, thereby improving feature representation. The standard scaled dot-product attention mechanism is defined as follows:
Attention ( Q , K , V ) = softmax Q K T d k V
where
-
Q (Query): Represents the current time step information.
-
K (Key): Stores information about the entire input sequence.
-
V (Value): Stores the feature representations of the input sequence.
-
d k is the dimensionality of the key vectors.
The computation process involves the following:
  • Computing similarity: Performing the dot product Q K T .
  • Scaling: Dividing by d k to prevent gradient explosion.
  • Softmax normalization: Generating attention weights.
  • Weighted sum: Computing the final attention output.
To enhance feature learning, the Transformer employs multi-head attention, allowing multiple independent attention heads to focus on different aspects of the input:
MultiHead ( Q , K , V ) = Concat ( head 1 , head 2 , , head h ) W O
where
-
h is the number of attention heads.
-
Each head i represents an independent attention mechanism.
-
W O is a learnable linear transformation matrix.
Advantages of multi-head attention:
  • Enables the model to capture different levels of features, improving representation capability.
  • Avoids local optima that may arise from a single attention head.

3.3.3. Positional Encoding

Since the Transformer lacks inherent recurrence, it explicitly incorporates sequential information via positional encoding, defined as follows:
P E ( p o s , 2 i ) = sin p o s 10,000 2 i / d m o d e l
P E ( p o s , 2 i + 1 ) = cos p o s 10,000 2 i / d m o d e l
where
-
p o s is the position index.
-
i is the dimension index.
-
d m o d e l is the embedding dimension.
Functionality:
  • Ensures temporally close time steps have similar representations.
  • Enables the Transformer to recognize sequence order, improving time series modeling.

3.3.4. Multi-Variable Transformer Architecture

Since fault diagnosis is a classification task, we use only the Transformer encoder without the decoder. Given that multi-variable data includes both numerical (analog) and categorical (state) data, we design two separate Transformer encoders for feature extraction. Figure 6 illustrates the multi-variable Transformer structure.
Figure 6. Multi-variable Transformer architecture.
For analog variables, the embedding layer employs a fully connected Feed-Forward Network (FFN), whereas for categorical state variables, we utilize a Word2Vec-based embedding approach. Word2Vec maps discrete state variables into a continuous vector space, allowing them to be processed effectively by the Transformer encoder.
Finally, the outputs from both Transformer encoders are concatenated into a feature matrix (feature size M × sequence length N) and passed into an adaptive classifier to generate the final classification result.

3.4. Discussion on Time Complexity

During our research, we observed significant variations in computational time when using different types of neural networks as the backbone model for MAML, even with the same parameter scale. In fault diagnosis practices, if conditions permit, longer input sequences are often preferred to achieve higher diagnostic accuracy. Therefore, this section discusses the time complexity of MAML when LSTM, 1D CNN, and Transformer are used as the backbone models concerning the input sequence length N.
LSTM (Long Short-Term Memory) has a computational complexity of O ( d ) per time step, where d represents the hidden layer dimension. Therefore, for a sequence of length N, the total computational complexity of LSTM is O ( N d ) .
For a 1D Convolutional Neural Network (1D CNN), the computational complexity is primarily determined by the convolution operations. Assuming the input sequence length is N, the convolution kernel size is k, and the output dimension of the convolution layer is d (i.e., the number of channels), the computational complexity of the convolution operation is O ( N k d ) .
The computational complexity of a Transformer model is mainly attributed to the self-attention mechanism. The self-attention operation at each position involves pairwise interactions, leading to a complexity of O ( N 2 ) . Considering the feature dimension d, the computational complexity per layer is O ( N 2 d ) , resulting in an overall complexity of O ( N 2 d ) .
In MAML, gradient computation requires two forward passes and one backward pass. Therefore, the time complexity for different models combined with MAML is the following:
O ( N k d ) for CNN + MAML
O ( N d ) for LSTM + MAML
O ( N 2 d ) for Transformer + MAML
From a theoretical perspective, the influence of sequence length N is most significant for the Transformer + MAML approach. This implies that the computational cost of MVMT increases with longer sequences. However, in practical experiments, we observed that when the sequence length is relatively short, the absolute training time of a Transformer with the same scale is comparable to that of LSTM, while CNN remains the fastest. As the sequence length increases, the computational burden of LSTM rises more sharply than that of the Transformer, whereas the computational time of the Transformer and CNN increases at a slower rate. This discrepancy between theoretical complexity and actual runtime can be attributed to GPU parallelism.
  • 1D CNN: Convolution operations exhibit high parallelism, especially on GPU, allowing for more efficient acceleration compared to LSTM and Transformer models. This advantage is particularly evident when processing long sequences and large batches.
  • LSTM: While LSTM can leverage GPU for batch computations, its inherently sequential nature limits parallel efficiency. Since each time step must be computed before the next one begins, LSTM computations remain fundamentally serial, resulting in a complexity of O ( N d ) even on GPU.
  • Transformer: Owing to its fully parallelizable self-attention mechanism, Transformer models benefit significantly from GPU acceleration. Since attention computations at each position are independent, Transformers generally outperform LSTMs in terms of efficiency, particularly as sequence length N increases.
In conclusion, although Transformer models theoretically have higher computational complexity, GPU acceleration significantly mitigates their runtime cost in practice. Compared to LSTM, the Transformer becomes increasingly advantageous as the sequence length grows. Moreover, since MAML involves both inner-loop and outer-loop updates, the integration of the MAML algorithm further enhances the training efficiency of the Transformer model.

4. Experimental Setup

4.1. Datasets

We evaluate our approach on two datasets: Tennessee Eastman Process (TEP) dataset and Power-Supply System dataset.

4.1.1. Dataset 1: TEP

The Tennessee Eastman (TE) process is a well-established chemical process simulation system that represents a typical industrial chemical production process involving multiple unit operations and a set of adjustable variables. It simulates a large-scale chemical plant, with its core centered around an ethylene production process. Through precise modeling and simulation, the TE process captures common dynamic and nonlinear behaviors in chemical operations, including improper control actions, process faults, and other operational anomalies. It is widely utilized in research on process control, optimization, monitoring, and fault diagnosis. The TE dataset encompasses 20 typical fault modes observed in industrial processes, making it a standard benchmark for process monitoring and fault diagnosis studies. The process schematic diagram of TE process is shown in Figure 7. Further details can be found in Appendix A Table A1.
Figure 7. Tennessee Eastman process schematic diagram.
The TE process dataset consists of multiple subsets, with each fault mode containing both a training set and a testing set. The dataset size is as follows:
  • Training Set: The training set for each fault mode contains 480 samples.
  • Testing Set: The testing set for each fault mode contains 960 samples.
Each data sample contains 53 variables, including the following:
  • 40 Process Variables
  • 13 Manipulated Variables
A complete list of all variables is provided in Appendix A Table A2.
Considering that different fault types vary in detection difficulty, we allocated fault types with similar difficulty levels between the meta-training set and the meta-testing set based on diagnostic accuracy under sufficient sample conditions. The division of the fault types is shown in Table 1.
Table 1. TEP dataset: fault types in meta-training and meta-testing sets.
After dividing the fault types into meta-training and meta-testing sets, we construct two types of few-shot learning tasks:
  • Task 1: Each task contains 5 fault types. In the meta-training set, each fault type includes 1 support sample and 5 query samples, whereas in the meta-testing set, each fault type includes 1 support sample and 30 query samples. The length of each sample is 50.
  • Task 2: Each task contains 5 fault types. In the meta-training set, each fault type includes 3 support samples and 10 query samples, whereas in the meta-testing set, each fault type includes 3 support samples and 30 query samples. The length of each sample is 30.
The task construction details are summarized in Table 2.
Table 2. Task settings for the TEP dataset.

4.1.2. Dataset 2: Power-Supply System

This dataset comprises operational data from a real-world Power-Supply System of in-orbit spacecraft, including variables related to voltage, current, temperature, and other relevant parameters. It includes labeled instances of different fault conditions that occur in the system.
The dataset contains 24 types of faults, as listed in Appendix A Table A3.
Since the number of samples varies across different faults, we divide them into meta-training and meta-testing sets based on actual sample availability. The selected fault types for each set are listed in Table 3.
Table 3. Satellite data: fault types in meta-training and meta-testing sets.
Based on the division of fault types, we construct three types of few-shot learning tasks:
  • Task Setting 1: Each task in both the meta-training and meta-testing sets contains 5 fault types. The meta-training set includes 5 support samples and 10 query samples per task, while the meta-testing set includes 5 support samples and 30 query samples per task. The length of each sample is 30.
  • Task Setting 2: The meta-training set contains 5 fault types per task, whereas the meta-testing set contains 7 fault types per task. In the meta-training set, each task consists of 5 support samples and 10 query samples, while in the meta-testing set, each task consists of 5 support samples and 30 query samples. The length of each sample is 30.
  • Task Setting 3: Each task in both the meta-training and meta-testing sets contains 5 fault types. The meta-training set includes 3 support samples and 10 query samples per task, while the meta-testing set includes 3 support samples and 30 query samples per task. The length of each sample is 50.
The details of task construction are summarized in Table 4.
Table 4. Task settings for the satellite telemetry dataset.

4.2. Evaluation Metrics

The evaluation metrics are categorized into two dimensions: efficiency and performance.
For the efficiency dimension, we adopt the total meta-training time, denoted as T training , as the evaluation metric.
For the performance dimension, we measure the classification accuracy of the trained model (with optimal initialization) on the query samples of all tasks in the test set.

4.3. Hardware Platform

To ensure the reproducibility of the experiments, all experiments were conducted on the following hardware platform shown in Table 5.
Table 5. Experimental hardware configuration.

4.4. Comparison Methods

To evaluate the effectiveness of the proposed method, we compare it against the following state-of-the-art meta-learning-based fault diagnosis approaches:
  • LSTM-based meta-learning (meta-LSTM) []
  • 1D CNN-based meta-learning (meta-CNN) []
  • ViT-based meta-learning (meta-ViT) []
Furthermore, in Dataset 2, we incorporate the meta-SGD strategy to explore its impact on both efficiency and performance across different models.

4.5. Hyperparameter Settings

To ensure a rigorous and fair comparison, we conducted preliminary experiments using a sufficient number of samples from both Dataset 1 and Dataset 2. Each base model was optimized using Optuna, and the final tuned hyperparameters are presented in Table 6 and Table 7.
Table 6. Optimized hyperparameters for Dataset 1.
Table 7. Optimized hyperparameters for Dataset 2.
It is noteworthy that Dataset 2 contains both continuous (analog) and discrete (state) variables. Hence, the input feature size is configured to accommodate both data types.

5. Results and Discussion

5.1. Experimental Results on Dataset 1 (TEP)

Table 8 presents the results of training and testing directly on the meta-test task set without employing the MAML learning strategy. Here, MVT refers to Multi-Variable Transformer, which can handle multivariate time series but does not utilize the MAML learning method for training on the meta-test task set from scratch. Traditional deep learning models, such as 1DCNN and LSTM, perform relatively poorly, with accuracy below 40% in both settings, highlighting their limitations in handling small-sample, complex time series data. In contrast, Transformer-based models—ViT and the proposed MVT—demonstrate higher performance, with accuracy exceeding 43%. Notably, MVT achieves the highest accuracy in Setting 1 (44.15%), while ViT slightly outperforms it in Setting 2 (44.53%). These results indicate that, even without the MAML strategy, the proposed multi-variable Transformer architecture is effective in capturing complex temporal patterns, offering improved generalization over conventional models, especially in more challenging settings with only one support sample.
Table 8. Experimental results on Dataset 1: baseline models.
Table 9 summarizes the training time and classification accuracy of models incorporating the MAML strategy.
Table 9. Experimental results on Dataset 1: models with MAML strategy.
From the results, it is evident that the proposed MVMT model consistently achieves the highest accuracy across both task settings while requiring only half the training time of LSTM + MAML. Although 1DCNN + MAML exhibits the fastest training speed, significantly outperforming other methods in terms of efficiency, its final performance remains inferior to Transformer-based models and is comparable to LSTM. On the other hand, LSTM attains moderate accuracy but incurs the longest training time. Overall, these findings indicate that MVMT offers a clear and consistent advantage in few-shot fault diagnosis tasks for chemical processes.
Moreover, regardless of the model architecture, incorporating the MAML strategy enables the models to learn better initial parameters from the source domain, facilitating rapid adaptation to new tasks in the target domain.

5.2. Experimental Results on Dataset 2 (Power-Supply System)

Table 10 presents the experimental results on Dataset 2, where both analog and state variables are included. Under these conditions, MVMT demonstrates an even greater advantage.
Table 10. Experimental results on Dataset 2.
In Task Setting 1 (5-way 5-shot classification), MVMT achieves the highest fault diagnosis accuracy while maintaining stable and efficient training performance. In contrast, LSTM requires the longest training time but fails to deliver accuracy improvements proportional to its computational cost. It is also noteworthy that the MAML algorithm consistently enhances each model’s adaptation to target-domain tasks, as parameters learned from the source domain significantly improve accuracy over random initialization.
When meta-SGD is introduced, the accuracy of 1DCNN and LSTM improves, while training time remains stable or even decreases slightly—likely due to hardware performance fluctuations. However, for Transformer-based models, the application of meta-SGD does not yield significant accuracy gains. In fact, for ViT, accuracy decreases from 78.40% to 72.66%, making it even less effective than LSTM+meta-SGD.
In Task Setting 2, the meta-test tasks involve 7-way classification to evaluate whether the adaptive classifier can handle varying class numbers. The results indicate that the classifier successfully adapts to this scenario, with MVMT again achieving the highest accuracy (73.00%). Although 1DCNN maintains a training time advantage, its accuracy drops significantly compared to Task Setting 1, suggesting that its combination with the adaptive classifier is less effective than MVMT. Regarding the impact of meta-SGD, there is a slight accuracy improvement for 1DCNN and ViT, while its effect on MVMT and LSTM is negligible. Additionally, in the baseline models without MAML, MVT achieves an accuracy close to that of MAML-enhanced 1DCNN, highlighting MVT’s inherent adaptability to different tasks.
Task Setting 3 involves longer sequences but fewer support samples (3-shot). Even under these conditions, MVMT remains the best-performing model, significantly outperforming others. When meta-SGD is introduced, 1DCNN and LSTM show notable accuracy improvements, whereas Transformer-based models see minimal gains. This observation suggests that Transformer models exhibit similar parameter update patterns during training, reducing the need for meta-SGD to adjust learning rates for individual parameters.
We can observe that MVMT demonstrates high and stable performance across two different datasets, highlighting the portability of the proposed method. This portability primarily arises from its Model-Agnostic Meta-Learning (MAML) foundation and the flexible Transformer architecture. MAML allows the model to learn generalizable initialization parameters, enabling rapid adaptation to new fault scenarios with limited data. Additionally, Transformers are inherently capable of handling variable-length, multivariate time series data without requiring task-specific customization. These characteristics make MVMT well suited for deployment in a wide range of large-scale systems.

6. Conclusions

This paper presents a novel approach to fault pattern recognition in complex engineering systems using meta-learning, with a specific focus on fault diagnosis and classification. The limitations of existing methods are analyzed, and improvements are made to the Transformer model. We introduce the Multi-Variable Transformer based on the Self-Attention Mechanism (MVMT), which combines the rapid adaptability of meta-learning with the powerful feature extraction capabilities of Transformer networks. MVMT effectively handles multi-variable fault pattern recognition, particularly in small-sample scenarios, demonstrating significant promise in both practical applications and theoretical advancements.
The key contributions of this paper are as follows:
  • MVMT demonstrates exceptional feature extraction capabilities, particularly for time series data, and benefits from GPU-accelerated parallel computation, making it more computationally efficient compared to traditional RNNs and LSTMs.
  • The model excels in rapidly adapting to new fault tasks with minimal data, addressing the challenge of fault categories with limited samples, which is crucial for complex systems.
  • MVMT successfully integrates various types of variables (e.g., analog and status variables), offering high flexibility for deployment in diverse, multi-variable complex systems.
  • The introduction of an adaptive classifier solves the issue of varying numbers of fault categories across tasks, overcoming a limitation in the original MAML approach.
  • Comparative experiments conducted on the TE chemical process dataset and spacecraft telemetry dataset demonstrate that MVMT outperforms existing methods in multi-variable, small-sample fault diagnosis tasks for complex systems.
Looking forward, as the scale and complexity of engineering systems continue to grow, fault diagnosis tasks will face increasingly diverse and sophisticated challenges. One key direction for future research is to address real-time multi-fault pattern recognition in more complex systems. While traditional methods focus on detecting single fault modes, real-world applications often involve multiple concurrent fault modes or interactions between them. This necessitates the development of models capable of simultaneously recognizing and distinguishing multiple fault modes. Multi-Task Learning (MTL) methods, which optimize multiple fault diagnosis tasks concurrently, represent a promising avenue for enhancing model adaptability in such systems.
Another crucial area for future research is dealing with the increasing diversity of fault diagnosis data types, such as time series, image, and text data. Achieving effective cross-modal fusion and analysis will be vital for advancing fault diagnosis techniques. While Transformer models have shown success with multi-variable time series data, incorporating other data forms will require innovative approaches. For example, combining Graph Neural Networks (GNNs) with Transformer models could provide a better representation of global system information and the relationships between system components, improving fault diagnosis accuracy.
Lastly, as computational power continues to advance, applying fault diagnosis models in practical industrial environments—especially in smart manufacturing and the Industrial Internet of Things (IIoT)—will become an important research focus. In these applications, model interpretability is crucial, as industrial systems require a clear understanding of the decision-making process. Future research can focus on enhancing the transparency and explainability of deep learning models, aiming to develop fault diagnosis systems that are not only robust and efficient but also interpretable and understandable, meeting the needs of engineering practice.
By integrating and advancing these technologies, future complex engineering systems will be able to perform more intelligent fault detection and prediction, significantly enhancing system reliability and safety.

Author Contributions

Conceptualization, W.L. and F.Y.; methodology, W.L.; validation, W.L. and Y.N.; resources, F.Y.; data curation, W.L. and Y.N.; writing—original draft preparation, W.L.; writing—review and editing, W.L. and Y.N.; supervision, F.Y.; funding acquisition, F.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China, grant number 61873142.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data supporting the reported results are available at Github under the following link: https://github.com/YPCJ/MVMT.git (accessed on 20 March 2025).

Acknowledgments

The authors would like to thank China Satellite Network Group Co., Ltd for the data support during the experiments.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A. Dataset Detail

Table A1. Fault description in TEP dataset.
Table A1. Fault description in TEP dataset.
FaultDescription
IDV(1)A/C Feed Ratio, B Composition Constant (Stream 4) Step
IDV(2)B Composition, A/C Ratio Constant (Stream 4) Step
IDV(3)D Feed Temperature (Stream 2) Step
IDV(4)Reactor Cooling Water Inlet Temperature Step
IDV(5)Condenser Cooling Water Inlet Temperature Step
IDV(6)A Feed Loss (Stream 1) Step
IDV(7)C Header Pressure Loss - Reduced Availability (Stream 4) Step
IDV(8)A, B, C Feed Composition (Stream 4) Random Variation
IDV(9)D Feed Temperature (Stream 2) Random Variation
IDV(10)C Feed Temperature (Stream 4) Random Variation
IDV(11)Reactor Cooling Water Inlet Temperature Random Variation
IDV(12)Condenser Cooling Water Inlet Temperature Random Variation
IDV(13)Reaction Kinetics Slow Drift
IDV(14)Reactor Cooling Water Valve Sticking
IDV(15)Condenser Cooling Water Valve Sticking
IDV(16)Unknown
IDV(17)Unknown
IDV(18)Unknown
IDV(19)Unknown
IDV(20)Unknown
Table A2. Variables in TEP dataset.
Table A2. Variables in TEP dataset.
VariableDescriptionVariableDescription
XMV(1)D Feed Flow (stream 2) (Corrected Order)XMEAS(15)Stripper Level
XMV(2)E Feed Flow (stream 3) (Corrected Order)XMEAS(16)Stripper Pressure
XMV(3)A Feed Flow (stream 1) (Corrected Order)XMEAS(17)Stripper Underflow (stream 11)
XMV(4)A and C Feed Flow (stream 4)XMEAS(18)Stripper Temperature
XMV(5)Compressor Recycle ValveXMEAS(19)Stripper Steam Flow
XMV(6)Purge Valve (stream 9)XMEAS(20)Compressor Work
XMV(7)Separator Pot Liquid Flow (stream 10)XMEAS(21)Reactor Cooling Water Outlet Temp
XMV(8)Stripper Liquid Product Flow (stream 11)XMEAS(22)Separator Cooling Water Outlet Temp
XMV(9)Stripper Steam ValveXMEAS(23)Component A
XMV(10)Reactor Cooling Water FlowXMEAS(24)Component B
XMV(11)Condenser Cooling Water FlowXMEAS(25)Component C
XMV(12)Agitator SpeedXMEAS(26)Component D
XMEAS(1)A Feed (stream 1)XMEAS(27)Component E
XMEAS(2)D Feed (stream 2)XMEAS(28)Component F
XMEAS(3)E Feed (stream 3)XMEAS(29)Component A
XMEAS(4)A and C Feed (stream 4)XMEAS(30)Component B
XMEAS(5)Recycle Flow (stream 8)XMEAS(31)Component C
XMEAS(6)Reactor Feed Rate (stream 6)XMEAS(32)Component D
XMEAS(7)Reactor PressureXMEAS(33)Component E
XMEAS(8)Reactor LevelXMEAS(34)Component F
XMEAS(9)Reactor TemperatureXMEAS(35)Component G
XMEAS(10)Purge Rate (stream 9)XMEAS(36)Component H
XMEAS(11)Product Sep TempXMEAS(37)Component D
XMEAS(12)Product Sep LevelXMEAS(38)Component E
XMEAS(13)Prod Sep PressureXMEAS(39)Component F
XMEAS(14)Prod Sep Underflow (stream 10)XMEAS(40)Component G
XMEAS(41)Component H
Table A3. Fault descriptions of Power-Supply System.
Table A3. Fault descriptions of Power-Supply System.
NumberDeviceUnitFault Description
0Power ControllerBCRB ModuleSingle Drive Transistor Open or Short in MEA Circuit
1Power ControllerBCRB ModuleSingle Operational Amplifier Open or Short in MEA Circuit
2Power ControllerBCRB ModuleIncorrect Battery Charging Voltage Command
3Power ControllerPower Lower MachineCDM Module Current Telemetry Circuit Fault
4Power ControllerPower Lower MachinePOWER Module Bus Current Measurement Circuit Fault
5Power ControllerPower Lower MachineSolar Array Current Measurement Circuit Fault
6Power ControllerPower Lower MachineCommunication Fault
7Power ControllerGlobalLoad Short Circuit
8Power ControllerPower ModuleS3R Circuit Diode Short Circuit
9Power ControllerPower ModuleS3R Circuit Shunt MOSFET Open Circuit
10Power ControllerPower ModuleAbnormal S3R Shunt Status
11Power ControllerDistribution (Heater) ModuleY Board Heater Incorrectly On
12Power ControllerDistribution (Heater) ModuleBattery Heater Band Incorrectly Off
13Power ControllerDistribution (Heater) ModuleBattery Heater Band Incorrectly On
14Solar ArrayIsolation DiodeShort Circuit
15Solar ArrayIsolation DiodeOpen Circuit
16Solar ArrayInterconnect RibbonOpen Circuit
17Solar ArrayBus BarSolder Joint Open Circuit
18Solar ArraySolar CellSingle Cell Short Circuit
19Solar ArraySolar CellSingle Cell Open Circuit
20Solar ArraySolar CellSolar Cell Performance Degradation
21Solar ArraySolar WingSingle Subarray Open Circuit
22Solar ArraySolar WingSingle Wing Open Circuit
23Solar ArraySolar WingSolar Wing Single Subarray Open Circuit

References

  1. Hospedales, T.; Antoniou, A.; Micaelli, P.; Storkey, A. Meta-learning in neural networks: A survey. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5149–5169. [Google Scholar] [CrossRef] [PubMed]
  2. Finn, C.; Abbeel, P.; Levine, S. Model-agnostic meta-learning for fast adaptation of deep networks. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 1126–1135. [Google Scholar]
  3. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  4. Jin, Y.; Hou, L.; Chen, Y. A time series transformer based method for the rotating machinery fault diagnosis. Neurocomputing 2022, 494, 379–395. [Google Scholar] [CrossRef]
  5. Hou, S.; Lian, A.; Chu, Y. Bearing fault diagnosis method using the joint feature extraction of Transformer and ResNet. Meas. Sci. Technol. 2023, 34, 075108. [Google Scholar] [CrossRef]
  6. Fang, H.; Deng, J.; Bai, Y.; Feng, B.; Li, S.; Shao, S.; Chen, D. CLFormer: A lightweight transformer based on convolutional embedding and linear self-attention with strong robustness for bearing fault diagnosis under limited sample conditions. IEEE Trans. Instrum. Meas. 2021, 71, 1–8. [Google Scholar] [CrossRef]
  7. Hou, Y.; Wang, J.; Chen, Z.; Ma, J.; Li, T. Diagnosisformer: An efficient rolling bearing fault diagnosis method based on improved Transformer. Eng. Appl. Artif. Intell. 2023, 124, 106507. [Google Scholar] [CrossRef]
  8. Zhang, S.; Zhou, J.; Ma, X.; Pirttikangas, S.; Yang, C. TSViT: A Time Series Vision Transformer for Fault Diagnosis of Rotating Machinery. Appl. Sci. 2024, 14, 10781. [Google Scholar] [CrossRef]
  9. Nascimento, E.G.S.; Liang, J.S.; Figueiredo, I.S.; Guarieiro, L.L.N. T4PdM: A deep neural network based on the transformer architecture for fault diagnosis of rotating machinery. arXiv 2022, arXiv:2204.03725. [Google Scholar]
  10. Yang, Z.; Cen, J.; Liu, X.; Xiong, J.; Chen, H. Research on bearing fault diagnosis method based on transformer neural network. Meas. Sci. Technol. 2022, 33, 085111. [Google Scholar] [CrossRef]
  11. Kong, X.; Cai, B.; Zou, Z.; Wu, Q.; Wang, C.; Yang, J.; Wang, B.; Liu, Y. Three-model-driven fault diagnosis method for complex hydraulic control system: Subsea blowout preventer system as a case study. Expert Syst. Appl. 2024, 247, 123297. [Google Scholar] [CrossRef]
  12. Kong, X.; Cai, B.; Liu, Y.; Zhu, H.; Yang, C.; Gao, C.; Liu, Y.; Liu, Z.; Ji, R. Fault diagnosis methodology of redundant closed-loop feedback control systems: Subsea blowout preventer system as a case study. IEEE Trans. Syst. Man Cybern. Syst. 2022, 53, 1618–1629. [Google Scholar] [CrossRef]
  13. Li, C.; Li, S.; Zhang, A.; He, Q.; Liao, Z.; Hu, J. Meta-learning for few-shot bearing fault diagnosis under complex working conditions. Neurocomputing 2021, 439, 197–211. [Google Scholar] [CrossRef]
  14. Chang, L.; Lin, Y.H. Meta-Learning With Adaptive Learning Rates for Few-Shot Fault Diagnosis. IEEE/ASME Trans. Mechatron. 2022, 27, 5948–5958. [Google Scholar] [CrossRef]
  15. Li, X.; Li, X.; Li, S.; Zhou, X.; Jin, L. Small sample-based new modality fault diagnosis based on meta-learning and neural architecture search. Control Decis. 2023, 38, 3175–3183. [Google Scholar]
  16. Ren, L.; Mo, T.; Cheng, X. Meta-learning based domain generalization framework for fault diagnosis with gradient aligning and semantic matching. IEEE Trans. Ind. Inform. 2023, 20, 754–764. [Google Scholar] [CrossRef]
  17. Wikimedia Commons, the Free Media Repository. File:The-Transformer-model-architecture.png. 2025. Available online: https://commons.wikimedia.org/wiki/File:The-Transformer-model-architecture.png (accessed on 25 March 2025).
  18. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  19. LeCun, Y.; Bengio, Y. Convolutional networks for images, speech, and time series. In The Handbook of Brain Theory and Neural Networks; MIT Press: Cambridge, MA, USA, 1995; p. 3361. [Google Scholar]
  20. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 4 May 2021. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.