Next Article in Journal
DistMLLM: Enhancing Multimodal Large Language Model Serving in Heterogeneous Edge Computing
Previous Article in Journal
Compensation Strategies in Post-Stroke Individuals: Insights from Upper Body Kinematics Analysis Based on Inertial Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Meta Variational Memory Transformer for Anomaly Detection of Multivariate Time Series

1
Shanghai Aerospace Electronic Communication Equipment Institute, Shanghai 201109, China
2
National Key Laboratory of Radar Signal Processing, Xidian University, Xi’an 710071, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(24), 7611; https://doi.org/10.3390/s25247611
Submission received: 18 November 2025 / Revised: 10 December 2025 / Accepted: 12 December 2025 / Published: 15 December 2025
(This article belongs to the Section Physical Sensors)

Abstract

Detecting anomalies in multivariate time series (MTS) is a crucial task in areas like financial fraud detection and industrial equipment monitoring. Recent research has focused on developing unsupervised probabilistic models to identify anomalous patterns within MTS. However, many of these methods rely on fixed parameter mappings for each MTS, resulting in high computational costs and limited adaptability. To overcome these challenges, we introduce a novel Meta Variational Memory Transformer (MVMT). MVMT captures the diverse patterns across various MTS by encoding them into a set of memory units using a specially developed meta memory attention (MMA) module. Utilizing these learned memory units, we introduce a memory-guided probabilistic generative model that selects relevant memories as priors for latent states, resulting in more expressive MTS representations. A key feature of MVMT is that MMA provides a diversified prior in the latent space, ensuring the generation of various patterns. Finally, we implement a Transformer-based upward–downward variational inference process to estimate the posterior distribution of latent variables. Our extensive experiments on six datasets demonstrate the effectiveness of MVMT in one-for-all anomaly detection tasks.

1. Introduction

Anomaly detection in MTS stands as a critical scenario in Information Technology (IT) operations, particularly vital for financial fraud detection and industrial equipment monitoring, such as server machines within cloud infrastructures [1,2,3], content delivery networks [4], and more. Traditional statistical techniques are effective in simpler cases, such as univariate time-series anomaly detection [5,6,7,8,9]. However, their effectiveness may wane in more intricate and massive MTS scenarios. Consequently, attention has shifted towards machine learning-driven anomaly detection techniques, which are generally categorized into supervised learning approaches [5,10,11] and unsupervised methods [1,2,8,12,13,14,15,16].
Since anomalies are often infrequent and hidden within large volumes of normal data, identifying and labeling them is both challenging and costly. Moreover, supervised anomaly detection models often suffer from suboptimal performance due to severe class imbalance [17,18]. Consequently, unsupervised anomaly detection has gained significant traction in recent years. A major obstacle in unsupervised anomaly detection lies in effectively modeling the normal patterns in MTS. Among the various unsupervised methods, probabilistic dynamical models like OmniAnomaly [1] and SDFVAE [4] have excelled by capturing temporal correlations and variations inherent in time series. The emergence of Transformer architectures [19] and their ability to capture long-range dependencies [20,21,22,23] have further motivated their adoption for time-series modeling, and recent Transformer-based anomaly detectors [24] have shown strong representational capability.
Since anomalies are often infrequent and hidden within large volumes of normal data, identifying and labeling them is both challenging and costly. Moreover, supervised anomaly detection models often suffer from suboptimal performance due to severe class imbalance [17,18]. Consequently, unsupervised anomaly detection has gained significant traction in recent years. A major obstacle in unsupervised anomaly detection lies in effectively modeling the normal patterns in MTS. Among the various unsupervised methods, probabilistic dynamical models like OmniAnomaly [1] and SDFVAE [4] have excelled by capturing temporal correlations and variations inherent in time series. The emergence of Transformer architectures [19] and their ability to capture long-range dependencies [20,21,22,23] have further motivated their adoption for time series modeling, and recent Transformer-based anomaly detectors [24] have shown strong representational capability. In contrast to physics-embedded MTS models [25], which rely on physical constraints or structure-preserving mappings, our method adopts a purely data-driven probabilistic framework. Rather than enforcing physics-guided latent spaces, we introduce a meta memory prior to capture diverse temporal patterns across heterogeneous MTSs, making the approach suitable when explicit physical laws are unavailable. However, these models typically learn a single shared representation and lack an explicit mechanism to encode and differentiate the diverse normal patterns that arise across heterogeneous MTS sources. As illustrated in Figure 1, different devices or subsystems may exhibit fundamentally different operational modes, making it difficult for a single Transformer encoder to generalize. Without structural priors, attention alone must implicitly capture all such variations, which often leads to underfitting of rare-but-normal behaviors and reduced robustness when intra-class variability is high [26]. These limitations motivate the introduction of a meta memory prior that can store, organize, and retrieve prototypical temporal patterns learned across multiple MTSs. By providing diversified and informative priors for latent states, such a mechanism fills the gap left by existing Transformer-based anomaly detectors and enables more expressive and adaptable representations in a unified one-for-all anomaly detection framework.
To overcome the aforementioned limitations, we introduce an innovative memory-guided probabilistic generative model tailored to capture intricate and diverse temporal dependencies across multiple MTSs. Our approach draws inspiration from meta-learning [27], aiming to extract transferable patterns across related tasks. We devised the Meta Memory Attention (MMA) mechanism to identify and retain complex and diverse temporal information shared among multiple MTSs, storing it within the memory units as illustrated in Figure 2. Subsequently, we developed a memory-guided probabilistic generative model within a hierarchical Bayesian framework. This model leverages the diverse temporal dependencies by employing the stored memory from MMA as a prior, thereby facilitating the acquisition of expressive representations crucial for anomaly detection within MTS data. To optimize the inference process for our model, we implement an innovative upward-downward autoencoding variational inference strategy. This approach integrates both bottom-up likelihood assessments and top-down prior knowledge of parameters to achieve a more accurate posterior approximation. Additionally, we enhance the inference network by integrating an anomaly transformer, which refines the latent representations. The key contributions of our work are summarized as follows:
  • To consider diverse nomal patterns within MTSs and achieve anomaly detection, we propose a hierarchical Bayesian network named MVMT. MVMT consists of a developed MMA module that captures and stores various temporal dependencies of MTSs, a memory-guided generative module to capture the shared diverse temporal information, as well as a transformer-powered inference module for accurate posterior approximation.
  • We carefully design an upward–downward variational inference model built upon the transformer blocks for MVMT to approximate the posterior for the latent variables.
  • While most existing methods rely on distinct parameter sets for each MTS, the proposed MVMT utilizes a single group of parameters to model diverse MTSs, allowing it to effectively capture multiple patterns within a unified framework.
  • We perform thorough experiments on six real-world datasets. The quantitative comparison results demonstrate that MVMT surpasses state-of-the-art methods in terms of F1-score with fewer parameters. The qualitative analysis reveals that it effectively captures diverse dynamic memory patterns within different MTSs.
Figure 2. Meta Memory Attention (MMA). We constructed this module by incorporating a multi-headed attention mechanism. This approach significantly enhances our model’s memory recall capability.
Figure 2. Meta Memory Attention (MMA). We constructed this module by incorporating a multi-headed attention mechanism. This approach significantly enhances our model’s memory recall capability.
Sensors 25 07611 g002

2. Related Work

Recently, unsupervised anomaly detection has become a focal point for operational engineers. Many of these methods adopt an encoding-reconstruction approach, leveraging architectures like AE, VAE, or GAN. To model the characteristics of MTS, these methods are typically based on either RNNs [28] or Transformers [19]. For instance, EncDec-AD [15] uses an LSTM-based AE architecture to capture temporal dependencies in MTS, detecting anomalies through deviations from reconstruction. Similarly, Telemanom [16] leverages LSTM to predict spacecraft telemetry channel values and detects anomalies by evaluating the residual error between predicted and observed values. Additionally, MAD-GAN [13] employs LSTM-RNN within a GAN to model typical temporal and spatial patterns. OmniAnomaly [1] uses a stochastic recurrent neural network (SRNN) to robustly learn representations, while SDFVAE [4] introduces a static and dynamic factorized VAE approach specifically addressing both time-invariant and time-varying elements. VAE-based models, such as OmniAnomaly and SDFVAE, detect anomalies by evaluating likelihood, with lower likelihood values suggesting a higher probability of an anomaly. It is important to highlight that these unsupervised approaches are largely RNN-based, concentrating on capturing temporal dependencies in MTSs. However, there is a recent inclination towards leveraging transformers, known for their efficiency in capturing intricate and long-range temporal dependencies, as well as in MTS anomaly detection [29]. Anomaly transformer [24] has emerged as the top performer in MTS anomaly detection thus far.
Although current unsupervised anomaly detection methods have shown effectiveness in practical applications, capturing diverse temporal dependencies across multiple MTSs remains a significant challenge. The common approach of training separate models for each MTS limits comprehensive modeling, leading to high computational costs and reduced adaptability. To address this issue, GmVRNN [30] seeks to handle diverse MTSs with a single set of parameters by using a Gaussian mixture distribution for latent states. Despite its promising results, this approach often struggles to capture all patterns effectively, as a single parameter set may not be sufficient, requiring adjustments for different MTSs to ensure optimal performance. This paper introduces MVMT, a pioneering Transformer-based probabilistic model. MVMT captures the diverse patterns across various MTS by encoding them into a set of memory units using a specially developed meta memory attention (MMA) module. Utilizing these learned memory units, we introduce a memory-guided probabilistic generative model that selects relevant memories as priors for latent states, resulting in more expressive MTS representations.

3. Preliminaries

This section presents a brief overview of the problem definition, the MVMT-based anomaly detection approach, and the data processing methodology. Additionally, we clarify the meanings of the symbols used in the paper, as detailed in Table 1.

3.1. Problem Definition

Consider the MTS as X = { x 1 , x 2 , . . . , x T l } , where T l represents the total length of X , and  x t R V is the observation at time t, with V indicating the dimensionality at each timestep. Thus, X R T l × V . Anomaly detection in MTS involves determining whether a specific observation at a given time, denoted as x t , is anomalous. In an unsupervised approach, the model is trained exclusively on normal data to capture normal patterns in MTS. During the testing phase, reconstruction errors are used as anomaly scores. The anomaly score function should produce higher values for anomalous MTS compared to normal ones.

3.2. Anomaly Detection Based on MVMT

The overarching structure for unsupervised anomaly detection utilizing MVMT is depicted in Figure 3. Our framework consists of three components: MTS preprocessing, representation model, and anomaly detection modules. Specifically, MTS preprocessing is introduced in Section 3.3, and MVMT is employed as the representation model. The reconstruction probability inferred by MVMT is utilized to guide the detection of anomalies.
As the model learns the typical patterns of normal MTS, observations that closely match these learned behaviors are reconstructed with high accuracy and confidence. Therefore, we use the reconstruction error as an anomaly score to assess deviations from expected behavior [1,2,8,31]. The anomaly score is computed as follows:
S t = log p ( x t z t , m a )
where S t is the anomaly score, and  m a and z are two components in the latent space of the proposed model, which will be explained in detail in Section 4. Because our model is built within a Bayesian probabilistic framework, the reconstruction term is naturally expressed as the log-likelihood. This choice ensures consistency with the hierarchical VAE derivation, whereas MSE residuals correspond to deterministic autoencoders and are not directly aligned with our probabilistic formulation. To classify an observation x t as anomalous, we compare its anomaly score S t against a predefined threshold. We use the POT method [32] to set this threshold. In MVMT, smaller anomaly scores are deemed more likely to indicate outliers. Consequently, similar to [1], we select thresholds based on lower-bound criteria to identify extreme values.

3.3. Data Preprocess

Similar to the approach in [33], for observed MTSs, we initially normalize them as
x t x t min ( X ) max ( X ) min ( X ) + ϵ
where max ( X ) and min ( X ) denote the maximum and minimum values for each feature across the training time series, respectively, and  ε is a small constant (set to 10 6 ) introduced to avoid division-by-zero instability, and the results are insensitive to its value within a reasonable numerical range. After normalizing the MTSs, we employ a sliding window technique with overlap, following [4], for dataset preprocessing. Here, the window size T = 20 , and there is an overlap of l = T 1 between two consecutive observed variables. Notably, to capture both time-varying and time-invariant information, we augment each timestep with its w neighbors, set to 5 for anomaly detection tasks.

4. Methodology

In this section, we outline the key concepts of the proposed MVMT framework, followed by an in-depth discussion of the MMA model, including its generative and inference components. To facilitate understanding, Table 1 provides a summary of the symbols used throughout the paper.

4.1. Meta Variational Memory Transformer

As depicted in Figure 4, MVMT establishes a memory model based on the transformer architecture from a variational Bayesian perspective. Our approach introduces a latent space comprising two components: m a and z . To capture and retain diverse patterns, we devised the MMA module, constructed upon a set of memory units organized using the attention mechanism. m a serves as a memory prior for z , ensuring an expressive latent representation and facilitating the generation of MTS with diverse temporal information. In contrast to the conventional VAE [34], MVMT defines the latent variable z with a memory prior ( m a ), integrating diverse temporal information from multiple MTSs. This structuring of the latent space is represented as follows in equation form:
p ( x ) = m a z p ( x z , m a ) p ( z m a ) p ( m a ) d m a d z
Based on the latent space defined in Equation (3), we present a refined MMA model alongside a Bayesian generative framework to effectively capture and represent the full latent space.

4.1.1. Meta Memory Attention Module

Global diverse temporal information within multiple MTSs is crucial for modeling multiple MTSs with a single model [30]. Inspired by Concept Transformer (CT) [35], we aim to capture and store diverse temporal information within MTS through a meta memory module, employing an attention module to facilitate memory operations. Specifically, considering N memory units M R N × d m and MTS features h R d m , where d m represents the vector dimension, M consists of vectors optimized by a well-designed loss function proposed later. Subsequently, an attention mechanism is utilized to derive the memory embedding M a . As illustrated in Figure 2, the input feature h initially undergoes mapping via a MLP to generate the query matrix Q R N × d m for the query-key-value cross-attention module. Following this, the memory units M serve as input to predict the key matrix K R N × d m and value matrix V R N × d m . The attention weights matrix A = [ α ] R N × N can be obtained from Q and K as
α = softmax 1 d m Q K
Final result from the MMA is computed by applying the attention map A to the value matrix V , yielding the weighted output.
M a = i = 1 N [ A V ] i
As shown in Figure 4a, with the memory embedding ( M a ) defined in Equation (5), we introduce a memory latent variable m a :
p ( m a M a ) N m a μ m a , diag σ m a 2 μ m a = G μ M a ( M a ) = f ( W μ M a M a + b μ M a ) σ m a 2 = G σ M a ( M a ) = f ( W σ M a M a + b σ M a )
where m a is treated as a Gaussian random variable, G μ M a and G σ M a are two functions used to generate m a from M a . Specifically, we utilize a fully connected network as G M a , and  f ( · ) represents a deterministic non-linear transition function, where W M a R d m × d m denotes the transition matrices, and  b M a R d m represents the bias weights. From Equation (6), m a is conditioned on the memory embedding M a . The advantage of the attention mechanism in MMA allows m a to provide powerful memory information for subsequent models. The memory units are initialized with small random values and are jointly updated with all model parameters through standard back-propagation under the variational objective, which ensures stable training and practical convergence.

4.1.2. Probablistic Memory Generative Model

Figure 4a illustrates how we model the diverse temporal dependencies in MTS by utilizing the multiple patterns stored in the MMA module. To achieve this, we design a probabilistic memory generative model based on a hierarchical Bayesian framework [36], where the latent variable is defined as
p ( z m a ) N z μ z , diag σ z 2 μ z = F μ m a ( m a ) = f ( W μ m a m a + b μ m a ) σ z 2 = F σ m a ( m a ) = f ( W σ m a m a + b σ m a )
where F μ m a and F σ m a are two functions used to generate z from m a . Specifically, we employ a fully connected network as F m a , and  f ( · ) represents a deterministic non-linear transition function. Here, W m a R d m × d m denotes the transition matrices, and  b m a R d m represents the bias weights. Unlike traditional VAEs, where the latent variable z is typically conditioned on a standard Gaussian distribution, in our model, z is conditioned on the latent variable m a . This enables MVMT to consider diverse temporal dependencies based on the meta-information stored in MMA. As depicted in Equation (6), p ( m a M a ) provides globally diverse temporal information for each latent variable. Similar to [30], for considering the stochasticity within MTS, we assign the generated x to follow a Gaussian distribution, defined as
p ( x z ) N x μ x , d i a g σ x 2
where μ x and σ x 2 are mean and variance parameters that are conditioned on the hidden variable z as
μ x = F μ z ( z ) = f ( W μ z z + b μ z ) σ x = F σ z ( z ) = f ( W σ z z + b σ z )
F μ z and F z σ are two functions used to generate x from z . Similar to Equation (7), we utilize a fully connected network as F m a , where model parameters W z and b z have the same settings. This approach enables the model to transfer the diverse temporal information stored in the MMA from m a to z , ultimately aiding in reconstructing x . The hierarchical information generation allows the model to consider the diversity within multiple MTS and acquire more expressive and robust representations for them. This is essential for unsupervised anomaly detection methods to enhance their performance.

4.1.3. Inference Model

As with most Bayesian models, directly computing the exact posterior distribution is intractable, a challenge common in VAEs [34,37]. To address this, we define the variational distribution q ( m a , z x , M a ) , which must be sufficiently flexible to closely approximate the true posterior. To efficiently approximate the posterior for the hierarchically structured latent variables in MVMT, we design an inference network that integrates both bottom-up likelihood information and top-down prior information, allowing for a more accurate approximation of the generative process. The variational distribution is defined as
q ( m a , z M a , x ) = q ( m a M a , x ) q ( z m a , x )
Specifically, to capture the complex temporal characteristics of MTSs, we initially employ an anomaly-transformer-like block (using the model structure but without the minimax strategy proposed by [24]) to extract the features of MTSs, and the parameters of the block are defined as W x h 0 . As depicted in Figure 4c, for the upward flow of information, we compute the variable h 0 through the anomaly transformer and then utilize a fully connected network to obtain the variables h 1 and h as follows:
h 0 = Anomaly   Transformer ( x ) h 1 = f ( W h 0 h 1 h 0 + b h 0 h 1 ) h = f ( W h 0 h h 0 + b h 0 h )
As depicted in Figure 4b,c, we estimate m a by integrating the extracted latent features with prior knowledge from the MMA module, constructing the variational posterior for m a as follows:
q ( m a M a , h 1 ) N m a μ ˜ m a , diag σ ˜ m a 2 μ ˜ m a = Φ ( M a , h 1 ) = f ( W ˜ m a μ M a + V ˜ m a μ h 1 + b ˜ m a μ ) σ ˜ m a = Φ ( M a , h 1 ) = f ( W ˜ m a σ M a + V ˜ m a σ h 1 + b ˜ m a σ )
where { W ˜ m a μ , V ˜ m a μ , W ˜ m a σ , V ˜ m a σ   } R d m × d m and { b ˜ m a μ , b ˜ m a σ }   R d m are the parameters that the inference model learns during training. As a hierarchical VAE, the inference model requires separate latent flows: h 0 estimates the posterior of the lower-level latent variable z , while h 1 provides a higher-level approximation for m a , ensuring that deeper latent layers are inferred from correspondingly more abstract features. This approach allows m a to be estimated by integrating both the bottom-up likelihood information present in h 1 from the inference network and the contextual memory from the MMA module. By combining these sources, the proposed inference network effectively captures long-range and diverse temporal dependencies, leading to a rich latent representation for MVMT. The latent variable q ( m a M a , h 1 ) is then reparameterized as
m a = μ ˜ m a k + σ ˜ m a k ϵ , ϵ Uniform ( 0 , 1 )
which allows for back-propagation through z . MVMT combines the obtained latent features by the inference network with the prior from m a to construct the variational posterior as
q ( z m a , h 0 ) N z μ ˜ z , diag σ ˜ z 2 μ ˜ z = Ψ ( m a , h 0 ) = f ( W ˜ z μ m a + V ˜ z μ h 0 + b ˜ z μ ) σ ˜ z = Ψ ( m a , h 0 ) = f ( W ˜ z σ m a + V ˜ z σ h 0 + b ˜ z σ )
where { W ˜ z μ , V ˜ z μ , W ˜ z σ , V ˜ z σ   } R d m × d m and { b ˜ z μ , b ˜ z σ }   R d m denote the learnable parameters that map m a and h 0 to z . This allows z to be inferred by integrating the likelihood information from h 0 with the prior knowledge provided by m a , thus facilitating the generation of rich latent representations for MVMT. Additionally, q ( z m a , h 0 ) can be reparameterized as
z = μ ˜ z + σ ˜ z ϵ , ϵ Uniform ( 0 , 1 )
which facilitates the use of the back-propagation algorithm. We note that MVMT incorporates transformer structure into inference processes, which enables it to model complex temporal dependencies within MTS efficiently in both processes, thus guaranteeing expressive latent representation learning and promising anomaly detection performance.

4.2. Model Properties

MVMT leverages the advantages of the hierarchical Bayesian model and the meta memory model enhanced by the attention mechanism, as detailed below.

4.2.1. Learnable Memory Prior

The general VAE [34] uses the standard normal distribution as the prior, which does not contain information. We designed the MMA to organize and store the information from the big data and select the corresponding memory for different MTS to generate a dedicated prior. This prior gives our model the ability to adapt quickly to the new MTS. Moreover, we designed an attention-based meta memory module to efficiently manage global information and provide high-quality meta memory priors.

4.2.2. One Model for All MTSs

Diverse MTSs often exhibit varying characteristics, making it difficult for anomaly detection models to fully capture this complexity using a single parameter set. MVMT overcomes this by utilizing the MMA and memory generative modules to capture and store a range of dynamics, enabling the model to handle diverse MTSs with a unified set of parameters—a trait we refer to as the one-for-all property. This property can not only ensure MVMT to manage large MTS data but also adapt to new MTS with few observations, which will be proved in the experiment.

4.2.3. Probability Modeling of Hidden Space and Generative Process

Previous deterministic transformer-based approaches may face challenges in modeling the variability present in highly structured MTS data, characterized by complex and strong dependencies across different timesteps. To overcome this, we introduce uncertainty into transformers by applying probabilistic modeling in both the hidden space and the generative process. Following similar arguments made by others [38], we propose that this probabilistic framework is particularly useful when the true data distribution is multimodal, which reflects the diverse nature of multivariate time-series data.

4.2.4. Transformer for Time Series

Transformers have demonstrated efficiency in capturing complex, long-range temporal dependencies. MVMT applies transformer structure in both memory and inference modules, which enables our model to learn expressive memories and latent representations. Especially for multiple MTSs data, there are diverse patterns within them; traditional recurrent-structured networks may not be powerful enough to capture there dynamics, while our model can handle them efficiently with transformer structure.

4.2.5. Beneficial to Process Large Data

MVMT is designed to efficiently handle large-scale MTS data. The learnable prior allows it to capture and store diverse characteristics within the data, while the mixture-distributed latent states help model various structural patterns and non-stationary temporal dependencies across different MTSs. By incorporating probabilistic modeling in the generative process, MVMT accounts for the variability present in complex MTS data. Additionally, the transformer architecture in both the inference and memory modules enhances its ability to capture diverse patterns. The ’one-for-all’ setting of MVMT can avoid the drawback of previous methods that the parameters grow exponentially facing to large data.

5. Model Training

To train the model end-to-end, we establish an objective function. This reconstruction-based approach captures the latent representation of the entire MTS. For MVMT, given the model parameters W , the marginal likelihood of the MTS dataset D is formulated as
p ( D W ) = m a z p ( x z , m a ) p ( z m a ) p ( m a ) d m a d z
Consistent with VAEs, MVMT aims to optimize by maximizing the evidence lower bound (ELBO) on the log marginal likelihood. The log marginal likelihood for MVMT is then expressed as
log p ( x ) = E z q ( z m a , x )   m a q ( m a M a , x ) [ log p ( x ) ]
Applying Bayes’ rule [39] for p ( x ) , we incorporate the terms q ( z m a , x ) , p ( z m a ) , p ( m a M a ) , and q ( m a M a , x ) to the log p ( x ) as
log p ( x ) = E [ log p ( x z ) p ( z ) p ( z x ) q ( z m a , x ) q ( z m a , x ) p ( z m a ) p ( z m a ) p ( m a M a ) p ( m a M a ) q ( m a M a , x ) q ( m a M a , x ) ]
Organizing Equation (17) can conclude
log p ( x ) = E [ log p ( x z ) ] E [ log q ( z m a , x ) p ( z m a ) ] E [ log q ( m a M a , x ) p ( m a M a ) ] + E [ log q ( m a M a , x ) p ( m a M a ) ] + E log q ( z m a , x ) p ( z m a ) + E [ log p ( z ) p ( z x ) ]
Next, applying the Kullback–Leibler (KL) Divergence, defined as D K L ( p q ) = E [ log ( p / q ) ] , we can further reformulate Equation (18) as
log p ( x ) = E [ log p ( x z ) ] D K L ( q ( z m a , x ) p ( z m a ) ) D K L ( q ( m a M a , x ) p ( m a M a ) ) + D K L ( q ( m a M a , x ) p ( m a M a ) ) + D K L ( q ( z m a , x ) p ( z m a ) ) + D K L ( p ( z ) p ( z x ) )
By applying Jensen’s inequality, we can derive the following expression for the ELBO:
log p ( x ) E [ log p ( x z ) ] D K L ( q ( z m a , x ) p ( z m a ) ) D K L ( q ( m a M a , x ) p ( m a M a ) )
Drawing inspiration from beta-VAE [40], we incorporate two hyperparameters, β 1 > 0 and β 2 > 0 , which function as temperature coefficients. These coefficients are used in a ‘warm-up’ procedure, where we initially train the model focusing solely on reconstruction error. Gradually, over the first N training epochs, the KL divergence loss is introduced, applying an additional warm-up factor that increases linearly from 0 to 1 during training. This gradual introduction helps address the issue of posterior collapse, resulting in the final optimization target:
L = E q ( z ) , q ( m a ) [ log p ( x z ) ] β 1 D K L ( q ( z m a , x ) p ( z m a ) ) β 2 D K L ( q ( m a M a , x ) p ( m a M a ) )
Unlike the conventional ELBO, which primarily focuses on the expected log-likelihood for maintaining reconstruction accuracy and the KL divergence for aligning the variational distribution of z with the prior, the ELBO of MVMT learns a set of meta memory prior (M) to guide latent variable m a . The information stored in the memory units passes to z through a KL constraint between m a and z . The prior of z is obtained from the posterior probability of the m a , so an additional KL is needed to give the m a prior. As demonstrated in Algorithm 1, the model parameters are optimized using stochastic gradient descent, allowing for an end-to-end training process. The algorithm implements an autoencoding variational inference strategy, employing an upward–downward scheme, to train and fine-tune the parameters essential for the MVMT model. Leveraging a pre-processed Multivariate Time Series (MTS) training dataset, this process encompasses encoding, memory management, latent variable inference, decoding, and iterative parameter updates, ensuring comprehensive model training and optimization.
Algorithm 1 Upward–Downward Autoencoding Variational Inference for MVMT
  • Input:
  • The pre-processed MTS training dataset D ( x 1 : T ) .
  • Parameter: The encoder parameters Ψ = { W x h 0 , W h 0 h 1 , b h 0 h 1 , W h 0 h , b h 0 h , W ˜ m a μ , V ˜ m a μ , b ˜ m a μ , W ˜ m a σ , V ˜ m a σ , b ˜ m a , W ˜ z μ , V ˜ z μ , b ˜ z μ , W ˜ z σ , V ˜ z σ , b ˜ z }
  • The decoder parameters Φ = { W μ M a , b μ M a , W σ M a , b σ M a ,   W μ m a , b μ m a ,   W σ m a , b σ m a , W μ z , b μ z , W σ z , b σ z , }
  • The meta memory units parameters M .
  • Output:
  • Parameters after training Ψ , Φ , M .
1:
Set mini-banch size as N and hyperparameters β 1 , β 2 ;
2:
Initialize the encoder parameters Ψ , decoder parameters Φ and meta memory units parameters M ;
3:
while epoch do
4:
   Randomly select a mini-batch { x 1 : T i } i = 1 N in D ( x 1 : T ) ;
5:
   Extract features h = { h 0 , h 1 , h } using encoder;
6:
   Inference the meta memory posterior q ( m a ) through MMA (Equation (12));
7:
   Inference the latent variable z through Equation (14);
8:
   Input the latent variable to get the decoder result;
9:
   Update the parameters( Ψ , Φ , M ) by Equation (21);
10:
end while
11:
return Parameters after training Ψ , Φ , M .

6. Experimental Evaluation

This section begins by detailing the experimental setup, followed by a comprehensive evaluation of our model through various experiments.

6.1. Experiment Setup

6.1.1. Datasets

Our experiments involve five datasets: (1) SMD [1], a 5-week dataset collected from a large internet company, consisting of 38 dimensions. (2) MSL (Mars Science Laboratory rover) and SMAP (Soil Moisture Active Passive satellite), which are publicly available NASA datasets [16] with 55 dimensions. Both datasets contain telemetry anomaly data sourced from Incident Surprise Anomaly (ISA) reports for spacecraft monitoring. (3) PSM [41], an internally collected dataset from multiple application server nodes at eBay, containing 26 dimensions. (4) SWaT [42], collected from 51 sensors monitoring a critical infrastructure system during continuous operation. (5) DND [43], a real-world multivariate KPI dataset from a major internet company in China, monitoring 12 websites with 36 KPIs per site. These websites offer various services, such as Video on Demand (VoD) and live streaming. The KPIs span roughly 1.5 months, with data points collected every 60 s. For each website, the first half of the KPIs are used for training and the second half for testing. Human operators confirmed the ground-truth anomalies in the DND testing data. Table 2 provides the basic statistical details of the datasets.

6.1.2. Baselines

We compare our model with 19 baselines across various categories. For reconstruction-based models, we evaluate against InterFusion [44], BeatGAN [45], OmniAnomaly [1], and LSTM-VAE [46]. For density-estimation methods, we include DAGMM [47], MPPCACD [48], and LOF [49]. Clustering-based approaches like ITAD [50], THOC [51], and Deep-SVDD [52] are also considered. We further compare with autoregression-based models, including CL-MPPCA [53], LSTM [16], and VAR [54], as well as classic methods such as OC-SVM [55] and IsolationForest [56]. State-of-the-art deep learning models such as Anomaly Transformer [24] and InterFusion [44] are part of the evaluation, along with GmVRNN [30], a leading probabilistic model for anomaly detection, and TranAD [33], a recent anomaly detection method.

6.1.3. Implementation Details

In our experiments, we implement a three-layer transformer similar to the Anomaly Transformer (without the minimax strategy from [24]), with hidden states of 128 dimensions serving as the encoder. The MLP used to map features h 0 , h 1 , and h also has a dimension of 128. The MMA module comprises 15 memory units, each with a dimensionality of 128 across all datasets. Both the memory prior m a and latent variable z are set to a dimension of 128. We use hyperparameters β 1 and β 2 at 0.1 to balance the reconstruction and KL divergence terms across all datasets. The model is optimized using the Adam optimizer with a learning rate of 0.0002 and a batch size of 256. For the threshold in the POT method, we empirically set p to 0.003. Consistent with [51], a sliding window technique is applied to obtain sub-series, with a window size of 20 across all datasets. All experiments were carried out using PyTorch [57] on an NVIDIA RTX 3090 24 GB GPU.

6.1.4. Metric and One-for-All Setting

We measure the model’s performance using Precision, Recall, and F1 score, as commonly used in similar studies [1,12,14]. The F1 score, in particular, offers a balanced view by combining both precision and recall. An anomaly is classified as a true positive if any point within the ground-truth anomaly segment is identified correctly [30]. In our experiments, we split each MTS dataset into two halves: the first half for training and the second half for testing. Please note that in the one-for-one experiment, we train individual models specifically for the MTS of each device, effectively preventing any negative impact resulting from mode variations across different devices. Conversely, the one-for-all experiment poses a greater challenge. In this task, we employ a single model for training and testing across all MTS from various devices.

6.2. Main Result

6.2.1. Quantitative Comparison

We assessed our model’s performance on five public datasets and one real-world dataset, benchmarking it against a range of competitive methods. The results presented in Table 3 highlight that our model surpasses reconstruction-based methods like OmniAnomaly, density-estimation techniques such as MPPCACD, and even state-of-the-art deep learning models like Anomaly Transformer. Although TranAD [33], the best baseline anomaly detection method, is the one-for-one model, MVMT still exhibits 2.14 % higher performance (F1) on average. Table 3 underscores the effectiveness of MVMT in time-series anomaly detection. Our model achieves superior performance across all five datasets, significantly reducing both false alarms and missed detections compared to other methods. This highlights MVMT’s potential for real-world applications where accuracy and reliability are crucial. Although MVMT achieves the best F1 score across all datasets (Table 3), the performance margin is noticeably smaller on datasets such as DND. These datasets exhibit relatively homogeneous normal patterns and limited temporal variability, allowing existing baselines to already model the underlying dynamics effectively. In such cases, the advantage of MVMT’s meta memory mechanism is less pronounced, as the need for capturing diverse temporal concepts is reduced. This observation provides insight into scenarios where MVMT may offer smaller gains despite overall superior performance.

6.2.2. Ablation Study

To evaluate the influence of each component within our framework, we carried out ablation studies. The results of these studies are summarized in Figure 5. MVMT-without-VAE refers to using a basic MLP as a decoder, serving as the baseline in our model. Transformer + VAE represents a model where we introduce a single latent variable z without incorporating the m a component. Notably, Transformer + VAE exhibits an average performance 6.61% better than MVMT-without-VAE in Figure 5, demonstrating the effectiveness of employing probabilistic modeling. The complete version of our model, MVMT, integrates the proposed MMA model and the latent variable m a . The MMA model, along with the latent variable m a , achieves the best F1 score, surpassing Transformer + VAE by an average of 8.80% in Figure 5. This result underscores the significance of the MMA module, capable of capturing and storing various temporal dependencies within MTSs. Consequently, MVMT efficiently captures multiple patterns of MTSs with a single set of parameters. Furthermore, introducing the meta memory component in our model results in additional performance enhancements. Our experiments confirm that the diverse temporal information introduced by the MMA module significantly contributes to improving MTS anomaly detection.

6.2.3. Meta Memory Attention

In this paper, we aim to improve model performance by leveraging memory units to capture diverse temporal information within MTSs. Efficient retrieval of relevant memory units for various MTSs is crucial. To evaluate this, we analyzed the attention weights of the MMA, as shown in Figure 6. We selected five memory units at random and examined their attention weights. The upper part of the figure depicts MTS data with four distinct patterns, while the lower part displays a bar chart of the mean attention weights, outlined in the red box. The analysis in Figure 6 illustrates that the meta memory attention weights can be tailored to generate specific memories for different patterns in the MTS data. For instance, the yellow bar shows a relatively lower proportion for the first three patterns but becomes more prominent for the fourth pattern.
In this section, we provide visual insights into the attention weights of the MMA. Figure 6 showcases the statistical properties across four different patterns of MTSs. Additionally, in Figure 7, we delve into observing subtle changes in a specific MTS pattern. For observation purposes, we randomly selected attention weights from five memory units. The upper section of the figure presents segments of the MTS data with four distinct patterns, while the lower section shows the corresponding attention weights.
In contrast to Figure 6, we visualize the attention weights for individual points within each MTS. As depicted in the purple box in Figure 7, the memory attention weights of MMA exhibit four distinct cases across the four patterns of MTS. However, within each pattern, MMA finely tunes the weights corresponding to specific time timesteps. This phenomenon illustrates that MVMT not only provides different memory priors for significant pattern changes but also adjusts the memory priors’ information based on slight variations in each MTS. Regarding abnormal states, in cases where there is no corresponding information in memory, the memory attention weights generate a distinctive response, enhancing the model’s performance in detecting anomalies.

6.2.4. Information Capacity of MVMT

A key contribution of this paper is the introduction of a meta variational memory module aimed at learning diverse characteristics across multiple MTSs. The MMA is designed to store and recall historical MTS data, allowing MVMT to adapt quickly to new datasets. To examine how the number of memory units, denoted as N, affects model performance, we tested different N values, as shown in Figure 8. The results indicate that increasing N from 5 to 15 enhances the model’s capacity and performance. However, when N exceeds 15, the model may begin to overfit due to the accumulation of less relevant information.

6.2.5. Adapt to New MTSs

In practical settings involving MTS data, such as CDN systems, there is often a need to adapt detection models to new MTSs with limited data availability. Previous methods face challenges in achieving this adaptation, especially in the one-for-one setting, as they often necessitate training individual models for new MTSs. This approach may falter when confronted with limited observations. Our proposed MVMT trains a unified set of parameters for all MTSs and harnesses global information using the MMA module. This capability empowers MVMT to adapt to new MTSs despite limited observations. To demonstrate the efficacy of MVMT in adapting to new MTSs, we designed two new experiments to evaluate the model’s performance in scenarios with minimal data.
For the experiments in this section, we allocated 80% of the MTSs as historical data and used the remaining 20% as new data. In practical settings, such as deploying a one-for-one anomaly detection model for new equipment, we might only have a limited number of MTS samples for training. Given that each sample takes about 60 s to acquire and setup time for the equipment is approximately 20 min, we trained the one-for-one models using the initial i samples ( i 1 , 5 , 10 , 20 , 200 ) from each new MTS and evaluated the model on the remaining samples.
For the one-for-all models (GmVRNN and MVMT), parameters were initially trained on historical MTS data and later fine-tuned using the first i samples ( i 1 , 5 , 10 , 20 , 200 ) from each new MTS. Table 4 outlines the detection performance across methods. The results show that one-for-one models, when trained on limited data, fail to accurately capture normal patterns, resulting in reduced performance. In contrast, one-for-all models leverage previously stored information from historical MTSs, leading to a noticeable improvement over the one-for-one approaches.
Additionally, to further validate the design of our methods tailored for the one-for-all setting, we conducted an additional experiment by pretraining the one-for-one models on the training data, similar to the pretraining strategy applied in the one-for-all models. As depicted in Table 4, although the one-for-one models perform relatively better due to the involvement of historical MTSs, they fail to capture and retain the diverse characteristics within multiple MTSs. In contrast, our proposed MVMT, equipped with multiple components for preserving and recalling historical information, exhibits significantly superior performance compared to the one-for-one models in this setting.

6.2.6. Qualitative Analysis

Beyond quantitative analysis, we also conducted qualitative assessments. We compared the anomaly scores produced by MVMT, Anomaly Transformer, MVMT without m a , and a standard Transformer, as shown in Figure 9. Deterministic models like Transformers tend to generate more fluctuating anomaly scores, as they do not account for the inherent stochasticity in MTS data. In contrast, probabilistic models, such as MVMT without m a , provide smoother anomaly scores due to their more structured probabilistic design. However, dealing with certain situations using a simple Gaussian hidden variable remains challenging. Introducing a memory module into the anomaly detection model plays a vital role in improving performance. By incorporating diverse temporal information, the model can more accurately reconstruct normal MTS patterns, leading to enhanced detection accuracy and a lower false alarm rate. As shown in Figure 9a, the temporal diversity introduced by MMA allows MVMT to generate noticeably lower anomaly scores in normal data scenarios compared to other models. MMA’s retrieval of relevant memory is akin to referencing a vast historical dataset, helping MVMT produce smoother anomaly scores during normal conditions and more pronounced scores when anomalies occur. Figure 9b–d illustrate several cases where MVMT detects anomalies while other methods fail to do so. In (b), Transformer and MVMT without m a fail to generate sufficiently high anomaly scores, leading to detection failure. The more sophisticated model designs of Anomaly Transformer and MVMT yield higher anomaly scores, successfully detecting anomalies. In (c), when two anomalies are proximate, other models may misconstrue them as usual or ignore one, whereas our proposed model can compare memory information to differentiate between these modes using MMV. In (d), the model easily disregards inconspicuous anomalies as noise, while MVMT can sensitively detect them.

6.2.7. Time Efficiency

In line with previous studies [30], we assessed the time efficiency of various methods, including our MVMT, transformer-based models like TranAD and Anomaly Transformer, and RNN-based approaches such as GmVRNN and LSTM-VAE. The results, shown in Table 5, indicate that MVMT had notably shorter training and testing times compared to the other methods when using the same GPUs. All models were able to detect anomalies in less than a second per sample, which is beneficial given the 60 s interval between data points. This suggests that these methods are well-suited for real-time anomaly detection.

6.2.8. Number of Parameters

We analyzed and compared the number of model parameters across various methods, as shown in Figure 10. For this analysis, we selected LSTM-VAE, OmniAnomaly and Anomaly Transformer as baselines, representing typical RNN-based and Transformer-based methods, respectively. Notably, the parameter count of OmniAnomaly is considerably smaller than that of Anomaly Transformer due to the complexity of the transformer structure. However, with the incorporation of a mixture-distributed latent variable and a meta memory module, our proposed MVMT can efficiently account for the diverse characteristics within various MTSs. This capability allows MVMT to effectively model all MTSs using a single set of parameters, resulting in superior performance compared to previous methods, even surpassing RNN-based methods, with significantly fewer parameters.

7. Conclusions

To address anomaly detection in large-scale Multivariate Time Series (MTS), we propose the Meta Variational Memory Transformer (MVMT), which integrates self-attention with a meta memory module and a memory-guided probabilistic generative framework to effectively capture, organize, and utilize temporal concepts across diverse MTSs. The upward–downward autoencoding inference scheme further enhances posterior estimation through the integration of bottom-up likelihood and top-down priors. Across five public datasets and one real-world dataset, MVMT consistently attains the top F1 performance among all compared methods. Beyond accuracy, the ability to apply a single set of parameters to heterogeneous MTS collections highlights MVMT’s practicality and scalability for real-world online anomaly detection. These results collectively demonstrate the effectiveness and applicability of MVMT in large-scale MTS environments.

Author Contributions

Conceptualization, K.Q., Y.L., W.C., X.H., B.C. and H.L.; Methodology, Y.L.; Software, Y.L.; Validation, Y.L.; Formal analysis, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 62576266 and U21B2006; in part by the Fundamental Research Funds for the Central Universities QTZX24003 and QTZX23018; in part by the 111 Project under Grant B18039.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author(s).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Su, Y.; Zhao, Y.; Niu, C.; Liu, R.; Sun, W.; Pei, D. Robust anomaly detection for multivariate time series through stochastic recurrent neural network. In Proceedings of the ACM SIGKDD 2019, Anchorage, AK, USA, 4–8 August 2019; pp. 2828–2837. [Google Scholar]
  2. Su, Y.; Zhao, Y.; Sun, M.; Zhang, S.; Wen, X.; Zhang, Y.; Liu, X.; Liu, X.; Tang, J.; Wu, W.; et al. Detecting outlier machine instances through gaussian mixture variational autoencoder with one dimensional cnn. IEEE Trans. Comput. 2021, 71, 892–905. [Google Scholar] [CrossRef]
  3. Sun, M.; Su, Y.; Zhang, S.; Cao, Y.; Liu, Y.; Pei, D.; Wu, W.; Zhang, Y.; Liu, X.; Tang, J. Ctf: Anomaly detection in high-dimensional time series with coarse-to-fine model transfer. In Proceedings of the IEEE INFOCOM 2021, Vancouver, BC, Canada, 10–13 May 2021; pp. 1–10. [Google Scholar]
  4. Dai, L.; Lin, T.; Liu, C.; Jiang, B.; Liu, Y.; Xu, Z.; Zhang, Z. SDFVAE: Static and dynamic factorized VAE for anomaly detection of multivariate CDN kpis. In Proceedings of the WWW, Ljubljana, Slovenia, 19–23 April 2021; pp. 3076–3086. [Google Scholar]
  5. Liu, D.; Zhao, Y.; Xu, H.; Sun, Y.; Pei, D.; Luo, J.; Jing, X.; Feng, M. Opprentice: Towards practical and automatic anomaly detection through machine learning. In Proceedings of the ACM IMC, Tokyo, Japan, 28–30 October 2015; pp. 211–224. [Google Scholar]
  6. Chen, Y.; Mahajan, R.; Sridharan, B.; Zhang, Z. A provider-side view of web search response time. In Proceedings of the ACM SIGCOMM, Hong Kong, China, 12–16 August 2013. [Google Scholar]
  7. Chen, W.; Xu, H.; Li, Z.; Pei, D.; Chen, J.; Qiao, H.; Feng, Y.; Wang, Z. Unsupervised anomaly detection for intricate kpis via adversarial training of VAE. In Proceedings of the IEEE INFOCOM, Paris, France, 29 April–2 May 2019; pp. 1891–1899. [Google Scholar]
  8. Xu, H.; Chen, W.; Zhao, N.; Li, Z.; Bu, J.; Li, Z.; Liu, Y.; Zhao, Y.; Pei, D.; Feng, Y.; et al. Unsupervised anomaly detection via variational auto-encoder for seasonal kpis in web applications. In Proceedings of the WWW 18, Lyon, France, 23–27 April 2018; pp. 187–196. [Google Scholar]
  9. Chandola, V.; Banerjee, A.; Kumar, V. Anomaly detection: A survey. ACM Comput. Surv. 2009, 41, 15:1–15:58. [Google Scholar] [CrossRef]
  10. Shon, T.; Moon, J. A hybrid machine learning approach to network anomaly detection. Inf. Sci. 2007, 177, 3799–3821. [Google Scholar] [CrossRef]
  11. Yamada, M.; Kimura, A.; Naya, F.; Sawada, H. Change-point detection with feature selection in high-dimensional time-series data. In Proceedings of the IJCAI, Beijing, China, 3–9 August 2013; pp. 1827–1833. [Google Scholar]
  12. Zhang, C.; Song, D.; Chen, Y.; Feng, X.; Lumezanu, C.; Cheng, W.; Ni, J.; Zong, B.; Chen, H.; Chawla, N.V. A deep neural network for unsupervised anomaly detection and diagnosis in multivariate time series data. In Proceedings of the AAAI, Beijing, China, 3–9 August 2019; pp. 1409–1416. [Google Scholar]
  13. Li, D.; Chen, D.; Jin, B.; Shi, L.; Goh, J.; Ng, S. MAD-GAN: Multivariate anomaly detection for time series data with generative adversarial networks. In Proceedings of the ICANN, Munich, Germany, 17–19 September 2019; pp. 703–716. [Google Scholar]
  14. Audibert, J.; Michiardi, P.; Guyard, F.; Marti, S.; Zuluaga, M.A. USAD: Unsupervised anomaly detection on multivariate time series. In Proceedings of the ACM SIGKDD, Virtual, 6–10 July 2020; pp. 3395–3404. [Google Scholar]
  15. Malhotra, P.; Ramakrishnan, A.; Anand, G.; Vig, L.; Agarwal, P.; Shroff, G. Lstm-based encoder-decoder for multi-sensor anomaly detection. In Proceedings of the ICML, New York, NY, USA, 19–24 June 2016. [Google Scholar]
  16. Hundman, K.; Constantinou, V.; Laporte, C.; Colwell, I.; Söderstrxoxm, T. Detecting spacecraft anomalies using lstms and nonparametric dynamic thresholding. In Proceedings of the ACM SIGKDD, London, UK, 19–23 August 2018; pp. 387–395. [Google Scholar]
  17. Chalapathy, R.; Chawla, S. Deep learning for anomaly detection: A survey. arXiv 2019, arXiv:1901.03407. [Google Scholar] [CrossRef]
  18. Li, Y.; Chen, W.; Chen, B.; Wang, D.; Tian, L.; Zhou, M. Prototype-oriented unsupervised anomaly detection for multivariate time series. In Proceedings of the International Conference on Machine Learning, PMLR, Honolulu, HI, USA, 23–29 July 2023; pp. 19407–19424. [Google Scholar]
  19. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the NeurIPS 30, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  20. Wen, Q.; Zhou, T.; Zhang, C.; Chen, W.; Ma, Z.; Yan, J.; Sun, L. Transformers in time series: A survey. arXiv 2022, arXiv:2202.07125. [Google Scholar]
  21. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16 × 16 words: Transformers for image recognition at scale. In Proceedings of the ICLR, Vienna, Austria, 3–7 May 2021. [Google Scholar]
  22. Dong, L.; Xu, S.; Xu, B. Speech-transformer: A no-recurrence sequence-to-sequence model for speech recognition. In Proceedings of the ICASSP, Calgary, AB, Canada, 15–20 April 2018; pp. 5884–5888. [Google Scholar]
  23. Chen, M.; Peng, H.; Fu, J.; Ling, H. Autoformer: Searching transformers for visual recognition. In Proceedings of the CVPR, Nashville, TN, USA, 20–25 June 2021; pp. 12270–12280. [Google Scholar]
  24. Xu, J.; Wu, H.; Wang, J.; Long, M. Anomaly transformer: Time series anomaly detection with association discrepancy. In Proceedings of the ICLR, Virtual, 25–29 April 2022. [Google Scholar]
  25. Song, L.-K.; Tao, F.; Li, X.-Q.; Yang, L.-C.; Wei, Y.-P.; Beer, M. Physics-embedding multi-response regressor for time-variant system reliability assessment. Reliab. Eng. Syst. Saf. 2025, 263, 111262. [Google Scholar] [CrossRef]
  26. You, Z.; Cui, L.; Shen, Y.; Yang, K.; Lu, X.; Zheng, Y.; Le, X. A unified model for multi-class anomaly detection. arXiv 2022, arXiv:2206.03687. [Google Scholar] [CrossRef]
  27. Vilalta, R.; Drissi, Y. A perspective view and survey of meta-learning. Artif. Intell. Rev. 2002, 18, 77–95. [Google Scholar] [CrossRef]
  28. Sepp, H.; Jurgen, S. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  29. Tang, B.; Matteson, D.S. Probabilistic transformer for time series analysis. In Proceedings of the NeurIPS, Online, 6–14 December 2021; pp. 23592–23608. [Google Scholar]
  30. Dai, L.; Chen, W.; Liu, Y.; Argyriou, A.; Liu, C.; Lin, T.; Wang, P.; Xu, Z.; Chen, B. Switching gaussian mixture variational rnn for anomaly detection of diverse cdn websites. In Proceedings of the INFOCOM 2022, London, UK, 2–5 May 2022; pp. 300–309. [Google Scholar]
  31. An, J.; Cho, S. Variational autoencoder based anomaly detection using reconstruction probability. Spec. Lect. IE 2015, 2, 1–18. [Google Scholar]
  32. Siffer, A.; Fouque, P.-A.; Termier, A.; Largouet, C. Anomaly detection in streams with extreme value theory. In Proceedings of the ACM SIGKDD, Halifax, NS, Canada, 13–17 August 2017; pp. 1067–1075. [Google Scholar]
  33. Tuli, S.; Casale, G.; Jennings, N.R. Tranad: Deep transformer networks for anomaly detection in multivariate time series data. arXiv 2022, arXiv:2201.07284. [Google Scholar] [CrossRef]
  34. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  35. Rigotti, M.; Miksovic, C.; Giurgiu, I.; Gschwind, T.; Scotton, P. Attention-based interpretability with concept transformers. In Proceedings of the ICLR, Vienna, Austria, 3–5 May 2021. [Google Scholar]
  36. Lee, T.S.; Mumford, D. Hierarchical bayesian inference in the visual cortex. JOSA A 2003, 20, 1434–1448. [Google Scholar] [CrossRef] [PubMed]
  37. Blei, D.M.; Jordan, M.I. Variational inference for dirichlet process mixtures. Bayesian Anal. 2006, 1, 121–143. [Google Scholar] [CrossRef]
  38. Chung, J.; Kastner, K.; Dinh, L.; Goel, K.; Courville, A.C.; Bengio, Y. A recurrent latent variable model for sequential data. In Proceedings of the Advances in Neural Information Processing Systems 28, Montreal, QC, Canada, 7–12 December 2015. [Google Scholar]
  39. Lindley, D.V. Fiducial distributions and bayes’ theorem. J. R. Soc. Ser. B (Methodol.) 1958, 20, 102–107. [Google Scholar] [CrossRef]
  40. Higgins, I.; Matthey, L.; Pal, A.; Burgess, C.; Glorot, X.; Botvinick, M.; Mohamed, S.; Lerchner, A. beta-vae: Learning basic visual concepts with a constrained variational framework. In Proceedings of the ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  41. Abdulaal, A.; Liu, Z.; Lancewicki, T. Practical approach to asynchronous multivariate time series anomaly detection and localization. In Proceedings of the ACM SIGKDD, Online, 14–18 August 2021; pp. 2485–2494. [Google Scholar]
  42. Mathur, A.P.; Tippenhauer, N.O. Swat: A water treatment testbed for research and training on ics security. In Proceedings of the 2016 International Workshop on Cyber-Physical Systems for Smart Water Networks (CySWater), Vienna, Austria, 11 April 2016; IEEE: New York, NY, USA, 2016; pp. 31–36. [Google Scholar]
  43. Chen, W.; Tian, L.; Chen, B.; Dai, L.; Duan, Z.; Zhou, M. Deep variational graph convolutional recurrent network for multivariate time series anomaly detection. In Proceedings of the International Conference on Machine Learning, PMLR, Baltimore, MD, USA, 17–23 July 2022; pp. 3621–3633. [Google Scholar]
  44. Li, Z.; Zhao, Y.; Han, J.; Su, Y.; Jiao, R.; Wen, X.; Pei, D. Multivariate time series anomaly detection and interpretation using hierarchical inter-metric and temporal embedding. In Proceedings of the ACM SIGKDD, Online, 14–18 August 2021; pp. 3220–3230. [Google Scholar]
  45. Zhou, B.; Liu, S.; Hooi, B.; Cheng, X.; Ye, J. Beatgan: Anomalous rhythm detection using adversarially generated time series. In Proceedings of the IJCAI, Macao, China, 10–16 August 2019; pp. 4433–4439. [Google Scholar]
  46. Park, D.; Hoshi, Y.; Kemp, C.C. A multimodal anomaly detector for robot-assisted feeding using an lstm-based variational autoencoder. IEEE Robot. Autom. Lett. 2018, 3, 1544–1551. [Google Scholar] [CrossRef]
  47. Zong, B.; Song, Q.; Min, M.R.; Cheng, W.; Lumezanu, C.; Cho, D.; Chen, H. Deep autoencoding gaussian mixture model for unsupervised anomaly detection. In Proceedings of the ICLR, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  48. Yairi, T.; Takeishi, N.; Oda, T.; Nakajima, Y.; Nishimura, N.; Takata, N. A data-driven health monitoring method for satellite housekeeping data based on probabilistic clustering and dimensionality reduction. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 1384–1401. [Google Scholar] [CrossRef]
  49. Breunig, M.M.; Kriegel, H.-P.; Ng, R.T.; Sander, J. Lof: Identifying density-based local outliers. In Proceedings of the ACM SIGMOD, Dallas, TX, USA, 15–18 May 2000; pp. 93–104. [Google Scholar]
  50. Shin, Y.; Lee, S.; Tariq, S.; Lee, M.S.; Jung, O.; Chung, D.; Woo, S.S. Itad: Integrative tensor-based anomaly detection system for reducing false positives of satellite systems. In Proceedings of the 29th CIKM, Virtual Event, 19–23 October 2020; pp. 2733–2740. [Google Scholar]
  51. Shen, L.; Li, Z.; Kwok, J. Timeseries anomaly detection using temporal hierarchical one-class network. In Proceedings of the NeurIPS 33, Virtual, 6–12 December 2020; pp. 13016–13026. [Google Scholar]
  52. Ruff, L.; Vandermeulen, R.; Goernitz, N.; Deecke, L.; Siddiqui, S.A.; Binder, A.; Müller, E.; Kloft, M. Deep one-class classification. In Proceedings of the ICML, Stockholm, Sweden, 10–15 July 2018; pp. 4393–4402. [Google Scholar]
  53. Tariq, S.; Lee, S.; Shin, Y.; Lee, M.S.; Jung, O.; Chung, D.; Woo, S.S. Detecting anomalies in space using multivariate convolutional lstm with mixtures of probabilistic pca. In Proceedings of the ACM SIGKDD, Anchorage, AK, USA, 4–8 August 2019; pp. 2123–2133. [Google Scholar]
  54. Clements, M.P.; Mizon, G.E. Empirical analysis of macroeconomic time series: Var and structural models. Eur. Econ. Rev. 1991, 35, 887–917. [Google Scholar] [CrossRef]
  55. Tax, D.M.; Duin, R.P. Support vector data description. Mach. Learn. 2004, 54, 45–66. [Google Scholar] [CrossRef]
  56. Liu, F.T.; Ting, K.M.; Zhou, Z.-H. Isolation forest. In Proceedings of the IEEE ICDM, Pisa, Italy, 15–19 December 2008; pp. 413–422. [Google Scholar]
  57. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. In Proceedings of the NeurIPS 32, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
Figure 1. Visualization of pattern changes in five randomly selected channels from the SMD dataset at different time points. The figure highlights the pattern changes in five fixed channels across four different time points. Each column represents an MTS at a specific time, while each row corresponds to the same channel observed at different times. As shown in the figure, the same measured MTS can exhibit different patterns at different times, and these patterns may have significant distribution differences.
Figure 1. Visualization of pattern changes in five randomly selected channels from the SMD dataset at different time points. The figure highlights the pattern changes in five fixed channels across four different time points. Each column represents an MTS at a specific time, while each row corresponds to the same channel observed at different times. As shown in the figure, the same measured MTS can exhibit different patterns at different times, and these patterns may have significant distribution differences.
Sensors 25 07611 g001
Figure 3. Overall framework of anomaly detection for MTS based on MVMT. Solid lines represent the inference process, whereas dashed lines denote the generation process.
Figure 3. Overall framework of anomaly detection for MTS based on MVMT. Solid lines represent the inference process, whereas dashed lines denote the generation process.
Sensors 25 07611 g003
Figure 4. Graphical illustration of MVMT: (a) generation process of x ; (b) inference of the variational distribution of m a and z ; (c) overview of the proposed MVMT. Solid lines represent the inference process, whereas dashed lines denote the generation process.
Figure 4. Graphical illustration of MVMT: (a) generation process of x ; (b) inference of the variational distribution of m a and z ; (c) overview of the proposed MVMT. Solid lines represent the inference process, whereas dashed lines denote the generation process.
Sensors 25 07611 g004
Figure 5. Ablation study of MVMT on four datasets.
Figure 5. Ablation study of MVMT on four datasets.
Sensors 25 07611 g005
Figure 6. Visualization of the meta memory unit’s attention weights. Our memory module adeptly recalls relevant memories for different patterns. The upper section of the figure depicts segments of the MTS channels with four distinct patterns, while the lower section displays the attention weights for the corresponding positions, highlighted in red boxes.
Figure 6. Visualization of the meta memory unit’s attention weights. Our memory module adeptly recalls relevant memories for different patterns. The upper section of the figure depicts segments of the MTS channels with four distinct patterns, while the lower section displays the attention weights for the corresponding positions, highlighted in red boxes.
Sensors 25 07611 g006
Figure 7. Visualization of attention weights in the meta memory units. Our memory module retrieves relevant memories for different patterns. The upper part of the figure depicts segments of MTS channels featuring four distinct patterns, whereas the lower part shows the associated attention weights.
Figure 7. Visualization of attention weights in the meta memory units. Our memory module retrieves relevant memories for different patterns. The upper part of the figure depicts segments of MTS channels featuring four distinct patterns, whereas the lower part shows the associated attention weights.
Sensors 25 07611 g007
Figure 8. Visualization of how performance metrics change with varying numbers of memory units on the SMD and MSL datasets.
Figure 8. Visualization of how performance metrics change with varying numbers of memory units on the SMD and MSL datasets.
Sensors 25 07611 g008
Figure 9. Anomaly score analysis on the SMD dataset. The blue-highlighted regions denote the ground-truth anomalies, and the red lines indicate the thresholds set by POT. The top row displays a segment of the MTS data, followed by the anomaly scores for MVMT, Anomaly Transformer, VAE, and a simple Transformer in the rows below. Panels (ad) represent different time periods of the data.
Figure 9. Anomaly score analysis on the SMD dataset. The blue-highlighted regions denote the ground-truth anomalies, and the red lines indicate the thresholds set by POT. The top row displays a segment of the MTS data, followed by the anomaly scores for MVMT, Anomaly Transformer, VAE, and a simple Transformer in the rows below. Panels (ad) represent different time periods of the data.
Sensors 25 07611 g009
Figure 10. Comparison of model parameters.
Figure 10. Comparison of model parameters.
Sensors 25 07611 g010
Table 1. Symbol description for MVMT.
Table 1. Symbol description for MVMT.
EncoderDecoder
SymbolsDescriptionSymbolsDescription
W x h 0 The model parameter symbol of the Anomaly Transformer W μ M a MLP weights mapped from M a to the mean ( μ ) of m a
W h 0 h 1 The MLP weights of feature h 0 to h 1 . b μ M a MLP bias mapped from M a to the mean ( μ ) of m a
b h 0 h 1 The MLP bias of feature h 0 to h 1 . W σ M a MLP weights mapped from M a to the variance ( σ ) of m a
W h 0 h The MLP weights of feature h 0 to h . b σ M a MLP bias mapped from M a to the variance ( σ ) of m a
b h 0 h The MLP bias of feature h 0 to h . W μ m a MLP weights mapped from m a to the mean ( μ ) of z.
W ˜ m a μ The MLP weights that approximate the mean ( μ ) of m a with M a b μ m a MLP bias mapped from m a to the mean ( μ ) of z.
V ˜ m a μ The MLP weights that approximate the mean ( μ ) of m a with h 1 W σ m a MLP weights mapped from m a to the variance ( σ ) of z.
b ˜ m a μ The MLP bias that approximate the mean ( μ ) of m a b σ m a MLP bias mapped from m a to the variance ( σ ) of z.
W ˜ m a σ The MLP weights that approximate the variance ( σ ) of m a with M a W μ z MLP weights mapped from z to the mean ( μ ) of x.
V ˜ m a σ The MLP weights that approximate the variance ( σ ) of m a with h 1 b μ z MLP bias mapped from z to the mean ( μ ) of x.
b ˜ m a The MLP bias that approximate the variance ( σ ) of m a W σ z MLP weights mapped from z to the variance ( σ ) of x.
W ˜ z μ The MLP weights that approximate the mean ( μ ) of z with M a b σ z MLP bias mapped from z to the variance ( σ ) of x.
V ˜ z μ The MLP weights that approximate the mean ( μ ) of z with h 1       
b ˜ z μ The MLP bias that approximate the mean ( μ ) of z      
W ˜ z σ The MLP weights that approximate the variance ( σ ) of z with M a       
V ˜ z σ The MLP weights that approximate the variance ( σ ) of z with h 1       
b ˜ z The MLP bias that approximate the variance ( σ ) of z      
Table 2. Basic statistics of datasets.
Table 2. Basic statistics of datasets.
DatasetSMDMSLSMAPPSMDNDSWaT
Dimension385555253251
Window202020202020
Training708,40558,317135,181105,984344,843496,800
Test (labeled)708,42073,729427,61787,841344,843449,919
Anomaly ratio (%)4.110.713.1327.83.4411.98
Table 3. We report the quantitative performance of different methods across five public datasets and one real-world dataset. The metrics used include Precision (P), Recall (R), and F1 score (in percentages), with the F1 score representing the harmonic mean of precision and recall. Higher values across these metrics indicate superior performance. Boldface indicates the best F1 score among all compared methods.
Table 3. We report the quantitative performance of different methods across five public datasets and one real-world dataset. The metrics used include Precision (P), Recall (R), and F1 score (in percentages), with the F1 score representing the harmonic mean of precision and recall. Higher values across these metrics indicate superior performance. Boldface indicates the best F1 score among all compared methods.
DatasetSMDMSLPSMSMAPDNDSWaT
MetricPRF1PRF1PRF1PRF1PRF1PRF1
OC-SVM44.3476.7256.1959.7886.8770.8262.7580.8970.6753.8559.0756.3468.2370.3669.2845.3949.2247.23
IsolationForest42.3173.2953.6453.9486.5466.4576.0992.4583.4852.3959.0755.5369.6374.2171.8549.2944.9547.02
LOF56.3439.8646.6847.7285.2561.1857.8990.4970.6158.9356.3357.6070.2474.6672.3872.1565.4368.62
Deep-SVDD78.5479.6779.1091.9276.6383.5895.4186.4990.7389.9356.0269.0473.0979.0275.9480.4284.4582.39
DAGMM67.3049.8957.3089.6063.9374.6293.4970.0380.0886.4556.7368.5172.7277.6775.1189.9257.8470.40
MMPCACD71.2079.2875.0281.4261.3169.9576.2678.3577.2988.6175.8481.7371.5275.7573.5782.5268.2974.73
VAR78.3570.2674.0874.6881.4277.9090.7183.8287.1381.3853.8864.8372.9277.7275.2481.5960.2969.34
LSTM78.5585.2881.7885.4582.5083.9576.9389.6482.8089.4178.1383.3973.5279.6776.4786.1583.2784.69
CL-MPPCA82.3676.0779.0973.7188.5480.4456.0299.9371.8086.1363.1672.8873.1278.7175.8176.7881.5079.07
ITAD86.2273.7179.4869.4484.0976.0772.8064.0268.1382.4266.8973.8573.8775.4374.6463.1352.0857.08
LSTM-VAE75.7690.0882.3085.4979.9482.6273.6289.9280.9692.2067.7578.1075.1079.0577.0276.0089.5082.20
BeatGAN72.9084.0978.1089.7585.4287.5390.3093.8492.0492.3855.8569.6176.6480.1878.3764.0187.4673.92
OmniAnomaly83.6886.8285.2289.0286.3787.6788.3974.4680.8392.4981.9986.9279.4782.3780.9081.4284.3082.83
InterFusion87.0285.4386.2281.2892.7086.6283.6183.4583.5289.7788.5289.1478.2185.3481.6280.5985.5883.01
THOC79.7690.9584.9988.4590.9789.6988.1490.9989.5492.0689.3490.6880.9783.4582.1983.9486.3685.13
GmVRNN96.0791.2393.5690.8192.1091.4195.6298.3696.9796.5194.5495.5183.8087.6785.5890.1194.6992.34
Anomaly Transformer89.4095.4592.3392.0995.1593.5996.9198.9097.8994.1399.4096.6982.1386.3184.1691.5596.7394.07
TranAD92.6299.7496.0590.3899.9994.9495.3698.6596.9780.4399.9989.1582.5986.6084.5497.6069.9781.51
Ours93.7398.6896.1491.8998.3795.0196.6599.3998.0094.2699.2896.7084.5888.5486.5194.0897.5895.79
Table 4. The quantitative results for adapting to new MTSs across three public datasets are shown. These results were obtained by training the models with a limited amount of new MTS data and then testing their performance directly on this data. Boldface indicates the best F1 score among all compared methods.
Table 4. The quantitative results for adapting to new MTSs across three public datasets are shown. These results were obtained by training the models with a limited amount of new MTS data and then testing their performance directly on this data. Boldface indicates the best F1 score among all compared methods.
MethodsDatasetSMDMSLSMAP
Data Number151020200151020200151020200
MetricF1F1F1F1F1F1F1F1F1F1F1F1F1F1F1
Random
Initialization
DAGMM43.1144.1644.8744.0263.3850.6650.5251.3651.4565.1553.3853.2654.8960.1665.68
MMPCACD54.4055.4955.6356.2167.0551.3151.9452.5752.7565.0852.9052.7656.0161.0565.87
VAR54.4054.2954.6555.7566.3953.9953.0253.6053.8266.3155.2955.1856.3662.9066.00
LSTM56.1756.6856.8756.8967.7660.5860.6261.2961.3769.5956.7356.5757.5861.9867.74
CL-MPPCA57.0057.0557.1757.8768.1459.6759.3460.3060.6970.0858.7658.4358.2865.5568.36
ITAD56.3656.0856.5156.9066.9362.9862.7163.1763.3569.5062.7462.7163.8566.9568.85
LSTM-VAE58.3858.7458.0859.4869.1561.8961.1061.9062.1671.0561.6161.2362.1967.4169.16
BeatGAN61.3061.1861.0462.0268.7363.0963.5764.2365.8874.3061.4961.1861.7266.8971.10
OmniAnomaly64.2965.1563.5362.0568.3164.3664.4365.1468.7773.8063.1365.5565.6168.7270.27
InterFusion63.6862.6662.3762.2869.8363.6063.5664.3265.8362.1663.5063.3463.9568.7972.44
THOC64.4163.4763.1263.9471.0164.3364.1364.2266.0074.6562.2462.4663.9270.6675.22
Anomaly Transformer64.2465.0971.8576.9881.1165.1664.9968.5270.4677.0263.8266.7267.3471.4578.88
TranAD66.2867.5872.2677.0181.5467.6666.2669.8372.7478.9464.2767.2468.7372.6580.78
Pretraining with
History MTSs
DAGMM72.0872.5572.8273.0473.3968.4767.7168.2868.6869.1770.8171.0471.6671.9472.07
MMPCACD73.1572.9073.9073.3173.1169.9969.6870.3470.5770.0472.9573.0173.0573.2573.44
VAR74.5475.0174.1174.7874.1070.8871.1471.0971.6771.9671.8772.1672.6772.3872.56
LSTM73.7773.6173.7873.9273.7071.7971.3272.0172.1771.9474.0774.5774.4473.7674.81
CL-MPPCA75.2375.1874.9275.3175.8773.5772.9973.5173.3673.1276.5075.9676.1576.7476.64
ITAD77.0376.8577.7677.6776.9977.0877.0077.4977.1877.6478.8078.8979.3178.8279.34
LSTM-VAE80.6580.7380.1381.6881.5176.1976.3176.4776.2077.0277.9277.8377.9678.1978.50
BeatGAN82.9083.0283.0882.9382.5779.6779.6680.5680.2180.8481.6681.6182.1282.5082.36
OmniAnomaly85.4085.4285.7886.1985.8381.1280.5882.1381.8681.6484.2884.7184.6184.9785.49
InterFusion84.6084.7384.7484.8585.0982.0981.8782.2882.4683.0086.7086.8087.2687.2787.12
THOC86.0186.0586.2486.6286.5481.5681.5581.3882.7382.8688.9589.2888.9288.5488.95
Anomaly Transformer91.5891.9291.4691.7891.6980.6582.8782.1583.5584.0193.3293.8993.5194.0194.11
TranAD91.9691.6991.5791.8791.9480.8581.4482.4284.0884.1293.0194.0393.6294.3594.37
One-For-AllGmVRNN91.0389.5190.0190.7890.3481.6881.1582.1281.5981.2293.4793.4193.0192.2194.12
Ours93.3192.8593.9093.7893.0889.2289.9089.9789.0390.3095.2795.2295.2695.4795.61
Table 5. Training and testing time of different methods.
Table 5. Training and testing time of different methods.
MethodsAverage Training Times per Epoch (s)Testing Times per Sample (s × 105)
SMDMSLSMAPPSMDNDSMDMSLSMAPPSMDND
LSTM-VAE26.713.886.2240.2937.598.3714.348.908.519.76
OmniAnomaly276.9721.3127.05243.24230.8135.0867.5240.3940.7844.82
GmVRNN25.202.675.0437.9635.267.0213.548.178.268.89
Anomaly Transformer28.802.645.4039.676.156.5111.347.217.327.95
TranAD43.765.5710.5563.5860.309.1717.6310.4910.6611.71
Ours25.282.034.9736.7334.845.3010.076.126.166.71
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Qin, K.; Li, Y.; Chen, W.; Hu, X.; Chen, B.; Liu, H. Meta Variational Memory Transformer for Anomaly Detection of Multivariate Time Series. Sensors 2025, 25, 7611. https://doi.org/10.3390/s25247611

AMA Style

Qin K, Li Y, Chen W, Hu X, Chen B, Liu H. Meta Variational Memory Transformer for Anomaly Detection of Multivariate Time Series. Sensors. 2025; 25(24):7611. https://doi.org/10.3390/s25247611

Chicago/Turabian Style

Qin, Kun, Yuxin Li, Wenchao Chen, Xinyue Hu, Bo Chen, and Hongwei Liu. 2025. "Meta Variational Memory Transformer for Anomaly Detection of Multivariate Time Series" Sensors 25, no. 24: 7611. https://doi.org/10.3390/s25247611

APA Style

Qin, K., Li, Y., Chen, W., Hu, X., Chen, B., & Liu, H. (2025). Meta Variational Memory Transformer for Anomaly Detection of Multivariate Time Series. Sensors, 25(24), 7611. https://doi.org/10.3390/s25247611

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop