Next Article in Journal
Mulvernet: Nucleus Segmentation and Classification of Pathology Images Using the HoVer-Net and Multiple Filter Units
Next Article in Special Issue
Task Offloading and Resource Allocation for Tasks with Varied Requirements in Mobile Edge Computing Networks
Previous Article in Journal
Next-Gen Mulsemedia: Virtual Reality Haptic Simulator’s Impact on Medical Practitioner for Higher Education Institutions
Previous Article in Special Issue
Robust Hierarchical Federated Learning with Anomaly Detection in Cloud-Edge-End Cooperation Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Decompose Auto-Transformer Time Series Anomaly Detection for Network Management †

1
School of Software, Nanchang Hangkong University, Nanchang 330063, China
2
Key Laboratory of Computer Network and Information Integration, Ministry of Education, Southeast University, Nanjing 211102, China
3
Purple Mountain Laboratories, Nanjing 211111, China
4
School of Information and Communication Engineering, Beijing University of Technology, Beijing 100124, China
5
Institute of Microelectronics, Chinese Academy of Sciences, Beijing 100029, China
6
Shandong Future Network Research Institute, Jinan 250002, China
7
Department of Computer Science and Engineering, Washington University in St. Louis, St. Louis, MO 63130, USA
*
Author to whom correspondence should be addressed.
Opening Foundation of Key Laboratory of Computer Network and Information Integration, Ministry of Education, Southeast University (K93-9-2021-05).
Electronics 2023, 12(2), 354; https://doi.org/10.3390/electronics12020354
Submission received: 26 September 2022 / Revised: 7 November 2022 / Accepted: 8 November 2022 / Published: 10 January 2023
(This article belongs to the Special Issue Resource Allocation in Cloud–Edge–End Cooperation Networks)

Abstract

:
Time series anomaly detection through unsupervised methods has been an active research area in recent years due to its enormous potential for networks management. The representation and reconstruction of time series have made extraordinary progress in existing works. However, time series is known to be complex in terms of their temporal dependency and stochasticity, which makes anomaly detection difficult. To this end, we propose a novel approach based on a decomposition auto-transformer networks(DATN) for time series anomaly detection. The time series is decomposed into seasonal and trend components, and renovated as a basic inner block deep model. With this design, transformers can decompose complex time series in a progressive manner. We also design an auto-transfomer block that determines dependencies and representation aggregation at the sub-series level based on series seasonal and trend components. Moreover, the complex transformer decoder is replaced by a simple linear decoder, which makes the model more efficient. Extensive experiments on various public benchmarks demonstrate that our method has achieved state-of-the-art performance.

1. Introduction

Time series anomaly detection, which is a widely used technology [1,2], plays an increasingly important role in academia and industry, such as network management. Anomaly detection refers to identifying rare items, events, or observations in data that are significantly different from expectation. Network management focuses on the optimal operation of resources to ensure the rational use of performance indicators and resources. Anomaly detection is an important part of network management. the effectiveness and function of anomaly detection is directly related to the effectiveness of network management. In order to improve the availability and reliability of the network, when the network is abnormal, the network manager can find the anomaly as soon as possible and take measures in advance to avoid the impact on the service. Therefore, efficient anomaly detection technology can improve the network intelligent operation.
Time series anomaly detection has become increasingly difficult in practice due to the explosion of raw data that can be captured. In recent decades, numerous efforts have been made to detect time series anomalies. Statistical methods, supervised learning methods, and unsupervised learning methods can be roughly divided into three categories. In the first class of methods, systems produce normal data based on a family of models, and the model parameters are learned using the normal data.
A time series anomaly detection problem is typically treated as a classification problem in supervised learning methods. Through shallow or deep models, they extract various features and perform two-class or multiclass classification. Although the above two kinds of methods provide impressive and promising results, unsupervised time-series anomaly detection methods have drawn much attention in recent years due to two main reasons. Despite the availability of large numbers of time series, obtaining the corresponding label information is difficult. It is a time-consuming and expensive process to manually annotate labels by domain experts. The anomalies are sparse even with labels, and they cannot cover every type of anomaly. Both normal and anomalous samples are highly imbalanced, which makes supervised methods inefficient. The core idea of unsupervised detection methods is to learn the normal pattern of data using a large amount of available normal data.
There may be a few serious abnormal samples in the training data, which adversely affect the learning of normal patterns, although many of the training data are normal samples. Therefore, unsupervised learning methods are usually employed to handle this task due to anomaly diversities, the lack of positive samples, and the high cost of annotations.
In practice, time-series anomaly detection has become increasingly challenging because of the explosion of raw data that can be captured. The solution to this problem has been addressed extensively over the past several years. In the early stage, anomalies were detected using classical machine learning techniques such as Gaussian models. Handcrafted features at the low level, however, cannot represent data in high dimensions. As deep neural networks have developed rapidly over the past few years, they have made significant progress. It is now possible to learn highly sophisticated representations using advanced machine learning techniques.
When modeling time series, it is essential to consider their temporal characteristics. Time-series data modeling provides natural advantages for improving recurrent neural network (RNN)-based methods [3]. However, the inherent limitations of recurrent models prevent them from running simultaneously. As an alternative, models based on convolutional neural networks were proposed. For sequential data modeling, temporal convolutional networks (TCNs) are a typical representative. This is a characteristic of parallel computing that gives it an advantage. To capture long-term dependence, attention mechanisms are used in another type of temporal context modeling. With the use of self-attention, the transformer successfully illustrates sequential representations.
However, the time-series anomaly detection task is extremely challenging under long-term settings. First, using long-term time series directly requires the extraction of temporal dependencies, which is difficult due to their entangled temporal patterns. We examine the intricate temporal patterns using decomposition, a standard method for time series analysis. Second, since the length of sequences for canonical transformers is quadratic, long-term forecasting cannot be done with them because of their computational complexity.
The model cannot identify reliable relationships due to long-term temporal patterns [4]. Due to the entangled temporal patterns, it is difficult to extract the temporal dependencies directly from the long-term time series. Using decomposition, a standard method for time series analysis, we try to sort out the intricate temporal patterns.
Based on the above information, we propose a decomposition auto-transformer network (DATN) for time-series anomaly detection. DATN still follows residual and encoder-decoder structures. In the encoder layer, we decompose the time series into seasonal and trend components. DATN embeds decomposition blocks as the inner operators. DATN is able to progressively separate long-term trend information by embedding our proposed decomposition blocks as the inner operators. DATN introduces an autoattention module to isolate and represent the periodicity of time series. The frequency-domain dominating periodic patterns are extracted using Fast Fourier Transforms (FFTs) and inverse FFTs.
The contributions of our paper are as follows:
(1)
To empower the deep anomaly detection model with imminent decomposition capability, we propose a decomposition auto-attention network as a decomposition architecture;
(2)
To extract the dominating periodic patterns in the frequency domain, we propose an auto-attention module to discover the period-based dependencies of time series;
(3)
Extensive experimental results on multiple public datasets show that the proposed method achieves state-of-the-art performance.
The remainder of the paper is structured as follows. Section 2 reviews related works. Section 3 examines the problem formulation and Transformer. Section 4 illustrates the proposed decomposition auto-transformer network (DATN). Section 5 describes the experimental process and results analysis of the method. Finally, we summarize the paper.

2. Related Work

Anomaly detection in time series has gained an ever-increasing interest in academia and industry [5,6,7]. There are three categories of existing methods: traditional statistical models, supervised learning methods, and unsupervised learning methods. Statistical models [8,9] such as MA, ARIMA, and Holter Winter are usually used to develop algorithms, and usually make strong assumptions regarding the time series being studied. Different types of time series require different algorithms and fine tuning parameters. As a result, they are not inappropriate for detecting anomalies in complex time series, which are typically found in real applications.
Anomaly detection is considered as a classification problem by supervised learning methods [8,10,11]. A classification model is trained using random forest and detection results are outputted by the aggregation functions. Although the supervised learning methods are powerful and show good results, they require labeled training data, which is difficult in time series anomaly detection. Despite the collection of abnormal data, the data may not include all types of anomalies. Due to this limitation, supervised detection methods have limited applications.
Because of the absence of label information, unsupervised learning methods have received a great deal of attention and popularity. We summarize a few recent works employing RNN, graph, GAN, VAE, and transformer-based methods. RNN-based methods have the advantage of processing time series. In Ref. [3], a variational autoencoder (VAE) was combined with the LSTM to enhance temporal feature representation. In Ref. [12], a stochastic RNN method was proposed to identify anomalies by modeling the data distribution through stochastic latent variables. Graph-based methods are able to model the relationships between signals [13]. In Ref. [14], two parallel GAT modules were employed to capture featurewise and temporal relationships respectively. In Ref. [15], a novel attention-based GNN approach was proposed to learn a graph of dependence relationships between signals. Then, they can identify and explain deviations from these relationships. GAN-based methods [16] train a generator and a discriminator to detect anomalies by the output of the discriminator or the reconstructed error. However, GAN-based methods are typically difficult to train and suffer from the mode collapse problem. In Ref. [17], one encoder and two decoders were designed to reconstruct the input series. Xu et al. [18] proposed an unsupervised anomaly detection algorithm, called Donut, which adopts a CNN as the basic unit for the encoder and decoders, and the training and inference times are sharply reduced. Li et al. [19] proposed a robust and unsupervised anomaly detection algorithm for seasonal key performance indices (KPIs), called Bagel, that adopts CVAE as the basis of the model. Transformer-based methods show great power in sequential data owing to the self-attention mechanism [20]. In Ref. [21], a time series representation learning framework was proposed based on the transformer encoder architecture. In Ref. [22], a self-attention mechanism was renovated to anomaly attention. This mechanism was well designed to represent the prior association and series association of each time stamp. We choose Transformer as the basic framework considering the advantages offered by its multiheaded self-attention.

3. Problem Formulation

A time series is generally considered as a collection of observations indexed in time order [6]. It can be defined as X = { x 0 , . . . , x T 1 } , where T is the length of the time series, x t is an observation at timestamp t, and 0 t < T . Define the number of signals as m, x t R m and X R T × m .
Given a training series X R T × m and a test series X ^ R T ^ × m with length T ^ , the task of unsupervised anomaly detection is to predict an output Y = { y 0 , . . . , y T ^ 1 } where y t { 0 , 1 } , 0 t < T ^ to denote whether X ^ t is an anomaly.
Time series can exhibit a variety of patterns. It is helpful to decompose time series into several components, each representing an underlying pattern. Generally, we consider a time series as the superposition of two different components: a seasonal component, a trend component. The seasonal component at timestamp t, defined as s t , represents the periodic features of the series. The trend component at timestamp t, defined as p t , reflects the long-term progression. The remaining component rt contains anything else in the series. We employ an additive decomposition, then the observation x t can be written as:
x t = s t + p t
The transformer [23] is an encoder-decoder structure composed of multiple identical blocks. The multi-headed self-attention module (MHSA) and a feed forward network (FFN) are two essential modules in each block. For a time series X R w × m within look-back window w, its embedding is defined as E R w × d , where d are the dimensions of the embedding features. The attention function firstly projects E into query Q R w × d ( q ) , key K R w × d ( k ) , and value V R w × d ( v ) with different learned linear projections to d ( q ) , d ( k ) and d ( v ) dimensions. Then, we compute their scaled dotproduct attention as follows:
A t t e n t i o n ( Q , K , V ) = S o f t m a x ( Q K T d ( k ) ) V
Multi-headed self-attention employs a few different sets of learned projections instead of a single attention. The position-wise feed-forward module is a fully connected network defined as:
F F N ( A ) = L N ( ( Re L U ( A W 1 + b 1 ) W 2 + b 2 ) + A )
where A is the output of the previous layer, W 1 , W 2 , b 1 , and b 2 are the trainable parameters, and LN is the Layer normalization operator.

4. Decompose Auto-Transformer

The proposed Decompose Auto-Transformer network (DATN) framework is shown in Figure 1. It is constructed by series decompose, auto-attention, multi-head attention, feed forward, and linear decoder. Series decompose block is designed to decompose the time series into seasonal component( X s ) and trend component( X t ). ’+’represent feature additive fusion. With X s and X t as input, auto-attention module responsible for embedding multi-scale correlation. Next, the features are fed into the following MHSA. Then, seasonal component embedding and trend component embedding are carried out feature fusion. After linear decoder, reconstruction modules are employed.
Series decomposition block. Time series can exhibit a variety of patterns. It is helpful to decompose time series into several components, each representing an underlying pattern. Generally, we consider of a time series as the superposition of two different components: a seasonal component and a trend component. The seasonal component ( X s ) represents the periodic features of the series. The trend component X t reflects the long-term progression. We employ an additive decomposition, then the observation X can be written as:
X = X t + X s
Moving averages are adapted to smooth out the seasonal component and the trend component. For length-w input time series X R w × m , m denotes the time series dimension, and the decomposition formula is as follows:
X t = M A ( X ) X s = X X t
where X s , X t denote the seasonal and trend part, respectively. M A ( ) denote the moving average function. We use X s , X t = S e r i e s D e c o m p ( X ) to summarize Equations (4).
Encoder: As shown in Figure 1, suppose we have N encoder layers. The overall equations for l-th encoder layer are summarized as X l = E n c o d e r ( X l 1 ) . Details are shown as follows:
X s l , X t l = S e r i e s D e c o m p ( X l 1 ) S s l = A u t o _ a t t e n t i o n ( X s l ) S t l = A u t o _ a t t e n t i o n ( X t l ) O s l = FFN ( Attention ( S s l , S s l , S s l ) ) O t l = FFN ( Attention ( S t l , S t l , S t l ) ) X l = O s l + O t l
where X l , l { 1 , , N } denotes the output of l-th encoder layer. X s l and X t l denote the seasonal and trend component. S s l and S t l denote the output of l-th A u t o _ a t t e n t i o n ( ) . O s l and O t l denote the output of A t t e n t i o n ( ) . Attention is a Multi-Headed Self-Attention module. We will give a detailed description of A u t o _ A t t e n t i o n in the following paragraph.
Decode. Our model reconstructs the time series using the embeddings X N (N is the number of encoder layers), i.e.,
X ^ = D e c o d e r ( X N )
where X ^ is the restructure time series, Decoder ( ) is a simple linear model.
Anomaly Criterion. With the reconstructed time series, we define a score function s t to flag anomalous behaviors by summing up the prediction errors in time t, i.e.,
s t = i = 1 m x ^ t x t 2
where x ^ t is the reconstructed value in time t, x t is input value in time t. When the score exceeds a certain threshold, the corresponding data point is identified as an anomaly.
Auto-Attention. To extract the period-based feature, we propose an auto-attention module as shown in Figure 2. The input of the l-th encoder layer is processed by FFTs (Fast Fourier Transforms). The frequency domain is converted from the time domain. Multi-head attention involves scaled dot-product attention. To discover the dominating periodic patterns, we select top K amplitudes and convert them back to time domain D s l . Then, multi-head attention is employed to enhance the feature extraction of D s l . For l-layer series X l is decomposed into seasonal component X s l and trend component X t l . FFT F and the inverse FFT F 1 are formalized as:
Γ k = F ( x n ) = n = 0 w 1 x n e j 2 π w k n k = 0 , . . . , w 1 x n = F 1 ( Γ k ) = 1 w n = 0 w 1 Γ k e j 2 π w k n k = 0 , . . . , w 1
For the seasonal component, auto_attention can be expressed as:
D s l = F 1 ( T o p k ( C o n c a t ( F ( X s l ) , F ( X s l ) , F ( X s l ) ) ) ) M s l = A t t e n t i o n ( D s l , D s l , D s l ) A u t o _ a t t e n t i o n ( X s l ) = L N ( M s l )
where T o p k ( ) obtains arguments of the Topk autocorrelations and let k = c × log w , where c is a hyperparameter, A t t e n t i o n ( ) is a Multi-Headed self-Attention module, and L N ( ) is a linear module. The A u t o _ a t t e n t i o n of the trend component is the same as the above formula.

5. Experiments

5.1. Datasets and Model Details

The purpose of this study is to verify the effectiveness of the proposed method, we choose five public datasets:
(1)
SMD (Server Machine Dataset) [12] collected from a large Internet company, where each observation is equally spaced by 1 min;
(2)
SMAP (Soil Moisture Active Passive satellite) [24] has 562,798 time points, of which the training set size is 135,182 and the testing set size is 427,616;
(3)
MSL (Mars Science Laboratory rover) [24] has 132,046 time points, of which the training set size is 58,317 and the testing set size is 73,729;
(4)
SWaT (Secure Water Treatment) [25] collected from a real-world water treatment plant with 51 dimensions;
(5)
PSM (Pooled Server Metrics) [26] collected from multiple application server nodes at eBay with 26 dimensions.
Dataset statistics are shown in Table 1.
Experimental environment: The operating system is Ubuntu 18.04 LTS, the GPU is a NVIDIA GEFORCE RTX 2080, and the implemented is program is PyTorch 3.6. All experiments are repeated three times.
Data normalization and cleaning are first performed on all datasets. We follow the threshold selection method proposed in Ref. [22]. Anomaly scores on the validation subset are clustered to find the threshold. We also employ the point-adjust strategy, which assumes that anomalies are correctly detected in a segment if any anomaly in this segment is detected correctly [12]. Hyperparameter c (in Equation (10)) is set as 4. We use a fixed look-back series w = 192 for training and testing. We stack 4 encoder layers in the model. The Adam optimizer is employed with an initial learning rate of 10 4 .

5.2. Baselines

We compare our model with seven state-of-the-art baselines:
(1)
LSTM-VAE [3]: The model utilizes both VAE and LSTM for anomaly detection;
(2)
OmniAnomaly [12]: The model is a stochastic recurrent neural network model that glues Gated Recurrent Unit (GRU) and VAE;
(3)
MTAD-GAT [14]: The model considers each univariate time-series as an individual feature and includes two graph attention layers in parallel;
(4)
USAD [17]: The model use adversarial training and its architecture allows it to isolate anomalies;
(5)
TransFram [21]: The model uses the transformer as the base architecture;
(6)
AnomalyTran [22]: The model combines the transformer and the Gaussian prior association makes it easier to distinguish rare anomalies;
(7)
TranAD [27]: The model uses an adversarial training program to amplify the reconstruction error, because the simple transformer based networks often miss small abnormal deviations.
These comparison baseline methods are the most representative at present.

5.3. Performance Comparison

Precision (P), Recall (R), and F1 score are employed to evaluate the detection performance of all methods. To comprehensively evaluate the five datasets mentioned above, we reproduce part of the results for the baselines. The quantitative evaluation is reported in Table 2, where the best results of the F1 score are in red and the second-best results are blue. From Table 2, we can make the following observations. We achieve the best F1 scores on the SMAP, MSL, SWaT, and PSM datasets and achieve the second-best F1 score on the SMD dataset. Our method achieves high accuracy, which indicates that the false-positive rate is low. Our method performs well in terms of accuracy and recall, with F1 scoring higher than most other methods. Both methods bring significant improvements and new directions. In contrast, we decompose the time series into seasonal components and trend components. Long-term robustness can be retained by auto-attention. We argue that it can improve temporal pattern representation. This hypothesis is confirmed by the results.
To evaluate the efficiency of the proposed method, we provide a comparison of the training time (per epoch) shown in Figure 3. Compared with RNN- and graph-based methods, we need less time per epoch training, as depicted in Figure 3. RNN-based models are weak in parallel computing. The graph-based model is time-consuming due to its complex matrix operations. The graph-based method consumes more time when there are more signals.

5.4. Ablation Study

We analyze the effectiveness of the main components proposed in the work, including series decompose and auto-attention, as shown in Table 3. For auto-attention, × indicates that the original transformer framework is used. For series decompose, × means that the proposed series decompose is replaced by a simple linear layer. From Table 3, we can make the following observations. Using the proposed series decompose, the average F1 rises from 84.20% to 91.41%. Replacing multi-head attention by our auto-attention, the average F1 rises from 91.41% to 92.45%. That is to say, series decompose and auto-attention can obviously improve the performance by achieving 8.25% (92.45% → 84.20%) and 3.88% (88.08% → 84.20%) absolute average F1 promotions, respectively. In addition, our method yields a 13.52% (84.20% → 92.45%) absolute average F1 promotion compared with the original Transformer framework.

5.5. Case Study

We leverage a piece of the SMD dataset for the case study, shown in Figure 4. There are four sub-figures. Figure 4a describes a selected signal with the reconstruction results. The signal is selected due to its greater contribution to anomaly scores. The signal values far from the reconstruction results will generate large anomaly scores (see Figure 4b). Then, with the anomaly scores and threshold, we can easily find the anomalies depicted in the Figure 4c. We produce results essentially in agreement with the anomaly labels shown in the Figure 4d. The differences can be partially avoided by the point-adjust strategy. Generally, the proposed method is effective.

6. Conclusions

In this study, we presented a decomposition auto-attention network for time-series anomaly detection. Different from previous methods, we combined the time-series decomposition architecture to learn the intricate temporal patterns in a divide-and-conquer way. Auto-attention based on FFT was carefully designed to isolate and represent the periodicity of time series. We proved the effectiveness of our detection algorithm on five real datasets. These experiments demonstrated the efficacy and robustness of DATN and the key module auto-attention, promising general use for time series reconstruction without the need for model customization and greedy hyperparameter tuning. The only limitation is the requirement of a certain amount of training data, as in typical neural networks.
In future work, we plan to expand our DATN to other intelligent tasks for IoT systems including forecasting and smart control. We also plan to make use of the feature extraction ability of auto-attention and explore an alternative usage of auto-attention from the perspective of representation learning for time series.

Author Contributions

Conceptualization, Y.C.; Data curation, C.F.; Funding acquisition, Y.T.; Methodology, B.W.; Resources, Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Opening Foundation of Key Laboratory of Computer Network and Information Integration (Southeast University), Ministry of Education grant number K93-9-2021-05.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Simon, D.; Duque, A.; Lohfink, A.P.; Garth, C.; Schotten, H.D. Security in process: Detecting attacks in industrial process data. arXiv 2019, arXiv:1909. [Google Scholar]
  2. Wang, S.Y. Ensemble2: Anomaly Detection via EVT-Ensemble Framework for Seasonal KPIs in Communication Network. arXiv 2022, arXiv:2205.14305. [Google Scholar]
  3. Hoshi, Y.; Park, D.; Kemp, C.C. A multimodal anomaly detector for robot-assisted feeding using an lstm-based variational autoencoder. IEEE Robot. Autom. Lett. 2018, 3, 1544–1551. [Google Scholar]
  4. Wang, J.; Wu, H.; Xu, J.; Long, M. Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Adv. Neural Inf. Process. Syst. 2021, 34, 22419–22430. [Google Scholar]
  5. Shaukat, K.; Alam, T.M.; Luo, S.; Shabbir, S.; Hameed, I.A.; Li, J.; Abbas, S.K.; Javed, U. A review of time-series anomaly detection techniques: A step to future perspectives. In Proceedings of the Future of Information and Communication Conference, Virtual, 29–30 April 2021; pp. 865–877. [Google Scholar]
  6. Yi, J.; Park, C.; Yoon, S.; Choi, K. Deep learning for anomaly detection in time-series data: Review, analysis, and guidelines. IEEE Access 2021, 9, 120043–120065. [Google Scholar]
  7. Sgueglia, A.; Di Sorbo, A.; Visaggio, C.A.; Canfora, G. A systematic literature review of iot time series anomaly detection solutions. Future Gener. Comput. Syst. 2022, 134, 170–186. [Google Scholar] [CrossRef]
  8. Wolf, T.; Shanbhag, S. Accurate anomaly detection through parallelism. IEEE Netw. 2009, 23, 22–28. [Google Scholar]
  9. Lemes, M.; Proena, E.; Pena, H.M.; de Assis, M.V.O. Anomaly detection using forecasting methods arima and hwds. In Proceedings of the International Conference of the Chilean Computer Science Society (SCCC), Temuco, Chile, 11–15 November 2013; pp. 63–66. [Google Scholar]
  10. Flint, I.; Laptev, N.; Amizadeh, S. Generic and scalable framework for automated time-series anomaly detection. In Proceedings of the 21th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Sydney, NSW, Australia, 10–13 August 2015; pp. 1939–1947. [Google Scholar]
  11. Wang, J.; Jing, Y.; Qi, Q.; Feng, T.; Liao, J. Alsr: An adaptive label screening and relearning approach for interval-oriented anomaly detection. Expert Syst. Appl. 2019, 136, 94–104. [Google Scholar] [CrossRef]
  12. Su, Y.; Zhao, Y.; Niu, C.; Liu, R.; Sun, W.; Pei, D. Robust anomaly detection for multivariate time series through stochastic recurrent neural network. Proc. ACM SIGKDD Int. Conf. Knowl. Discovery Data Mining 2019, 3, 2828–2837. [Google Scholar]
  13. Wu, B.; Liu, Y.; Lang, B.; Huang, L. Dgcnn: Disordered graph convolutional neural network based on the gaussian mixture model. Neurocomputing 2018, 321, 346–356. [Google Scholar] [CrossRef] [Green Version]
  14. Zhao, H.; Wang, Y.; Duan, J.; Huang, C.; Cao, D.; Tong, Y.; Xu, B.; Bai, J.; Tong, J.; Zhang, Q. Multivariate time-series anomaly detection via graph attention network. In Proceedings of the 2020 IEEE International Conference on Data Mining (ICDM), Sorrento, Italy, 17–20 November 2020; pp. 841–850. [Google Scholar]
  15. Deng, A.; Hooi, B. Graph neural network-based anomaly detection in multivariate time series. Proc. AAAI Conf. Artif. Intell. 2021, 35, 4027–4035. [Google Scholar] [CrossRef]
  16. Bashar, M.A.; Nayak, R. Time series anomaly detection with generative adversarial networks. In Proceedings of the IEEE Symposium Series on Computational Intelligence (SSCI), Canberra, Australia, 1–4 December 2020; pp. 1778–1785. [Google Scholar]
  17. Audibert, J.; Michiardi, P.; Guyard, F.; Marti, S.; Zuluaga, M.A. Usad: Unsupervised anomaly detection on multivariate time series. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual, 6–10 July 2020; pp. 3395–3404. [Google Scholar]
  18. Xu, H.; Chen, W.; Zhao, N.; Li, Z.; Bu, J.; Li, Z.; Liu, Y.; Zhao, Y.; Pei, D.; Feng, Y.; et al. Unsupervised anomaly detection via variational auto-encoder for seasonal kpis in web applications. In Proceedings of the 2018 World Wide Web Conference, Lyon, France, 23–27 April 2018; pp. 187–196. [Google Scholar]
  19. Li, Z.; Chen, W.; Pei, D. Robust and unsupervised kpi anomaly detection based on conditional variational autoencoder. In Proceedings of the 2018 IEEE 37th International Performance Computing and Communications Conference (IPCCC), Orlando, FL, USA, 17–19 November 2018. [Google Scholar]
  20. Yan, H.; Li, Z.; Li, W.; Wang, C.; Wu, M.; Zhang, C. Contnet: Why not use convolution and transformer at the same time? arXiv 2021, arXiv:2104.13497. [Google Scholar]
  21. Zerveas, G.; Jayaraman, S.; Patel, D.; Bhamidipaty, A.; Eickhoff, C. A transformer-based framework for multivariate time series representation learning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Virtual, 14–18 August 2021; pp. 2114–2124. [Google Scholar]
  22. Xu, J.; Wu, H.; Wang, J.; Long, M. Anomaly transformer: Time series anomaly detection with association discrepancy. arXiv 2021, arXiv:2110.02642. [Google Scholar]
  23. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30–45. [Google Scholar]
  24. Hundman, K.; Constantinou, V.; Laporte, C.; Colwell, I.; Soderstrom, T. Detecting spacecraft anomalies using lstms and nonparametric dynamic thresholding. In Proceedings of the 24th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, London, UK, 19–23 August 2018; pp. 387–395. [Google Scholar]
  25. Mathur, A.P.; Tippenhauer, N.O. Swat: A water treatment testbed for research and training on ics security. In Proceedings of the 2016 International Workshop on Cyber-physical Systems for Smart Water Networks (CySWater), Vienna, Austria, 11 April 2016; pp. 31–36. [Google Scholar]
  26. Abdulaal, A.; Liu, Z.; Lancewicki, T. Practical approach to asynchronous multivariate time series anomaly detection and localization. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Virtual, 14–18 August 2021; pp. 2485–2494. [Google Scholar]
  27. Tuli, S.; Casale, G.; Jennings, N.R. Tranad: Deep transformer networks for anomaly detection in multivariate time series data. arXiv 2022, arXiv:2201.07284. [Google Scholar] [CrossRef]
Figure 1. Decompose Auto-Transformer network (DATN), framework.
Figure 1. Decompose Auto-Transformer network (DATN), framework.
Electronics 12 00354 g001
Figure 2. The proposed auto-attention module.
Figure 2. The proposed auto-attention module.
Electronics 12 00354 g002
Figure 3. Efficiency comparison of the baseline methods.
Figure 3. Efficiency comparison of the baseline methods.
Electronics 12 00354 g003
Figure 4. Visualization analysis of anomaly detection. (a) The reconstruction results. (b) Anomaly scores. (c) Anomaly detection results. (d) Anomaly labels.
Figure 4. Visualization analysis of anomaly detection. (a) The reconstruction results. (b) Anomaly scores. (c) Anomaly detection results. (d) Anomaly labels.
Electronics 12 00354 g004
Table 1. Dataset statistics.
Table 1. Dataset statistics.
DatasetTrainTestDimensionsAnomalies (%)
SMD708,405708,420384.16
SMAP135,183427,6172513.13
MSL58,31773,7295510.72
SWaT475,200449,9195112.14
PSM132,48187,8412527.76
Table 2. Performance comparison of the baseline methods (as %, red: best, blue: second best).
Table 2. Performance comparison of the baseline methods (as %, red: best, blue: second best).
MethodSMDSMAPMSLSWaTPSM
PRF1PRF1PRF1PRF1PRF1
LSTM-VAE [3]77.6389.1282.9888.6572.1879.5792.3278.6184.9281.7878.5580.1390.1674.4881.57
Omni Anomaly [12]83.3494.4988.5774.1697.7684.3488.6791.1789.8986.3376.9481.3691.6171.3680.23
MTAD-GAT [14]88.2884.9286.5789.0691.2390.1387.5494.4090.8492.4675.1282.8995.2875.6584.34
USAD [17]93.1496.1793.8276.9798.3181.8688.1097.8691.0998.7074.0284.6074.4299.0184.97
TransFram [21]91.6086.4488.9485.3687.4886.4197.2190.3393.6492.4775.8883.3688.1486.9987.56
Anomaly Tran [22]88.4296.9092.4792.1689.7988.9695.8392.6694.2286.3892.1488.1794.7588.5990.56
TranAD [27]92.6299.7496.0580.4399.9989.1590.3899.9994.9497.6069.9781.5188.1584.9189.72
DATN93.5094.3793.9296.2586.4591.1596.7593.2694.9795.3885.3290.1297.3987.4992.25
Table 3. Ablation study of series decompose and auto-attention modules.
Table 3. Ablation study of series decompose and auto-attention modules.
ComponentsF1 Measures (as %)
Series
Decompose
Auto-AttentionSMDSMAPMSLSWaTPSMAvg
××86.2185.2383.1382.2184.2284.20
×89.2388.1487.1986.3689.4388.08
×92.3890.3892.7589.1792.1591.41
93.9591.1394.9990.0592.1592.45
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wu, B.; Fang, C.; Yao, Z.; Tu, Y.; Chen, Y. Decompose Auto-Transformer Time Series Anomaly Detection for Network Management. Electronics 2023, 12, 354. https://doi.org/10.3390/electronics12020354

AMA Style

Wu B, Fang C, Yao Z, Tu Y, Chen Y. Decompose Auto-Transformer Time Series Anomaly Detection for Network Management. Electronics. 2023; 12(2):354. https://doi.org/10.3390/electronics12020354

Chicago/Turabian Style

Wu, Bo, Chao Fang, Zhenjie Yao, Yanhui Tu, and Yixin Chen. 2023. "Decompose Auto-Transformer Time Series Anomaly Detection for Network Management" Electronics 12, no. 2: 354. https://doi.org/10.3390/electronics12020354

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop