Next Article in Journal
On the Optimal File Size of Capacity-Achieving Byzantine-Resistant Private Information Retrieval Schemes
Previous Article in Journal
Comparative Evidence-Based Model Choice: A Sketch of a Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improving Multivariate Time-Series Anomaly Detection in Industrial Sensor Networks Using Entropy-Based Feature Aggregation

School of Electronics and Information Engineering, Beihang University, Beijing 100191, China
Entropy 2026, 28(1), 14; https://doi.org/10.3390/e28010014
Submission received: 23 September 2025 / Revised: 3 December 2025 / Accepted: 15 December 2025 / Published: 23 December 2025
(This article belongs to the Section Multidisciplinary Applications)

Abstract

Anomaly detection using multivariate time-series data remains a significant challenge for complex industrial systems, such as Cyber–Physical Systems (CPSs), Industrial Control Systems (ICSs), Intrusion Detection Systems (IDSs), the Internet of Things (IoT), and Remote Sensing Monitoring Platforms, including satellite Earth observation systems and Mars Rovers. In these systems, sensors are highly interconnected, and local anomalies frequently affect multiple components. Because these interconnections are often implicit and involve complex interactions, systematic characterization is required. To address this, our study employs graph neural networks with a structure-entropy-based attention mechanism, which models multi-element relationships and formally represents implicit relationships within complex industrial systems using a network-based structural model. Specifically, our method distinguishes the weights of different high-order neighbor nodes based on their locations, rather than treating all nodes equally. Through this formalization, we identify and represent key adjacent elements by analyzing system entropy. We validate our method on SMAT, MSL, SWaT, and WADI datasets, and experimental results demonstrate improved detection performance compared to baseline approaches.

1. Introduction

Anomaly detection [1,2] is a critical interdisciplinary domain focused on identifying deviations from standard operational patterns within complex industrial environments. This capability is essential across various industrial systems engineering fields, such as Cyber–Physical Systems (CPSs) [3], Industrial Control Systems (ICSs) [4], Intrusion Detection Systems (IDSs) [5], the Internet of Things (IoT) [6], and Remote Sensing Monitoring Platforms (e.g., satellite Earth observation systems and Mars Rovers). The core goal of anomaly detection is to identify potential faults in these industrial systems, which are usually evidenced by the oscillation of time-series signals collected by multiple sensors throughout the system. For example, in the NASA Mars Science Laboratory (MSL) [7] (shown in Figure 1), multiple measurements from cameras, spectrometers, and environmental sensors record time-series signals. Anomalous sensor data in the MSL system can disrupt operational logic, indicating possible malfunctions of the Mars Rover.
Current industrial anomaly detection faces three primary challenges that require the attention of researchers. The first challenge is the inherent complexity of industrial systems, which consist of numerous interconnected components such as machinery, sensors, control systems, and human–machine interfaces. These components interact non-linearly, resulting in systems that exhibit dynamic and evolving behaviors [8,9,10]. The second challenge involves the multifactorial origin of faults. Although diverse time series can be used to characterize the temporal evolution of each element (sensor) separately for detection and analysis, neglecting the system’s architecture risks disregarding the impact of physical and logical coupling among variables on fault detection. The third challenge is the stability of industrial system sensors, which usually remain stable for extended periods, changing only during specific time intervals. These shifts can alter overall system properties, potentially leading to faults. This phenomenon suggests that a minor alteration in a critical parameter can induce a qualitative shift in the system’s state [11,12].
Recently, there has been increasing interest in structurally modelling various industrial systems using multivariate time series correlations. In such modelling, the significance of sensors is characterised by the structure’s importance, and structural changes show sensor variations over time. A particularly promising direction is graph modelling, which has gained attention for directly bridging the gap between complex industrial systems and various machine learning techniques. For instance, basic Graph Convolutional Networks (GCNs) [13] treat each node as a sensor and learn structural relationships among sensors through information aggregation and propagation. Building upon this, recent research has integrated time series modelling with GCNs. For example, GTAD [14] introduced an automated graph modelling approach for IoT that utilises a transformer architecture to capture the temporal characteristics of data. Similarly, MTGNN [15] employs a novel mix-hop propagation layer and a dilated inception layer to capture spatial and temporal dependencies within the time series. In addition, StemGNN [16] applies graph spectral theory to multivariate time series modelling, explicitly using the Graph Fourier Transform (GFT) [17] to model inter-series correlations and the Discrete Fourier Transform (DFT) [18] to model temporal dependencies within an end-to-end framework. Finally, LGAT [19] automatically learns graph structures and leverages an enhanced Anomaly Transformer architecture to capture temporal dependencies. Another strategy to tackle anomaly detection is the imputation-based method, which offers a straightforward solution for incomplete data by filling gaps. This strategy assumes that signals could be accurately modeled. Typical methods in this field are ImDiffusion [20], DiffAD [21], and DSDI [22].
Graph modeling theory and its associated methodologies provide innovative strategies for anomaly detection in complex industrial systems [23]. Building on these foundations, we employ graph models to extract and represent data from multiple sensors through a neural network architecture. We introduce an entropy-based attention mechanism to generate node representations within the network. For clarity, our approach is organized into four main sections. First, we describe the Graph Structure of Sensors, utilizing directed graphs to identify implicit relationships among sensors and selecting the top K related nodes for each sensor. Second, we present the Entropy Centrality-based Graph Attention, which integrates all modules within a graph neural network framework and incorporates an entropy-based attention module to aggregate information and represent each node. Third, we detail the Loss and Training process to explain how the model is trained and its parameters are updated using gradient descent. Finally, we perform anomaly detection, which calculates the mean deviation between predicted and expected values to identify anomalies.
The innovations of this paper correspond to the main modules, which are as follows:
  • This work employs a graph model to reveal hidden structures in complex industrial systems by examining sensor interaction relationships.
  • This work introduces Entropy Centrality-based Graph Attention, which screens each node’s neighbors by analyzing entropy decay strength in the graph’s structure.
  • Experiments have shown that this method enhances the efficiency of anomaly detection tasks.
The rest of this paper is structured as follows. Section 2 introduces the definitions of multivariate time-series data and graph models. Section 3 introduces the framework of this work, including the Graph Structure of Sensors, Entropy Centrality-based Graph Attention, and anomaly detection. Section 4 presents various experiments, essential details, results, and discussions. Section 5 summarizes this work and provides a future outlook.

2. Problem Statement

2.1. Multivariate Time-Series Data

In this work, we focus on multivariate time-series anomaly detection. Let x ( t ) R N × 1 denote the original multivariate time-series data at any time stamp t, where N is the total number of sensors and x i ( t ) is the i-th sensor’s record data. The high imbalance between normal data and anomalies is a common characteristic of various anomaly detection tasks, where the amount of normal data is significantly greater than that of anomalies. Anomaly detection aims to predict anomalies in the test data based on normal data.
In mathematics, anomaly detection aims to learn a prediction function F that maps the input to the label field based on historical ω -step time series. Specifically, the m-th input is denoted as the ω -step time series
X m : = [ X 1 m , , X N m ] = [ x ( m ω + ( m 1 ) τ ) , , x ( ( m 1 ) ω + ( m 1 ) τ ) ] .
where τ is the sliding step. Here, X i m is the i-th sensor’s ω -step record time series. This prediction function is denoted as
F : X m Y , X m Y ,
where Y is the binary label pending prediction and ω is the window size. Then, the training dataset collects a set of M different inputs:
X T r a i n = { X 1 , , X M } .
The visualization of the training set is presented in Figure 2.

2.2. Graph Model

Graph is a kind of structural data that organizes various items within a system with edges connecting them. Formally, a graph is denoted as G = { V , E } , where V = { v 1 , , v n } consists of n elements and E V × V is the edge set.
Graphs can be categorized as undirected or directed based on the directionality of their edges. Undirected graphs have edges without directionality, meaning that ( v i , v j ) = ( v j , v i ) . Directed graphs have edges with explicit directionality, implying that ( v i , v j ) ( v j , v i ) . Specifically, for an edge ( v i , v j ) with direction v i v j , the node v i refers to the source node and v j is the target node. The directionality of graph edges is typically captured using an n × n adjacency matrix A , where the matrix entry at row i and column j indicates the presence or absence of an edge from node v i to node v j . A is determined as follows:
A i j = 1 , i f   ( v i , v j ) E , 0 , o t h e r w i s e .

3. Method

The method includes three parts: Graph Structure of Sensors, Entropy Centrality-based Graph Attention, and anomaly detection. The framework is shown in Figure 3.
  • Step 1. Sensor Graph Generation utilizes directed graphs to extract implicit relationships among sensors in complex industrial systems, selecting the top K related nodes for each node.
  • Step 2. Entropy Centrality-based Feature Aggregation integrates all preceding modules within a graph neural network framework by employing an entropy-based attention module that gathers information from related nodes and represents each node accordingly.
  • Step 3. Loss and Training trains and updates parameters vis gradient descent.
  • Step 4. Anomaly Detection implements anomaly detection by calculating the mean deviation between predicted and expected values.

3.1. Sensor Graph Generation

This section aims to model the graph structure of complex industrial systems based on multivariate time-series data. It primarily involves representing system participants as nodes and modeling the relationships between these participants within industrial systems. Specifically, we denote G m as the Sensor Graph derived from m-th ω -length multivariate time-series X m .

3.1.1. Representing Nodes

In complex industrial systems, including the NASA Mars Science Laboratory (MSL) rover dataset, data is collected from multiple sensors monitoring various rover components. When a fault occurs, several sensors often trigger alerts at once, indicating correlations among specific sensors. These correlations are difficult to spot in multivariate time-series data. This section aims to represent sensors as graph nodes using learnable vectors. These vectors are then optimized by model training in later sections. In this context, we employ the graph model G = ( V , E ) to depict a sensor network. The set V = { v 1 , , v n } represents the nodes, and the set E = { ( v 1 , v 2 ) , , ( v n 1 , v n ) } represents the connections between them. To obtain a vector representation for each node, we utilize learnable vectors encapsulated in an n × d feature matrix V . Each row v i R d within this matrix corresponds to the feature vector for node v i in the m-th graph, G m , where i ranges from 1 to n.

3.1.2. Representing Edges

To elucidate the interdependencies within complex industrial systems, this paper employs directed edges and graphs to model the influence exerted by one sensor upon others. Since structural relationships among sensors are generally absent, the most straightforward approach is to construct a fully connected graph, in which each sensor is influenced by all others. However, this representation is overly coarse. Therefore, this paper focuses solely on the top K sensors that significantly influence a given sensor.
For a node v i G and any other node v j G , the weight of any possible directed edge ( v j , v i ) can be represented by a normalization function:
W j i = v j , v i v j · v i .
Then, the real edges are selected via arranging the top-k weight scores of all possible edges connected to v i . That is,
E ( v i ) = { ( v j , v i ) | v j TopK ( v i ) } ,
where TopK ( v i ) is the set of nodes that derives the top K weighted scores. Consequently, the directed edge structure of the graph can be obtained and represented by the directed adjacency matrix A , where A j i = 1 if v j E ( v i ) .

3.2. Entropy Centrality-Based Feature Aggregation

In industrial systems, components and sensors construct complex networks through the Industrial Internet and the Industrial Internet of Things. Each sensor acts as a node within the system’s overall graph structure, while the connections between them represent system interactions. This complex system builds upon the previously described graph model, which captures its topological features. Because structure embodies information, this section will assess the significance of each node (each sensor) within the entire graph (the industrial complex system) from an information entropy perspective.
Entropy is a fundamental concept in statistical physics, and the second law of thermodynamics indicates that it is challenging to decrease the entropy of a macroscopic system [24]. Shannon introduced the idea of entropy into information theory as a measure to quantify the information associated with a system [25,26]. A key conclusion is that the more essential an element is, the greater its contribution to the system’s entropy [27]. Therefore, removing a critical element results in a more significant decrease in the system’s overall entropy (Figure 4).
For a discrete system S and each element s S , the basic definition entropy of S is denoted as
H ( S ) = s S p ( s ) log p ( s ) ,
where p ( s ) is the occurring probability of element s and I ( s ) = log p ( s ) is the corresponding self-information [28].
Various definitions exist for the probability of the occurrence of elements s. In this graph modeling context, graph G represents the discrete system S , with each node v corresponding to an element s. This study leverages the graph’s structural characteristics to redefine entropy as
H ( G ) = v i V p ( v i ) log p ( v i ) .
Here
p ( v i ) = f ( v i ) v j V f ( v j ) ,
where
f ( v i ) = ( v i , v j ) E v i , v j .
Thus, the entropy centrality of node v i is denoted as the difference between original graph and the graph without v i ,
H ( v i ) = H ( G ) H ( G { v i } ) .
The pseudocode is shown in Algorithm 1.
Algorithm 1: greenGraph Entropy Function H ( G )
Entropy 28 00014 i001
In graph neural network computations, each node in the graph obtains its representation by aggregating information from its own and its surrounding low-order or high-order neighbors. Common graph-attention-mechanism neural networks aggregate patterns [29,30,31], which adjust the weights of surrounding neighbors adaptively to obtain target node aggregation. Their basic form is
z i m = ReLU ( α i , i Θ 1 X i m + ( v i , v j ) E α i , j Θ 1 X j m ) .
Here, X i m R ω is the input of node (sensor) v i , which is the i-th row of input X i m in Formula (1). Θ 1 is the weight matrix pending training, and α i , i , α i , j are the attention coefficients from attention mechanism.
In this work, we propose a novel attention calculation using entropy centrality. Specifically,
α i , j = exp ( π ( v i , v j ) ) v r { v i } { v k | ( v i , v k ) E } exp ( π ( v i , v r ) ) ,
where
π ( v i , v j ) = LeakyReLU ( H ( v j ) ) = LeakyReLU ( H ( G ) H ( G { v j } ) ) .

3.3. Loss and Training

In the feature extractor layer, this work get the aggregated representations for all nodes of m-th sensor graph G m , which are Z m = { z 1 m , , z N m } . Then, this work uses the m-th input of overall sensor values by stacked fully-connected layers with element-wise multiply of z i m and v i as input. The output is
s ^ m = { s ^ 1 m , , s ^ N m } = f Θ 2 ( [ z 1 m v 1 , , z N m v N ] ) ,
where ∘ is the element-wise multiply, and Θ 2 is the learned parameters.
This work uses the overall Mean Squared Error (MSE) as the regression loss function to evaluate the prediction on the training data. The format is
L = 1 M m = 1 M s ^ m s m 2 2 = 1 M m = 1 M i = 1 N ( s ^ i m s i m ) ,
where s m = { s 1 m , , s N m } is the real observed data during training.

3.4. Anomaly Detection

Following a malfunction in a complex system, abrupt anomalies often appear in the time-series data of its components. This paper individually computes the reconstruction through anomaly scores for each sensor to identify these anomalies. Subsequently, these sensor-specific anomaly scores are aggregated to derive an overall system anomaly score for the specified time period. For i-th sensor v i , the anomaly score is
E R R O R ( v i , m ) = | s i m s ^ i m | .
Given the potential for significant variability in the characteristics of different sensors, their bias values may also exhibit disparate dimensions. We implement a robust normalization procedure for each sensor’s error values to mitigate the undue influence of individual sensor biases on the overall outcome.
N E R R O R ( v i , m ) = E R R O R ( v i , m ) μ ˜ i σ ˜ i .
Here, μ ˜ i and σ ˜ i are the mean and standard deviations of E R R O R ( v i , m ) .
After that, this work identifies the sensor with the largest anomaly score for the m-th input as the signal of anomalousness. This is because only a few sensors affect the overall anomalousness of the entire industrial complex system. The overall Anomaly Detection Score is
M E R R O R ( m ) = max v i G N E R R O R ( v i , m ) .
Consequently, by setting an appropriate threshold η , M E R R O R ( m ) values exceeding this threshold can be classified as anomalies, which satisfy M E R R O R ( m ) η . The most straightforward approach involves utilizing the expected value derived from the training set’s statistical distribution, which represents the typical behavior in the absence of anomalies. The anomaly detection is shown in Algorithm 2.

4. Experiments

This section presents a comparative analysis of the proposed methodology against several state-of-the-art models for multivariate time-series anomaly detection on the SMAP, MSL, SWaT, and WADI datasets. The experimental implementations, findings, analysis, and discussion are subsequently proposed.

4.1. Datasets

This paper uses four publicly available multivariate time-series anomaly detection datasets, whose details are summarized in Table 1. The details of these datasets are as follows, including the volumes of the training and test sets, their dimensions, and the anomaly rates within the datasets.
  • The Soil Moisture Active Passive (SMAP) dataset [32] encompasses soil samples and telemetry data from NASA’s Mars Rover.
  • The Mars Science Laboratory (MSL) dataset [33] is analogous to SMAP. MSL includes sensor and actuator data from the Mars Rover. However, the dataset is distinguished by a high prevalence of trivial sequences.
  • The Secure Water Treatment (SWaT) dataset [34] is derived from a real-world water treatment plant, comprising 7 days of normal operation and 4 days of abnormal operation. It incorporates sensor readings (e.g., water level, flow rate) and actuator commands (e.g., valve and pump control).
  • The Water Distribution (WADI) dataset [35] is an extension of the SWaT system. The WADI dataset integrates more than twice the number of sensors and actuators compared to the SWaT model. The dataset spans a longer period, with 14 days of normal operation and 2 days of attack scenarios.
Algorithm 2: greenAnomaly Detection
Entropy 28 00014 i002

4.2. Baselines

This study implements twelve baseline methodologies: IsolationForest, LSTM, LSTM-NDT, MAD-GAN, MTAD-GAT, DAGMM, GDN, OmniAnomaly, USAD, ATUAD, MSCRED, and MSAD. Then, the experimental outcomes of these baseline algorithms are compared with the results of the proposed methodology.
  • IsolationForest [36]: IsolationForest is an unsupervised anomaly detection algorithm that focuses on isolating anomalies rather than profiling normal data. It constructs random decision trees to separate outliers, which require fewer splits due to their unique characteristics. This algorithm is efficient, exhibiting low linear time complexity, and performs well on high-dimensional datasets without requiring distance calculations.
  • LSTM [37]: LSTM (Long Short-Term Memory) is a recurrent neural network (RNN) architecture designed to capture long-term dependencies in sequential data. It utilizes specialized memory cells with input, output, and forget gates to control information flow, thereby preventing vanishing gradients in deep learning.
  • LSTM-NDT [38]: LSTM-NDT enhances traditional LSTM by integrating adaptive thresholding for anomaly detection. It dynamically adjusts decision boundaries in time-series data, improving robustness against noise while maintaining sensitivity to true anomalies.
  • MAD-GAN [39]. MAD-GAN is a GAN-based multivariate time-series anomaly detection method. It employs adversarial training between a generator and a discriminator to model the normal data distribution using reconstruction error for anomaly identification. Its strength lies in capturing intricate spatiotemporal dependencies, making it suitable for applications like industrial IoT. In a typical implementation, an LSTM encoder analyzes local and global anomaly patterns within high-dimensional time-series data.
  • MTAD-GAT [40]: MTAD-GAT detects anomalies in interdependent sensors by modeling relationships via GAT. It combines feature reconstruction and forecasting errors, excelling in industrial IoT and system monitoring.
  • DAGMM [41]: DAGMM is an unsupervised anomaly detection model that combines deep autoencoders with Gaussian Mixture Models (GMMs). It jointly optimizes dimensionality reduction and density estimation via end-to-end training, leveraging reconstruction error and low-dimensional features to identify anomalies effectively
  • GDN [42]: GDN is a novel GNN framework that learns sensor dependencies via attention-based graph structure learning and detects anomalies by scoring deviations from learned patterns, achieving interpretability and superior accuracy on water treatment datasets
  • OmniAnomaly [43]: OmniAnomaly is a robust stochastic recurrent neural network for multivariate time-series anomaly detection. It learns standard patterns through stochastic variable connections and a planar normalizing flow, utilizing reconstruction probabilities to identify and interpret anomalies, achieving a state-of-the-art F1-Score.
  • USAD [44]: USAD is an unsupervised anomaly detection framework characterized by an adversarial training architecture. USAD is capable of achieving both efficient learning and high detection accuracy.
  • ATUAD [45]: ATUAD is an unsupervised anomaly detection model that integrates adversarial training with the Transformer architecture to amplify subtle anomalies and enhance robustness.
  • MSCRED [46]: MSCRED leverages a multi-scale convolutional recurrent encoder–decoder architecture to achieve effective spatio-temporal modeling for end-to-end anomaly detection.
  • MSAD [47]: MSAD refers to the multivariate time series based on multi-standard fusion, which understands the data by analyzing the data density and sample spacing.

4.3. Parameters

The proposed methodology was implemented using a Python 3.6 and PyTorch 1.7 environment. During the experimental phase, 20% of the training dataset was designated as the validation set, with the remaining 80% allocated for training purposes. The training set is used to train the model, while the validation set assesses the performance of the trained model. This work utilizes the Adam optimizer [48] to train our model, employing an initial learning rate of 0.01 and a step scheduler with a step size of 0.5. The number of layers in the Transformer encoders is two, the number of layers in the feed-forward unit of the encoders is two, there are 64 hidden units in the encoder layers, and the dropout rate is 0.1. The most important parameters are the sliding window w and the sliding step m, which are crucial for extracting dynamic and localized information. The values used in the experiments for these two parameters are covered in Table 2.

4.4. Data Preprocessing and Evaluation Metrics

4.4.1. Data Preprocessing

Data standardization was used before training to improve the robustness of all models. Data standardization balances and resizes each data point x to the maximum and minimum of the training data.
x ¯ = x min ( X ) max ( X ) min ( X ) .

4.4.2. Evaluation Metrics

The detection performance of all models was evaluated using Precision , Recall , and F 1 -Score [49], in which
Precision = TP TP + FP ; Recall = TP TP + FN ; F 1 = 2 × Precision × Recall Precision + Recall .
In this context, TP denotes the true positives, which are the anomalies that are correctly identified. FP signifies the false positives, representing instances incorrectly flagged as anomalies. TN refers to the true negatives, indicating the accurate classification of normal samples. Finally, FN represents the false negatives, which are the normal samples that were incorrectly classified as anomalies.

4.5. Results

The anomaly detection results, including Precision, Recall, and F1-Score, for the proposed method and baseline models are presented in Table 3. Specifically, the results for the baselines are sourced from the published literature.
For the SMAP dataset, regarding Precision, the highest score of 0.8941 is achieved by LSTM, while the proposed method scores 0.8552, ranking third after LSTM and DAGMM, whose results are 0.8941 and 0.8645, respectively. Regarding Recall, the proposed method scored 0.9790, ranking third behind the top scores of 0.9991 by MTAD-GAT and 0.9891 by GDN. For the composite F1-Score, the proposed method achieves 0.9071, ranking the top. This shows that the proposed method achieves a favorable balance between Precision and Recall. For the MSL dataset, the proposed method still achieves a balanced performance. Concerning Precision, the highest score of 0.9308 was obtained by GDN, with the proposed method scoring 0.9182, ranking second, 0.0126 lower than GDN. Regarding Recall, the highest score of 1.0000 was achieved by LSTM-NDT, while the proposed method scored 0.9941, ranking second. For the composite F1-Score, the proposed method achieved 0.9546, ranking second and slightly below GDN’s 0.9591. For the SWaT dataset, the proposed method shows the best Precision over all baselines. In terms of Precision, the proposed method achieves the highest score of 0.99828, which is 0.0046 higher than the second result, achieved by OmniAnomaly. For Recall, the highest score is achieved by MAD-GAN at 0.6957, while the proposed method scores 0.6563, which is 0.0394 lower than MAD-GAN. For the composite F1-Score, the proposed method achieves 0.7870. For the WADI dataset, concerning Precision, the highest score is achieved by OmniAnomaly at 0.3158, with the proposed method scoring 0.3029, which is 0.0129 lower than OmniAnomaly. Regarding Recall, the proposed method scored 0.8196, lower than DAGMM’s 0.9981 and MAD-GAN’s 0.9124. For the F1-Score, the proposed method achieved a score of 0.4423, ranking first and outperforming the second-place GDN and OmniAnomaly at 0.4260.
In summary, the proposed method performs favorably on the SMAP, MSL, and WADI datasets. On these datasets, the proposed method ranks first or second overall in terms of the F1-Score. On SWaT, the proposed method fails to yield a favorable result in terms of the overall F1-Score, but demonstrates a powerful discrimination ability regarding Precision.

4.6. Ablation Experiment

The primary innovation of this paper lies in the complex network association of sensors, coupled with node feature aggregation via graph neural network computation, leveraging a method of entropy analysis for complex systems. This section removes the influence of Entropy Centrality-based Feature Aggregation, where in Formula (12), α i , i and α i , j are assigned equal weights, i.e., α i , i = α i , j .
E n t r o p y C e n t r a l i t y : z i m = ReLU ( α i , i Θ 1 X i m + ( v i , v j ) E α i , j Θ 1 X j m ) , R e m o v e E n t r o p y C e n t r a l i t y : z i m = ReLU ( Θ 1 X i m + ( v i , v j ) E Θ 1 X j m ) .
Ablation experiments were conducted by comparing the computational results before and after removing entropy centrality to validate the effectiveness of the entropy centrality design. The compared results are shown in Figure 5 and Table 4. It is evident that the proposed method enhances performance compared to the method that removes entropy centrality.

4.7. Discussion

This paper employs unsupervised experiments and an ablation study to evaluate the effectiveness of the proposed method. In the unsupervised experiments, our approach demonstrates superior performance, achieving excellent results on key metrics including Precision and F1-Score. The ablation study further validates the core innovation of this work: the entropy-centric attention mechanism. Unlike the ablation baseline, which fuses data using neural networks alone and treats all neighboring nodes equally, our proposed mechanism assigns differentiated weights to high-order neighboring nodes based on their topological positions. The results conclusively show that this entropy-based mechanism is vital for the effective aggregation of node features and constitutes the principal advantage of our method over other network-based approaches.
Regarding benchmarks of graph-based methods, existing approaches often rely on pre-defined or learned graph structures, such as using correlation or proximity measures, to model relationships among time-series sensors. While these methods incorporate graph neural networks for information aggregation, they typically treat neighboring nodes uniformly or assign attention weights based on superficial node feature similarities. In contrast, our method introduces an entropy-based method by quantifying the informational importance of nodes through entropy centrality. This allows the model to assign differentiated attention weights not merely based on features, but on the relative uncertainty and influence of a node within the higher-order neighborhood structure. Consequently, our mechanism can effectively capture connection in sensor networks, which is often critical for identifying anomalies that are missed by methods using static or feature-only attention.
The proposed model consists of four main components, which are Sensor Graph Generation, Entropy Centrality-based Feature Aggregation, Loss and Training, and anomaly detection. Among the four modules, the core module is Entropy Centrality-based Feature Aggregation. The advantage of this method lies in optimizing the attention module solely through entropy centrality, which can be generalized to other attention-based methods. For an n-node graph, the complexity of Sensor Graph Generation is O ( d · n 2 + l o g K · n 2 ) , where d denotes the dimension of the feature vector of each node v i . The complexity of calculating entropy centrality is O ( n · | E | · d + n 2 ) ; the complexity of Feature Aggregation is O ( n · d i n · d o u t + n 2 ) . The complexity of Loss and Training is O ( m · n ) . The complexity of anomaly detection is O ( n ) . Thus, the overall upper bound of the proposed model is O ( n 2 ) .
The hyperparameter that influences anomaly detection is threshold η . As for the selection of threshold η , we used the common “defining anomalies by proportion” method [50]. This method defines the threshold by the proportion of the data the work is willing to label as anomalous. For instance, in this manuscript, we assume each dataset is expected to contain 1% anomalies, so the system will rank all anomaly scores in ascending order and demarcate a cutoff point immediately following the top 1%. All data points with scores exceeding this threshold will be classified as anomalies. Figure 6 shows the ROC and AUC curves based on different proportions of anomalous data in the MSL dataset.

5. Conclusions

In various industrial scenarios, researchers diagnose anomalies within industrial systems by analysing time-series data recorded by sensors. Specifically, sensors interact and correlate to form complex systems, the structure of which is determined by the characteristics of the sensors, such as sensor feature data and recorded data. This paper utilizes a graph-based network model to model the implicit structure within complex industrial systems. Further, it represents and analyzes the graph structure through graph neural networks for multivariate time series. Addressing the information aggregation representation problem of graph nodes in neural network calculations, this paper proposes an entropy-based attention mechanism to induce the aggregation and representation of information from neighboring nodes to target nodes. The advantage of the proposed method lies in its ability to mine the implicit interaction relationships between multivariate time series from the perspective of complex system structure modeling and to determine the implicit anomalous characteristics in multivariate time series from the perspective of complex system graph structure analysis.
Several directions warrant further discussion in future work. One important area is the problem of effective anomaly detection under small-sample conditions. Industrial scenarios often have only a small amount of labeled data; therefore, unsupervised, semi-supervised, and few-shot learning have greater applicability. Specifically, effectively augmenting a small number of labels or anomalous samples is necessary to ensure the model’s effectiveness. Another related point of interest involves the window length and sliding distance hyperparameters for time-series training-sample sampling, which are often optimized through multiple experiments. These parameters often determine the amount of information contained in the data used for training. Therefore, analyzing the quantitative relationship between these hyperparameters and the amount of information is of great significance for quickly estimating key hyperparameters under new datasets.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on reasonable request from the corresponding author.

Acknowledgments

Beihang University generously supported the research. The authors sincerely thank the editors and anonymous reviewers for their diligent efforts and kind comments.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Rafique, S.H.; Abdallah, A.; Musa, N.S.; Murugan, T. Machine learning and deep learning techniques for internet of things network anomaly detection—Current research trends. Sensors 2024, 24, 1968. [Google Scholar] [CrossRef] [PubMed]
  2. Li, Q.; Tang, Y.; Chu, L. Generative adversarial networks for prognostic and health management of industrial systems: A review. Expert Syst. Appl. 2024, 253, 124341. [Google Scholar] [CrossRef]
  3. Rani, S.; Kataria, A.; Kumar, S.; Karar, V. A new generation cyber-physical system: A comprehensive review from security perspective. Comput. Secur. 2025, 148, 104095. [Google Scholar] [CrossRef]
  4. Rawat, H.; Rawat, A.; Rajavat, A.; Rawat, R. Literature Review Analysis for Cyberattacks at Management Applications and Industrial Control Systems. In Information Visualization for Intelligent Systems; Wiley: Hoboken, NJ, USA, 2025; pp. 461–487. Available online: https://www.wiley.com/en-us/Information+Visualization+for+Intelligent+Systems+-p-9781394305797 (accessed on 2 December 2025).
  5. Rahman, M.M.; Al Shakil, S.; Mustakim, M.R. A survey on intrusion detection system in IoT networks. Cyber Secur. Appl. 2025, 3, 100082. [Google Scholar] [CrossRef]
  6. Mishra, R.; Mishra, A. Current research on Internet of Things (IoT) security protocols: A survey. Comput. Secur. 2025, 151, 104310. [Google Scholar] [CrossRef]
  7. Denison, J.; Maimone, M. Improving Sequence Traceability During Testing and Review for the Mars Science Laboratory. In Proceedings of the 2025 IEEE Aerospace Conference, Big Sky, MT, USA, 1–8 March 2025; pp. 1–13. [Google Scholar]
  8. Janson, N.B. Non-linear dynamics of biological systems. Contemp. Phys. 2012, 53, 137–168. [Google Scholar] [CrossRef]
  9. Karmeshu; Jain, V. Non-linear models of social systems. Econ. Political Wkly. 2003, 38, 3678–3685. [Google Scholar]
  10. Horváth, I.; Rusák, Z.; Li, Y. Order beyond chaos: Introducing the notion of generation to characterize the continuously evolving implementations of cyber-physical systems. In Proceedings of the International Design Engineering Technical Conferences and Computers and Information in Engineering Conference, Cleveland, OH, USA, 6–9 August 2017; American Society of Mechanical Engineers: New York City, NY, USA, 2017; Volume 58110, p. V001T02A015. [Google Scholar]
  11. Papon, P.; Leblond, J.; Meijer, P.H. Physics of Phase Transitions; Springer: Berlin/Heidelberg, Germany, 2002. [Google Scholar]
  12. Landau, L. The theory of phase transitions. Nature 1936, 138, 840–841. [Google Scholar] [CrossRef]
  13. Sadasivan, A.; Gananathan, K.; Pal Nesamony Rose Mary, J.D.; Balasubramanian, S. A Systematic Survey of Graph Convolutional Networks for Artificial Intelligence Applications. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2025, 15, e70012. [Google Scholar]
  14. Guan, S.; Zhao, B.; Dong, Z.; Gao, M.; He, Z. GTAD: Graph and temporal neural network for multivariate time series anomaly detection. Entropy 2022, 24, 759. [Google Scholar] [CrossRef]
  15. Zheng, S.; Hou, Y.; Elwakil, E. Enhanced Prediction of Water Pipeline Failures Using a Novel MTGNN-Based Classification Model for Smart City Infrastructure. In Proceedings of the 2025 IEEE Conference on Technologies for Sustainability (SusTech), Los Angeles, CA, USA, 20–23 April 2025; pp. 1–6. [Google Scholar]
  16. Cao, D.; Wang, Y.; Duan, J.; Zhang, C.; Zhu, X.; Huang, C.; Tong, Y.; Xu, B.; Bai, J.; Tong, J.; et al. Spectral temporal graph neural network for multivariate time-series forecasting. Adv. Neural Inf. Process. Syst. 2020, 33, 17766–17778. [Google Scholar]
  17. Zhang, Y.; Li, B.Z. The graph fractional Fourier transform in Hilbert space. arXiv 2025, arXiv:2403.10527. [Google Scholar] [CrossRef]
  18. Smith, J.O. Mathematics of the Discrete Fourier Transform (DFT): With Audio Applications; Smith, J., Ed.; Amazon Digital Services: Seattle, WA, USA, 2007. [Google Scholar]
  19. Wen, M.; Chen, Z.; Xiong, Y.; Zhang, Y. LGAT: A novel model for multivariate time series anomaly detection with improved anomaly transformer and learning graph structures. Neurocomputing 2025, 617, 129024. [Google Scholar] [CrossRef]
  20. Chen, Y.; Zhang, C.; Ma, M.; Liu, Y.; Ding, R.; Li, B.; He, S.; Rajmohan, S.; Lin, Q.; Zhang, D. Imdiffusion: Imputed diffusion models for multivariate time series anomaly detection. arXiv 2023, arXiv:2307.00754. [Google Scholar] [CrossRef]
  21. Xiao, C.; Gou, Z.; Tai, W.; Zhang, K.; Zhou, F. Imputation-based time-series anomaly detection with conditional weight-incremental diffusion models. In Proceedings of the 29th ACM SIGKDD Conference on Knowledge Discovery and Data Mining, Long Beach, CA, USA, 6–10 August 2023; pp. 2742–2751. [Google Scholar]
  22. Xiao, C.; Jiang, X.; Du, X.; Yang, W.; Lu, W.; Wang, X.; Chetty, K. Boundary-enhanced time series data imputation with long-term dependency diffusion models. Knowl.-Based Syst. 2025, 310, 112917. [Google Scholar] [CrossRef]
  23. Li, X.; Cheng, D.; Zhang, L.; Zhang, C.; Feng, Z. Few-Shot Graph Anomaly Detection via Dual-Level Knowledge Distillation. Entropy 2025, 27, 28. [Google Scholar] [CrossRef]
  24. Liu, X.; Sun, D.; Wei, W. Alleviating the over-smoothing of graph neural computing by a data augmentation strategy with entropy preservation. Pattern Recognit. 2022, 132, 108951. [Google Scholar] [CrossRef]
  25. Bromiley, P.; Thacker, N.; Bouhova-Thacker, E. Shannon entropy, Renyi entropy, and information. Stat. Inf. Ser. 2004, 9, 2–8. [Google Scholar]
  26. Wei, W.; Li, H.; Zhou, S.; Li, B.; Liu, X. Attention With System Entropy for Optimizing Credit Assignment in Cooperative Multi-Agent Reinforcement Learning. IEEE Trans. Autom. Sci. Eng. 2025, 22, 14775–14787. [Google Scholar] [CrossRef]
  27. Wei, G.; Liu, Y. A Graph Contrastive Learning Method for Enhancing Genome Recovery in Complex Microbial Communities. Entropy 2025, 27, 921. [Google Scholar] [CrossRef]
  28. Greven, A.; Keller, G.; Warnecke, G. Entropy; Princeton University Press: Princeton, NJ, USA, 2014. [Google Scholar]
  29. Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Lio, P.; Bengio, Y. Graph attention networks. arXiv 2017, arXiv:1710.10903. [Google Scholar]
  30. Elahi, E.; Anwar, S.; Al-kfairy, M.; Rodrigues, J.J.; Ngueilbaye, A.; Halim, Z.; Waqas, M. Graph attention-based neural collaborative filtering for item-specific recommendation system using knowledge graph. Expert Syst. Appl. 2025, 266, 126133. [Google Scholar] [CrossRef]
  31. Zhang, X.; Zhou, Y.; Wang, J.; Lu, X. Personal Interest Attention Graph Neural Networks for Session-Based Recommendation. Entropy 2021, 23, 1500. [Google Scholar] [CrossRef] [PubMed]
  32. O’Neill, P.; Entekhabi, D.; Njoku, E.; Kellogg, K. The NASA soil moisture active passive (SMAP) mission: Overview. In Proceedings of the 2010 IEEE International Geoscience and Remote Sensing Symposium, Honolulu, HI, USA, 25–30 July 2010; pp. 3236–3239. [Google Scholar]
  33. Malin, M.C.; Ravine, M.A.; Caplinger, M.A.; Tony Ghaemi, F.; Schaffner, J.A.; Maki, J.N.; Bell, J.F., III; Cameron, J.F.; Dietrich, W.E.; Edgett, K.S.; et al. The Mars Science Laboratory (MSL) mast cameras and descent imager: Investigation and instrument descriptions. Earth Space Sci. 2017, 4, 506–539. [Google Scholar] [CrossRef] [PubMed]
  34. Mathur, A.P.; Tippenhauer, N.O. SWaT: A water treatment testbed for research and training on ICS security. In Proceedings of the 2016 International Workshop on Cyber-Physical Systems for Smart Water Networks (CySWater), Vienna, Austria, 11 April 2016; pp. 31–36. [Google Scholar]
  35. Sahin, A.K.; Cavdar, B.; Dogan, R.O.; Ayas, S.; Ozgenc, B.; Ayas, M.S. A hybrid CNN-LSTM framework for unsupervised anomaly detection in water distribution plant. In Proceedings of the 2023 Innovations in Intelligent Systems and Applications Conference (ASYU), Sivas, Turkey, 11–13 October 2023; pp. 1–6. [Google Scholar]
  36. Liu, F.T.; Ting, K.M.; Zhou, Z.H. Isolation forest. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 413–422. [Google Scholar] [CrossRef]
  37. Hochreiter, S.; Schmidhuber, J. Long short-term memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  38. Malhotra, P.; Ramakrishnan, A.; Anand, G.; Vig, L.; Agarwal, P.; Shroff, G. LSTM-based encoder-decoder for multi-sensor anomaly detection. ACM Trans. Knowl. Discov. Data 2020, 14, 1–20. [Google Scholar] [CrossRef]
  39. Li, D.; Chen, D.; Jin, B.; Shi, L.; Goh, J.; Ng, S.K. MAD-GAN: Multivariate anomaly detection for time series data with generative adversarial networks. In International Conference on Artificial Neural Networks; Springer: Cham, Switzerland, 2019; pp. 703–716. [Google Scholar]
  40. Zhao, H.; Wang, Y.; Duan, J.; Huang, C.; Cao, D.; Tong, Y.; Xu, B.; Bai, J.; Tong, J.; Zhang, Q. Multivariate time-series anomaly detection via graph attention network. In Proceedings of the 2020 IEEE International Conference on Data Mining (ICDM), Sorrento, Italy, 17–20 November 2020; pp. 841–850. [Google Scholar]
  41. Zong, B.; Song, Q.; Min, M.R.; Cheng, W.; Lumezanu, C.; Cho, D.; Chen, H. Deep Autoencoding Gaussian Mixture Model for Unsupervised Anomaly Detection. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  42. Deng, A.; Hooi, B. Graph Neural Network-Based Anomaly Detection in Multivariate Time Series. arXiv 2021, arXiv:2106.06947. [Google Scholar] [CrossRef]
  43. Su, Y.; Zhao, Y.; Niu, C.; Liu, R.; Sun, W.; Pei, D. Robust Anomaly Detection for Multivariate Time Series through Stochastic Recurrent Neural Network. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 2828–2837. [Google Scholar] [CrossRef]
  44. Audibert, J.; Michiardi, P.; Guyard, F.; Marti, S.; Zuluaga, M.A. Usad: Unsupervised anomaly detection on multivariate time series. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual Event, 6–10 July 2020; pp. 3395–3404. [Google Scholar]
  45. Yu, X.; Zhang, K.; Liu, Y.; Zou, B.; Wang, J.; Wang, W.; Qian, R. Adversarial Transformer-Based Anomaly Detection for Multivariate Time Series. IEEE Trans. Ind. Inform. 2025, 21, 2471–2480. [Google Scholar] [CrossRef]
  46. Zhang, C.; Song, D.; Chen, Y.; Feng, X.; Lumezanu, C.; Cheng, W.; Ni, J.; Zong, B.; Chen, H.; Chawla, N.V. A deep neural network for unsupervised anomaly detection and diagnosis in multivariate time series data. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; Volume 33, pp. 1409–1416. [Google Scholar]
  47. Tian, H.; Kong, H.; Lu, S.; Li, K. Unsupervised anomaly detection of multivariate time series based on multi-standard fusion. Neurocomputing 2025, 611, 128634. [Google Scholar] [CrossRef]
  48. Hospodarskyy, O.; Martsenyuk, V.; Kukharska, N.; Hospodarskyy, A.; Sverstiuk, S. Understanding the adam optimization algorithm in machine learning. In Proceedings of the CITI’2024: 2nd International Workshop on Computer Information Technologies in Industry 4.0, Ternopil, Ukraine, 12–14 June 2024. [Google Scholar]
  49. Goutte, C.; Gaussier, E. A probabilistic interpretation of precision, recall and F-score, with implication for evaluation. In European Conference on Information Retrieval; Springer: Berlin/Heidelberg, Germany, 2005; pp. 345–359. [Google Scholar]
  50. Wang, Y.; Wu, H.; Dong, J.; Liu, Y.; Long, M.; Wang, J. Deep Time Series Models: A Comprehensive Survey and Benchmark. arXiv 2024, arXiv:2407.13278. [Google Scholar] [CrossRef]
Figure 1. The 27 sensors on the Mars Rover record time-series signals, including cameras, spectrometers, and environmental sensors. All sensors construct a complex network with edges indicating potential connections.
Figure 1. The 27 sensors on the Mars Rover record time-series signals, including cameras, spectrometers, and environmental sensors. All sensors construct a complex network with edges indicating potential connections.
Entropy 28 00014 g001
Figure 2. The training data for the work. Each training dataset is an ω -length multivariate time series with a sliding step of τ .
Figure 2. The training data for the work. Each training dataset is an ω -length multivariate time series with a sliding step of τ .
Entropy 28 00014 g002
Figure 3. The proposed method comprises four main steps: (1) Sensor Graph Generation uses directed graphs to find hidden links among sensors in industrial systems, selecting the top K related nodes for each node. (2) Entropy Centrality-Based Feature Aggregation puts all modules into a graph neural network and uses an Entropy-Based Attention module to combine information and represent each node. (3) Loss and Training trains and updates parameters using gradient descent. (4) Anomaly detection finds anomalies by measuring the mean difference between predicted and expected values.
Figure 3. The proposed method comprises four main steps: (1) Sensor Graph Generation uses directed graphs to find hidden links among sensors in industrial systems, selecting the top K related nodes for each node. (2) Entropy Centrality-Based Feature Aggregation puts all modules into a graph neural network and uses an Entropy-Based Attention module to combine information and represent each node. (3) Loss and Training trains and updates parameters using gradient descent. (4) Anomaly detection finds anomalies by measuring the mean difference between predicted and expected values.
Entropy 28 00014 g003
Figure 4. The entropy-based centrality for each node v i within graph G, which is H ( v i ) = H ( G ) H ( G { v i } ) .
Figure 4. The entropy-based centrality for each node v i within graph G, which is H ( v i ) = H ( G ) H ( G { v i } ) .
Entropy 28 00014 g004
Figure 5. Ablation experiments for entropy centrality (P for Precision, R for Recall, and F1 for F1-Score).
Figure 5. Ablation experiments for entropy centrality (P for Precision, R for Recall, and F1 for F1-Score).
Entropy 28 00014 g005
Figure 6. ROC and AUC curves based on different proportions of anomalous data in the MSL dataset.
Figure 6. ROC and AUC curves based on different proportions of anomalous data in the MSL dataset.
Entropy 28 00014 g006
Table 1. The details of four publicly available multivariate time-series anomaly detection datasets (SMAP, MSL, SWaT, and WADI).
Table 1. The details of four publicly available multivariate time-series anomaly detection datasets (SMAP, MSL, SWaT, and WADI).
DatasetsTrainTestDimensionsAnomaly Rate
SMAP [32]135,183427,61725 (55)13.13%
MSL  [33]58,31773,72955 (3)10.72%
SWaT [34]496,800449,91951 (1)11.98%
WADI [35]1,048,571172,801123 (1)5.99%
Table 2. Hyperparameters (sliding window w and sliding step m) used in different datasets.
Table 2. Hyperparameters (sliding window w and sliding step m) used in different datasets.
DatasetsSliding Window wSliding Step m
SMAP12010
MSL9510
SWaT5010
WADI13510
Table 3. Quantitative results for the proposed method and baselines on four real-world datasets. A higher value indicates a better performance for these three metrics (P for Precision, R for Recall, and F1 for F1-Score). For the results of this work, the upper-right number indicates the ranking within the column.
Table 3. Quantitative results for the proposed method and baselines on four real-world datasets. A higher value indicates a better performance for these three metrics (P for Precision, R for Recall, and F1 for F1-Score). For the results of this work, the upper-right number indicates the ranking within the column.
MethodSMAPMSL
PRF1PRF1
IsolationForest [36]0.52390.59070.55530.53940.86540.6645
LSTM [37]0.89410.78130.83390.85450.82500.8359
LSTM-NDT [38]0.85230.73260.78790.62881.00000.7721
MAD-GAN [39]0.81570.92160.86540.85160.99300.9169
MTAD-GAT [40]0.79910.99910.88800.79170.98240.8768
DAGMM [41]0.86450.56730.68510.89600.63930.7462
GDN [42]0.74800.98910.85180.93080.98920.9591
OmniAnomaly [43]0.81300.94190.87280.78480.99240.8765
USAD [44]0.74800.96270.84190.87100.95360.9104
ATUAD [45]0.81570.99990.89850.90410.99990.9496
MSCRED [46]0.81750.92160.86640.89120.98620.9363
MSAD [47]0.85020.90660.87750.90650.96260.9337
This Work 0.8452   3 0.9790   3 0.9071   1 0.9182   2 0.9941   2 0.9546   2
MethodSWaTWADI
PRF1PRF1
IsolationForest [36]0.49290.44950.47020.18220.78240.2954
LSTM [37]0.86150.83270.84690.20630.76320.3358
LSTM-NDT [38]0.77780.51090.61670.21380.78230.3358
MAD-GAN [39]0.95930.69570.80650.22330.91240.3588
MTAD-GAT [40]0.97180.69570.81090.28180.80120.4169
DAGMM [41]0.89920.57840.70400.27600.99810.4324
GDN [42]0.96970.69570.81010.29120.79310.4260
OmniAnomaly [43]0.97820.69570.81310.31580.65410.4260
USAD [44]0.99770.68790.81430.18730.82960.3056
ATUAD [45]0.96960.69570.81010.30050.82860.4403
MSCRED [46]0.93180.98250.95650.25130.73190.3741
MSAD [47]0.95320.99430.97390.28320.83560.4230
This Work 0.9828   1 0.6563   6 0.7870   6 0.3029   2 0.8196   3 0.4423   1
Table 4. Ablation experiments for entropy centrality (P for Precision, R for Recall, and F1 for F1-Score).
Table 4. Ablation experiments for entropy centrality (P for Precision, R for Recall, and F1 for F1-Score).
MethodSMAPMSL
PRF1PRF1
Entropy Centrality0.84520.97900.90710.91820.99410.9546
Without Entropy Centrality0.73860.87750.80210.87250.91120.89143
MethodSWaTWADI
PRF1PRF1
Entropy Centrality0.98280.65630.78700.30290.81960.4423
Without Entropy Centrality0.9450.64890.76420.27810.82000.4152
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, B. Improving Multivariate Time-Series Anomaly Detection in Industrial Sensor Networks Using Entropy-Based Feature Aggregation. Entropy 2026, 28, 14. https://doi.org/10.3390/e28010014

AMA Style

Wang B. Improving Multivariate Time-Series Anomaly Detection in Industrial Sensor Networks Using Entropy-Based Feature Aggregation. Entropy. 2026; 28(1):14. https://doi.org/10.3390/e28010014

Chicago/Turabian Style

Wang, Bowen. 2026. "Improving Multivariate Time-Series Anomaly Detection in Industrial Sensor Networks Using Entropy-Based Feature Aggregation" Entropy 28, no. 1: 14. https://doi.org/10.3390/e28010014

APA Style

Wang, B. (2026). Improving Multivariate Time-Series Anomaly Detection in Industrial Sensor Networks Using Entropy-Based Feature Aggregation. Entropy, 28(1), 14. https://doi.org/10.3390/e28010014

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop