Next Article in Journal
A Degradation Warning Method for Ultra-High Voltage Energy Devices Based on Time-Frequency Feature Prediction
Previous Article in Journal
Classification Algorithms for Early Tooth Demineralization Assessment by Impedance Spectroscopy
Previous Article in Special Issue
Rolling Bearing Fault Diagnosis Based on a Synchrosqueezing Wavelet Transform and a Transfer Residual Convolutional Neural Network
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Data Quality Monitoring for the Hadron Calorimeters Using Transfer Learning for Anomaly Detection

by
Mulugeta Weldezgina Asres
1,*,
Christian Walter Omlin
1,*,
Long Wang
2,
David Yu
3,
Pavel Parygin
4,
Jay Dittmann
5 and
the CMS-HCAL Collaboration
1
Centre for Artificial Intelligence Research, Department of Information and Communication Technology, University of Agder, 4879 Grimstad, Norway
2
Department of Physics, University of Maryland, College Park, MD 20742, USA
3
Department of Physics, Brown University, Providence, RI 02912, USA
4
Department of Physics and Astronomy, University of Rochester, Rochester, NY 14627, USA
5
Department of Physics, Baylor University, Waco, TX 76706, USA
*
Authors to whom correspondence should be addressed.
The CMS-HCAL Collaboration author list is given at Supplementary Materials.
Sensors 2025, 25(11), 3475; https://doi.org/10.3390/s25113475
Submission received: 1 May 2025 / Revised: 26 May 2025 / Accepted: 28 May 2025 / Published: 31 May 2025
(This article belongs to the Special Issue AI-Assisted Condition Monitoring and Fault Diagnosis)

Abstract

:
The proliferation of sensors brings an immense volume of spatio-temporal (ST) data in many domains, including monitoring, diagnostics, and prognostics applications. Data curation is a time-consuming process for a large volume of data, making it challenging and expensive to deploy data analytics platforms in new environments. Transfer learning (TL) mechanisms promise to mitigate data sparsity and model complexity by utilizing pre-trained models for a new task. Despite the triumph of TL in fields like computer vision and natural language processing, efforts on complex ST models for anomaly detection (AD) applications are limited. In this study, we present the potential of TL within the context of high-dimensional ST AD with a hybrid autoencoder architecture, incorporating convolutional, graph, and recurrent neural networks. Motivated by the need for improved model accuracy and robustness, particularly in scenarios with limited training data on systems with thousands of sensors, this research investigates the transferability of models trained on different sections of the Hadron Calorimeter of the Compact Muon Solenoid experiment at CERN. The key contributions of the study include exploring TL’s potential and limitations within the context of encoder and decoder networks, revealing insights into model initialization and training configurations that enhance performance while substantially reducing trainable parameters and mitigating data contamination effects.

1. Introduction

Spatio-temporal (ST) anomaly detection (AD) a promising monitoring application of deep learning (DL) in several fields [1,2,3,4,5,6]. A unique quality of ST data is the presence of dependencies among measurements induced by the spatial and temporal attributes, where data correlations are more complex to capture using conventional techniques [1]. A spatio-temporal anomaly can thus be defined as a data point or cluster of data points that violate the nominal ST correlation structure of the normal data points. DL models dominate the recent AD studies, as AD models capture complex structures, extract end-to-end automatic features, and scale for large-volume data sets [5,6,7,8,9,10,11]. AD models can broadly be categorized as: (1) supervised methods requiring labeled anomaly observations [10,11] and (2) unsupervised approaches using unlabeled data, which are more pragmatic in many real-world applications, as data labeling is tedious and expensive [5,6,7,8,9]. Unsupervised AD models trained with only healthy observations are often called semi-supervised approaches [6]. Semi-supervised AD models have accomplished promising performance in reliability, safety, and health monitoring applications in several domains [6,7,8,9].
The deployment of ST DL models in a new environment is often circumscribed by the limited amount of clean data [12]. Data curation for DL modeling remains cumbersome and particularly challenging for temporal data despite abundant availability. Transfer learning (TL) mechanisms have been proposed for DL models to mitigate the challenge of data insufficiency; it accelerates model training and enhances accuracy [10,11,12,13,14,15,16,17,18,19,20,21,22,23]. It aims to achieve in-domain and cross-domain learning by extracting useful information from the model or data of the source task and transferring it to the target tasks [13,20,21,22,23]. TL is widely employed in computer vision (e.g., a large image classifier trained on over 1000 classes with ImageNet1K [24] is fine-tuned to classify a few types of fruit categories) [20] and natural language processing (e.g., BERT [25], initially trained on a massive and diverse text corpus to learn general language features like syntax and semantics, is fine-tuned with smaller task-specific data sets, specialized in question-answering tasks) [21]. It has also been proposed for temporal sensor data related to machine monitoring [14], electricity load monitoring [15], medical applications [16], dynamic systems [17], and ST data for crowd prediction [12,18,26,27], finance [28], environment monitoring [26], and fault diagnosis [11,19]. TL on ST data for AD application remains limited, and deeper investigation on autoencoder (AE) models, assessing both the encoder and decoder networks, is lacking [11,12,18,19,27,28].
Our study discusses ST AD modeling for the Compact Muon Solenoid (CMS) experiment at the Large Hadron Collider (LHC) [29,30]. The CMS experiment, one of the two high-luminosity general-purpose detectors at the LHC, consists of a tracker to reconstruct particle paths accurately, two calorimeters—the electromagnetic (ECAL) and the hadronic (HCAL)—to detect electrons, photons, and hadrons, and a muon system [30,31]. The CMS experiment employs the Data Quality Monitoring (DQM) system to guarantee high-quality physics data through online monitoring that provides live feedback during data acquisition, as well as offline monitoring that certifies the data quality after offline processing [32]. The online DQM identifies emerging problems using reference distributions and predefined tests to detect known failure modes using summary histograms, such as digi-occupancy maps of the calorimeters [33,34]. A digi-occupancy map contains the histogram record of particle hits of the data-taking channels of the calorimeters at the digitization level. The CMS calorimeters may encounter problems during data taking, such as issues with the front-end particle sensing scintillators, digitization and communication systems, back-end hardware, and algorithms [6,35]. These problems are usually reflected in the digi-occupancy maps. The growing complexity of detectors and the variety of physics experiments make data-driven AD systems essential tools for CMS to automate the detection, identification, and localization of detector anomalies [6,36,37,38]. Recent efforts in DQM at CMS have presented DL for AD applications [6,32,35,38,39,40,41]. The synergy in AD has thus far achieved promising results on spatial 2D histogram maps of the DQM for the ECAL [35,38], the muon detectors [40], and ST 3D maps of the HCAL [6].
Further study of TL for ST AD models—often involving combinations of spatial and temporal learning networks—is essential, considering the achievements of TL in other domains [21]. Recent ST DL models are hybrid and commonly made of combinations of variants of convolutional neural networks (CNNs), recurrent neural networks (RNNs), graph neural networks (GNNs), and transformers for various data mining tasks [12,18,27,42,43]. Our study investigates the potential strengths and limitations of TL on high-dimensional ST semi-supervised AD models. Although there are several ST AD architectures in the literature, most operate with 2D spatial data, such as images [10,12,18,26,27], and the ones that incorporate GNN deal with a limited number of nodes ranges from tens to a few hundreds [3,43]. Hence, we limit our discussion to the GraphSTAD system [6]—an AE model made of a CNN, RNN, and GNN operating on high-dimensional 3D spatial data—to deeply investigate the potential of TL for ST AD in the context of calorimeters for the CMS experiment. The GraphSTAD has been proposed for online DQM to automate monitoring of the thousands of HCAL channels through DL in Ref. [6]. The model captures abnormal events using spatial appearance and temporal context on digi-occupancy maps of the DQM. GraphSTAD employs CNNs to capture the behavior of adjacent channels exposed to regional collision particle hits, GNNs to learn local electrical and environmental characteristics due to a shared back-end circuit of the channels, and RNNs to detect temporal degradation on faulty channels [6]. We have transferred a pre-trained GraphSTAD model on the source HCAL Endcap (HE) subsystem into another target subsystem of the HCAL Barrel (HB) for the TL experiment. The HE and HB are subdetectors of the HCAL; they are designed to capture hadron particles at different positions of the calorimeter. The subdetectors share similarities but also have differences in design, technology, and configuration, such as detector segmentation [44].
Brute-forcing the knowledge from the source into the target, irrespective of their divergence, and thorough investigation of the several network-building modules would cause certain performance degeneration [20,22]. Hence, we have provided insights on TL using various training modes with different network hierarchies of the AE of the GraphSTAD system. The experiment has demonstrated the potential of TL when applied to the feature extraction encoder and the reconstruction decoder networks with different fine-tuning mechanisms on the target dataset. We have also examined the impact of TL with RNN state preservation within and across sliding time windows on ST reconstruction. TL has achieved promising ST reconstruction and AD while reducing the trainable parameters and providing better robustness against anomaly contamination in the training dataset. Our study demonstrates the efficacy of ST TL in overcoming training data sparsity and model training computation.
The key contributions of this study can be summarized as follows:
  • This study explores the potential and limitations of TL in the context of high-dimensional ST models for AD application, at scale with 3000–7000 spatial nodes.
  • This study, different from existing TL studies, assesses both encoder and decoder networks of a hybrid AE—evaluated on each main building block with various configurations. Related TL studies primarily focus on the feature extraction encoder or fine-tuning the entire network, as highlighted in this study. We present deeper insights and considerations on previously underexplored angles of TL.
  • We demonstrate the robustness of TL, with limited training data sets, in improving model accuracy, reducing training parameters, and better mitigation against vulnerability to training data contamination.
We discuss the related work in TL and the CMS DQM system in Section 2. We describe our datasets in Section 3 and the AD and TL methodologies in Section 4. Section 5 presents the performance evaluation and discussion of the results. We provide the conclusion and review the impact of our results in Section 6.

2. Background

This section discusses TL in DL models and provides an overview of the DQM system of the CMS experiment.

2.1. Transfer Learning on Deep Learning

In the last decade, the effectiveness of DL in handling large datasets has caught the attention of both academia and industry. Its ability to learn nonlinear behavior, along with end-to-end automatic feature extraction, allows it to find complex patterns within high-dimensional large data sets. However, most DL models are complex and require extensive data sizes for modeling, which can be expensive and time-consuming to curate, especially in the case of temporal data. Transfer learning approaches, which incorporate pre-trained models into new tasks, are potential solutions for developing DL models when clean data are limited [14,15,16,17,19,45]. TL is a paradigm where knowledge from a source model or data on different domains (e.g., different data sources or datasets) or tasks (e.g., different model applications) is utilized to improve the efficacy of a target model [13,20,22,23,45].
The TL techniques in the literature can broadly be categorized into various taxonomies [13,20,21]. One of the typical categorizations is based on the similarity of the task and domain between the source and target [13,20,21,22]: (1) inductive TL: the source and target tasks are different, but their domains may remain the same; (2) transductive TL: the tasks remain the same, but the domains are different; and (3) unsupervised TL: similar to inductive transferring on different but related tasks with unlabeled datasets. TL can be carried out on (1) model parameters, where all or some parameters are transferred from a pre-trained source model, and (2) data, where all or part of the source domain data instances are utilized to train the target model [13,21,22]. In this study, TL signifies the use of learned network parameters (weights and biases from a source model pre-trained on adequate datasets) on a target model for a related task on a different dataset, with or without fine-tuning of the parameters [13]. We refer readers to a survey study in refs. [13,21,22] for further discussion and progress on recent deep TL approaches.
The recent successes of generative models on image and text data have ameliorated the adoption of TL methods for several applications [13,20,21]. The notable contribution of TL is significant in transferring feature extraction networks (encoders) that are trained on immense datasets with very expensive computation grids [26,27]. Robustly extracted features reduce the model complexity and training cost of the fine-tuned decision networks while enhancing accuracy [14,16,18,21,26,27]. We refer to this TL mechanism as the freeze and fine-tune approach [13]. Although abundant studies are available for images and language text, TL is relatively less explored for temporal data, such as sensor measurement datasets [19]; TS datasets are often not readily available or accessible on the internet, unlike images and text, and the datasets are often multidimensional and so diverse that they require domain-specific knowledge for data curation and preparation. TL on a temporal data has been investigated in various applications [11,14,15,16,17]. The efforts towards adopting TL for ST data are even more limited [12,18,19,26,27,28]. Hijazi et al. [19] proposed a TL approach that integrates CNN and temporal long short-term memory (LSTM) (referred to as ConvLSTM) to efficiently train a new stability prediction model when the power system undergoes topological changes. Wang et al. [18] applied TL for cross-city crowd-flow prediction where feature extraction ConvLSTM of the forecasting model trained on one city is fine-tuned on another city’s dataset. Wang et al. [12] extended TL on ConvLSTM using a deep adaptation mechanism for crowd-flow prediction. The adaptation network matches the embedding representations of the source and target domain distributions to learn the transferable features between the two domains. Guo et al. [28] fine-tuned an autoencoder for a store recommendation system from a model trained on a different city dataset. Sarker et al. [26] and Natha et al. [27] adopted pretrained 3D CNNs for ST feature extraction to improve anomaly detection on video datasets. Yang et al. [11] proposed unsupervised TL utilizing the fault knowledge learned from labeled sensor fault datasets to perform online anomaly monitoring on unmanned aerial vehicle sensor data. Some studies have employed TL to increase training data from multiple sources and address training with diverse data issues, such as catastrophic forgetting, using different regularization techniques [45]. Recent DL models built on hybrids of CNNs [2,6,18,19,26,27,46,47], RNNs [6,18,19,27,45,48,49], and GNNs [3,6,43,49] have gained momentum for TS and ST data in AD and other data mining applications. Thus far, most TL studies have focused on feature extraction encoding networks and predominantly on forecasting tasks [12,18,19,45]. We have studied the transferability of CNNs, GNNs, and RNNs on both the encoder and decoder networks of an autoencoder and qualitatively evaluated the effectiveness of the TL on ST reconstruction and AD tasks.

2.2. The Hadron Calorimeter of the CMS Detector

Figure 1a shows the CMS experiment and the HCAL detector inside CMS [30,31]. The calorimeters of the CMS detector are highly segmented to improve the accuracy of energy-deposition profile-measurement and particle identification [30,31,50]. The segmentation geometry of the detector is represented using η and ϕ spaces, which correspond to pseudo-rapidity and azimuth, respectively (as shown in Figure 1b). The z-axis lies along the incident beam direction, ϕ is the azimuthal angle between the x and y axis, and η is calculated from the polar angle θ c m between the z and x y -planes as follows:
η = ln ( tan ( θ c m / 2 ) )
where x, y, and z are orthogonal axes of the cylinder, θ c m is the center-of-mass scattering angle, and ln is a natural log function. The η ϕ space corresponds to a rectangular coordinate system representing an outgoing particle’s direction from the center of the detector (where the collision occurs). Particles traveling in the same direction lie near each other in η ϕ space.
Figure 2a illustrates the four major subdetectors of the HCAL covering different segments in the CMS detector: the HB, the HE, the HCAL Outer (HO), and the HCAL Forward (HF). Since this study’s datasets are from the LHC Run-2 collision experiment, we will describe the HCAL system configurations from 2018 below. The HB and HE are sampling calorimeters with a brass absorber and active plastic scintillators to measure the energy depositions [31]. The subdetectors surround the ECAL and are fully immersed within the strong magnetic field of the solenoid: the HB are joined hermetically with the barrel extending out to η = 1.4 and the HE covering the overlapping range 1.3 < η < 3.0 (as shown in Figure 2b). The HF is located 11.2 m from the interaction point and extends the pseudo-rapidity coverage (overlapping with the HE) from η = 2.9 to η = 5 . The central shower containment in the region η < 1.26 is improved with the HO, an array of scintillators located outside the magnet.
The front-end electronics of the HCAL, responsible for sensing and digitizing optical signals of the collision particles, are divided into sectors of readout boxes (RBXes) that house the electronics and provide voltage, backplane communications, and cooling. Our study’s use cases, the HE and HB, consist of 36 RBXes arranged on the plus (HE[HB]P) and minus (HE[HB]M) hemispheres of the CMS detector. The front-end acquisition systems transmit the photons produced in the plastic scintillators through the wavelength-shifting fibers to the silicon photomultipliers (SiPMs) [HE] or the hybrid photodiode transducers (HPD) [HB] [31]. Each RBX houses frontend electronics that include four digitization readout modules (RMs), the next-generation clock and control module, and the calibration unit [31]. Each RM is made of SiPMs [HE] or HPD [HB], a SiPM control card, and four readout charge integrator and encoder (QIE) cards, each with several QIE chips and field-programmable gate array (FPGA) modules. A QIE chip integrates charge from one SiPM [HE] or HPD [HB] at 40 MHz, and the FPGA serializes and encodes the data from the QIE chips (channels).

2.3. CMS Data Quality Monitoring

The collision data of the LHC are organized into runs, where each run contains thousands of luminosity sections (also called lumisections). A lumisection (LS) corresponds to approximately 23 s of data collection and comprises hundreds or thousands of collision events containing particle hit records across the CMS detector. The DQM system in CMS provides feedback on detector performance and data reconstruction; it generates a list of certified runs for physics analyses and stores it in the “Golden JSON” [32]. The DQM employs online and offline monitoring mechanisms: (1) online monitoring is real-time DQM during data acquisition, and (2) offline monitoring provides the final fine-grained data quality analysis for data certification 48 h after the collisions were recorded. The online DQM populates a set of histogram-based maps on a selection of events and provides summary plots with alarms that DQM experts inspect to spot problems. The digi-occupancy map is one of the histogram maps generated by the online DQM, and it contains particle hit histogram records of the particle readout channel sensor of the calorimeters. A digi, also called a hit, is a reconstructed and calibrated collision physics signal of the calorimeter. Several errors can arise in the calorimeter affecting the front-end particle sensing scintillators, the digitization and communication systems, the back-end hardware, or the algorithms. These errors appear in the digi-occupancy map as holes, under- or over-populated bins, or saturated bins. Previous efforts by the authors of refs. [6,32,35,39,40] demonstrate the promising AD efficacy of using digi-occupancy maps for calorimeter channel monitoring using machine learning. Our GraphSTAD has extended the efforts in AD for the HCAL with ST modeling of the 3D digi-occupancy maps of the DQM [6]. The GraphSTAD incorporates both CNNs and GNNs to capture Euclidean and non-Euclidean spatial characteristics, respectively, as well as RNNs for temporal learning for the HCAL channels.

3. Dataset Description

We utilized the digi-occupancy data of the online DQM system of the CMS experiment to train and validate our models. The data contain healthy digi-occupancy maps with a 20 fC minimum threshold and were selected from certified good collision runs, as referred to by the “Golden JSON” of CMS. The digi-occupancy datasets were collected in 2018 during the LHC Run-2 collision experiment with a received luminosity per lumisection of up to 0.4 pb−1 and up to 2250 events. The source and target datasets contain three-dimensional digi-occupancy maps for the HE and HB subsystems of the HCAL, respectively (as shown in Figure 3).
The digi-occupancy map contains a particle hit count of the calorimeter readout channels for a given period of time. The HCAL covers a considerable volume of CMS and has a fine segmentation along three axes ( i η [ 32 , , 32 ] , i ϕ [ 1 , , 72 ] and depth  [ 1 , , 7 ] ). The i η and i ϕ denote integer notation of the towers covering ranges of η and ϕ of the CMS detector, respectively [31]. The digi-occupancy measurement corresponds to a hit record of the readout channels at the segmentation positions. The similarities between the source and target datasets and tasks have been established in the literature to be essential factors that impact the performance of TL [54]. The source system HE (as shown in Figure 3b) and the target system HB (as shown in Figure 3c) share a similar task but cover different segments of the HCAL. Another major difference between the HE and HB in the 2018 LHC collision run is the front-end data acquisition optical-to-electrical technology, i.e., the HE was upgraded to SiPMs with QIE11 technology, and the HB utilized HPD with QIE8. We compare the source and target datasets in Table 1.

4. Methodology

This section presents the GraphSTAD modeling and the experimental setups for the transfer learning study.

4.1. Data Preprocessing

This section describes the data preprocessing stages of the proposed approach, i.e., digi-occupancy renormalization and graph-adjacency matrix generation.

4.1.1. Digi-Occupancy Map Renormalization

We apply digi-occupancy renormalization in the data preprocessing stages to normalize the values for the variation in the luminosity and the number of event configurations in the collision experiments [6]. The digi-occupancy ( γ ) map data of the HCAL vary with the received luminosity ( β ) and the number of events ( ξ ) (as shown in Figure 4). The per-channel γ s ( i ) can range γ s ( i ) = [ 0 , ξ s ] , where s denotes the s th LS (3D map) in the dataset and i denotes the i th channel in the s th map. ξ s is usually adjusted with β s , but not always. β and γ are retrieved from different systems on the existing CMS system; directly accessing β for real-time γ AD monitoring requires further effort. We renormalize the maps ( γ s γ ^ s ) based on ξ s to obtain consistent interpretation of the γ maps across lumisections:
γ ^ s = γ s ξ s
The renormalization of γ with only ξ does not entirely avoid the data distribution variations across collision runs, and the distribution shifts and unpredictable spikes due to β s may affect the AD model training performance on the ST data (see Figure 4). We employ additional reversible renormalization (RN) before and after invoking the AD model to mitigate the non-linearity of the γ ST data. The renormalization exploits the symmetric property of the i ϕ axis (the γ channels are less diverse along the i ϕ axis); it divides the γ channels per each i η and d e p t h coordinate by the median values along the i ϕ axis on the model input and reverses the action on the model output. The remaining impact of β s is left to be learned by the AD model from the training data.

4.1.2. Adjacency Matrix Generation

We deployed an undirected graph network G ( V , Θ ) to represent the HCAL channels in a graph network based on their connections to a shared RBX system. The graph G contains nodes υ V , with edges ( υ i , υ j ) Θ in a binary adjacency matrix A R M × M , where M is the number of nodes (the channels). An edge indicates the channels sharing the same RBX as follows:
A ( υ i , υ j ) = 1 , if Ω ( υ i ) = Ω ( υ j ) 0 , otherwise
where Ω ( υ ) returns the RBX identification of the channel υ . There are approximately 7000 channels for the HE and 2600 for the HB in a graph representation of the digi-occupancy map. We retrieved the channel-to-RBX mapping from the 2018 HCAL’s calorimeter segmentation map.

4.2. Anomaly Detection Mechanism

We denote the AE model of the AD system as F . It takes ST data X R T × N i η × N i ϕ × N d × N f as a sequence in a time window t x [ t T , t ] , where N i η × N i ϕ × N d is the spatial dimension corresponding to the i η , i ϕ , and d e p t h axes, respectively, and N f is the number of input variables ( N f = 1 , as we monitored only a digi-occupancy quantity in the spatial data). The F θ , ω : X X ¯ , parametrized by θ and ω , attempts to reconstruct the input ST data X and outputs X ¯ . The encoder network of the model E θ : X Z provides low-dimension latent space, Z = E θ ( X ) , and the decoder D ω : Z X ¯ reconstructs the ST data from Z , X ¯ = D ω ( Z ) as follows:
X ¯ = F θ , ω ( X ) = D ω ( E θ ( X ) )
Anomalies can live for a short time on a single digi-occupancy map, or they can persist over time, affecting a sequence of maps. Aggregated spatial reconstruction error is calculated over a time window T using mean absolute error (MAE) to capture a time-persistent anomaly as follows:
e i , M A E = 1 T t = t T t | x i ( t ) x ¯ i ( t ) |
where x i X and x ¯ i X ¯ are the input and the reconstructed digi-occupancy of the i th channel. We standardized e i , M A E to homogenize the reconstruction accuracy variations among the channels when generating the anomaly score a i as follows:
a i = e i , M A E σ i
where σ i is the standard deviation of the e i , M A E on the training dataset. The standardized anomaly score allows us to use a single AD decision threshold α for all the channels in the spatial map. The anomaly flags are generated after applying α to the anomaly scores ( a i > α ). The α value can be tuned to control the detection sensitivity.
The use-case GraphSTAD AE model is made of CNN, GNN, and RNN networks; it employs a CNN and GNN with a pooling mechanism to extract relevant features from spatial DQM data followed by RNN to capture the temporal characteristics of the extracted features (see Figure 5). It integrates a variational layer [55] at the end of the encoder for regularization of overfitting by enforcing continuous and normally distributed latent representations [6,36]. We refer readers to ref. [6] for further discussion of the mathematical formulation and architecture of the GraphSTAD model.
We trained the AE on healthy digi-occupancy maps (without significant anomaly contamination, see Section 3) of the target HB system. We normalized the spatial data per channel into a [ 0 , 1 ] range to train the model across the variations in calorimeter channels effectively. We utilized a mean squared error (MSE) loss function as follows:
L M S E = 1 M i ( x i x ¯ i ) 2
where x i and x ¯ i are the input and the reconstructed values of the normalized γ ^ of the i th channel, respectively, and M is the total number of channels. The variational layer of the AE (denoted as VAE in Figure 5) regularizes the training MSE loss using the Kullback–Leibler divergence (KL) distance D K L [55] to achieve the normally distributed latent space as follows:
L = argmin W R L M S E λ D K L N ( μ z , σ z ) , N ( 0 , I ) + ρ W 2 2
where N is a normal distribution with zero mean and unit variance, and . 2 2 is a squared Frobenius norm of L 2  regularization for the trainable model parameters W [56]. λ = 0.003 and ρ = 10 7 are tunable regularization hyperparameters. We employed the Adam optimizer [57] for training.

4.3. Transfer Learning Approach

Model parameter TL generally consists of four basic steps: (1) selection of a source task with a related modeling problem and an abundance of data where we can exploit the mapping knowledge from the inputs to outputs, (2) development of the source model that performs well in the source task, (3) transfer source model to target model where whole or part of the source model is employed as part of the target model, and (4) fine-tuning the target model on the target dataset if necessary. We present knowledge transfer on GraphSTAD AE models, i.e., an AD model trained on digi-occupancy maps of the source HE subsystem is transferred to the target HB subsystem. Direct transfer of knowledge from the source into the target, irrespective of their divergence and thorough investigation of the model network layers, would limit the efficacy of TL in the target domain [13,20,23].
We have thus investigated several transferring cases when employing TL in two principal model training phases: initialization and training (see Figure 6).
  • Init mode ( T init ): the trainable network parameters (weights and bias) of the source model are transferred into the target model initialization. The target model is further trained on the target HB dataset, resulting in fine-tuning.
  • Train mode ( T train ): The model parameters of the source model are directly reused as the final inference parameters of the target model; the parameters are frozen and excluded from fine-tuning on the target HB dataset.
Let M ( Ψ , Ω ) be an AD model with parameters Ψ and Ω that represent the model networks that can be affected and not affected by TL, respectively. M e ( Ψ e , Ω e ) and M b ( Ψ b , Ω b ) are the source and target models for the HE and HB, respectively. The TL modes of T can be formulated mathematically as follows:
T init : M b ( Ψ e , Ω b ) fine - tuning M b ( Ψ e , Ω b ) T train : M b ( Ψ e , Ω b ) fine - tuning M b ( Ψ e , Ω b )
where the superscript ′ denotes the parameters that are updated after fine-tuning the M b model on the target dataset.
F θ , ω of the GraphSTAD is made of CNNs and GNNs with a pooling mechanism to extract relevant features from high-dimensional spatial data, followed by RNNs to capture the temporal characteristics of the extracted features (as shown in Figure 5). Table 2 presents the TL mechanisms that we apply to the different deep networks of the encoder and decoder to study the impacts on ST digi-occupancy map reconstruction and AD accuracy. We also analyze the effects of RNN state preservation within and across time windows. We further investigate variations in training iterations and learning rate scheduling methods. The discussion includes the impact of the TL on reconstruction and AD accuracy, saturation, training stability, and the reduction in the number of model trainable parameters.
The implementation of parameter transferring on DL networks can be accomplished in two ways: (1) start with the source model and then reset (remove and add) the networks that are not included in the TL, and (2) start with the target model with random initialization and update the parameter values of the networks included in the TL from the corresponding source networks. The first approach is widely utilized in DL TL literature and employed for feature extraction; however, it may not be suitable for flexibly choosing layers at different hierarchies, as the target models might have slight variations. Several configuration setups of the AE are derived from the spatial configuration of the input 3D map, which differs for the source HE and target HB systems—e.g., variation in the depth spatial dimension between HE and HB. We have found the second approach more convenient for our study, as we intend to apply TL on different networks of the encoder and decoder of the AE model.

5. Results and Discussion

This section will discuss the results of TL on different network layers of the GraphSTAD AE models. We will investigate the effects of TL on reconstruction accuracy, trainable parameter reduction, and AD performance on the target HB digi-occupancy map dataset. We applied TL for model initialization ( T init ) and training or fine-tuning ( T train ) on the target HB dataset. We trained the models on NVIDIA Tesla V100 with 4 GPUs using 4000 digi-occupancy maps from LS 1 to 500 and evaluated them on a test set that contains approximately 3000 maps from LS from 500 to 1500. We utilized 20% of the training dataset for validation loss calculation during training to determine the best states for the models. We set the LR at 10 3 to train the models with five LSs per time window.

5.1. Spatio-Temporal Reconstruction Performance

We will discuss below the reconstruction performance (using L M S E ) of TL applied on spatial (CNNs and GNNs) and temporal (RNNs) learning networks. We will also briefly present comparison results for the LR scheduling choice.

5.1.1. Transfer Learning on Spatial Learning Networks

We have assessed the transferability of the DL AD model at the initialization and inference phases for the spatial learning networks (CNNs and GNNs) on both the encoder and decoder networks on different numbers of training epochs (see Figure 7). The TL has reduced the reconstruction error L M S E of healthy maps by 32.5% to 20.7% when the number of epochs is varied from 75 to 200 (as shown in Figure 7b). The minimum gain of 13% is achieved at epoch 150, just before the performance of the no-TL model starts to saturate. The complete fine-tuning—TL for initialization, followed by fine-tuning the whole network—provided around 20% improvement. The L M S E generally decreases, while the relative TL gain roughly decreases as the epoch increases to 150. The results are not entirely unexpected; the DL models may improve performance as the training epoch increases, reducing the gap caused by the difference in the initialization and training mechanisms. When the epoch increased beyond 150, the randomly initialized model using no-TL achieves only slight improvement, whereas the L M S E continues to drop for the TL models, increasing the relative gain of the TL. The initialization TL on all the spatial learning networks of the AE using T init = TL-4 and training only the decoder while freezing the encoder using T train = TL-3 achieves the best improvement, from 26% to 32.5%. The TL gain of the GNNs is limited compared to CNNs; the CNNs are the primary networks that learn the input spatial data, and they have 15 times more parameters than the GNNs in the use-case GraphSTAD AE model. Transferring and freezing the CNNs of the encoder (TL-2 and TL-3) results in stable performance on repeated experiments. Although TL-2 outperforms TL-3 slightly (by 3 % ) at epoch 200, TL-3 provides computational leverage ( 7.2 % , see Section 5.1.2), bypassing the training overhead of the GNNs.
Table 3 further provides the average and best ST reconstruction performance at e p o c h = 200 . Inference TL on the decoder networks without fine-tuning T train = TL-2d fails to reconstruct the target data adequately. In an AE architecture, the encoder maps the input into low-dimensional latent space (information compression), while the decoder attempts to reconstruct (information expansion) the target data from the latent space. The decoder networks thus require fine-tuning on the target dataset to adjust their parameters to the target reconstruction effectively. Boulle et al. [17] investigated TL and DL for a univariate chaotic time series classification model; they argued that BN without fine-tuning limits the transferability of CNNs. The scaling and shifting parameters for BN and bias parameters are estimated from the training dataset and strongly correlate to the data. We further studied TL on the decoder when the BN layer and the bias parameters of the CNNs are fine-tuned on the target dataset. L M S E is substantially improved by 50% compared to the frozen decoder (see Table 4). However, the error is still 20 times higher than that without TL, indicating that the CNNs of the decoder also require fine-tuning to achieve reasonable accuracy. The results demonstrate the promising leverage of TL for AE model initialization on both feature extraction encoder and reconstruction decoder networks, whereas fine-tuning with the target dataset is essential for the decoder networks.

5.1.2. Transfer Learning on Spatio-Temporal Learning Networks

We investigated TL on the temporal RNNs (LSTM layers) in both the encoder and decoder networks, along with the spatial learning networks (CNNs and GNNs), using T init = TL-4 and T train = TL-3—the best performing TL for spatial networks across epochs (see Figure 7).
Table 5 presents L M S E when TL is applied to the ST networks. We evaluated the models by preserving the RNN states across time windows that leverage the accuracy. When the TL involves freezing the RNNs of the decoder ( T train = TL-6), the L M S E improves by 22.6 32.6 % while considerably reducing the model trainable parameters by 97.77 % , mainly due to the frozen LSTM networks (see Table 5).
However, the performance of T train = TL-6 suffers substantially, increasing L M S E by more than 50% if the state memory of the RNNs is not preserved across the sliding time window, i.e., memory reset at every non-overlapping sliding time window start (as shown in Figure 8). Figure 8 presents the L M S E values on multiple epochs when the TL includes the RNNs with and without state preservation across time windows. The plots show a significant enhancement by preserving the states on the frozen decoder RNNs using T train = TL-6 but a limited impact when the target dataset fine-tunes the decoder RNNs using T train = TL-5. Figure 9a demonstrates that the TL-6 model struggles to reconstruct the map at the first time step in each sliding time window when states are not preserved across time windows. This is caused by the model’s reliance solely on the input map for the first time-step reconstruction with reset memory states (zeros), while the states are adjusted and improved for the subsequent maps. The reconstruction improves when utilizing previous states, even for the first maps in the time windows (see Figure 9b).

5.1.3. Applying Learning Rate Scheduling

The AE L M S E reaches saturation after e p o c h > 150 when trained without TL (as illustrated in Figure 7a). Learning rate (LR) scheduling mechanisms, e.g., lowering the LR when the loss flattens, or fast convergence methods, could mitigate training stagnation. We have investigated the impact of scheduling on the TL by training the model with super-convergence one-cyclic LR scheduling [59]. The LR scheduling sets the LR according to a one-cycle policy that anneals the LR from an initial LR ( i n i t _ l r = 4 × 10 5 ) to a maximum LR ( m a x _ l r = 10 3 ) and then from that maximum LR to a minimum LR ( m i n _ l r = 4 × 10 7 ). We utilize a cosine annealing mechanism along with the other settings of the scheduler, such as d i v _ f a c t o r = 25 and f i n a l _ d i v _ f a c t o r = 100 , where d i v _ f a c t o r determines the initial LR by dividing m a x _ l r , and f i n a l _ d i v _ f a c t o r estimates m i n _ l r by dividing the initial LR. We have kept the default values of the remaining hyperparameters given in the PyTorch = 1.12.0 implementation [59].
Table 6 shows that the LR scheduling has improved the L M S E compared to the fixed LR (provided in Table 5) by 19% for without TL, and 7.1% and 4.3 % for TL with T train = TL-5, and T train = TL-6, respectively. The relative progress of the TL is approximately 9% with the LR scheduling, which is lower than the 22.6% achieved with the fixed LR. The results are consistent with Figure 7b, showing a narrowing of the performance difference as the number of epochs increases past e p o c h > 150 , with performance saturating for the model without TL, T init = RANDOM. The cyclic LR scheduling method may require more configuration tuning effort to improve the performance compared to fixed LR or other simpler LR scheduling approaches.

5.2. Anomaly Detection Performance

Machine learning studies performed thus far in the CMS DQM system have primarily employed simulated anomaly data to evaluate the efficacy of the developed AD models [6,35]; a small fraction of the DQM data is affected by real anomalies that are inadequate for comprehensive model validation. We validate the AD models on synthetic anomalies simulating real channel anomalies of the HCAL [6]. We have generated synthetic anomalies simulating dead, hot, and degraded channels and injected them into healthy digi-occupancy maps of the test dataset. We formulate the simulated channel anomalies as follows:
γ a = R D γ h , where R D 1 and γ a ξ γ a = ξ , where R D = 1 and γ h < ξ
where γ a [ 0 , ξ ] and γ h ( 0 , ξ ] are the digi-occupancy of the generated anomaly channel and its corresponding expected healthy reading, respectively. The R D is the degradation factor, and the channel anomalies are defined as follows:
Dead : γ a = 0 , using R D = 0 Degraded : 0 < γ a < γ h , using 0 < R D < 1 Noisy hot : γ h < γ a ξ , using R D > 1 Fully hot : γ h < γ a = ξ , using R D = 1
The algorithm that generates the synthetic anomaly samples involves three steps [6]: (1) selection of a random set of LSs from the test set, (2) random selection of spatial locations φ for each LS, where φ [ i η × i ϕ  × depth] on the HB axes (see Figure 3c), and (3) injection of the simulated anomalies into digi-occupancy maps of the LSs. The simulated anomalies include dead, degraded, noisy hot, and fully hot channels. For consistency, we have kept the same spatial locations for all the anomaly types. We have evaluated the performance on several classification metrics using three anomaly thresholds set to capture 90%, 95%, and 99% of the injected anomalies.
We evaluate the AD accuracy on 14,000 digi-occupancy maps (2000 maps for each anomaly type) for the dead ( R D = 0 % ), decaying anomalies ( R D = [ 80 % , 60 % , 40 % , 20 % ] ), noisy hot ( R D = 200 % ), and fully-hot ( γ a = ξ ) channels. We investigate persistent channel anomalies that affect consecutive maps in a time window. We have processed 70,000 digi-occupancy maps (with the generated 1.17% abnormal channels) that include five history maps in the time window for each of the 14,000 maps.
We compare the AD performance of models without TL and the best TL from Table 6. The models are denoted as follows:
M N : T init = RANDOM and T train = No - TL M T : T init = TL - 7 and T train = TL - 6
where M N and M T are the models trained without TL and with TL, respectively.
Table 7 presents the AD accuracy of the models on the dead, degrading, fully hot, and noisy hot channel abnormalities. Both models perform well in the area under the receiver operating characteristic curve (AUC) and false positive rate (FPR). The TL model significantly improves dead and fully hot channel detection but performs slightly lower for the noisy hot channels. Figure 10 demonstrate the ability of the models to detect and localize the different anomaly types that have been injected at sample channels located at { 4 < i η < 11 , 11 < i ϕ < 19  depth}; the M T model accomplishes better detection on the fully hot channels with less dispersion in its anomaly score values.
Figure 11 portrays the distribution of the AD reconstruction error score ( e i , M A E ) and the overlap region between the healthy and faulty channels at the different degradation rates. We observed an increase in the error score of the healthy channels for the noisy hot channel anomalies at R D = 200 % . Close investigation reveals that the healthy channels have higher anomaly scores due to their proximity to the abnormal channels (as illustrated in Figure 12); the channels are filtered out from the anomaly simulation generation in Equation (11) due to γ a = R D γ h > ξ . Since channels positioned in proximity to the HB segmentation may share common RBX and are exposed to similar collision particles, the AD AE exploits the correlation for spatial data reconstruction. Figure 13 elaborates further proof of the proximity, showing that the false positive healthy channels with higher anomaly scores ( e i , M A E > 0.2 ) belong to R D γ h > ξ .
The degrading and dead channels are another major difference between the M N and M T models. Table 7 shows that the AD slightly deteriorates for the dead channels ( R D = 0 % ) compared to the degraded channel at R D = 20 % , defying the expectation of getting better AD on stronger anomalies. The error score of the M N model drops to zero for dead channels, although the channels have higher error scores at R D = 60 % (see Figure 11a); this is influenced by the presence of real dead channels in the training dataset at the location of { i η [ 16 , 15 , 13 ] ,   i ϕ = 8 ,  depth  = 1 } (see Figure 14). Figure 14 depicts that the M N model learned to reconstruct the real dead channels as healthy (very low error score), whereas M T provides a high error, signifying it is detecting the channels as anomalies. The results demonstrate TL robustness when a semi-supervised model’s training dataset is contaminated with real anomalies.

6. Conclusions

We have presented transfer learning on anomaly detection models in the context of high-dimensional spatio-temporal autoencoders. We have discussed TL using semi-supervised AD models designed to monitor the Hadron Calorimeter using three-dimensional digi-occupancy maps of the Data Quality Monitoring system. We have successfully transferred the AD model, employing convolutional, graph, and recurrent neural networks, from the source HCAL Endcap to the target HCAL Barrel calorimeter using several ST TL configurations. This study has provided insights into several TL scenarios at the model initialization and training phases in both the encoder and decoder networks. Applying TL to the feature extraction networks of the encoder and inner reconstruction networks of the decoder provided promising results in the ST AD. The approach has also demonstrated potential leverage for small training datasets, a significant reduction in training computation, and an enhancement in robustness against data contamination in the training dataset. In addition to the similarity between the source and target datasets, the choice of model settings, such as the target network layers, the number of training iterations, the learning rate schedule, and the temporal state preservation during inference, can influence the performance of TL. Our study remains relevant to applications in other domains, as it provides essential understanding of TL on widely utilized hybrid neural network layers applied to ST AD datasets.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/s25113475/s1, Membership of the CMS-HCAL Collaboration.

Author Contributions

Conceptualization, M.W.A., L.W., and D.Y.; data curation, M.W.A., L.W., and D.Y.; formal analysis, M.W.A.; investigation, M.W.A., L.W., P.P., and J.D.; methodology, M.W.A. and C.W.O.; resources and collision experiment, L.W., D.Y., P.P., J.D., and the CMS-HCAL Collaboration; software, M.W.A.; supervision, C.W.O.; validation, M.W.A., C.W.O., L.W., and D.Y.; visualization, M.W.A.; writing—original draft, M.W.A.; writing—review and editing, M.W.A., C.W.O., L.W., P.P., J.D, and the CMS-HCAL Collaboration. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Acknowledgments

We sincerely appreciate the CMS collaboration—specifically the HCAL data performance group, the HCAL operation group, the CMS data quality monitoring groups, and the CMS machine learning core teams. Their technical expertise, diligent follow-up on our work, and thorough manuscript review have been invaluable. We also thank the collaborators for building and maintaining the detector systems used in our study. We extend our appreciation to CERN for the operations of the LHC accelerator. The teams at CERN have also received support from the Belgian Fonds de la Recherche Scientifique and Fonds voor Wetenschappelijk Onderzoek; the Brazilian Funding Agencies (CNPq, CAPES, FAPERJ, FAPERGS, and FAPESP); SRNSF (Georgia); the Bundesministerium für Bildung und Forschung, the Deutsche Forschungsgemeinschaft (DFG), under Germany’s Excellence Strategy—EXC 2121 ”Quantum Universe”—390833306, and under project number 400140256—GRK2497, and Helmholtz-Gemeinschaft Deutscher Forschungszentren, Germany; the National Research, Development and Innovation Office (NKFIH) (Hungary) under project numbers K 128713, K 143460, and TKP2021-NKTA-64; the Department of Atomic Energy and the Department of Science and Technology, India; the Ministry of Science, ICT and Future Planning, and National Research Foundation (NRF), Republic of Korea; the Lithuanian Academy of Sciences; the Scientific and Technical Research Council of Turkey, and Turkish Energy, Nuclear and Mineral Research Agency; the National Academy of Sciences of Ukraine; the US Department of Energy.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AD
Anomaly Detection
AE
Autoencoder
AUC
Area Under the Curve
CMS
Compact Muon Solenoid
CNN
Convolutional Neural Network
DL
Deep Learning
DQM
Data Quality Monitoring
ECAL
Electromagnetic Calorimeter
FC
Fully Connected Neural Network
FPR
False Positive Rate
GNN
Graph Neural Network
GraphSTAD
Graph-based ST AD model
HCAL
Hadron Calorimeter
HB, HE, HF, HO
HCAL Barrel, HCAL Endcap, HCAL Forward, and HCAL Outer subdetectors
HPD
Hybrid Photodiode Transducers
KL
Kullback–Leibler divergence
LHC
Large Hadron Collider
LR
Learning Rate
LS
Luminosity Section
LSTM
Long Short-Term Memory
MAE, MSE
Mean Absolute Error, Mean Squared Error
QIE
Charge Integrating and Encoding
RBX
Readout Box
RNN
Recurrent Neural Network
SiPM
Silicon Photo Multipliers
ST
Spatio-Temporal
TL
Transfer Learning
γ , γ ^ , γ h , γ a
Digi-occupancy map, renomalized γ by ξ , healthy γ , anomalous γ
β , ξ
Received luminosity, number of collision events
i η , i ϕ , d e p t h
i e t a , i p h i , and d e p t h axes of the HCAL channels
F θ , ω , E θ , D ω
Mathematical notation of an AE model, encoder of F θ , ω , decoder of F θ , ω
L M S E , e i , M A E
AE reconstruction MSE loss score, MAE AD score
R D
Degradation factor of channel anomaly
M e , M b
AE model of the TL source HE system, AE model of the TL target HB system
M N , M T
Model trained without TL, model trained with TL
T init , T train
TL during initialization phase, TL during training phase (frozen parameters)

References

  1. Atluri, G.; Karpatne, A.; Kumar, V. Spatio-temporal data mining: A survey of problems and methods. ACM Comput. Surv. 2018, 51, 83. [Google Scholar] [CrossRef]
  2. Chang, Y.; Tu, Z.; Xie, W.; Luo, B.; Zhang, S.; Sui, H.; Yuan, J. Video anomaly detection with spatio-temporal dissociation. Pattern Recognit. 2022, 122, 108213. [Google Scholar] [CrossRef]
  3. Deng, L.; Lian, D.; Huang, Z.; Chen, E. Graph convolutional adversarial networks for spatiotemporal anomaly detection. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 2416–2428. [Google Scholar] [CrossRef]
  4. Tišljarić, L.; Fernandes, S.; Carić, T.; Gama, J. Spatiotemporal road traffic anomaly detection: A tensor-based approach. Appl. Sci. 2021, 11, 12017. [Google Scholar] [CrossRef]
  5. Fathizadan, S.; Ju, F.; Lu, Y.; Yang, Z. Deep spatio-temporal anomaly detection in laser powder bed fusion. IEEE Trans. Autom. Sci. Eng. 2023, 21, 5227–5239. [Google Scholar] [CrossRef]
  6. Asres, M.W.; Omlin, C.W.; Wang, L.; Yu, D.; Parygin, P.; Dittmann, J.; Karapostoli, G.; Seidel, M.; Venditti, R.; Lambrecht, L.; et al. Spatio-temporal anomaly detection with graph networks for data quality monitoring of the Hadron Calorimeter. Sensors 2023, 23, 9679. [Google Scholar] [CrossRef]
  7. Zhao, Y.; Deng, L.; Chen, X.; Guo, C.; Yang, B.; Kieu, T.; Huang, F.; Pedersen, T.B.; Zheng, K.; Jensen, C.S. A comparative study on unsupervised anomaly detection for time series: Experiments and analysis. arXiv 2022, arXiv:2209.04635. [Google Scholar]
  8. Chalapathy, R.; Chawla, S. Deep learning for anomaly detection: A survey. arXiv 2019, arXiv:1901.03407. [Google Scholar]
  9. Cook, A.A.; Mısırlı, G.; Fan, Z. Anomaly detection for IoT time-series data: A survey. IEEE Internet Things J. 2019, 7, 6481–6494. [Google Scholar] [CrossRef]
  10. Wang, Y.; Liu, G. Self-supervised dam deformation anomaly detection based on temporal–spatial contrast learning. Sensors 2024, 24, 5858. [Google Scholar] [CrossRef]
  11. Yang, J.; Chu, H.; Guo, L.; Ge, X. A Weighted-Transfer Domain-Adaptation Network Applied to Unmanned Aerial Vehicle Fault Diagnosis. Sensors 2025, 25, 1924. [Google Scholar] [CrossRef] [PubMed]
  12. Wang, S.; Miao, H.; Li, J.; Cao, J. Spatio-temporal knowledge transfer for urban crowd flow prediction via deep attentive adaptation networks. IEEE Trans. Intell. Transp. Syst. 2021, 23, 4695–4705. [Google Scholar] [CrossRef]
  13. Yu, F.; Xiu, X.; Li, Y. A survey on deep transfer learning and beyond. Mathematics 2022, 10, 3619. [Google Scholar] [CrossRef]
  14. Shao, S.; McAleer, S.; Yan, R.; Baldi, P. Highly accurate machine fault diagnosis using deep transfer learning. IEEE Trans. Ind. Inform. 2018, 15, 2446–2455. [Google Scholar] [CrossRef]
  15. Laptev, N.; Yu, J.; Rajagopal, R. Reconstruction and regression loss for time-series transfer learning. In Proceedings of the Special Interest Group on SIGKDD, London, UK, 19–23 August 2018; Volume 20. [Google Scholar]
  16. Gupta, P.; Malhotra, P.; Vig, L.; Shroff, G. Transfer learning for clinical time series analysis using recurrent neural networks. arXiv 2018, arXiv:1807.01705. [Google Scholar]
  17. Boullé, N.; Dallas, V.; Nakatsukasa, Y.; Samaddar, D. Classification of chaotic time series with deep learning. Phys. D Nonlinear Phenom. 2020, 403, 132261. [Google Scholar] [CrossRef]
  18. Wang, L.; Geng, X.; Ma, X.; Liu, F.; Yang, Q. Cross-city transfer learning for deep spatio-temporal prediction. arXiv 2018, arXiv:1802.00386. [Google Scholar]
  19. Hijazi, M.; Dehghanian, P.; Wang, S. Transfer learning for transient stability predictions in modern power systems under enduring topological changes. IEEE Trans. Autom. Sci. Eng. 2023. [Google Scholar] [CrossRef]
  20. Shao, L.; Zhu, F.; Li, X. Transfer learning for visual categorization: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2014, 26, 1019–1034. [Google Scholar] [CrossRef]
  21. Niu, S.; Liu, Y.; Wang, J.; Song, H. A decade survey of transfer learning (2010–2020). IEEE Trans. Artif. Intell. 2020, 1, 151–166. [Google Scholar] [CrossRef]
  22. Adama, D.A.; Lotfi, A.; Ranson, R. A survey of vision-based transfer learning in human activity recognition. Electronics 2021, 10, 2412. [Google Scholar] [CrossRef]
  23. Chato, L.; Regentova, E. Survey of transfer learning approaches in the machine learning of digital health sensing data. J. Pers. Med. 2023, 13, 1703. [Google Scholar] [CrossRef] [PubMed]
  24. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  25. Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of deep bidirectional transformers for language understanding. arXiv 2018, arXiv:1810.04805. [Google Scholar]
  26. Sarker, M.I.; Losada-Gutiérrez, C.; Marron-Romera, M.; Fuentes-Jiménez, D.; Luengo-Sánchez, S. Semi-supervised anomaly detection in video-surveillance scenes in the wild. Sensors 2021, 21, 3993. [Google Scholar] [CrossRef] [PubMed]
  27. Natha, S.; Ahmed, F.; Siraj, M.; Lagari, M.; Altamimi, M.; Chandio, A.A. Deep BiLSTM attention model for spatial and temporal anomaly detection in video surveillance. Sensors 2025, 25, 251. [Google Scholar] [CrossRef]
  28. Guo, B.; Li, J.; Zheng, V.W.; Wang, Z.; Yu, Z. CityTransfer: Transferring inter-and intra-city knowledge for chain store site recommendation based on multi-source urban data. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 2018, 1, 135. [Google Scholar] [CrossRef]
  29. Evans, L.; Bryant, P. LHC machine. J. Instrum. 2008, 3, S08001. [Google Scholar] [CrossRef]
  30. The CMS Collaboration. Development of the CMS detector for the CERN LHC Run 3. arXiv 2023, arXiv:2309.05466. [Google Scholar]
  31. The CMS Collaboration; Chatrchyan, S.; Hmayakyan, G.; Khachatryan, V.; Sirunyan, A.; Adam, W.; Bauer, T.; Bergauer, T.; Bergauer, H.; Dragicevic, M.; et al. The CMS experiment at the CERN LHC. J. Instrum. 2008, 3, S08004. [Google Scholar]
  32. Azzolini, V.; Bugelskis, D.; Hreus, T.; Maeshima, K.; Fernandez, M.J.; Norkus, A.; Fraser, P.J.; Rovere, M.; Schneider, M.A. The data quality monitoring software for the CMS experiment at the LHC: Past, present and future. EPJ Web Conf. 2019, 214, 02003. [Google Scholar] [CrossRef]
  33. Tuura, L.; Meyer, A.; Segoni, I.; Della Ricca, G. CMS data quality monitoring: Systems and experiences. J. Phys. Conf. Ser. 2010, 219, 072020. [Google Scholar] [CrossRef]
  34. De Guio, F.; The CMS Collaboration. The CMS data quality monitoring software: Experience and future prospects. J. Phys. Conf. Ser. 2014, 513, 032024. [Google Scholar] [CrossRef]
  35. Azzolin, V.; Andrews, M.; Cerminara, G.; Dev, N.; Jessop, C.; Marinelli, N.; Mudholkar, T.; Pierini, M.; Pol, A.; Vlimant, J.R. Improving data quality monitoring via a partnership of technologies and resources between the CMS experiment at CERN and industry. EPJ Web Conf. 2019, 214, 01007. [Google Scholar] [CrossRef]
  36. Asres, M.W.; Cummings, G.; Parygin, P.; Khukhunaishvili, A.; Toms, M.; Campbell, A.; Cooper, S.I.; Yu, D.; Dittmann, J.; Omlin, C.W. Unsupervised deep variational model for multivariate sensor anomaly detection. In Proceedings of the International Conference on Progress in Informatics and Computing, Shanghai, China, 17–19 December 2021; pp. 364–371. [Google Scholar]
  37. Asres, M.W.; Cummings, G.; Khukhunaishvili, A.; Parygin, P.; Cooper, S.I.; Yu, D.; Dittmann, J.; Omlin, C.W. Long horizon anomaly prediction in multivariate time series with causal autoencoders. Phm Soc. Eur. Conf. 2022, 7, 21–31. [Google Scholar] [CrossRef]
  38. The CMS-ECAL Collaboration. Autoencoder-based Anomaly Detection System for Online Data Quality Monitoring of the CMS Electromagnetic Calorimeter. arXiv 2023, arXiv:2309.10157. [Google Scholar]
  39. Pol, A.A.; Azzolini, V.; Cerminara, G.; De Guio, F.; Franzoni, G.; Pierini, M.; Sirokỳ, F.; Vlimant, J.R. Anomaly detection using deep autoencoders for the assessment of the quality of the data acquired by the CMS experiment. EPJ Web Conf. 2019, 214, 06008. [Google Scholar] [CrossRef]
  40. Pol, A.A.; Cerminara, G.; Germain, C.; Pierini, M.; Seth, A. Detector monitoring with artificial neural networks at the CMS experiment at the CERN Large Hadron Collider. Comput. Softw. Big Sci. 2019, 3, 3. [Google Scholar] [CrossRef]
  41. Parra, O.J.; Pardiñas, J.G.; Pérez, L.D.P.; Janisch, M.; Klaver, S.; Lehéricy, T.; Serra, N. Human-in-the-loop reinforcement learning for data quality monitoring in particle physics experiments. arXiv 2024, arXiv:2405.15508. [Google Scholar]
  42. Xie, P.; Li, T.; Liu, J.; Du, S.; Yang, X.; Zhang, J. Urban flow prediction from spatiotemporal data using machine learning: A survey. Inf. Fusion 2020, 59, 1–12. [Google Scholar] [CrossRef]
  43. Lai, Y.; Zhu, Y.; Li, L.; Lan, Q.; Zuo, Y. STGLR: A spacecraft anomaly detection method based on spatio-temporal graph learning. Sensors 2025, 25, 310. [Google Scholar] [CrossRef]
  44. Strobbe, N. The upgrade of the CMS Hadron Calorimeter with Silicon photomultipliers. J. Instrum. 2017, 12, C01080. [Google Scholar] [CrossRef]
  45. Huber, F.; Inderka, A.; Steinhage, V. Leveraging remote sensing data for yield prediction with deep transfer learning. Sensors 2024, 24, 770. [Google Scholar] [CrossRef] [PubMed]
  46. Wu, P.; Liu, J.; Li, M.; Sun, Y.; Shen, F. Fast sparse coding networks for anomaly detection in videos. Pattern Recognit. 2020, 107, 107515. [Google Scholar] [CrossRef]
  47. Hasan, M.; Choi, J.; Neumann, J.; Roy-Chowdhury, A.K.; Davis, L.S. Learning temporal regularity in video sequences. In Proceedings of the Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 733–742. [Google Scholar]
  48. Luo, W.; Liu, W.; Lian, D.; Tang, J.; Duan, L.; Peng, X.; Gao, S. Video anomaly detection with sparse coding inspired deep neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 43, 1070–1084. [Google Scholar] [CrossRef] [PubMed]
  49. Hsu, D. Anomaly detection on graph time series. arXiv 2017, arXiv:1708.02975. [Google Scholar]
  50. Focardi, E. Status of the CMS detector. Phys. Procedia 2012, 37, 119–127. [Google Scholar] [CrossRef]
  51. Neutelings, I. CMS Coordinate System. 2023. Available online: https://tikz.net/axis3d_cms/ (accessed on 14 December 2023).
  52. Cheung, H.W.; The CMS Collaboration. CMS: Present status, limitations, and upgrade plans. Phys. Procedia 2012, 37, 128–137. [Google Scholar] [CrossRef]
  53. Virdee, T.; The CMS Collaboration. The CMS experiment at the CERN LHC. In Proceedings of the 6th International Symposium on Particles, Strings and Cosmology, Boston, MA, USA, 22–29 March 1999.
  54. Wen, T.; Keyes, R. Time series anomaly detection using convolutional neural networks and transfer learning. arXiv 2019, arXiv:1905.13628. [Google Scholar]
  55. Kingma, D.P.; Welling, M. Auto-encoding variational bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  56. Van Laarhoven, T. L2 regularization versus batch and weight normalization. arXiv 2017, arXiv:1706.05350. [Google Scholar]
  57. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  58. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 1026–1034. [Google Scholar]
  59. Smith, L.N.; Topin, N. Super-convergence: Very fast training of neural networks using large learning rates. In Proceedings of the Artificial Intelligence and Machine Learning for Multi-Domain Operations Applications, Baltimore, MD, USA, 14–18 April 2019; Volume 11006, pp. 369–386. [Google Scholar]
  60. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. JMLR 2008, 9, 2579–2605. [Google Scholar]
Figure 1. Schematic of the CMS detector: (a) CMS with its major systems [50], and (b) geometry axes and angles of the CMS with respect to the collision intersection point [51].
Figure 1. Schematic of the CMS detector: (a) CMS with its major systems [50], and (b) geometry axes and angles of the CMS with respect to the collision intersection point [51].
Sensors 25 03475 g001
Figure 2. The subdetectors of the HCAL: (a) longitudinal view of the HB, HE, HF, and HO subdetectors on CMS [52]; and (b) longitudinal view of one quadrant of CMS with segmentation angle specifications of the η , where the origin denotes the interaction point [31,53].
Figure 2. The subdetectors of the HCAL: (a) longitudinal view of the HB, HE, HF, and HO subdetectors on CMS [52]; and (b) longitudinal view of one quadrant of CMS with segmentation angle specifications of the η , where the origin denotes the interaction point [31,53].
Sensors 25 03475 g002
Figure 3. A sample digi-occupancy map (year = 2018, RunId = 325,170, LS = 15): (a) digi-occupancy map for the HE and HB together; (b) the source system HE channels are placed in i η [ 16 , , 29 ] , i ϕ [ 1 , , 72 ] , and depth  [ 1 , , 7 ] ; and (c) the target system HB channels are placed in i η [ 1 , , 16 ] , i ϕ [ 1 , , 72 ] , and depth  [ 1 , 2 ] . The HE and HB share similarities and differences in tasks, calorimeter technology, and data characteristics. The missing sector at (b) corresponds to the two failed HE-RBX sectors during the 2018 collision runs.
Figure 3. A sample digi-occupancy map (year = 2018, RunId = 325,170, LS = 15): (a) digi-occupancy map for the HE and HB together; (b) the source system HE channels are placed in i η [ 16 , , 29 ] , i ϕ [ 1 , , 72 ] , and depth  [ 1 , , 7 ] ; and (c) the target system HB channels are placed in i η [ 1 , , 16 ] , i ϕ [ 1 , , 72 ] , and depth  [ 1 , 2 ] . The HE and HB share similarities and differences in tasks, calorimeter technology, and data characteristics. The missing sector at (b) corresponds to the two failed HE-RBX sectors during the 2018 collision runs.
Sensors 25 03475 g003
Figure 4. Total digi-occupancy data distribution of the HB and run settings per map (s): the received luminosity ( β s ) and the number of events ( ξ s ). N 1 is the renormalization of γ s based on ξ s , and N 2 is the reversible renormalization based on the median γ along the i ϕ axis. The colors correspond to different collision runs.
Figure 4. Total digi-occupancy data distribution of the HB and run settings per map (s): the received luminosity ( β s ) and the number of events ( ξ s ). N 1 is the renormalization of γ s based on ξ s , and N 2 is the reversible renormalization based on the median γ along the i ϕ axis. The colors correspond to different collision runs.
Sensors 25 03475 g004
Figure 5. The architecture of the proposed AE for the GraphSTAD system [6]. The GNN and CNN provide spatial feature extraction for each time step, and the RNN network captures the temporal behavior of the extracted features. The feature extraction E θ incorporates the GNN for back-end physical connectivity among the spatial channels, CNN for regional spatial proximity of the channels, and RNN for temporal behavior extraction. D ω contains RNNs and deconvolutional neural networks to reconstruct the ST input data from the low-dimensional latent features.
Figure 5. The architecture of the proposed AE for the GraphSTAD system [6]. The GNN and CNN provide spatial feature extraction for each time step, and the RNN network captures the temporal behavior of the extracted features. The feature extraction E θ incorporates the GNN for back-end physical connectivity among the spatial channels, CNN for regional spatial proximity of the channels, and RNN for temporal behavior extraction. D ω contains RNNs and deconvolutional neural networks to reconstruct the ST input data from the low-dimensional latent features.
Sensors 25 03475 g005
Figure 6. General framework of the proposed transfer learning mechanism.
Figure 6. General framework of the proposed transfer learning mechanism.
Sensors 25 03475 g006
Figure 7. Reconstruction L M S E performance of the TL on spatial networks across different epochs: (a) test MSE loss, where the bars show the dispersion of five repeated experiments; (b) the average relative difference between the MSE loss and no-TL. TL is applied with T init on the encoder and decoder using TL-4: ENCODER[CNN, GNN], DECODER[CNN], and T train on the encoder using TL-1: ENCODER[GNN], TL-2: ENCODER[CNN], and TL-3: ENCODER[CNN, GNN]. The no-TL model starts to saturate at e p o c h > 150 .
Figure 7. Reconstruction L M S E performance of the TL on spatial networks across different epochs: (a) test MSE loss, where the bars show the dispersion of five repeated experiments; (b) the average relative difference between the MSE loss and no-TL. TL is applied with T init on the encoder and decoder using TL-4: ENCODER[CNN, GNN], DECODER[CNN], and T train on the encoder using TL-1: ENCODER[GNN], TL-2: ENCODER[CNN], and TL-3: ENCODER[CNN, GNN]. The no-TL model starts to saturate at e p o c h > 150 .
Sensors 25 03475 g007
Figure 8. Reconstruction L M S E performance of TL on the ST networks. TL is applied to train the encoder and decoder with TL-5: (ENCODER: [CNN, GNN, RNN]) and TL-6: (ENCODER: [CNN, GNN, RNN], DECODER: [RNN]). The MSE loss in (a) non-preserved LSTM states that reset for each time window, and (b) preserved LSTM states across consecutive time windows. The bars show the dispersion of five repeated experiments. Lower epochs have higher performance variations among repeated experiments, and the stabilization is better at higher epochs.
Figure 8. Reconstruction L M S E performance of TL on the ST networks. TL is applied to train the encoder and decoder with TL-5: (ENCODER: [CNN, GNN, RNN]) and TL-6: (ENCODER: [CNN, GNN, RNN], DECODER: [RNN]). The MSE loss in (a) non-preserved LSTM states that reset for each time window, and (b) preserved LSTM states across consecutive time windows. The bars show the dispersion of five repeated experiments. Lower epochs have higher performance variations among repeated experiments, and the stabilization is better at higher epochs.
Sensors 25 03475 g008
Figure 9. Digi-occupancy map reconstruction on sample ST data from the test set. The model was trained using TL-6, and the inference was executed (a) without and (b) with LSTM state preservation across time windows. The AE operates on ST γ maps, but the curves in these plots correspond to the aggregate renormalized γ per LS to illustrate the model’s performance in handling the fluctuation across lumisections.
Figure 9. Digi-occupancy map reconstruction on sample ST data from the test set. The model was trained using TL-6, and the inference was executed (a) without and (b) with LSTM state preservation across time windows. The AE operates on ST γ maps, but the curves in these plots correspond to the aggregate renormalized γ per LS to illustrate the model’s performance in handling the fluctuation across lumisections.
Sensors 25 03475 g009
Figure 10. Spatial AD e i , M A E on a sample digi-occupancy map at depth  = 1 with degraded ( 0 < R D < 100 % ), dead ( R D = 0 % ), noisy hot ( R D = 200 % ), and fully hot ( γ ^ a = 1.0 ) anomaly types: (a) renormalized digi-occupancy map with simulated anomaly channels; the reconstruction error maps of the (b) M N model and (c) the M T model. The anomaly region is localized well with proportional strength to the severity of the anomaly in both models. The M T model has better localization with relatively less dispersion in its anomaly score map.
Figure 10. Spatial AD e i , M A E on a sample digi-occupancy map at depth  = 1 with degraded ( 0 < R D < 100 % ), dead ( R D = 0 % ), noisy hot ( R D = 200 % ), and fully hot ( γ ^ a = 1.0 ) anomaly types: (a) renormalized digi-occupancy map with simulated anomaly channels; the reconstruction error maps of the (b) M N model and (c) the M T model. The anomaly region is localized well with proportional strength to the severity of the anomaly in both models. The M T model has better localization with relatively less dispersion in its anomaly score map.
Sensors 25 03475 g010
Figure 11. AD reconstruction e i , M A E distribution of healthy and anomalous channels at different degradation rates of the simulated anomalies. The models are (a M N and (b) M T . The overlap region decreases substantially as the channel deterioration increases for R D < 100 % . However, the overlap increases for R D = 200 % , as the error increases for the normal channels due to the correlation to adjacent anomalies (as shown in Figure 12).
Figure 11. AD reconstruction e i , M A E distribution of healthy and anomalous channels at different degradation rates of the simulated anomalies. The models are (a M N and (b) M T . The overlap region decreases substantially as the channel deterioration increases for R D < 100 % . However, the overlap increases for R D = 200 % , as the error increases for the normal channels due to the correlation to adjacent anomalies (as shown in Figure 12).
Sensors 25 03475 g011
Figure 12. Spatio-temporal location embedding for channels with high e i , M A E in the presence of noisy hot anomalies ( R D = 200 % ) with M T : (left to right) location embedding for the anomaly channels ( A n o m a l y = T r u e ), the normal channels ( A n o m a l y = F a l s e ), and both, respectively. We applied t-SNE embedding [60] to the channels’ locations (coordinates: LS, i η , i ϕ , and depth) to generate the 2D representation. The normal channels ( A n o m a l y = F a l s e ) with high reconstruction error occur near the anomalous channels.
Figure 12. Spatio-temporal location embedding for channels with high e i , M A E in the presence of noisy hot anomalies ( R D = 200 % ) with M T : (left to right) location embedding for the anomaly channels ( A n o m a l y = T r u e ), the normal channels ( A n o m a l y = F a l s e ), and both, respectively. We applied t-SNE embedding [60] to the channels’ locations (coordinates: LS, i η , i ϕ , and depth) to generate the 2D representation. The normal channels ( A n o m a l y = F a l s e ) with high reconstruction error occur near the anomalous channels.
Sensors 25 03475 g012
Figure 13. Proximity effect explanation for false positives in noisy hot channel anomaly ( R D = 200 % ) detection. The healthy channels with higher AD scores ( e i , M A E > 0.2 ) belong to the filtered out channels from the anomaly injection R D γ h > ξ (see Equation (11)) and generate a high score due to their proximity to the abnormal channels. * Example of anomaly region selection where channels with R D γ h > ξ are excluded from the selection to meet the requirement of γ a [ 0 , ξ ] .
Figure 13. Proximity effect explanation for false positives in noisy hot channel anomaly ( R D = 200 % ) detection. The healthy channels with higher AD scores ( e i , M A E > 0.2 ) belong to the filtered out channels from the anomaly injection R D γ h > ξ (see Equation (11)) and generate a high score due to their proximity to the abnormal channels. * Example of anomaly region selection where channels with R D γ h > ξ are excluded from the selection to meet the requirement of γ a [ 0 , ξ ] .
Sensors 25 03475 g013
Figure 14. Spatial AD e i , M A E per channel map at depth  = 1 averaged over the training dataset for the (a) M N model, and (b) M T model. M N reconstructs the real dead channels (in red boxes) as normal, with very low error scores. In contrast, M T produces a high error, which signifies detecting the channels as anomalies.
Figure 14. Spatial AD e i , M A E per channel map at depth  = 1 averaged over the training dataset for the (a) M N model, and (b) M T model. M N reconstructs the real dead channels (in red boxes) as normal, with very low error scores. In contrast, M T produces a high error, which signifies detecting the channels as anomalies.
Sensors 25 03475 g014
Table 1. Description of source and target datasets.
Table 1. Description of source and target datasets.
DatasetSensor TechnologyNo. of Channels per RBXNo. of RBXesCalorimeter Segmentation RangesSample Size
Source (HE)SiPM19236 i η [ 16 , , 29 ] , i ϕ [ 1 , , 72 ] , depth  [ 1 , , 7 ] 20,000
Target (HB)HPD7236 i η [ 1 , , 16 ] , i ϕ [ 1 , , 72 ] , depth  [ 1 , 2 ] 7000
Table 2. Transfer learning experiment configurations.
Table 2. Transfer learning experiment configurations.
Config.Init Mode ( T init )Train Mode ( T train )
Notation Description Notation Description
1RANDOM M b is initialized randomly (weights: using Kaiming uniform [58], and biases: zero)No-TLComplete training (fine-tuning)
2TL-4 M d is initialized randomly, except the spatial learning networks (CNN and GNN) are initialized by TL from M e No-TLComplete training (fine-tuning)
3TL-1GNN of E θ is frozen (not fine-tuned)
4TL-2CNN of E θ is frozen
5TL-2dCNN of D ω is frozen
6TL-3CNN and GNN of E θ are frozen
7TL-7All the spatial and temporal learning networks (CNN, GNN, and RNN) of the M b are initialized by TL from M e TL-5CNN, GNN, and RNN of E θ are frozen
8TL-6CNN, GNN, and RNN of E θ , and RNN of D ω are frozen
TL: transfer learning is applied. TL-1: ENCODER[GNN], TL-2: ENCODER[CNN], TL-2d: DECODER[CNN], TL-3: ENCODER[CNN, GNN], TL-4: ENCODER[CNN, GNN], DECODER[CNN], TL-5: ENCODER[CNN, GNN, RNN], TL-6: ENCODER[CNN, GNN, RNN], DECODER[RNN], TL-7: ENCODER[CNN, GNN, RNN], DECODER[CNN, RNN].
Table 3. ST reconstruction L M S E of TL on spatial networks ( e p o c h = 200 ).
Table 3. ST reconstruction L M S E of TL on spatial networks ( e p o c h = 200 ).
T init T train L MSE Δ L MSE w.r.t T init = RANDOM
Average Score
RANDOMNo-TL 3.361 × 10 4
TL-4No-TL 2.666 × 10 4 −20.7%
TL-4TL-1 2.604 × 10 4 −22.5%
TL-4TL-2 2 . 463 × 10 4 −26.7%
TL-4TL-3 2.489 × 10 4 −25.9%
TL-4TL-2d 1.530 × 10 2 4452.2%
Best Score
RANDOMNo-TL 3.085 × 10 4
TL-4No-TL 2.569 × 10 4 −16.7%
TL-4TL-1 2.502 × 10 4 −18.9%
TL-4TL-2 2 . 420 × 10 4 −21.6%
TL-4TL-3 2.451 × 10 4 −20.5%
TL-4TL-2d 1.5255 × 10 2 4844.9%
TL-1: ENCODER[GNN], TL-2: ENCODER[CNN], TL-2d: DECODER: [CNN], TL-3: ENCODER[CNN, GNN], TL-4: ENCODER[CNN, GNN], DECODER[CNN]. The bold font is the best score and, the down arrow (↓) indicates that lower is better.
Table 4. ST reconstruction L M S E of TL for spatial networks ( T init = TL−4, e p o c h = 200 , average score).
Table 4. ST reconstruction L M S E of TL for spatial networks ( T init = TL−4, e p o c h = 200 , average score).
T init T train L MSE Δ L MSE w.r.t TL- 2 d
TL-4TL-2d 1.530 × 10 2
TL-4TL-2d/[BN] 7.200 × 10 3 −53.0%
TL-4TL-2d/[BN, BIAS] 7.354 × 10 3 −51.9%
TL-2d: DECODER[CNN], TL-4: ENCODER[CNN, GNN], DECODER[CNN], and / denotes excluding.
Table 5. ST reconstruction L M S E of TL on ST networks.
Table 5. ST reconstruction L M S E of TL on ST networks.
T init T train L MSE Δ L MSE w.r.t T init = RANDOM Δ W w.r.t T init = RANDOM
Best Score at e p o c h = 75
RANDOMNo-TL 3.826 × 10 4
TL-4No-TL 3.180 × 10 4 −16.9%0.00%
TL-4TL-1 3.082 × 10 4 −19.5%−0.17%
TL-4TL-2 2.686 × 10 4 −29.8%−2.23%
TL-4TL-3 2.705 × 10 4 -29.3%−2.39%
TL-7TL-5 2.667 × 10 4 −30.3%−8.38%
TL-7TL-6 2 . 577 × 10 4 −32.6%−97.77%
Best Score at e p o c h = 200
RANDOMNo-TL 3.085 × 10 4
TL-4No-TL 2.569 × 10 4 −16.7%0.00%
TL-4TL-1 2.502 × 10 4 −18.9%−0.17%
TL-4TL-2 2.420 × 10 4 −21.6%−2.23%
TL-4TL-3 2.451 × 10 4 −20.5%−2.39%
TL-7TL-5 2.457 × 10 4 −20.4%−8.38%
TL-7TL-6 2 . 389 × 10 4 −22.6%−97.77%
TL-1: ENCODER[GNN], TL-2: ENCODER[CNN], TL-3: ENCODER[CNN, GNN], TL-4: ENCODER[CNN, GNN], DECODER[CNN], TL-5: ENCODER[CNN, GNN, RNN], TL-6: ENCODER[CNN, GNN, RNN], DECODER[RNN], TL-7: ENCODER[CNN, GNN, RNN], DECODER[CNN, RNN]. Δ W is the reduction in the number of trainable model parameters. The bold font is the best score and, the down arrow (↓) indicates that lower is better.
Table 6. ST reconstruction L M S E of TL with LR scheduling mechanism ( e p o c h = 200 , best score).
Table 6. ST reconstruction L M S E of TL with LR scheduling mechanism ( e p o c h = 200 , best score).
T init T train L MSE Δ L MSE w.r.t T init  =  RANDOM
RANDOMNo-TL 2.500 × 10 4
TL-4No-TL 2.400 × 10 4 −4.0%
TL-4TL-3 2.460 × 10 4 −1.6%
TL-7TL-5 2 . 283 × 10 4 −8.7%
TL-7TL-6 2.286 × 10 4 −8.6%
TL-3: ENCODER[CNN, GNN], TL-4: ENCODER[CNN, GNN], DECODER[CNN], TL-5: ENCODER[CNN, GNN, RNN], TL-6: ENCODER[CNN, GNN, RNN], DECODER[RNN], TL-7: ENCODER[CNN, GNN, RNN], DECODER[CNN, RNN]. The bold font is the best score and, the down arrow (↓) indicates that lower is better.
Table 7. AD performance DQM abnormal channels.
Table 7. AD performance DQM abnormal channels.
Channel Anomaly TypeFPR (90%) ↓FPR (95%) ↓FPR (99%) ↓AUC ↑
M N : T init = RANDOM and  T train = No-TL
Degraded ( R D = 80 % ) 6.281 × 10 4 1.519 × 10 3 8.741 × 10 3 0.993
Degraded ( R D = 60 % ) 5.991 × 10 5 1.438 × 10 4 8.242 × 10 4 1.000
Degraded ( R D = 40 % ) 4.881 × 10 5 5.628 × 10 5 1.466 × 10 4 1.000
Degraded ( R D = 20 % ) 5.870 × 10 5 6.273 × 10 5 7.342 × 10 5 1.000
Dead ( R D = 0 % ) 6.636 × 10 5 7.080 × 10 5 8.331 × 10 5 1.000
Noisy hot ( R D = 200 % ) 3 . 300 × 10 4 5 . 732 × 10 4 1 . 765 × 10 3 1.000
Fully hot ( γ a = ξ , γ a > γ h ) 1.220 × 10 4 1.317 × 10 4 1.606 × 10 4 1.000
M T : T init = TL-7  and  T train = TL-6
Degraded ( R D = 80 % ) 5 . 019 × 10 4 1 . 320 × 10 3 6 . 527 × 10 3 0.996
Degraded ( R D = 60 % ) 2 . 118 × 10 5 9 . 642 × 10 5 8 . 141 × 10 4 1.000
Degraded ( R D = 40 % ) 1 . 614 × 10 6 4 . 034 × 10 6 7 . 161 × 10 5 1.000
Degraded ( R D = 20 % ) 1 . 614 × 10 6 3 . 833 × 10 6 8 . 472 × 10 6 1.000
Dead ( R D = 0 % ) 1 . 815 × 10 6 4 . 236 × 10 6 8 . 472 × 10 6 1.000
Noisy hot ( R D = 200 % ) 1.380 × 10 3 2.143 × 10 3 4.099 × 10 3 1.000
Fully hot ( γ a = ξ , γ a > γ h )0.000 6 . 051 × 10 7 3 . 631 × 10 5 1.000
AUC: area under the receiver operating characteristic curve and FPR: false positive rate. The FPR ( ρ % ) denotes the FPR score of capturing the ρ % of the anomalies. The bold font is the better score between M N and M T . The down arrows (↓) indicates that lower is better, and vice versa for th up arrows (↑).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Asres, M.W.; Omlin, C.W.; Wang, L.; Yu, D.; Parygin, P.; Dittmann, J.; the CMS-HCAL Collaboration. Data Quality Monitoring for the Hadron Calorimeters Using Transfer Learning for Anomaly Detection. Sensors 2025, 25, 3475. https://doi.org/10.3390/s25113475

AMA Style

Asres MW, Omlin CW, Wang L, Yu D, Parygin P, Dittmann J, the CMS-HCAL Collaboration. Data Quality Monitoring for the Hadron Calorimeters Using Transfer Learning for Anomaly Detection. Sensors. 2025; 25(11):3475. https://doi.org/10.3390/s25113475

Chicago/Turabian Style

Asres, Mulugeta Weldezgina, Christian Walter Omlin, Long Wang, David Yu, Pavel Parygin, Jay Dittmann, and the CMS-HCAL Collaboration. 2025. "Data Quality Monitoring for the Hadron Calorimeters Using Transfer Learning for Anomaly Detection" Sensors 25, no. 11: 3475. https://doi.org/10.3390/s25113475

APA Style

Asres, M. W., Omlin, C. W., Wang, L., Yu, D., Parygin, P., Dittmann, J., & the CMS-HCAL Collaboration. (2025). Data Quality Monitoring for the Hadron Calorimeters Using Transfer Learning for Anomaly Detection. Sensors, 25(11), 3475. https://doi.org/10.3390/s25113475

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop