Next Article in Journal
Joint Relay Selection and Power Allocation through a Genetic Algorithm for Secure Cooperative Cognitive Radio Networks
Next Article in Special Issue
A New Method of Priority Assignment for Real-Time Flows in the WirelessHART Network by the TDMA Protocol
Previous Article in Journal
A Feasibility Study of Transformer Winding Temperature and Strain Detection Based on Distributed Optical Fibre Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Protection of Superconducting Industrial Machinery Using RNN-Based Anomaly Detection for Implementation in Smart Sensor †

by
Maciej Wielgosz
1,2,*,
Andrzej Skoczeń
3 and
Ernesto De Matteis
4
1
Faculty of Computer Science, Electronics and Telecommunications, AGH University of Science and Technology, al. Adama Mickiewicza 30, 30-059 Cracow, Poland
2
Academic Computer Centre CYFRONET AGH, ul. Nawojki 11, 30-072 Cracow, Poland
3
Faculty of Physics and Applied Computer Science, AGH University of Science and Technology, al. Adama Mickiewicza 30, 30-059 Cracow, Poland
4
CERN European Organization for Nuclear Research, CH-1211 Geneva 23, Switzerland
*
Author to whom correspondence should be addressed.
This paper is an extended version of “Looking for a Correct Solution of Anomaly Detection in the LHC Machine Protection System” published in the Proceedings of the 2018 International Conference on Signals and Electronic Systems (ICSES), Kraków, Poland, 10–12 September 2018.
Sensors 2018, 18(11), 3933; https://doi.org/10.3390/s18113933
Submission received: 27 October 2018 / Revised: 9 November 2018 / Accepted: 11 November 2018 / Published: 14 November 2018
(This article belongs to the Special Issue Real-Time Sensor Networks and Systems for the Industrial IoT)

Abstract

:
Sensing the voltage developed over a superconducting object is very important in order to make superconducting installation safe. An increase in the resistive part of this voltage (quench) can lead to significant deterioration or even to the destruction of the superconducting device. Therefore, detection of anomalies in time series of this voltage is mandatory for reliable operation of superconducting machines. The largest superconducting installation in the world is the main subsystem of the Large Hadron Collider (LHC) accelerator. Therefore a protection system was built around superconducting magnets. Currently, the solutions used in protection equipment at the LHC are based on a set of hand-crafted custom rules. They were proved to work effectively in a range of applications such as quench detection. However, these approaches lack scalability and require laborious manual adjustment of working parameters. The presented work explores the possibility of using the embedded Recurrent Neural Network as a part of a protection device. Such an approach can scale with the number of devices and signals in the system, and potentially can be automatically configured to given superconducting magnet working conditions and available data. In the course of the experiments, it was shown that the model using Gated Recurrent Units (GRU) comprising of two layers with 64 and 32 cells achieves 0.93 accuracy for anomaly/non-anomaly classification, when employing custom data compression scheme. Furthermore, the compression of proposed module was tested, and showed that the memory footprint can be reduced four times with almost no performance loss, making it suitable for hardware implementation.

1. Introduction

The benefit of using superconductivity in industrial applications is well understood. However actual application of superconducting devices is still limited by difficulties in the maintenance of cryogenic stability of superconducting cables and coils. In many cases, it is hard to design safe superconducting circuit. A small mechanical rearrangement releases enough energy to initiate local avalanche process (quench) leading to loss of superconducting state and next to overheating of the machine and then even to melting. Therefore superconducting machines require a special protection system.
In general, such a system consists of data producers and processing servers. The producers are electronic devices located close to individual superconducting machines which require protection. This device (producer) is capable of collecting and processing data and generating the activation signal for local actuators. The servers are capable of storing and analyzing data delivered by producers through the network.
The overarching research goal is to improve data processing within the producers and to move a part of the analysis task into them in order to reduce network load. The preliminary results of this research were presented in Ref. [1], which focused on testing the suitability of Recurrent Neural Networks (RNNs) for this application. In this work, the additional aspects for hardware implementation, especially model compression, are explored.
The presented research main contributions are as follows:
  • development of a neural algorithm dedicated to detecting anomaly occurring in the voltage time series acquired on the terminals of superconducting machines in electrical circuits,
  • design and verification of the complete processing flow,
  • introduction of the RNN-based solution for edge computing which paves the way for low-latency and low-throughput hardware implementation of the presented solution,
  • development of a system level model suited for future experiments with the adaptive grid-based approach; the software is available online (see Supplementary Material section).

1.1. Protection System for Superconducting Machinery

One of the biggest superconducting systems is installed at the Large Hadron Collider (LHC) accelerator at the European Organization for Nuclear Research (CERN). Despite its scientific purpose, the LHC should be considered as a huge industrial system. The final product of this factory is the total number of particle collisions. During the steady operation, every second, 40,000,000 collisions are performed in three different interaction points. The LHC consists of a chain of superconducting magnets. The chain is located in an underground circular tunnel, 100 m under the Earth’s surface. The Figure 1 presents the view on the magnet’s chain. The perimeter of the tunnel is about 27 km long. Superconducting magnets, responsible for shaping the beam trajectory, are crucial elements of the accelerator that require permanent monitoring. The details of the design of the LHC accelerator are described in Ref. [2].
When the LHC collides particles, even a tiny fraction of energy stored in each proton beam can cause a magnet to leave the superconducting state. Such an occurrence is named a quench, and it can severely damage the magnets in case of machine protection procedures failure. These procedures mainly relay on triggering the power down of the whole accelerator when resistive part of voltage on one superconducting element exceeds a predefined threshold. This study concerns only a protection system of superconducting elements inside the LHC.
A protection system known as Quench Protection System (QPS) has been installed at the LHC since the beginning and it successfully works since ten years of the LHC operation. The detailed description of the existing system can be found in Refs. [3,4]. The protection unit, visible in Figure 1, installed under the magnet performs acquisition, processing and buffering of samples of voltage existing between terminals of the superconducting coil. These units are end-points of a massive distributed system covering the whole length of the LHC. The acquisition is performed using Analog-to-Digital Converter (ADC). The processing relies on filtering and compensation of inductive voltage. The pure resistive voltage is compared with the threshold, and the output of comparator activates actuators and generates trigger signals for other subsystems of the LHC. The actuators are also included to the yellow rack installed under each magnet in the tunnel (see Figure 1). The task of actuators is to inject energy to the objective coil in order to heat it homogeneously.
The data is continuously buffered in a circular buffer. Some of the samples (both total and resistive voltage) are directly transmitted to the CERN Accelerator Logging Service (CALS) database in the cloud. The CALS serves permanent monitoring of almost every device in the CERN’s accelerator’s complex. The sampling rate for this service is very low, in the best case it is 10 Hz ( 100 ms ). The circular buffer consists of two parts. The first part is filled with samples all the time in a circular manner. The second part is filled only in case of triggering or in case of a request sent by an operator. A trigger (or a request) freezes first part of the buffer. Then the whole buffer is transmitted to a cloud and stored in a dedicated database Post Mortem (PM) System. Examples of voltage time series taken from this database are presented in Figure 2. The sampling rate of the PM data is much higher, and in our case, it is 500 Hz ( 2 ms ).
The presented protection system underwent many upgrades introduced during breaks of the LHC operation. However, the emergence of new superconducting materials opens a question concerning detection algorithms for application in the future protection system again. Such experiments are conducted in SM18 CERN’s facility (see Figure 3), currently testing the magnets for the High Luminosity LHC phase.

1.2. State of the Art

Anomaly and novelty detection methods have been researched over many years [5,6,7] which resulted in the development of many successful algorithms. They may be in general divided into three different categories of density-based, distance-based and parametric methods. In addition to the standard procedures, neural algorithms in a majority of cases employing RNN-based architectures [8,9,10,11] slowly pave the way to the basic set of the anomaly detection procedures.
Training dataset is different for novelty and outlier detection. In novelty detection, it is not contaminated by anomalies—all outliers need to be removed beforehand. On the contrary, the training procedure of the outlier detection model involves incorporating anomalous data into the training dataset. Both flows employ an unsupervised approach as a training procedure, although some of the procedures may be boosted using some external hyper-parameters such as contamination factor, thresholds or max features that are taken into account in the training process.
In this work three different algorithms were used as a baseline for the RNN-based approach proposed by the authors: Elliptic Envelope, Isolation Forest (IF), and One-Class Support Vector Machine (OC-SVM) [12,13,14,15,16,17,18]. Elliptic Envelope belongs to a set of methods with an underlying assumption of known distribution (usually Gaussian) for normal data and all the points distant from the center of the ellipse are considered outliers. The Mahalanobis distance [19] is used as a measure of distance and an indicator that a given data point may be considered as an outlier.
Another useful method, invented in 2008, is the Isolation Forest [20] which performs anomaly detection using the random forest. The underlying concept of this approach is based on an idea of a random selection of features and a random selection of split within the tree nodes between maximum and minimum values of the selected features. A concept of the decision function in IF algorithm defines deviation of an averaged path over a forest of random trees. Random partitioning creates significantly shorter paths for anomalies which results from the fact that outliers are concentrated close to extreme values of features which in turn locates them on the border of the trees.
OC-SVM is usually considered to be a novelty-detection method, and the training data should not contain outliers. It performs well in high-dimensional space where there is no assumption regarding the distribution of the underlying data. However, if the data is well distributed (e.g., Gaussian) the Isolation Forest or Elliptic Envelope may perform better.
The presented methods assume the spatial structure of the data with no temporal relationships. There is a set of methods such as ESD, ARIMA, Holt-Winters [21,22] which take time component into account and have proven to be effective. However, due to their complexity, it is challenging to implement them in very low-latency systems in hardware.
When detecting anomalies in time series, RNNs are both scalable and adaptable which is critical when it comes to design and implementation of complex anomaly detection systems [23,24,25]. RNNs were introduced long ago [26] and have been successfully applied in many domains [8], according to the authors’ knowledge, there are no studies on the performance of compressed RNNs in anomaly detection. Nevertheless, several works on effective quantization and compression of RNNs are available [27,28]. Consequently, decision was made to explore the feasibility and performance of several compression techniques of RNNs in low-latency anomaly detection domain of LHC machinery monitoring. Furthermore, the adopted approach addresses both data and coefficients quantization with in-depth analysis of correlation of employing different techniques.

2. Materials and Methods

2.1. Quantization Algorithm

2.1.1. Previous Work

In the authors’ previous works concerning superconducting magnets monitoring Root-Mean-Square Error (RMSE) [24] and both static [23] and adaptive data quantization [1,25] approaches were used. Based on experiments conducted and described therein, a conclusion can be drawn that RNNs can be used to model magnets behavior and detect anomalous occurrences.
The initially introduced RMSE approach had several drawbacks, the main of which was a necessity to select an arbitrary detection threshold. In Ref. [23], the static quantization was used, mapping the input data into a set of m equal-width bins. This method, however, resulted in sub-par results, stemming from the uneven distribution of the samples in the bins, up to the point where nearly all samples occupied only one or two bins (see static samples counts in Figure 4).
In Ref. [25], an approach based on adaptive data quantization and automatic thresholds selection was introduced. Adaptive data quantization resulted in much better use of bins and consequently significantly improved the accuracy results. Its principle of operation is mapping the input space to a fixed number of categories (bins) in such a way, that all categories have (ideally) the same samples cardinality (see Appendix A.1 for more details). Resulting bins widths are uneven, explicitly adjusted to each of the input signal channels (see adaptive samples counts in Figure 4 and Figure 5 to compare bin edges generated with various approaches).

2.1.2. Other Quantization Approaches

The drawback of adaptive algorithm is that it can effectively generate fewer bins than requested when some values occur in the dataset in significant numbers (see Figure 5). To mitigate this effect, which became apparent when working with PM data, a modification of adaptive algorithm, called recursive_adaptive, was introduced. The m + 1 initially found edges are treated as candidates, and if duplicates are detected, the recursive process is started. At first, the duplicated edges are added to the final edges list, and all repeating values are removed from the dataset. Then, the remaining data is used as an entry point to find m + 1 number of final edges new edge candidates. The process is repeated until there are no duplicates in candidate edges or there is no more data left in the dataset. As a result of this process, the bins are more evenly used.
An alternative edge-finding algorithm, called cumulative_amplitude, is based on the idea of equalizing the sum of the samples amplitudes in each bin. As in adaptive algorithm, before edges selection, the samples are normalized and sorted. Then, the threshold value is computed as a sum of amplitudes of samples left in the dataset divided by the required edges number (see Appendix A.2 for equations). Contrary to the adaptive approach of determining the edges based on the samples count, in cumulative_amplitude the edge value is chosen when the sum of samples’ amplitudes crosses this threshold. As a result, the maximal values are not grouped with smaller ones. It may, however, level the differences between smaller values, that may contain crucial information. In the implementation, the concept described above was modified to also use recursive duplication removal, with the threshold value determined anew for each recursion level.

2.2. Implementation Overview

The presented anomaly detection system was created in Python, using Keras [29] library, with both Theano [30] and Tensorflow [31] backends depending on availability. The reference methods were implemented using scikit-learn [32] library. It is prepared to work with normalized data, with all the available data (both training and testing) used during the normalization process. The focal system modules and data flow can be seen in Figure 6.
The number of input categories ( in _ grid ), the bins’ edges calculation algorithm ( in _ algorithm ), the history window length ( look _ back ), the model and its hyper-parameters used during the anomaly detection process and other options are specified in the configuration file. The particular setup is also saved while results are reported, ensuring the particular test environment can be recreated even if configuration included several possible values. Each of the input channels is quantized using the same grid/algorithm combination.
The model is an abstraction layer over the actual classifier. Currently implemented models include Random (for baseline testing), Elliptic Envelope, Isolation Forest, OC-SVM, Long Short-Term Memory (LSTM) and Gated Recurrent Unit (GRU).
Depending on the system configuration (Figure 6), some models can be used for either classification or regression. In the regression mode, the model is trained on data without anomalies and yields the output that needs to be further processed by the analyzer to obtain anomaly detection results.
In the classification mode, used in the experiments presented in this paper, the model is trained using data containing anomalies (except for models belonging to the novelty detection category). Instead of trying to predict the next (quantized) value, it directly classifies the sample as either anomaly or not.
The models can be roughly divided into the ones working with either spatial (Elliptic Envelope, Isolation Forest, OC-SVM) or temporal data (LSTM, GRU). Relevant data preprocessing and structuring, as well as model training and testing, is coordinated by one of the possible detectors. The model/detector combination used in particular setup is defined in configuration file. This option ensures the system extensibility since the detector does not need to know about all possible models in advance.
Currently, the experiments are carried out using the software implementation of the system. The target system, however, will need to be implemented in hardware to ensure it complies with latency requirements. To fit the Neural Network (NN) model onto the Field-Programmable Gate Array (FPGA) or Application-Specific Integrated Circuit (ASIC) board, it needs to be compressed while retaining the high accuracy (Figure 7). The ready module can potentially be used as a stand-alone detector or in conjunction with the currently used system (Figure 8).

2.3. Model Complexity Reduction

Deep Learning models have a range of features which render them superior to other similar Machine Learning models. However, they usually have high memory footprint as well as require substantial processing power [33]. Computing requirements are especially crucial when it comes to embedded implementation of Deep Learning models in edge processing nodes like in the case of the system described in this paper. Fortunately, there are multiple ways to mitigate these issues and preserve all the benefits of the models, for example by using techniques such as pruning and quantization [28].
During a quantization, a floating-point number x from a quasi-continuous space of IEEE-754 notation is mapped to fixed-point value q, represented using total bits. The total is conventionally equal to 8 or 16, which are bit-widths supported by GPUs and the latest embedded processors. For FPGA and ASIC it is, however, possible to use an arbitrary number of bits. The quantization is done separately for each layer’s weights W .

2.3.1. Linear Quantization

The linear quantization used during experiments can be described by the following equation:
q = s · clip x s + 1 2 , 1 2 total 1 , 2 total 1 1 ,
where scaling factor:
s = 1 2 total 1 log 2 max ( W )
and clipping function:
clip ( a , m i n , m a x ) = m i n if a < m i n , a if m i n a m a x , m a x otherwise .

2.3.2. MinMax Quantization

The minmax quantization used during experiments can be described by the following equation:
q = s · x min ( W ) s + 1 2 + min ( W ) ,
where scaling factor:
s = max ( W ) min ( W ) 2 total 1 .
Also tested was log_minmax quantization, where:
q = sign ( x ) · e minmax ( ln | x | )

2.3.3. Hyperbolic Tangent Quantization

The tanh quantization used during experiments can be described by the following equation:
q = arctanh s · tanh ( x ) + 1 s + 1 2 1 ,
where scaling factor:
s = 2 2 total 1 .
The main idea behind coefficients quantization is using a dynamic range of the available number representation to its fullest, and meet the requirements of the hardware platform to be used for deploying the system at the same time. In our implementation, the so-called dynamic fixed-point notation was used. It allows emulating fixed-point number representation using floating-point container. It is worth noting that most of edge computing platforms require linear quantization due to the fixed, a priori defined size of internal registers and arithmetic processing elements. FPGA and custom-designed ASIC which this work is targeting have no such limitation.

3. Results

A series of experiments were conducted to practically examine performance of the proposed methods. Different configurations of the module setup were used in order to expose impact of different parameters of the proposed algorithm on the overall performance of the anomaly detection system. Furthermore, the performance of the proposed solution was compared with a range of state-of-the-art algorithms.

3.1. Dataset

The dataset used in the experiments contained 2500 series retrieved from PM database, with 64-16-20 training-validation-testing split. 2415 of those series had a length of 1248 samples, while the remaining 85 series had a length of 1368 samples. For each of the series, the four input channels were available:
  • U DIFF —total voltage measured between terminals of superconducting coil,
  • U RES —resistive voltage extracted from the total voltage U DIFF using the electric current I DCCT ,
  • I DCCT —current flowing through superconducting coil measured using Hall sensor, and
  • I DIDT —time derivative of the electric current I DIDT calculated numerically.
Anomalies were marked based on the value of QUENCHTIME field found in PM data, with each anomaly starting at the indicated point and continuing until the end of a series. As such, the data can be considered to be weakly labelled. 874 training and 225 testing series contained anomalies and over 26% of samples in the dataset were marked. Over 84% of the anomalies had a length of 750 samples, over 7%—566, and over 4%—1320. The length of the remaining anomalies varied between 214 and 908 samples.
Before the start of experiments, the data was normalized. The example data series (and results) visualizations can be seen in Figure 9 and Figure 10. Even in just those two figures, it can be seen that the quenches vary in shape and it is really difficult to find apparent similarities just by visual examination. This makes tasks of data labeling and manual verification of the detection results not feasible without heavy experts involvement.

3.2. Quality Measures

During the experiments two main quality measures were used: F-measure and accuracy. While F-measure is better suited to evaluate the results of anomaly detection, in case of PM data the relative lack of imbalance between anomalous and normal samples (especially factoring in the required history length) makes the accuracy also a viable metric.
Additionally, the NN models quantization results are usually measured in terms of accuracy, so its usage allows to relate our results with others found in literature. For example, the drop in accuracy resulting from quantization should be no higher than one percentage point [34].
An accuracy can be defined as:
accuracy = tp + tn tp + tn + fp + fn ,
where:
  • tp —true positive—item correctly classified as an anomaly,
  • tn —true negative—item correctly classified as a part of normal operation,
  • fp —false positive—item incorrectly classified as an anomaly,
  • fn —false negative—item incorrectly classified as a part of normal operation.
An F-measure is calculated using two helper metrics, a recall (10), also called sensitivity, and a precision, also called specificity (11):
recall = tp tp + fn ,
precision = tp tp + fp .
The β parameter controls the recall importance in relevance to the precision when calculating an F-measure:
F β = ( 1 + β 2 ) · recall · precision recall + β 2 · precision .
During the experiments two β values were used, 1 and 2, to show the impact of the recall on the final score. Recall as a quality assessment measure reflect an ability of an algorithm to find all entities. On the other hand precision, describes several found entities were correctly classifier. Those measures have to some extent opposite effect on each other. This means that raise of precision usually leads to a drop of recall and vice-versa.
The Receiver Operating Characteristic (ROC) curve is a graph used to analyze the operation of the binary classifiers as the one presented in this work. It shows the performance of the model taking into account all classification thresholds. True Positive Rate ( recall ) is plotted as a function of False Positive Rate ( 1 precision ). When a classification threshold is lowered, the classifier tends to classify more input data items as positive, which leads to an increase of both fp and tp .
To derive quantitative conclusions from ROC curve Area Under Curve (AUC) may be employed. It measures the whole two-dimensional area under the ROC curve. It may be considered as an integral operation performed from points (0,0) to (1,1) on a ROC graph. AUC values fall into a range between 0 and 1. A model whose predictions are 100% wrong has an AUC of 0.0, the one which works perfectly hasAUC of 1.0.

3.3. History Length and Data Quantization

The initial experiments attempted to determine the impact of history length ( look _ back ) and quantization levels ( in _ grid ) on RNN models performance. Models were trained on full dataset and four channels for 7 epochs, with batch size equal to 16,384.
Use of dynamic range of the data representation is one of the most important indicators of the quantization algorithm effectiveness since it affects the potential information loss due to the lack of a proper representation capacity, i.e., ‘wasting’ resources on empty bins, while other bins contain ‘too many’ of the values or the number of bins could be reduced altogether. Figure 11 shows that recursive and cumulative adaptive approaches provide full grid use in contrast to adaptive quantization method which exhibits significant grid underuse.
Experiments with different sizes of in _ grid were conducted and it turned out that this parameter has a very little impact on the performance of the model (see Figure 12). Aside from in _ grid = 8 , which universally performs the worst regardless of the used algorithm, other values yield similar results starting with look _ back = 128 . This allows reducing the size of the input ( in _ grid ) to 32 which can be encoded using 5 bits. The biggest impact on the performance has look _ back which was presented in Figure 13, Figure 14 and Figure 15 and Table 1. The models with look _ back of 512 reach AUC close to 0.98 and significantly outperform the models with look _ back of 16. It also can be seen that in current tests only setups with look _ back = 256 and look _ back = 512 were capable of reaching the recall = 1 , ensuring all anomalies were found, while retaining high precision. The avoidance of false negatives is crucial in this use case, since quench after-effects, resulting in the equipment destruction, can be both dangerous and extremely costly.
It is also worth to keep in mind that the developed model is supposed to work in highly demanding environment where response latency is a critical factor which decides how much time other sections of the global protection system have to execute their procedures. Thus the work needs to be done towards reducing discrepancy in AUC value between setups of different look _ back . For instance, AUC for look _ back = 64 is 0.87 and for look _ back = 256 equals to 0.97 (see Figure 13).
The comparative tests were run using full training set ( samples _ percentage = 1 ). The goal was to study the behavior of the RNN-based methods and other classic anomaly detection methods with respect to the quantization algorithm. Most of the tests were run using single U RES input channel (Table 2), with additional experiments using four input channels run for RNN-based methods (Table 3).
As a baseline, the Random model was used. It generates predictions by respecting the training set’s class distribution (“stratified” strategy). Since it ignores the input data entirely, its results do not depend on the in _ algorithm choice. This model performance turned out to be at the level of 0.6334.
For the other models’ tests, based on the previous experiments, the values of look _ back = 256 and in _ grid = 32 were selected as providing a good tradeoff between resources consumption and the results quality. For the spatial methods, the history vector was flattened. The amount of contamination (the proportion of outliers in the data set) was calculated based on the whole training set and passed to the methods. For the Elliptic Envelope, all points were included in the support of the raw Minimum Covariance Determinant (MCD) estimate. The OC-SVM was tested with both 332 Radial Basis Function (RBF) and linear kernel, with RBF kernel coefficient ( γ ) equal to 0.1 and an upper bound on the fraction of training errors and a lower bound of the fraction of support vectors ( ν ) equal to 0.95 contamination + 0.05 . Each of the RNNs had the same testing architecture (two layers, with 64 cells in the first one and 32 cells in second, followed by fully connected layer) and was trained for seven epochs.
It is worth noting that the presented methods such as OC-SVM, Isolation Forest, or Elliptic Envelope are more sensitive to the data distribution and perform well for specific kind of underlying data (i.e., specific distribution or feature extraction which can bring the original data to this distribution before the methods are applied). In the case of the presented CERN data, the distribution is not always stable (varies across setups and magnets). It is also worth noting that the distribution as such does not tell much about temporal aspects of the analyzed data. It may happen that the signals of the same distribution have different temporal shape. This is even more pronounced for more temporarily complex signals (see Figure 9 and Figure 10).

3.4. Coefficients Quantization

Table 4 shows results of coefficients quantization for the neural models from Table 2 and Table 3 using several different methods. For all methods, the quantization above the ten bits yields results nearly identical to original. A significant drop in accuracy is observed below 8 bits of representation. It may be noted (Table 3) that for lower number of bits the performance oscillates between ≈ 0.7 and ≈ 0.3 which means classifying all the features as one category.

4. Discussion and Conclusions

The protection system for superconducting machinery existing at the LHC is a vast distributed system installed around the whole circular tunnel. It consists of many individual units connected with a dedicated network. The approach used to design protection units is hard to scale and requires laborious manual adjustment of working parameters. The presented work explores the possibility of using the RNN to build a protection device of a new generation. The idea is to perform on-line analysis inside local protection unit using data acquired with a much higher sampling rate without sending such a massive amount of data to the cloud.
One of the main advantages of the proposed methodology is the simplicity of the parameters setting and adjustment through a complete workflow. Only the model architecture and data quantization levels need to be selected, and even those can be automatically optimized. Additionally, since the solution is based on NNs, it can be extended (scaled) to use more sensors (or data streams) or even incorporate text tags (present in many cases in historical CERN superconducting magnets data), while keeping the overall architectural design the same, even for various types of magnets. It is also possible to fine-tune or retrain RNN-based modules when data has changed (e.g., underlying architecture was modified or aged) when in the traditional approach it would require reconsideration and restructuring of the existing solution. This architectural uniformity makes it a good candidate for implementation in a distributed edge-computing cluster of sensors, trained in an end-to-end fashion. Such a holistic approach can significantly reduce overall resources consumption, latency, and throughput.
The conducted experiments showed that using large look _ back significantly boost the performance of the model, while the number of quantization levels ( in _ grid ) as low as 32 is sufficient for the task. The framework demonstrated to be capable of achieving 93% of testing accuracy for GRU (two layers, 64 and 32 cells). The accuracy results are affected by the weak labeling of the data, e.g., sometimes the system labels as anomalous samples occurring for a bit before QUENCHTIME marker. Such results lower the accuracy, while in fact being the desired outcome. The proposed system also often creates a ‘gap’ in the anomaly, at which point the system shutdown signal would already be sent (such example can be seen in Figure 9). Considering that, the availability of the data manually labeled by experts could improve the system performance.
The coefficients quantization level should also be considered a meta-parameter of the model optimization. Experiments showed that the selection of any value equal to or above 8 bits does not lead to noticeable performance degradation. Careful choice of the quantization level may allow reducing memory footprint even more; however, it must be noted that below 8 bits the accuracy of the model oscillates.
Overall, due to the relatively small size of the neural models and the possibility of significantly reducing their memory footprint ( 4 × ) with a minimal performance loss, the presented model is a good candidate for hardware implementation in FPGA or ASIC.

Supplementary Materials

The software of presented anomaly detection system is available online at https://bitbucket.org/maciekwielgosz/anomaly_detection.

Author Contributions

Conceptualization, M.W., A.S. and E.D.M.; methodology, M.W.; software, M.W.; writing—original draft preparation, M.W. and A.S.; writing—review and editing, M.W. and A.S.; resources, M.W.; visualization, M.W., data curation, A.S. and E.D.M.; supervision, M.W.; funding acquisition, E.D.M.

Funding

This work was partially financed (supported) by the statutory tasks of Faculty of Physics & Applied Computer Science and Faculty of Computer Science, Electronics & Telecommunications of AGH University of Science and Technology in Cracow (Poland), within the subsidy of Polish Ministry of Science and Higher Education. The publication fee was funding by CERN.

Acknowledgments

We give special thanks to Andrzej Siemko the Group Leader of Machine Protection and 398 Electrical Integrity Group (MPE) at Technology Department (TE) at CERN and Reiner Denz the Section Leader 399 of Electronics for Protection Section (EP) at MPE-TE at CERN. They spend their time on long discussions and 400 revealed a great interest in our investigations.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ADCAnalog-to-Digital Converter
ASICApplication-Specific Integrated Circuit
AUCArea Under Curve
CALSCERN Accelerator Logging Service
CERNEuropean Organization for Nuclear Research
EPElectronics for Protection Section
FPGAField-Programmable Gate Array
GRUGated Recurrent Unit
IFIsolation Forest
LHCLarge Hadron Collider
LSTMLong Short-Term Memory
MCDMinimum Covariance Determinant
MPEMachine Protection and Electrical Integrity Group
NNNeural Network
OC-SVMOne-Class Support Vector Machine
PMPost Mortem
QPSQuench Protection System
RBFRadial Basis Function
RMSERoot-Mean-Square Error
RNNRecurrent Neural Network
ROCReceiver Operating Characteristic
TETechnology Department

Appendix A. Data Quantization

This section presents the equations that can help to better understand the data quantization performed during experiments.

Appendix A.1. Adaptive Data Quantization

Adaptive data quantization principle of operation is mapping the input space to a fixed number of categories (bins) in such a way, that all categories have (ideally) the same samples cardinality as described by Equations (A1) and (A3) (see Table A1 for notation).
S norm Π qa ( m ) S qa : 0 m 1 1 × n ,
Π qa ( m ) : x S norm y S qa y = edges y x · m < edges y + 1 if x < 1 y = m 1 if x = 1 ,
edges : 0 y m edges y = 0 if y = 0 srt _ samples y · n m if 0 < y < m 1 if y = m .

Appendix A.2. Cumulative Amplitude Data Quantization

Cumulative_amplitude method is based on the idea of equalizing the sum of the samples amplitudes in each bin. As in adaptive algorithm, before edges selection, the samples are normalized and sorted. Then, the threshold value Θ m is computed as a sum of amplitudes of samples left in the dataset divided by the required edges number (A4).
Θ m = srt _ samples m ,
edges : 0 y m edges y = 0 if y = 0 srt _ samples idx ( y ) if 0 < y < m 1 if y = m ,
idx ( y ) = 1 if y = 0 min ( k ) : i = idx ( y 1 ) + 1 k srt _ samples i Θ m if 0 < y < m .
Contrary to the adaptive approach of determining the edges based on the samples count, in cumulative_amplitude the edge value is chosen when the sum of samples’ amplitudes crosses this threshold (A5)–(A6).
Table A1. Notation used in quantizaton Equations (A1)–(A6).
Table A1. Notation used in quantizaton Equations (A1)–(A6).
SymbolMeaning
nnumber of samples
mnumber of classes (categories, bins); m N > 0
S norm normalized input space
S qa signal space after adaptive quantization
edges i i-th quantization edge, see (A3)
srt _ samples i i-th sample in the ascending sorted array of all available signal samples

References

  1. Wielgosz, M.; Skoczeń, A.; Wiatr, K. Looking for a Correct Solution of Anomaly Detection in the LHC Machine Protection System. In Proceedings of the 2018 International Conference on Signals and Electronic Systems (ICSES), Kraków, Poland, 10–12 September 2018; pp. 257–262. [Google Scholar]
  2. Evans, L.; Bryant, P. LHC Machine. J. Instrum. 2008, 3, S08001. [Google Scholar] [CrossRef]
  3. Denz, R. Electronic Systems for the Protection of Superconducting Elements in the LHC. IEEE Trans. Appl. Supercond. 2006, 16, 1725–1728. [Google Scholar] [CrossRef] [Green Version]
  4. Steckert, J.; Skoczen, A. Design of FPGA-based Radiation Tolerant Quench Detectors for LHC. J. Instrum. 2017, 12, T04005. [Google Scholar] [CrossRef]
  5. Chandola, V.; Mithal, V.; Kumar, V. Comparative Evaluation of Anomaly Detection Techniques for Sequence Data. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 743–748. [Google Scholar]
  6. Chandola, V.; Banerjee, A.; Kumar, V. Anomaly Detection: A Survey. ACM Comput. Surv. 2009, 41, 15. [Google Scholar] [CrossRef]
  7. Pimentel, M.A.; Clifton, D.A.; Clifton, L.; Tarassenko, L. A Review of Novelty Detection. Signal Process. 2014, 99, 215–249. [Google Scholar] [CrossRef]
  8. Graves, A. Supervised Sequence Labelling with Recurrent Neural Networks; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  9. Morton, J.; Wheeler, T.A.; Kochenderfer, M.J. Analysis of Recurrent Neural Networks for Probabilistic Modelling of Driver Behaviour. IEEE Trans. Intell. Transp. Syst. 2016, 18, 1289–1298. [Google Scholar] [CrossRef]
  10. Pouladi, F.; Salehinejad, H.; Gilani, A.M. Recurrent Neural Networks for Sequential Phenotype Prediction in Genomics. In Proceedings of the 2015 International Conference on Developments of E-Systems Engineering (DeSE), Duai, UAE, 13–14 December 2015; pp. 225–230. [Google Scholar]
  11. Chen, X.; Liu, X.; Wang, Y.; Gales, M.J.F.; Woodland, P.C. Efficient Training and Evaluation of Recurrent Neural Network Language Models for Automatic Speech Recognition. IEEE Trans. Audio Speech Lang. Process. 2016, 24, 2146–2157. [Google Scholar] [CrossRef]
  12. Ma, J.; Perkins, S. Time-series Novelty Detection Using One-Class Support Vector Machines. In Proceedings of the International Joint Conference on Neural Networks, Portland, OR, USA, 20–24 July 2003; Volume 3, pp. 1741–1745. [Google Scholar]
  13. Zhang, R.; Zhang, S.; Muthuraman, S.; Jiang, J. One Class Support Vector Machine for Anomaly Detection in the Communication Network Performance Data. In Proceedings of the 5th Conference on Applied Electromagnetics, Wireless and Optical Communications; World Scientific and Engineering Academy and Society (WSEAS) ELECTROSCIENCE’07, Stevens Point, WI, USA, 14–16 December 2007; pp. 31–37. [Google Scholar]
  14. Su, J.; Long, Y.; Qiu, X.; Li, S.; Liu, D. Anomaly Detection of Single Sensors Using OCSVM_KNN. In Proceedings of the Big Data Computing and Communications: First International Conference, BigCom 2015, Taiyuan, China, 1–3 August 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 217–230. [Google Scholar]
  15. Ruiz-Gonzalez, R.; Gomez-Gil, J.; Gomez-Gil, F.J.; Martínez-Martínez, V. An SVM-Based Classifier for Estimating the State of Various Rotating Components in Agro-Industrial Machinery with a Vibration Signal Acquired from a Single Point on the Machine Chassis. Sensors 2014, 14, 20713–20735. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Hornero, R.; Escudero, J.; Fernández, A.; Poza, J.; Gómez, C. Spectral and Nonlinear Analyses of MEG Background Activity in Patients With Alzheimer’s Disease. IEEE Trans. Biomed. Eng. 2008, 55, 1658–1665. [Google Scholar] [CrossRef] [PubMed]
  17. Schölkopf, B.; Williamson, R.C.; Smola, A.J.; Shawe-Taylor, J.; Platt, J.C. Support Vector Method for Novelty Detection. In Proceedings of the 12th International Conference on Neural Information Processing Systems, Denver, CO, USA, 29 November–4 December 2000; pp. 582–588. [Google Scholar]
  18. Schölkopf, B.; Platt, J.C.; Shawe-Taylor, J.C.; Smola, A.J.; Williamson, R.C. Estimating the Support of a High-Dimensional Distribution. Neural Comput. 2001, 13, 1443–1471. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  19. Masnan, M.J.; Mahat, N.I.; Shakaff, A.Y.M.; Abdullah, A.H.; Zakaria, N.Z.I.; Yusuf, N.; Subari, N.; Zakaria, A.; Aziz, A.H.A. Understanding Mahalanobis distance criterion for feature selection. AIP Conf. Proc. 2015, 1660, 050075. [Google Scholar] [CrossRef]
  20. Liu, F.T.; Ting, K.M.; Zhou, Z.H. Isolation Forest. In Proceedings of the 2008 Eighth IEEE International Conference on Data Mining, Pisa, Italy, 15–19 December 2008; pp. 413–422. [Google Scholar]
  21. Wang, C.; Viswanathan, K.; Choudur, L.; Talwar, V.; Satterfield, W.; Schwan, K. Statistical Techniques for Online Anomaly Detection in Data Centers. In Proceedings of the 12th IFIP/IEEE International Symposium on Integrated Network Management (IM 2011) and Workshops, Dublin, Ireland, 23–27 May 2011; pp. 385–392. [Google Scholar]
  22. Ekberg, J.; Ylinen, J.; Loula, P. Network behaviour anomaly detection using Holt-Winters algorithm. In Proceedings of the 2011 International Conference for Internet Technology and Secured Transactions, Abu Dhabi, UAE, 11–14 December 2011; pp. 627–631. [Google Scholar]
  23. Wielgosz, M.; Skoczeń, A. Recurrent Neural Networks with Grid Data Quantization for Modeling LHC Superconducting Magnets Behaviour. In Contemporary Computational Science; Kulczycki, P., Kowalski, P.A., Łukasik, S., Eds.; AGH University of Science and Technology: Kraków, Poland, 2018. [Google Scholar]
  24. Wielgosz, M.; Skoczeń, A.; Mertik, M. Using LSTM recurrent neural networks for detecting anomalous behavior of LHC superconducting magnets. Nuclear Inst. Methods Phys. Res. A 2017, 867, 40–50. [Google Scholar] [CrossRef]
  25. Wielgosz, M.; Mertik, M.; Skoczeń, A.; Matteis, E.D. The model of an anomaly detector for HiLumi LHC magnets based on Recurrent Neural Networks and adaptive quantization. Eng. Appl. Artif. Intell. 2018, 74, 166–185. [Google Scholar] [CrossRef]
  26. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  27. He, Q.; Wen, H.; Zhou, S.; Wu, Y.; Yao, C.; Zhou, X.; Zou, Y. Effective Quantization Methods for Recurrent Neural Networks. arXiv, 2016; arXiv:1611.10176. [Google Scholar]
  28. Shin, S.; Hwang, K.; Sung, W. Fixed-Point Performance Analysis of Recurrent Neural Networks. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 976–980. [Google Scholar]
  29. Chollet, F. Keras. 2015. Available online: https://keras.io (accessed on 10 November 2018).
  30. Theano Development Team. Theano: A Python Framework for Fast Computation of Mathematical Expressions. 2016. Available online: http://deeplearning.net/software/theano/ (accessed on 10 November 2018).
  31. Abadi, M. TensorFlow: Large-Scale Machine Learning on Heterogeneous Systems. 2015. Available online: https://tensorflow.org (accessed on 10 November 2018).
  32. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  33. Sze, V.; Chen, Y.H.; Emer, J.; Suleiman, A.; Zhang, Z. Hardware for Machine Learning: Challenges and Opportunities. In Proceedings of the 2017 IEEE Custom Integrated Circuits Conference (CICC), Austin, TX, USA, 30 April–3 May 2017. [Google Scholar]
  34. Wu, J.; Leng, C.; Wang, Y.; Hu, Q.; Cheng, J. Quantized Convolutional Neural Networks for Mobile Devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 4820–4828. [Google Scholar]
Figure 1. The LHC tunnel. The blue cryostat contains superconducting main dipole magnets. The protection unit is visible on the floor under the magnet (yellow rack). The photo taken by A.S. in 2007.
Figure 1. The LHC tunnel. The blue cryostat contains superconducting main dipole magnets. The protection unit is visible on the floor under the magnet (yellow rack). The photo taken by A.S. in 2007.
Sensors 18 03933 g001
Figure 2. The presentation of a contents of U RES field of two PM data files for one of the superconducting magnets (with electrical current 600 A ). The voltage range of the ADC is from 256 mV to 256 mV . Time 0 ms refers to trigger (request) time stored in the field QUENCHTIME in the PM data files.
Figure 2. The presentation of a contents of U RES field of two PM data files for one of the superconducting magnets (with electrical current 600 A ). The voltage range of the ADC is from 256 mV to 256 mV . Time 0 ms refers to trigger (request) time stored in the field QUENCHTIME in the PM data files.
Sensors 18 03933 g002
Figure 3. Test facilities of SM18 for testing MQXFS inner triplet quadrupole magnet, including the rack used for the data acquisition and tests (photos provided by E.M.).
Figure 3. Test facilities of SM18 for testing MQXFS inner triplet quadrupole magnet, including the rack used for the data acquisition and tests (photos provided by E.M.).
Sensors 18 03933 g003
Figure 4. Samples per bin for PM dataset U RES channel ( m = 32 ). Note the logarithmic scale.
Figure 4. Samples per bin for PM dataset U RES channel ( m = 32 ). Note the logarithmic scale.
Sensors 18 03933 g004
Figure 5. Full (a) and zoomed-in (b) bin edges for PM dataset U RES channel ( m = 32 ). Please note that adaptive quantization algorithm effectively yields only 10 bins, since some edges values occur multiple times.
Figure 5. Full (a) and zoomed-in (b) bin edges for PM dataset U RES channel ( m = 32 ). Please note that adaptive quantization algorithm effectively yields only 10 bins, since some edges values occur multiple times.
Sensors 18 03933 g005
Figure 6. High-level system architecture. © 2018 IEEE. Reprinted, with permission, from Wielgosz, M.; Skoczeń, A.; Wiatr, K. Looking for a Correct Solution of Anomaly Detection in the LHC Machine Protection System. 2018 International Conference on Signals and Electronic Systems (ICSES), 2018, pp. 257–262 [1].
Figure 6. High-level system architecture. © 2018 IEEE. Reprinted, with permission, from Wielgosz, M.; Skoczeń, A.; Wiatr, K. Looking for a Correct Solution of Anomaly Detection in the LHC Machine Protection System. 2018 International Conference on Signals and Electronic Systems (ICSES), 2018, pp. 257–262 [1].
Sensors 18 03933 g006
Figure 7. Design flow for hardware implementation. © 2018 IEEE. Reprinted, with permission, from Wielgosz, M.; Skoczeń, A.; Wiatr, K. Looking for a Correct Solution of Anomaly Detection in the LHC Machine Protection System. 2018 International Conference on Signals and Electronic Systems (ICSES), 2018, pp. 257–262 [1].
Figure 7. Design flow for hardware implementation. © 2018 IEEE. Reprinted, with permission, from Wielgosz, M.; Skoczeń, A.; Wiatr, K. Looking for a Correct Solution of Anomaly Detection in the LHC Machine Protection System. 2018 International Conference on Signals and Electronic Systems (ICSES), 2018, pp. 257–262 [1].
Sensors 18 03933 g007
Figure 8. Proposed system. © 2018 IEEE. Reprinted, with permission, from Wielgosz, M.; Skoczeń, A.; Wiatr, K. Looking for a Correct Solution of Anomaly Detection in the LHC Machine Protection System. 2018 International Conference on Signals and Electronic Systems (ICSES), 2018, pp. 257–262 [1].
Figure 8. Proposed system. © 2018 IEEE. Reprinted, with permission, from Wielgosz, M.; Skoczeń, A.; Wiatr, K. Looking for a Correct Solution of Anomaly Detection in the LHC Machine Protection System. 2018 International Conference on Signals and Electronic Systems (ICSES), 2018, pp. 257–262 [1].
Sensors 18 03933 g008
Figure 9. Example single series results visualization ( in _ grid = 32 , in _ algorithm = adaptive , look _ back = 256 ). Red line across all subplots marks the QUENCHTIME and gray spans indicate the anomalies found by the system.
Figure 9. Example single series results visualization ( in _ grid = 32 , in _ algorithm = adaptive , look _ back = 256 ). Red line across all subplots marks the QUENCHTIME and gray spans indicate the anomalies found by the system.
Sensors 18 03933 g009
Figure 10. Example single series results visualization ( in _ grid = 32 , in _ algorithm = recursive _ adaptive , look _ back = 128 ). Red line across all subplots marks the QUENCHTIME and gray spans indicate the anomalies found by the system.
Figure 10. Example single series results visualization ( in _ grid = 32 , in _ algorithm = recursive _ adaptive , look _ back = 128 ). Red line across all subplots marks the QUENCHTIME and gray spans indicate the anomalies found by the system.
Sensors 18 03933 g010
Figure 11. Grid use for various in _ grid values and in _ algorithm . ad—adaptive, ra—recursive_adaptive, ca—cumulative_amplitude.
Figure 11. Grid use for various in _ grid values and in _ algorithm . ad—adaptive, ra—recursive_adaptive, ca—cumulative_amplitude.
Sensors 18 03933 g011
Figure 12. F 1 score as a function of look _ back for several in _ grid and in _ algorithm values. Dashed line shows Random baseline model performance for the same look _ back .
Figure 12. F 1 score as a function of look _ back for several in _ grid and in _ algorithm values. Dashed line shows Random baseline model performance for the same look _ back .
Sensors 18 03933 g012
Figure 13. The ROC curve for algorithm adaptive ( in _ algorithm = adaptive , in _ grid = 32 , GRU (two layers, 64 and 32 cells) + Dense).
Figure 13. The ROC curve for algorithm adaptive ( in _ algorithm = adaptive , in _ grid = 32 , GRU (two layers, 64 and 32 cells) + Dense).
Sensors 18 03933 g013
Figure 14. The ROC curve for algorithm recursive_adaptive ( in _ algorithm = recursive _ adaptive , in _ grid = 32 , GRU (two layers, 64 and 32 cells) + Dense).
Figure 14. The ROC curve for algorithm recursive_adaptive ( in _ algorithm = recursive _ adaptive , in _ grid = 32 , GRU (two layers, 64 and 32 cells) + Dense).
Sensors 18 03933 g014
Figure 15. The ROC curve for algorithm cumulative_amplitude ( in _ algorithm = cumulative _ amplitude , in _ grid = 32 , GRU (two layers, 64 and 32 cells) + Dense).
Figure 15. The ROC curve for algorithm cumulative_amplitude ( in _ algorithm = cumulative _ amplitude , in _ grid = 32 , GRU (two layers, 64 and 32 cells) + Dense).
Sensors 18 03933 g015
Table 1. The parameters of NN built with GRU cells for three different algorithms (two layers, 64 and 32 cells + Dense, in _ grid = 32 ).
Table 1. The parameters of NN built with GRU cells for three different algorithms (two layers, 64 and 32 cells + Dense, in _ grid = 32 ).
In _ Algorithm Look _ Back Accuracy F 1 Score F 2 Score
adaptive160.84620.67220.6167
320.85060.70310.6687
640.86110.73760.7124
1280.88380.79730.7835
2560.91620.87430.8796
5120.95430.94740.9522
recursive_adaptive160.85070.69200.6481
320.85430.70220.6561
640.86520.73500.6928
1280.88680.80400.7939
2560.91720.87460.8749
5120.95710.95060.9560
cumulative_amplitude160.84360.66090.5999
320.84730.66640.5968
640.85620.71150.6620
1280.88530.79270.7622
2560.92310.88300.8805
5120.96690.96250.9779
Table 2. Testing accuracy (20% of dataset). All models were run with in _ grid = 32 and look _ back = 256 , using single input channel ( U RES ), NNs were trained for 7 epochs. The best result is marked in bold.
Table 2. Testing accuracy (20% of dataset). All models were run with in _ grid = 32 and look _ back = 256 , using single input channel ( U RES ), NNs were trained for 7 epochs. The best result is marked in bold.
Model In _ Algorithm
AdaptiveRecursive_AdaptiveCumulative_Amplitude
Random (stratified)0.63340.63340.6334
Elliptic Envelope0.67000.77750.6700
Isolation Forest0.79470.75960.8094
OC-SVM (RBF kernel)0.33000.82320.3300
OC-SVM (linear kernel)0.29590.78810.2528
GRU (two layers, 64 and 32 cells)0.89280.90050.8842
LSTM (two layers, 64 and 32 cells)0.82710.85520.7402
Table 3. Testing accuracy (20% of dataset). Models were run with in _ grid = 32 and look _ back = 256 , using four input channels ( U RES , U DIFF , I DCCT , I DIDT ), NNs were trained for 7 epochs. The best result is marked in bold.
Table 3. Testing accuracy (20% of dataset). Models were run with in _ grid = 32 and look _ back = 256 , using four input channels ( U RES , U DIFF , I DCCT , I DIDT ), NNs were trained for 7 epochs. The best result is marked in bold.
Model In _ Algorithm
AdaptiveRecursive_AdaptiveCumulative_Amplitude
GRU (two layers, 64 and 32 cells)0.92350.93000.8842
LSTM (two layers, 64 and 32 cells)0.91940.90920.9023
Table 4. Coefficients Quantization Results for GRU (two layers, 64 and 32 cells) + Dense, trained on four input channels. Accuracy as a function of bit-width.
Table 4. Coefficients Quantization Results for GRU (two layers, 64 and 32 cells) + Dense, trained on four input channels. Accuracy as a function of bit-width.
BitsMethod In _ Algorithm
AdaptiveRecursive_AdaptiveCumulative_Amplitude
Original Model 0.92350.93000.8842
10linear0.92360.92870.8841
minmax0.92330.93000.8841
log_minmax0.92350.92980.8842
tanh0.92320.92830.9232
9linear0.92360.92790.8838
minmax0.92370.92950.8842
log_minmax0.92310.92930.8843
tanh0.92190.92600.8842
8linear0.92060.92570.8830
minmax0.92380.93110.8838
log_minmax0.92070.92830.8844
tanh0.91610.91430.8836
7linear0.91770.39890.8850
minmax0.91940.92500.8841
log_minmax0.92180.92360.8833
tanh0.91310.90330.8851
6linear0.89520.90080.8871
minmax0.91440.88390.8842
log_minmax0.91110.90760.8844
tanh0.87020.87820.8788
5linear0.37220.84420.8802
minmax0.90310.90580.8810
log_minmax0.39480.88780.8812
tanh0.82470.33060.8670
4linear0.85000.27450.8587
minmax0.86780.87020.8775
log_minmax0.86490.38480.8734
tanh0.74910.84640.3017
3linear0.79280.81350.8190
minmax0.33910.79000.8530
log_minmax0.76640.80230.8564
tanh0.69220.28330.7985
2linear0.30060.67000.7065
minmax0.73710.33910.3466
log_minmax0.79080.73690.3110
tanh0.72160.75490.2309
1linear0.67000.33000.3300
minmax0.67060.70030.6717
log_minmax0.71710.74590.2121
tanh0.71710.74590.2121

Share and Cite

MDPI and ACS Style

Wielgosz, M.; Skoczeń, A.; De Matteis, E. Protection of Superconducting Industrial Machinery Using RNN-Based Anomaly Detection for Implementation in Smart Sensor. Sensors 2018, 18, 3933. https://doi.org/10.3390/s18113933

AMA Style

Wielgosz M, Skoczeń A, De Matteis E. Protection of Superconducting Industrial Machinery Using RNN-Based Anomaly Detection for Implementation in Smart Sensor. Sensors. 2018; 18(11):3933. https://doi.org/10.3390/s18113933

Chicago/Turabian Style

Wielgosz, Maciej, Andrzej Skoczeń, and Ernesto De Matteis. 2018. "Protection of Superconducting Industrial Machinery Using RNN-Based Anomaly Detection for Implementation in Smart Sensor" Sensors 18, no. 11: 3933. https://doi.org/10.3390/s18113933

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop