Next Article in Journal
Spectrum Sensing Method Based on Information Geometry and Deep Neural Network
Previous Article in Journal
On Unitary t-Designs from Relaxed Seeds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Concept Paper

Introduction to Extreme Seeking Entropy

1
Department of Computing and Control Engineering, Faculty of Chemical Engineering, University of Chemistry and Technology, 166 28 Prague, Czech Republic
2
Department of Process Control, Faculty of Electrical Engineering and Informatics, University of Pardubice, 530 02 Pardubice, Czech Republic
*
Authors to whom correspondence should be addressed.
Entropy 2020, 22(1), 93; https://doi.org/10.3390/e22010093
Submission received: 9 December 2019 / Revised: 31 December 2019 / Accepted: 8 January 2020 / Published: 12 January 2020

Abstract

:
Recently, the concept of evaluating an unusually large learning effort of an adaptive system to detect novelties in the observed data was introduced. The present paper introduces a new measure of the learning effort of an adaptive system. The proposed method also uses adaptable parameters. Instead of a multi-scale enhanced approach, the generalized Pareto distribution is employed to estimate the probability of unusual updates, as well as for detecting novelties. This measure was successfully tested in various scenarios with (i) synthetic data, (ii) real time series datasets, and multiple adaptive filters and learning algorithms. The results of these experiments are presented.

1. Introduction

Novelty detection (ND) plays an important role in signal processing. Many research groups have dealt with both the methods and applications because there are many complex tasks where accurate ND is needed. However, the success of this method depends on the type of data, so the current methods usually give good performance and results only for specific datasets. As more data are being analyzed currently, there is a greater need for new methods of ND. Furthermore, the increasing computational power provides more possibilities and methods that were not possible to use a few decades ago, but can now be performed for real-time tasks easily. For these reasons, we consider the topic of ND to be vital.
Two different approaches have been established over the last few decades. The first approach is based on the statistical features of the data [1], and some methods also use extreme value theory to estimate the novelty of the data [2,3,4,5]. The second approach uses learning systems [6,7,8]: the attributes of a learning system are used to obtain information about novelties in the data. Over the last decade, many new methods have been proposed in the field of machine learning [9]. The set membership algorithm [10,11,12] uses the prediction error for better accuracy, reducing the computational resources required and assuring a greater robustness with the proper filter, especially for data without drift. Bukovsky et al. proved that the learning effort of a learning system can be used to estimate a measure of the novelty for each data point [13,14], but a shortcoming of that method is that it is hard to interpret the ND score. A similar approach, combining the prediction error with adaptive weight increments, was proposed in [15]. That method also lacks the possibility of a meaningful interpretation of the ND score. It was also already shown that the accuracy of the learning system is not necessarily correlated with the accuracy of the ND [16] and that simple predictors are useful even for signals that are produced by complex systems (e.g., EEG, ECG).
ND brings a new point of view to complex signal analysis. Research groups have started dealing with the early diagnosis of different diseases where ND plays an important role. Taoum et al. presented ND and data fusion methods to identify acute respiratory problems [17]. Rad introduced ND for gait and movement monitoring to diagnosis Parkinson’s disease and autism spectrum disorders [18]. Burlina used ND algorithms in the diagnosis of different muscle diseases [19].
Other fields where ND can be found are information and mechanical engineering. Hu introduced ND as an appropriate tool for monitoring the health of mechanical systems, where it is usually impossible to know every potential fault [20]. Surace described the application of ND to the simulation of an offshore steel platform [21].
In this article, a new method for ND is introduced. The proposed method combines both a statistics based approach and a learning systems based approach. The changes of the adaptive parameters of the learning system obtained via an incremental learning algorithm are evaluated. A new measure, called extreme seeking entropy, is then estimated. It is shown that the proposed measure corresponds to different types of novelties in various datasets and how it may be useful for diagnostics and failure detection tasks. It also outperforms the other unsupervised adaptive ND methods.
This paper is organized as follows. Section 2 describes the specifications of the learning system and learning algorithm used during the experiments. Section 3 recalls the learning entropy algorithm and an error and learning based novelty detection method. Then, the general suitable properties of learning based information are discussed. Section 4 introduces the new measure of novelty, and the ND algorithm based on this measure is presented. Section 5 describes a case study where both synthetic and real datasets are used to show the usability of the proposed algorithm and also contains the rationale behind the selection of the experiments. Section 6 contains the rate detection of the proposed algorithm in two cases, namely detection of a change in the trend and the detection of a step change of a signal generator. The last two sections are dedicated to limitations and further challenges, Section 7, and then our conclusions, Section 8.

2. Review of the Learning Systems Used

All the supervised learning systems used in the experimental analysis are introduced in this section. In general, assume that the output of the learning system is a function of weights and the input data:
y = f ( w , x )
where y R denotes the output, w R n is the vector of its adaptable parameters, x R n is a vector that contains the input data, and f is the mapping function that maps the input data and weights to the output. The following adaptation is done in order to minimize the error:
e ( k ) = d ( k ) y ( k ) ,
where k is a discrete time index and d ( k ) R is the target of the supervised learning system (the desired output). The update of the weights w is done with every new sample as follows:
w ( k + 1 ) = w ( k ) + Δ w ( k )
where Δ w R n is a vector that contains the updates of the adaptive parameters. This update depends on the learning algorithm used. The learning algorithms will be discussed later.

2.1. Adaptive Models

The adaptive models used during the experiments are described briefly in this section.

2.1.1. Linear Adaptive Filter

One of the simplest adaptive models is the linear adaptive filter, also known as the linear neural unit (LNU), with finite impulse response (FIR). The output of this model at a discrete time index k is given by:
y ( k ) = i = 1 n w i · x i ( k ) ,
which is equivalent to
y ( k ) = w T ( k ) · x ( k ) ,
where w T ( k ) = [ w 1 ( k ) , w 2 ( k ) , , w n ( k ) ] R n is the row vector of adaptive weights and x T ( k ) = [ x 1 ( k ) , x 2 ( k ) , , x n ( k ) ] R n is the column input vector. The vector of adaptive weights is updated with every new sample obtained, and the size of the update depends on the learning algorithm used. In general, x may contain the history of a single input or even the history of multiple inputs.

2.1.2. An Adaptive Filter Based on Higher Order Neural Units

The quadratic neural unit (QNU) [22,23,24] (also known as a second order neural unit) is a non-linear predictive model. The output of the QNU is:
y ( k ) = i = 0 n j = i n w i , j · x i ( k ) · x j ( k )
where often, x 0 = 1 . This is equivalent to:
y ( k ) = w · colx ,
where the column input vector colx for n inputs has the general form:
colx = [ 1 , x 1 , , x n , x 1 2 , x 1 · x 2 , , x 1 · x n , x 2 2 , x 2 · x 3 , , x 2 · x n , , , x n 1 · x n , x n 2 ] T
and w is a row vector of adaptive weights that has the same length as colx . Note that the first term in colx , x 0 = 1 , should be used when the data have a non-zero offset.

2.2. Learning Algorithms

To prove the generality of the adaptive weight evaluation approach for novelty detection, different learning algorithms have been tested. Both algorithms are heavily used in signal processing and machine learning.

2.2.1. Normalized Least Mean Squares Algorithm

The normalized least mean squares (NLMS) algorithm [25] is a variant of the least mean squares algorithm. The problem with the selection of the learning rate is solved by normalizing by the power of the input. It is a stochastic gradient approach. The update of this adaptive algorithm is given by:
Δ w ( k ) = μ · x ( k ) · e ( k ) ϵ + x T ( k ) · x ( k ) ,
where ϵ R is a small positive constant used to avoid division by zero, μ R is the learning rate, and e R is the error defined as in (2). According to the normalization of the learning rate shown in (9), it is necessary to choose a learning rate μ satisfying 0 μ 2 to preserve the stability of the NLMS algorithm.

2.2.2. Generalized Normalized Gradient Descent

The generalized normalized gradient descent (GNGD) [26] algorithm is another algorithm for linear adaptive FIR filters. Due to its adaptation of the learning rate based on the signal dynamics, it converges in places where the NLMS algorithm diverges. The update of this adaptive algorithm is given by:
Δ w ( k ) = η ( k ) e ( k ) x ( k )
with:
η ( k ) = μ x ( k ) 2 2 + ϵ ( k )
ϵ ( k ) = ϵ ( k 1 ) ρ μ e ( k ) e ( k 1 ) x T ( k ) x ( k 1 ) ( x ( k 1 ) 2 2 + ϵ ( k 1 ) ) 2
where η R is the adaptive learning rate, ϵ R is a compensation term, and ρ is the step size adaptation parameter, which should be chosen so as to satisfy 0 ρ 1 .

3. On the Evaluation of the Increments in the Adaptive Weights in Order to Estimate the Novelty in the Data

This section recalls two ND methods that evaluate the increments in the adaptive weights, namely learning entropy, and error and learning based novelty detection. Those methods are compared with the proposed algorithm in Section 4 and Section 6. Then, the general properties of the learning based information measure will be discussed.

3.1. Learning Entropy: A Direct Algorithm

The recent publication on Learning Entropy [14] specifies a direct algorithm to estimate the learning entropy ( L E ) as follows.
L E ( k ) = i = 1 n w z ( | Δ w i ( k ) | )
Here, z is a special Z-score, given as follows:
z ( | Δ w i ( k ) | ) = | Δ w i ( k ) | | Δ w i M ( k 1 ) | ¯ σ ( | Δ w i M ( k 1 ) | )
where | Δ w i M ( k 1 ) | ¯ is the mean of the last M increments of w i , σ ( | Δ w i M ( k 1 ) | ) is their standard deviation, and n w is the number of adaptive weights. According to Equation (15), the function f in this case corresponds to the special Z-score function z, and the function A is represented by the sum over the adaptive weights.

3.2. Error and Learning Based Novelty Detection

Another recently published method that evaluates the increments of the adaptive weight together with the prediction error is ELBND [15]. ELBND describes every sample with the value obtained as follows:
E L B N D ( k ) = max 1 i n w | Δ w i ( k ) · e ( k ) |
or, alternatively,
E L B N D ( k ) = i = 1 n w | Δ w i ( k ) · e ( k ) | .
In this case, the function f is represented by multiplying the ith adaptive weight increment Δ w i by the prediction error e. The function A is the maximum of the vector in the case of E L B N D given by Equation (13) and the sum over the weights in the case of the E L B N D given by Equation (14).

3.3. General Properties of A Suitable Learning Based Information Measure

Learning entropy was proposed in [13,14]. It is a learning based information measure L that, in general, evaluates unusually large learning increments, as follows:
L ( k ) = A ( f ( Δ w ( k ) ) )
where A is a general aggregation function and f is a function that quantifies the irregularity in the learning effort [14].
Another form for f and A will be presented in the present paper. Firstly, the function f is presented.
Assume that the value of f should be high when the increments Δ w are unusually high. Furthermore, this function also takes the history of those increments as input. As stated, some cumulative distribution function of each weight increment seems suitable. This cumulative distribution function (cdf) is discussed later in this paper. The question is how to deal with the aggregation function A. Under the assumption that each weight is independent of the others, it is possible to choose the aggregation function A as follows:
A ( f ( Δ w ( k ) ) = log i = 1 n ( 1 f c d f i ( | Δ w i ( k ) | ) ) .
The function A in the stated form is high for high cdf values of the weight updates, and hence for the values where the cdf is close to one. The function 1 f c d f i can be viewed as the complementary cumulative distribution function (or the survival function, also known as the reliability function). This approach clearly avoids the need for a multi-scale approach. The result is that much fewer parameters are needed for detecting potential novelties. Only the crucial choice of the cdf remains. In the next section, a suitable probability distribution will be presented, together with the new novelty detection algorithm.

4. Extreme Seeking Entropy

4.1. The Generalized Pareto Distribution

A normal distribution is used in some novelty detection algorithms [27,28,29]. However, the normal distribution cannot always be used, especially when the description of the data by a mean and a symmetric range of variation would be misleading [30]. Let us mention the Pickands–Balkema–de Haan theorem [31,32], which states that if we have a sequence X 1 , X 2 , of independent and identically distributed random variables and F u is their conditional excess distribution function (over the threshold u), then:
F u ( x ) G P D ( ξ , μ , σ ) ( x ) , as u
where G P D is the generalized Pareto distribution and F u is defined by:
F u ( x ) = P ( X u x , X > u ) = F ( u + x ) F ( x ) 1 F ( x )
for 0 x x F u , where x F is the right endpoint of the underlying unknown distribution F. The probability density function of the GPDtakes the form:
f ( ξ , μ , σ ) ( x ) = 1 σ 1 + ξ ( x μ ) σ 1 ξ 1 for ξ 0 , exp x μ σ for ξ = 0 .
where in general, μ ( , + ) is a location parameter, σ ( 0 , ) is the scale, and ξ ( , ) is a shape parameter. The corresponding cumulative distribution function then takes the form:
F ( ξ , μ , σ ) ( x ) = 1 1 + ξ ( x μ ) σ 1 ξ for ξ 0 , 1 exp x μ σ for ξ = 0 .
Note that the support is x μ if ξ 0 , and μ x μ σ / ξ if ξ < 0 where μ R , σ > 0 , and ξ R . In Figure 1, we show the ability of the GPD to deal with many possible shapes of the tails of the distributions. Note that if ξ = 1 , it is equivalent to the uniform distribution; if ξ = 0 , it is equivalent to the exponential distribution; if ξ = 0.5 , it is the triangular distribution; if 0.5 < ξ < 0 , it is a light tailed distribution (e.g., the normal distribution or the Gumbel distribution); if ξ > 0 , it is a heavy tailed distribution (e.g., the Pareto distribution, the log-normal distribution, or Student’s t-distribution); and if ξ < 1 , it is a monotonically increasing distribution with compact support (e.g., the beta distribution).
As long as we do not know the distribution of increments of the adaptive weights, it is appropriate to use the GPD due to its universality in modeling the tails of other distributions [33,34,35]. As the aim is to evaluate unusually high increments of an adaptive system, the need for some threshold arises: denote this threshold by z. This threshold should divide the weight increments into two sets. An increment that is lower than the threshold should belong to the set that contains the usual high increments; denote this by L. However, an increment that is greater than or equal to this threshold should belong to the set H. Assume that both sets exist for every adaptable parameter, so for the ith adaptable parameter w i , we should set a threshold z i so the weight updates belong to the sets as follows.
| Δ w i | < z i L i
| Δ w i | z i H i
The increments belonging to L i will be unlikely to contain any information about a novelty in the adaptation, so we are not going to evaluate them. The set H i should contain the weight increments that are drawn from the GPD if the choice of the threshold was appropriate. The threshold z i depends on the method chosen, peaks over threshold, which will be discussed in the following subsection.

4.2. The Peaks over Threshold Method

The main issue in GPD fitting is the estimation of a suitable threshold, z. If the threshold is too high (i.e., there are only a few points that exceed it), then the parameters of the GPD suffer a high variance. If the threshold is too low, then the GPD approximation is not reliable. Therefore, the proper choice of threshold is crucial for the performance of the ND algorithm. There are many approaches to estimating the threshold [36]. To show the usability of the proposed ND algorithm, multiple rules of thumb [37,38,39] for the choice of the threshold have been used. Let l be the number of samples used for the GPD fitting and n s be the total number of samples available:
l 1 = 0.1 · n s
l 2 = n s
l 3 = n s 2 3 l o g ( l o g ( n s ) )
Note that we use the highest adaptive weight increment to estimate the GPD parameters. The peaks over threshold (POT) method is crucial for deciding whether | Δ w k ( k ) | belongs to H i or to L i . In Section 5 are presented the results with different techniques of choosing the threshold.

4.3. Extreme Seeking Entropy Algorithm

In this subsection, the new novelty measure and the new novelty detection algorithm are presented. We will introduce the extreme seeking entropy measure, which is given as follows:
E S E ( | Δ w ( k ) | ) = log i = 1 n ( 1 f c d f i ( | Δ w i ( k ) | ) )
where:
f c d f i ( | Δ w i ( k ) | ) = 0 , | Δ w i ( k ) | L i F ( ξ i , μ i , σ i ) ( | Δ w i ( k ) | ) , | Δ w i ( k ) | H i .
The proposed algorithm evaluates the value of E S E for every newly obtained weight increment. Note that if the weight increment is smaller than the threshold from the POT method, the addition to the novelty measure E S E is zero. Small probability increments, which are highly likely to contain a novelty, have a high value of E S E . To estimate the parameters of the GPD pdf, it is possible to process all available history samples, or only the n s newest samples, with the POT method. The proposed algorithm is described by the following pseudocode (Algorithm 1).
Algorithm 1 Extreme seeking entropy algorithm.
1:
set n s , and choose the POT method
2:
initial estimation of the parameters of the GPD: ξ i , μ i , σ i for each adaptable parameter
3:
for each new d ( k ) do
4:
 update the adaptive model to get Δ w ( k )
5:
 proceed with the POT method
6:
if | Δ w i | ( k ) H i then
7:
  update the parameters ξ i , μ i , σ i
8:
end if
9:
 compute E S E according to (26)
10:
end for
The proposed ND algorithm needs only one parameter to be set, which avoids the need for a multi-scale approach and overcomes the issues arising from setting multiple parameters. The parameter n s can also take all available samples, if needed. Furthermore, there is the need to choose the proper POT method. Choosing the POT method depends strongly on the nature of the data. The limitation of the proposed method is the need for an initial estimate of the parameters of the GPD. We need a priori information about ξ , σ , and μ for each adaptive weight. If there are n w adaptive weights, then we need 3 · n w parameters to start the extreme seeking entropy algorithm. If there is no a priori information about the parameters, we need at least n s samples to obtain the first results. Another problem may arise if the type of underlying unknown distribution F or its parameters are significantly varying in time.

5. Experimental Results

5.1. The Design of the Experiments

The proposed E S E algorithm was studied in various testing schemes with synthetic data and with one real dataset. For each experiment, we also show the results of the ELBND and LE methods, for the sake of comparison. The parameter M that specifies the number of increments for the L E evaluation was set as M = n s in all experiments. The first experiment was the detection of perturbed data in the Mackey–Glass time series. This experiment was chosen due to the possibility of comparing it with the results published in [13]. The second experiment, with synthetic data, showed the ability of the ESE algorithm to detect a change in the standard deviation of the noise in a random data stream, which can be viewed as a novelty in the data. It was inspired by a problem that arises in hybrid navigation systems that use both GPS and dead-reckoning sensors [40]. The third experiment, involving a step change in the parameters of a signal generator, was an analogue to a problem that may arise in evaluating multiple stream random number generators [41], where we may detect and evaluate the probability of changes in the parameters of those generators. The fourth experiment was the detection of the disappearance of noise. This experiment was chosen as neither of the compared methods (LE, ELBND) were able to deal with this problem, where the disappearance of the noise could be also viewed as a novelty in the signal. The fifth experiment was the detection of a change in trend; this is a common problem in fault detection and diagnosis [42]. The last experiment was performed on the mouse EEG dataset. The aim of this experiment was to show that the proposed ESE algorithm was suitable even for real-world complex phenomena that are characterized by non-linear dynamics [43,44]. This dataset contained the start of an epileptic seizure, and we wanted to show that it was possible to detect this seizure with the proposed ESE algorithm. All of the experiments were carried out in the programming language Python [45], with the libraries Numpy [46], Scipy [47], and Padasip [48]. The graphs were plotted with the Matplotlib library [49]. The codes with the experiments can be obtained via email from the authors.

5.2. Mackey–Glass Time Series Perturbation

The first experiment was the detection of a perturbed sample in a deterministic chaotic time series. The time series data were obtained as the solution of the Mackey–Glass equation [50].
d y ( t ) d t = β · y ( t τ ) 1 + y α ( t τ ) γ y ( t )
with parameters α = 10 , β = 0.2 , γ = 0.1 , and τ = 17 . In all, 701 data samples were generated. The data sample at discrete time index k = 523 contained the perturbation, as follows:
y ( 523 ) = y ( 523 ) + 0.05 · y ( 523 )
The data series and detailed perturbation are depicted in Figure 2.
The QNU was chosen for the data processing. The number of inputs to the QNU was set to n = 4 , so the inputs are:
x = [ y ( k 1 ) , y ( k 2 ) , y ( k 3 ) , y ( k 4 ) ]
and hence, the adaptive filter had in all 15 adaptive weights. The parameters were updated with every newly obtained sample by means of the NLMS algorithm. The setting was the same as in [13]. The learning rate during the experiment was constantly set to μ = 1 . The POT method was chosen according to (23) with n s = 300 . The details of the adaptive filters and prediction error are depicted in Figure 3. The results of the ND are shown in Figure 4. Note that the global maximum in the E S E corresponds to the perturbed sample. The global maxima of the E L B N D and L E methods correspond to the biggest prediction error, and not to the perturbed sample.

5.3. Change of the Standard Deviation of the Noise in a Random Data Stream

The detection of a change in the standard deviation of the noise in the obtained data was carried out in the following experiment. Assume there are two inputs x 1 ( k ) and x 2 ( k ) and that the output y ( k ) is related to them by:
y ( k ) = x 1 ( k ) + x 2 ( k ) + x 1 ( k ) · x 2 ( k ) + v ( k )
where v ( k ) represents a Gaussian noise that is added to w ( k ) . The Gaussian noise has zero mean and standard deviation 0.1 , υ N ( 0 , 0.1 ) . The values of x 1 ( k ) and x 2 ( k ) are drawn from a uniform distribution, so that x ( k ) 0 and x ( k ) 1 for every k. At the discrete time index k = 500 , the standard deviation of the noise changes to 0.2 , so υ N ( 0 , 0.2 ) . The QNU was chosen for the data processing. The number of inputs to the QNU was set to n = 2 , so the inputs are:
x = [ x 1 ( k 1 ) , x 2 ( k 1 ) ]
and hence, the adaptive filter had three adaptive weights in all. The structure of the QNU corresponds to the structure of the data generator described by Equation (31). The parameters were updated with every newly obtained sample using the GNGD algorithm. The learning rate during the experiment was set to μ = 1 . The POT method was chosen according to (24) with n s = 500 . The results of the novelty detection and details about the adaptive filters are depicted in Figure 5. The a priori values of GPD for ESE and for LE were obtained using 500 samples, which are not shown in Figure 5. Note that the global maximum of the ESE corresponded to the change in standard deviation. The detection by the ELBND and LE was delayed.

5.4. Step Change in the Parameters of a Signal Generator

The scheme of this experiment was similar to the previous one. Assume there are two inputs x 1 ( k ) and x 2 ( k ) and one output y ( k ) , related by:
y ( k ) = x 1 ( k ) + x 2 ( k ) + x 1 ( k ) · x 2 ( k ) + v ( k )
where v ( k ) represents a Gaussian noise that is added to y ( k ) . The Gaussian noise has zero mean and standard deviation 0.1 , υ N ( 0 , 0.1 ) . The values of x 1 ( k ) and x 2 ( k ) are drawn from a uniform distribution, so x ( k ) 0 and x ( k ) 1 for every k. At the discrete time index k = 500 , the equation is changed to the following one:
y ( k ) = 0.4 x 1 ( k ) + 1.6 x 2 ( k ) + 0.99 x 1 ( k ) · x 2 ( k ) + v ( k ) .
The QNU was chosen for the data processing. The number of inputs to the QNU was set to n = 2 , so the inputs are:
x = [ x 1 ( k 1 ) , x 2 ( k 1 ) ]
and hence, the adaptive filter had three adaptive weights in all. Note that the structure of the QNU corresponded to the structure of the signal generator. The parameters were updated with every newly obtained sample, using the GNGD algorithm. The learning rate during the experiment was constantly set to μ = 1 . The POT method was chosen according to (23) with n s = 500 . The a priori values of GPD for ESE and for LE were obtained using 500 samples, which are not shown in Figure 6. The results of the novelty detection and details about the adaptive filters are depicted in Figure 6. Note that the ESE successfully detected the change in the parameters of the signal generator. The LE failed to detect this change, and the detection by ELBND was delayed. Furthermore, the value of the peak in E S E was significantly higher than that in the E L B N D case.

5.5. Noise Disappearance

In this experiment, it was shown that the slightly reformulated algorithm could also deal with an immediate decrease of the learning effort. Assume that instead of an unusually high learning effort, we want to focus on an unusually low learning effort. The only change in the proposed algorithm was that we used the POT method to get l the smallest weight updates, and based on those, the parameters of the GPD would be estimated. The scheme of this experiment was similar to the previous one. We assumed there were two inputs x 1 ( k ) and x 2 ( k ) and one output y ( k ) , which were related by (31). However, in this case, at discrete time index k = 500 , the noise was removed, so Equation (31) for k 500 takes the form:
y ( k ) = x 1 ( k ) + x 2 ( k ) + x 1 ( k ) x 2 ( k ) .
The QNU was chosen for the data processing. The number of inputs to the QNU was set to n = 2 , so the inputs are:
x = [ x 1 ( k 1 ) , x 2 ( k 1 ) ]
and so, the adaptive filter had three adaptive weights in all. The structure of the adaptive filter was chosen to correspond to the structure of the signal generator. The parameters were updated with every newly obtained sample using the GNGD algorithm. The learning rate during the experiment was constantly set to μ = 1 . The POT method was chosen according to (23) with n s = 500 . Figure 7 shows that the peak in E S E corresponded to the disappearance of the noise. The L E and E L B N D methods failed to detect the disappearance of the noise. For E L B N D , these results were to be expected, as the values of the E L B N D were high for a high prediction error and high adaptive weight increments.

5.6. Trend Change

The last experiment with artificial data was the detection of a change in trend. Assume that there are two inputs x 1 ( k ) and x 2 ( k ) and one output y ( k ) , related by:
y ( k ) = x 1 ( k ) + x 2 ( k ) + 0.01 · k + v ( k )
where v ( k ) represents a Gaussian noise that is added to y ( k ) . The Gaussian noise had zero mean and standard deviation 0.1 . At the discrete time index k = 500 , there was a change in the trend, so Equation (38) changes to:
y ( k ) = x 1 ( k ) + x 2 ( k ) + 0.0105 · k + v ( k ) ,
where k 500 . The LNU was chosen for the data processing. The number of inputs to the LNU was set to n = 3 , so the inputs are:
x = [ x 1 ( k 1 ) , x 2 ( k 1 ) , 1 ]
and the adaptive filter had three adaptive weights in all. The structure of the adaptive filter was chosen in accordance with the structure of the signal generator. The parameters were updated with every newly obtained sample by means of the GNGD algorithm. The learning rate during the experiment was constantly set to μ = 1 . The POT method was chosen according to (23) with n s = 500 . Figure 8 shows that the peak in E S E corresponded to the trend change point, which was the same as the peak in L E and E L B N D . Note that the value of the peak in E S E was significantly higher than in L E and E L B N D .

5.7. Detection of Epilepsy in Mouse EEG

The last experiment was with a mouse EEG signal. Three channels of the EEG data were chosen, which contained a significant seizure. According to the expert, the seizure started at about k 1700 , as is shown in Figure 9, which shows the z-scores of the EEG data.
The LNU was chosen for the data processing. The number of inputs to LNU was set to n = 10 , so the inputs are:
x = [ x ( k 1 ) , x ( k 2 ) , , x ( k 10 ) ]
and the adaptive filter had 10 adaptive weights in all. The number of inputs and filter structure were chosen experimentally. The parameters were updated with every newly obtained sample using the NLMS algorithm. The learning rate during the experiment was set to μ = 1 . The POT method was chosen according to (25) with n s = 1000 . Figure 10 shows that the peak in E S E approximately corresponded to the beginning of the seizure. Especially in channel C3, the peak in E S E was significant. The position of the peaks was at k = 1735 for channel C3, k = 1698 for channel Pz, and k = 1727 for channel Fp1.

6. Evaluation of the ESE Detection Rate

This section is dedicated to evaluating the detection rate in two different cases. The first case was a step change in the parameters of a signal generator (similar to the experiment described in Section 5.4). The second case was the detection of a change in trend.

6.1. Step Change in the Parameters of a Signal Generator: Evaluation of the Detection Rate

Assume there are two inputs x 1 ( k ) and x 2 ( k ) , one output y ( k ) , and weights a 1 , a 2 , and a 3 , related by:
y ( k ) = a 1 · x 1 ( k ) + a 2 · x 2 ( k ) + a 3 · x 1 ( k ) · x 2 ( k ) + v ( k )
where v ( k ) represents a Gaussian noise that is added to y ( k ) . The Gaussian noise had zero mean and standard deviation σ . The initial values of a 1 , a 2 , a 3 were drawn from the uniform distribution U ( 1 , 1 ) . At discrete time index k = 200 , there was a step change in a 1 , a 2 , and a 3 , and their new values were drawn again from U ( 1 , 1 ) . The structure of the adaptive filter was the same as described in Section 5.4. The parameters were updated with every newly obtained sample using the GNGD algorithm. The POT method was chosen according to (23) with n s = 1200 . The performance of the ESE algorithm was compared with those of LE, ELBND, and plain prediction error evaluation. The a priori values of GPD for ESE and LE were obtained using 1200 samples with initial values for the parameters a 1 , a 2 , a 3 . For each experiment, the signal-to-noise ratio (SNR) was evaluated as follows:
S N R = 10 log 10 σ s 2 σ 2
where σ s is the standard deviation of the output of the system and σ is the standard deviation of the noise. The evaluation of the rate detection was performed as follows:
  • choose noise standard deviation σ
  • for given noise standard deviation σ , perform 1000 experiments, and at the beginning of each experiment, choose new parameters a 1 , a 2 , and a 3
  • successful detection was when the global peak in ESE, LE, ELBND, or prediction error was between discrete time index k 200 and k 210 ; compute the detection rate
  • compute the SNR for each experiment according to (43), and compute the average SNR for all experiments for given noise standard deviation σ
The evaluation of the detection rate was performed for the inputs x 1 , x 2 whose values were drawn from the uniform distribution U ( 1 , 1 ) and from the normal distribution N ( 0 , 1 ) . The results for the inputs drawn from the uniform distribution are depicted in Figure 11. The corresponding table with results for various SNRs is Table A2 (see Appendix A). The results for inputs drawn from the normal distribution are depicted in Figure 12. The corresponding table with results for various SNRs is Table A3 (see Appendix A).

6.2. Detection of a Change in Trend: Evaluation of the Detection Rate

Assume there are two inputs x 1 ( k ) and x 2 ( k ) and one output y ( k ) , related by:
y ( k ) = x 1 ( k ) + x 2 ( k ) + 0.01 · k + v ( k )
where v ( k ) represents a Gaussian noise that is added to y ( k ) . The Gaussian noise has zero mean and standard deviation σ . At discrete time index k, the trend changed, so the output of the system y ( k ) for k 200 is given by:
y ( k ) = x 1 ( k ) + x 2 ( k ) + ( 0.01 + a ) · k + v ( k )
where a is drawn from the uniform distribution U ( 0.02 , 0.02 ) . The structure of the adaptive filter was the same as in the experiment described in Section 5.6. The parameters were updated with every newly obtained sample using the GNGD algorithm. The POT method was chosen according to (23) with n s = 1200 . The performance of the ESE algorithm was compared with LE, ELBND, and plain prediction error evaluation. The a priori values of the GPD for ESE and LE were obtained using 1200 samples where the output of the system was described by Equation (44). For each experiment, the SNR was evaluated according to (43). The evaluation of the rate detection was performed as follows:
  • choose noise standard deviation σ
  • for given noise standard deviation σ , perform 1000 experiments where at k = 200 , there is a change in trend
  • successful detection is when the global peak in ESE, LE, ELBND, or prediction error is between discrete time index k 200 and k 210 ; compute the detection rate
  • compute the SNR for each experiment according to (43), and compute the average SNR for all experiments for given noise standard deviation σ
The evaluation of the detection rate was performed for inputs x 1 , x 2 whose values were drawn from the uniform distribution U ( 1 , 1 ) . The results are depicted in Figure 13. The corresponding table with the results for various SNRs is Table A1 (see Appendix A).

7. Limitations and Further Challenges

There is a significant limitation to using the ESE algorithm. As was already mentioned in Section 4, before we could obtain the first results, we needed to get a priori information about the parameters of the GPD or obtain a suitably large sample size to compute those parameters. This limitation arose from the nature of using the probability distribution and is common to many statistical approaches to ND. This was the main drawback compared to, e.g., the ELBND method, which was able to produce the results immediately. Another limitation of the presented algorithm is the selection of a suitable POT method, as the estimation of the parameters of the GPD and the selection of the threshold were strongly related to this. To avoid this issue, it was possible to implement some sophisticated parameter estimator that could deal with the optimal threshold selection (e.g., Zhang’s method [51], an estimator based on generalized probability weighted moment equations [52], or a method that combines the method of moments and the likelihood moment [53]), but these are outside the scope of this article. Another challenge was how to combine the ESE of unusually low and unusually high increments together, because both could correspond to a novelty in the data. Further work will be oriented toward using adaptive filters whose adaptive parameters are non-linearly related to the output, e.g., fuzzy adaptive filters or non-linear adaptive Kalman filters. Furthermore, more learning algorithms should be tested. Another topic, which was not mentioned in this article, is that of deciding whether the value of the ESE implies a novelty in the data or not, so we need some threshold. To evaluate the precision of the classification, the area under the receiver operating characteristics [54,55] should be estimated. Due to the scope of this article, this was omitted, but it will be part of further work on the ESE.

8. Conclusions

This paper introduced a new measure of data novelty, called extreme seeking entropy, and a detection algorithm that used this measure. An experimental study was also presented. The algorithm evaluated the absolute value of the increments in the adaptive system weights that were unusually high. The generalized Pareto distribution was used to model those increments, and we tested whether a low probability of a weight increment corresponded to a novelty in the data. It was also shown that the prediction error did not need to be correlated with a novelty in the data, so relatively simple, even inaccurate, adaptive models could be used. Five experiments with synthetic data including novelties and one experiment with a real mouse EEG signal were presented. It was shown that the proposed novelty detection algorithm was able to detect novelties in both kinds of data (real and synthetic) and that the proposed approach using simple adaptive models should be suitable for adaptive novelty detection. The detection rate of the proposed algorithm was evaluated for various SNRs in the scenarios of trend change detection and of a step change in the parameters of a signal generator. These scenarios were also tested with LE, ELBND, and prediction error evaluation. It was shown that for higher SNRs, the proposed ESE algorithm outperformed the other tested algorithms in terms of a successful detection rate in both scenarios.

Author Contributions

Conceptualization, J.V.; method, J.V.; software, J.V.; validation, J.V. and J.M.; investigation, J.V.; resources, J.V. and J.M.; writing, original draft preparation, J.V.; writing, review and editing, J.M.; visualization, J.V.; supervision, J.M.; project administration, J.M.; funding acquisition, J.M. All authors have read and agreed to the published version of the manuscript.

Funding

Financial support was from specific university research (MSMTNo 21-SVV/2019).

Acknowledgments

Jan Vrba would like to thank Matouš Cejnek for developing the PADASIP (Python Adaptive Signal Processing library) and Ivo Bukovský for helpful discussions about learning entropy and learning systems.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
NDnovelty detection
ESEextreme seeking entropy
LElearning entropy
ELBNDerror and learning based novelty detection
GEVgeneralized extreme value
GPDgeneralized Pareto distribution
GNGDgeneralized normalized gradient descent
NLMSnormalized least mean squares
pdfprobability density function
cdfcumulative density function
POTpeak over threshold
LNUlinear neural unit
QNUquadratic neural unit
SNRsignal-to-noise ratio

Appendix A

Table A1. Trend change detection rates.
Table A1. Trend change detection rates.
SNR (dB)ESE (%)ELBND (%)LE (%)Err (%)
41.6399.068.198.954.8
35.6298.468.398.454.2
29.6197.370.297.356.9
26.1196.174.695.759.3
23.6494.374.094.161.5
21.7492.974.992.762.1
20.2091.674.091.462.1
18.9290.474.289.463.9
17.8189.473.489.165.8
16.8588.872.688.065.8
16.0087.574.587.066.7
15.2386.574.485.567.3
14.5585.574.084.068.2
13.9284.575.383.870.4
13.3584.275.383.269.9
12.8283.675.682.470.9
12.3482.676.781.272.1
11.8881.676.280.472.8
11.4680.975.679.372.0
11.0779.875.278.471.8
10.7078.774.777.372.4
10.0276.975.075.673.0
9.4275.674.274.072.3
8.8873.972.871.372.3
8.3972.272.169.171.9
7.9570.470.466.970.3
7.5469.168.365.268.8
7.1666.567.163.067.7
6.8264.466.760.766.7
6.2059.263.256.163.9
5.6755.458.652.160.6
4.9949.052.546.255.6
Table A2. Step change detection rates for inputs drawn from uniform distribution U ( 1 , 1 ) .
Table A2. Step change detection rates for inputs drawn from uniform distribution U ( 1 , 1 ) .
SNR (dB)ESE (%)ELBND (%)LE (%)Err (%)
39.70100.094.699.595.2
33.68100.094.698.795.3
30.1699.894.798.195.4
27.6799.594.796.795.3
25.7399.294.696.495.4
25.0799.194.696.395.4
24.1598.994.596.295.2
23.3398.694.395.995.0
22.8298.594.295.794.9
22.1198.194.295.795.1
21.6797.894.395.294.8
20.6697.494.094.894.5
19.7596.993.993.994.2
18.9396.793.893.694.2
18.1996.493.593.194.0
17.7795.793.192.593.6
17.5195.392.792.293.2
17.1294.892.191.692.7
16.8894.592.191.192.4
16.5294.091.890.792.1
16.2994.191.690.392.1
15.9693.591.590.291.7
15.7592.891.489.591.2
15.2491.791.088.491.0
15.0491.490.888.090.8
14.7690.790.087.590.6
14.4890.389.586.990.3
14.3190.289.286.390.1
14.1390.188.985.689.9
13.8889.188.485.589.5
10.6078.880.275.282.3
10.0976.677.973.680.4
9.6274.675.971.878.0
9.1972.173.669.675.6
8.7869.971.668.073.6
8.4168.270.465.771.2
8.0566.568.864.269.1
7.7263.967.462.067.2
7.4061.564.259.664.8
3.6430.132.430.131.4
1.3311.711.211.59.7
Table A3. Step change detection rates for inputs drawn from normal distribution N ( 0 , 1 ) .
Table A3. Step change detection rates for inputs drawn from normal distribution N ( 0 , 1 ) .
SNR (dB)ESE (%)ELBND (%)LE (%)Err (%)
43.86100.096.199.994.8
37.84100.096.399.294.8
34.32100.096.198.494.9
31.8399.895.898.194.7
29.8999.696.097.794.6
28.3199.496.097.194.6
26.9799.395.796.994.4
25.8299.195.796.994.5
24.8098.995.796.493.9
23.8998.595.696.193.6
23.0698.395.695.993.2
22.3197.995.095.592.7
21.9697.694.895.392.5
21.6297.594.795.392.4
20.9896.594.294.591.6
20.3996.194.193.991.5
19.8495.893.593.891.2
19.3295.693.193.390.6
18.8395.292.792.690.0
18.3794.992.592.389.7
17.9394.792.291.888.5
17.5294.391.891.687.5
17.1293.491.090.487.1
16.3992.189.989.486.1
15.7190.188.687.884.6
15.0988.687.186.883.6
14.5287.386.485.882.0
13.9885.084.984.279.5
13.7384.383.983.778.9
13.4883.683.382.778.1
13.0282.181.781.676.4
12.1679.778.978.873.6
11.7778.677.877.972.3
11.4077.676.176.070.5
11.0575.974.874.169.3
10.8875.074.173.668.8
10.5573.372.471.967.2
10.0971.370.469.765.5
9.6768.868.767.063.9
9.2767.165.965.062.2
9.0165.864.063.560.9
8.7764.262.062.059.0
8.5462.460.260.757.0
8.1059.958.157.854.9
7.8958.256.856.454.1
7.4155.555.153.850.5
6.9752.152.250.846.6
6.5749.049.648.644.2
6.2045.947.345.642.1
5.5441.040.740.137.8
4.9836.336.536.534.7
4.5033.032.432.330.6
4.0830.028.828.427.3
3.7127.426.025.924.5
3.3924.224.123.922.3
3.1021.622.221.420.5
2.8519.920.319.219.4
2.6318.518.117.517.8
2.4316.916.516.316.6
2.2515.815.415.315.2
2.0914.714.314.014.0

References

  1. Markou, M.; Singh, S. Novelty detection: A review—Part 1: Statistical approaches. Signal Process. 2003, 83, 2481–2497. [Google Scholar] [CrossRef]
  2. Clifton, D.A.; Hugueny, S.; Tarassenko, L. Novelty detection with multivariate extreme value statistics. J. Signal Process. Syst. 2011, 65, 371–389. [Google Scholar] [CrossRef]
  3. Hugueny, S.; Clifton, D.A.; Tarassenko, L. Probabilistic patient monitoring with multivariate, multimodal extreme value theory. In International Joint Conference on Biomedical Engineering Systems and Technologies; Springer: Berlin/Heidelberg, Germnay, 2010; pp. 199–211. [Google Scholar]
  4. Clifton, D.A.; Clifton, L.; Hugueny, S.; Tarassenko, L. Extending the generalised Pareto distribution for novelty detection in high-dimensional spaces. J. Signal Process. Syst. 2014, 74, 323–339. [Google Scholar] [CrossRef] [Green Version]
  5. Luca, S.; Clifton, D.A.; Vanrumste, B. One-class classification of point patterns of extremes. J. Mach. Learn. Res. 2016, 17, 6581–6601. [Google Scholar]
  6. Markou, M.; Singh, S. Novelty detection: A review—Part 2: Neural network based approaches. Signal Process. 2003, 83, 2499–2521. [Google Scholar] [CrossRef]
  7. Marsland, S. Novelty detection in learning systems. Neural Comput. Surv. 2003, 3, 157–195. [Google Scholar]
  8. Polycarpou, M.M.; Trunov, A.B. Learning approach to nonlinear fault diagnosis: Detectability analysis. IEEE Trans. Autom. Control 2000, 45, 806–812. [Google Scholar] [CrossRef]
  9. Pimentel, M.A.; Clifton, D.A.; Clifton, L.; Tarassenko, L. A review of novelty detection. Signal Process. 2014, 99, 215–249. [Google Scholar] [CrossRef]
  10. Yazdanpanah, H.; Lima, M.V.; Diniz, P.S. On the robustness of set-membership adaptive filtering algorithms. EURASIP J. Adv. Signal Process. 2017, 2017, 72. [Google Scholar] [CrossRef] [Green Version]
  11. Diniz, P.S.R. Adaptive Filtering; Springer: New York, NY, USA, 2013. [Google Scholar] [CrossRef]
  12. Gollamudi, S.; Nagaraj, S.; Kapoor, S.; Huang, Y.F. Set-membership filtering and a set-membership normalized LMS algorithm with an adaptive step size. IEEE Signal Process. Lett. 1998, 5, 111–114. [Google Scholar] [CrossRef] [Green Version]
  13. Bukovsky, I. Learning Entropy: Multiscale Measure for Incremental Learning. Entropy 2013, 15, 4159–4187. [Google Scholar] [CrossRef]
  14. Bukovsky, I.; Kinsner, W.; Homma, N. Learning Entropy as a Learning-Based Information Concept. Entropy 2019, 21, 166. [Google Scholar] [CrossRef] [Green Version]
  15. Cejnek, M.; Bukovsky, I. Concept drift robust adaptive novelty detection for data streams. Neurocomputing 2018, 309, 46–53. [Google Scholar] [CrossRef]
  16. Bukovsky, I.; Kinsner, W.; Bila, J. Multiscale analysis approach for novelty detection in adaptation plot. In Proceedings of the Sensor Signal Processing for Defence (SSPD 2012), London, UK, 25–27 September 2012; pp. 1–6. [Google Scholar]
  17. Taoum, A.; Mourad-chehade, F.; Amoud, H. Early-warning of ARDS using novelty detection and data fusion. Comput. Biol. Med. 2018, 102, 191–199. [Google Scholar] [CrossRef] [PubMed]
  18. Rad, N.M.; van Laarhoven, T.; Furlanello, C.; Marchiori, E. Novelty Detection using Deep Normative Modeling for IMU-Based Abnormal Movement Monitoring in Parkinson’s Disease and Autism Spectrum Disorders. Sensors 2018, 18, 3533. [Google Scholar] [CrossRef] [Green Version]
  19. Burlina, P.; Joshi, N.; Billings, S.; Wang, I.J.; Albayda, J. Deep embeddings for novelty detection in myopathy. Comput. Biol. Med. 2019, 105, 46–53. [Google Scholar] [CrossRef]
  20. Hu, L.; Hu, N.; Fan, B.; Gu, F. Application of novelty detection methods to health monitoring and typical fault diagnosis of a turbopump. J. Phys. Conf. Ser. 2012, 364, 012128. [Google Scholar] [CrossRef] [Green Version]
  21. Surace, C.; Worden, K. A novelty detection method to diagnose damage in structures: An application to an offshore platform. In The Eighth International Offshore and Polar Engineering Conference; International Society of Offshore and Polar Engineers: Mountain View, CA, USA, 1998. [Google Scholar]
  22. Bukovsky, I.; Homma, N.; Smetana, L.; Rodriguez, R.; Mironovova, M.; Vrana, S. Quadratic neural unit is a good compromise between linear models and neural networks for industrial applications. In Proceedings of the 9th IEEE International Conference on Cognitive Informatics (ICCI’10), Beijing, China, 7–9 July 2010; pp. 556–560. [Google Scholar] [CrossRef]
  23. Zhang, M. (Ed.) Artificial Higher Order Neural Networks for Modeling and Simulation; IGI Global: Hershey, PA, USA, 2013. [Google Scholar] [CrossRef] [Green Version]
  24. Bukovsky, I.; Voracek, J.; Ichiji, K.; Noriyasu, H. Higher Order Neural Units for Efficient Adaptive Control of Weakly Nonlinear Systems. In Proceedings of the 9th International Joint Conference on Computational Intelligence; SciTePress—Science and Technology Publications: Setúbal, Portugal, 2017. [Google Scholar] [CrossRef]
  25. Haykin, S.S. Adaptive Filter Theory; Pearson Education India: Delhi, India, 2005. [Google Scholar]
  26. Mandic, D.P. A generalized normalized gradient descent algorithm. IEEE Signal Process. Lett. 2004, 11, 115–118. [Google Scholar] [CrossRef]
  27. Grubbs, F.E. Procedures for detecting outlying observations in samples. Technometrics 1969, 11, 1–21. [Google Scholar] [CrossRef]
  28. Ma, J.; Perkins, S. Time-series novelty detection using one-class support vector machines. In Proceedings of the International Joint Conference on Neural Networks, 2003, Portland, OR, USA, 20–24 July 2003; Volume 3, pp. 1741–1745. [Google Scholar]
  29. Ma, J.; Perkins, S. Online novelty detection on temporal sequences. In Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Washington, DC, USA, 24–27 August 2003; pp. 613–618. [Google Scholar]
  30. Limpert, E.; Stahel, W.A. Problems with Using the Normal Distribution—And Ways to Improve Quality and Efficiency of Data Analysis. PLoS ONE 2011, 6, e21403. [Google Scholar] [CrossRef] [Green Version]
  31. Pickands, J., III. Statistical inference using extreme order statistics. Ann. Stat. 1975, 3, 119–131. [Google Scholar]
  32. Balkema, A.A.; De Haan, L. Residual life time at great age. Ann. Probab. 1974, 2, 792–804. [Google Scholar] [CrossRef]
  33. Forbes, C.; Evans, M.; Hastings, N.; Peacock, B. Statistical Distributions; Wiley: Hoboken, NJ, USA, 2010. [Google Scholar]
  34. Coles, S. An Introduction to Statistical Modeling of Extreme Values; Springer Series in Statistics; Springer: London, UK, 2013. [Google Scholar]
  35. Lee, H.J.; Roberts, S.J. On-line novelty detection using the Kalman filter and extreme value theory. In Proceedings of the 2008 19th International Conference on Pattern Recognition, Tampa, FL, USA, 8–11 December 2008. [Google Scholar] [CrossRef]
  36. Scarrott, C.; MacDonald, A. A review of extreme value threshold estimation and uncertainty quantification. REVSTAT-Stat. J. 2012, 10, 33–60. [Google Scholar]
  37. DuMouchel, W.H. Estimating the stable index α in order to measure tail thickness: A critique. Ann. Stat. 1983, 11, 1019–1031. [Google Scholar] [CrossRef]
  38. Ferreira, A.; de Haan, L.; Peng, L. On optimising the estimation of high quantiles of a probability distribution. Statistics 2003, 37, 401–434. [Google Scholar] [CrossRef]
  39. Loretan, M.; Phillips, P.C. Testing the covariance stationarity of heavy-tailed time series: An overview of the theory with applications to several financial datasets. J. Empir. Financ. 1994, 1, 211–248. [Google Scholar] [CrossRef]
  40. Spangenberg, M.; Calmettes, V.; Julien, O.; Tourneret, J.Y.; Duchateau, G. Detection of variance changes and mean value jumps in measurement noise for multipath mitigation in urban navigation. Navigation 2010, 57, 35–52. [Google Scholar] [CrossRef] [Green Version]
  41. L’Ecuyer, P. History of uniform random number generation. In Proceedings of the 2017 Winter Simulation Conference (WSC), Las Vegas, NV, USA, 3–6 December 2017; pp. 202–230. [Google Scholar]
  42. Maurya, M.R.; Rengaswamy, R.; Venkatasubramanian, V. Fault diagnosis using dynamic trend analysis: A review and recent developments. Eng. Appl. Artif. Intell. 2007, 20, 133–146. [Google Scholar] [CrossRef]
  43. Strogatz, S.H. Nonlinear Dynamics and Chaos: With Applications to Physics, Biology, Chemistry, and Engineering; CRC Press: Boca Raton, FL, USA, 2018. [Google Scholar]
  44. Rodriguez-Bermudez, G.; Garcia-Laencina, P.J. Analysis of EEG signals using nonlinear dynamics and chaos: A review. Appl. Math. Inf. Sci. 2015, 9, 2309. [Google Scholar]
  45. Van Rossum, G.; Drake, F.L., Jr. Python Tutorial; Centrum voor Wiskunde en Informatica: Amsterdam, The Netherlands, 1995. [Google Scholar]
  46. Oliphant, T.E. A Guide to NumPy; Trelgol Publishing USA, 2006; Volume 1, Available online: https://ecs.wgtn.ac.nz/foswiki/pub/Support/ManualPagesAndDocumentation/numpybook.pdf (accessed on 11 January 2020).
  47. Jones, E.; Oliphant, T.; Peterson, P. SciPy: Open Source Scientific Tools for Python. 2001. Available online: https://www.bibsonomy.org/bibtex/21b37d2cc741af879d7958f2f7c23c420/microcuts (accessed on 11 January 2020).
  48. Cejnek, M. Padasip—Open source library for adaptive signal processing in language Python. In Studentská Tvůrčí činnost 2017; Department of Instrumentation and Control Engineering: Prague, Czech Rebpublic, 2017. [Google Scholar]
  49. Hunter, J.D. Matplotlib: A 2D graphics environment. Comput. Sci. Eng. 2007, 9, 90–95. [Google Scholar] [CrossRef]
  50. Mackey, M.; Glass, L. Oscillation and chaos in physiological control systems. Science 1977, 197, 287–289. [Google Scholar] [CrossRef] [PubMed]
  51. Zhang, J. Likelihood moment estimation for the generalized Pareto distribution. Aust. N. Z. J. Stat. 2007, 49, 69–77. [Google Scholar] [CrossRef]
  52. Chen, H.; Cheng, W.; Zhao, J.; Zhao, X. Parameter estimation for generalized Pareto distribution by generalized probability weighted moment-equations. Commun. Stat.-Simul. Comput. 2017, 46, 7761–7776. [Google Scholar] [CrossRef]
  53. Zhao, X.; Zhang, Z.; Cheng, W.; Zhang, P. A New Parameter Estimator for the Generalized Pareto Distribution under the Peaks over Threshold Framework. Mathematics 2019, 7, 406. [Google Scholar] [CrossRef] [Green Version]
  54. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 143, 29–36. [Google Scholar] [CrossRef] [Green Version]
  55. Bradley, A.P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef] [Green Version]
Figure 1. GPD probability density function with various parameters ξ , and fixed parameters σ = 1 , μ = 0 .
Figure 1. GPD probability density function with various parameters ξ , and fixed parameters σ = 1 , μ = 0 .
Entropy 22 00093 g001
Figure 2. The (upper) plot displays an overview of the data with the perturbation. The (bottom) plots show the detailed perturbation.
Figure 2. The (upper) plot displays an overview of the data with the perturbation. The (bottom) plots show the detailed perturbation.
Entropy 22 00093 g002
Figure 3. The graph (a) shows the data series with perturbation (solid black) and the output of the predictor (dotted green). The sample with perturbation is marked with the black arrow. The graph (b) shows the absolute value of the error of the predictor. The graph (c) shows the absolute value of the increments of the adaptable parameters.
Figure 3. The graph (a) shows the data series with perturbation (solid black) and the output of the predictor (dotted green). The sample with perturbation is marked with the black arrow. The graph (b) shows the absolute value of the error of the predictor. The graph (c) shows the absolute value of the increments of the adaptable parameters.
Entropy 22 00093 g003
Figure 4. The graph (a) shows the ESE novelty score. Note that n s = 300 , so the first 300 samples are needed to obtain the first value of E S E . The graph (b) shows the ELBND novelty score. The first 300 samples are set to zero for easier comparison with the other methods. The graph (c) shows the LE novelty score.
Figure 4. The graph (a) shows the ESE novelty score. Note that n s = 300 , so the first 300 samples are needed to obtain the first value of E S E . The graph (b) shows the ELBND novelty score. The first 300 samples are set to zero for easier comparison with the other methods. The graph (c) shows the LE novelty score.
Entropy 22 00093 g004
Figure 5. The graph (a) shows the data series (blue) and the output of the predictor (green). The graph (b) shows the error of the predictor. The graph (c) shows the absolute value of the increments of the adaptable parameters. The graph (d) shows the E S E novelty score. At discrete time index k = 500 , there is a step change in the standard deviation of the noise and a corresponding global peak in E S E . Graphs (e) and (f) contain the results of the ELBND and LE methods.
Figure 5. The graph (a) shows the data series (blue) and the output of the predictor (green). The graph (b) shows the error of the predictor. The graph (c) shows the absolute value of the increments of the adaptable parameters. The graph (d) shows the E S E novelty score. At discrete time index k = 500 , there is a step change in the standard deviation of the noise and a corresponding global peak in E S E . Graphs (e) and (f) contain the results of the ELBND and LE methods.
Entropy 22 00093 g005
Figure 6. The graph (a) shows the data series (blue) and the output of the predictor (green). The graph (b) shows the value of predictors error. The graph (c) shows the absolute value of increments in the adaptable parameters. The graph (d) shows the E S E novelty score. At discrete time index k = 500 , there is a step change in the parameters and a corresponding global maximum in E S E . Graphs (e) and (f) contain the results of the ELBND and LE methods. The detection with ELBND is delayed. The LE method failed to detect the change in this case, as the global maximum of LE is at k = 338 .
Figure 6. The graph (a) shows the data series (blue) and the output of the predictor (green). The graph (b) shows the value of predictors error. The graph (c) shows the absolute value of increments in the adaptable parameters. The graph (d) shows the E S E novelty score. At discrete time index k = 500 , there is a step change in the parameters and a corresponding global maximum in E S E . Graphs (e) and (f) contain the results of the ELBND and LE methods. The detection with ELBND is delayed. The LE method failed to detect the change in this case, as the global maximum of LE is at k = 338 .
Entropy 22 00093 g006
Figure 7. Noise disappearance detection. The graph (a) shows the data series (blue) and the output of the predictor (green). The graph (b) shows the error of the predictor. The graph (c) shows the absolute value of the increments of the adaptable parameters. The graph (d) shows the E S E novelty score. For the discrete time index k 500 , the noise is removed from the signal, which corresponds to the peak in E S E . Graphs (e) and (f) contain the results of the ELBND and LE methods.
Figure 7. Noise disappearance detection. The graph (a) shows the data series (blue) and the output of the predictor (green). The graph (b) shows the error of the predictor. The graph (c) shows the absolute value of the increments of the adaptable parameters. The graph (d) shows the E S E novelty score. For the discrete time index k 500 , the noise is removed from the signal, which corresponds to the peak in E S E . Graphs (e) and (f) contain the results of the ELBND and LE methods.
Entropy 22 00093 g007
Figure 8. Trend change detection. The graph (a) shows the data series (blue) and the output of the predictor (green). The black arrow indicates the trend change. The graph (b) shows the error of the predictor. The graph (c) shows the absolute values of the increments in the adaptable parameters. The graph (d) shows the E S E novelty score. At discrete time index k = 500 , there is a step change in the trend, which corresponds to the peak in E S E . Graphs (e) and (f) contain the results of the ELBND and LE methods and peaks corresponding to the trend change.
Figure 8. Trend change detection. The graph (a) shows the data series (blue) and the output of the predictor (green). The black arrow indicates the trend change. The graph (b) shows the error of the predictor. The graph (c) shows the absolute values of the increments in the adaptable parameters. The graph (d) shows the E S E novelty score. At discrete time index k = 500 , there is a step change in the trend, which corresponds to the peak in E S E . Graphs (e) and (f) contain the results of the ELBND and LE methods and peaks corresponding to the trend change.
Entropy 22 00093 g008
Figure 9. Selected mouse EEG channels with significant seizure. The data were standardized. The start of the seizure is approximately at discrete time index k 1700 .
Figure 9. Selected mouse EEG channels with significant seizure. The data were standardized. The start of the seizure is approximately at discrete time index k 1700 .
Entropy 22 00093 g009
Figure 10. ESE value for mouse EEG data channels containing a seizure. The peaks approximately correspond to the beginning of the seizure. Note that channel C3 contains a significant peak in ESE compared to the other channels.
Figure 10. ESE value for mouse EEG data channels containing a seizure. The peaks approximately correspond to the beginning of the seizure. Note that channel C3 contains a significant peak in ESE compared to the other channels.
Entropy 22 00093 g010
Figure 11. Detection of the step change in the parameters of a signal generator. The inputs of the signal generator are drawn from the uniform distribution U ( 1 , 1 ) . For S N R > 15 dB, the ESE algorithm outperforms in the detection rate the LE, ELBND, and error evaluation. For S N R > 33 dB, the ESE achieved a 100% detection rate.
Figure 11. Detection of the step change in the parameters of a signal generator. The inputs of the signal generator are drawn from the uniform distribution U ( 1 , 1 ) . For S N R > 15 dB, the ESE algorithm outperforms in the detection rate the LE, ELBND, and error evaluation. For S N R > 33 dB, the ESE achieved a 100% detection rate.
Entropy 22 00093 g011
Figure 12. Detection of the step change in parameters of a signal generator. The inputs of the signal generator are drawn from the normal distribution N ( 0 , 1 ) . For S N R > 8 dB, the ESE algorithm outperforms in the detection rate the LE, ELBND, and error evaluation. For S N R > 34 dB, the ESE achieved a 100% detection rate.
Figure 12. Detection of the step change in parameters of a signal generator. The inputs of the signal generator are drawn from the normal distribution N ( 0 , 1 ) . For S N R > 8 dB, the ESE algorithm outperforms in the detection rate the LE, ELBND, and error evaluation. For S N R > 34 dB, the ESE achieved a 100% detection rate.
Entropy 22 00093 g012
Figure 13. Detection of the trend change. The inputs of the signal generator are drawn from the uniform distribution U ( 1 , 1 ) . For S N R > 8 dB, the ESE algorithm outperforms in the detection rate the LE, ELBND, and error evaluation.
Figure 13. Detection of the trend change. The inputs of the signal generator are drawn from the uniform distribution U ( 1 , 1 ) . For S N R > 8 dB, the ESE algorithm outperforms in the detection rate the LE, ELBND, and error evaluation.
Entropy 22 00093 g013

Share and Cite

MDPI and ACS Style

Vrba, J.; Mareš, J. Introduction to Extreme Seeking Entropy. Entropy 2020, 22, 93. https://doi.org/10.3390/e22010093

AMA Style

Vrba J, Mareš J. Introduction to Extreme Seeking Entropy. Entropy. 2020; 22(1):93. https://doi.org/10.3390/e22010093

Chicago/Turabian Style

Vrba, Jan, and Jan Mareš. 2020. "Introduction to Extreme Seeking Entropy" Entropy 22, no. 1: 93. https://doi.org/10.3390/e22010093

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop