Next Article in Journal
Predicting Premature Video Skipping and Viewer Interest from EEG Recordings
Next Article in Special Issue
Ordinal Pattern Based Entropies and the Kolmogorov–Sinai Entropy: An Update
Previous Article in Journal
Communication Enhancement through Quantum Coherent Control of N Channels in an Indefinite Causal-Order Scenario
Previous Article in Special Issue
Two Tests for Dependence (of Unknown Form) between Time Series
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Permutation Entropy: Enhancing Discriminating Power by Using Relative Frequencies Vector of Ordinal Patterns Instead of Their Shannon Entropy

by
David Cuesta-Frau
1,2,*,
Antonio Molina-Picó
1,2,
Borja Vargas
3 and
Paula González
3
1
Technological Institute of Informatics, Universitat Politècnica de València, 03801 Alcoi Campus, Spain
2
Innovatec Sensorización y Comunicación S.L., Avda. Elx, 3, 03801 Alcoi, Spain
3
Department of Internal Medicine, Móstoles Teaching Hospital, Móstoles, 28935 Madrid, Spain
*
Author to whom correspondence should be addressed.
Entropy 2019, 21(10), 1013; https://doi.org/10.3390/e21101013
Submission received: 21 September 2019 / Revised: 15 October 2019 / Accepted: 16 October 2019 / Published: 18 October 2019

Abstract

:
Many measures to quantify the nonlinear dynamics of a time series are based on estimating the probability of certain features from their relative frequencies. Once a normalised histogram of events is computed, a single result is usually derived. This process can be broadly viewed as a nonlinear I R n mapping into I R , where n is the number of bins in the histogram. However, this mapping might entail a loss of information that could be critical for time series classification purposes. In this respect, the present study assessed such impact using permutation entropy (PE) and a diverse set of time series. We first devised a method of generating synthetic sequences of ordinal patterns using hidden Markov models. This way, it was possible to control the histogram distribution and quantify its influence on classification results. Next, real body temperature records are also used to illustrate the same phenomenon. The experiments results confirmed the improved classification accuracy achieved using raw histogram data instead of the PE final values. Thus, this study can provide a very valuable guidance for the improvement of the discriminating capability not only of PE, but of many similar histogram-based measures.

1. Introduction

Classification of time series is one of the main applications of pattern recognition and machine learning [1]. From a collection of features extracted from each time series, a varied and diverse set of methods have been proposed to assign a class label to a group considered homogeneous using a certain dissimilarity criterion [2]. These methods have been applied to any scientific or technological field where a temporal sequence of data is generated [3,4,5,6,7,8].
The present paper is focused specifically on the classification of time series using signal complexity features. Such features exhibit a very high discriminating power, and that is why they have been successfully employed in many applications [9,10,11]. However, there are ongoing efforts to further improve this discriminating power with new complexity estimation algorithms or by tweaking current ones [12,13,14,15,16,17], as is the case in this paper.
Many measures or methods have been described in the scientific literature so far to quantify time series complexity, irregularity, uncertainty, predictability or disorder, among other similar terms devised to characterise their dynamical behaviour [18]. The present study deals specifically with the concept of entropy, one of the previous terms related to expected information, and often employed as a single distinguishing feature for time series classification. Entropy, as an information gauge, has also been defined in many forms, with higher entropy values accounting for higher uncertainty.
The concept of entropy was first applied in the context of thermodynamics [19]. A few decades later, other scientific and technological areas adapted and customised this concept to information theory [20]. The Shannon entropy introduced in [21] soon became one of the most used and successful of these measures. In its discrete and generic form, it is given by
h S = - ω Ω p ω log p ω
where ω is a random variable with probability distribution p ω over the probability space Ω , and log is usually a base 2 logarithm, although natural logarithms are fairly common too.
For the focus of interest of the present paper, classification of time series, Equation (1) must be customised for sequences of numerical values. Assuming a stationary ergodic time series, x , containing amplitude values of a continuous process with a sampling period of T, this time series can be written as the discrete time vector x = x [ 0 T ] , x [ 1 T ] , x [ 2 T ] , , x [ ( N - 1 ) T ] , or, in a more compact and simple form, as x = x 0 , x 1 , x 2 , , x N - 1 , with length N. At this point, Equation 1 could be applied to the samples in x . However, as in most real applications p is unknown and N is finite, an estimation, p ^ , based on event counting or relative frequency has to be used instead.
There are many possible available choices as for the event type related to x from which to obtain p ^ . We adopted a block approach, but instead of directly using samples or subsequences from the time series, we used associated ordinal patterns (more details in Section 2.1) from consecutive m overlapping windows, x i m = x i , x i + 1 , x i + 2 , , x i + m - 1 . This way, the number of different blocks is finite and known in advance, and the relative frequencies can be estimated robustly provided the length of the blocks m is significantly shorter than the length of the time series, m < < N [18]. If each possible ordinal pattern is referred to as Π j m , with 0 j < m ! , Equation (1) takes the form
h ^ S m = - j = 0 m ! - 1 p ^ Π j m log p ^ Π j m
where p ^ Π i m is the relative frequency of the ordinal pattern Π i m . For example, for subsequences of length 3, with sample position indices ranging from 0 to 2, the ordinal patterns Π j 3 that can be found are (0,1,2), (0,2,1), (1,0,2), (1,2,0), (2,1,0) and (2,0,1).
Figure 1 depicts the histograms for a few popular synthetic time series, as an example of the distribution of these relative frequencies. Specifically, the histogram in Figure 1a belongs to a random time series with a uniform distribution, as can be inferred from the equality of the bins. At the opposite end of the spectrum of randomness, Figure 1b shows the histogram of a sinusoidal time series, with a more polarised distribution of motifs, some of them forbidden (zero probability) [22]. Figure 1c,d shows the histograms of other time series with a different degree of determinism, a Logistic (coefficient 3.5) and a Lorenz (parameters 10, 8/3 and 28) time series, respectively [23,24]. When computing the associated entropy to the relative frequencies shown in Figure 1, the single value obtained will be arguably very different in each case.
However, the mapping of relative frequencies into a single scalar (in the sense of a size 1 vector) may entail a significant information loss, as there are infinite vectors of estimated probabilities p ^ = p ^ Π 0 m , p ^ Π 1 m , , p ^ Π m ! - 1 m that can yield the same h ^ S m . One of the most successful applications of entropy statistics is time series classification, and this information loss could have a detrimental impact on the accuracy achieved in these applications.
The present paper is aimed at characterising those situations where a single entropy measure may fail in providing discriminating information due to histogram compensation. This study will be based on a specific Shannon entropy embodiment: permutation entropy (PE) [25]. Despite the good results achieved using this measure, there are still cases where the histogram differences are lost when computing the final PE value. We illustrate this situation with synthetic records generated using hidden Markov models (HMM), and with real body temperature records that exhibit this same behaviour. We then propose a different approach based on a clustering scheme that uses all the relative frequencies as time series features to overcome this problem. The results for both types of experimental time series demonstrate how critical this information loss may become, and how the proposed clustering method can solve it.
There is no similar study in the scientific literature, as far as we know. However, there are a few works that could use a multidimensional classification approach as ours. For example, it is worth describing the approach taken by [26]. This paper presents a method to classify beat-to-beat intervals data series from healthy and congestive heart failure patients using classical RR features, a symbolic representation and ordinal pattern statistics. It first considers each feature individually, and then a combination of two using a linear support vector machine. The main difference with the approach described in the present paper is that the method is fully supervised, and customised for a specific biomedical record, without taking into account the full picture of the PE context.
Another key concept that will play an important role in this work is the concept of forbidden ordinal pattern [18]. In a general sense, we consider a forbidden pattern with an ordinal pattern with a relative frequency of 0 and a forbidden transition i j , the impossibility to generate an ordinal pattern j if the previous one was an ordinal pattern i. Forbidden patterns have already demonstrated their usefulness to detect determinism in time series [27,28], and have even been used for classification purposes already [22,23], as additional distinguishing features. Specifically, the present study will use forbidden transitions as a tool to generate synthetic time series including forbidden patterns or certain histogram distributions.

2. Materials and Methods

2.1. Permutation Entropy

PE is the Shannon entropy of the estimated probabilities of ordinal patterns found in input time series x [25]. To obtain these ordinal patterns, consecutive subsequences of length m and commencing at sample i have to be drawn from x . Formally, these subsequences can be defined as x i m = x i , x i + 1 , , x i + m - 1 , where the index i ranges from 0 to N - m , with N being the total length of x . Initially, when the subsequences are first extracted from the time series, the order of x i , x i + 1 , , x i + m - 1 can be any, but the indices of the samples are always taken as 0 , 1 , , m - 1 by default. When x i , x i + 1 , , x i + m - 1 are sorted in ascending order, the resulting new order is translated to the vector of indices. This new order of indices is considered the ordinal pattern or motif that represents x i m . We will refer to this vector as π i m . For example, if x i 3 = 1.7 , - 0.3 , 2.5 , the associated motif is ( 1 , 0 , 2 ) , as the minimum, −0.3, was located at index 1, the next value in ascending order, 1.7, at index 0, and the maximum value, 2.5, at index 2. Therefore, π i 3 = 1 , 0 , 2 . The corresponding counter (histogram bin) of motif 1 , 0 , 2 , Π j 3 , will therefore be increased, c Π j 3 c Π j 3 + 1 (supposing Π j 3 = 1 , 0 , 2 ). For simplification purposes, m will be 3 in all the tests conducted.
This process is repeated for all the possible subsequences in x , and the final relative frequency of each j-th motif is computed, obtaining p ^ using the normalised histogram of such motifs. Thus, p ^ specifically becomes in PE p ^ = p ^ ( Π 0 m ) , p ^ ( Π 1 m ) , p ^ ( Π 2 m ) , , p ^ ( Π m ! - 1 m ) , as described in the previous section. Finally, p ^ is used to calculate PE as
PE ( x , m , N ) = - k = 0 m ! - 1 p ^ ( Π k m ) log p ^ ( Π k m ) , p ^ ( Π k m ) > 0
which essentially coincides with Equation (2). This entropy can be further normalised by the embedded dimension m, enabling the PE comparison for different m and by log m ! . There are versions based on Reny or Tsallis entropies instead [14].

2.2. Clustering Algorithm

The objective of a clustering algorithm is to find in an unsupervised way natural groups in any dataset according to certain features and a dissimilarity measure or metric [29]. Specifically, the purpose of using a clustering algorithm in this study is to separate automatically the classes in the experimental datasets based on the relative frequencies of motifs.
The selection of a specific method of clustering among the thousands available depends on the features of the input data, the achievable accuracy, implementation issues, and computational cost, among other possible requirements [30]. In any case, the goal is to minimise the intracluster distance or dissimilarity between objects (time series) and maximise the intercluster one [31].
We specifically chose the k-means clustering algorithm [32]. This is probably the most widely used and known clustering algorithm, and, after more than 50 years since its introduction [33], it remains fully in use, in its original form or in its many spawned evolutions [34]. It is also simple to implement and available in many software tools and libraries. We have also previously used this method successfully [4,35,36,37].
The input to the clustering algorithm in this case is a set of time series defined by their features, their estimated probabilities p ^ ( Π 0 m ) , p ^ ( Π 1 m ) , p ^ ( Π 2 m ) , , p ^ ( Π m ! - 1 m ) . The dissimilarity d between two time series, i and j, will correspond to the distance between their two associated estimated probabilities vectors d p ^ i , p ^ j . This distance will be based on the Euclidean metric [38]. The number of clusters is set in advance as 2 (to account for the two classes included in all the experimental datasets), avoiding one of the main problems in many clustering algorithms [39].
Before the k-means iteration takes place, it is necessary to obtain an initial set of centroids from which an initial partition is computed. There are also many methods to achieve this initial set [40], including random centroid selection [41]. The clustering performance may greatly depend on this [42], as k-means does not guarantee global optimality. For repeatability, we chose a max–min scheme to choose the two centroids [43]. The first record in the experimental database was taken as the centroid of one class, and the furthest to this one in terms of d was the initial centroid for the other class (max approach). The corresponding partition can then be obtained assigning each time series to its closest centroid (min approach). This way, the results will not alter across the experiments, at the expense of a possible suboptimal centroid selection [42].
At this point the k-Means iterations can take place. The centroids are replaced by an arithmetic partition average (stated another way, the histograms of all the time series assigned to a centroid are averaged), and the partition is built again using the new centroids. This process is repeated until a convergence criterion is met or a number of iterations is reached [44]. In this work, a fixed number of 10 iterations was used to keep the computational cost low and constant. This can also come at the expense of a loss of performance, but the goal of the present study was not to design a classifier, but to illustrate how the estimated motif probabilities as a feature vector can enhance the discriminating power of PE.
In general, the time complexity of the standard k-means clustering algorithm is O ( l s k m ! ) , with l being the number of iterations, s being the number of objects (size of the dataset), k being the number of clusters and m ! being the dimension of the input vector [45]. In each iteration, the distance to each centroid from each object has to be computed. That makes s k distances. As each distance entails an arithmetic operation with each vector component, then each time a partition is computed needs at least s k m ! operations. As this is repeated at each iteration, the global computational cost can be roughly estimated as O ( l s k m ! ) , as stated. However, this is a very conservative upper bound for the algorithm complexity since this a worst case approach [46]. Moreover, further optimisations can be applied [47,48].
In practice, l < < s , k < < s and m < < s , being the number of time series s the main factor influencing the computational cost. However, m ! grows very fast, and the clustering task may soon become a high-dimensional clustering problem [49]. Fortunately, as already demonstrated in [23], only a few features, namely, histogram bins in the present paper, are responsible for differences among classes (or many bins are just 0), and therefore, more efficient sparse clustering algorithm versions can be used instead [50]. Besides, as m grows, the compensation probability among histogram bins decreases, as occurs for the sum of several dice [51] when the number of dice increases.
The performance of the method was quantified in terms of classification accuracy in all cases, given by the percentage of correctly classified times series. Further details of the k-means algorithm can be found in many scientific papers, such as in [32,52]. Improvements to the standard k-means algorithm used in this paper are also extensively described elsewhere [53,54,55,56].

2.3. Hidden Markov Models

HMM are stochastic processes that can be used to generate symbol sequences [57]. They consist of a number of nodes representing M states, Q = q 0 , q 1 , , q M - 1 . These nodes are interconnected by links related to a transition probability between any two states, in other words, the probability of being at state q j at time t + 1 if we were at state q i at previous time t. In principle, any HMM includes a bidimensional transition probability matrix A = a i j with M rows and M columns, in its simplest and more generic case. Thus, a i j = p S t a t e t + 1 = q j | S t a t e t = q i . There also exists another set of probabilities of emitting a specific symbol at each state q i ; but, in this work, each state generates a single entire motif (ordinal pattern) with probability 1, and therefore these probabilities will not be considered.
To discover the values of all the parameters involved in HMMs, three problems have to be solved first [57]:
  • To compute the probability of observing a certain input vector (Evaluation).
  • To find a state transition sequence that maximizes the probability of a certain input vector (Generation).
  • To induce a model that maximises the probability of a certain input vector (Learning).
However, this study uses HMM only for generation of synthetic time series based on constrains related to possible ordinal patterns. Moreover, the generation will be used in the opposite direction. That is, the input to the HMM will be the transition probabilities between states, and with a random excitation, a sequence of ordinal patterns will be obtained to generate a synthetic time series accordingly.
For m = 3 , the HMM used will have six states, from q 0 to q 5 . The correspondence between states and ordinal patterns will be q 0 ( 0 , 1 , 2 ) , q 1 ( 0 , 2 , 1 ) , q 2 ( 1 , 0 , 2 ) , q 3 ( 1 , 2 , 0 ) , q 4 ( 2 , 1 , 0 ) and q 5 ( 2 , 0 , 1 ) . This means that if, at time t, for example, state q 3 is reached, the ordinal pattern to consider must be (1,2,0), and therefore a synthetic random sample subsequence x i 3 following this order will have to be generated and appended to the complete time series. The topology will be ergodic, considering state transition probabilities a i j between any two states in general. The structure of this model is shown in Figure 2a. The initial state of the HMM, S t a t e t = 0 = q i , will be chosen randomly.
However, not all the transitions between ordinal patterns are possible, as there are two samples overlapping between consecutive subsequences. Therefore, the order and values of the last two samples of subsequence x i - 1 3 have to be the same for the two first samples of x i 3 . In practical terms, these forbidden transitions will be represented by a specific a i j = 0 , in contrast to the possible or admissible transitions, where a i j > 0 . This will be explained in more detail in next section. The resulting model applying these constraints is shown in Figure 2b, and this is the model that will be used to generate synthetic time series.

2.4. Experimental Dataset

The experimental dataset was composed of synthetic and real-life records. The synthetic time series were devised to fine tune the content in terms of ordinal patterns and enable the study of specific effects. This way, the study would be capable of unveiling the possible weaknesses of the methods assessed. The real-life dataset was chosen with the objective of translating into practical terms the lessons learnt with the synthetic data.

2.4.1. Synthetic Dataset

The creation of a controlled set of ordinal patterns is not a straightforward task. Even for low m values, there are many constraints that have to be taken into account. For example, as described in [58], and also stated above, not all the transitions between consecutive motifs in PE are possible. These transitions are summarised for m = 3 in Table 1. The rationale of this behaviour is that consecutive subsequences overlap. Namely, given an initial subsequence x i = x i , x i + 1 , x i + 2 , , x i + m - 1 , the next consecutive subsequence x i + 1 shares all the values with x i except the first one, as x i + 1 = x i + 1 , x i + 2 , x i + 3 , , x i + m . In other words, the last m - 1 samples of x i are the first m - 1 samples of x i + 1 . Consequently, the order found for x i + 1 , x i + 2 , x i + 3 , , has to be the same both in x i and in x i + 1 . For example, let x i 3 = 2.13 , 4.56 , 6.29 for m = 3 , with an associated ordinal pattern π i 3 = 0 , 1 , 2 . The next subsequence will have to be of the form x i + 1 = 4.56 , 6.29 , λ .
There are three possible regions where the value of λ can fall: λ < min ( 4.56 , 6.29 ) , min ( 4.56 , 6.29 ) < λ < max ( 4.56 , 6.29 ) or λ > max ( 4.56 , 6.29 ) (equal values are not included to avoid ties [59,60]). In the first case, it would correspond to an ordinal pattern π i 3 = 2 , 0 , 1 ; in the second case, to π i 3 = 0 , 2 , 1 and; in the last case, to π i 3 = 0 , 1 , 2 , being any other ordinal pattern impossible to achieve. Analytically, this could have been anticipated from all the motifs that respect the relative order of ( 4.56 , 6.29 ) ( 0 , 1 ): 2 , 0 , 1 , 0 , 2 , 1 and 0 , 1 , 2 . Although more complex, this reasoning can be extrapolated for m > 3 .
This behaviour of the transitions between consecutive motifs fits very well into a HMM [57]. The synthetic time series to be created can be considered an overlapping sequence of random values whose order is given by a motif from a finite and known alphabet. This knowledge of the motifs, their sequence and their probabilities of occurrence, enables the definition of a discrete time invariant stochastic model able to generate synthetic sequences. These sequences can exhibit a desired statistical distribution of motifs given the appropriate input excitation and state transition probabilities.
Of the initially possible 36 state transition probabilities, those 18 corresponding to forbidden transitions were permanently fixed to 0 in the HMM. The other 18 non-null state transition probabilities were grouped according to the state of origin of the transition. Thus, a 00 , a 01 and a 05 were considered a group of state transition probabilities from state q 0 , a 12 , a 13 and a 14 for q 1 , a 20 , a 21 and a 25 for q 2 , a 30 , a 31 and a 35 for q 3 , a 42 , a 43 , and a 44 for q 4 , and finally, a 52 , a 53 , a 54 for q 5 . To avoid configuring 18 probabilities for each realisation of the experiments with synthetic records, the probabilities were set using these groups of three. For example, when stated that probabilities were set as a i j = 1 10 , 2 10 , 7 10 , a 12 = 1 10 , a 13 = 2 10 , a 14 = 7 10 , a 20 = 1 10 , a 21 = 2 10 , a 25 = 7 10 and so on.
Using the model in Figure 2b, and a suitable random excitation, the time series of any length with subsequences whose ordinal patterns are equally or unequally probable can be generated easily. As an example, see Figure 3a, whose histogram in Figure 3b resembles that of a random time series, like the one shown in Figure 1d. The random excitation determines the final transitions that take place at each t from an initial random state. The path followed, in terms of q i , chains the ordinal patterns emitted, and from these ordinal patterns, subsequences of time series values are also randomly generated, ensuring they overlap with the previous ones and abide by the order emitted. These subsequences are generated ensuring also that ties do not take place, and therefore a clear falling or rising trend may be exhibited, depending on the dominant ordinal pattern, if any.
If the transition probabilities are not equal, what we obtain is a biased model instead that results in a nonuniform histogram of ordinal patterns. For example, Figure 4a shows a synthetic time series resulting from probabilities a i j = 1 10 , 1 10 , 8 10 . Its histogram is depicted on its right, in Figure 4b.
In the experiments, two classes of synthetic records were generated using the approach described, termed Class A and Class B, with two sets of different probabilities. A total of 100 sequences of length 1000 samples were randomly created for each class.

2.4.2. Real Dataset

The real dataset used in the experiments is composed of body temperature records. This dataset is the same previously used in other works [16], where additional details can be found. It contains 30 records and two classes. One class corresponds to a group of healthy individuals monitored during their daily activities: the control group. The size of this group is 16. The second group, the pathological group, contains 14 records of admitted patients at the Internal Medicine ward of the Teaching Hospital of Móstoles, Madrid. These patients developed a fever at least 24h before the temperature data acquisition started. An example of records of each class is shown in Figure 5.

3. Experiments and Results

The discriminating power of PE has been well demonstrated in a number of publications [16,24,60,61,62]. As a single feature, using a threshold, it has been possible to successfully classify a disparity of records of different types.
Initially, the experiments were devised to assess the sensitivity of PE in terms of histogram differences that led to PE differences as well. The two class synthetic records used in these experiments where generated using unbiased and biased versions of the HMM described above. The unbiased model using equal transition probabilities a i j = 1 3 , 1 3 , 1 3 (Class A) and biased versions, with unbalanced probabilities defined by the relationship a i j = α , α , 1 - 2 α , with 0 α 0.5 (Class B). In this case, PE was expected to achieve a classification accuracy close to 100% using records from the synthetic dataset, given that Class A and B are very different.
The results of this test are shown in Table 2. Class A was always fixed to an unbiased version (only the initial state was chosen at random), compared against different probability combinations featuring a biased histogram for Class B. The results confirmed that these two classes were easily separable, with poor classification performance only for values of α really close to the unbiased version of 1 3 . The classification accuracy is given as the average value and standard deviation of the 10 realisations of each experiment, where at least the initial state could vary.
In the next set of experiments, transition probabilities were more varied for both synthetic classes. The main objective in this case was to find histograms with similar amplitude bins but at different locations (motifs) in order to assess the possible PE loss of discriminating power. Note that the relationship between probabilities and motifs is given by the graphical structure of the model shown in Figure 2b. In other words, an asymmetry in the transition probabilities does not necessarily entail the same asymmetry in the histogram bins, since the state (motif emitter) can be reached from multiple paths. These results are shown in Table 3.
Some of the experiments in Table 3 are representative of the PE weakness addressed in the present paper: very similar relative frequencies in different motifs lead to the same PE value, and therefore classes become indistinguishable from a classification perspective. Specifically, this is the case generated by means of transition probabilities 1 2 , 0 , 1 2 and 0 , 1 2 , 1 2 . The time series generated using these probabilities are shown in Figure 6a,b, respectively. Despite having a very different motif distribution, the p ^ values are almost identical, at different positions, yielding similar PE values.
For example, for 1 2 , 0 , 1 2 , the main motifs were (0,1,2) and (2,0,1), followed by (1,0,2) and (1,2,0). Motifs (0,2,1) and (2,1,0) were negligible, as depicted in Figure 6c. On the other hand, for 0 , 1 2 , 1 2 , the main motifs were (1,0,2) and (2,1,0), then (0,2,1) and (2,1,0), with no ordinal patterns (0,1,2) and (1,2,0) (see Figure 6d). However, bin amplitudes were very similar, with an equal mapping on a single PE value.
However, the results in Table 3 for 1 2 , 0 , 1 2 and 1 2 , 1 2 , 0 , despite having a similar transition probability set of values, are completely different from an histogram perspective, and that is why the classification accuracy was 100%. This is very well illustrated in Figure 7a,b.
Figure 6 and Figure 7 very well summarise what can occur in a real case using PE as the distinguishing feature of a classification procedure (or any other similar mapping approach). Although the case in Figure 7 is the most frequent case, it is worth exploring alternatives to deal with poor time series classification performances based on PE, as this study proposes. Along this line, the results shown in Table 4 correspond to the same experiments as in Table 2 and Table 3, but using the relative frequencies instead and the clustering algorithm as described in Section 2.2.
The classification performance achieved in the experiments in Table 4 were clearly superior to those achieved using only PE, as hypothesised. For those cases where PE had a high discriminating power already, using the raw estimated probabilities vector and the k-Means clustering algorithm, such power was maintained or even slightly increased. For the specific case described in Figure 6, where PE failed as commented above, this time the performance achieved an expected 100%, given the differences in the motif distribution. Only when there were no significant differences, with equal transition probabilities, both methods obviously failed.
The experiments were repeated using real body temperature time series from the Control and Pathological datasets described in Section 2.4.2. The relative frequencies of each motif for the first subset are numerically shown in Table 5 for each record. The same for pathological records, in Table 6.
Temperature records of the experimental dataset exhibit the same behaviour than synthetic records with transition probabilities 1 2 , 0 , 1 2 and 0 , 1 2 , 1 2 . Using PE as the single classification feature, it was not possible to find differences between classes, with a nonsignificant accuracy of 60%. With the clustering and probability vector approach, this accuracy raised up to the 90%, with only three misplaced time series (19 objects assigned to the Control class, 16 really Control ones and the three errors, and 11 to the pathological one). The final centroids for each class were 0.3812 , 0.0928 , 0.0992 , 0.1166 , 0.1885 , 0.1229 and 0.3667 , 0.0931 , 0.0976 , 0.1177 , 0.2035 , 0.1218 .
Although it was not the objective of the present study to design a classifier, to better support the results using the estimated probabilities instead of PE, a leave-one-out (LOO) [63] classification analysis was conducted on the temperature data. A total of 100 realisations are used. In each realisation, a record from each class (with replacement) was randomly omitted in the k-means analysis. Then, the resulting centroids were used for classification of the omitted records using a nearest neighbour approach. The number of errors in this case were 18, that is, 18 records of the Pathological group were incorrectly classified as Control records, whereas no Control record was misclassified as Pathological. Therefore, the global accuracy using LOO and the probability vectors of the temperature records was 82%.

4. Discussion

The initial results in Table 2 give a sense of the PE general behaviour, in terms of discriminating power. For very significantly different transition probabilities, the classification performance easily reaches 100%. As these probabilities become closer, the error increases, and obviously, for the same models, it is impossible to distinguish between the two classes (54.8%). It is important to note that the probability window for which performance is not 100% is very narrow, from 0.33 to 0.3 accuracy goes from 54.8% up to 96.1%, which, in principle, confirms the high discriminating power of PE.
When both state transition probabilities are biased, the classification results are not 100% so often, as Table 3 illustrates. Moreover, although the probabilities are the same but in a different order, they result in significant differences in the histograms translated into a still high classification accuracy due to the asymmetry of the model (Figure 2b). In a few transition probabilities cases, as for 1 2 , 0 , 1 2 and 0 , 1 2 , 1 2 , these histogram differences can be numerically compensated and lose the discriminating capability, as hypothesised in this study.
Table 4 demonstrates this point by applying a clustering algorithm to the set of relative frequency values instead of the resulting PE value. For all the theoretically separable cases, this scheme yields a performance of at least 96%, and, in all cases, the performance is higher than that achieved using only PE. It is especially significant the 1 2 , 0 , 1 2 and 0 , 1 2 , 1 2 case, a clear representative of the possible detrimental effects of mapping to a single feature. From a nonsignificant classification accuracy of 55%, the clustering approach achieves a 100% accuracy, since the class differences are very apparent in the set of relative frequencies (Figure 6c,d). Obviously, all methods fail when the state transition probabilities are the same (51.1% accuracy in this case for 1 3 probabilities). Although with PE the result was 54.8% accuracy, it should not be interpreted as an improvement over the 51.1%, since both results were not significant, and correspond to a plain random guess that can exhibit a minor bias due to the limited number of realisations, 10 in this case.
Table 5 and Table 6 demonstrate, with real data, the possible detrimental impact on PE discriminating power mapping all the relative frequencies on a single scalar may exert, confirming what synthetic records already showed. The numerical values listed enable a detailed comparison of each motif contribution to the possible differences between classes. The main differences take place for patterns (0,1,2) and (2,1,0), at the first and fifth data columns, respectively. When using PE values, these differences become less noticeable, and that is why it is not possible to separate significantly the Control and Pathological groups of these records. Moreover, these relative frequencies go in opposite directions, that is, they are greater for the Control group in the first case ( 0.3905 > 0.3584 ) and smaller in the second case ( 0.1813 < 0.2087 ) , with relatively small standard deviations, and therefore compensation is more likely. It is also important to note that the final clustering centroids very well captured the histogram structure of both classes.
The LOO analysis provided a more realistic picture of the possible performance if a classifier had to be designed for the experimental dataset of temperature records. The accuracy went down slightly, from 90% using the entire dataset, down to 82% when a record of each class was omitted in the centroid computation. Anyway, this is still a very high classification accuracy, and it is important to note that PE alone was unable to find significant differences in this case, even when using all the records.
The computational cost using m ! features instead of a single PE value was obviously expected to be higher. Therefore, a trade–off between accuracy and execution time should be considered. Using a Windows© 8 computer with an Intel Core i7-4720HQ© 2.60GHz processor, with 16GB of RAM and C++ programming language, the execution time for the 100 LOO iterations using temperature records was 5.92s for m = 3 , 21.1s for m = 4 and 104.1s for m = 5 , without any algorithm optimisation. With a single feature, PE, the execution time was 2.48s.
The embedded dimension m is also related to the amount of information captured by a subsequence. A multiscale approach, with different time scales for PE [64,65], could also contribute to gain more insights into the dynamics of the time series. In this regard, it has been demonstrated that higher values of m frequently provide more discriminating power [24,60], as well as an optimal set of time delays [66]. However, using the approach proposed, it would be first necessary to assess the significance of the m ! features, even for each time delay, to keep the computational cost reasonably low.

5. Conclusions

The distinguishing power of PE is very high and is sufficient for many time series classification applications. PE has been successfully employed as a single feature in a myriad of technological and scientific fields.
However, sometimes the distinguishing features embedded in the histogram bins can become blurred when Equation 3 is applied. In order to avoid this problem, before discarding PE as the classification feature, or any other measure based on this kind of many-to-one mapping approach, we propose to look at the previous step of the PE algorithm and analyse the discriminating power of the relative frequencies instead.
The experimental set had to be chosen with this problem in mind. To control, up to a certain extent, the distribution of ordinal patterns in the time series used, we first developed a method based on HMM to create synthetic time series that satisfied certain motif emission constraints, and caused PE to fail at least in some cases. In addition, real-time series and body temperature records were also included, as they also exhibited this same problem under study.
The use of a vector of features instead of a single value can be seen as a typical multifeature extraction stage of a pattern recognition task. Almost any clustering algorithm is perfectly suited to deal with this task. For illustrative purposes only—not for optimisation—we applied a classical k - Means algorithm and a standard configuration.
The results using this approach confirmed, for both datasets, that using the original histogram values, the discriminating power of PE can be enhanced. As a consequence, we propose, when possible, to analyse the information provided by each histogram bin jointly, as is the case in the present study, or separately, in a kind of feature selection analysis, as in [23], to maximise the information provided by ordinal pattern distribution beyond the scalar PE value. This approach can be arguably be exported to many other methods, and open a new line of research combining event-counting metrics with pattern recognition or machine learning algorithms.
Obviously, if there are no differences at the histogram level; for example, when temporal correlations are the same, any method based on such information will fail, that is, both PE and the method proposed. This scenario can be illustrated trying to distinguish time series with random Gaussian or uniform amplitude distributions. Despite the amplitude differences, it was not possible to distinguish between both classes using PE or the method proposed. For such cases, we conducted a few additional tests using measures better suited to amplitude differences in order to propose alternatives to overcome this drawback. The most straightforward approach is to use methods such as ApEn or SampEn, both amplitude-based, which achieved a classification accuracy close to 100% using Gaussian and uniform amplitude distributions for classes. A solution more related to the present paper was to use a PE derivative that included amplitude information [24]. In this regard, Weighted Permutation Entropy [67] also achieved a very high classification accuracy, well above 80%.

Author Contributions

D.C.-F. designed and conducted the experiments, and wrote the paper. A.M.-P. implemented and tweaked the algorithms. B.V. acquired, preprocessed and selected the body temperature records. Conceptualization, D.C.-F.; Data curation, B.V. and P.G.; Methodology, D.C.-F.; Software, A.M.-P.; Validation, B.V.; Writing-original draft, D.C.-F.

Funding

Ministerio de Ciencia, Innovación y Universidades: PTQ–16–08538.

Acknowledgments

David Cuesta-Frau and Antonio Molina were supported by the Torres Quevedo program of the Spanish Ministry of Science, code PTQ-16-08538. Borja Vargas and Paula González were supported by Instituto de Salud Carlos III through the project PI17/00856 and cofinanced by the European Development Fund “A way to make Europe”. We also thank the reviewers for their useful comments, which have greatly contributed to improve the quality and soundness of the paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Esling, P.; Agon, C. Time-series Data Mining. ACM Comput. Surv. 2012, 45, 12:1–12:34. [Google Scholar] [CrossRef]
  2. Bagnall, A.; Lines, J.; Bostrom, A.; Large, J.; Keogh, E. The great time series classification bake off: A review and experimental evaluation of recent algorithmic advances. Data Min. Knowl. Discov. 2017, 31, 606–660. [Google Scholar] [CrossRef]
  3. Tabar, Y.R.; Halici, U. A novel deep learning approach for classification of EEG motor imagery signals. J. Neural Eng. 2016, 14, 016003. [Google Scholar] [CrossRef]
  4. Cuesta-Frau, D.; Biagetti, M.; Quinteiro, R.; Micó, P.; Aboy, M. Unsupervised classification of ventricular extrasystoles using bounded clustering algorithms and morphology matching. Med. Biol. Eng. Comput. 2007, 45, 229–239. [Google Scholar] [CrossRef] [PubMed]
  5. Dakappa, P.H.; Prasad, K.; Rao, S.B.; Bolumbu, G.; Bhat, G.K.; Mahabala, C. Classification of Infectious and Noninfectious Diseases Using Artificial Neural Networks from 24-Hour Continuous Tympanic Temperature Data of Patients with Undifferentiated Fever. Crit. Rev. Biomed. Eng. 2018, 46, 173–183. [Google Scholar] [CrossRef]
  6. Wang, C.C.; Kang, Y.; Shen, P.C.; Chang, Y.P.; Chung, Y.L. Applications of fault diagnosis in rotating machinery by using time series analysis with neural network. Expert Syst. Appl. 2010, 37, 1696–1702. [Google Scholar] [CrossRef]
  7. Fong, S.; Lan, K.; Wong, R. Classifying Human Voices By Using Hybrid SFX Time-series Pre-processing and Ensemble Feature Selection. Biomed Res. Int. 2013, 2013, 1–27. [Google Scholar]
  8. Lines, J.; Bagnall, A.; Caiger-Smith, P.; Anderson, S. Classification of Household Devices by Electricity Usage Profiles. In Intelligent Data Engineering and Automated Learning-IDEAL; Yin, H., Wang, W., Rayward-Smith, V., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; pp. 403–412. [Google Scholar] [Green Version]
  9. Papaioannou, V.E.; Chouvarda, I.G.; Maglaveras, N.K.; Baltopoulos, G.I.; Pneumatikos, I.A. Temperature multiscale entropy analysis: A promising marker for early prediction of mortality in septic patients. Physiol. Meas. 2013, 34, 1449. [Google Scholar] [CrossRef]
  10. Cuesta-Frau, D.; Miró-Martínez, P.; Jordán-Núnez, J.; Oltra-Crespo, S.; Molina-Picó, A. Noisy EEG signals classification based on entropy metrics. Performance assessment using first and second generation statistics. Comput. Biol. Med. 2017, 87, 141–151. [Google Scholar] [CrossRef]
  11. Li, P.; Karmakar, C.; Yan, C.; Palaniswami, M.; Liu, C. Classification of 5-S Epileptic EEG Recordings Using Distribution Entropy and Sample Entropy. Front. Physiol. 2016, 7, 136. [Google Scholar] [CrossRef] [Green Version]
  12. Azami, H.; Escudero, J. Amplitude-aware permutation entropy: Illustration in spike detection and signal segmentation. Comput. Meth. Programs Biomed. 2016, 128, 40–51. [Google Scholar] [CrossRef] [PubMed]
  13. Chen, Z.; Yaan, L.; Liang, H.; Yu, J. Improved Permutation Entropy for Measuring Complexity of Time Series under Noisy Condition. Complexity 2019, 2019, 1403829. [Google Scholar] [CrossRef]
  14. Manis, G.; Aktaruzzaman, M.; Sassi, R. Bubble Entropy: An Entropy Almost Free of Parameters. IEEE Trans. Biomed. Eng. 2017, 64, 2711–2718. [Google Scholar] [PubMed]
  15. Simons, S.; Espino, P.; Abásolo, D. Fuzzy Entropy Analysis of the Electroencephalogram in Patients with Alzheimer’s Disease: Is the Method Superior to Sample Entropy? Entropy 2018, 20, 21. [Google Scholar] [CrossRef]
  16. Cuesta-Frau, D.; Miró-Martínez, P.; Oltra-Crespo, S.; Jordán-Núñez, J.; Vargas, B.; González, P.; Varela-Entrecanales, M. Model Selection for Body Temperature Signal Classification Using Both Amplitude and Ordinality-Based Entropy Measures. Entropy 2018, 20, 853. [Google Scholar] [CrossRef]
  17. Karmakar, C.; Udhayakumar, R.K.; Li, P.; Venkatesh, S.; Palaniswami, M. Stability, Consistency and Performance of Distribution Entropy in Analysing Short Length Heart Rate Variability (HRV) Signal. Front. Physiol. 2017, 8, 720. [Google Scholar] [CrossRef]
  18. Amigó, J. Permutation Complexity in Dynamical Systems: Ordinal Patterns, Permutation Entropy and All That; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  19. Greven, A.; Keller, G.; Warnecke, G. Entropy; Princeton University Press: Princeton, NJ, USA, 2014. [Google Scholar]
  20. Cruces, S.; Martín-Clemente, R.; Samek, W. Information Theory Applications in Signal Processing. Entropy 2019, 21, 653. [Google Scholar] [CrossRef]
  21. Shannon, C.E.; Weaver, W. The Mathematical Theory of Communication; The University of Illinois Press: Urbana, IL, USA, 1949. [Google Scholar]
  22. Zunino, L.; Zanin, M.; Tabak, B.M.; Pérez, D.G.; Rosso, O.A. Forbidden patterns, permutation entropy and stock market inefficiency. Physica A 2009, 388, 2854–2864. [Google Scholar] [CrossRef]
  23. Cuesta-Frau, D.; Murillo-Escobar, J.P.; Orrego, D.A.; Delgado-Trejos, E. Embedded Dimension and Time Series Length. Practical Influence on Permutation Entropy and Its Applications. Entropy 2019, 21, 385. [Google Scholar] [CrossRef]
  24. Cuesta–Frau, D. Permutation entropy: Influence of amplitude information on time series classification performance. Math. Biosci. Eng. 2019, 16, 6842. [Google Scholar] [CrossRef]
  25. Bandt, C.; Pompe, B. Permutation Entropy: A Natural Complexity Measure for Time Series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef] [PubMed]
  26. Parlitz, U.; Berg, S.; Luther, S.; Schirdewan, A.; Kurths, J.; Wessel, N. Classifying cardiac biosignals using ordinal pattern statistics and symbolic dynamics. Comput. Biol. Med. 2012, 42, 319–327. [Google Scholar] [CrossRef] [PubMed]
  27. Zanin, M. Forbidden patterns in financial time series. Chaos 2008, 18, 013119. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Kulp, C.; Chobot, J.; Niskala, B.; Needhammer, C. Using Forbidden Patterns To Detect Determinism in Irregularly Sampled Time Series. Chaos 2016, 26, 023107. [Google Scholar] [CrossRef]
  29. Tzortzis, G.; Likas, A. The MinMax k–Means clustering algorithm. Pattern Recognit. 2014, 47, 2505–2516. [Google Scholar] [CrossRef]
  30. Xu, D.; Tian, Y. A Comprehensive Survey of Clustering Algorithms. AODS 2015, 2, 165–193. [Google Scholar] [CrossRef] [Green Version]
  31. Rodriguez, M.Z.; Comin, C.H.; Casanova, D.; Bruno, O.M.; Amancio, D.R.; Costa, L.d.F.; Rodrigues, F.A. Clustering algorithms: A comparative approach. PLoS ONE 2019, 14, e0210236. [Google Scholar] [CrossRef]
  32. Yu, S.S.; Chu, S.W.; Wang, C.M.; Chan, Y.K.; Chang, T.C. Two improved k-means algorithms. Appl. Soft. Comput. 2018, 68, 747–755. [Google Scholar] [CrossRef]
  33. Jain, A.K. Data clustering: 50 years beyond K-means. Pattern Recognit. Lett. 2010, 31, 651–666. [Google Scholar] [CrossRef]
  34. Wu, J. Advances in K-means Clustering: A Data Mining Thinking; Springer: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  35. Cuesta-Frau, D.; Pérez-Cortes, J.C.; García, G.A. Clustering of electrocardiograph signals in computer-aided Holter analysis. Comput. Meth. Programs Biomed. 2003, 72, 179–196. [Google Scholar] [CrossRef]
  36. Rodríguez-Sotelo, J.; Peluffo-Ordoñez, D.; Cuesta-Frau, D.; Castellanos-Domínguez, G. Unsupervised feature relevance analysis applied to improve ECG heartbeat clustering. Comput. Meth. Programs Biomed. 2012, 108, 250–261. [Google Scholar] [CrossRef] [PubMed]
  37. Rodríguez-Sotelo, J.L.; Osorio-Forero, A.; Jiménez-Rodríguez, A.; Cuesta-Frau, D.; Cirugeda-Roldán, E.; Peluffo, D. Automatic Sleep Stages Classification Using EEG Entropy Features and Unsupervised Pattern Analysis Techniques. Entropy 2014, 16, 6573. [Google Scholar]
  38. Gower, J.C.; Legendre, P. Metric and Euclidean properties of dissimilarity coefficients. J. Classif. 1986, 3, 5–48. [Google Scholar] [CrossRef]
  39. Pakhira, M.K. Finding Number of Clusters before Finding Clusters. Proc. Tech. 2012, 4, 27–37. [Google Scholar] [CrossRef] [Green Version]
  40. Poomagal, S.; Saranya, P.; Karthik, S. A Novel Method for Selecting Initial Centroids in K-means Clustering Algorithm. Int. J. Intell. Syst. Technol. Appl. 2016, 15, 230–239. [Google Scholar] [CrossRef]
  41. Kuncheva, L.I.; Vetrov, D.P. Evaluation of Stability of k-Means Cluster Ensembles with Respect to Random Initialization. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28, 1798–1808. [Google Scholar] [CrossRef]
  42. Fränti, P.; Sieranoja, S. How much can k-means be improved by using better initialization and repeats? Pattern Recognit. 2019, 93, 95–112. [Google Scholar] [CrossRef]
  43. Yuan, B.; Zhang, W.; Yuan, Y. A Max-Min clustering method for k-Means algorithm of data clustering. J. Ind. Manag. Optim. 2012, 8, 565. [Google Scholar] [CrossRef]
  44. Pérez, O.J.; Pazos, R.R.; Cruz, R.L.; Reyes, S.G.; Basave, T.R.; Fraire, H.H. Improving the Efficiency and Efficacy of the K-means Clustering Algorithm Through a New Convergence Condition. In International Conference on Computational Science and Its Applications; Gervasi, O., Gavrilova, M.L., Eds.; Springer: Berlin/Heidelberg, Germany, 2007; pp. 674–682. [Google Scholar]
  45. Osamor, V.C.; Adebiyi, E.F.; Oyelade, J.O.; Doumbia, S. Reducing the Time Requirement of k-Means Algorithm. PLoS ONE 2012, 7, e49946. [Google Scholar] [CrossRef]
  46. Har-Peled, S.; Sadri, B. How Fast Is the k-Means Method? Algorithmica 2005, 41, 185–202. [Google Scholar] [CrossRef]
  47. Lai, J.Z.; Huang, T.J.; Liaw, Y.C. A fast k–means clustering algorithm using cluster center displacement. Pattern Recognit. 2009, 42, 2551–2556. [Google Scholar] [CrossRef]
  48. Celebi, M.E.; Kingravi, H.A.; Vela, P.A. A comparative study of efficient initialization methods for the k-means clustering algorithm. Expert Syst. Appl. 2013, 40, 200–210. [Google Scholar] [CrossRef] [Green Version]
  49. Sun, W.; Wang, J.; Fang, Y. Regularized k-means clustering of high-dimensional data and its asymptotic consistency. Electron. J. Stat. 2012, 6, 148–167. [Google Scholar] [CrossRef]
  50. Gong, W.; Zhao, R.; Grünewald, S. Structured sparse K-means clustering via Laplacian smoothing. Pattern Recognit. Lett. 2018, 112, 63–69. [Google Scholar] [CrossRef]
  51. The Probability Distribution of the Sum of Several Dice: Slot Applications. UNLV Gaming Res. Rev. J. 2011, 15, 10.
  52. Jain, A.K.; Murty, M.N.; Flynn, P.J. Data Clustering: A Review. ACM Comput. Surv. 1999, 31, 264–323. [Google Scholar] [CrossRef]
  53. Karimov, J.; Ozbayoglu, M. Clustering Quality Improvement of k-means Using a Hybrid Evolutionary Model. Procedia. Comput. Sci. 2015, 61, 38–45. [Google Scholar] [CrossRef] [Green Version]
  54. Rodriguez-Sotelo, J.L.; Cuesta-Frau, D.; Castellanos-Dominguez, G. An improved method for unsupervised analysis of ECG beats based on WT features and J-means clustering. In 2007 Computers in Cardiology; IEEE: Piscataway, NJ, USA, 2007; pp. 581–584. [Google Scholar] [Green Version]
  55. Panda, S.; Sahu, S.; Jena, P.; Chattopadhyay, S. Comparing Fuzzy-C Means and K-Means Clustering Techniques: A Comprehensive Study. In Advances in Computer Science, Engineering & Applications; Wyld, D.C., Zizka, J., Nagamalai, D., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 451–460. [Google Scholar]
  56. Bahmani, B.; Moseley, B.; Vattani, A.; Kumar, R.; Vassilvitskii, S. Scalable K-means++. Proc. VLDB Endow. 2012, 5, 622–633. [Google Scholar] [CrossRef]
  57. Rabiner, L.R. A tutorial on hidden Markov models and selected applications in speech recognition. Proc. IEEE 1989, 77, 257–286. [Google Scholar] [CrossRef]
  58. Unakafova, V.; Keller, K. Efficiently Measuring Complexity on the Basis of Real-World Data. Entropy 2013, 15, 4392–4415. [Google Scholar] [CrossRef] [Green Version]
  59. Zunino, L.; Olivares, F.; Scholkmann, F.; Rosso, O.A. Permutation entropy based time series analysis: Equalities in the input signal can lead to false conclusions. Phys. Lett. A 2017, 381, 1883–1892. [Google Scholar] [CrossRef] [Green Version]
  60. Cuesta-Frau, D.; Varela-Entrecanales, M.; Molina-Picó, A.; Vargas, B. Patterns with Equal Values in Permutation Entropy: Do They Really Matter for Biosignal Classification? Complexity 2018, 2018, 1324696. [Google Scholar] [CrossRef]
  61. Keller, K.; Unakafov, A.M.; Unakafova, V.A. Ordinal Patterns, Entropy, and EEG. Entropy 2014, 16, 6212–6239. [Google Scholar] [CrossRef]
  62. Cuesta-Frau, D.; Miró-Martínez, P.; Oltra-Crespo, S.; Jordán-Núñez, J.; Vargas, B.; Vigil, L. Classification of glucose records from patients at diabetes risk using a combined permutation entropy algorithm. Comput. Meth. Programs Biomed. 2018, 165, 197–204. [Google Scholar] [CrossRef]
  63. Arlot, S.; Celisse, A. A survey of cross-validation procedures for model selection. Statist. Surv. 2010, 4, 40–79. [Google Scholar] [CrossRef]
  64. Li, D.; Li, X.; Liang, Z.; Voss, L.J.; Sleigh, J.W. Multiscale permutation entropy analysis of EEG recordings during sevoflurane anesthesia. J. Neural Eng. 2010, 7, 046010. [Google Scholar] [CrossRef]
  65. Liu, T.; Yao, W.; Wu, M.; Shi, Z.; Wang, J.; Ning, X. Multiscale permutation entropy analysis of electrocardiogram. Physica A 2017, 471, 492–498. [Google Scholar] [CrossRef]
  66. Tao, M.; Poskuviene, K.; Alkayem, N.; Cao, M.; Ragulskis, M. Permutation Entropy Based on Non-Uniform Embedding. Entropy 2018, 20, 612. [Google Scholar] [CrossRef]
  67. Fadlallah, B.; Chen, B.; Keil, A.; Príncipe, J. Weighted-permutation entropy: A complexity measure for time series incorporating amplitude information. Phys. Rev. E 2013, 87, 022911. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Examples of histograms from well-known synthetic time series. Some features of these series are clearly reflected in the histograms, with a clear correlation between the distribution of the motifs and the determinism degree of the records.
Figure 1. Examples of histograms from well-known synthetic time series. Some features of these series are clearly reflected in the histograms, with a clear correlation between the distribution of the motifs and the determinism degree of the records.
Entropy 21 01013 g001
Figure 2. Basic Markov models for consecutive ordinal patterns in a time series for m = 3 .
Figure 2. Basic Markov models for consecutive ordinal patterns in a time series for m = 3 .
Entropy 21 01013 g002
Figure 3. Example of a synthetic record generated using an unbiased HMM and its resulting uniform motif relative frequencies.
Figure 3. Example of a synthetic record generated using an unbiased HMM and its resulting uniform motif relative frequencies.
Entropy 21 01013 g003
Figure 4. Example of a synthetic record generated using a biased version of the Markov model and its resulting nonuniform motif relative frequencies.
Figure 4. Example of a synthetic record generated using a biased version of the Markov model and its resulting nonuniform motif relative frequencies.
Entropy 21 01013 g004
Figure 5. Example of records from the two classes of body temperature data. Top: Pathological. Bottom: Control.
Figure 5. Example of records from the two classes of body temperature data. Top: Pathological. Bottom: Control.
Entropy 21 01013 g005
Figure 6. Equivalent histograms from the PE point of view despite clear differences in ordinal pattern distribution. When the estimated probability is computed, the PE measure is more or less the same since the differences in bin locations are lost, only amplitudes matter.
Figure 6. Equivalent histograms from the PE point of view despite clear differences in ordinal pattern distribution. When the estimated probability is computed, the PE measure is more or less the same since the differences in bin locations are lost, only amplitudes matter.
Entropy 21 01013 g006
Figure 7. Contrary to case depicted in Figure 6, differences in ordinal pattern distribution are supported by differences in histogram amplitudes, which results in significant PE differences as well.
Figure 7. Contrary to case depicted in Figure 6, differences in ordinal pattern distribution are supported by differences in histogram amplitudes, which results in significant PE differences as well.
Entropy 21 01013 g007
Table 1. Admissible and forbidden transitions between consecutive motifs for m = 3 .
Table 1. Admissible and forbidden transitions between consecutive motifs for m = 3 .
Initial MotifAdmissible Next MotifsForbidden Next Motifs
0 , 1 , 2 0 , 1 , 2 1 , 2 , 0
0 , 2 , 1 2 , 1 , 0
2 , 0 , 1 1 , 0 , 2
0 , 2 , 1 1 , 2 , 0 0 , 1 , 2
2 , 1 , 0 0 , 2 , 1
1 , 0 , 2 2 , 0 , 1
1 , 2 , 0 0 , 1 , 2 1 , 2 , 0
0 , 2 , 1 2 , 1 , 0
2 , 0 , 1 1 , 0 , 2
2 , 0 , 1 1 , 2 , 0 0 , 1 , 2
2 , 1 , 0 0 , 2 , 1
1 , 0 , 2 2 , 0 , 1
2 , 1 , 0 1 , 2 , 0 0 , 1 , 2
2 , 1 , 0 0 , 2 , 1
1 , 0 , 2 2 , 0 , 1
1 , 0 , 2 0 , 1 , 2 1 , 2 , 0
0 , 2 , 1 2 , 1 , 0
2 , 0 , 1 1 , 0 , 2
Table 2. Classification performance of permutation entropy (PE) as a single feature when one class of synthetic records was generated using a balanced set of probabilities, and the other class using a varied set of probabilities.
Table 2. Classification performance of permutation entropy (PE) as a single feature when one class of synthetic records was generated using a balanced set of probabilities, and the other class using a varied set of probabilities.
Class AClass BClassification
Transition ProbabilitiesTransition ProbabilitiesAccuracy (%)
1 3 , 1 3 , 1 3 1 10 , 1 10 , 8 10 100 ± 0
1 3 , 1 3 , 1 3 1 9 , 1 9 , 7 9 100 ± 0
1 3 , 1 3 , 1 3 1 8 , 1 8 , 6 8 100 ± 0
1 3 , 1 3 , 1 3 1 7 , 1 7 , 5 7 100 ± 0
1 3 , 1 3 , 1 3 1 6 , 1 6 , 4 6 100 ± 0
1 3 , 1 3 , 1 3 1 5 , 1 5 , 3 5 100 ± 0
1 3 , 1 3 , 1 3 2 7 , 2 7 , 3 7 99.5 ± 0.57
1 3 , 1 3 , 1 3 4 13 , 4 13 , 5 13 87.4 ± 1.88
1 3 , 1 3 , 1 3 1 3 , 1 3 , 1 3 54.8 ± 1.20
1 3 , 1 3 , 1 3 3 10 , 3 10 , 4 10 96.1 ± 1.65
1 3 , 1 3 , 1 3 1 4 , 1 4 , 2 4 100 ± 0
Table 3. Classification performance of PE as a single feature when both synthetic classes are generated using a biased set of probabilities.
Table 3. Classification performance of PE as a single feature when both synthetic classes are generated using a biased set of probabilities.
Class AClass BClassification
Transition ProbabilitiesTransition ProbabilitiesAccuracy (%)
12 20 , 5 20 , 3 20 3 20 , 5 20 , 12 20 87.5 ± 2.87
13 20 , 4 20 , 3 20 3 20 , 4 20 , 13 20 95.3 ± 0.83
10 20 , 6 20 , 4 20 9 20 , 3 20 , 8 20 85.4 ± 1.43
1 2 , 0 , 1 2 0 , 1 2 , 1 2 55.0 ± 2.89 (Figure 6)
1 2 , 0 , 1 2 1 2 , 1 2 , 0 100 ± 0 (Figure 7)
Table 4. Classification performance achieved using the 6 estimated probability values as a feature vector instead of the single PE value, for the previous experiments.
Table 4. Classification performance achieved using the 6 estimated probability values as a feature vector instead of the single PE value, for the previous experiments.
Class AClass BClassification
Transition ProbabilitiesTransition ProbabilitiesAccuracy (%)
1 3 , 1 3 , 1 3 1 10 , 1 10 , 8 10 100 ± 0
1 3 , 1 3 , 1 3 1 5 , 1 5 , 3 5 100 ± 0
1 3 , 1 3 , 1 3 2 7 , 2 7 , 3 7 99.7 ± 0.51
1 3 , 1 3 , 1 3 4 13 , 4 13 , 5 13 96.0 ± 2.23
1 3 , 1 3 , 1 3 1 3 , 1 3 , 1 3 51.1 ± 1.74
1 3 , 1 3 , 1 3 3 10 , 3 10 , 4 10 99.0 ± 0.57
12 20 , 5 20 , 3 20 3 20 , 5 20 , 12 20 100 ± 0
13 20 , 4 20 , 3 20 3 20 , 4 20 , 13 20 100 ± 0
10 20 , 6 20 , 4 20 9 20 , 3 20 , 8 20 100 ± 0
1 2 , 0 , 1 2 0 , 1 2 , 1 2 100 ± 0
1 2 , 0 , 1 2 1 2 , 1 2 , 0 100 ± 0
Table 5. Individual results for the estimated frequencies for each motif using the control records.
Table 5. Individual results for the estimated frequencies for each motif using the control records.
Record(0,1,2)(0,2,1)(1,0,2)(1,2,0)(2,1,0)(2,0,1)
Control000.40580.08360.09410.12550.15480.1359
Control010.53760.08990.08360.08780.11920.0815
Control020.38490.07530.08360.10250.24260.1108
Control030.38700.07320.08780.11290.20920.1297
Control040.32420.12760.12550.11710.19030.1150
Control050.35980.09620.09830.12130.20080.1234
Control060.39120.08150.06480.14010.19870.1234
Control070.34100.10040.13170.08150.23220.1129
Control080.33680.09410.09830.11710.23010.1234
Control090.31590.11710.12550.12340.18410.1338
Control100.45600.09200.09620.11290.12760.1150
Control110.35560.10040.11710.12130.16940.1359
Control120.53340.07320.07530.08990.13800.0899
Control130.26150.10660.12340.13170.23010.1464
Control140.40580.08360.09410.12550.15480.1359
Control150.45180.07740.08780.12760.11920.1359
Mean0.39050.09200.09920.11490.18130.1218
StdDev0.07510.01570.01980.01650.04210.0174
Table 6. Individual results for the estimated frequencies for each motif using the pathological records.
Table 6. Individual results for the estimated frequencies for each motif using the pathological records.
Record(0,1,2)(0,2,1)(1,0,2)(1,2,0)(2,1,0)(2,0,1)
Pathological000.40540.08380.09200.10850.19380.1161
Pathological010.32600.09600.10360.12070.22440.1290
Pathological020.44960.06590.06830.11630.18220.1175
Pathological030.34050.09800.09260.10890.25340.1062
Pathological040.21910.12710.13480.13740.23470.1465
Pathological050.27360.09790.10830.13480.24070.1444
Pathological060.38050.09790.09920.12040.17930.1224
Pathological070.33310.09500.09500.13470.20720.1347
Pathological080.43610.07110.06600.12190.18730.1174
Pathological090.31030.10650.10190.12870.22830.1241
Pathological100.43420.07560.09190.09680.18810.1131
Pathological110.34470.12240.11550.12650.17160.1190
Pathological120.35900.09720.10720.10540.21540.1154
Pathological130.40560.08380.09430.09500.21540.1056
Mean0.35840.09420.09790.11830.20870.1222
StdDev0.06570.01740.01730.01370.02550.0125

Share and Cite

MDPI and ACS Style

Cuesta-Frau, D.; Molina-Picó, A.; Vargas, B.; González, P. Permutation Entropy: Enhancing Discriminating Power by Using Relative Frequencies Vector of Ordinal Patterns Instead of Their Shannon Entropy. Entropy 2019, 21, 1013. https://doi.org/10.3390/e21101013

AMA Style

Cuesta-Frau D, Molina-Picó A, Vargas B, González P. Permutation Entropy: Enhancing Discriminating Power by Using Relative Frequencies Vector of Ordinal Patterns Instead of Their Shannon Entropy. Entropy. 2019; 21(10):1013. https://doi.org/10.3390/e21101013

Chicago/Turabian Style

Cuesta-Frau, David, Antonio Molina-Picó, Borja Vargas, and Paula González. 2019. "Permutation Entropy: Enhancing Discriminating Power by Using Relative Frequencies Vector of Ordinal Patterns Instead of Their Shannon Entropy" Entropy 21, no. 10: 1013. https://doi.org/10.3390/e21101013

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop