Next Article in Journal / Special Issue
Permutation Entropy: An Ordinal Pattern-Based Resilience Indicator for Industrial Equipment
Previous Article in Journal
Exact Expressions for Kullback–Leibler Divergence for Univariate Distributions
Previous Article in Special Issue
Legendre Polynomial Fitting-Based Permutation Entropy Offers New Insights into the Influence of Fatigue on Surface Electromyography (sEMG) Signal Complexity
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Generalized Gaussian Distribution Improved Permutation Entropy: A New Measure for Complex Time Series Analysis

1
Institute of Visual Informatics, National University of Malaysia (UKM), Bangi 43600, Selangor, Malaysia
2
School of Information and Communication, Guilin University of Electronic Technology, Guilin 541004, China
3
School of AI and Advanced Computing, XJTLU Entrepreneur College (Taicang), Xi’an Jiaotong—Liverpool University, Suzhou 215400, China
4
School of Intelligent Manufacturing Ecosystem, XJTLU Entrepreneur College (Taicang), Xi’an Jiaotong—Liverpool University, Suzhou 215400, China
5
Cognitive Radio and Information Processing Key Laboratory Authorized by China’s Ministry of Education Foundation, Guilin University of Electronic Technology, Guilin 541004, China
*
Authors to whom correspondence should be addressed.
Entropy 2024, 26(11), 960; https://doi.org/10.3390/e26110960
Submission received: 10 September 2024 / Revised: 10 October 2024 / Accepted: 15 October 2024 / Published: 7 November 2024
(This article belongs to the Special Issue Ordinal Pattern-Based Entropies: New Ideas and Challenges)

Abstract

:
To enhance the performance of entropy algorithms in analyzing complex time series, generalized Gaussian distribution improved permutation entropy (GGDIPE) and its multiscale variant (MGGDIPE) are proposed in this paper. First, the generalized Gaussian distribution cumulative distribution function is employed for data normalization to enhance the algorithm’s applicability across time series with diverse distributions. The algorithm further processes the normalized data using improved permutation entropy, which maintains both the absolute magnitude and temporal correlations of the signals, overcoming the equal value issue found in traditional permutation entropy (PE). Simulation results indicate that GGDIPE is less sensitive to parameter variations, exhibits strong noise resistance, accurately reveals the dynamic behavior of chaotic systems, and operates significantly faster than PE. Real-world data analysis shows that MGGDIPE provides markedly better separability for RR interval signals, EEG signals, bearing fault signals, and underwater acoustic signals compared to multiscale PE (MPE) and multiscale dispersion entropy (MDE). Notably, in underwater target recognition tasks, MGGDIPE achieves a classification accuracy of 97.5% across four types of acoustic signals, substantially surpassing the performance of MDE (70.5%) and MPE (62.5%). Thus, the proposed method demonstrates exceptional capability in processing complex time series.

1. Introduction

The primary objective of complex time series analysis is to classify and identify signals by extracting their distinctive features or patterns. To perform this task, entropy-based methods, which enhance the understanding of underlying dynamics, have garnered more and more attention in recent years [1,2]. These methods have been successfully applied across various complex engineering domains, including electrocardiogram (ECG) and electroencephalographic (EEG) records analysis [3,4,5], bearing fault diagnosis [6], and underwater target recognition [7,8].
The prevalent entropy metrics can be broadly classified into two main categories: those that are conditional entropy-based and those that are Shannon entropy-based. The former includes widely used methods such as Approximate Entropy (ApEn) [9], Sample Entropy (SampEn) [10], and Fuzzy Entropy (FuzzyEn) [11]. The core of these measures is to compare the probabilities of similar patterns occurring in m dimensional and m + 1 dimensional phase spaces to infer the emergence of new patterns. Pattern similarity is determined by whether the distance between vectors falls below a predefined threshold. For example, if D X i m , X j m r and D X i m + 1 , X j m + 1 > r , where D x , y represents the distance between x and y , and r is a constant, it suggests the emergence of a new pattern when the embedding dimension is enlarged from m to m + 1 . Generally, the more components and complexity within a time series, the higher the probability of new patterns emerging, resulting in a higher entropy value. The main drawback of the aforementioned methods lies in their requirement for extensive distance calculation within the phase space, which results in high computational complexity. Furthermore, when dealing with short data sequences, these methods are prone to generate undefined entropy values.
Without being exhaustive, Shannon entropy-based approaches include Permutation Entropy (PE) [12], Distribution Entropy (DistEn) [13], Weighted Permutation Entropy (WPE) [14], Modified Permutation Entropy (mPE) [15], Dispersion Entropy (DispEn) [16], Slope entropy (SlopEn) [17], and Attention Entropy (AE) [18], among others. Permutation entropy, introduced by Bandt et al. [12] estimates probability density based on the frequency of ordinal patterns (OP). Its conceptual simplicity and computational efficiency have quickly made it one of the most prominent entropy algorithms [2,17,19,20]. Subsequently, WPE was proposed to address the limitation of PE in neglecting absolute amplitude information. In WPE, weights are assigned to each OP to calculate the weighted probability density [14]. Bian and the coauthors addressed the distortion in PE estimation that can occur under low-resolution conditions. They accomplished this by mapping equal values within sub-sequences to the same symbol and then calculating the probability distribution of the resulting symbolic patterns [15]. In 2016, Rostaghi et al. [16] developed DispEn, which estimates entropy by calculating the probability density of newly designed symbolic patterns known as dispersion patterns. Compared to PE and its variants, DispEn demonstrates a superior ability to distinguish complex time series [16,19,20]. In 2019, Cuesta-Frau invented SlopEn, which utilizes a new encoding technique based on the gradient between two consecutive sample points [17]. However, a key challenge with SlopEn is the lack of theoretical guidance for determining certain parameters, which may affect the algorithm’s effectiveness. Unlike PE, DispEn, and SlopEn, AE exclusively focuses on the frequency distribution of intervals between key observations in a time series, such as the intervals between local maxima [18]. This parameter-free approach demonstrates promising potential in the analysis of heart rate signals. In summary, the primary differences among these Shannon entropy-based algorithms are rooted in their approaches to estimating the probability density distribution of time series.
Entropy metrics are still continually evolving, with researchers exploring various strategies to enhance their capacity to analyze complex time series. Recognizing that real-world data often exhibit intricate structures across multiple temporal scales, Costa et al. proposed integrating entropy quantification methods with coarse-graining techniques, leading to the development of Multiscale Entropy (MSE) [21]. This innovative approach has been widely adopted by scholars and has become a vital tool for analyzing practical engineering data [22,23]. Additionally, some researchers have developed complexity-entropy plane algorithms [24,25,26,27,28,29,30,31,32,33], which map time series onto a two-dimensional plane as points or curves, thereby improving the discriminability of entropy features. However, it is important to note that the effectiveness of both Multiscale Entropy and complexity-entropy plane approaches is still influenced by the choice of entropy measure employed.
This highlights that traditional entropy algorithms possess unique advantages and limitations. The effectiveness of newer methods, such as DispEn, SlopEn, and AE, in various complex engineering applications has not yet been fully established. To address this gap, this paper presents a new entropy measure known as Generalized Gaussian Distribution Improved Permutation Entropy (GGDIPE). Firstly, the generalized Gaussian distribution cumulative distribution function (GGDCDF) with a specific shape parameter β is employed for data normalization. This technique ensures that our method is adaptable to a range of time series with different underlying distributions. Secondly, a new encoding approach for generating symbolic patterns is presented. This strategy combines the strengths of SampEn, PE, and its modifications, effectively preserving both temporal correlation and absolute amplitude information. Moreover, it resolves the issue of equal values that is encountered in PE. Finally, the probability density of our newly designed symbolic patterns is calculated, and the GGDIPE is derived based on Shannon entropy. To enhance the proposed algorithm’s capability in handling complex time series, this paper also introduces a multiscale version of GGDIPE that utilizes the coarse-graining technique.
The remainder of this paper is organized as follows: a detailed description of GGDIPE is provided in Section 2; synthetic and experimental data processing results are demonstrated and discussed in Section 3 and Section 4, respectively; the paper is concluded in Section 5.

2. GGDIPE

As illustrated in Figure 1, the proposed GGDIPE is developed through a series of steps: data normalization, improved permutation entropy (IPE) calculation, and the integration of entropy feature vectors. Each of these processes is described in detail in the following subsections. Additionally, a numerical example is provided in Appendix B to enhance clarity.

2.1. Data Normalization

The benefits of data normalization are well established in numerous engineering applications. DispEn’s strong data processing capabilities can be partially attributed to its use of the Gaussian cumulative distribution function for normalization. However, real-world data often deviates from Gaussian distributions. To address this issue, the GGDCDF with various shape parameters are employed for data normalization in this paper. This approach ensures that our method is adaptable to a wide range of time series with different underlying distributions.
As shown in Equation (1), given a shape parameter β and a time series x = x 1 , x 2 , x 3 , x N , the GGDCDF will transform x into a new sequence y , where the elements of y range from 0 to 1.
y = 1 2 + sgn ( x μ ) γ 1 / β , x μ ρ β 2 Γ ( 1 / β )
In Equation (1), sgn represents the Sign Function, as shown in Equation (2).
sgn ( x ) = 1   if   x < 0 0   if   x = 0 1   if   x > 0
μ denotes the average of x ; Γ stands for the famous gamma function; ρ = σ ( Γ ( 1 / β ) / Γ ( 3 / β ) ) , with σ the standard deviation of x ; and γ denotes the lower incomplete gamma function. Obviously, an element y i represents the probability that a random variable from the Generalized Gaussian Distribution (GGD) is less than x i . The GGD can be tailored to different forms by varying the shape parameter β . Specifically, when β = 2 , the GGD approximates a normal distribution, whereas when β = , it transforms into a uniform distribution [34]. A brief introduction to GGD can be found in Appendix B.

2.2. IPE

The IPE employs a new encoding strategy for generating symbolic patterns. Let y = y 1 , y 2 , y 3 , y N be the normalized time series under β i , the data is first segmented into several m dimensional vectors, with the j-th vector denoted as Y ( j , : ) = [ y j , y j + 1 , y j + ( m 1 ) ] , j = 1 , 2 , N m + 1 . This process is also known as phase space reconstruction, and m is called the embedding dimension.
To incorporate the absolute amplitude information of y , the first element of the j-th m dimensional vector Y ( j , 1 ) is symbolized using a Linear partition operator (LPO), as shown in Equation (3), yielding S ( j , 1 ) . Herein, y min and y max represent the minimum and maximum values of y , respectively. L corresponds to the number of symbols used to characterize the time series, and Δ = ( y max y min ) / L . In fact, the role of L in our algorithm is analogous to the tolerance r in SampEn. In SampEn, two vectors are considered similar if D X i m , X j m r . Similarly, in our algorithm, Y ( i , 1 ) and Y ( j , 1 ) are assigned the same symbol if they fall within the same region. For example, if y min Y ( i , 1 ) < Δ + y min and y min Y ( j , 1 ) < Δ + y min , the symbol 0 will be assigned to both Y ( i , 1 ) and Y ( j , 1 ) . As will be shown in Section 3, a smaller L increases information loss during symbolization, reducing the precision of entropy estimation but enhancing noise resistance. Conversely, a larger L minimizes information loss and leads to more accurate entropy estimation. In summary, selecting L requires a balance between the accuracy of entropy estimation and the robustness to noise.
LPO ( y ) = 0 y min y < Δ + y min 1 y min + Δ y < 2 Δ + y min L 1 y max Δ < y y max
To take into account the temporal correlation information, the k-th ( 2 k m ) element of the j-th vector, represented by Y ( j , k ) , is symbolized by Equation (4), yielding S ( j , k ) . The function ⌊ ⌋ returns the largest integer less than or equal to the input. Evidently, S ( j , k ) is derived from S ( j , 1 ) and the difference between Y ( j , k ) and Y ( j , 1 ) . Similar to OP, temporal correlation information is encoded into S ( j , k ) through comparing the two points.
S ( j , k ) = S ( j , 1 ) + ( Y ( j , k ) Y ( j , 1 ) ) / Δ
After symbolization, each m dimensional vector is mapped to a symbol pattern, denoted as S ( j , k ) , 1 k m , j = 1 , 2 , N m + 1 . There are L m underlying symbol patterns in the IPE algorithm. The probability of each symbol pattern p i ( 1 i L m ) is then counted, and the IPE is finally calculated using Equation (5), where ln ( L m ) denotes the largest IPE value.
I P E = i = 1 L m p i ln ( p i ) / ln ( L m )
In comparison to PE and its variants, IPE preserves both absolute amplitude and temporal correlation information and utilizes a larger set of potential symbolic patterns to characterize the signal. This suggests that IPE is likely to achieve better signal differentiation. Additionally, the symbolization process described in Equations (3) and (4) minimizes the need for extensive handling of equal values in the sequence, effectively addressing the equal value issue present in PE.

2.3. Entropy Feature Vector Integration

Section 2.2 outlines the process for calculating IPE with a specific shape parameter β i . As previously mentioned, the shape parameter determines the form of the GGD, enabling it to adapt to signals with varying distributions. In accordance with the recommendations in [34], in this paper, β is incremented from 0.1 to 2.9 in steps of 0.2. The IPE is computed for each value of β , and the resulting entropy values are combined into a 1 × 15 entropy feature vector, yielding the GGDIPE.

2.4. Multiscale GGDIPE

To enhance GGDIPE’s capability in handling complex time series, a multiscale version of GGDIPE is also proposed. Let x = x 1 , x 2 , x 3 , x N be the time series to be analyzed. The coarse-graining technique is employed to decompose the signal into multiple subsequences, as shown in Equation (6), where s denotes the scale factor, r i s represents the i-th element of the subsequence under scale s , and 1 j N / s . It is important to note that as the scale factor increases, the length of subsequence decreases to N / s , leading to potential instability in entropy estimation. Therefore, the scale factor should be carefully chosen based on the length of the original time series being analyzed.
r j s = 1 s i = ( j 1 ) s + 1 j s x i
After coarse-graining, GGDIPE is applied to each subsequence. Assuming the maximum scale factor is s max , the Multiscale Generalized Gaussian Distribution Improved Permutation Entropy (MGGDIPE) is ultimately represented as a s max × 15 entropy feature matrix.

3. Synthetic Data Analysis

In this section, the proposed GGDIPE algorithm is utilized to analyze various types of noise and chaotic signals to validate its effectiveness.

3.1. Noise

Noise is pervasive in practical engineering applications. The 1 / f α noise has been extensively studied in previous research and is an important benchmark for evaluating the performance of entropy algorithms [2,20]. For α = 0 , the power spectrum density is flat across all frequencies, corresponding to white noise. For α = 1 and α = 2 , the power of the signal decreases at varying rates as the frequency increases. Typically, 1 / f noise is referred to as pink noise, while 1 / f 2 noise is also known as brown noise.
In this subsection, we investigate the impact of various parameters—including embedding dimension m , the number of symbols L , and data length N —on GGDIPE using three types of noise.

3.1.1. Impact of Embedding Dimension

The embedding dimension determines the length of the reconstructed vectors. To explore its impact on GGDIPE, we set N = 10000 and L = 4 , then varied m from 3 to 6, observing how the entropy changes with different values of m . In this experiment, 20 independent sets were generated for each type of noise, and the entropy analysis results are presented using error bar plots. The center of each error bar represents the mean entropy value across multiple trials, while the length of the error bar indicates the standard deviation of these results. The analysis results are shown in Figure 2.
The GGDIPE results for the three types of noise are quite similar across different embedding dimensions, with only minor variations in entropy values. For example, when m = 3 , the entropy values for Brown noise remain stable around 0.3 across different β values, whereas for m = 6 , these values slightly decrease to approximately 0.15. Furthermore, the GGDIPE characteristics for the three noise types exhibit significant differences when β exceeds 0.5, with the differences becoming more pronounced as m increases. This suggests that GGDIPE is relatively insensitive to changes in the embedding dimension, demonstrating its general applicability. In subsequent study, m = 4 is selected.
As observed in Figure 2, a local maximum in entropy occurs when β is near 0.5. This is because when β 0.5 , the probability density of the generalized Gaussian distribution becomes sharply peaked, leading to the normalized values clustering around specific points. As a result, the distribution of certain patterns becomes more uniform, contributing to an increase in the entropy value.
For comparison, Figure 3 presents the analysis results of PE and DispEn for three types of noise across various embedding dimensions. It appears that the embedding dimension has minimal impact on DispEn, while it significantly affects PE. Specifically, for m = 3 , PE assigns very high entropy values (>0.94) to all three types of noise, obscuring their differences. However, as m increases, these differences become more obvious. Thus, a larger m enhances PE’s ability to distinguish between different signals.
The comparison results between Figure 2 and Figure 3 indicate that, compared to PE, GGDIPE is more stable and less affected by changes in the embedding dimension.

3.1.2. Impact of L

The parameter L defines the number of symbols used to represent the time series. To examine the effect of L on GGDIPE, we fixed m = 4 and N = 10000 and varied L across values of 2, 4, 6, and 8. We then observed how changes in L affect the entropy measurements. Again, for each type of noise, 20 independent trials were generated.
As illustrated in Figure 4, when L = 2 , GGDIPE’s entropy quantification for the three types of noise exhibits significant distortion. For instance, in the case of white noise, which is inherently random and theoretically should yield the maximum entropy value, it shows an entropy value below 0.7 for certain β values, as depicted in Figure 4a.
This issue is mitigated as L increases, with the entropy value for white noise approaching 1 as L becomes larger. This improvement is attributed to the fact that a higher L reduces information loss during the symbolization process, allowing GGDIPE to capture more potential patterns of the signal and thereby enhance the accuracy of entropy estimation. Nonetheless, L should not be excessively large. As will be discussed in Section 3.2, the selection of L requires a balance between entropy estimation accuracy and noise resistance.

3.1.3. Impact of Data Length

To investigate the impact of data length on GGDIPE, we set m = 4 and L = 4 and varied N from 100 to 2000 in increments of 10. For clarity, Figure 5 displays experimental results for selected values of β . Again, for each type of noise, 20 independent trials were generated.
The results shown in Figure 5 demonstrate that GGDIPE is capable of distinguishing between the three types of noise even with relatively short data lengths, such as N = 100 . As the data length increases, the entropy value for white noise converges towards 1, indicating enhanced accuracy in GGDIPE’s entropy estimation. Furthermore, the ability to differentiate among the three noise types improves with longer data lengths. The GGDIPE results stabilize for all three noise types when N exceeds 1000.
For comparison, Figure 6 illustrates the experimental results for PE and DispEn. With shorter data lengths, the permutation entropy values for white noise and pink noise are very close, and the larger error bars indicate lower consistency among multiple experimental trials. As the data length increases, this issue improves; however, the permutation entropy values for all three noise types remain above 0.9, suggesting they are still quite close. DispEn generally outperforms PE, but even at a data length of 2000, the DispEn value for brown noise has not yet stabilized, and the consistency among multiple experiments is somewhat lower compared to GGDIPE.
In summary, GGDIPE exhibits superior stability with shorter data lengths compared to PE and DispEn, offering enhanced signal differentiation capabilities and demonstrating reduced sensitivity to variations in data length.

3.2. Noisy Lorenz Signal

To assess the noise resistance of GGDIPE, Lorenz chaotic signals were generated using Equation (7), with only the z-axis component considered. White noise of varying intensities was added to simulate different signal-to-noise ratio (SNR) conditions. For each SNR condition, 20 independent experiments were conducted. The Lorenz system was solved using a fourth-order Runge–Kutta scheme with Δ t = 0.001 , and 10,000 data points were recorded.
x ˙ = 10 y x y ˙ = x 28 z y z ˙ = x y 8 3 z
Figure 7 illustrates the analysis results of GGDIPE applied to noisy Lorenz signals with varying values of L . When L = 2 , the results across different signal-to-noise ratio (SNR) conditions are largely consistent, with only the entropy curve for SNR = −10 dB showing a slight deviation from the clean signal’s curve. However, as L increases, these entropy curves start to diverge at higher SNR levels. For instance, when L = 8 , the entropy curve for 0 dB already shows a noticeable difference from that of the clean signal. As noted in Section 3.1.2, a larger L improves the accuracy of entropy estimation, which is why the GGDIPE curve for the clean Lorenz signal in Figure 7d is more precise. Comparing Figure 7a,d reveals that although the results for L = 2 are more consistent across different SNRs, the accuracy of the entropy estimation is significantly reduced. This reduction in accuracy is due to the greater information loss during the symbolization process with a smaller L , despite the increased noise resistance that comes with it. Taking into account the findings from both Figure 7 and Figure 4, it is clear that the selection of L requires a balance between entropy estimation accuracy and noise resistance. In the subsequent sections, unless otherwise specified, L = 4 will be used.
For comparison, Figure 8 presents the analysis results for PE and DispEn, displayed using box plots. The results clearly show that PE has minimal resistance to noise, with entropy values at 10 dB already significantly deviating from those of the pure Lorenz signal. In contrast, DispEn demonstrates stronger noise resistance, with noticeable deviations only occurring at 0 dB. These comparisons highlight that the proposed GGDIPE algorithm exhibits significantly better noise resistance than PE.

3.3. Logistic Map

The ability to detect periodicity and nonlinearity is a crucial metric for assessing the performance of entropy algorithms. To assess GGDIPE’s effectiveness in this area, we analyzed the Logistic model, represented by x n + 1 = μ x n ( 1 x n ) , where μ is a critical parameter influencing the model’s dynamic behavior. As μ varies from 3.5 to 3.99, the model exhibits period-doubling bifurcations, chaos, and intermittent periodic states, respectively. Figure 9 shows the results of applying various entropy algorithms to the Logistic model. In this analysis, μ is incremented by 0.0001 within the specified range. For each μ value, the Logistic model generates 10,000 data points, and their corresponding entropy values are calculated. For clarity, the figure displays GGDIPE results for only few specific β values.
The results show that while PE generally captures the overall dynamics of the Logistic model as μ varies, it falls short in fully describing the transition processes. In contrast, both GGDIPE and DispEn exhibit an increasing trend in the region marked by the arrow, around μ 3.57 , reflecting the system’s shift from a periodic state to chaos, consistent with the findings in [2]. Furthermore, in the elliptical region near μ 3.84 , both GGDIPE and DispEn show a marked decrease, revealing the brief periodic state in this area. These results in Figure 9 highlight GGDIPE’s strong capability in uncovering the dynamical behavior of the system.

3.4. Computation Cost

Computational complexity is a crucial metric for evaluating entropy algorithms, as it directly influences their suitability for real-time processing. To examine this, we analyzed 20 groups of white noise with varying data lengths using multiple entropy algorithms on the same computing device and recorded the computation time. All simulations in this section were performed using MATLAB R2024a on a computer with an i9-14900HX CPU (2.2 GHz) and 16 GB of RAM, produced by Xiaomi Communication Technology Co., LTD., in Beijing, China. For ease of comparison, the computation time for GGDIPE was recorded for a single β value. This strategy is justified because β can be selected according to specific application needs when utilizing GGDIPE.
The results in Table 1 indicate that DispEn consistently exhibits the fastest processing speed across various data lengths, with GGDIPE trailing slightly behind and PE showing the slowest processing speed. Given that PE is typically considered a real-time processing algorithm, the computational time of GGDIPE should not pose a significant barrier to its practical application.

4. Experimental Data Analysis

To assess the effectiveness of GGDIPE in managing complex real-world time series, the proposed algorithm was applied to a diverse set of intricate signals, including RR intervals data, EEG signals, bearing fault signals, and underwater acoustic signals.

4.1. RR Intervals

RR intervals are the time intervals between successive R-wave peaks in ECG signals. These intervals are crucial for analyzing heart rate variability. The experimental data used in the subsection is sourced from the Fantasia database [35], which includes data from 20 young and 20 elderly subjects. The primary objective is to determine whether the age of the subjects can be inferred from their RR interval data. Achieving this goal necessitates significant differences in the RR interval data characteristics between the young and elderly groups.
The GGDIPE analysis results for the RR intervals of healthy young and elderly subjects are presented in Figure 10. The results indicate that, for most values of β , there are significant differences in the features between the RR intervals of young and elderly subjects. This demonstrates that GGDIPE is capable of effectively distinguishing age-related variations in RR interval data. For comparison, Figure 11 presents the analysis results obtained using PE and DispEn. Visually, both methods demonstrate less effective separability between the two data categories compared to GGDIPE.
To quantify the separability of the entropy features, the non-parametric Mann–Whitney U-test was applied, with the results summarized in Table 2. Due to space limitations, Table 2 displays only a subset of the p-values for GGDIPE under some specific β values. In the table, * indicates a p-value less than 0.05, ** denotes a p-value less than 0.01, and *** represents a p-value less than 0.001. It is evident that, compared to PE and DispEn, the GGDIPE features reveal more pronounced differences between the two subject groups.

4.2. EEG Signals

EEG signals refer to the electrical activity recorded from the brain using electroencephalography, which are widely employed in prediction of epileptic seizures. The experimental data for this section are obtained from the Bonn EEG database [36], encompassing EEG signals from four distinct subject categories: healthy subjects with eyes open (Class 0), healthy subjects with eyes closed (Class 1), subjects during epileptic interictal periods (Class 2), and subjects experiencing seizure attacks (Class 3). The data was sampled at a frequency of 173.61 Hz, with 100 segments collected for each subject group, each segment having a duration of 23.6 s and comprising 4097 sample points. The challenge is to distinguish between Class 0 and Class 1, as well as between Class 2 and Class 3. To this end, multiscale entropy analysis was applied to the data. Given that each segment contains only 4097 sample points, the maximum scale factor was set to 5 to ensure the accuracy of the multiscale entropy estimates.
The analysis results for MGGDIPE, MPE, and MDE are shown in Figure 12 and Figure 13, respectively. For clarity, Figure 12 presents MGGDIPE results for only a subset of β values. Across all scales, Class 1 exhibits the highest average MGGDIPE values, followed by Class 0, Class 3, and Class 2. This distinct trend is absent in the MPE and MDE results. In particular, MDE shows minimal differentiation between Class 0 and Class 1 at lower scales, while MPE values for Class 2 and Class 3 nearly overlap at higher scales. Visually, MGGDIPE provides markedly better discrimination among the different signal groups compared to MPE and MDE.
To further substantiate this observation, the non-parametric Mann–Whitney U-test was applied to quantify the significance of differences between the feature sets, with the results detailed in Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8. The results are highly consistent with the feature extraction outcomes. Across all scales, the p-values for MGGDIPE are less than 0.001, indicating significant differences between the MGGDIPE features of class 0 versus class 1 and class 2 versus class 3. In contrast, MPE and MDE exhibit significant differences only at specific scales. These results suggest that MGGDIPE provides superior discriminative capability for differentiating between various EEG signal types compared to MPE and MDE.

4.3. Bearing Fault Signals

Bearings are essential components in many mechanical systems, and failures in bearings can lead to significant operational disruptions and severe consequences. Therefore, the early detection of bearing faults is crucial and presents a substantial challenge in the industrial sector. This section assesses the effectiveness of MGGDIPE for bearing fault diagnosis using data from the Jiangnan University Database [37,38]. The database includes four types of signals: inner-race defect, outer-race defect, rolling element defect, and normal signals. Data were collected at three different rotational speeds—600, 800, and 1000 rpm—with a sampling frequency of 50 kHz. Each signal type was recorded for 10 s at each speed. Due to the long duration of the signals, each dataset was segmented into non-overlapping chunks with 10,000 samples, and the maximum scale factor was set to 10.
The analysis results for MGGDIPE, MPE, and MDE are shown in Figure 14 and Figure 15, respectively. For clarity, Figure 14 presents MGGDIPE results for only a subset of β values. The MGGDIPE features demonstrate clear discriminative ability among the four signal types at low scales, with similar patterns observed in the MDE analysis. In contrast, MPE features for the four signal types exhibit substantial overlap at low scales, though their discriminative power improves with increasing scale factors. Nevertheless, MPE values for all signal types remain clustered between 0.84 and 0.98, reflecting weaker separability compared to MDE and MGGDIPE. These findings indicate that MGGDIPE has significant potential for effective bearing fault diagnosis.

4.4. Underwater Acoustic Signals

Underwater acoustic signals refer to sound waves transmitted through water. The ocean ambient noise and ship-radiated noise are the most frequently studied signals. This section assesses the effectiveness of MGGDIPE in classifying underwater acoustic signals using data from the ShipsEar database [39]. The dataset, sampled at 52,734 Hz, comprises four signal categories: Passenger, Ocean liner, Motorboat, and Ambient noise, with each category including recordings from multiple vessels. In contrast to some prior studies, this analysis requires feature extraction and identification from a larger number of targets, which presents a greater challenge. To facilitate analysis, the dataset was divided into numerous segments, each lasting 3 s. A comprehensive description of the signals is provided in Table 9.
The analysis results for MGGDIPE, MPE, and MDE are shown in Figure 16 and Figure 17, respectively. For clarity, Figure 16 presents MGGDIPE results for only a subset of β values. Given the extended length of the data, a larger maximum scale factor was chosen to capture a broader range of multiscale features, with a value of 40 applied in this section. The results in Figure 16 illustrate the advantages of using GGDCDF for data normalization. When β is large (e.g., 2.1 and 2.9), the entropy values of Motorboat and Ambient noise overlap at higher scales. However, with smaller β values (e.g., 0.9 and 1.5), the entropy values of these two signal types show clear separability at higher scales. This demonstrates that adjusting β enhances the feature extraction capabilities of GGDIPE. Additionally, the MGGDIPE feature trends, which describe how entropy varies with the scale factor, demonstrate distinct separability across the four types of underwater acoustic signals. As the scale factor increases from 1 to 40, the entropy of the Ocean liner steadily rises from 0.25 to 0.7, while the Passenger shows a rapid increase from 0.31 to 0.8, stabilizing after scale 20. The entropy of Ambient noise gradually decreases from approximately 0.7 to 0.6, leveling off after scale 15. In contrast, the entropy of the Motorboat signal increases sharply from 0.5 to 0.72, experiences a slight dip around scale 15, and then gradually ascends to 0.8.
In Figure 17a, it is evident that the separability of MPE features for the four types of underwater acoustic signals is significantly weaker than that of MGGDIPE. This is partly due to PE’s omission of signal amplitude information and its relatively poor noise resistance. In Figure 17b, the feature extraction results of MDE closely resemble those shown in Figure 16d, indicating consistency between the two methods. However, due to the absence of the β parameter, MDE exhibits weaker discrimination between Motorboat and Ambient noise compared to MGGDIPE.
To further quantitatively compare the effectiveness of the three multiscale entropy algorithms in underwater target classification, a probabilistic neural network (PNN) was used to train and test the extracted features, with recognition rate serving as the performance metric. The training and testing samples were non-overlapping, with 100 randomly selected segments from each signal type used for testing, while the remaining segments were used for training. The recognition results are presented in Table 10, Table 11 and Table 12. To highlight the impact of the parameter β on data classification, the classification results of MGGDIPE at specific β values are also provided in Appendix A.
The recognition results align closely with the previous analysis, with MGGDIPE achieving an accuracy rate of 97.5%, substantially exceeding those of MDE (70.5%) and MPE (62.5%). These findings underscore the significant potential of the proposed MGGDIPE algorithm for underwater target recognition.

5. Conclusions

To enhance the performance of entropy algorithms in analyzing complex time series, this paper introduces the GGDIPE algorithm and its multiscale variant. GGDIPE employs the GGDCDF for data normalization, making it versatile across various distributions. The algorithm further processes the normalized data with the IPE approach, which maintains both the absolute magnitude and temporal correlations of the signals, overcoming the equal value issue found in traditional PE. Simulation results indicate that GGDIPE is less sensitive to parameter variations, exhibits strong noise resistance, accurately reveals the dynamic behavior of chaotic systems, and operates significantly faster than PE, with speed comparable to DispEn. Real-world data analysis shows that MGGDIPE provides markedly better separability for RR interval signals, EEG signals, bearing fault signals, and underwater acoustic signals compared to conventional MPE and the recently proposed MDE algorithm. Notably, in underwater target recognition tasks, MGGDIPE achieved a classification accuracy of 97.5% across four types of acoustic signals, substantially surpassing the performance of MDE (70.5%) and MPE (62.5%). Thus, the proposed method demonstrates exceptional capability in processing complex time series data.

Author Contributions

Conceptualization, Z.C. and J.K.C.; methodology, K.Z.; investigation, K.Z. and H.-S.G.; writing—original draft preparation, K.Z.; writing—review and editing, H.-S.G., J.K.C., and S.-H.T.; funding acquisition, Z.C. and S.-H.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Guangxi Science and Technology Base and Talent Project (No. GuikeAD21220098) and the 2021 Open Fund project of the Key Laboratory of Cognitive Radio and Information Processing of the Ministry of Education (No. CRKL210102).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

To better understand the algorithm proposed in this paper, this appendix provides an explanation through a simple example. Supposing that the signal to be analyzed is x = [ 1.5 , 1.06 , 0.57 , 2.21 , 0.43 , 0.7 , 0.74 , 0.32 , 1.5 , 1.2 ] and β = 0.9 , Equation (1) transforms x into y = [ 0.8730 ,   0.2979 ,   0.6558 ,   0.0726 ,   0.6186 ,   0.3937 ,   0.3834 ,   0.5907 ,   0.8730 ,   0.2611 ] . Setting m = 4 and L = 4 , phase space reconstruction results can be expressed as Equation (A1), and symbolization results can be obtained through Equations (3) and (4), as shown in Equation (A2). Obviously, p 4231 = p 2313 = p 3132 = p 1322 = p 3223 = p 2224 = p 2342 = 1 7 , hence I P E = i = 1 7 1 7 ln ( 1 7 ) ln ( 4 4 ) = 0.3510.
1.5 1.06 0.57 2.21 1.06 0.57 2.21 0.43 0.57 2.21 0.43 0.7 2.21 0.43 0.7 0.74 0.43 0.7 0.74 0.32 0.7 0.74 0.32 1.5 0.74 0.32 1.5 1.2
4 2 3 1 2 3 1 3 3 1 3 2 1 3 2 2 3 2 2 3 2 2 2 4 2 3 4 2

Appendix B

The probability density function of the generalized Gaussian distribution is given by Equation (A3), where μ and σ denote the mean and the standard deviation of the distribution, Γ stands for the gamma function, and ρ = σ ( Γ ( 1 / β ) / Γ ( 3 / β ) ) . Figure A1 shows the probability density function of the generalized Gaussian distribution and the Gaussian distribution.
f ( x ; μ , ρ , β ) = β 2 ρ Γ ( 1 / β ) exp x μ ρ β
Figure A1. Probability density function of generalized Gaussian distribution.
Figure A1. Probability density function of generalized Gaussian distribution.
Entropy 26 00960 g0a1

Appendix C

To demonstrate the impact of the parameter β on data classification, the classification results of MGGDIPE at specific β values are provided in Table A1 and Table A2.
The results indicate that utilizing a single β value for data normalization significantly decreases the recognition rate of Motorboat. Specifically, when β = 2.1 , the normalization closely approximates the cumulative distribution function of the normal distribution, yielding a recognition rate of 74% for Motorboat. In contrast, using β = 1.5 improves the recognition rate to 82%. This finding highlights the considerable impact of different normalization methods on recognition accuracy. In the absence of prior knowledge about the true distribution of the experimental data, employing multiple functions for data normalization is a prudent approach that aligns with the principles underlying the proposed GGDIPE method. Furthermore, the results presented in Table 10 demonstrate that combining multiscale entropy across various β values into a feature vector enhances data separability.
Table A1. PNN recognition results for MGGDIPE with β = 1.5 .
Table A1. PNN recognition results for MGGDIPE with β = 1.5 .
CategoriesRecognized AsClassification Accuracy
PassengerOcean LinerMotorboatAmbient Noise
Passenger100000100%
Ocean liner010000100%
Motorboat13582082%
Ambient noise0049696%
In total\\\\94.5%
Table A2. PNN recognition results for MGGDIPE with β = 2.1 .
Table A2. PNN recognition results for MGGDIPE with β = 2.1 .
CategoriesRecognized AsClassification Accuracy
PassengerOcean LinerMotorboatAmbient Noise
Passenger100000100%
Ocean liner010000100%
Motorboat20474274%
Ambient noise0029898%
In total\\\\93%

References

  1. Deka, B.; Deka, D. An improved multiscale distribution entropy for analyzing complexity of real-world signals. Chaos Solitons Fractals 2022, 158, 112101. [Google Scholar] [CrossRef]
  2. Hamed, A.; Saeid, S.; Tarek, K.R. Ensemble entropy: A low bias approach for data analysis. Knowl. -Based Syst. 2022, 256, 109876. [Google Scholar]
  3. Sharma, K.; Sunkaria, R.K. Cardiac Arrhythmia Detection Using Cross-Sample Entropy Measure Based on Short and Long RR Interval Series. J. Arrhythm. 2023, 39, 412–421. [Google Scholar] [CrossRef] [PubMed]
  4. Deng, B.; Cai, L.; Li, S.; Wang, R.; Yu, H.; Chen, Y.; Wang, J. Multivariate multi-scale weighted permutation entropy analysis of EEG complexity for Alzheimer’s disease. Cogn. Neurodyn. 2017, 11, 217–231. [Google Scholar] [CrossRef]
  5. Li, P.; Karmakar, C.; Yan, C.; Palaniswami, M.; Liu, C. Classification of 5-S Epileptic EEG Recordings Using Distribution Entropy and Sample Entropy. Front. Physiol. 2016, 7, 136. [Google Scholar] [CrossRef]
  6. Ragavesh, D.; Scott, M.; Gordon, M. A Novel Bearing Faults Detection Method Using Generalized Gaussian Distribution Refined Composite Multiscale Dispersion Entropy. IEEE Trans. Instrum. Meas. 2022, 71, 3517112. [Google Scholar]
  7. Lei, Z.; Lei, X.; Zhou, C.; Qing, L.; Zhang, Q. Compressed Sensing Multiscale Sample Entropy Feature Extraction Method for Underwater Target Radiation Noise. IEEE Access 2022, 10, 77688–77694. [Google Scholar] [CrossRef]
  8. Li, Z.; Li, Y.; Zhang, K.; Guo, J. Hierarchical Refined Composite Multiscale Fluctuation-Based Dispersion Entropy: Application to Feature Extraction of Underwater Target Signal. Nonlinear Dyn. 2023, 111, 22399–22417. [Google Scholar] [CrossRef]
  9. Pincus, S.M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. USA 1991, 88, 2297–2301. [Google Scholar] [CrossRef]
  10. Richman, J.S.; Moorman, J.R. Physiological time-series analysis using approximate entropy and sample entropy. Am. J. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef]
  11. Chen, W.; Wang, Z.; Xie, H.; Yu, W. Characterization of Surface EMG Signal Based on Fuzzy Entropy. IEEE Trans. Neural Syst. Rehabilitation Eng. 2007, 15, 266–272. [Google Scholar] [CrossRef] [PubMed]
  12. Bandt, C.; Pompe, B. Permutation entropy: A natural complexity measure for time series. Phys. Rev. Lett. 2002, 88, 174102. [Google Scholar] [CrossRef] [PubMed]
  13. Li, P.; Liu, C.; Li, K.; Zheng, D.; Liu, C.; Hou, Y. Assessing the complexity of short-term heartbeat interval series by distribution entropy. Med. Biol. Eng. Comput. 2015, 53, 77–87. [Google Scholar] [CrossRef] [PubMed]
  14. Fadlallah, B.; Chen, B.D.; Keil, A.; Principe, J. Weighted-permutation entropy: A complexity measure for time series incorporating amplitude information. Phys. Rev. E 2013, 88, 022911. [Google Scholar] [CrossRef]
  15. Bian, C.; Qin, C.; Ma, Q.D.; Shen, Q. Modified permutation-entropy analysis of heartbeat dynamics. Phys. Rev. E 2012, 85, 021906. [Google Scholar] [CrossRef]
  16. Rostaghi, M.; Azami, H. Dispersion entropy: A measure for time-series analysis. IEEE Signal Process. Lett. 2016, 23, 610–614. [Google Scholar] [CrossRef]
  17. Cuesta-Frau, D. Slope Entropy: A New Time Series Complexity Estimator Based on Both Symbolic Patterns and Amplitude Information. Entropy 2019, 21, 1167. [Google Scholar] [CrossRef]
  18. Yang, J.; Choudhary, G.I.; Rahardja, S.; Franti, P. Classification of Interbeat Interval Time-Series Using Attention Entropy. IEEE Trans. Affect. Comput. 2023, 14, 321–330. [Google Scholar] [CrossRef]
  19. Azami, H.; Escudero, J. Amplitude- and Fluctuation-Based Dispersion Entropy. Entropy 2018, 20, 210. [Google Scholar] [CrossRef]
  20. Chen, Z.; Li, Y.A.; Liang, H.T.; Yu, J. Improved permutation entropy for measuring complexity of time series under noisy condition. Complexity 2019, 2019, 1403829. [Google Scholar] [CrossRef]
  21. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of biological signals. Phys. Rev. E 2005, 71, 021906. [Google Scholar] [CrossRef] [PubMed]
  22. Jiang, Y.; Peng, C.K.; Xu, Y.S. Hierarchical entropy analysis for biological signals. J. Comput. Appl. Math. 2011, 236, 728–742. [Google Scholar] [CrossRef]
  23. Azami, H.; Rostaghi, M.; Abásolo, D.; Escudero, J. Refined Composite Multiscale Dispersion Entropy and its Application to Biomedical Signals. IEEE Trans. Biomed. Eng. 2017, 64, 2872–2879. [Google Scholar] [PubMed]
  24. López-Ruiz, R.; Mancini, H.L.; Calbet, X. A statistical measure of complexity. Phys. Lett. A. 1995, 209, 321–326. [Google Scholar] [CrossRef]
  25. Ribeiro, H.; Jauregui, M.; Zunino, L.; Lenzi, E.K. Characterizing time series via complexity-entropy curves. Phys. Rev. E. 2017, 95, 062106. [Google Scholar] [CrossRef]
  26. Jauregui, M.; Zunino, L.; Lenzi, E.K.; Mendes, R.; Ribeiro, H. Characterization of time series via Rényi complexity–entropy curves. Phys. A 2018, 498, 74–85. [Google Scholar] [CrossRef]
  27. Peng, K.; Shang, P. Characterizing ordinal network of time series based on complexity-entropy curve. Pattern Recognit. 2022, 124, 108464. [Google Scholar] [CrossRef]
  28. Lee, M. Early warning detection of thermoacoustic instability using three-dimensional complexity-entropy causality space. Exp. Therm. Fluid Sci. 2022, 130, 110517. [Google Scholar] [CrossRef]
  29. Zhang, Y.; Shang, P.; He, J.; Xiong, H. Cumulative Tsallis entropy based on power spectrum of financial time series. Chaos An. Interdiscip. J. Nonlinear Sci. 2019, 29, 103118. [Google Scholar] [CrossRef]
  30. Rosso, O.A.; Larrondo, H.A.; Martin, M.T.; Plastino, A.; Fuentes, M.A. Distinguishing Noise from Chaos. Phys.Rev.Lett. 2007, 99, 154102. [Google Scholar] [CrossRef]
  31. Olivares, F.; Plastino, A.; Rosso, O.A. Contrasting Chaos with Noise via Local versus Global Information Quantifiers. Phys.Lett. A 2012, 376, 1577–1583. [Google Scholar] [CrossRef]
  32. Zunino, L.; Soriano, M.C.; Rosso, O.A. Distinguishing Chaotic and Stochastic Dynamics from Time Series by Using a Multiscale Symbolic Approach. Phys. Rev. E 2012, 86, 046210. [Google Scholar] [CrossRef] [PubMed]
  33. Mateos, D.M.; Zozor, S.; Olivares, F. Contrasting Stochasticity with Chaos in a Permutation Lempel–Ziv Complexity-Shannon Entropy Plane. Phys. A 2020, 554, 124640. [Google Scholar] [CrossRef]
  34. Yu, S.; Zhang, A.; Li, H. A review of estimating the shape parameter of generalized Gaussian distribution. J. Comput. Inf. Syst. 2012, 21, 9055–9064. [Google Scholar]
  35. Iyengar, N.; Peng, C.K.; Morin, R.; Goldberger, A.L.; Lipsitz, L.A. Age-related alterations in the fractal scaling of cardiac interbeat interval dynamics. Am. J. Physiol. 1996, 271, 1078–1084. [Google Scholar] [CrossRef]
  36. Andrzejak, R.G.; Lehnertz, K.; Mormann, F.; Rieke, C.; David, P.; Elger, C.E. Indications of nonlinear deterministic and finite-dimensional structures in time series of brain electrical activity: Dependence on recording region and brain state. Phys. Rev. E 2001, 64, 061907. [Google Scholar] [CrossRef]
  37. Li, K.; Ping, X.; Wang, H.; Chen, P.; Cao, Y. Sequential Fuzzy Diagnosis Method for Motor Roller Bearing in Variable Operating Conditions Based on Vibration Analysis. Sensors 2013, 13, 8013–8041. [Google Scholar] [CrossRef]
  38. Zhao, Z.; Li, T.; Wu, J.; Sun, C.; Wang, S.; Yan, R.; Chen, X. Deep Learning Algorithms for Rotating Machinery Intelligent Diagnosis: An Open Source Benchmark Study. ISA Trans. 2020, 107, 224–255. [Google Scholar] [CrossRef]
  39. Santos-Domínguez, D.; Torres-Guijarro, S.; Cardenal-López, A. ShipsEar: An underwater vessel noise database. Appl. Acoust. 2016, 113, 64–69. [Google Scholar] [CrossRef]
Figure 1. Calculation flow chart of generalized Gaussian distribution improved permutation entropy (GGDIPE).
Figure 1. Calculation flow chart of generalized Gaussian distribution improved permutation entropy (GGDIPE).
Entropy 26 00960 g001
Figure 2. The GGDIPE analysis for three types of noise across varying embedding dimensions. (a) m = 3 (b) m = 4 (c) m = 5 (d) m = 6 .
Figure 2. The GGDIPE analysis for three types of noise across varying embedding dimensions. (a) m = 3 (b) m = 4 (c) m = 5 (d) m = 6 .
Entropy 26 00960 g002
Figure 3. Analysis results of PE and DispEn for three types of noise across various embedding dimensions. (a) PE analysis result; (b) DispEn analysis result.
Figure 3. Analysis results of PE and DispEn for three types of noise across various embedding dimensions. (a) PE analysis result; (b) DispEn analysis result.
Entropy 26 00960 g003
Figure 4. The GGDIPE analysis for three types of noise across varying L (a) L = 2; (b) L = 4; (c) L = 6; (d) L = 8.
Figure 4. The GGDIPE analysis for three types of noise across varying L (a) L = 2; (b) L = 4; (c) L = 6; (d) L = 8.
Entropy 26 00960 g004
Figure 5. The GGDIPE analysis for three types of noise across varying data length. (a) β   =   0.9 (b) β   =   1.5 (c) β   =   2.1 (d) β   =   2.9 .
Figure 5. The GGDIPE analysis for three types of noise across varying data length. (a) β   =   0.9 (b) β   =   1.5 (c) β   =   2.1 (d) β   =   2.9 .
Entropy 26 00960 g005
Figure 6. Analysis results of PE and DispEn for three types of noise across various data length. (a) PE analysis result; (b) DispEn analysis result.
Figure 6. Analysis results of PE and DispEn for three types of noise across various data length. (a) PE analysis result; (b) DispEn analysis result.
Entropy 26 00960 g006
Figure 7. The GGDIPE analysis for noisy Lorenz signals. (a) L = 2 (b) L = 4 (c) L = 6 (d) L = 8 .
Figure 7. The GGDIPE analysis for noisy Lorenz signals. (a) L = 2 (b) L = 4 (c) L = 6 (d) L = 8 .
Entropy 26 00960 g007
Figure 8. Analysis results of PE and DispEn for noisy Lorenz signals. (a) PE analysis result; (b) DispEn analysis result.
Figure 8. Analysis results of PE and DispEn for noisy Lorenz signals. (a) PE analysis result; (b) DispEn analysis result.
Entropy 26 00960 g008
Figure 9. Analysis results of applying various entropy algorithms to the Logistic model.
Figure 9. Analysis results of applying various entropy algorithms to the Logistic model.
Entropy 26 00960 g009
Figure 10. GGDIPE analysis results for the RR intervals of healthy young and healthy elderly subjects.
Figure 10. GGDIPE analysis results for the RR intervals of healthy young and healthy elderly subjects.
Entropy 26 00960 g010
Figure 11. Analysis results of PE and DispEn for the RR intervals of healthy young and healthy elderly subjects. (a) PE analysis result; (b) DispEn analysis result.
Figure 11. Analysis results of PE and DispEn for the RR intervals of healthy young and healthy elderly subjects. (a) PE analysis result; (b) DispEn analysis result.
Entropy 26 00960 g011
Figure 12. The MGGDIPE analysis for EEG signals. (a) β   =   0.9 (b) β   =   1.5 (c) β   =   2.1 (d) β   =   2.9 .
Figure 12. The MGGDIPE analysis for EEG signals. (a) β   =   0.9 (b) β   =   1.5 (c) β   =   2.1 (d) β   =   2.9 .
Entropy 26 00960 g012
Figure 13. The MPE and MDE analysis for EEG signals. (a) MPE analysis result; (b) MDE analysis result.
Figure 13. The MPE and MDE analysis for EEG signals. (a) MPE analysis result; (b) MDE analysis result.
Entropy 26 00960 g013
Figure 14. The MGGDIPE analysis for bearing fault signals. (a) β   =   0.9 (b) β   =   1.5 (c) β   =   2.1 (d) β   =   2.9 .
Figure 14. The MGGDIPE analysis for bearing fault signals. (a) β   =   0.9 (b) β   =   1.5 (c) β   =   2.1 (d) β   =   2.9 .
Entropy 26 00960 g014
Figure 15. The MPE and MDE analysis for bearing fault signals. (a) MPE analysis result; (b) MDE analysis result.
Figure 15. The MPE and MDE analysis for bearing fault signals. (a) MPE analysis result; (b) MDE analysis result.
Entropy 26 00960 g015
Figure 16. The MGGDIPE analysis for underwater acoustic signals. (a) β   =   0.9 (b) β   =   1.5 (c) β   =   2.1 (d) β   =   2.9 .
Figure 16. The MGGDIPE analysis for underwater acoustic signals. (a) β   =   0.9 (b) β   =   1.5 (c) β   =   2.1 (d) β   =   2.9 .
Entropy 26 00960 g016
Figure 17. The MPE and MDE analysis for underwater acoustic signals. (a) MPE analysis result; (b) MDE analysis result.
Figure 17. The MPE and MDE analysis for underwater acoustic signals. (a) MPE analysis result; (b) MDE analysis result.
Entropy 26 00960 g017
Table 1. The computation time required by different entropy algorithms to complete 20 sets of noise analysis experiments.
Table 1. The computation time required by different entropy algorithms to complete 20 sets of noise analysis experiments.
N = 10,000N = 50,000N = 100,000
PE0.43 s2.04 s4.01 s
DispEn0.13 s0.30 s0.51 s
GGDIPE0.13 s0.39 s0.69 s
Table 2. The non-parametric Mann–Whitney U-test results for the RR intervals of healthy young and healthy elderly subjects. * indicates a p-value less than 0.05, ** denotes a p-value less than 0.01, and *** represents a p-value less than 0.001.
Table 2. The non-parametric Mann–Whitney U-test results for the RR intervals of healthy young and healthy elderly subjects. * indicates a p-value less than 0.05, ** denotes a p-value less than 0.01, and *** represents a p-value less than 0.001.
β = 0.9β = 1.5β = 2.1β = 2.9PEDispEn
p-values 2.34 × 10 4 *** 2.61 × 10 4 *** 5.35 × 10 4 ***0.0019 **0.27920.0038 *
Table 3. The significance of differences in MGGDIPE ( β = 0.9 ) features between distinct categories. *** denotes p-values less than 0.001.
Table 3. The significance of differences in MGGDIPE ( β = 0.9 ) features between distinct categories. *** denotes p-values less than 0.001.
Scale 1Scale 2Scale 3Scale 4Scale 5
Class0 vs. Class1 p = 4.575 × 10 8 *** p = 7.869 × 10 11 *** p = 3.026 × 10 17 *** p = 4.815 × 10 21 *** p = 3.922 × 10 19 ***
Class2 vs. Class3 p = 1.839 × 10 22 *** p = 7.352 × 10 18 *** p = 2.370 × 10 15 *** p = 3.923 × 10 14 *** p = 8.281 × 10 14 ***
Table 4. The significance of differences in MGGDIPE ( β = 1.5 ) features between distinct categories. *** denotes p-values less than 0.001.
Table 4. The significance of differences in MGGDIPE ( β = 1.5 ) features between distinct categories. *** denotes p-values less than 0.001.
Scale 1Scale 2Scale 3Scale 4Scale 5
Class0 vs. Class1 p = 5.857 × 10 8 *** p = 6.063 × 10 11 *** p = 8.717 × 10 18 *** p = 2.687 × 10 21 *** p = 9.414 × 10 20 ***
Class2 vs. Class3 p = 1.250 × 10 22 *** p = 6.322 × 10 18 *** p = 2.667 × 10 15 *** p = 3.639 × 10 14 *** p = 1.551 × 10 13 ***
Table 5. The significance of differences in MGGDIPE ( β = 2.1 ) features between distinct categories. *** denotes p-values less than 0.001.
Table 5. The significance of differences in MGGDIPE ( β = 2.1 ) features between distinct categories. *** denotes p-values less than 0.001.
Scale 1Scale 2Scale 3Scale 4Scale 5
Class0 vs. Class1 p = 3.076 × 10 6 *** p = 1.552 × 10 9 *** p = 5.904 × 10 16 *** p = 3.395 × 10 21 *** p = 1.817 × 10 18 ***
Class2 vs. Class3 p = 2.688 × 10 16 *** p = 1.322 × 10 12 *** p = 5.496 × 10 11 *** p = 6.388 × 10 10 *** p = 3.699 × 10 9 ***
Table 6. The significance of differences in MGGDIPE ( β = 2.9 ) features between distinct categories. *** denotes p-values less than 0.001.
Table 6. The significance of differences in MGGDIPE ( β = 2.9 ) features between distinct categories. *** denotes p-values less than 0.001.
Scale 1Scale 2Scale 3Scale 4Scale 5
Class0 vs. Class1 p = 4.966 × 10 5 *** p = 5.857 × 10 8 *** p = 3.047 × 10 13 *** p = 3.651 × 10 17 *** p = 6.527 × 10 16 ***
Class2 vs. Class3 p = 2.402 × 10 9 *** p = 2.125 × 10 8 *** p = 2.545 × 10 7 *** p = 8.088 × 10 6 *** p = 1.592 × 10 4 ***
Table 7. The significance of differences in MPE features between distinct categories. *, and *** denotes p-values less than 0.05 and 0.001, respectively.
Table 7. The significance of differences in MPE features between distinct categories. *, and *** denotes p-values less than 0.05 and 0.001, respectively.
Scale 1Scale 2Scale 3Scale 4Scale 5
Class0 vs. Class1 p = 7.006 × 10 14 *** p = 1.841 × 10 14 *** p = 2.313 × 10 14 *** p = 5.988 × 10 5 *** p = 0.0118 *
Class2 vs. Class3 p = 4.7184 × 10 29 *** p = 9.922 × 10 6 *** p = 0.9932 p = 0.0925 p = 0.1484
Table 8. The significance of differences in MDE features between distinct categories. **, *** denotes p-values less than 0.01, and 0.001, respectively.
Table 8. The significance of differences in MDE features between distinct categories. **, *** denotes p-values less than 0.01, and 0.001, respectively.
Scale 1Scale 2Scale 3Scale 4Scale 5
Class0 vs. Class1 p = 0.1582 p = 0.0739 p = 0.0043 ** p = 6.173 × 10 7 *** p = 4.288 × 10 9 ***
Class2 vs. Class3 p = 0.0025 ** p = 6.438 × 10 5 *** p = 8.265 × 10 4 *** p = 0.0081 ** p = 0.1314
Table 9. Detailed information of underwater acoustic signals.
Table 9. Detailed information of underwater acoustic signals.
CategoriesShip NameNumber of Segments
PassengerMar de Cangas267
Mar de Onza124
Pirata de Salvora65
Arrois103
Ocean linerMSC Opera160
Adventure of the sea89
Costa Voyager397
MotorboatSmall Yacht76
Motorboat286
High speed motorboat36
Zodiac96
Ambient noise\267
Table 10. PNN recognition results for MGGDIPE.
Table 10. PNN recognition results for MGGDIPE.
CategoriesRecognized AsClassification Accuracy
PassengerOcean LinerMotorboatAmbient Noise
Passenger100000100%
Ocean liner010000100%
Motorboat1690390%
Ambient noise000100100%
In total\\\\97.5%
Table 11. PNN recognition results for MPE.
Table 11. PNN recognition results for MPE.
CategoriesRecognized AsClassification Accuracy
PassengerOcean LinerMotorboatAmbient Noise
Passenger48005248%
Ocean liner19810081%
Motorboat04217521%
Ambient noise000100100%
In total\\\\62.5%
Table 12. PNN recognition results for MDE.
Table 12. PNN recognition results for MDE.
CategoriesRecognized AsClassification Accuracy
PassengerOcean LinerMotorboatAmbient Noise
Passenger100000100%
Ocean liner19810081%
Motorboat64216921%
Ambient noise00208080%
In total\\\\70.5%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zheng, K.; Gan, H.-S.; Chaw, J.K.; Teh, S.-H.; Chen, Z. Generalized Gaussian Distribution Improved Permutation Entropy: A New Measure for Complex Time Series Analysis. Entropy 2024, 26, 960. https://doi.org/10.3390/e26110960

AMA Style

Zheng K, Gan H-S, Chaw JK, Teh S-H, Chen Z. Generalized Gaussian Distribution Improved Permutation Entropy: A New Measure for Complex Time Series Analysis. Entropy. 2024; 26(11):960. https://doi.org/10.3390/e26110960

Chicago/Turabian Style

Zheng, Kun, Hong-Seng Gan, Jun Kit Chaw, Sze-Hong Teh, and Zhe Chen. 2024. "Generalized Gaussian Distribution Improved Permutation Entropy: A New Measure for Complex Time Series Analysis" Entropy 26, no. 11: 960. https://doi.org/10.3390/e26110960

APA Style

Zheng, K., Gan, H.-S., Chaw, J. K., Teh, S.-H., & Chen, Z. (2024). Generalized Gaussian Distribution Improved Permutation Entropy: A New Measure for Complex Time Series Analysis. Entropy, 26(11), 960. https://doi.org/10.3390/e26110960

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop