Next Article in Journal
A-DVM: A Self-Adaptive Variable Matrix Decision Variable Selection Scheme for Multimodal Problems
Next Article in Special Issue
Effects of Tau and Sampling Frequency on the Regularity Analysis of ECG and EEG Signals Using ApEn and SampEn Entropy Estimators
Previous Article in Journal
Effect of Finite-Size Heat Source’s Heat Capacity on the Efficiency of Heat Engine
Previous Article in Special Issue
New Fast ApEn and SampEn Entropy Algorithms Implementation and Their Application to Supercomputer Power Consumption
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Composite Multiscale Partial Cross-Sample Entropy Analysis for Quantifying Intrinsic Similarity of Two Time Series Affected by Common External Factors

1
Key Laboratory of Intelligent Computing and Information Processing of Ministry of Education and Hunan Key Laboratory for Computation and Simulation in Science and Engineering, Xiangtan University, Xiangtan 411105, China
2
School of Electrical Engineering and Computer Science, Queensland University of Technology, GPO Box 2434, Brisbane QLD 4000, Australia
*
Author to whom correspondence should be addressed.
Entropy 2020, 22(9), 1003; https://doi.org/10.3390/e22091003
Submission received: 6 August 2020 / Revised: 4 September 2020 / Accepted: 6 September 2020 / Published: 8 September 2020
(This article belongs to the Special Issue Information Theoretic Measures and Their Applications)

Abstract

:
In this paper, we propose a new cross-sample entropy, namely the composite multiscale partial cross-sample entropy (CMPCSE), for quantifying the intrinsic similarity of two time series affected by common external factors. First, in order to test the validity of CMPCSE, we apply it to three sets of artificial data. Experimental results show that CMPCSE can accurately measure the intrinsic cross-sample entropy of two simultaneously recorded time series by removing the effects from the third time series. Then CMPCSE is employed to investigate the partial cross-sample entropy of Shanghai securities composite index (SSEC) and Shenzhen Stock Exchange Component Index (SZSE) by eliminating the effect of Hang Seng Index (HSI). Compared with the composite multiscale cross-sample entropy, the results obtained by CMPCSE show that SSEC and SZSE have stronger similarity. We believe that CMPCSE is an effective tool to study intrinsic similarity of two time series.

1. Introduction

Complex systems with interacting constituents exist in all aspects of nature and society, such as geophysics [1], solid state physics, climate system, ecosystem, financial system [2,3], and so forth. These complex systems are constantly generating a large number of time signals. Fortunately, in recent decades, numerous creative methods have been proposed to explore the operation mechanism of these complex systems. Among them, entropy-based methods are very powerful modern analysis technology. The concept of ’entropy’ was first proposed by Clausius to deal with thermodynamic problems, and then Boltzmann gave a microscopic explanation from the perspective of statistical mechanics and proposed Boltzmann entropy. Gibbs proposed Gibbs entropy when determining uncertain system. In 1948, Shannon introduced the concept of entropy into information theory and put forward Shannon entropy (information entropy) [4]. Shortly after that, Renyi extended it and proposed Renyi entropy [5]. In 1988, Tsallis gave a Generalization of Boltzmann-Gibbs Statistics and proposed Tsallis entropy [6]. Although Gibbs entropy and Shannon entropy have the same mathematical expression, Shannon entropy has a broader meaning than thermodynamic entropy, as all the basic laws of thermodynamics can be derived from information entropy [7]. Since information entropy and Shannon entropy were proposed, many entropy-based methods have been proposed to explore the system complexity through studying the time series generated from them [8,9]. In order to quantify the changing complexity of real finite time series, Picnus proposed the approximate entropy (ApEn) [10,11,12], which had been used to study biological time series [13,14]. In 2002, Richman et al. analyzed the deficiencies of ApEn and proposed the concept of sample entropy (SampEn). Compared with ApEn, SampEn agreed with theoretical results much closer than ApEn over a broad range of conditions, and has been successfully applied to clinical cardiovascular study [15,16]. Cross-sample entropy (Cross-SampEn) was also proposed for comparing two different time series to assess their degree of similarity [15]. And in 2010, when Liu et al. studied the correlation of foreign exchange time series, they found that cross-SampEn is superior to correlation coefficient in describing the correlation between the foreign exchange time series [17]. In 2003, Costa et al. found that an increase in the entropy of a system is usually but not always associated with an increase of complexity, so the traditional entropy-based algorithms may lead to misleading results [18]. And in order to avoid this situation, they introduced the multiscale sample entropy (MSE), which had been successfully used to study various dynamical systems [19,20,21,22,23]. Not long after that, MSE was extended to multiscale cross-sample entropy (MCSE) to measure the cross-sample entropy over different time scales. Unfortunately, in the process of multi-scale analysis, the coarse-grained procedure sets a higher requirement for the length of the time series, that is, when the length of the sequence is not long enough, it will get inaccurate results. In addition, in some cases, the insufficiency of sequence length will lead to no template vector matched to another, and hence the cross-sample entropy can not be defined. In order to overcome this shortcoming, Wu et al. proposed the composite multiscale sample entropy (CMSE) [24] and refined composite multiscale entropy (RCMSE) [25] successively. Inspired by CMSE and RCMSE, Yin et al. introduced composite multiscale cross-sample entropy (CMCSE) and Refined composite multiscale cross-sample entropy (RCMCSE) [26], which reduced the probability of undefined entropy and has been successfully used to study structural health monitoring system [27]. In 2018, in order to better study the time series from the stock market, Wu and his coworkers introduced modified multiscale sample entropy measure based on symbolic representation and similarity (MSEBSS) [28]. Recently, Wang et al. proposed multiscale cross-trend sample entropy (MCTSE) to study the similarity of two time series that with potential trends [29]. In addition, multivariate multiscale sample entropy algorithm has been proposed to deal with multivariate data [30,31,32]. Recently, Jamin and Humeau-Heurtier offered a state-of-the-art on cross-entropy measures and their multiscale approaches in [33].
On the other hand, when some scholars studied the long-range correlation between time series, they found that if two non-stationary time series are driven by a common third-party force or by common external factors, the result without considering the common third-party force may not reflect their intrinsic relationship [34,35,36]. Fortunately, Baba et al. [37] found that if two time series affected by the external factors are additive, the levels of intrinsic cross-correlation between two time series can be measured by the partial cross-correlation coefficient. In 2015, Yuan et al. [38] and Qian et al. [39] introduced partial cross-correlation analysis to deal with this kind of situation from different departure points.
Inspired by the above works, we propose the composite multiscale partial cross-sample entropy (CMPCSE) to measure the intrinsic similarity of two time series affected by the third common external factor simultaneously in this paper. We first test CMPCSE on three sets of artificial data, and find that it can reveal the intrinsic similarity of the time series come from the models, and then apply it to a set of stock market indices.

2. Composite Multiscale Partial Cross-Sample Entropy

In this section, based on CMCSE [26], we propose a new method-composite multiscale partial cross-sample entropy (CMPCSE), which can be used to quantify the intrinsic similarity of two time series linearly affected by a common external factor.
Consider two time series recorded simultaneously, { x ( t ) : t = 1 , 2 , , N } and { y ( t ) : t = 1 , 2 , , N } linearly affected by { z ( t ) : t = 1 , 2 , , N } , the main steps of CMPCSE are as follows:
Step 1: First we eliminate the effect of z ( t ) on x and y, respectively. The additive model for models of x ( t ) and y ( t ) can be given respectively as:
x ( t ) = β x , 0 + β x , 1 z ( t ) + r x ( t ) , y ( t ) = β y , 0 + β y , 1 z ( t ) + r y ( t ) ,
where, t = 1 , 2 , , N . When using regression analysis to estimate the value r x ( t ) , r y ( t ) , in a window of length s, we use the idea in MF-TWXDFA [40] to remove the effect of the sequence z ( t ) on x ( t ) and y ( t ) point by point as follows. For a given integer s ( s 2 ), the points j contained in a sliding window M W i corresponding to the point i should satisfy | i j | s . When the length of time series is different, we take different value for s. Usually the value of s is determined by experience. Accordingly, the weight function of the geographic weighted regression model is:
ω i j = [ 1 ( ( i j ) s ) 2 ] 2 , i f | i j | s , 0 , o t h e r w i s e .
In the window M W i , we perform linear regression for { ω i j x j } on { z j } or { ω i j y j } on { z j } , respectively. We can get the regression values x ^ ( z i ) and y ^ ( z i ) of x ( i ) and y ( i ) , respectively. Then we get the corresponding estimates of r x ( t ) , r y ( t ) :
r ^ x ( i ) = x ( i ) x ^ ( z i ) r ^ y ( i ) = y ( i ) y ^ ( z i ) .
Then the normalized data of r ^ x ( t ) , r ^ y ( t ) are defined as r ¯ x ( t ) = ( r ^ x ( t ) < r ^ x ( t ) > ) / δ r ^ x ( t ) and r ¯ y ( t ) = ( r ^ y ( t ) < r ^ y ( t ) > ) / δ r ^ y ( t ) , respectively. Here < . > and δ are the corresponding mean and standard deviation. Next, we calculate the CMCSE of r ¯ x ( t ) and r ¯ y ( t ) .
Step 2: Construct coarse-grained time series from the series r ¯ x ( t ) and r ¯ y ( t ) with the scale factor τ , respectively. Then we get { u k τ ( t ) } and { v k τ ( t ) } . Each point of the k-th coarse-grained time series at a scale factor of τ is defined as
u k τ ( j ) = 1 / τ i = ( j 1 ) τ + k j τ + k 1 r ¯ x ( i ) , 1 j ( N k + 1 ) / τ , 1 k τ .
v k τ ( j ) = 1 / τ i = ( j 1 ) τ + k j τ + k 1 r ¯ y ( i ) , 1 j ( N k + 1 ) / τ , 1 k τ .
For scale one ( τ = 1 ) , the times series u 1 1 and v 1 1 are the original series r ¯ x and r ¯ y . For τ > 1 , Figure 1 and Figure 2 show two more intuitive examples of the coarse-grained procedure.
Step 3: According to the following formula, construct vector sequences with length m
m h k τ ( i ) = ( u k τ ( i ) , u k τ ( i + 1 ) , , u k τ ( i + m 1 ) ) , { i : 1 i ( N k + 1 ) / τ m + 1 } ,
m w k τ ( j ) = ( v k τ ( j ) , v k τ ( j + 1 ) , , u k τ ( j + m 1 ) ) , { j : 1 j ( N k + 1 ) / τ m + 1 } ,
from { u k τ ( t ) } and { v k τ ( t ) } respectively. Let m n k τ ( i ) be the number of vectors m w k τ ( j ) whose distance with m h k τ ( i )
d ( m h k τ ( i ) , m w k τ ( j ) ) = m a x { | u k τ ( i + t ) v k τ ( j + t ) | : 0 t m 1 }
is within the tolerance r. And then m n k τ = i m n k τ ( i ) represents the total number of m-dimensional matched vector pairs and is obtained from the two k-th coarse-grained time series at a scale factor of τ . Similarly, m + 1 n k τ is the total number of matches of length m + 1 . Finally, the CMPCSE is calculated with the equation:
C M P C S E ( x , y , z , τ , m , r ) = C M C S E ( r ¯ x , r ¯ y , τ , m , r ) = 1 τ k C S E ( u k τ , v k τ , m , r ) = 1 τ k l n m + 1 n k τ m n k τ ,
where k means that neither m + 1 n k τ nor m n k τ is zero, that is, l n m + 1 n k τ m n k τ makes sense, and τ is the number that makes l n m + 1 n k τ m n k τ meaningful at a scale factor τ .
A more intuitive procedure of CMPCSE is shown in Figure 3.
In this paper, the entropies are calculated from scale 1 to 20, that is τ = 1 , 2 , 3 , , 20 . And the cross-sample entropy of each pair of coarse-grained series is calculated with m = 2 and r = r , where r is the value selected from the candidate set { 0.05 , 0.1 , 0.15 , , 0.95 } according to the criterion proposed by Lake et al. [16].

3. Numerical Experiments for Artificial Time Series

In this section, we use a additive model of x and y as Equation (9) to perform numerical simulation and verify the effectiveness of the CMPCSE.
x ( t ) = 2 + 3 z ( t ) + r x ( t ) , y ( t ) = 2 + 3 z ( t ) + r y ( t ) .
In the following numerical simulations, the series r x ( t ) , r y ( t ) are generated from the Bivariate Fractional Brownian Motion (BFBMs), TWO-component ARFIMA process and Multifractal binomial measures, respectively, and all the third party interference factor series z ( t ) are pink( 1 / f ) noise generated by the DSP System Toolbox in MATLAB 2016. In the experiments, all the results about the sequences with random terms are the average of 100 repeated results with series length N = 2 12 .

3.1. Bivariate Fractional Brownian Motion (BFBMs)

In this subsection, in order to test the performance of CMPCSE, we first use it to calculate the partial cross-sample entropy of BFBMs in the two sets of the above additive models (Equation (9)). The r x and r y are the incremental series of the two components of BFBMs with Hurst indices H r x and H r y . Extensive research on BFMS has been made. We know that BFBMs is a single fractal process and there is a relationship H r x r y = ( H r x + H r y ) / 2 [41,42,43]. Wei et al. studied the long-range power cross-correlations between r x and r y in 2017 [40]. In the simulations, we set: (left) H r x = 0.6 , H r y = 0.7 , ρ = 0.7 ; (right) H r x = 0.6 , H r y = 0.9 , ρ = 0.7 ; where ρ is the cross-correlation coefficient between r x and r y .
We apply the CMPCSE method to the series simulated by BFBMs and pink noise. Figure 4 shows the results between the series simulated by the pink noise and BFBMs with (left) H r x = 0.6 , H r y = 0.7 , ρ = 0.7 ; (right) H r x = 0.6 , H r y = 0.9 , ρ = 0.7 . From Figure 4 we can know that the entropy values of x y : z and r x r y are very close at all time scales, but there are obviously discrepancy between the values of x y : z and x y except when the time scale equal to 1, which indicates that, when r x , r y are affected by the third party factor z simultaneously, the CMPCSE method can capture the intrinsic cross-sample entropy values of r x , r y by eliminating the influence of z.

3.2. TWO-Component ARFIMA Process

ARFIMA process is a monofractal process [40] and often used to model the power-law auto-correlations in stochastic variables [44]. It is defined as follows:
g ( t ) = G ( t ) + ε g ( t ) ,
where d ( 0 , 0.5 ) is a memory parameter, ε g is an independent and identically distributed Gaussian variable, and G ( d , t ) = n = 1 a n ( d ) g ( t n ) , in which a n ( d ) is the weight a n ( d ) = d Γ ( n d ) / [ Γ ( 1 d ) Γ ( n + 1 ) ] . The Hurst index H G G is related to the memory parameters [45,46]. For the two-component ARFIMA processes discussed below, we take G = X or Y. The two-component ARFIMA process is defined as follows [47]:
r x ( t ) = W X ( d 1 , t ) + ( 1 W ) Y ( d 2 , t ) + ε r x ( t ) , r y ( t ) = ( 1 W ) X ( d 1 , t ) + ( 1 W ) Y ( d 2 , t ) + ε r y ( t ) ,
where W [ 0.5 , 1 ] quantifies the coupling strength between the two processes r x ( t ) and r y ( t ) . When W = 1 , r x ( t ) and r y ( t ) are fully decoupled and become two separate ARFIMA processes as defined in Equation (11). The cross-correlation between r x ( t ) and r y ( t ) increases when W decreases from 1 to 0.5 [47].
In the process of our calculations, we choose W = 0.8 and the parameters ( d 1 , d 2 ) of ARFIMA as d 1 = 0.1 , d 2 = 0.2 and d 1 = 0.1 , d 2 = 0.4 respectively, and corresponding two error terms ε r x ( t ) and ε r y ( t ) share one independent and identically distributed Gaussian variable with zero mean and unit variance. The CMPCSE method was used to the series simulated by two-component ARFIMA process and pink noise.
Figure 5 also shows that the entropy values of x y : z and r x r y are very close at all time scales, but there are obviously discrepancy between the values of x y : z and x y except when the time scale equal 1. It also means that, when r x , r y are affected by the third party factor z simultaneously, one can use the CMPCSE to get intrinsic cross-sample entropy values of r x , r y .

3.3. Multifractal Binomial Measures

In this subsection, the series r x , r y to be tested come from the binomial measures generated by p model with known analytic multifractal properties [40]. We combine them with pink noise to test the performance of CMPCSE. Each binomial measure or multifractal signal can be generated by iteration. We start with the iteration k = 0 , where the data set g ( i ) consists of one value, g ( 0 ) ( 1 ) = 1 . In the kth iteration, the data set { g ( k ) ( i ) , i = 1 , 2 , , z k } is obtained from g ( k ) ( 2 i 1 ) = p g ( k 1 ) ( i ) and g ( k ) ( 2 i ) = ( 1 p ) g ( k 1 ) ( i ) . When k , g ( k ) ( i ) approaches to a binomial measures, and the scaling exponent function H g g ( q ) is:
H g g ( q ) = 1 / q l o g 2 [ p q + ( 1 p ) q ] / q .
In our simulation, we iterated 12 times with p 1 = 0.2 , p 2 = 0.3 , p 3 = 0.4 and then get 3 binomial measures g p 1 ( i ) , g p 2 ( i ) , g p 3 ( i ) . In our actual calculation process, we set r x = diff ( g p ( i ) ) , here diff means the first order difference.
We present CMCSE results of the series x y , r x r y and the CMPCSE x y : z in Figure 6 with p x = 0.2 , p y = 0.3 and p x = 0.3 , p y = 0.4 . From the two pictures in Figure 6, we can easily find out that the entropy values of x y : z and r x r y are very close at all time scales, but there are obviously discrepancy between the values of x y : z and x y . It also indicates that, when r x , r y are affected by the third party factor z simultaneously, one can use the CMPCSE method to get intrinsic cross-sample entropy values of r x , r y by eliminating the influence of z on x , y .

4. Application to Stock Market Index

In order to validate the applicability of the CMPCSE method for empirical time series, we then apply it to stock market indices. The analyzed data sets consist of three Chinese stock indices: Shanghai securities composite index (SSEC), Shenzhen Stock Exchange Component Index (SZSE) and Hang Seng Index (HSI). All the raw data were download from https://finance.yahoo.com/. Then the daily closing data for the indices from 26 December 1999, to 17 July 2020, were used. Due to the different opening dates in mainland and Hong Kong, we exclude the data recorded on different dates and then reconnect the remaining parts of the original series to obtain time series with same length. As a result, the final daily closing data length is 5000.
In practice, we usually apply normalized time series. Denoting the closing index on the tth days as x ( t ) , the daily index return is defined by: g ( t ) = l n ( x ( t ) ) l n ( x ( t 1 ) ) . Then the normalized daily return is defined as R ( t ) = ( g ( t ) < g ( t ) > ) / δ , where < g ( t ) > and δ are the mean value and standard deviation of the series g ( t ) , respectively.
In 2015, Shi and Shang studied the multisacle cross-correlation coefficient and multisacle cross-sample entropy between SSEC, SZSE and HSI [48]. From their results, we can know that there is a strong correlation between the return data of SSEC and SZSE, and both them have weak correlation with HSI. The results of our estimation and comparison of the cross-sample entropy of the two return time series SSEC and SZSE, which includes two cases of including and excluding the influence of the HSI index, are shown in Figure 7. From the entropy measure results of return data in Figure 7, one can easily find that the entropy values of SSEC-SZSE are always bigger than SSEC-SZSE:HSI at all scales, which means that if the entropy values of SSEC-SZSE calculated by CMCSE are used to estimate the degree of similarity between SSEC and SZSE, the similarity between them will be underestimated. That is to say, the partial cross-sample entropy SSEC-SZSE:HSI can deliver more reasonable and real synchronization between the two return time series of SSEC and SZSE. We believe this result is reasonable, as SSEC and SZSE are the two most important stock indices in the mainland of china, so their daily return data should have strong synchronicity, especially under large time scales.

5. Discussion and Conclusions

In this paper, we proposed CMPCSE for quantifying intrinsic similarity of two time series affected by common external factors. Firstly, we described the calculation process of CMPCSE in detail. And then, in order to test the validity of CMPCSE, we applied it to three sets of artificial data. These three sets of artificial data were constructed by linear superposition of BFBMs, TWO-component ARFIMA process and Multifractal binomial measures with pink( 1 / f ) noise respectively. The results of each set of the artificial data show that CMPCSE can accurately measure the intrinsic cross-sample entropy of two simultaneously recorded time series by removing the effects that come from pink noise. At last, CMPCSE was employed to investigate the partial cross-sample entropy of SSEC and SZSE by eliminating the effect of HSI. Compared with the conclusion from CMCSE, the results from CMPCSE show that SSEC and SZSE have stronger similarity. Because SSEC and SZSE are the two most important stock indices in the mainland of China, they should have strong consistency, especially under large time scales, so we think the result is reasonable and it is necessary to consider partial cross-sample entropy when one wants to measure the similarity of SZSE and SSEC.
On the other hand, we must also note that the first step in the calculation of CMPCSE is crucial to the result of CMPCSE. Maybe there are other ways to eliminate the influence of the third party on the two time series that we studied. In our work, we adopted the idea from Reference [40] and satisfactory results were obtained in our artificial data examples. At the same time, in our research process, we also notice that when CMPCSE is used to study the linear combination of NBVP times series mentioned in Reference [26] and pink noise, which is constructed in the way mentioned above, we can not get satisfactory results. Therefore, we think that the way to eliminate the third-party influence in this paper can not achieve good results for the sequence with violent oscillation. Meanwhile, we expect to see better methods to deal with similar times series.
All in all, we think the partial cross-sample entropy analysis is necessary when one wants to measure the similarity of two times series affected by common external factors and, at present, CMPCSE is a good choice.

Author Contributions

B.L. contributed to the conception and design of the study, developed the method and wrote the manuscript. Z.Y. gave the ideas and supervised the project. G.H. and S.J. analyzed the data and results. All authors discussed the results and reviewed the manuscript, and approved the final manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This project was supported by the National Natural Science Foundation of China (Grant No. 11871061), Collaborative Research project for Overseas Scholars (including Hong Kong and Macau) of the National Natural Science Foundation of China (Grant No. 61828203), the Chinese Program for Changjiang Scholars and Innovative Research Team in University (PCSIRT) (Grant No. IRT_15R58) and the Hunan Provincial Innovation Foundation for Postgraduate (Grant No. CX2017B265).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Campillo, M.; Paul, A. Long-Range Correlations in the Diffuse Seismic Coda. Science 2003, 299, 547–549. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Auyang, S.Y. Foundations of Complex-System Theories: In Economics, Evolutionary Biology, and Statistical Physics; Cambridge University Press: Cambridge, UK, 1998. [Google Scholar]
  3. Plerou, V.; Stanley, H.E. Stock return distributions: Tests of scaling and universality from three distinct stock markets. Phys. Rev. E 2008, 77, 037101. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–423. [Google Scholar] [CrossRef] [Green Version]
  5. Rényi, A. On measures of entropy and information. In Proceedings of the 4th Berkeley Symposium on Mathematical Statistics and Probability, Berkeley, CA, USA, 20 June–30 July 1960; University of California Press: Berkeley, CA, USA, 1961; Volume 5073, pp. 547–561. [Google Scholar]
  6. Tsallis, C. Possible Generalization of Boltzmann-Gibbs Statistics. J. Stat. Phys. 1988, 52, 479–487. [Google Scholar] [CrossRef]
  7. Tribus, M. The Maximum Entropy Formalism; MIT Press: Cambridge, MA, USA, 1979. [Google Scholar]
  8. Grassberger, P.; Procaccia, I. Estimation of the Kolmogorov entropy from a chaotic signal. Phys. Rev. A 1983, 28, 2591–2593. [Google Scholar] [CrossRef] [Green Version]
  9. Eckmann, J.P.; Ruelle, D.; Ruelle, D. Ergodic theory of chaos and strange attractors. Rev. Mod. Phys. 1985, 57, 617–656. [Google Scholar] [CrossRef]
  10. Pincus, S.M. Approximate entropy as a measure of system complexity. Proc. Natl. Acad. Sci. USA 1991, 88, 2297–2301. [Google Scholar] [CrossRef] [Green Version]
  11. Pincus, S.M. Approximate entropy (ApEn) as a complexity measure. Chaos 1995, 5, 110–117. [Google Scholar] [CrossRef]
  12. Pincus, S.M. Quantifying complexity and regularity of neurobiological systems. Methods Neurosci. 1995, 28, 336–363. [Google Scholar]
  13. Pincus, S.M.; Viscarello, R.R. Approximate entropy: A regularity measure for fetal heart rate analysis. Obstet. Gynecol. 1992, 79, 249–255. [Google Scholar]
  14. Schuckers, S.A.C. Use of approximate entropy measurements to classify ventricular tachycardia and fibrillation. J. Electrocardiol. 1998, 31, 101–105. [Google Scholar] [CrossRef]
  15. Richman, J.S.; Moorman, J.R. Physiological time–series analysis using approximate entropy and sample entropy. Am. J. Physiol. Heart Circ. Physiol. 2000, 278, H2039–H2049. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Lake, D.E.; Richman, J.S.; Griffin, M.P.; Moorman, J.R. Sample entropy analysis of neonatal heart rate variability. Am. J. Physiol. Regul. Integr. Comp. Physiol. 2002, 283, R789–R797. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  17. Liu, L.Z.; Qian, X.Y.; Lu, H.Y. Cross-sample entropy of foreign exchange time series. Physica A 2010, 389, 4785–4792. [Google Scholar] [CrossRef]
  18. Costa, M.; Peng, C.K.; Goldberger, A.L.; Hausdorff, J.M. Multiscale entropy analysis of human gait dynamics. Physica A 2003, 330, 53–60. [Google Scholar] [CrossRef]
  19. Costa, M.; Goldberger, A.L.; Peng, C.K. Multiscale entropy analysis of biological signals. Phys. Rev. E 2005, 71, 021906. [Google Scholar] [CrossRef] [Green Version]
  20. Thuraisingham, R.A.; Gottwald, G.A. On multiscale entropy analysis for physiological data. Physica A 2006, 366, 323–332. [Google Scholar] [CrossRef]
  21. Peng, C.K.; Costa, M.; Goldberger, A.L. Adaptive data analysis of complex fluctuations in physiologic time series. Adv. Adapt. Data Anal. 2009, 1, 61–70. [Google Scholar] [CrossRef]
  22. Zhang, L.; Xiong, G.L.; Liu, H.S.; Zou, H.J.; Guo, W.Z. Bearing fault diagnosis using multi-scale entropy and adaptive neuro–fuzzy inference. Expert Syst. Appl. 2010, 37, 6077–6085. [Google Scholar] [CrossRef]
  23. Lin, J.L.; Liu, J.Y.C.; Li, C.W.; Tsai, L.F.; Chung, H.Y. Motor shaft misalignment detection using multiscale entropy with wavelet denoising. Expert Syst. Appl. 2010, 37, 7200–7204. [Google Scholar] [CrossRef]
  24. Wu, S.D.; Wu, C.W.; Lin, S.G.; Wang, C.C.; Lee, K.Y. Time series analysis using composite multiscale entropy. Entropy 2013, 15, 1069–1084. [Google Scholar] [CrossRef] [Green Version]
  25. Wu, S.D.; Wu, C.W.; Lin, S.G.; Lee, K.Y.; Peng, C.K. Analysis of complex time series using refined composite multiscale entropy. Phys. Lett. A 2014, 378, 1369–1374. [Google Scholar] [CrossRef]
  26. Yin, Y.; Shang, P.J.; Feng, G.C. Modified multiscale cross-sample entropy for complex time series. Appl. Math. Comput. 2016, 289, 98–110. [Google Scholar]
  27. Lin, T.K.; Chien, Y.H. Performance evaluation of an entropy-based structural health monitoring system utilizing composite multiscale cross-sample entropy. Entropy 2019, 21, 41. [Google Scholar] [CrossRef] [Green Version]
  28. Wu, Y.; Shang, P.J.; Li, Y.L. Multiscale sample entropy and cross-sample entropy based on symbolic representation and similarity of stock markets. Commun. Nonlinear Sci. Numer. Simul. 2018, 56, 49–61. [Google Scholar] [CrossRef]
  29. Wang, F.; Zhao, W.C.; Jiang, S. Detecting asynchrony of two series using multiscale cross–trend sample entropy. Nonlinear Dyn. 2020, 99, 1451–1465. [Google Scholar] [CrossRef]
  30. Ahmed, M.U.; Li, L.; Cao, J.; Mandic, D.P. Multivariate multiscale entropy for brain consciousness analysis. In Proceedings of the IEEE Engineering in Medicine and Biology Society (EMBC), Boston, MA, USA, 30 August–3 September 2011; pp. 810–813. [Google Scholar]
  31. Ahmed, M.U.; Mandic, D.P. Multivariate multiscale entropy: A tool for complexity analysis of multichannel data. Phys. Rev. E 2011, 84, 061918. [Google Scholar] [CrossRef] [Green Version]
  32. Looney, D.; Adjei, T.; Mandic, D.P. A Novel Multivariate Sample Entropy Algorithm for Modeling Time Series Synchronization. Entropy 2018, 20, 82. [Google Scholar] [CrossRef] [Green Version]
  33. Jamin, A.; Humeau-Heurtier, A. (Multiscale) Cross-Entropy Methods: A Review. Entropy 2020, 22, 45. [Google Scholar] [CrossRef] [Green Version]
  34. Kenett, D.Y.; Shapira, Y.; Ben-Jacob, E. RMT assessments of the market latent information embedded in the stocks’ raw, normalized, and partial correlations. J. Probab. Stat. 2009, 2009, 249370. [Google Scholar] [CrossRef] [Green Version]
  35. Kenett, D.Y.; Tumminello, M.; Madi, A.; Gur-Gershgoren, G.; Mantegna, R.N.; Ben-Jacob, E. Dominating clasp of the financial sector revealed by partial correlation analysis of the stock market. PLoS ONE 2010, 5, e15032. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Shapira, Y.; Kenett, D.Y.; Ben-Jacob, E. The index cohesive effect on stock market correlations. Eur. Phys. J. B 2009, 72, 657. [Google Scholar] [CrossRef]
  37. Baba, K.; Shibata, R.; Sibuya, M. Partial correlation and conditional correlation as measures of conditional independence. Aust. N. Z. J. Stat. 2004, 46, 657–664. [Google Scholar] [CrossRef]
  38. Yuan, N.M.; Fu, Z.T.; Zhang, H.; Piao, L.; Xoplak, E.I.; Luterbacher, J. Detrended Partial-Cross-Correlation Analysis: A New Method for Analyzing Correlations in Complex System. Sci. Rep. 2015, 5, 8143. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Qian, X.Y.; Liu, Y.M.; Jiang, Z.Q.; Podobnik, B.; Zhou, W.X.; Stanley, H.E. Detrended partial cross-correlation analysis of two nonstationary time series influenced by common external forces. Phys. Rev. E 2015, 91, 062816. [Google Scholar] [CrossRef] [Green Version]
  40. Wei, Y.L.; Yu, Z.G.; Zou, H.L.; Anh, V.V. Multifractal temporally weighted detrended cross-correlation analysis to quantify power-law cross-correlation and its application to stock markets. Chaos 2017, 27, 063111. [Google Scholar]
  41. Lavancier, F.; Philippe, A.; Surgailis, D. Covariance function of vector self–similar processes. Stat. Probab. Lett. 2009, 79, 2415–2421. [Google Scholar] [CrossRef] [Green Version]
  42. Coeurjolly, J.F.; Amblard, P.O.; Achard, S. On multivariate fractional Brownian motion and multivariate fractional Gaussian noise. In Proceedings of the 2010 18th European Signal Processing Conference, Aalborg, Denmark, 23–27 August 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1567–1571. [Google Scholar]
  43. Amblard, P.O.; Coeurjolly, J.F. Identification of the multivariate fractional Brownian motion. IEEE Trans. Signal Process. 2011, 59, 5152–5168. [Google Scholar] [CrossRef] [Green Version]
  44. Hosking, J.R.M. Fractional differencing. Biometrika 1981, 68, 165–176. [Google Scholar] [CrossRef]
  45. Podobnik, B.; Ivanov, P.; Biljakovic, K.; Horvatic, D.; Stanley, H.E.; Grosse, I. Fractionally integrated process with power-law correlations in variables and magnitudes. Phys. Rev. E 2005, 72, 026121. [Google Scholar] [CrossRef] [Green Version]
  46. Podobnik, B.; Stanley, H.E. Detrended cross-correlation analysis: A new method for analyzing two nonstationary time series. Phys. Rev. Lett. 2008, 100, 084102. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  47. Podobnik, B.; Horvatic, D.; Ng, A.L.; Stanley, H.E.; Ivanov, P.C. Modeling long-range cross-correlations in two-component ARFIMA and FIARCH processes. Physica A 2008, 387, 3954–3959. [Google Scholar] [CrossRef] [Green Version]
  48. Shi, W.B.; Shang, P.J. The multiscale analysis between stock market time series. Int. J. Mod. Phys. C 2015, 26, 1550071. [Google Scholar] [CrossRef]
Figure 1. Schematic illustration of the coarse-grained procedure of composite multiscale partial cross-sample entropy (CMPCSE) when τ = 2 . Modified from Reference [24].
Figure 1. Schematic illustration of the coarse-grained procedure of composite multiscale partial cross-sample entropy (CMPCSE) when τ = 2 . Modified from Reference [24].
Entropy 22 01003 g001
Figure 2. Schematic illustration of the coarse-grained procedure of CMPCSE when τ = 3 . Modified from Reference [24].
Figure 2. Schematic illustration of the coarse-grained procedure of CMPCSE when τ = 3 . Modified from Reference [24].
Entropy 22 01003 g002
Figure 3. Flow charts of the CMPCSE algorithms.
Figure 3. Flow charts of the CMPCSE algorithms.
Entropy 22 01003 g003
Figure 4. The CMPCSE results between the series simulated by the pink noise and bivariate fractional Brownian motion (BFBMs) with (left) H r x = 0.6 , H r y = 0.7 , ρ = 0.7 ; (right) H r x = 0.6 , H r y = 0.9 , ρ = 0.7 .
Figure 4. The CMPCSE results between the series simulated by the pink noise and bivariate fractional Brownian motion (BFBMs) with (left) H r x = 0.6 , H r y = 0.7 , ρ = 0.7 ; (right) H r x = 0.6 , H r y = 0.9 , ρ = 0.7 .
Entropy 22 01003 g004
Figure 5. The CMPCSE results between the series simulated by the pink noise and two-component ARFIMA process with (left) d 1 = 0.1 , d 2 = 0.2 , W = 0.8 ; (right) d 1 = 0.1 , d 2 = 0.4 , W = 0.8 .
Figure 5. The CMPCSE results between the series simulated by the pink noise and two-component ARFIMA process with (left) d 1 = 0.1 , d 2 = 0.2 , W = 0.8 ; (right) d 1 = 0.1 , d 2 = 0.4 , W = 0.8 .
Entropy 22 01003 g005
Figure 6. The CMPCSE results between the series simulated by the pink noise and first order difference series of the binomial measures (left) p x = 0.2 , p y = 0.3 ; (right) p x = 0.3 , p y = 0.4 .
Figure 6. The CMPCSE results between the series simulated by the pink noise and first order difference series of the binomial measures (left) p x = 0.2 , p y = 0.3 ; (right) p x = 0.3 , p y = 0.4 .
Entropy 22 01003 g006
Figure 7. Estimation and comparison of the cross-sample entropy between the two return time series Shanghai securities composite index (SSEC) and Shenzhen Stock Exchange Component Index (SZSE) when including and excluding the influence of the Hang Seng Index (HSI) index.
Figure 7. Estimation and comparison of the cross-sample entropy between the two return time series Shanghai securities composite index (SSEC) and Shenzhen Stock Exchange Component Index (SZSE) when including and excluding the influence of the Hang Seng Index (HSI) index.
Entropy 22 01003 g007

Share and Cite

MDPI and ACS Style

Li, B.; Han, G.; Jiang, S.; Yu, Z. Composite Multiscale Partial Cross-Sample Entropy Analysis for Quantifying Intrinsic Similarity of Two Time Series Affected by Common External Factors. Entropy 2020, 22, 1003. https://doi.org/10.3390/e22091003

AMA Style

Li B, Han G, Jiang S, Yu Z. Composite Multiscale Partial Cross-Sample Entropy Analysis for Quantifying Intrinsic Similarity of Two Time Series Affected by Common External Factors. Entropy. 2020; 22(9):1003. https://doi.org/10.3390/e22091003

Chicago/Turabian Style

Li, Baogen, Guosheng Han, Shan Jiang, and Zuguo Yu. 2020. "Composite Multiscale Partial Cross-Sample Entropy Analysis for Quantifying Intrinsic Similarity of Two Time Series Affected by Common External Factors" Entropy 22, no. 9: 1003. https://doi.org/10.3390/e22091003

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop