Next Article in Journal
Ultrasoft Classical Systems at Zero Temperature
Next Article in Special Issue
Bayesian Analysis of Tweedie Compound Poisson Partial Linear Mixed Models with Nonignorable Missing Response and Covariates
Previous Article in Journal
BRAQUE: Bayesian Reduction for Amplified Quantization in UMAP Embedding
Previous Article in Special Issue
Detecting Non-Overlapping Signals with Dynamic Programming
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Change-Point Detection in a High-Dimensional Multinomial Sequence Based on Mutual Information

1
School of Management, University of Science and Technology of China, Heifei 230026, China
2
Department of Mathematics and Statistics, York University, Toronto, ON M3J 1P3, Canada
*
Author to whom correspondence should be addressed.
Entropy 2023, 25(2), 355; https://doi.org/10.3390/e25020355
Submission received: 11 January 2023 / Revised: 4 February 2023 / Accepted: 13 February 2023 / Published: 14 February 2023
(This article belongs to the Special Issue Statistical Methods for Modeling High-Dimensional and Complex Data)

Abstract

:
Time-series data often have an abrupt structure change at an unknown location. This paper proposes a new statistic to test the existence of a change-point in a multinomial sequence, where the number of categories is comparable with the sample size as it tends to infinity. To construct this statistic, the pre-classification is implemented first; then, it is given based on the mutual information between the data and the locations from the pre-classification. Note that this statistic can also be used to estimate the position of the change-point. Under certain conditions, the proposed statistic is asymptotically normally distributed under the null hypothesis and consistent under the alternative hypothesis. Simulation results show the high power of the test based on the proposed statistic and the high accuracy of the estimate. The proposed method is also illustrated with a real example of physical examination data.

1. Introduction

The change-point problem was first proposed by Page [1]. It considers a model in which the distribution of the observed data changes abruptly at some point in time, which is common in biology [2], finance [3], literature [4] and epidemiology [5]. Change-point detection can be employed as a tool in time series segmentation. A typical reference in the field of study is [6]. Once a change-point is detected in a data sequence, it is used to split the data sequence into two segments so that both segments are modeled separately. On the other hand, from a practical point of view, behavior and policies can be adjusted based on changes in events of interest. So, it is very important to perform change-point detection.
There are mainly two problems in the change-point model: checking the existence of change points and estimating the positions of these change points. These issues have been studied in substantial literature. For example, see Sen and Srivastava [7] for the mean change in a normal distribution and Worsley [8] for a change in an exponential family using the maximum likelihood ratio method. Others include Bai [9] for the least squares estimate of mean shift in linear processes, Vexler [10] for the change-point problem in a linear regression model and Gombay [11] for the change-point in autoregressive time series, etc. See [12,13,14] for details.
A study has shown that most work on change-point problems has been done for continuous data [14]. In real life, however, many data are observed on a discrete scale. Common discrete distributions include binomial, multinomial and Poisson distributions. In this article, we consider the change-point problem in a multinomial sequence, which originated from the transcription of the Gospels [15]. The Lindisfarne Gospels were divided into several sections, assuming that only one author contributed to the writing of any section and that the sections written by any one author were continuous. The goal was to test whether a single author wrote the Gospels. The data may be the frequency of vocabulary or grammar used by the author of each section. In general, suppose that X 1 = ( X 11 , , X 1 m ) , , X K = ( X K 1 , , X K m ) are K independent multinomial variables with parameter ( n 1 , p 1 ) , , ( n K , p K ) , where p i = ( p i 1 , , p i m ) , j = 1 m p i j = 1 , i = 1 , , K . On the i t h point, there are n i experiments with m outcomes, and X i records the frequencies of m outcomes. We want to test
H 0 : p 1 = = p K v . s . H 1 : p 1 = = p k * p k * + 1 = = p K
where k * is the true change-point, 1 < k * < K . If H 0 is rejected, we further estimate k * .
To solve this problem, Wolfe and Chen [16] proposed several statistics based on the cumulative sum (CUSUM) method. Horváth and Serbinowska [17] used the maximum likelihood ratio and maximum chi-square statistic to test the existence of change points and derived their transformed limit distribution. Batsidis and Horváth [18] extended it and proposed a family of phi-divergence tests that involves broad statistics. Riba and Ginebra [19] performed a graphical exploration of a sequence of polynomial observations and found a break point. Note that they all assumed that the number of categories m is fixed.
In recent years, the rise of big data has made the high-dimensional change-point problem more important. Thus, it becomes urgent to consider high-dimensional multinomial data as the categories of one thing in life can be quite large, such as the type of stores selling certain items on a shopping platform and the type of illness of patients in an outpatient clinic during a day. In this paper, we consider problem (1) with m tending to infinity. Recently, Wang et al. [20] proposed a procedure based on Pearson’s Chi-square test under the above scenario. Their idea is to pre-divide categories into two based on their probability magnitudes and to use the original and modified Pearson’s Chi-square statistic for large and small categories, respectively. This pre-classification can balance sparse and dense signals, resulting in good statistical performance. So, here, we use the pre-classification idea to construct a test statistic for problem (1) with m tending to infinity.
Another tool used in this article is based on information entropy. Entropy, originally a concept in statistical physics, was introduced into information theory by Shannon [21]. It has been widely applied in change-point problems. Unakafov and Keller [22] used ordinal mode conditional entropy to detect change points. Ma and Sofronov [23] proposed a cross-entropy algorithm to estimate the number and positions of change points. Vexler and Gurevic [24] applied empirical likelihood method to change-point detection, in which the essence of empirical likelihood estimation is a density-based entropy estimation. Mutual information, denoted by MI, computed as the difference between entropy and conditional entropy, is popular in deep learning, for example, see [25,26]. In the area of machine learning, MI is similar to information gain, which is often used as a measure of the goodness of a step in an algorithm, such as the selection of node splitting in a tree. Therefore, MI is naturally used as a metric in event detection problems. Relevant works include [27,28], etc. We utilize the MI between data and their position in this paper, given that a large value of MI means a high probability that a change point occurs.
In this paper, we consider the offline change-point problem. We propose a test statistic based on the mutual information for the at most one change-point (AMOC) problem (1) with m tending to infinity as the sample size tends to infinity. We adopt the pre-classification idea in [20] here. The optimal change-point position can also be estimated by MI. We show that the proposed statistic has an asymptotic normal distribution under the null distribution, and the power of the test converges to one under the alternative hypothesis. Meanwhile, we point out the relationship between MI and the likelihood ratio. In fact, the proposed statistic is based on the likelihood ratio method. As is widely acknowledged, although there is no uniformly most powerful test for change-point detection in general [29,30], the test based on the likelihood ratio structure has a high power [31]. Simulation studies demonstrate the excellent power of the test based on the proposed statistic as well as the high accuracy of the estimation. The innovations we have made in this article are that we replace the Pearson Chi-square statistic in Wang et al. [20] with mutual information and achieve better performance in terms of power and accuracy compared to their method.
The remaining structure of this paper is as follows. In Section 2, we present the proposed test statistic and the estimation method of a change point. In Section 3, we provide simulation results. In Section 4, we illustrate the method with an example based on physical examination data. In Section 5, we conclude the paper with some remarks. The proofs of the theorems are given in Appendix A.

2. Methods

2.1. Entropy and Mutual Information

We first briefly introduce some concepts about entropy and mutual information.
Definition 1.
Suppose that x 1 , , x u are the possible values taken by a random variable X, where u can be infinity. Let P X ( x i ) be the probability that X = x i . The Shannon entropy of X is defined as
H ( X ) = i = 1 u P X ( x i ) log P X ( x i ) .
when P X ( x i ) = 0 , define P X ( x i ) log P X ( x i ) = 0 .
Definition 2.
Let Y be a random variable that takes values in { y 1 , , y v } , where v can be infinity. The conditional entropy of X given Y is defined as
H ( X | Y ) = i = 1 u j = 1 v P X , Y ( x i , y j ) log P X | Y ( x i | y j ) ,
where P X , Y , P X | Y are the joint probabilities of X and Y and the conditional probability of X given Y, respectively.
Definition 3.
Assume that X and Y are the same as in Definitions 1 and 2. The mutual information (MI) of X relative to Y is defined as
MI ¯ ( X ; Y ) = i = 1 u j = 1 v P X , Y ( x i , y j ) log P X , Y ( x i , y j ) P X ( x i ) P Y ( y j ) .
The entropy value is larger when the data distribution is more symmetric. On the contrary, when the data are skewed, they have a small entropy [32]. The conditional entropy measures how much uncertainty is eliminated in X by observing Y. Obviously, mutual information can be written as the difference between entropy and conditional entropy, that is, MI ¯ ( X ; Y ) = H ( X ) H ( X | Y ) . It represents the average amount of information about X that can be gained or the amount of reduction of uncertainty in X by observing Y. MI ¯ ( X ; Y ) 0 , and it becomes zero if X and Y are independent of each other.

2.2. Pre-Classification

For multinomial data, when the number of categories m is large, it is sometimes not realistic to treat all categories equally. For example, of all the cities in China, only a few of them account for half of the economy, which means that the rest of the cities have a small average share. The well-known Pareto principle [33] that 20% of the population owns 80% of the wealth in society also illustrates this phenomenon. Therefore, it is reasonable to classify the categories with different orders of magnitude.
Consider problem (1), i.e.,
H 0 : p 1 = = p K v . s . H 1 : p 1 = = p k * p k * + 1 = = p K .
We denote p 1 = = p K = q 0 under H 0 , and p 1 = = p k * = q 0 , p k * = = p K = q 1 under H 1 . Denote q l = ( q l 1 , , q l m ) , l = 0 , 1 . Note that q 0 q 1 .
Similar to Wang et al. [20], let B 0 be a subset of { 1 , , m } such that max j B 0 q 0 j a m 0 , B 1 be a subset of { 1 , , m } such that max j B 1 q 1 j a m 0 , where a m 1 is O ( 1 ) satisfying some conditions as m . Let A 0 = B 0 c and A 1 = B 1 c , where the superscript c stands for the complement operator. Assume that min j A 0 q 0 j a m > ε and min j A 1 q 1 j a m > ε for some ε > 0 as m . Let A = A 1 A 0 and B = A c . Then, m categories are divided into large and small orders of magnitude by a m denoted by A and B. A change from q 0 to q 1 q 0 might occur either in A or B.
Let X i A be the component of X i in A for i = 1 , , K and q 0 A be the component of q 0 in A. Let X i B and q 0 B be similarly defined. Then, the marginal distributions of X i A and X i B under the null assumption are
( X i A , j B X i j ) Multi ( n i , ( q 0 A , 1 j A q 0 j ) ) ,
and
( X i B , j A X i j ) Multi ( n i , ( q 0 B , 1 j B q 0 j ) ) .
In the next subsection, we construct a statistic built on the marginal distributions (5) and (6).
Here are some additional notations. Denote N = i = 1 K n i , N 0 k = i = 1 k n i , N 1 k = i = k + 1 K n i as the number of experiments in total, before and after time k, and Z = i = 1 K X i , Z 0 k = i = 1 k X i , Z 1 k = i = k + 1 K X i as the number of successful trials in total, before and after time k. Let q ^ = Z N , q ^ 0 k = Z 0 k N 0 k , and q ^ 1 k = Z 1 k N 1 k be the corresponding frequencies.
For the data in A, let Z A = i = 1 K X i A , Z 0 k A = i = 1 k X i A , Z 1 k A = i = k + 1 K X i A be the number of successful trials in total, before and after time k. Define Z B S = i = 1 K j B X i j , Z 0 k B S = i = 1 k j B X i j , and Z 1 k B S = i = k + 1 K j B X i j as the sum of successful trials in B of total, before, and after k. Let q ^ A j = Z A j N , q ^ 0 k A j = Z 0 k A j N 0 k , q ^ 1 k A j = Z 1 k A j N 1 k , j A , q ^ B S = Z B S N , q ^ 0 k B S = Z 0 k B S N 0 k , q ^ 1 k B S = Z 1 k B S N 1 k be the corresponding frequencies. Subscript S denotes the sum of frequencies. Similarly, we define Z B , Z 0 k B , Z 1 k B , Z A S , Z 0 k A S , Z 1 k A S , q ^ B j , q ^ 0 k B j , q ^ 1 k B j , j B , q ^ A S , q ^ 0 k A S , q ^ 1 k A S . We illustrate some of the above notations in Table 1 in a more structured fashion.

2.3. Test Statistic

We use MI between the data X = ( X 1 , , X K ) and the location of the data to construct the statistic. For the data in A, the entropy is
H A = q ^ B S log q ^ B S j A q ^ A j log q ^ A j .
The entropies in A before and after k are
H 0 k A = q ^ 0 k B S log q ^ 0 k B S j A q ^ 0 k A j log q ^ 0 k A j
and
H 1 k A = q ^ 1 k B S log q ^ 1 k B S j A q ^ 1 k A j log q ^ 1 k A j ,
respectively.
Denote Y k = I { the location of X i is before k } as the indicator function of the position of a sample relative to k. Note that, given the observations, P ( Y k = 1 ) = N 0 k N by the independence. By Section 2.1, the MI between X and Y k in A is
H A H A ( X | Y k ) = ^ MI ¯ k A ,
where H A ( X | Y k ) = P ( Y k = 0 ) H 0 k A + P ( Y k = 1 ) H 1 k A = N 0 k N H 0 k A + N 1 k N H 1 k A is the conditional entropy of X given Y k . Similarly, MI ¯ k B = H B N 0 k N H 0 k B N 1 k N H 1 k B , where H B , H 0 k B and H 1 k B are defined similarly as in (7)–(9).
The uncertainty of X given Y k would reach the largest reduction if k is at the true break point k * ; hence, either MI ¯ k * A or MI ¯ k * B should be large. On the contrary, if the sequence is stable, the value of MI ¯ k should be small for any k { 1 , , K } .
Since A and B are unknown, in light of Wang et al. [20], we use A ^ = { j : q ^ j a m > C ε } to estimate A. Here, C > 0 is some constant. As shown in [20], A ^ is a consistent estimator of A if a m satisfies certain assumptions. Let B ^ = A ^ c . Construct the test statistic
G m , A ^ = 2 K k = 1 K N 0 k N 1 k N MI ¯ k B ^ + e m I ( max k = 1 , , K 2 N MI ¯ k A ^ > r m )
for (1). Summation and maximization are conducted respectively for the MI of A ^ and B ^ in G m , A ^ . The first term in G m , A ^ is the weighted log-likelihood ratio estimate, as pointed out after Lemma 1. The second term in G m , A ^ is based on the maximum norm of MI. It is widely acknowledged that the max-norm test is more suitable for sparse and strong signals, see [34,35]. r m is a threshold for A ^ , which ensures that the second term in G m , A ^ converges to zero under H 0 . e m is a large number. Note that the statistic in [20] is based on the Pearson Chi-square statistic. Since in reality, the frequencies of small categories might be zeros, the Pearson Chi-square statistic for B ^ is hence modified. The statistic presented here does not need to take into account the fact that a frequency may be zero, since by the definition of entropy, p log p = 0 if p = 0 . In order to study the properties of G m , A ^ better, we first give a lemma about MI .
Lemma 1.
Denote L k A = 2 log j A q ^ A j Z A j q ^ B S Z B S j A q ^ 0 k A j Z 0 k A j q ^ 0 k B S Z 0 k B S j A q ^ 1 k A j Z 1 k A j q ^ 1 k B S Z 1 k B S . Then, 2 N MI ¯ k A = L k A . It is also true by replacing A with B in all the subscripts, that is, 2 N MI ¯ k B = L k B .
Note that L k A and L k B in Lemma 1 are estimations of minus two log likelihood ratios for data in A and B when the change-point is at k. Therefore, the problem based on MI can be transformed into the problem based on likelihood ratios.
By Lemma 1, the second term in (11) is e m I ( max k = 1 , , K L k A ^ > r m ) , and hence the existing limit theorems on likelihood ratios can be applied to it directly. The first term in (11) is 1 K k = 1 K N 0 k N 1 k N 2 L k B ^ , which has the form of a weighted log likelihood ratio estimation. In Appendix A, we show that it is only an infinitesimal quantity away from some CUSUM statistic [36] using Taylor expansion and then prove the asymptotic distribution of G m , A ^ from related conclusions.
The sum of L k B ^ without weighting, k = 1 K L k B ^ , is closely related to the Shiryayev–Roberts procedure [37,38]. It uses k = 1 K Λ k as a statistic, where Λ k is the likelihood ratio when the change point is at k. It is widely applied to determine the best stopping criterion in sequential change-point monitoring (see, e.g., [39]). However, replacing unknown parameters in k = 1 K Λ k with their maximum likelihood estimation, which leads to k = 1 K L k B ^ in this paper, would result in a complex asymptotic analysis [40]. So, here we use the weighted version 1 K k = 1 K N 0 k N 1 k N 2 L k B ^ instead of k = 1 K L k B ^ .
Theorem 1.
Let | A | denote the cardinality of any set A and sup | A | denote the maximal cardinality of the set A. Assume that sup | A | = d < , and
(i) 
N a m 2 ( log a m ) 1 as ( m , K ) ,
(ii) 
[ a ( log N ) ] 2 r m [ b d ( log N ) ] 2 and log e m a ( log N ) r m 1 / 2 b d ( log N ) 0 as ( m , K ) ,
(iii) 
lim sup K m max 1 k K 2 ( n k + 1 N 0 k ) 1 2 log N 0 k < ,
lim sup K m max K 2 k K ( n k + 1 N 1 k ) 1 2 log N 1 k < .
Then, under H 0 ,
G m , A ^ m d + 1 6 m d + 1 45 d N ( 0 , 1 )
as ( m , K ) , where a ( x ) = ( 2 log x ) 1 2 , b d ( x ) = 2 log x + ( d / 2 ) log log x log Γ ( d / 2 ) .
Theorem 1 shows that G m , A ^ is asymptotically normally distributed under the null hypothesis. The condition (i) in Theorem 1 ensures the consistency of A ^ , which was also assumed in Theorem 1 of [20]. The condition ( i i ) in Theorem 1 requires the threshold r m to be large enough in order to guarantee that e m I ( max k = 1 , , K 2 N MI ¯ k A ^ > r m ) converges to zero with probability one under the null hypothesis. Condition ( i i i ) means that every n i is much less than N. Next, we focus on the properties of the statistic under the alternative hypothesis.
Theorem 2.
Assume that the conditions (i)–( i i i ) in Theorem 1 hold. Let δ j = q 1 j q 0 j , j = 1 , , m . Further assume that
(i) 
e m > c m as ( m , K ) , where c > 1 6 .
(ii) 
N k * N κ 0 as ( m , K ) , κ 0 ( 0 , 1 ) , and there exist 0 < c 1 , c 2 < such that c 1 < N 0 k N , N 0 k N < c 2 for k = 1 , , K as ( m , K ) .
If the shift sizes δ j s satisfy either of the following two conditions,
(iii) 
N δ j 2 r m 1 for some j A ,
(iv) 
| δ j | > 0 for some j B ,
then as ( m , K ) ,
P G m , A ^ m d + 1 6 m d + 1 45 > z 1 α 1 ,
where z 1 α is the critical value of the standard normal distribution at level α.
Theorem 2 establishes the consistency of the test under certain conditions when the probability in A or B changes. Condition (i) in Theorem 2 means that e m tends to infinity at a certain rate. It aims to ensure that G m , A ^ tends to infinity when the parameters in A change. Condition ( i i ) requires comparable sample sizes before and after the change point. The proofs of Theorem 1 and Theorem 2 are provided in Appendix A.
Once H 0 is rejected, we further use MI to estimate k * . If max k = 1 , , K 2 N MI ¯ k A ^ > r m , then k ^ = arg max k = 1 , , K MI ¯ k A ^ ; otherwise, k ^ = arg max k = 1 , , K MI ¯ k B ^ . Numeric studies in the next section show that the power of the new statistic increases rapidly as the difference between the alternative hypothesis and the null hypothesis increases. At the same time, the precision of k ^ using pre-classification is also satisfactory.

3. Simulation

We conduct simulation experiments to assess the performance of the test procedures in empirical size, power and estimation in finite samples. All results are based on 1000 replications. We use R to obtain simulation results. The necessary R code is given in Appendix B.
To analyze the empirical size, we simulate multinomial data with parameter q 0 = ( ω d 1 d , 1 ω m d 1 m d ) under the null hypothesis without break with reference to [20]. The first d probabilities are much greater than those of the latter. Hence, in reality, A ^ can be chosen as { 1 , , d ^ } . Following [20], we use d ^ = arg max i = 1 , , m 1 1 q ^ ( i ) q ^ ( i + 1 ) 1 + q ^ 2 ( i ) 1 + q ^ 2 ( i + 1 ) , where q ^ ( 1 ) q ^ ( m ) are the sorted values of q ^ j . We consider different situations with the sample size K arranged from 50 to 500, and let m = K in each situation. For simplicity, we fix n i = 100 , i = 1 , , K . For the formula of G m , A ^ , we choose e m = m , r m = ( 2 log log N + d 2 log log log N ) 2 according to the conditions in the above section. The simulation results with various combinations of ( ω , d) are reported in Table 2. We observe that the empirical size of the test is 4.5–6.7%, which is thus around the nominal 5% level in different situations. Here, we show the case of ω 0.5 . We also performed simulations for ω > 0.5 and found empirical values slightly higher than 5% (data not shown).
To evaluate the power of the test, the alternative hypotheses stipulate a single break in the data sequence. We first consider parameters of two forms:
(i)
q 1 = ( ( 1 + s ) ω d 1 d 2 , ( 1 s ) ω d 1 d 2 , 1 ω m d 1 m d ) d % 2 = 0 , ( ( 1 + s ) ω d 1 d 1 2 , ω [ 1 1 + s 2 d 1 d ] d + 1 2 1 d + 1 2 , 1 ω m d 1 m d ) otherwise ;
(ii)
q 1 = ( ω d 1 d , ( 1 + s ) 1 ω m d 1 m d 2 , ( 1 s ) 1 ω m d 1 m d 2 ) ( m d ) % 2 = 0 , ( ω d 1 d , ( 1 + s ) 1 ω m d 1 m d 1 2 , ( 1 ω ) [ 1 1 + s 2 m d 1 m d ] m d + 1 2 1 m d + 1 2 ) otherwise .
ω is the proportion in A, 0 < ω < 1 . 1 d denotes a d-dimensional vector with all components equal to 1. s represents the shift size. We consider different values of s when evaluating power and accuracy to better observe changes in efficiency and accuracy as the gap between the alternative and null hypotheses increases. % is the mod operation. The two alternative hypotheses assume that the change point is located in A and B, respectively. We consider k * = 0.2 K and 0.5 K to capture breaks in the beginning and middle of a sample. For comparison, we use two competitors:
  • The weighted maximum likelihood ratio statistic L = max k = 1 , , K N 0 k N 1 k N 2 Λ ^ k where Λ ^ k = 2 log ( j = 1 , m q ^ j Z j j = 1 , m q ^ 0 k j Z 0 k j j = 1 , m q ^ 1 k j Z 1 k j ) proposed by Horváth and Serbinowska [17];
  • The statistic Q = k = 1 K j B ^ ( L k j L k j ( 0 ) ) + e m I ( max k = 1 , , K max j A ^ R k j > r m ) in [20], in which L k j = N 0 k N 1 k N ( Z 0 k j N 0 k Z 1 k j N 1 k ) 2 , L k j ( 0 ) = N 0 k N 1 k N ( Z 0 k j N 0 k 2 + Z 1 k j N 1 k 2 ) and R k j = L k j q ^ j , e m = K 4 3 and r m = log K log K in the simulations.
The results are summarized in Figure 1 for level α = 0.05 . The size of L is on the high side, as seen from the curve at small s in Figure 1. The new test is very powerful, as evidenced by the rapid rate of convergence to 1 when s increases. In most cases, the empirical power of G m , A ^ is larger than the other two for alternative hypothesis (i). For the alternative hypothesis (ii), the three statistics perform equally well. These results further show that our test has higher power to detect a change located in the middle of the sample than in the beginning while the power is also still high.
We also briefly investigate how well the change-point location k * is approximated by the estimator k ^ . We choose k * = 0.2 K and 0.5 K as the change-point location. In Table 3 and Table 4, we report the mean and standard deviation of the absolute errors | k ^ k * | for the different choices of s and m under the alternative hypothesis (i) or (ii), respectively. We compare our estimate with the maximum likelihood ratio estimate k ^ L = arg max k = 1 , , K Λ ^ k and k ^ Q in [20].
The corresponding absolute errors in Table 3 and Table 4 underscore the considerable precision of k ^ , which improves when s is increased 0.3 from to 0.8. For the alternative (i), in almost all situations, k ^ is better than the other two competitors. Small changes (for example, s = 0.3) are found with greater difficulty by using k ^ L and k ^ Q , while the precision of k ^ remains high. For the alternative hypothesis (ii), k ^ and k ^ L have similar performance, and they are both slightly better than k ^ Q . Alternative Hypothesis (i): Assume that the large probability changes while the small probability remains the same. Alternative Hypothesis (ii): Assume that the small probability changes while the large probability remains the same. Under alternative hypothesis (i), our method has better performance than the other two methods, probably because entropy as a non-linear function can increase the difference between frequencies, and it is more pronounced when the difference is small (e.g., s = 0.3).
Finally, we simulate the power and estimation precision for alternative hypothesis (iii):
q 1 = ( ( 1 + s ) ω d 1 d 2 , ( 1 s ) ω d 1 d 2 , ( 1 + s ) 1 ω m d 1 m d 2 , ( 1 s ) 1 ω m d 1 p d 2 ) , d % 2 = 0 , ( m d ) % 2 = 0 , ( ( 1 + s ) ω d 1 d 1 2 , ω [ 1 1 + s 2 d 1 d ] d + 1 2 1 d + 1 2 , ( 1 + s ) 1 ω p d 1 m d 1 2 , ( 1 ω ) [ 1 1 + s 2 m d 1 m d ] m d + 1 2 1 m d + 1 2 ) , otherwise ,
where parameters in A and B change simultaneously, which was not mentioned in [20]. We compare our statistic and k ^ with Q and k ^ Q in this case. The results are displayed in Figure 2, Table 5 and Table 6, from which we see that the power of G m , A ^ is slightly better than that of Q, and the precision of k ^ is obviously higher than that of k ^ Q .

4. Example

In this section, we use a data set to address the applicability of our method. The data concern the medical examination results of people working in Hefei’s financial sector (including banks and insurance companies) from 27 September 2017 to 25 August 2021, which includes each person’s age, gender, the date of examination and the disease detected. From the perspective of health analysis and disease prevention, it is thus important to understand the diseases in terms of how often they are detected.
Our goal is to test whether the proportion of people who have been diagnosed with some diseases change over time. After removing gender-specific diseases, we finally choose 210 diseases. Because in some weeks there is no person to have the examination, we eliminate those weeks and finally keep 173 weeks. Let X i be a 210-dimension vector with each component indicating the frequency of a certain disease detected during the i t h time period. Then, there are K = 173 vectors X 1 , , X 173 of dimension m = 210 with N = 16596 outcomes.
Figure 3 shows the numbers of the top 30 diseases detected. The weekly sample size n i s are provided in Figure 4. We find from Figure 3 that the numbers of the first six diseases, Fatty Liver (FL), Overweight (OW), Thyroid Nodule (TN), Pulmonary Nodule (PN), Hepatic Cyst (HC) and Thyroid Cyst (TC), were much higher than those of the other diseases. By calculation, their proportions were, respectively, 0.088, 0.086, 0.06, 0.06, 0.038 and 0.032, which accounted for 35.8% of all the detected diseases. Hence, we choose d ^ = 6 . The value of the statistic is G m A ^ = 200.2577 , and hence the null hypothesis that there is no change in the proportions of diseases detected is rejected.
Because max k = 1 , , K 2 N MI ¯ k A ^ > r m , we find that k ^ = arg max k = 1 , , K MI ¯ k A ^ = 14 , corresponding to 27 December 2017. This suggests that the proportions of diseases detected vary before and after 2018. Table 7 displays the proportions of the first six diseases before and after 2018. The proportions of Overweight and Thyroid Nodule were the highest before 2018. However, after 2018, the proportion of Fatty Liver jumped to the highest, and the proportion of Pulmonary Nodule also increased significantly. A possible explanation is that some unexpected events lead to changes in people’s lifestyles, which lead to changes in the proportion of the population suffering from different diseases. For example, the start of the Sino–US trade war in early February 2018 led to a continuous decline in the price of China’s A-shares, which was the trigger for the change in the lifestyle of financial practitioners after 2018. The study into the proportions of people with different diseases in the financial sector can reveal which disease is on the rise in this sector, and hence proper recommendations can be made for disease prevention.

5. Conclusions

This paper develops a change-point test based on MI for multinomial data when the number of categories is comparable to the sample size. We show that under certain conditions, the proposed statistic is asymptotically normal under the null hypothesis and consistent under the alternative hypothesis. The simulation results suggest that the test based on the proposed statistic has a high power. The proposed inference procedures are used to analyze the change in proportions of diseases detected in physical examination data during a period.

Author Contributions

Conceptualization and methodology, B.J., X.X. and Y.W.; software and writing—original draft preparation, X.X.; writing—review and editing, B.J. and Y.W.; supervision, B.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation grant number 7201101228, 12231017, Natural Science Foundation of Anhui Province of China grant number 2108085J02 and Natural Science and Engineering Research Council of Canada grant number RGPIN-2017-05720.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors are grateful to the referees for their insightful comments in revising this paper.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MIMutual Information
FLFatty Liver
OWOverweight
TNThyroid Nodule
PNPulmonary Nodule
HCHepatic Cyst
TCThyroid Cyst

Appendix A

Proof of Lemma 1.
2 N MI ¯ k A = 2 N ( H A N 0 k N H 0 k A N 1 k N H 1 k A ) = 2 N H A 2 N 0 k H 0 k A 2 N 1 k H 1 k A = 2 [ j A Z A j log q ^ A j + Z B S log q ^ B S ( j A Z A j log q ^ 0 k A j + Z 0 k B S log q ^ 0 k B S ) ( j A Z 1 k A j log q ^ 1 k A j + Z 1 k B S log q ^ 1 k B S ) ] = L k A .
Lemma A1.
If N a m 2 ( log a m ) 1 as ( m , K ) , then
(i) 
under H 0 , P ( A ^ = A 0 ) 1 as ( m , K ) for any 0 < C < 1 ;
(ii) 
under H 1 , P ( A ^ = A 0 A 1 ) 1 as ( m , K ) for any 0 < C < min ( κ 0 , 1 κ 0 ) 2 , where κ 0 is defined in condition ( i i ) of Theorem 2.
Proof of Lemma A1.
See the proof of Theorem 1 in Wang et al. [20] □
Proof of Theorem 1.
By Lemma 1 and Lemma A1, it suffices to deduce the distribution of G m , A = 1 K k = 1 K N 0 k N 1 k N 2 L k B + e m I ( max k = 1 , , K L k A > r m ) .
We first show that under H 0 , e m I ( max k = 1 , , K L k A > r m ) = o p ( 1 ) . Then, we only need prove E ( e m 2 I ( max k = 1 , , K L k A > r m ) ) = e m 2 P ( max k = 1 , , K L k A > r m ) 0 as ( m , K ) .
Because sup | A | = d < , according to Theorem 1.1 in [17],
lim m K { P ( a ( log N ) ( max k = 1 , , K L k A ) 1 2 t + b d ( log N ) } = exp ( 2 e t ) .
Then
P ( max k = 1 , , K L k A > r m ) = P ( a ( log N ) ( max k = 1 , , K L k A ) 1 2 b d ( log N ) > a ( log N ) r m 1 2 b d ( log N ) ) = 1 P ( a ( log N ) ( max k = 1 , , K L k A ) 1 2 b d ( log N ) a ( log N ) r m 1 2 b d ( log N ) ) ,
and by condition (ii), we have
lim m K e m 2 P ( max k = 1 , , K L k A > r m ) = lim m K e m 2 ( 1 exp { 2 e ( a ( log N ) r m 1 2 b d ( log N ) ) } ) = 0 .
For simplicity, we suppress the subscript B in L k B , i.e., write L k B as L k . By a second-order Taylor expansion,
L k = 2 [ j Z 0 k j log q ^ j q ^ 0 k j + j Z 1 k j log q ^ j q ^ 1 k j ] = 2 [ j Z 0 k j log ( q ^ j q ^ 0 k j 1 + 1 ) + j Z 1 k j log ( q ^ j q ^ 1 k j 1 + 1 ) ] = 2 [ j Z 0 k j ( q ^ j q ^ 0 k j 1 ) 1 2 j Z 0 k j ( q ^ j q ^ 0 k j 1 ) 2 + j Z 1 k j ( q ^ j q ^ 1 k j 1 ) 1 2 j Z 1 k j ( q ^ j q ^ 1 k j 1 ) 2 ] + o p ( 1 ) = j Z 0 k j ( q ^ j q ^ 0 k j 1 ) 2 + j Z 0 k j ( q ^ j q ^ 1 k j 1 ) 2 + o p ( 1 ) = N 0 k 2 N 1 k 2 N 2 j Z j Z 0 k j Z 1 k j ( q ^ 0 k j q ^ 1 k j ) 2 + o p ( 1 ) = j N 0 k N 1 k N ( q ^ 0 k j q ^ 1 k j ) 2 / q ^ j + o p ( 1 ) .
The last inequality follows from the fact that Z j N / Z 0 k j N 0 k 1 and Z 1 k j N 1 k / Z 0 k j N 0 k 1 a s ( m , K ) under H 0 .
Let P k j = N 0 k N 1 k N ( q ^ 0 k j q ^ 1 k j ) 2 / q ^ j , and P k = j P k j . Then from the above, we have L k = P k + o p ( 1 ) . Rewrite q ^ 0 k and q ^ 1 k as 1 N 0 k i = 1 N 0 k Y i and 1 N 0 k i = 1 N 0 k Y i , respectively, where Y i = ( Y i 1 , , Y i m ) i . i . d Multi ( 1 , q 0 ) . Denote S k = 1 N ( i = 1 N 0 k Y i N 0 k N i = 1 N Y i ) , Σ ^ 1 n = diag ( q ^ 1 B , , q ^ m d + 1 B ) , where the q ^ j B denotes the estimated proportions by ( X i B , j A X i j ) . Then N 0 k N 1 k N 2 P k = S k T Σ ^ n 1 S k . By Remark 2.1 in [36],
1 K k = 1 K N 0 k N 1 k N 2 P k B m d + 1 6 m d + 1 45 d N ( 0 , 1 ) , a s ( m , K ) .
Hence, Theorem 1 follows from L k B = P k B + o p ( 1 ) . □
Proof of Theorem 2.
First assume that δ j s satisfy the condition ( i i i ) in Theorem 2. By Theorem 1.7.2 in [41], max k = 1 , , K L k A ^ = max k = 1 , , K P k A ^ + O p ( exp ( ( log N ) 1 ε ) ) for any 0 < ε < 1 . Since exp ( ( log N ) 1 ε ) 0 as ( m , K ) , by the proof of Theorem 3 in [20], we have
P ( max k = 1 , , K 2 N MI ¯ k A ^ > r m ) = P ( max k = 1 , , K L k A ^ > r m ) = P ( max k = 1 , , K P k A ^ + O p ( exp ( ( log N ) 1 ε ) ) > r m ) P ( P k * A ^ + o p ( 1 ) > r m ) P ( P k * A ^ j + o p ( 1 ) > r m ) 1 , a s ( m , K ) .
Hence, with probability one, G m , A ^ > e m . By the condition that e m > c m as ( m , K ) , where c > 1 6 , we obtain that
G m , A ^ m d + 1 6 m d + 1 45 , a s ( m , K ) .
Therefore, the power converges to 1 as ( m , K ) .
Now, assume that δ j s satisfy the condition ( i v ) in Theorem 2. By Theorem 2.2 in [36], for δ j > 0 for some j B , we have 2 K k = 1 K N 0 k N 1 k N M I k B ^ converges to a nonzero limit with probability one, which suggests that the power converges to 1 as ( m , K ) . □

Appendix B

Necessary R code related to this article can be found online at https://github.com/xxrdragonfly/ChangePoint/blob/main/Rcode.R (accessed on 3 Feburary 2023).

References

  1. Page, E.S. Continuous inspection schemes. Biometrika 1954, 41, 100–115. [Google Scholar] [CrossRef]
  2. Fletcher, R.J.; Robertson, E.P.; Poli, C.; Dudek, S.; Gonzalez, A.; Jeffery, B. Conflicting nest survival thresholds across a wetland network alter management benchmarks for an endangered bird. Biol. Conserv. 2021, 253, 108893. [Google Scholar] [CrossRef]
  3. Fryzlewicz, P. Wild binary segmentation for multiple change-point detection. Ann. Stat. 2014, 42, 2243–2281. [Google Scholar] [CrossRef]
  4. Ross, G.J.; Chevalier, A.; Sharples, L. Tracking the evolution of literary style via Dirichlet-multinomial change point regression. J. R. Stat. Soc. Ser. A-Stat. Soc. 2019, 183, 149–167. [Google Scholar] [CrossRef]
  5. Jiang, F.; Zhao, Z.; Shao, X. Time series analysis of COVID-19 infection curve: A change-point perspective. J. Econom. 2020, 232, 1–17. [Google Scholar] [CrossRef]
  6. Palivonaite, R.; Lukoseviciute, K.; Ragulskis, M. Algebraic segmentation of short nonstationary time series based on evolutionary prediction algorithms. Neurocomputing 2013, 121, 354–364. [Google Scholar] [CrossRef]
  7. Sen, A.K.; Srivastava, M.S. On tests for detecting change in mean. Ann. Stat. 1975, 3, 98–108. [Google Scholar] [CrossRef]
  8. Worsley, K.J. Confidence regions and tests for a change-point in a sequence of exponential family of random variables. Biometrika 1986, 73, 91–104. [Google Scholar] [CrossRef]
  9. Bai, J. Least squares estimation of a shift in linear processes. J. Time Ser. Anal. 1994, 15, 453–472. [Google Scholar] [CrossRef]
  10. Vexler, A. Guaranteed testing for epidemic changes of a linear regression model. J. Stat. Plan. Inference 2006, 136, 3101–3120. [Google Scholar] [CrossRef]
  11. Gombay, E. Change detection in autoregressive time series. J. Multivar. Anal. 2008, 99, 451–464. [Google Scholar] [CrossRef]
  12. Truong, C.; Oudre, L.; Vayatis, N. Selective review of offline change point detection methods. Signal Process. 2020, 167, 107299. [Google Scholar] [CrossRef]
  13. Aue, A.; Horváth, L. Structural breaks in time series. J. Time Ser. Anal. 2013, 34, 1–16. [Google Scholar] [CrossRef]
  14. Chen, J.; Gupta, A.K. Parametric Statistical Change Point Analysis; Birkhäuser: Boston, MA, USA, 2000. [Google Scholar]
  15. Ross, A.S.C. Philological probability problems. J. R. Stat. Soc. Ser.-Stat. Methodol. 1950, 12, 19–59. [Google Scholar] [CrossRef]
  16. Wolfe, D.A.; Chen, Y.S. The changepoint problem in a multinomial sequence. Commun.-Stat.-Simul. Comput. 1990, 19, 603–618. [Google Scholar] [CrossRef]
  17. Horváth, L.; Serbinowska, M. Testing for changes in multinomial observations: The Lindisfarne scribes problem. Scand. J. Stat. 1995, 22, 371–384. [Google Scholar]
  18. Batsidis, A.; Horváth, L.; Martín, N.; Pardo, L.; Zografos, K. Change-point detection in multinomial data using phi-divergence test statistics. J. Multivar. Anal. 2013, 118, 53–66. [Google Scholar] [CrossRef]
  19. Riba, A.; Ginebra, J. Change-point estimation in a multinomial sequence and homogeneity of literary style. J. Appl. Stat. 2005, 32, 61–74. [Google Scholar] [CrossRef]
  20. Wang, G.H.; Zou, C.L.; Yin, G.S. Change-point detection in multinomial data with a large number of categories. Ann. Stat. 2018, 46, 2020–2044. [Google Scholar] [CrossRef]
  21. Shannon, C.E. A mathematical theory of communication. Bell Syst. Tech. J. 1948, 27, 379–432. [Google Scholar] [CrossRef]
  22. Unakafov, A.M.; Keller, K. Change-Point Detection Using the Conditional Entropy of Ordinal Patterns. Entropy 2018, 20, 709. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  23. Ma, L.J.; Sofronov, G. Change-point detection in autoregressive processes via the Cross-Entropy method. Algorithms 2020, 13, 128. [Google Scholar] [CrossRef]
  24. Vexler, A.; Gurevich, G. Density-Based Empirical Likelihood Ratio Change Point Detection Policies. Commun.-Stat.-Simul. Comput. 2010, 39, 1709–1725. [Google Scholar] [CrossRef]
  25. Kamimura, R. Supposed maximum mutual information for improving generalization and interpretation of multi-layered neural networks. J. Artif. Intell. Soft Comput. Res. 2019, 9, 123–147. [Google Scholar] [CrossRef]
  26. Liu, L.X. Image multi-threshold method based on fuzzy mutual information. Comput. Eng. Appl. 2009, 45, 166–168, 197. [Google Scholar]
  27. Oh, B.S.; Sun, L.; Ahn, C.S.; Yeo, Y.K.; Yang, Y.; Liu, N.; Lin, Z.P. Extreme learning machine based mutual information estimation with application to time-series change-points detection. Neurocomputing 2017, 261, 204–216. [Google Scholar] [CrossRef]
  28. Kopylova, Y.; Buell, D.A.; Huang, C.T.; Janies, J. Mutual information applied to anomaly detection. J. Commun. Netw. 2008, 10, 89–97. [Google Scholar] [CrossRef]
  29. Gurevich, G. Retrospective parametric tests for homogeneity of data. Commun.-Stat.-Theory Methods 2007, 36, 2841–2862. [Google Scholar] [CrossRef]
  30. James, B.; James, K.L.; Siegmund, D. Tests for a change-point. Biometrika 1987, 74, 71–83. [Google Scholar] [CrossRef]
  31. Lai, T.L. Sequential changepoint detection in quality control and dynamical systems. J. R. Stat. Soc. Ser. B Stat. Methodol. 1995, 57, 613–658. [Google Scholar] [CrossRef]
  32. Lee, W. A Data Mining Framework for Constructing Features and Models for Intrusion Detection Systems. Ph.D. Thesis, Columbia University, New York, NY, USA, 1999. [Google Scholar]
  33. Pareto, V. Cours d’Economie Politique; Droz: Geneva, Switzerland, 1896. [Google Scholar]
  34. Chen, S.X.; Qin, Y.L. A two-sample test for high-dimensional data with applications to gene-set testing. Ann. Stat. 2010, 38, 808–835. [Google Scholar] [CrossRef]
  35. Fan, J.; Liao, Y.; Yao, J. Power enhancement in high-dimensional cross-sectional tests. Econometrica 2015, 83, 1497–1541. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Aue, A.; Hrmann, S.; Horváth, L.; Reimherr, M. Break detection in the covariance structure of multivariate time series models. Ann. Stat. 2009, 37, 4046–4087. [Google Scholar] [CrossRef]
  37. Shiryayev, A.N. On optimum methods in quickest detection problems. Theory Probab. Its Appl. 1963, 13, 22–46. [Google Scholar] [CrossRef]
  38. Roberts, S.W. A comparison of some control chart procedures. Technometrics 1966, 8, 411–430. [Google Scholar] [CrossRef]
  39. Krieger, A.M.; Pollak, M.; Yakir, B. Surveillance of a Simple Linear Regression. J. Am. Stat. Assoc. 2003, 98, 456–469. [Google Scholar] [CrossRef]
  40. Vexler, A.; Gregory, G. Average Most Powerful Tests for a Segmented Regression. Commun.-Stat.-Theory Methods 2009, 38, 2214–2231. [Google Scholar] [CrossRef]
  41. Csörgo, M.; Horváth, L. Limit Theorems in Change-Point Analysis; Wiley: New York, NY, USA, 1997. [Google Scholar]
Figure 1. Empirical power of three statistics for level α = 0.05 . (a) The power under the alternative hypothesis (i). (b) The power under the alternative hypothesis (ii). G denotes the proposed statistic. L is the weighted maximum likelihood ratio statistic in [17]. Q is the statistic in [20]. ω = 0.3 , d = 5 .
Figure 1. Empirical power of three statistics for level α = 0.05 . (a) The power under the alternative hypothesis (i). (b) The power under the alternative hypothesis (ii). G denotes the proposed statistic. L is the weighted maximum likelihood ratio statistic in [17]. Q is the statistic in [20]. ω = 0.3 , d = 5 .
Entropy 25 00355 g001
Figure 2. Empirical power of G m A ^ and Q under alternative hypothesis (iii). ω = 0.3 , d = 5 .
Figure 2. Empirical power of G m A ^ and Q under alternative hypothesis (iii). ω = 0.3 , d = 5 .
Entropy 25 00355 g002
Figure 3. The numbers of top 30 diseases detected.
Figure 3. The numbers of top 30 diseases detected.
Entropy 25 00355 g003
Figure 4. Weekly sample size.
Figure 4. Weekly sample size.
Entropy 25 00355 g004
Table 1. Explanation of some notations.
Table 1. Explanation of some notations.
TotalBefore kAfter k
ExperimentN N 0 k N 1 k
CategoryABABAB
Successful trials ( Z A , Z B S ) ( Z B , Z A S ) ( Z 0 k A , Z 0 k B S ) ( Z 0 k B , Z 0 k A S ) ( Z 1 k A , Z 1 k B S ) ( Z 1 k B , Z 1 k A S )
Frequency ( q ^ A , q ^ B S ) ( q ^ B , q ^ A S ) ( q ^ 0 k A , q ^ 0 k B S ) ( q ^ 0 k B , q ^ 0 k A S ) ( q ^ 1 k A , q ^ 1 k B S ) ( q ^ 1 k B , q ^ 1 k A S )
Table 2. Empirical sizes of G m , A ^ at the nominal test size 5% under different situations.
Table 2. Empirical sizes of G m , A ^ at the nominal test size 5% under different situations.
ω , dm
50100200300500
(0.3, 5)0.0460.0530.0520.0450.053
(0.3, 6)0.0520.0580.0590.0520.052
(0.3, 10)0.0530.0390.0540.0600.057
(0.5, 6)0.0500.0340.0470.0660.067
(0.5, 8)0.0530.0450.0540.0620.062
(0.5, 10)0.0530.0510.0610.0660.054
Table 3. Mean and standard deviation (in parentheses) of | k ^ k * | , | k ^ L k * | and | k ^ Q k * | under the alternative hypothesis (i) or (ii) with ω = 0.3 , d = 5 , and k * = 0.5 K .
Table 3. Mean and standard deviation (in parentheses) of | k ^ k * | , | k ^ L k * | and | k ^ Q k * | under the alternative hypothesis (i) or (ii) with ω = 0.3 , d = 5 , and k * = 0.5 K .
msAlternative Hypothesis (i)Alternative Hypothesis (ii)
| k ^ k * | | k ^ L k * | | k ^ Q k * | | k ^ k * | | k ^ L k * | | k ^ Q k * |
2000.31.88(3.35)15.95(27.15)25.97(34.74)0.64(1.16)0.64(1.18)0.66(1.20)
0.40.77(1.32)2.52(5.02)3.94(6.90)0.20(0.50)0.20(0.50)0.21(0.53)
0.50.49(0.91)1.11(1.93)2.11(3.16)0.06(0.25)0.06(0.25)0.06(0.27)
0.60.31(0.67)0.53(0.96)1.67(2.59)0.01(0.11)0.01(0.11)0.01(0.11)
0.70.16(0.44)0.29(0.63)1.12(1.94)000
0.80.16(0.52)0.16(0.48)0.87(1.63)000
5000.31.57(2.32)10.17(28.95)6.40(9.60)0.57(1.11)0.56(1.10)0.57(1.07)
0.40.89(1.49)2.24(3.41)3.52(4.59)0.16(0.46)0.16(0.46)0.18(0.48)
0.50.50(0.92)1.08(1.68)2.18(3.15)0.07(0.26)0.07(0.26)0.07(0.27)
0.60.31(0.67)0.51(1.06)1.61(2.46)0.01(0.11)0.01(0.10)0.02(0.13)
0.70.15(0.40)0.31(0.67)1.10(1.83)000
0.80.09(0.32)0.14(0.42)0.88(1.39)000
Table 4. Mean and standard deviation (in parentheses) of | k ^ k * | , | k ^ L k * | and | k ^ Q k * | under the alternative hypothesis (i) or (ii) with ω = 0.3 , d = 5 , and k * = 0.2 K .
Table 4. Mean and standard deviation (in parentheses) of | k ^ k * | , | k ^ L k * | and | k ^ Q k * | under the alternative hypothesis (i) or (ii) with ω = 0.3 , d = 5 , and k * = 0.2 K .
msAlternative Hypothesis (i)Alternative Hypothesis (ii)
| k ^ k * | | k ^ L k * | | k ^ Q k * | | k ^ k * | | k ^ L k * | | k ^ Q k * |
2000.310.01(31.28)33.16(48.90)66.82(58.79)0.86(1.71)0.87(1.70)0.89(1.87)
0.40.95(1.61)8.21(23.85)19.98(41.21)0.24(0.59)0.25(0.62)0.25(0.60)
0.50.73(0.26)1.82(5.59)2.45(3.94)0.07(0.27)0.08(0.28)0.09(0.33)
0.60.71(1.29)0.75(1.35)1.62(2.49)0.02(0.15)0.02(0.15)0.03(0.18)
0.70.45(0.88)0.33(0.70)1.2(2.17)000
0.80.30(0.66)0.20(0.49)0.93(1.73)000
5000.31.77(2.73)52.01(105.27)54.39(110.28)0.80(1.29)0.80(1.26)0.83(1.46)
0.40.94(1.48)3.42(6.47)3.82(5.64)0.22(0.54)0.23(0.56)0.24(0.57)
0.50.80(1.34)1.40(2.27)2.52(3.88)0.07(0.28)0.07(0.28)0.07(0.28)
0.60.68(1.22)0.73(1.22)1.56(2.27)0.01(0.09)0.01(0.09)0.02(0.13)
0.70.40(0.82)0.36(0.75)1.09(1.67)0.01(0.08)0.01(0.08)0
0.80.31(0.66)0.20(0.48)0.82(1.29)000
Table 5. Mean and standard deviation (in parentheses) of | k ^ k * | and | k ^ Q k * | under alternative hypothesis (iii) with ω = 0.3 , d = 5 and k * = 0.5 K .
Table 5. Mean and standard deviation (in parentheses) of | k ^ k * | and | k ^ Q k * | under alternative hypothesis (iii) with ω = 0.3 , d = 5 and k * = 0.5 K .
s m = 200 m = 500
| k ^ k * | | k ^ Q k * | | k ^ k * | | k ^ Q k * |
0.31.82(2.63)4.01(6.68)1.62(2.68)5.59(8.32)
0.40.82(1.42)3.42(4.76)0.82(0.82)3.06(4.49)
0.50.47(0.89)2.24(3.37)0.46(1.07)2.33 (3.33)
0.60.24(0.58)1.52(2.48)0.25(0.57)1.37(2.15)
0.70.17(0.48)1.10(1.72)0.12(0.41)1.16(1.77)
0.80.19(0.51)0.83(1.31)0.10(0.36)0.84(1.42)
Table 6. Mean and standard deviation (in parentheses) of | k ^ k * | and | k ^ Q k * | under alternative hypothesis (iii) with ω = 0.3 , d = 5 and k * = 0.2 K .
Table 6. Mean and standard deviation (in parentheses) of | k ^ k * | and | k ^ Q k * | under alternative hypothesis (iii) with ω = 0.3 , d = 5 and k * = 0.2 K .
s m = 200 m = 500
| k ^ k * | | k ^ Q k * | | k ^ k * | | k ^ Q k * |
0.31.67(2.61)1.70(3.69)1.80(2.85)4.89(7.62)
0.40.98(1.58)3.22(6.08)0.85(1.42)3.59(5.02)
0.50.84(1.41)2.44(3.78)0.55(1.11)2.54(3.80)
0.60.54(1.01)1.54(2.41)0.69(1.18)1.67(2.45)
0.70.42(0.91)1.28(2.12)0.44(0.87)1.16(1.91)
0.80.29(0.65)0.88(1.40)0.28(0.62)0.87(1.47)
Table 7. Proportions of the first six diseases before and after k ^ .
Table 7. Proportions of the first six diseases before and after k ^ .
DiseaseFLOWTNPNHCTC
Proportionbefore k ^ 0.0780.0950.1000.0210.0350.045
after k ^ 0.0900.0840.0510.0620.0380.029
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xiang, X.; Jin, B.; Wu, Y. Change-Point Detection in a High-Dimensional Multinomial Sequence Based on Mutual Information. Entropy 2023, 25, 355. https://doi.org/10.3390/e25020355

AMA Style

Xiang X, Jin B, Wu Y. Change-Point Detection in a High-Dimensional Multinomial Sequence Based on Mutual Information. Entropy. 2023; 25(2):355. https://doi.org/10.3390/e25020355

Chicago/Turabian Style

Xiang, Xinrong, Baisuo Jin, and Yuehua Wu. 2023. "Change-Point Detection in a High-Dimensional Multinomial Sequence Based on Mutual Information" Entropy 25, no. 2: 355. https://doi.org/10.3390/e25020355

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop