Next Article in Journal
Refined Discontinuous Trigger Scheme for Event-Based Synchronization of Chaotic Neural Networks
Previous Article in Journal
On the Convergence Order of Jarratt-Type Methods for Nonlinear Equations
Previous Article in Special Issue
A Learning Path Recommendation Approach in a Fuzzy Competence Space
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Change-Point Estimation and Detection for Mixture of Linear Regression Models

1
School of Science, Xi’an Polytechnic University, Xi’an 710048, China
2
School of Mathematics, Northwest University, Xi’an 710127, China
*
Author to whom correspondence should be addressed.
Axioms 2025, 14(6), 402; https://doi.org/10.3390/axioms14060402
Submission received: 15 April 2025 / Revised: 14 May 2025 / Accepted: 19 May 2025 / Published: 26 May 2025

Abstract

:
This paper studies the estimation and detection problems in the mixture of linear regression models with change point. An improved Expectation–Maximization (EM) algorithm is devised specifically for multi-classified mixture data with change points. Under appropriate conditions, the large-sample properties of the estimator are rigorously proven. This improved EM algorithm not only precisely locates the change points but also yields accurate parameter estimates for each class. Additionally, a detector grounded in the score function is proposed to identify the presence of change points in mixture data. The limiting distributions of the test statistics under both the null and alternative hypotheses are systematically derived. Extensive simulation experiments are conducted to assess the effectiveness of the proposed method, and comparative analyses with the conventional EM algorithm are performed. The results clearly demonstrate that the EM algorithm without considering change points exhibits poor performance in classifying data, often resulting in the misclassification or even omission of certain classes. In contrast, the estimation method introduced in this study showcases remarkable accuracy and robustness, with favorable empirical sizes and powers.

1. Introduction

Mixture models have found extensive applications in econometrics and social sciences, and the associated theoretical frameworks have been thoroughly investigated. Among them, finite mixtures of linear regression models, a particularly valuable class of mixture models, have been widely employed across diverse fields, including econometrics, marketing, epidemiology, and biology. Simultaneously, the statistical inference and parameter estimation based on mixture models play a pivotal role in practical applications. On another front, the change point problem remains a focal point in statistical research. Conceptually, a change point denotes the specific location or instant at which a sudden alteration in a statistical property occurs. Change point phenomena are pervasive in both natural and social systems, and have enabled crucial advancements in numerous domains such as industrial control, financial economics, biomedicine, and signal processing. The detection and estimation of change points within mixture datasets are essential for discerning underlying patterns and making informed decisions, thereby holding significant practical implications.
In recent decades, remarkable progress has been achieved in the analysis of mixture models. Goldfeld and Quandt provided an outstanding introduction to mixtures of linear regression models [1]. They delved into the parameter estimation problem of such mixtures and applied it to analyze the imbalance in the housing price market. Frühwirth and Schnatter summarized the Bayesian approach for mixtures of linear regression models in their paper [2]. Hurn and Justel proposed a generalized linear finite mixture model and overcame the label switching issue by normalizing the loss function [3]. Li, Chen, and their collaborators (Chen, Li, and Fu) concentrated on the hypothesis testing problem regarding the order of the mixture model and derived the limiting distribution of the test statistics [4,5]. Huang and Yao introduced a semiparametric mixture regression model, where the mixing proportion is a smooth function of covariates [6]. Huang, Li, and Wang developed a non-parametric finite mixture method for regression models and demonstrated the estimation procedure through an analysis of U.S. Housing Price Index (HPI) data [7].
The change point problem has attracted extensive research attention, with a vast body of literature accumulated over the years. It was first introduced by Page during his study of continuous inspection schemes, who proposed the Cumulative Sum (CUSUM) test method [8]. Subsequently, scholars like Sen and Srivastava, Hawkins, James et al., and Srivastava and Worsley investigated the detection of change points in the mean of normal random variable sequences [9,10,11,12,13]. Basseville and Nikiforov systematically expounded the theoretical foundation of change point detection and estimation algorithms [14]. Wang, Zou, and Yin delved into change point detection in multinomial data as the number of categories approaches infinity, rigorously proving the statistical properties of relevant test statistics [15]. Xia and Qiu proposed the JIC criterion to tackle the multi-change point problem in non-parametric regression models [16]. Bai employed the least squares method to estimate the mean change point of linear processes, deriving the consistency and convergence rate of the change point estimator [17]. Baranowski et al. utilized grid search algorithms based on binary segmentation to accurately determine the number and positions of change points [18]. Follain et al. developed a novel approach for estimating change points in partially observed, high-dimensional time series with simultaneous mean shifts in sparse coordinate subsets [19]. Gong et al. redefined the graph-based change point detection problem as a prediction task and proposed an innovative change point detection (CPD) method for dynamic graphs through a latent evolution model. Drabech et al. introduced a Markov Random Field (MRF) model to detect slope changes [20], while Ratnasingam et al. proposed an empirical likelihood-based non-parametric procedure for identifying structural changes in quantile regression models [21]. Notably, despite these abundant studies, the statistical analysis of change points within finite mixtures of linear models remains a relatively overlooked area [22].
This paper delves into the estimation and detection problems of the finite mixture of linear regression models with change points. Specifically, for multi-classified mixture data incorporating change points, we ingeniously propose an improved Expectation–Maximization (EM) algorithm. This algorithm is meticulously designed to accurately compute parameter estimators, such as the location of change points and regression coefficients. Through rigorous mathematical proofs, we firmly establish the consistency and asymptotic normality of these parameter estimators, providing a solid theoretical foundation. Moreover, to effectively detect the presence of change points in the mixture of linear regression models, we innovatively develop a detector based on the score function. By conducting in-depth analyses, we systematically derive the limiting distributions under both the null hypothesis and the alternative hypothesis, which significantly contributes to the theoretical framework of this study. To thoroughly verify the practical effectiveness of the proposed method, we carefully design and execute a series of simulation experiments.
The remaining sections of this paper are structured as follows. In Section 2, we introduce the estimation and detection methods, statistical models, along with their associated assumptions. In Section 3 is dedicated to the development of the estimation and detection techniques for change points within the mixture of linear regression models. Additionally, we obtain the large sample properties of the estimators and test statistics. In Section 4, the simulation results are presented. Section 5 provides some concluding remarks and discussions. In Section 6, the limitations and future research directions are presented. The proofs of the theorems are included in the appendix.

2. Statistical Model and Assumptions

Assume that  { ( X i , Y i ) , i = 1 , , n }  are independent random samples from the following data generating process:
Y i c = 1 C π c N X i T β c , σ c 2 , 1 i k , c = 1 C π c + N X i T β c + , σ c 2 + , k + 1 i n , k [ 1 , n 1 ] ,
where  k  is a change point, X i is a p-dimensional random covariate, C is the number of components,  β c , π c , σ c ,    β c + , π c + , σ c + are mixture coefficients for the cth component,  1 π c 0 , π c + 0 , c = 1 C π c = c = 1 C π c + = 1 . π c and π c + are called mixing proportions or weights. β c , β c + R p , σ c 2 , σ c 2 + R + and the parameters are unknown. Equivalently speaking, { Y i , i = 1 , , n } follow a finite mixture of normals,
Y i c = 1 C π c N ( X i T β c , σ c 2 ) · I { i k } + c = 1 C π c + N ( X i T β c + , σ c 2 + ) · I { i k + 1 } .
Naturally, the log-likelihood function corresponding to the dataset { ( X i , Y i ) , i = 1 , , n }  can be expressed as
l ( k , π c , β c , σ c 2 , π c + , β c + , σ c 2 + ) = i = 1 n log c = 1 C π c ϕ Y i | X i T β c , σ c 2 · I { i k } + π c + ϕ Y i | X i T β c + , σ c 2 + · I { i k + 1 }
where ϕ ( y | μ , σ 2 ) denotes the normal density.
For later use, we introduce some notations. Let θ = ( π ) T , ( β ) T , ( σ 2 ) T T denote the true parameter, where π = ( π 1 , , π C 1 ) T ,   β = ( β 1 T , , β C T ) T , σ 2 = ( σ 1 2 , , σ C 2 ) T  and  β i R P , i = 1 , , C . Let  θ , θ +  represent the parameters on the left and right sides of the change point, respectively. Similarly, we can define θ , θ , θ + .
Let k ^ n  and  θ ^ n be the maximum likelihood estimator of k and θ , which can be obtained by maximizing (2). In Section 3.1, an improved EM algorithm is proposed to carry out the estimation. The following assumptions are imposed.
Assumption 1.
k n τ n and  τ ( 0 , 1 ) .
The above assumption requires that  k  depends on the sample size n and is approximately proportional to n.
Assumption 2.
We suppose that θ Θ and parameter space Θ R q is a compact set, q = ( p + 2 ) · C 1 .
Assumption 2 restricts the parameter space to be compact, which means any open cover of Θ has a finite subcover.
Assumption 3.
{ ( X i , Y i ) , i = 1 , 2 , , n } are independent.

3. Parameter Inference

3.1. Estimation Procedure

In this subsection, we propose an effective EM algorithm to deal with the parameter estimation problem. We take  i k + 1  as an example to illustrate the derivation of the iterative formula (  i < k  can be obtained in the same way). At this time, the log-likelihood function corresponding to the data is
i = k + 1 n log c = 1 C π c + ϕ { Y i | X i T β c + , σ c 2 + } .
In the EM framework, the mixture problem is described as an incomplete-data problem. We view the observed data ( X i , Y i , i = 1 , 2 , , n ) as being incomplete, and then introduce unobserved  B e r n o u l l i  random variables
z i c = 1 , i f ( X i , Y i ) i s i n t h e c t h g r o u p , 0 , o t h e r w i s e ,
i = 1 , 2 , , n , c = 1 , 2 , , C .
and  z i = ( z i 1 , , z i C ) T , which is the component label of ( X i , Y i ) . Therefore, the complete data are  { ( X i , Y i , z i ) } i = 1 n , and the complete log-likelihood function corresponding to (3) is
i = k + 1 n c = 1 C z i c log π c + + log ϕ { Y i | X i T β c + , σ c 2 + } .
Supposing in the lth cycle of the EM algorithm iteration, we have π c + ( l ) , β c + ( l ) , σ c 2 + ( l ) ; then, in the E-step of  ( l + 1 ) th cycle, the expectation of the latent variable  z i c can be calculated by
r i c ( l + 1 ) = π c + ( l ) ϕ { Y i | X i T β c + ( l ) , σ c 2 + ( l ) } c = 1 C π c + ( l ) ϕ { Y i | X i T β c + ( l ) , σ c 2 + ( l ) } .
In the M-step of the  ( l + 1 ) th cycle, we maximize
i = k + 1 n c = 1 C r i c ( l + 1 ) log π c + + log ϕ { Y i | X i T β c + , σ c 2 + } .
The maximization of Equation (4) is equivalent to separately maximize   i = k + 1 n c = 1 C r i c ( l + 1 ) log π c + and i = k + 1 n c = 1 C r i c ( l + 1 ) log ϕ { Y i | X i T β c + , σ c 2 + } . Furthermore, the iterative formula can be easily derived as follows:
π c + ( l + 1 ) = i = k + 1 n r i c ( l + 1 ) n ,
β c + ( l + 1 ) = ( i = k + 1 n X i r i c ( l + 1 ) X i T ) 1 ( i = k + 1 n X i r i c ( l + 1 ) Y i ) ,
σ c 2 + ( l + 1 ) = i = k + 1 n r i c ( l + 1 ) ( Y i X i T β c + ( l ) ) ( Y i X i T β c + ( l ) ) T i = k + 1 n r i c ( l + 1 ) , c = 1 , 2 , , C
Let the initial value be  π c + ( 0 ) , β c + ( 0 ) , σ c + ( 0 ) , c = 1 , , C , repeat the above iteration algorithm, and when the iteration converges, we obtain the estimator of mixture parameters, denoted as π c + ( k ) , β c + ( k ) , σ c 2 + ( k ) , c = 1 , , C . At this time, the log-likelihood function corresponding to the data { ( X i , Y i ) } i = k + 1 n is
l + ( k , π c + , β c + , σ c 2 + ) = i = k + 1 n log c = 1 C π c + ( k ) ϕ { Y i | X i T β c + ( k ) , σ c 2 + ( k ) } .
Whenever a value of k is determined, the mixture data are divided into two parts:  { ( X i , Y i ) } i = 1 k  and  { ( X i , Y i ) } i = k + 1 n . Let the initial value be  π c ± ( 0 ) , β c ± ( 0 ) , σ c ± ( 0 ) , c = 1 , , C , and we can obtain the estimations of change point and mixture parameter according to Algorithm 1:
( k ^ , θ ^ , θ + ^ ) = arg max k , π c ± , β c ± , σ c ± l ( k , π c , β c , σ c 2 , π c + , β c + , σ c 2 + ) ,
where  θ ^ ± = ( ( π ^ ± ) T , ( β ^ ± ) T , ( σ ^ 2 ± ) T ) T , π ^ ± = ( π ^ 1 ± , , π ^ C ± ) T and β ^ ± = ( β ^ 1 ± , , β ^ C ± ) T , σ ^ 2 ± = ( σ ^ 1 2 ± , , σ ^ C 2 ± ) T , l ( k , π c , β c , σ c 2 , π c + , β c + , σ c 2 + ) = l + ( k , π c + , β c + , σ c 2 + ) + l ( k , π c , β c , σ c 2 ) .
Algorithm 1 EM algorithm considering change point
Intput:  { X i , Y i } i = 1 n , the number of mixture component C, the maximum iterations T.
Output:  k ^ , θ ^ , θ ^ +
  1:
for  k = 1 , 2 , , n 1  do
  2:
     For any given k, the mixture data are divided into two parts:  { ( X i , Y i ) } i = 1 k  and  { ( X i , Y i ) } i = k + 1 n . Using EM algorithm separately for two parts of data. For  { ( X i , Y i ) } i = 1 k , let the initial value be π c ( 0 ) , β c ( 0 ) , σ c ( 0 ) , c = 1 , , C .
  3:
     for t = 1 , 2 , , T do
  4:
          E-step: calculate  r i c ( l + 1 ) ( k ) ,
r i c ( l + 1 ) ( k ) = π c ( l ) ϕ { Y i | X i T β c ( l ) , σ c 2 ( l ) } c = 1 C π c ( l ) ϕ { Y i | X i T β c ( l ) , σ c 2 ( l ) } , i = 1 , , k , c = 1 , , C
  5:
          M-step: calculate
π c ( l + 1 ) ( k ) = i = 1 n r i c ( l + 1 ) n , β c ( l + 1 ) ( k ) = ( i = 1 n X i r i c ( l + 1 ) X i T ) 1 ( i = 1 n X i r i c ( l + 1 ) Y i )
σ c 2 ( l + 1 ) ( k ) = i = 1 n r i c ( l + 1 ) ( Y i X i T β c ( l ) ) ( Y i X i T β c ( l ) ) T i = 1 n r i c ( l + 1 )
  6:
     end for
  7:
     When the iteration reaches convergence, we obtain the estimator π c ( k ) , β c ( k ) , σ c 2 ( k ) , c = 1 , , C and the log-likelihood function value l ( k )
  8:
     For { ( X i , Y i ) } i = k + 1 n , repeat steps 3–7 and obtain π c + ( k ) , β c + ( k ) , σ c 2 + ( k ) , c = 1 , , C  and  l + ( k )  
  9:
end for
10:
k ^ = arg max 1 k n 1 ( l ( k ) + l + ( k ) ) , θ ^ = θ ^ ( k ^ ) , θ ^ + = θ + ^ ( k ^ ) ,

3.2. Hypothesis Test

Define the following notations for later use:
η ( Y | θ ) = c = 1 C π c ϕ { X , Y | β c , σ c 2 } , ( θ , Y ) = log η ( Y | θ ) ,
q 1 { θ , Y } = { θ , Y } θ , q 2 { θ , Y } = 2 { θ , Y } θ θ T ,
I ( θ ) = E [ q 2 { θ , Y } ] , B ( θ ) = V a r [ q 1 { θ , Y } ] = E [ q 1 { θ , Y } q 1 T { θ , Y } ] .
In the following, we are interested in testing the null hypothesis:
H 0 : θ i = θ ( i = 1 , , n )
against the local alternative hypothesis:
H A : θ i = θ + 1 n · g ( i n ) ,
where g = ( g 1 , g 2 , , g q ) T is a function of bounded variation on [0,1] which describes the patter of departure from stability of the parameter θ .
Firstly, we give the following lemmas and then we build an empirical process and infer its limiting process under the H 0 and H 1 hypothesis, respectively, as described by Theorem 1.
Theorem 1.
Under H 0 and Lemma A3 in the Appendix A, given by
Δ n ( t , θ ^ n ) = 1 n i = 1 [ n t ] q 1 ( θ ^ n , Y i ) , t ( 0 , 1 ) ,
then, we have
(1). under H 0 ,
B ^ n 1 / 2 Δ n ( t , θ ^ n ) d W 0 ( t ) , n ,
where W 0 ( t ) is a q-dimensional standard Brownian bridge with W 0 ( t ) = W ( t ) t W ( 1 ) and B ^ n some consistent covariance matrix estimation, such as B ^ n = 1 n i = 1 n q 1 ( θ ^ n , Y i ) q 1 ( θ ^ n , Y i ) T .
(2). Under H A ,
B ^ n 1 / 2 Δ n ( t , θ ^ n ) d W 0 ( t ) + B ( θ ) 1 / 2 G 0 ( t ) , n ,
where G 0 ( t ) = G ( t ) t G ( 1 ) with G ( t ) = 0 t g ( y ) d y .
Next, we construct test statistics according to the empirical process (5) stated above. The key idea is that the empirical process reflects some symptoms of structural changes, and its performance under H 0 and H A is significantly different; then, the test statistics can be constructed as follows:
T n = sup 1 k n B ^ n 1 / 2 Δ n ( k / n , θ ^ n ) 2 .
Further from Theorem 1 and continuous mapping theorem, the following corollary can be obtained:
Corollary 1.
Given the empirical process (5) in Theorem 1, we have
(1). under H 0 ,
T n d sup t [ 0 , 1 ] W 0 ( t ) 2 = sup t [ 0 , 1 ] i = 1 q W i 0 ( t ) 2 , n .
(2). Under H A ,
T n d sup t [ 0 , 1 ] W 0 ( t ) + B ( θ ) 1 / 2 G 0 ( t ) 2 , n
The above results show that the limiting distributions of the test statistics are different under the null and alternative hypothesis; thus, we can judge the existence of the change point according to it.

3.3. Consistency

In order to prove the consistency of the estimator, we first construct a function L n as follows:
L n ( τ , θ ) = 1 n i = 1 n f ˜ ( X i , Y i | t , θ ) = 1 n i = 1 [ n τ ] f ˜ ( X i , Y i | t , θ ) + 1 n i = [ n τ ] + 1 n f ˜ ( X i , Y i | t , θ ) ,
where [ n τ ] = k , [ a ] means the integer part of a and
f ˜ ( X i , Y i | t , θ ) = log [ c = 1 C { π c ϕ ( Y i | X i T β c , σ c 2 ) I { i [ n t ] } + π c + ϕ ( Y i | X i T β c + , σ c 2 + ) I { i [ n t ] + 1 } } ] .
Without loss of generality, we only consider the case k k for brevity; then, by K o l m o g o r o v ’s Law of Large Numbers and J e s s e n inequality, the Formula (7) can be rewritten as
L n ( τ , θ ) = 1 n i = 1 [ n τ ] f ( X i , Y i | τ , θ ) + 1 n i = [ n τ ] + 1 [ n τ ] f ( X i , Y i | τ , θ ) + 1 n i = [ n τ ] + 1 n f ( X i , Y i | τ , θ + ) = [ n τ ] n · 1 [ n τ ] i = 1 [ n τ ] f ( X i , Y i | τ , θ ) + [ n τ ] [ n τ ] n · 1 [ n τ ] [ n τ ] i = [ n τ ] + 1 [ n τ ] f ( X i , Y i | τ , θ ) + n [ n τ ] n · 1 n [ n τ ] i = [ n τ ] + 1 n f ( X i , Y i | τ , θ + ) p τ · E ( τ , θ ) f ( X , Y | τ , θ ) + ( τ τ ) · E ( τ , θ ) f ( X , Y | τ , θ ) + ( 1 τ ) · E ( τ , θ + ) f ( X , Y | τ , θ + ) L ( τ , θ | τ , θ )
where
f ( X i , Y i | τ , θ ) = f ˜ ( X i , Y i | τ , θ ) | θ = θ , f ( X i , Y i | τ , θ + ) = f ˜ ( X i , Y i | τ , θ ) | θ = θ + .
Now, we state our second theoretical result, which is the consistency of the obtained estimator.
Theorem 2.
Let τ ^ = k ^ n n and under the conditions of the Lemmas A1 and A2 in Appendix A; then, we have
( τ ^ , θ ^ n ) p ( τ , θ ) .

4. Simulation

To assess the performance of our proposed method, we shall consider two simulation experiments as follows.

4.1. Estimation Experiment

In Experiment 1, we focus on the scenario where the vector β is two-dimensional and the value of C is set to 2. Additionally, when dealing with the parameter estimation problem of mixed models that incorporate change points, the conventional approach is to overlook the presence of these change points. Instead, the Expectation–Maximization (EM) algorithm is directly employed to estimate the unknown parameters, as if the data were generated from a mixed model consisting of 2 C components. Consequently, we conduct a comparison between the estimation results obtained by the EM method that does not take change points into account (referred to as “NCPEM”) and those derived from our improved EM method (“CPEM”). We utilize the data generation mechanism of Model (1) to generate the data. Subsequently, we construct a two-component mixed model with a change point and proceed to specify the parameters as follows:
( β 1 ) T = ( 1 , 2 ) , σ 1 2 = 0.1 , π 1 = 0.4 , ( β 2 ) T = ( 1 , 3 ) , σ 2 2 = 0.2 , π 2 = 0.6 . ( β 1 + ) T = ( 3 , 4 ) , σ 1 2 + = 0.2 , π 1 + = 0.7 , ( β 2 + ) T = ( 3 , 5 ) , σ 2 2 + = 0.2 , π 2 + = 0.3 .
X i T = ( 1 , X i ) and X i U ( 0 , 1 ) for i = 1 , 2 , , n . Let the total sample size n from { 500 , 1000 , 2000 } and the change point k be { 250 , 500 , 1000 } , respectively. The change point estimators and the corresponding standard errors are k ^ = 250.45 ( 0.730 ) , 500.18 ( 0.458 ) ,   1000.5 ( 0.759 ) over 100 simulations. It can be seen that the change point estimator is close to the true value. The mixture parameter estimators and the corresponding standard errors are presented in Table 1, Table 2 and Table 3.
As is evident from Table 1, the estimation results obtained by the Change Point Expectation–Maximization (CPEM) method are much closer to the true parameter values compared to those derived from the Non-Change Point Expectation–Maximization (NCPEM) method. Moreover, the CPEM method yields a significantly smaller standard error. Based on the estimations of σ 2 and π presented in Table 2 and Table 3, it is clear that the estimation results of the NCPEM method deviate substantially from the true parameter values. Additionally, as the sample size increases, the majority of the estimation results tend to converge more closely to the true parameter values, accompanied by a smaller standard error. These findings strongly indicate that our improved CPEM method exhibits superior accuracy and enhanced robustness.

4.2. Detection Experiment

In experiment 2, we explore the performance of the detection procedure. Let p = 1 and consider the following mixture model:
Y i c = 1 2 π c N μ c , σ c 2 , 1 i k , c = 1 2 π c + N μ c + , σ c 2 + , k + 1 i n , k [ 1 , n ] ,
where
μ 1 = 1 , σ 1 2 = 0.1 , π 1 = 0.6 , μ 2 = 2 , σ 2 2 = 0.2 , π 2 = 0.4 ,
μ 1 + = 3 , σ 1 2 + = 0.2 , π 1 + = 0.3 , μ 2 + = 4 , σ 2 2 + = 0.3 , π 2 + = 0.7 .
The limiting distribution function in (6) is well known and tabulated by Kiefer [23]. In model (8), q = 5 , at the significance level of 0.05, the critical value is 2.0005. We set the values of n to be 500, 1000, 2000, k to be n / 4 , n / 2 , and 3 n / 4 to represent the case where the change point is located in the front, middle, and back end of the data. The empirical sizes and powers are shown in Table 4 and Table 5 under 100 simulations.
Based on the analysis of the data in the tables, the performance of the method proposed in this paper under large-sample scenarios can be summarized as follows:
(1) Empirical Level Performance: As clearly shown in Table 4, with the continuous increase in the sample size n, the empirical level gradually approaches the significance level of 0.05. This indicates that, under large sample conditions, the actual test level of this method is in good agreement with the preset theoretical level.
(2) Empirical Power Performance: According to the data results in Table 5, there is a significant positive correlation between the increase in sample size and the improvement in empirical power. That is, the larger the sample size, the closer the value of empirical power is to 1, suggesting that the method can more effectively detect true change point situations in large samples.
(3) Influence of Change Point Location: The position of the change point in the data sequence has an impact on the detection effect. When the change point is located in the middle or front part of the data, the detection effect of this method is significantly better than when the change point is located at the end of the data.
(4) Comprehensive Performance Advantages: In large sample scenarios, the method proposed in this paper demonstrates good empirical size and power.

5. Concluding Remarks

This paper proposes an improved EM method to solve the parameter estimation problem with a mixture of the linear regression model and change point. This method is more accurate and robust against usual EM algorithms. Furthermore, a detection method based on score function is obtained to test the change point in mixture data. Simulations demonstrate the effectiveness of the proposed methodology.

6. Limitations and Future Research Directions

The method presented in this article also has certain limitations and challenges. Firstly, there is the issue of robustness in complex noisy scenarios. The method proposed in this article performs well in dealing with normal noise, but the detection accuracy remains to be studied under complex noise interference. The algorithm mentioned in the article still needs further research in dealing with high-dimensional data point detection and estimation problems. Finally, the computational complexity of the method proposed in this article is relatively high, and it is worth studying whether there is a simpler and faster processing method. The above issues are also possible research directions for the future.

Author Contributions

Methodology, Z.X.; Software, W.Z.; Writing—original draft, T.C.; Writing—review & editing, W.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (Grant No. 12171391) and the Natural Science Basic Research Program of Shaanxi (Grant No. 2024JC-YBQN-0039), Shaanxi Fundamental Science Research Project for Mathematics and Physics (Grant No. 23JSY043).

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Appendix for Proof

Lemma A1.
Under the Assumption 3 and C r a m e ´ r-Rao regular conditions, we have
E [ q 1 ( θ , Y ) ] = 0 ,
E [ q 2 ( θ , Y ) ] = E [ q 1 ( θ , Y ) q 1 T ( θ , Y ) ] .
Proof. 
See McLachlan and Peel [24]. □
Lemma A2.
For the process given by
Δ n ( t , θ ) = 1 n i = 1 [ n t ] q 1 ( θ , Y i ) , t ( 0 , 1 ) .
and under the conditions of Lemma A1 and under H 0 the following functional central limit theorem holds:
Δ n ( · , θ ) d Z ( · ) ,
where Z ( · ) is the Gaussian process whose mean is E [ Z ( t ) ] = 0 and variance function is V a r [ Z ( t ) ] = B ( θ ) .
Furthermore, if B ( θ ) is reversible, then we have
B ( θ ) 1 / 2 Δ n ( · , θ ) d W ( · ) ,
where W ( · ) is standard Brownian motion.
Proof. 
The proof follows by direct application of Donsker’s theorem (Billingsley [25]). □
Usually in applications, the parameters θ under the null hypothesis are not known but have to be estimated. Let θ ^ n is the maximum likelihood estimator of θ ; therefore,
i = 1 n q 1 ( θ ^ n , Y i ) = 0 .
Taylor expansion of Γ n ( θ ) = 1 n i = 1 n q 1 ( θ , Y i ) at θ when θ = θ ^ n ; then,
0 = Γ n ( θ ^ n ) = Γ n ( θ ) + Γ n ( θ ) ( θ ^ n θ ) + R n .
Under suitable regularity conditions
Γ n ( θ ) = 1 n i = 1 n q 2 ( θ , Y i ) p I ( θ ) , n Γ n ( θ ) d N ( 0 , B ( θ ) ) , n R n p 0 .
Therefore, the following holds:
n ( θ ^ n θ ) d N ( 0 , V ( θ ) ) ,
where V ( θ ) = I ( θ ) 1 B ( θ ) { I ( θ ) 1 } T . Equivalently, we can write
n ( θ ^ n θ ) I ( θ ) 1 Δ n ( 1 , θ ) .
Lemma A3.
Under H 0 , the following holds:
Δ n ( · , θ ^ n ) d Z 0 ( · ) ,
where Z 0 ( t ) = Z ( t ) t Z ( 1 ) .
Proof. 
Taylor expansion of Δ n ( t , θ ^ n ) at θ ,
Δ n ( t , θ ^ n ) 1 n i = 1 [ n t ] q 1 ( θ , Y i ) + 1 n i = 1 [ n t ] q 2 ( θ , Y i ) · n ( θ ^ n θ ) Δ n ( t , θ ) [ n t ] n I ( θ ) · I ( θ ) 1 Δ n ( 1 , θ ) d Z ( t ) t Z ( 1 ) .
Next, we will provide the proof process for Theorem 1.
Proof. 
(1) On the basis of the three lemmas stated above and if B ( θ ) is reversible, we can obtain the first conclusion in Theorem 1:
B ^ n 1 / 2 Δ n ( · , θ ^ n ) d W 0 ( · ) ,
where W 0 ( τ ) = W ( τ ) τ W ( 1 ) is a standard Brownian bridge, B ^ n is the variance estimation, such as B ^ n = 1 n i = 1 n q 1 ( θ ^ n , Y i ) q 1 ( θ ^ n , Y i ) T .
(2) Next, we prove the second conclusion. Y i has the probability density function under H A hypothesis:
η ( Y | θ i ) η ( Y | θ ) + η ( Y | θ ) θ | θ = θ 0 = η ( Y | θ ) + log η ( Y | θ ) θ · η ( Y | θ ) | θ = θ 0 · 1 n · I { i [ n τ ] + 1 } = η ( Y | θ ) { 1 + q 1 ( θ , Y ) T · 1 n · I { i [ n τ ] + 1 } } ,
which can be derived from the Taylor expansion of η . Therefore, under the alternative hypothesis, the process (A1) no longer has zero mean in general but
E [ q 1 ( θ i , Y ) ] q 1 ( θ η ( y | θ ) d y + q 1 ( θ , y ) q 1 ( θ , y ) T η ( y | θ ) 1 n · I { i [ n t ] + 1 } d y = 0 + 1 n B ( θ ) · I { i [ n t ] + 1 } .
Similar to the Lemmas A2 and A3, we can deduce
Δ n ( · , θ ) d W A ( · ) ,
where W A ( t ) = W ( t ) + B ( θ ) G ( t ) , ( G ( t ) = 0 t · I { i n t + 1 n } d ( i n ) = t ( 1 t ) + o ( 1 ) ) .
Finally, the following limiting process can be derived, which is our second conclusion:
B ^ n 1 / 2 Δ n ( τ , θ ^ n ) B ^ n 1 / 2 { Δ n ( τ , θ ) τ Δ n ( 1 , θ ) } W 0 ( τ ) + B ^ ( θ ^ n ) 1 / 2 G 0 ( τ ) d W 0 ( t ) + B ( θ ) 1 / 2 G 0 ( t ) .
Next, we prove Theorem 2 and first give two useful lemmas.
Lemma A4.
Under the conditions of model (8) and Assumptions 1 and 2, L n uniformly converges to L with probability.
Proof. 
According to the structure of the function L , we can show that L n converges to L point by point. Next, we prove uniform convergence.
Due to the fact that the parameter space Θ is a compact set, i , δ i > 0 , Θ i = 1 m U ( ( τ i , θ i ) , δ i ) , m R , where U ( ( τ i , θ i ) , δ i ) is the neighborhood with center ( τ i , θ i ) and radius δ i . According to the continuity of L n , ϵ > 0 ,
| ( τ i , θ i ) ( τ , θ ) | < δ i | L n ( τ i , θ i ) L n ( τ , θ ) | < ϵ , ( τ , θ ) U ( ( τ i , θ i ) , δ i ) .
Consequently,
| L n ( τ , θ ) L ( τ , θ ) | = | L n ( τ , θ ) L n ( τ i , θ i ) + L n ( τ i , θ i ) L ( τ i , θ i ) + L ( τ i , θ i ) L ( τ , θ ) | | L n ( τ , θ ) L n ( τ i , θ i ) | + | L n ( τ i , θ i ) L ( τ i , θ i ) | + | L ( τ i , θ i ) L ( τ , θ ) | < 3 ϵ .
Lemma A5.
Under the condition of model (8) and Assumptions 1 and 3, the parameters are identifiable.
Proof. 
From the J e s s e n inequality,
L ( τ , θ | τ , θ ) L ( τ , θ | τ , θ ) = τ · E ( τ , θ ) f ( X , Y | τ , θ ) + ( τ τ ) · E ( τ , θ ) f ( X , Y | τ , θ + ) + ( 1 τ ) · E ( τ , θ + ) f ( X , Y | τ , θ + ) ( τ · E ( τ , θ ) f ( X , Y | τ , θ ) + ( τ τ ) · E ( τ , θ ) f ( X , Y | τ , θ ) + ( 1 τ ) · E ( τ , θ + ) f ( X , Y | τ , θ + ) ) = τ · [ E ( τ , θ ) f ( X , Y | τ , θ ) E ( τ , θ ) f ( X , Y | τ , θ ) ] + ( τ τ ) [ E ( τ , θ ) f ( X , Y | τ , θ + ) E ( τ , θ ) f ( X , Y | τ , θ ) ] + ( 1 τ ) · [ E ( τ , θ + ) f ( X , Y | τ , θ + ) E ( τ , θ + ) f ( X , Y | τ , θ + ) ] = τ · E ( τ , θ ) [ log g 1 ( X , Y | θ , θ ) ] + ( τ τ ) · E ( τ , θ ) [ log g 2 ( X , Y | θ + , θ ) ] + ( 1 τ ) · E ( τ , θ + ) [ log g 3 ( X , Y | θ + , θ + ) ] τ · log { E ( τ , θ ) [ g 1 ( X , Y | θ , θ ) ] } + ( τ τ ) · log { E ( τ , θ ) [ g 2 ( X , Y | θ + , θ ) ] } + ( 1 τ ) · log { E ( τ , θ + ) [ g 3 ( X , Y | θ + , θ + ) ] } = 0 ,
where
g 1 ( X , Y | θ , θ ) = c = 1 C π c 2 π σ c 2 exp { ( Y X T β c ) 2 2 σ c 2 } c = 1 C π c 2 π σ c 2 exp { ( Y X T β c ) 2 2 σ c 2 } , g 2 ( X , Y | θ + , θ ) = c = 1 C π c 2 π σ c 2 + exp { ( Y X T β c + ) 2 2 σ c 2 + } c = 1 C π c 2 π σ c 2 exp { ( Y X T β c ) 2 2 σ c 2 } , g 3 ( X , Y | θ + , θ + ) = c = 1 C π c 2 π σ c 2 + exp { ( Y X T β c + ) 2 2 σ c 2 + } c = 1 C π c 2 π σ c 2 + exp { ( Y X T β c + ) 2 2 σ c 2 + } .
Therefore, we have L ( τ , θ | τ , θ ) L ( τ , θ | τ , θ ) , and the equal sign holds only when ( τ , θ ) = ( τ , θ ) ; that is, L maximizes at ( τ , θ ) . Next, we prove that the maximum point is unique.
According to J e s s e n inequality, the three terms on the left side of the inequality are all less than or equal to 0. Therefore, the left side is equal to 0 if and only if the three terms are equal to 0. Since τ , τ ( 0 , 1 ) , both g 1 ( X , Y | θ , θ ) and g 3 ( X , Y | θ + , θ + ) are equal to 1 if and only if ( π c , β c , σ c 2 ) = ( π c , β c , σ c 2 ) and ( π c , β c + , σ c 2 + ) = ( π c , β c + , σ c 2 + ) ; meanwhile, because of β c + β c , σ c 2 + σ c 2 , g 2 ( X , Y | θ + , θ ) is equal to 1 if and only if τ = τ . Therefore, L ( τ , θ | τ , θ ) obtains the unique maximum at ( τ , θ ) . □
The proof of Theorem 2 is as follows.
Proof. 
For any ϵ > 0 , let
A = { L n ( τ , θ ) > L ( τ , θ ) ϵ / 3 } , B = { L n ( τ ^ , θ ^ n ) > L n ( τ , θ ) ϵ / 3 } , C = { L ( τ ^ , θ ^ n ) > L n ( τ ^ , θ ^ n ) ϵ / 3 } .
From Lemmas A4 and A5, L n uniformly converges to L ; furthermore, ( τ ^ , θ ^ n ) and ( τ , θ ) are the maximum points of L n and L , respectively; therefore, when n , P ( A ) , P ( B ) , P ( C ) all tend to 1. So when the event A B C occurs,
L ( τ , θ ) L ( τ ^ , θ ^ n ) = ( L ( τ , θ ) L n ( τ , θ ) ) + ( L n ( τ , θ ) L n ( τ ^ , θ ^ n ) ) + ( L n ( τ ^ , θ ^ n ) L ( τ ^ , θ ^ n ) ) < ϵ 3 + ϵ 3 + ϵ 3 = ϵ .
Furthermore, the parameter space Θ is a compact set and L is continuous, so for a sufficiently small δ , let N = U ( ( τ , θ ) , δ ) , we have
sup ( τ , θ ) Θ N c L ( τ , θ ) L ( τ 0 , θ 0 ) < L ( τ , θ ) .
We might as well take ϵ = L ( τ , θ ) L ( τ 0 , θ 0 ) > 0 , then ϵ < L ( τ , θ ) L ( τ ^ , θ ^ n ) < ϵ holds with probability 1.
Therefore, the following holds:
P ( | L ( τ , θ ) L ( τ ^ , θ ^ n ) | < ϵ ) 1 .
In other words, when ( τ ^ , θ ^ n ) N c and | ( τ , θ ) ( τ ^ , θ ^ n ) | δ , then | L ( τ , θ ) L ( τ ^ , θ ^ n ) | ϵ . It can be shown that
P ( | ( τ , θ ) ( τ ^ , θ ^ n ) | δ ) P ( | L ( τ , θ ) L ( τ ^ , θ ^ n ) | ϵ ) 0 .
i.e., P ( | ( τ , θ ) ( τ ^ , θ ^ n ) | δ ) 0 . Therefore, we obtain the desired result:
( τ ^ , θ ^ n ) p ( τ , θ ) .

References

  1. Goldfeld, S.M.; Quandt, R.E. A Markov model for switching regressions. J. Econom. 1973, 1, 3–15. [Google Scholar] [CrossRef]
  2. Frühwirth-Schnatter, S. Finite Mixture and Markov Switching Models; Springer: New York, NY, USA, 2006; pp. 241–275. [Google Scholar]
  3. Hurn, M.; Justel, A.; Robert, C.P. Estimating Mixtures of Regressions. J. Comput. Graph. Stat. 2003, 12, 55–79. [Google Scholar] [CrossRef]
  4. Li, P.; Chen, J. Testing the Order of a Finite Mixture. J. Am. Stat. Assoc. 2010, 105, 1084–1092. [Google Scholar] [CrossRef]
  5. Chen, J.; Li, P.; Fu, Y. Inference on the Order of a Normal Mixture. J. Am. Stat. Assoc. 2012, 107, 1096–1105. [Google Scholar] [CrossRef]
  6. Huang, M.; Yao, W. Mixture of Regression Models With Varying Mixing Proportions: A Semiparametric Approach. J. Am. Stat. Assoc. 2012, 107, 711–724. [Google Scholar] [CrossRef]
  7. Huang, M.; Li, R.; Wang, S. Nonparametric Mixture of Regression Models. J. Am. Stat. Assoc. 2013, 108, 929–941. [Google Scholar] [CrossRef]
  8. Page, E.S. Continuous inspection schemes. Biometrika 1954, 41, 100–115. [Google Scholar] [CrossRef]
  9. Sen, A.; Srivastava, M.S. On tests for detecting change in mean. Ann. Stat. 1975, 3, 98–108. [Google Scholar] [CrossRef]
  10. Sen, A.; Srivastava, M.S. Some one-sided tests for change in level. Technometrics 1975, 17, 61–64. [Google Scholar] [CrossRef]
  11. Hawkins, D.M. Testing a sequence of observations for a shift in location. J. Am. Stat. Assoc. 1977, 72, 180–186. [Google Scholar] [CrossRef]
  12. James, B.; James, K.; Siegmund, D. Tests for a change-point. Biometrika 1987, 74, 71–83. [Google Scholar] [CrossRef]
  13. Srivastava, M.S.; Worsley, K.J. Likelihood ratio tests for a change in the multivariate normal mean. J. Am. Stat. Assoc. 1986, 81, 199–204. [Google Scholar] [CrossRef]
  14. Bassevile, M.; Nikiforov, I. Detection of Abrupt Changes: Theory and Applications; Prentice Hall: Hoboken, NJ, USA, 1993. [Google Scholar]
  15. Wang, G.; Zou, C.; Yin, G. Change-point detection in multinomial data with a large number of categories. Ann. Stat. 2018, 46, 2020–2044. [Google Scholar] [CrossRef]
  16. Xia, Z.; Qiu, P. Jump information criterion for statistical inference in estimating discontinuous curves. Biometrika 2015, 102, 397–408. [Google Scholar] [CrossRef]
  17. Bai, J. Least squares estimation of a shift in linear processes. J. Time Ser. Anal. 1994, 15, 453–472. [Google Scholar] [CrossRef]
  18. Baranowski, R.; Chen, Y.; Fryzlewicz, P. Narrowest-over-threshold detection of multiple change points and change-point-like features. J. R. Stat. Soc. Ser. Stat. Methodol. 2019, 81, 649–672. [Google Scholar] [CrossRef]
  19. Follain, B.; Wang, T.; Samworth, R.J. High-dimensional changepoint estimation with heterogeneous missingness. J. R. Stat. Soc. Ser. Stat. Methodol. 2022, 84, 1023–1055. [Google Scholar] [CrossRef]
  20. Drabech, Z.; Douimi, M.; Zemmouri, E. A Markov random field model for change points detection. J. Comput. Sci. 2024, 83, 102429. [Google Scholar] [CrossRef]
  21. Ratnasingam, S.; Gamage, R. Empirical likelihood change point detection in quantile regression models. Comput. Stat. 2025, 40, 999–1020. [Google Scholar] [CrossRef]
  22. Gong, Y.; Dong, X.; Zhang, J.; Chen, M. Latent evolution model for change point detection in time-varying networks. Inf. Sci. 2023, 646, 119376. [Google Scholar] [CrossRef]
  23. Kiefer, J. K-sample analogues of the Kolmogorov-Smirnov and Cramér-Von Mises tests. Ann. Math. Stat. 1959, 30, 420–447. [Google Scholar] [CrossRef]
  24. McLachlan, G.; Peel, D. Finite Mixture Models; Wiley: New York, NY, USA, 2000. [Google Scholar]
  25. Billingsley, P. Convergence of Probability Measures, 2nd ed.; Wiley: New York, NY, USA, 1999. [Google Scholar]
Table 1. Estimated value and standard error of β .
Table 1. Estimated value and standard error of β .
β ( β 1 ) T = ( 1 , 2 ) ( β 2 ) T = ( 1 , 3 ) ( β 1 + ) T = ( 3 , 4 ) ( β 2 + ) T = ( 3 , 5 )
n = 500CPEM(0.994, 2.004)(0.975, 3.023)(2.993, 4.004)(2.999, 4.999)
(0.045, 0.029)(0.150, 0.129)(0.064, 0.041)(0.085, 0.057)
NCPEM(2.235, 1.911)(0.988, 3.048)(3.086, 3.874)(3.022, 4.941)
(9.906, 1.624)(7.933, 1.290)(9.285, 1.609)(0.511, 0.278)
n = 1000CPEM(0.982, 2.054)(1.006, 2.993)(2.992, 4.005)(2.996, 5.001)
(0.042, 0.017)(0.104, 0.077)(0.039, 0.024)(0.058, 0.038)
NCPEM(0.452, 2.220)(0.884, 3.044)(1.550, 4.192)(3.187, 4.939)
(8.490, 1.664)(3.710, 0.678)(8.211, 1.367)(2.540, 0.484)
n = 2000CPEM(1.000, 1.999)(1.014, 2.995)(2.930, 4.017)(3.001, 4.997)
(0.022, 0.014)(0.069, 0.046)(0.030, 0.019)(0.041, 0.035)
NCPEM(1.225, 2.054)(1.318, 2.975)(2.880, 3.921)(3.009, 4.975)
(4.212, 0.624)(2.383, 0.505)(0.557, 0.342)(0.182, 0.107)
Table 2. Estimated value and standard error of σ 2 .
Table 2. Estimated value and standard error of σ 2 .
  σ 2 σ 1 2 = 0.1 σ 2 2 = 0.2 σ 1 2 + = 0.1 σ 2 2 + = 0.2
n = 500CPEM0.099 (0.006)0.199 (0.012)0.100 (0.008)0.197 (0.011)
NCPEM0.193 (0.466)0.346 (0.626)0.171 (0.436)0.441 (0.786)
n = 1000CPEM0.098 (0.011)0.200 (0.005)0.099 (0.006)0.199 (0.008)
NCPEM0.143 (0.287)0.375 (0.686)0.147 (0.338)0.376 (0.661)
n = 2000CPEM0.100 (0.003)0.203 (0.015)0.099 (0.011)0.204 (0.005)
NCPEM0.153 (0.339)0.333 (0.616)0.096 (0.022)0.376 (0.687)
Table 3. Estimated value and standard error of π .
Table 3. Estimated value and standard error of π .
  π π 1 = 0.6 π 2 = 0.4 π 1 + = 0.3 π 2 + = 0.7
n = 500CPEM0.601 (0.031)0.399 (0.031)0.301 (0.030)0.699 (0.030)
NCPEM0.182 (0.053)0.310 (0.050)0.189 (0.045)0.320 (0.062)
n = 1000CPEM0.594 (0.063)0.406 (0.063)0.295 (0.019)0.705 (0.019)
NCPEM0.189 (0.060)0.310 (0.044)0.193 (0.039)0.308 (0.056)
n = 2000CPEM0.599 (0.014)0.401 (0.013)0.298 (0.033)0.702 (0.033)
NCPEM0.194 (0.035)0.311 (0.058)0.186 (0.049)0.310 (0.045)
Table 4. Empirical sizes.
Table 4. Empirical sizes.
n50010002000
size0.070.060.05
Table 5. Empirical powers.
Table 5. Empirical powers.
k n 4 n 2 3 n 4
n
5000.600.830.59
10000.960.880.67
200010.980.94
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, W.; Cheng, T.; Xia, Z. Change-Point Estimation and Detection for Mixture of Linear Regression Models. Axioms 2025, 14, 402. https://doi.org/10.3390/axioms14060402

AMA Style

Zhao W, Cheng T, Xia Z. Change-Point Estimation and Detection for Mixture of Linear Regression Models. Axioms. 2025; 14(6):402. https://doi.org/10.3390/axioms14060402

Chicago/Turabian Style

Zhao, Wenzhi, Tian Cheng, and Zhiming Xia. 2025. "Change-Point Estimation and Detection for Mixture of Linear Regression Models" Axioms 14, no. 6: 402. https://doi.org/10.3390/axioms14060402

APA Style

Zhao, W., Cheng, T., & Xia, Z. (2025). Change-Point Estimation and Detection for Mixture of Linear Regression Models. Axioms, 14(6), 402. https://doi.org/10.3390/axioms14060402

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop