Next Article in Journal
Semiparametric Estimation and Application of Realized GARCH Model with Time-Varying Leverage Effect
Previous Article in Journal
Optimizing Fuel Economy in Hybrid Electric Vehicles Using the Equivalent Consumption Minimization Strategy Based on the Arithmetic Optimization Algorithm
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Incorporating Prior Information in Latent Structures Identification for Panel Data Models

1
College of Tourism, Hunan Normal University, Changsha 410081, China
2
School of Management, Chongqing University of Science and Technology, Chongqing 401331, China
3
School of Economics, Xiamen University, Xiamen 361005, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(9), 1505; https://doi.org/10.3390/math13091505
Submission received: 14 March 2025 / Revised: 22 April 2025 / Accepted: 29 April 2025 / Published: 2 May 2025

Abstract

:
In this paper, we explore the latent structures for panel data models in presence of available prior information. The latent structure in panel models allows individuals to be classified into several distinct groups, where the individuals within the same group share the same slope parameters, while the group-specific parameters are heterogeneous. To incorporate the prior information, we design a new alternating direction method of multipliers (ADMM) algorithm based on the pairwise group fused Lasso penalty approach. The asymptotic properties and the convergence of ADMM algorithm are well established. Simulation studies demonstrate the advantages of the proposed method over existing methods in terms of both estimation efficiency and detection accuracy. We illustrate the practical utility of the proposed procedure by analyzing the relationship between electricity consumption and GDP in China.

1. Introduction

Panel data models are widely encountered in many substantial areas in both econometric and statistics. When conducting analyses for panel data, conventional approaches typically assume that the slopes of regressors are completely homogeneous across all individuals [1]. However, such homogeneity assumptions are not in accordance with reality through many practical analyses. See, for example,  [2,3,4], to name a few. From the perspective of model strategy, the complete homogeneity assumption ignores individual-specific heterogeneity, which is quite important in panel data analysis. On the other hand, estimating parameters for each individual separately is feasible for achieving complete heterogeneity, but the estimators and relevant inferences are inefficient or even imprecise, as the time dimension T is usually very short. To strike a balance between completely homogeneous and fully heterogeneous models, a natural way is to assume that the panel units can be classified into several unknown groups, where the members within a group share the same slope parameters, while the slopes across different groups are different. In this way, the convergence rate of estimators changes from T to N k T , where N k is the total number of individuals in the kth group. The group structure here is called latent structure [5], and the identification of latent structures is sometimes called homogeneity structure learning [6,7] or subgroup identification [8].

1.1. Literature Review

The existing literature on exploring the latent group structure can be roughly classified into three categories based on different perspectives. First, researchers have found that identifying the latent group can be formulated into a penalization estimation context, such that modern feature selection techniques [9,10] can be employed. A pivotal contribution in this area can be traced back to [5], who developed a novel variant of the Lasso estimator called Classifier-Lasso (C-Lasso). The penalty form of C-Lasso takes an additive-multiplicative form, which forces the parameters of individuals into several groups for a given number of groups. This method had subsequently been extended in various panel models, such as [11,12,13,14]. Recently, ref. [15] proposed a new penalized estimation procedure based on the pairwise adaptive group fused Lasso (PAGFL) penalty, which can automatically identify latent group structures by shrinking the difference in slopes between two individuals towards zero. Compared to C-Lasso, this approach is guaranteed to converge to a unique global minimizer. Second, note that clustering panel individuals into groups is essentially a supervised clustering problem. Hence, the classical K-means algorithm can be adapted to the panel regression framework with group structures with slight modifications. Research in this direction includes [2,16,17]. The third popular approach is based on the sequential binary segmentation algorithm (SBSA) of [18] which was originally designed for structural break detection. Ref. [19] employed the SBSA to detect latent group structures in nonlinear panel models, which consists of two steps: first, sorting the preliminary unconstrained consistent estimators of the regression coefficients, and then performing break point detection. The individuals within adjacent estimated change points therefore form a group. This idea has subsequently been extended by [20,21] to detect the latent structures in spatial dynamic panels and time-varying panels.
These methods make important contributions but often ignore the use of prior information. It is well-known that incorporating prior information in economic and statistical modeling can help improve estimation performance. For example, in the application of predicting clinical measures from neuroimages, ref. [22] proposed a sparse fused group Lasso method that incorporates spatial and group structure information to enhance prediction accuracy. Similarly, ref. [23] detected the sparsity and homogeneity of regression coefficients with imposed prior constraints. To quickly identify clustered patterns in spatial regression and reduce computational burden, ref. [24] built a minimum spanning tree (MST) based on the spatial neighborhood information, and then detected homogeneity in coefficients based on the tree. Along this line, ref. [25] proposed a new MST by leveraging the concept of network topology and the local model similarity, simultaneously. Numerical experiments and real-world data analysis in the literature demonstrated that incorporating more prior information can achieve better estimation and clustering efficiency.
In panel data analysis, researchers often have access to prior information before economic modeling, such as existing research conclusions or objective geographic information. For example, ref. [15] used the preliminary individual-specific estimator to construct the data-adaptive weights in PAGFL. The adaptive weights can be regarded as the indicator for similarity among units that implies the likelihood of two units being in the same group. One important example, which is closely related to this work, is [26], who extended the idea of [6] and exploited a panel-CARDS algorithm. This method first sorts all individuals based on a preliminary estimator, and then constructs an ordered partition set for the individuals. By applying shrinkage estimation between and within the segments, the latent structure can be identified in a data-driven manner. Notably, constructing the rank mapping from the preliminary estimates effectively borrows prior information. Another leading example arises from [2], who studied grouped fixed effects and assumed that some coefficients do not change across individuals. This model setup also posits the prior information that the coefficients of certain regressors are homogeneous across individuals. Hence, as noted by [2] in their supplementary material, adding prior information on the structure of unobserved homogeneity deserves further investigation, and we are currently working on this issue.

1.2. Contributions and Organization

This paper mainly focuses on incorporating the prior information to improve the latent structures detection performance. The contributions are as follows. First, we provide a general prior information framework that can be represented as a constrained region for regression coefficients for panel models. To  identify the latent group structure and utilize the prior information, we develop a novel penalized estimation scheme based on the concave pairwise fusion penalty combining the constraint information. Second, due to the restrictions imposed by the prior constraint, traditional numerical optimization algorithms cannot be directly applied here. To implement the proposed approach, we further design an alternating direction method of multipliers (ADMM) algorithm [27] for solving the constrained fused Lasso problem. The ADMM algorithm is straightforward and easy to implement using free software such as R. Additionally, we have developed a new R package phpadmm to facilitate the estimation procedures described in this paper, which is available upon request. Third, we also demonstrate the convergence property of the ADMM algorithm and the asymptotic results for the estimators. We validate the theoretical results through simulation studies and real-world application. Fourth, we apply the proposed latent structures detection procedure to investigate the relationship between electricity consumption and GDP in China, a significant topic that has garnered considerable interest in the field of energy economics, and find four subgroups among the 30 provinces.
The remainder of the paper is structured as follows. We illustrate the proposed ADMM estimation method in Section 2. The  convergence of the algorithm and the asymptotic theory for the estimators are examined in Section 3. Section 4 compares the finite sample performance of different methods. We investigate the effect of electricity consumption on GDP using proposed methodology in Section 5. Section 6 concludes the paper. The proofs and additional testing procedure are given in the Appendix.
Notation. Throughout this paper, we adopt the following notations. Denote 0 m as a m × 1 zero vector, 1 m as a m × 1 unit vector, and  I m as a m × m identity matrix. Let a , b = a T b be the inner product of two vector a and b with the same dimension. For an m × n matrix A = ( a 1 , , a n ) T , we denote its transpose as A T and vec ( A ) = ( a 1 T , , a n T ) T . Throughout this paper, a superscript zero on any quantity refers to the corresponding true quantity, which is a fixed number, vector or matrix.

2. Model and Proposed Estimation Method

2.1. Panel Heterogeneity with Prior Constraint Information

Let y i t be the real-valued dependent variable and x i t = ( x i t , 1 , , x i t , p ) a p × 1 vector of regressors for i = 1 , , N and t = 1 , , T , where i and t index the individual and time respectively. We consider the following linear panel model with grouped latent structures pattern:
y i t = μ i + x i t T β i 0 + ϵ i t , i = 1 , , N ; t = 1 , , T ,
where μ i is the fixed individual effect for the ith individual, ϵ i t is error term with mean zero, and  β i 0 is a p × 1 vector of unknown coefficients for the ith individual. The set of coefficients { β 1 0 , , β N 0 } admits the following latent grouping structure
β i 0 = α 1 0 , if   i G 1 0 α K 0 0 if   i G K 0 0
where K 0 is an unknown integer, α l 0 α k 0 for any l k , and  α k represents the common slope parameters for the kth subgroup. The set G 0 = { G 1 0 , , G K 0 } forms a partition of { 1 , , N } and G l 0 G k 0 = for any l k .
Let B = ( β 1 , , β N ) T be a K × p coefficient matrix, and let β = vec ( B ) . In this paper, we assume that the prior information can be formulated as a constrained region, i.e.,
β C ,
where C is a non-empty convex set, which is quite general. For example, in a cross-province panel data analysis, one may consider that the provinces in neighboring regions are more likely to belong to the same group. We can incorporate geographical information into C . Similarly, in financial studies, the financial returns of similar industry sectors tend to share the same scores on the common factors. In addition, when C = { β : β R N p } , group identification in (2) simplifies to the no-prior information situation, which has been discussed by e.g., [5,12,13]. The convex set C is user-specified that can include equality constraints, inequality constraints, or other constrained forms depending on practical considerations. To enhance the readability of this paper, we summarize the key notations in Table 1.
Prior information is typically known and significant, especially in the context of this paper. As mentioned earlier, several existing methods has utilized the prior information in identifying the latent group structures in panel models, including the ordering-based method [26], and the adaptive weights in the group fused Lasso penalty [15]. In essence, these two methods both make use of the distance between any two units’ initial slope estimators. In contrast, our method generalizes the definition of prior information including both the equality and inequality constraints. We summarize the differences between the existing works and our framework about prior information in Table 2. Effectively utilizing this prior information can enhance both homogeneity detection and parameter estimation accuracy. Our objective in this paper is to explore how to incorporate these prior constraints (3) into the process of detecting homogeneity.

2.2. A Regularized Approach

To detect homogeneity in panel data models, one may utilize the K-means clustering method, as demonstrated by [2,8], or the binary segmentation method proposed by [28]. However both methods struggle to incorporate prior information effectively. In  this paper, we consider a regularized approach with a concave pairwise fusion penalty to detect homogeneity in model (1). This method was firstly applied in [29] for subgroup analysis and has been extensively employed in various fields, such as precision medicine. After concentrating out the fixed effects, the objective function can be represented as
L ( β ; γ ) = 1 2 N T i = 1 N t = 1 T ( y ˜ i t x ˜ i t β i ) 2 + i < j p γ ( β i β j 1 ) s . t . β C
where x ˜ i t = x i t x ¯ i and y ˜ i t = y i t y ¯ i in which x ¯ i = 1 T t = 1 T x i t and y ¯ i = 1 T t = 1 T y i t , p γ ( · ) is penalty functions and γ > 0 is a tuning parameters that control the degree of penalty on β i β j 1 . There are a total of N ( N 1 ) 2 terms in i < j p γ ( β i β j 1 ) , which encourages shrinkages of the pairwise differences in slopes for any two subjects. This regularized adjustment can identify the latent homogeneous structures for the coefficients β . Ref. [30] pointed out that the non-convex penalties p γ ( · ) should satisfy four conditions, and the smoothly clipped absolute deviation (SCAD) [9] and minimax concave (MCP) [31] penalties are adequate choices.
To further incorporate the prior constraint condition (3), we introduce a new regularized term and rewrite the objective function (4) as
L ˜ = 1 2 N T i = 1 N t = 1 T ( y ˜ i t x ˜ i t β i ) 2 + i < j p γ ( β i β j 1 ) + I C ( β )
where I C ( · ) is an indicator function for β , equaling a constant when β belongs to C and + otherwise. It should be noted that directly minimizing the Equation (5) is computationally challenging, since the SCAD and MCP penalties are not convex, and the pairwise fusion penalization term is non-separable concerning β i in the estimation. To achieve computational feasibility, we formally propose an ADMM algorithm to minimize (5).

2.3. ADMM Implementation

ADMM algorithm is an optimization algorithm that is particularly useful for solving optimization problems with specific structures, such as those involving separable or block-separable objective functions [27]. The key idea of the ADMM algorithm is to decompose a global optimization problem that is hard to solve into several small pieces that can be solved more easily. In terms of our optimization problem, following the principle of ADMM, we first introduce two sets of auxiliary parameters δ i j = β i β j and η i = β i to reparameterize the objective function (5) as a constrained optimization problem
L ˜ ( β , δ ) = 1 2 N T i = 1 N t = 1 T ( y ˜ i t x ˜ i t β i ) 2 + i < j p γ ( δ i j 1 ) + I C ( η ) s . t . β i β j δ i j = 0 β η = 0
where δ = ( δ i j T , i < j ) T and η = ( η 1 T , , η N T ) T . To solve (6), we use the ADMM algorithm to solve the following augmented Lagrangian function
Q ( β , η , δ , ν , ϑ ) = L ˜ ( β , δ ) + i < j ν i j , β i β j δ i j + ρ 2 2 i < j β i β j δ i j 2 + I C ( η ) + j = 1 N ϑ j , β j η j + ρ 1 2 j = 1 N β j η j 2
where ν = ( ν i j T , i < j ) T , ϑ = ( ϑ j T , , ϑ N T ) T are Lagrange multipliers, and  ρ 1 , ρ 2 are two fixed augmented parameters. The ADMM algorithm has the advantage of decomposing (7) into several manageable subsets. Along this line, we consider alternating minimization for each block of parameters β , η , δ , ϑ and ν . Specifically, given the current values at the mth iteration, denoted as ( β ( m ) , η ( m ) , δ ( m ) , ϑ ( m ) , ν ( m ) ) , the ADMM computation at the ( m + 1 ) th iteration proceeds as follows:
β ( m + 1 ) = arg min β Q ( β , η ( m ) , δ ( m ) , ϑ ( m ) , ν ( m ) ) ,
η ( m + 1 ) = arg min η Q ( β ( m + 1 ) , η , δ ( m ) , ϑ ( m ) , ν ( m ) ) ,
δ ( m + 1 ) = arg min δ Q ( β ( m + 1 ) , η ( m + 1 ) , δ , ϑ ( m ) , ν ( m ) ) ,
ϑ ( m + 1 ) = ϑ ( m ) + ρ 1 ( β ( m + 1 ) ) η ( m + 1 ) ) ,
ν ( m + 1 ) = ν ( m ) + ρ 2 ( Ω β ( m + 1 ) δ ( m + 1 ) ) .
For the first minimization problem in (8), β ( m + 1 ) is the minimizer of
H ρ ( β ) = 1 2 N T i = 1 N t = 1 T ( y ˜ i t x ˜ i t β i ) 2 + ρ 2 2 i < j β i β j δ i j ( m ) + ρ 2 1 ν i j ( m ) 2 + ρ 1 2 j = 1 N β j η j ( m ) + ρ 1 1 ϑ j ( m ) 2 + constant .
By simple algebraic manipulation, we can rewrite it in matrix form
H ρ ( β ) = 1 2 N T y ˜ x ˜ β 2 + ρ 2 2 D β δ ( m ) + ρ 2 1 ν ( m ) 2 + ρ 1 2 β η ( m ) + ρ 1 1 ϑ ( m ) 2 + constant ,
where y ˜ = ( y ˜ 1 T , , y ˜ N T ) T is a N T × 1 vector with y ˜ i = ( y ˜ i 1 , , y ˜ i T ) T , x ˜ = diag ( x ˜ 1 , , x ˜ N ) is a N T × N p matrix with x ˜ i = ( x ˜ i 1 , , x ˜ i T ) T , and  D : = E I p is a N ( N 1 ) 2 p × N p matrix, where ⊗ represents the Kronecker product, E = { ( e i e j ) , i < j } N ( N 1 ) 2 × N and e i is the N × 1 vector with one at the ith component, and zeros elsewhere. Notably, the minimizer has a closed-form solution:
β ( m + 1 ) = 1 N T x ˜ T x ˜ + ρ 1 I N p + ρ 2 D T D 1 × 1 N T x ˜ y + ρ 1 ( η ( m ) ρ 1 1 ϑ ( m ) ) + ρ 2 D ( δ ( m ) ρ 2 1 ν ( m ) ) .
For η ( m + 1 ) in (9), it is actually a standard convex optimization problem, since we only need to minimize
ρ 1 2 ξ j ( m ) η j + I C ( η j )
where ξ j ( m ) = β j ( m + 1 ) + ρ 1 1 ϑ j ( m ) . In practical implementation, we employ the quadratic programming algorithm by the R function “solve.QP” available from the package quadprog.
For δ ( m + 1 ) in (10), by discarding the terms that are independent of δ , we just need to minimize
ρ 2 2 ζ i j ( m ) δ i j 2 + p γ ( δ i j 1 )
where ζ i j ( m ) = β i ( m + 1 ) β j ( m + 1 ) + ρ 2 1 ν i j ( m ) . For the MCP with γ > 1 / ρ 1 , the updated δ ( m + 1 ) has a closed-form solution
δ i j ( m ) = S ( ζ i j ( m ) , λ / ρ 1 ) 1 1 / ( γ ρ 1 ) if   ζ i j ( m ) γ λ , ζ i j ( m ) if   ζ i j ( m ) > γ λ .
For the SCAD penalty with γ > 1 / ρ 1 + 1 , the update is given by
δ i j ( m ) = S ( ζ i j ( m ) , λ / ρ 1 ) if   ζ i j ( m ) λ + λ / ρ 1 , S ( ζ i j ( m ) , γ λ / ( ρ 1 ( γ 1 ) ) ) 1 1 / ( γ ρ 1 ) if   λ + λ / ρ 1 < ζ i j ( m ) γ λ , ζ i j ( m ) if   ζ i j ( m ) > γ λ .
where S ( z , t ) = ( 1 t / z ) + z is the groupwise soft-thresholding operator with ( x ) + = x if x > 0 and 0 otherwise.
As suggested by [27], a reasonable termination criterion for an ADMM algorithm is that both the primal and dual residuals are sufficiently small. In this paper, the primal and dual residuals are defined as
r 1 ( m + 1 ) 2 = Ω β ( m + 1 ) δ ( m + 1 ) 2 , r 2 ( m + 1 ) 2 = β ( m + 1 ) η ( m + 1 ) 2
and
s 1 ( m + 1 ) 2 = ρ 1 ( η ( m + 1 ) η ( m ) ) 2 , s 2 ( m + 1 ) = ρ 2 Ω ( δ ( m + 1 ) δ ( m + 1 ) ) 2 ,
respectively. When lim m r k ( m ) 2 e primal for k = 1 , 2 and lim m s 1 ( m ) + s 2 ( m ) 2 e dual hold, where e primal and e dual are pre-specified feasible primal and dual tolerances, we stop the iteration. We summarize the above procedures in Algorithm 1.
Algorithm 1: ADMM algorithm for panel data models with prior constraints
Mathematics 13 01505 i001

2.4. Initial Values and Tuning Parameter

In nonconvex optimization, selecting an appropriate initialization for parameters is crucial, because this will not only produce an optimal solution but also significantly accelerate the iterations. As proposed by [26], we set the initial values for β as the estimators without the latent group structure, that is
L N T ( β ) = 1 2 N T i = 1 N t = 1 T ( y ˜ i t x ˜ i t β i ) 2 .
Solve the optimization problem yields the least square estimates
β i ( 0 ) = 1 T t = 1 T x ˜ i t x ˜ i t T 1 1 T t = 1 T x ˜ i t y ˜ i t ,
for i = 1 , , N . Then, set δ i j ( 0 ) = β i ( 0 ) β j ( 0 ) , η i ( 0 ) = β i ( 0 ) , ϑ ( 0 ) = 0 , and  ν ( 0 ) = 0 . The question arises regarding how to determine the tuning parameters ρ 1 , ρ 2 and λ . Following [29], we set ρ 1 = ρ 2 = 1 and λ = 3 .
The selection of γ is particularly important, since the performance of group structures identification depends on γ . Following the strategy of [5], we minimize the following BIC-type criterion function:
IC ( γ ) = log ( σ N T 2 ( γ ) ) + p K ( γ ) ψ N T
where σ N T 2 ( γ ) = L N T ( β ( γ ) ) is the sum of squared residuals, and  K ( γ ) denotes the number of groups corresponding to γ . Here, the positive constant ψ N T that depends on N and T serve as a balance between the degrees of freedom and the model fitting errors. To select the optimal γ , define a grid of points γ min = γ 0 < < γ W = γ max with equal spacing. To speed up the computation, we employ the warm start strategy. Specifically, for each γ { γ 0 , , γ W } , we compute the solution path β ^ ( γ w ) under the initial value β ^ ( γ w 1 ) . The optimal value for γ is defined as γ ^ = arg min γ w , w = 0 , , W IC ( γ w ) .

3. Theoretical Analysis

In this section, we mainly discuss the the convergence of the ADMM Algorithm 1, and establish the asymptotic property of the proposed estimator.

3.1. Convergence of the Algorithm

Under some mild conditions, the convergence of the ADMM algorithm can be guaranteed. For the iterated sequence { β ( m ) , η ( m ) , δ ( m ) , ϑ ( m ) , ν ( m ) } , there exists a global minimizer { β , η , δ , ϑ , ν } satisfying the first order condition for a stationary point The following proposition illustrates the convergence of Algorithm 1.
Proposition 1. 
Recall the primal and dual residuals defined in (16) and (17). It holds that lim m r j ( m ) 2 = 0 for j = 1 , 2 , and  lim m s 1 ( m ) + s 2 ( m ) 2 = 0 for both SCAD and MCP penalties. Therefore, the sequence { β ( m ) , η ( m ) , δ ( m ) , ϑ ( m ) , ν ( m ) } m = 1 will approach to an optimal point.
Proposition 1 indicates that Algorithm 1 converges if both the primal feasibility and dual feasibility hold. The proof of this proposition follows a strategy similar to that in [29], and the detailed proof is postponed to the Appendix.

3.2. Asymptotic Property

We now establish the theoretical properties of the penalized estimator. First, we introduce some notations. Let x i = ( x i 1 , , x i T ) T and y i = ( y i 1 , , y i n ) T . Let G ( K ) = ( G 1 ( K ) , , G K ( K ) ) be an arbitrary K-partition of { 1 , , N } . We define
σ ^ G ( K ) 2 = ( N T ) 1 k = 1 K i G k ( K ) t = 1 T y ˜ i t x ˜ i t T β ^ i , G k ( K ) 2 ,
where β ^ i , G k ( K ) = arg min β i ( N k T ) 1 i G k ( K ) t = 1 T ( y ˜ i t x ˜ i t T β i ) 2 . Denote N k as the number of elements in the kth group. Define ρ ( t ) = λ 1 p γ ( t , λ ) as the scale penalty function and ρ ( t ) = p λ ( | t | ) sgn ( t ) as its first-order derivative. We then introduce the following assumptions.
Assumption 1. 
(i) { ( x i t , y i t ) , t = 1 , , T } is stationary strong mixing for each i with mixing coefficients α i ( · ) . α ( · ) = max i α i ( · ) satisfies α ( τ ) c α ρ τ for some constant c α > 0 and ρ ( 0 , 1 ) . (ii) The random variables { x i , y i } i = 1 N are independent, and  max i , t E x i t 2 q < c 1 for some constant c 1 < and q > 4 . (iii) There exists two positive constants c 2 and c 3 such that min 1 k K μ min 1 T N k i G k 0 E ( x ˜ i T x ˜ i ) c 2 and max 1 i N μ max 1 T E ( x i T x i ) c 3 . (iv) The error term satisfies E ( ϵ i t ) = 0 , E ( x i t ϵ i t ) = 0 and max i , t E | ϵ i t | 2 q < c 1 . (v) N k either tends to infinity or remains fixed as T and N = O ( T 2 ) .
Assumption 2. 
(i) The function ρ ( t ) is symmetric, non-decreasing and concave on [ 0 , ) . It satisfies ρ ( 0 ) = 0 and is constant for all t a λ for some constant a > 0 . (ii) The derivative ρ ( t ) satisfies ρ ( 0 + ) = 1 and is continuous except for a finite number of values of t.
Assumption 3. 
(i) There exist positive definite matrices Ω k for k = 1 , , K such that 1 T N k i G k 0 t = 1 T x ˜ i t x ˜ i t T p Ω k as T or ( N k , T ) . (ii) Let B k = 1 N k T i G k 0 t = 1 T E ( x ˜ i t ϵ i t ) . Then N k T i G k 0 t = 1 T x ˜ i t ϵ i t B k d N ( 0 , Ψ k ) as T or ( N k , T ) where B k equals O p ( N k / T ) if x ˜ i t is not exogenous and 0 otherwise.
Assumption 4. 
min 1 K ^ K inf G ( K ^ ) σ ^ G ( K ^ ) 2 p σ ̲ 2 > σ 0 2 as ( N , T ) where σ 0 2 = lim ( N , T ) ( N T ) 1 k = 1 K 0 i G k 0 t = 1 T E ( y ˜ i t x ˜ i t T β i 0 ) .
Assumption 5. 
As ( N , T ) , ψ N T 0 and ψ N T T .
Assumption 1(i,ii) imposes some conditions on { ( x i t , y i t ) } , which are quite common in panel data models. Assumption 1(iii,iv) specifies moment conditions on the regressors x i t and the noise term ϵ i t , respectively. Assumption 1(v) allows the number of elements in each group to be finite, which distinguishes it from the conditions set forth in [5]. Assumption 2 is standard for concave penalty functions, such as SCAD and MCP. Assumption 4 is utilized to study the asymptotic normality of the proposed estimators. Finally, Assumptions 4 and 5 are adapted from [5] to demonstrate the classification consistency described in (18).
Once β ^ is given, the estimated group pattern can be directly derived by classifying the coefficients β ^ i ’s into groups. We denote K ^ as the estimated number of groups and α ^ k as the common slope shared by the kth group for k = 1 , , K ^ . By definition, the estimated group pattern G ^ = ( G ^ 1 , , G ^ K ^ ) is given by
G ^ k = { i { 1 , , N } : β ^ i = α ^ k } for   k = 1 , , K ^ .
We then have the following limiting distribution for α ^ k .
Theorem 1. 
Suppose that Assumption 1–3 hold. Then,
N k T ( α ^ k α k 0 ) Ω k 1 B k d N ( 0 , Ω k 1 Ψ k Ω k 1 )
for k = 1 , , K as T or ( N k , T ) .
Theorem 1 demonstrates that the group-specific estimator possesses the asymptotic normality properties with a convergence rate of N k T . If the group membership is known, the oracle estimator for α is given by
α ^ k o r = i G k 0 x ˜ i T x ˜ i 1 i G k 0 x ˜ i T y ˜ i , for   k = 1 , , K .
As pointed by [5,26], α ^ k o r and α ^ k share similar asymptotic normality. Given the estimated grouping structure G ^ , a post-Lasso version of α is obtained by
α ^ G ^ k = i G ^ k x ˜ i T x ˜ i 1 i G ^ k x ˜ i T y ˜ i , for   k = 1 , , K ^ .
The post-Lasso estimators perform at least as well as the Lasso estimators in terms of convergence rate, but they exhibit a smaller second-order bias, as shown by [32]. It would be interesting to investigate the higher-order asymptotic properties of the post-Lasso estimators within our framework.
We utilize the information criterion defined in (18) to determine the number of groups. The following theorem justifies its asymptotic validity.
Theorem 2. 
Suppose that Assumptions A1–A2 and A4–A5 hold. The information criterion in (18) will select a optimal tuning parameter γ o p t such that the estimated β ^ ( γ o p t ) approaches to the oracle estimator β ^ o r with probability approaching 1, and 
P ( K ^ ( γ o p t ) = K ) 1 , as   ( N , T ) .  
Theorem 2 indicates that under some mild condition, the minimizer of (18) can only be the one that produces the correct number of groups, as N or T or both go to infinity. Since ψ N T in the IC (18) plays a crucial role in determining the latent group structures, we follow [5] to pick ψ N T by trying a list of candidates. Similar to [26], we find that ψ N T = 1 2 N T has fairly good sample performance. The same setting is also aplied in the empirical study.

4. Simulation Studies

In this section, we conduct a series of simulation studies to assess the finite sample performance of the proposed method. We consider the following three data generating processes:

4.1. Basic Setup

DGP 1. The first data setting is similar to the GDP 1 in [5]. Here, the observations { ( y i t , x i t ) , i = 1 , , N , t = 1 , , T } are generated from a linear panel model where x i t = ( 0.2 μ i + e i t 1 , 0.2 μ i + e i t 2 ) T with the fixed effect μ i following an i.i.d. standard normal distribution across time. The error terms e i t 1 and e i t 2 are i.i.d. standard normal and independent of μ i . The slopes of individuals are divided into three groups with a ratio of N 1 : N 2 : N 3 = 4 : 3 : 3 . The true coefficients are
B a = 0.4 1 1.6 1.6 1 0.4 T .
DGP 2. The DGP 2 is similar to the DGP 3 in [26]. In this setting, the coefficients are divided into eight groups, where the first group accounts for 30% of the total individuals and each of the remaining seven groups shares 10% of individuals. The generation for x i t is similar to DGP 1. The group-specific coefficients take the values
B b = 4 3 2 1 1 2 3 4 4 3 2 1 1 2 3 4 T .
DGP 3. In the third DGP, we let x i t = ( x i t 1 , x i t 2 , x i t 3 ) T and x i t k = 0.2 μ i + e i t k for k = 1 , 2 , 3 . Individuals are also divided into three groups with a ratio of N 1 : N 2 : N 3 = 3 : 4 : 3 . The true parameter values for each group are
B c = 0.6 0.6 0.6 1.5 1 0.5 1 0 1 T
The constraint set for each DGP includes both equality or inequality constraints, which are described as follows.
(i)
(equality constraint:) For B a , the sum of coefficients for each individual is the same. For  B b , the sum of all elements equals zero, and for B c , the coefficients for the first regressor remain constant across all individuals.
(ii)
(inequality constraint:) For B a , all elements in the coefficient matrix are required to be greater than zero.
Hence, the constrained regions for β a : = vec ( B a ) is
C a = { β a : R a e T β a = 0 , R a i T β a > 0 6 }
where R a e = ( ( 1 2 T , 1 2 T , 0 2 T ) T , ( 0 2 T , 1 2 T , 1 2 T ) T ) T and R a i = I 6 . For  β b : = vec ( B b ) ,
C b = { β b : R b e T β b = 0 }
where R b e = 1 16 . For  β c : = vec ( B c ) ,
C c = { β c : R c T β c = 0 2 }
where R c = ( ( 1 , 0 2 , 1 , 0 5 ) T ; ( 0 3 , 1 , 0 2 , 1 , 0 2 ) T ) T .
Throughout these DGPs, we consider the individual sizes N = 100 , 200 and the time spans T = 10 , 20 and 40. The error terms ϵ i t are drawn from a standard normal distribution, independent across i and t, and are also independent of all regressors. The fixed effects μ i and the error terms ϵ i t are mutually independent. For each case, we conduct 300 replications. For comparison, we also consider the results obtained without prior information as reported by [5].

4.2. Simulation Results

We present the frequency with which a specific number of groups is obtained across 300 replications for all DGPs. While the accurate determination of the number of groups is undoubtedly crucial, this metric alone does not provide insight into the degree of similarity between the estimated groups and the true groups. Following [6], we employ the normalized mutual information (NMI) measure to assess the similarity between the estimated group structure G ^ and the true group structure G 0 . Specifically, for two sets of disjoint clusters A = { A 1 , A 2 , } and B = { B 1 , B 2 , } on the same set { 1 , , N } , the NMI is defined as
NMI ( A , B ) = I ( A ; B ) [ H ( A ) + H ( B ) ] / 2
where
I ( A ; B ) = i , j ( | A i B j | / N ) log N | A i B j | | A i | · | B j | and H ( A ) = i | A i | N log | A i | N ,
where | A | denotes the cardinality of set A. Obviously when two groupings are exactly the same, i.e.,  A = B , we have I ( A ; B ) = H ( A ) = H ( B ) , and  NMI ( A , B ) = 1 . In general, the NMI measure takes values on [ 0 , 1 ] , and larger NMI implies the two group structures are closer. To measure the estimation accuracy for slope parameters, we use the average model error, which is defined as 1 N ( β ^ β 0 ) T ( β ^ β 0 ) .
Figure 1 reports the classification results across DGPs 1–3 for difference combinations of N and T. The yellow and green boxes represent the detection results with and without prior constraints, respectively. It shows the NMI between the estimated group structure G ^ and the true group structure G 0 , and suggests that, as N or T increase, the NMI increases rapidly for both methods. More importantly, our method exhibits higher NMI values across all DGPs, which implies that incorporating prior information can substantially improve the classification performance. This phenomenon is particularly pronounced when T is small. We display the average model estimation error in Figure 2. As expected, the average model estimation error decreases as N or T increase. It is obvious that the proposed parameter estimators has smaller estimation biases, since some prior constraint information can help us to gain the extra efficiency of homogeneity detection.
Next, give the number of groups, we focus on the point estimation of the post-Lasso estimator. To save space, we report only the simulation results of the first group-specific parameter of all DGPs in Table 3, based on N = 200 individuals. The oracle estimators, which assume that the latent group is known, are also included for comparison. As shown in the Table, the biases of the three types of estimators become negligible as T increases, indicating that all the estimators are consistent. Additionally, the standard deviations (SD) are close to their corresponding estimated standard errors (ESE), and the empirical coverage probabilities (ECP) converge to the nominal level of 95%. This validates the asymptotic normality properties of the proposed estimator, as demonstrated in Theorem 1. The no-prior estimators yield relatively smaller ECP and larger biases compared to the other two types of estimators, especially when T is small. Undoubtedly, the oracle estimators perform reasonably well, and our method produce results closer to those of the oracle estimators. In summary, incorporating prior information significantly improves the finite sample performance in terms of estimation efficiency, particularly when the time span is small.

5. Empirical Analysis

With the deepening of China’s reform and opening up, the country has undergone profound changes over the past few decades, becoming the second-largest economic entity and the second-largest power consumer in the world. It is well known that electricity plays a crucial role in both human activity and industrial production, including sectors such as communication and engineering. Thus, electricity serves as a solid foundation for economic development. Consequently, many researchers have focused on the relationship between electricity consumption and economic growth. For example, ref. [33] found that real GDP and electricity consumption in China are cointegrated using data from 1971 to 2004. Additionally, ref. [34] investigated the relationship between electricity consumption and economic growth through a panel data analysis. China’s vast territory encompasses 34 provinces, municipalities, and autonomous regions, creating a significant domestic market. This paper primarily focuses on the homogeneous effect of electricity consumption on GDP across these provinces.
GDP is the most closely monitored statistical index in macroeconomics and serves as an important measure of economic development. At the beginning of each year, the Chinese government releases the provincial GDP data for the previous year. The electricity consumption data used in this study are derived from the Li Keqiang Index (https://en.wikipedia.org/wiki/Li_Keqiang_index#) (accessed on 1 February 2025), which has been cited by The Economist as a reliable measure of China’s economy. This index incorporates three indicators electricity consumption, railway cargo volume, and bank lending that reflect the operation of China’s economy more accurately. Due to significant data gaps and challenges in data collection, the Tibet Autonomous Region, Hong Kong, Macao, and Taiwan are excluded from this analysis. Therefore, the study includes a total of 30 provinces, with observations spanning from 2001 to 2019, resulting in N = 30 and T = 19 . All data utilized in this study are sourced from the China Statistical Yearbook Database, available at https://www.stats.gov.cn/sj/ndsj (accessed on 1 February 2025).
We first examine the existence of a homogeneous effect across provinces using a residual-based bootstrapping procedure. The detailed algorithm can be found in the Appendix. The resulting p-value is very close to zero, indicating the statistical significance of the homogeneous effect. Moreover, numerous empirical analyses suggest a positive relationship between GDP and electricity consumption. Therefore, we assume that the coefficients for all provinces are positive as prior information. To investigate this hidden homogeneity, we consider the following linear fixed effects model with a latent group structure:
y i t = μ i + x i t β i + ϵ i t , i = 1 , , N ; t = 1 , , T , s . t . β i > 0 ,
where y i t is the logarithm of GDP for the ith province at the tth year, x i t represents electricity consumption, μ i characterizes the fixed effect, and  ϵ i t is the error term. The prior information here includes a inequality constraint for parameters β i ’s, i.e.,  C = { β i > 0 , i = 1 , , N } .
Table 4 presents the classification results using two different approaches: the proposed estimator incorporating the inequality constraint C and the PGFAL estimator proposed by [15] that does not account for the prior constraint. Using our proposed ADMM algorithm described in Algorithm 1, the provinces are divided into two groups. As shown by the Table, Group 1 includes provinces such as Anhui, Fujian, Guangdong, Guangxi, Hainan, Hebei, Henan, Hubei, Hunan, Jiangsu, Liaoning, Inner Mongolia, Shandong, Shanxi, Shanghai, Sichuan, Xinjiang, Yunnan, and Zhejiang, with an estimated coefficient of 1.1692. Group 2 comprises Beijing, Gansu, Guizhou, Heilongjiang, Jilin, Jiangxi, Ningxia, Qinghai, Shaanxi, Tianjin, and Chongqing, with a significantly higher coefficient of 3.4770.
Group 1 consists largely of provinces from the Eastern and Central regions, characterized by advanced industrialization, technological innovation, and a strong emphasis on services and capital investment. In these regions, the direct relationship between electricity consumption and economic growth is weaker, as economic growth is increasingly driven by non-energy-intensive sectors. While electricity consumption remains significant, its marginal impact on GDP growth is reduced due to energy efficiency improvements and shifts toward high-value-added industries. Group 2, primarily consisting of provinces from the Western and Northeastern regions, shows a stronger dependency of economic growth on electricity consumption. These regions are still in the process of industrialization, infrastructure development, and urbanization, making electricity consumption a crucial driver of economic growth. Beijing, Tianjin, and Chongqing are notable inclusions in Group 2 despite being considered economically advanced. This classification can be explained by the ongoing infrastructure expansion and high-tech investments in these cities, which require substantial energy inputs. Beijing and Tianjin, while economically developed, are still heavily investing in energy-demanding sectors such as smart technologies, urban infrastructure, and research and development. Chongqing, traditionally an industrial city, continues to undergo substantial industrialization and urban expansion, further increasing its reliance on electricity consumption for growth. This observation is consistent with the findings of [35,36], who reported that in Western regions, energy consumption has a significant causal relationship with economic growth, while in Eastern and Central regions, this relationship is less pronounced. This pattern highlights the varying roles of electricity consumption in different stages of industrialization across China provinces, where more industrialized areas show diminishing returns from energy use, and less industrialized areas remain heavily reliant on electricity to drive economic growth.
However, if we do not consider the prior information, the provinces are classified into five groups, which consist of 2, 1, 1, 15 and 11 province(s) with the corresponding group-specific coefficients 0.6935, 0.8715, 0.9837, 1.0879 and 3.4351, respectively. This suggest that imposing the prior information helps to prevent overly divergent groupings, thus leading to more stable classifications. We also calculate the mean squared prediction errors (MSPE) of these two methods. Specifically, the observations of all individuals in the first 12 years are used as the training set, while the observations in the subsequent 7 years serves as the validation set. Based on the identified latent structures, we calculate the MSPE as
MSPE = 1 N | T val | i = 1 N t T val ( y ^ i t y i t ) 2 ,
where T val = { 13 , , 19 } , y i t is the true observation and y ^ i t is the prediction value. As a result, the MSPE of our estimator is 0.8396, while the MSPE without prior information is 3.9443, much larger than our method. The reduced prediction error illustrates that incorporating prior constraint information can improve the predictive performance in latent structures identification for panel models.

6. Discussion

Identifying the latent groups structures in panel data analysis has recently attracted significant research interest. In this paper, we explore a pairwise fusion approach incorporating prior constraint information to identify the group pattern, and design an efficient ADMM algorithm to solve the optimization problem. Simulation studies and a real data application show that the proposed estimators outperform the existing approaches in terms of both detection accuracy and predictive performance. Our work still has its limitations; for example, how to obtain the prior constraint information in practical data analysis. Besides, our asymptotic results are built on the condition that the prior information is correctly specified; however, it remains to be explored when the prior information is misspecified.
There are some interesting issues that warrant further research. First, we only consider the commonly-used linear panel model with individual fixed effects. It is possible to adapt our proposed estimation framework to other panel models with latent group structures, such as the panel models with endogenous regressors [5] or the nonlinear panel models [19]. The specific objective function and estimation routine need to be re-examined. Second, the selection consistency property in Theorem 2 assumes that the true number of groups K 0 is fixed. It is appealing to consider a general framework that allows K 0 increases with the sample size. However, it may bring additional technical difficulties. Third, it is reasonable to allow for the presence of cross-section dependence in the current models. For such case, utilizing the factor structure [37] or interactive fixed effects [38] is potentially feasible but would require specific efforts. We leave these issues as our future research topics.

Author Contributions

Conceptualization, Y.L. and X.L.; data curation, M.L.; formal analysis, X.L.; investigation, Y.L.; methodology, M.L.; project administration, Y.L.; software, M.L.; supervision, Y.L.; validation, X.L.; visualization, Y.L.; writing original draft, X.L. and M.L.; writing review and editing, Y.L. All authors have read and agreed to the published version of the manuscript.

Funding

This is work was supported by Natural Science Foundation of Hunan Province (No. 2023JJ40453), Scientific Research Project of Education Department of Hunan Province (No. 23B0086), and National Natural Science Foundation of China (No. 72374071).

Data Availability Statement

Data are available in a publicly accessible repository: https://www.stats.gov.cn/sj/ndsj (accessed on 14 March 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

Appendix A.1

This appendix mainly contains the technical proofs and the test procedure for heterogeneity.

Appendix B. Proofs

Appendix B.1

Proof of Proposition 1. 
By (11) and (12),
Q ( β ( m + 1 ) , η ( m + 1 ) , δ ( m + 1 ) , ν ( m ) , ϑ ( m ) ) Q ( β ( m + 1 ) , η ( m + 1 ) , δ ( m + 1 ) , ν ( m ) , ϑ ( m ) ) = ρ 1 1 ϑ ( m + 1 ) ϑ ( m ) 2 + ρ 2 1 ν ( m + 1 ) ν ( m ) 2 .
Then for any δ and η , we have
Q ( β ( m + 1 ) , η ( m + 1 ) , δ ( m + 1 ) , ν ( m + 1 ) , ϑ ( m + 1 ) ) Q ( β ( m + 1 ) , η ( m + 1 ) , δ ( m + 1 ) , ν ( m ) , ϑ ( m ) ) ,
from (10) and (13). Define a constraint set S ( η , δ ) = { ( η , δ ) : β ( m + 1 ) η = 0 ; Ω β ( m + 1 ) δ = 0 } . Define
f ( m + 1 ) = inf S ( η , δ ) 1 2 N T y x ˜ β ( m + 1 ) 2 + i < j p γ ( δ i j , λ ) + I C ( η ) = inf S ( η , δ ) Q ( β ( m + 1 ) , η , δ , ν ( m ) , ϑ ( m ) )
Hence,
Q ( β ( m + 1 ) , η ( m + 1 ) , δ ( m + 1 ) , ν ( m ) , ϑ ( m ) ) f ( m + 1 ) .
Let t be an arbitrary positive integer. From (10) and (13), we obtain
ϑ ( m + t 1 ) = ϑ ( m ) + ρ 1 i = 1 t 1 ( β ( m + i ) η ( m + i ) ) , ν ( m + t 1 ) = ν ( m ) + ρ 2 i = 1 t 1 ( Ω β ( m + i ) δ ( m + i ) ) .
It holds that
Q ( β ( m + t ) , η ( m + t ) , δ ( m + t ) , ν ( m + t 1 ) , ϑ ( m + t 1 ) ) = 1 2 N T y x ˜ β ( m + t ) 2 + p γ ( | δ ( m + t ) | , λ ) + I C ( η ( m + t 1 ) ) + ν ( m + t 1 ) T ( Ω β ( m + t ) δ ( m + t ) ) + ρ 2 2 Ω β ( m + t ) δ ( m + t ) 2 + ϑ ( m + t 1 ) T ( β ( m + t ) η ( m + t ) ) + ρ 1 2 β ( m + t ) η ( m + t ) 2 = 1 2 N T y x ˜ β ( m + t ) 2 + p γ ( | δ ( m + t ) | , λ ) + I C ( η ( m + t 1 ) ) + ν ( m ) T ( Ω β ( m + t ) δ ( m + t ) ) + ρ 2 i = 1 t 1 Ω β ( m + i ) δ ( m + i ) 2 + ρ 2 2 Ω β ( m + t ) δ ( m + t ) 2 + ϑ ( m ) T ( β ( m + t ) η ( m + t ) ) + ρ 1 i = 1 t 1 β ( m + i ) η ( m + i ) 2 + ρ 1 2 β ( m + t ) η ( m + t ) 2 f ( m + t ) .
Note that Q ( β , η ( m ) , δ ( m ) , ν ( m ) , ϑ ( m ) ) is strong convex and differentiable with respect to β , as the Hessian matrix ( 1 N T x ˜ T x ˜ + ρ 1 I N p + ρ 2 Ω T Ω ) is positive definite. Together with the fact that Q ( β , η , δ , ν , ϑ ) is convex in ( δ , η ) , the sequence { β ( m ) , η ( m ) , δ ( m + 1 ) , ϑ ( m ) , ν ( m ) } m = 1 converges to a stationary point { β , η , δ , ϑ , ν } by directly applying the Theorem 4.1 of [39]. Thus we have
Ω β δ = 0 , β η = 0 ,
and
f = lim m f ( m + 1 ) = lim m f ( m + t ) = inf S ( η , δ ) 1 2 N T y x ˜ β 2 + p γ ( | δ | , λ ) + I C ( η ) .
Moreover, for any t 0 , we have
lim m Q ( β ( m + t ) , η ( m + t ) , δ ( m + t ) , ν ( m + t 1 ) , ϑ ( m + t 1 ) ) = 1 2 N T y x ˜ β 2 + p γ ( | δ | , λ ) + I C ( η ) + ( t 1 2 ) ρ 2 Ω β δ 2 + lim m ν ( m ) T ( Ω β δ ) + lim m ϑ ( m ) T ( β η ) + ( t 1 2 ) ρ 1 β η 2 f .
Hence,
lim m r 1 ( m ) 2 = r 1 2 = Ω β δ 2 = 0 , lim m r 2 ( m ) 2 = r 2 2 = β η 2 = 0 .
By the convexity of β concerning Q ( β , η ( m ) , δ ( m ) , ν ( m ) , ϑ ( m ) ) , it then follows that
Q ( β , η ( m ) , δ ( m ) , ν ( m ) , ϑ ( m ) ) / β = 1 N T x ˜ T ( y x ˜ β ( m + 1 ) ) + Ω T ν ( m ) + ρ 2 Ω T ( Ω β ( m + 1 ) δ ( m + 1 ) ) + ϑ ( m + 1 ) + ρ 1 ( β ( m + 1 ) η ( m ) ) = 1 N T x ˜ T ( y x ˜ β ( m + 1 ) ) + Ω T ( ν ( m ) + ρ 2 ( Ω β ( m + 1 ) δ ( m ) ) + ϑ ( m ) + ρ 1 ( β ( m + 1 ) η ( m ) ) = 1 N T x ˜ T ( y x ˜ β ( m + 1 ) ) + Ω T ( ν ( m + 1 ) ρ 2 ( Ω β ( m + 1 ) δ ( m + 1 ) ) + ρ 2 ( Ω β ( m + 1 ) δ ( m ) ) ) + ϑ ( m + 1 ) ρ 1 ( β ( m + 1 ) η ( m + 1 ) ) + ρ 1 ( β ( m + 1 ) η ( m ) ) = 1 N T x ˜ T ( y x ˜ β ( m + 1 ) ) + Ω T ν ( m + 1 ) + ρ 2 Ω T ( δ ( m + 1 ) δ ( m ) ) + ϑ ( m + 1 ) + ρ 1 ( η ( m + 1 ) η ( m ) ) = 0
Therefore,
s 1 ( m + 1 ) + s 2 ( m + 1 ) = 1 N T x ˜ T ( y x ˜ β ( m + 1 ) ) Ω ν ( m + 1 ) ϑ ( m + 1 ) .
Since Ω β δ 2 = 0 and β η 2 = 0 from (A3), we have
lim m Q ( β , η ( m ) , δ ( m ) , ν ( m ) , ϑ ( m ) ) / β | β = β ( m + 1 ) = lim m 1 N T x ˜ T ( y x ˜ β ( m + 1 ) ) + Ω T ν ( m + 1 ) + ρ 2 Ω T ( δ ( m + 1 ) δ ( m ) ) + ϑ ( m + 1 ) + ρ 1 ( η ( m + 1 ) η ( m ) ) = 1 N T x ˜ T ( y x ˜ β ) + Ω T ν + ϑ = 0 .
So, lim m s 1 ( m + 1 ) + s 2 ( m + 1 ) = 0 . Combining with (A4), the proof of Proposition 1 is completed.    □

Appendix B.2

Proof of Proposition 1. 
We divide the proof of Theorem 1 into several steps:
Step 1: Define M G as the restricted subspace of R N p where
M G = { β C : β i = β j , for   any   i , j G k 0 ,   1 k K }
Suppose that the true latent structure G 0 is known, the oracle estimator for β is
β ^ o r = arg min β M G 0 1 N T y ˜ x ˜ β 2 ,
and the oracle for α denoted by α ^ o r is hence given by
α ^ k o r = arg min α k 1 N k T i G k 0 t = 1 T y ˜ i t x ˜ i t T α k 2 , for   k = 1 , , K .
Then we have the following lemma. □
Lemma A1. 
Suppose that Assumption 1-2 hold. Then with probability 1 o ( K / T ) , we have β ^ o r β 0 = O p ( K / T ) .
Proof of Lemma A1. 
To prove Lemma A1, it is sufficient to show
P β ^ o r β 0 C K ( ln T ) 2 / T 1 o ( K / T ) ,
where C is some positive constant. Let E 0 = μ min 1 T N k i G k 0 x ˜ i T x ˜ i > c 2 / 2 . For any event E , we denote its complement as E c . By applying the Lemma A.1 in [26], we have:
P N k α ^ o r α k 0 C ln T / T = P N k 1 T N k i G k 0 x ˜ i T x ˜ i 1 1 T N k i G k 0 x ˜ i T ϵ i C ln T / T P N k 1 T N k i G k 0 x ˜ i T x ˜ i 1 · 1 T N k i G k 0 x ˜ i T ϵ i C ln T / T , E 0 + P ( E 0 c ) P 1 T N k i G k 0 x ˜ i T ϵ i c 2 2 C ln T / N k T + o ( T 1 ) = o ( T 1 ) .
It is obvious that
P β ^ o r β 0 2 C 2 K ( ln T ) 2 / T = P k = 1 K N k α ^ k o r α k 0 2 C 2 K ( ln T ) 2 / T k = 1 K P N k α ^ k o r α k 0 2 C 2 ( ln T ) 2 / T = o ( K / T ) .
Hence, (A6) holds. Compared with Theorem 2 in [26], the oracle estimator β ^ o r of our method has a higher probability of 1 o ( K / T ) to approach the true β 0 , while their method only has the probability 1 o ( K / T ) ϵ 0 . The underlying reason is that we take the concave pairwise fusion penalty to directly detect the homogeneity while [26] adopted the panel-CARDS procedures which required a good preliminary estimates for β at first.    □
Step 2: Let
b N T = min i G k , j G m , k m β i 0 β j 0 = min k m α k 0 α m 0
be the minimal difference of the common coefficient between two groups. It is important to show the following lemma.
Lemma A2. 
Suppose that the Assumptions in Lemma A1 hold. If  b N T > a λ and λ ϕ N T where ϕ N T = C ln T K / T , the objective function (5) has a local minimizer β ^ ( λ ) such that
P β ^ ( λ ) = β ^ o r 1 .
Proof of Lemma A2. 
Define
L N T ( β ) = 1 2 N T y ˜ x ˜ β 2 , H N T ( β ) = λ i < j ρ ( β i β j ) + I C ( β ) .
Recall the definition of M G in (A5). We define the following two mappings:
T : M G R K p and T : R N p R K p .
where T ( β ) is a K p × 1 vector consisting of K blocks with dimension p and the kth block is the common slopes of β i for i G k , and  T ( β ) = { N k 1 i G k 0 β i , k = 1 , , K } T . Obviously, when β belong to M G , T ( β ) = T ( β ) .
From (A7), we have L ˜ ( β ) = L N T ( β ) + H N T ( β ) . Define
L N T G ( α ) = L N T ( T 1 ( α ) ) , H N T G ( α ) = H N T ( T 1 ( α ) ) .
Hence L ˜ G ( α ) = L N T G ( α ) + H N T G ( α ) . We also consider a small neighborhood of β 0 :
Θ = { β C : β β 0 < ϕ N T } .
From Lemma A1, there exists an event E 1 = { β ^ o r β 0 ϕ N T } such that P ( E 1 c ) o ( K / T ) . For any β C , let β = T 1 ( T ( β ) ) . It is sufficient to show that β ^ o r is a local minimizer of L ˜ ( β ) with probability approaching 1. For this, we need to show:
(1).
In the event E 1 , L ˜ ( β ) > L ˜ ( β ^ o r ) for any β Θ and β β ^ o r ;
(2).
There exists an event E 2 such that P ( E 2 c ) o ( T 1 ) . Over the event E 1 E 2 , there exists a neighborhood of β ^ o r , denoted by Θ N T such that L ˜ ( β ) L ˜ ( β ) for any β Θ N T Θ and the inequality strictly holds when β β .
If both (1) and (2) hold, we have L ˜ ( β ) > L ˜ ( β o r ) for any β Θ N T Θ and β β ^ o r , and therefore β ^ o r is a strict local minimizer of L ˜ ( β ) over the event E 1 E 2 .
We now prove the result in (1). We first show that L ˜ G ( T ( β ) ) equals a constant for any β Θ . Recall that T ( β ) = α = ( α 1 T , , α K T ) T . Let b N T = min k m α k 0 α m 0 . Then for any k and m, we have
min 1 k < m K α k α m = min 1 k < m K ( α k α k 0 ) ( α m α m 0 ) + ( α k 0 α m 0 ) min 1 k < m K α k 0 α m 0 2 max 1 k K α k α k 0 ,
and
max 1 k K α k α k 0 = max 1 k K N k 1 i G k ( β i β i 0 ) 2 max 1 k K N k 1 i G k β i β i 0 2 ϕ N T 2 .
Then min 1 k < m K α k α m b N T 2 ϕ N T > a λ . By Assumption 2, ρ ( α k α m ) is a constant on Θ and therefore H N T ( T ( β ) ) is also a constant on β Θ . So L ˜ G ( β ) = L N T G ( T ( β ) ) + Constant for all β Θ . Note that L N T G ( α ) is strictly convex w.r.t α and α ^ o r is the unique global minimizer of L N T G ( α ) , then L N T G ( T ( β ) ) > L N T G ( α ^ o r ) for all T ( β ) α ^ o r . From (A9), we have L ˜ G ( α ^ o r ) = L ˜ ( β ^ o r ) and L ˜ G ( T ( β ) ) = L ˜ ( T 1 ( T ( β ) ) ) = L ˜ ( β ) . Hence L ˜ ( β ) > L ˜ ( β ^ o r ) for all β β ^ o r . So (1) is obtained.
We next prove the result in (2). Let Θ N T = { β C : β β ^ o r t N T } where t N T is a positive sequence. For any β Θ N T Θ , we have
L ˜ ( β ) L ˜ ( β ) = L N T ( β ) L N T ( β ) + H N T ( β ) H N T ( β ) I 1 + I 2 .
By using Taylor expansion, we have
I 1 = L N T ( β ) L N T ( β ) = ( y ˜ x ˜ β ˘ ) T x ˜ ( β β ) , I 2 = H N T ( β ) H N T ( β ) ,
where β ˘ = κ β + ( 1 κ ) β lies between β and β for κ ( 0 , 1 ) .
For ( I 2 ) , we have
I 2 = λ { j > i } ρ ( β ˘ i β ˘ j ) β ˘ i β ˘ j 1 ( β ˘ i β ˘ j ) T ( β i β i ) + λ { j < i } ρ ( β ˘ i β ˘ j ) β ˘ i β ˘ j 1 ( β ˘ i β ˘ j ) T ( β i β i ) = λ { j > i } ρ ( β ˘ i β ˘ j ) β ˘ i β ˘ j 1 ( β ˘ i β ˘ j ) T { ( β i β i ) ( β j β j ) } .
For each i , j G k , then β i = β j and β ˘ i β ˘ j = κ ( β i β j ) . Thus
I 2 = λ k = 1 K { i , j G k , i < j } ρ ( β ˘ i β ˘ j ) β ˘ i β ˘ j 1 ( β ˘ i β ˘ j ) T ( β i β j ) + λ 1 k < m K { i G k , j G m } ρ ( β ˘ i β ˘ j ) β ˘ i β ˘ j 1 ( β ˘ i β ˘ j ) T { ( β i β i ) ( β j β j ) } .
Since sup i β i β i 0 2 = sup k α k α k 0 2 ϕ N T 2 and β ˘ i lies between β i and β i , then
sup i β ˘ i β i 0 κ sup i β i β i 0 + ( 1 κ ) sup i β i β i 0 κ ϕ N T + ( 1 κ ) ϕ N T = ϕ N T .
For any k m , i G k and j G m ,
β ˘ i β i 0 min i G k , j G m β i 0 β j 0 2 max i β ˘ i β i 0 b N T 2 ϕ N T > a λ ,
and hence ρ ( β ˘ i β ˘ j ) = 0 by assumption 2. Since β ˘ i β ˘ j = κ ( β i β j ) ,
I 2 = λ k = 1 K { i , j G k , i < j } ρ ( β ˘ i β ˘ j ) β ˘ i β ˘ j 1 ( β ˘ i β ˘ j ) T ( β i β j ) = λ k = 1 K { i , j G k , i < j } ρ ( β ˘ i β ˘ j ) β i β j .
By using the similar argument, it is easy to show that
sup i β i β ^ i o r = sup k α k α ^ k o r sup i β i β ^ i o r .
Since
sup i β ˘ i β ˘ j 2 sup i β ˘ i β i 2 sup i β i β i 2 ( sup i β i β ^ i o r + sup i β i β ^ i o r ) 4 sup i β i β ^ i o r 4 t N T ,
we have ρ ( β ˘ i β ˘ j ) ρ ( 4 t N T ) due to the concavity of ρ ( · ) . So
I 2 k = 1 K { i , j G k , i < j } λ ρ ( 4 t N T ) β i β j .
For ( I 1 ) , we let
B = ( B 1 , , B N ) T = [ ( y ˜ x ˜ β ˘ ) T x ˜ ] T
where B i = ( y ˜ i x ˜ i T β i ) T x ˜ i . Hence
I 1 = B T ( β β ) = k = 1 K { i , j G k } N k 1 B i T ( β i β j ) = k = 1 K { i , j G k } B i T ( β i β j ) 2 N k k = 1 K { i , j G k } B i T ( β i β j ) 2 N k = k = 1 K { i , j G k } ( B j B i ) T ( β j β i ) 2 N k = k = 1 K { i , j G k , i < j } N k 1 ( B j B i ) T ( β j β i ) .
Note that sup i B i sup i x ˜ i T ϵ i + x ˜ i · β i 0 β ˘ i . By assumption 1, we have sup i x ˜ i T ϵ i c 1 T p and sup i x ˜ i 2 c 3 . From (A11), we have sup i β i 0 β ˘ i ϕ N T . Hence there exists an event E 2 such that P ( E 2 c ) = o ( T 1 ) and over the set E 2 , sup i B i ( c 1 T p + 2 c 3 ϕ N T ) . Then
N k 1 ( B j B i ) T ( β j β i ) N k 1 B j B i · β i β j N k 1 2 sup i B i · β i β j 2 N k 1 ( c 1 T p + 2 c 3 ϕ N T ) β i β j .
Combining together (A12)–(A14), we have
L ˜ ( β ) L ˜ ( β ) k = 1 K { i , j G k , i < j } { λ ρ ( 4 t N T ) 2 N k 1 ( c 1 T p + 2 c 3 ϕ N T ) } β i β j .
By choosing sufficient small t N T , we have ρ ( 4 t N T ) 1 and 2 N k 1 ( c 1 T p + 2 c 3 ϕ N T ) λ . Therefore, L ˜ ( β ) L ˜ ( β ) 0 for any β Θ N T . The result in (2) is now completed.
Step 3: From Lemma A2, as  T and σ 0 2 0 , we have P ( K ^ = K ) 1 and P ( G ^ 1 = G 1 0 , , G ^ K = G k 0 | K ^ = K ) 1 . By conditional probability formula, it yields that
P ( G ^ 1 = G 1 0 , , G ^ K = G k 0 ) = P ( G ^ 1 = G 1 0 , , G ^ K = G K 0 | K ^ = K ) P ( K ^ = K ) 1 ,
as T . Next, let B be a Borel-measureable set, then
P N k T ( α ^ k α k 0 ) B = P N k T ( α ^ k α k 0 ) B | β ^ = β ^ o r P β ^ = β ^ o r + P N k T ( α ^ k α k 0 ) B | β ^ β ^ o r P β ^ β ^ o r = P N k T ( α ^ k α k 0 ) B { 1 o ( 1 ) } + o ( 1 )   P N k T ( α ^ k α k 0 ) B as   T .
This means that N k T ( α ^ k α k 0 ) shares a similar asymptotic distribution with N k T ( α ^ k o r α k 0 ) , where
N k T ( α ^ k o r α k 0 ) = 1 T N k i G k 0 x ˜ i T x ˜ i 1 1 N k T i G k 0 x ˜ i T ϵ i .
By Assumption 3, we have N k T ( α ^ k o r α k 0 ) Ω k 1 B k d N ( 0 , Ω k 1 Ψ k Ω k 1 ) . Thus the results in Theorem 1 is obtained.   □

Appendix B.3

Proof of Theorem 2. 
For a given tuning parameter γ , denote the corresponding parameters estimator as β ^ ( γ ) and the mean square error as σ ^ G ^ ( γ ) 2 . Then
σ ^ N T 2 ( γ ) = 2 L N T ( β ^ ( γ ) ) [ I { G ^ ( γ ) = G 0 } + I { G ^ ( γ ) G 0 } ] = 2 L N T ( β ^ ( γ ) ) I { G ^ ( γ ) = G 0 } + o p ( 1 )   σ 0 2 , as   ( N , T ) .
By the definition, IC ( γ o p t ) = ln σ ^ N T 2 ( γ o p t ) + p K ( γ o p t ) ψ N T , which approaches ln ( σ 0 2 ) as ( N , T ) by Assumption 5. For any γ , we identify two situations:
Case 1: Under-fitted model. Suppose that K ^ = K ^ ( γ ) with 1 K ^ < K . From Assumption 4, we have
IC ( γ ) = ln σ ^ G ^ K ^ 2 + p K ^ ψ N T min 1 K ^ < K min G ^ K ^ ln σ ^ G ^ K 2 + p K ^ ψ N T   ln σ ̲ 2 > ln ( σ 0 2 ) , as   ( N , T ) .
Hence IC ( γ ) > IC ( γ o p t ) as ( N , T ) .
Case 2: Over-fitted model. Consider any other γ such that K ^ = K ^ ( γ ) > K . Following Lemma S1.14 in [5], we can obtain N T ln σ ^ G 0 2 ln σ ^ G ^ K ^ 2 = O p ( 1 ) . Then by Assumption 5, we have
P ( K ^ > K ) = P IC ( γ ) < IC ( γ o p t ) = P N T ln σ ^ G 0 2 ln σ ^ G ^ K ^ 2 > ( K ^ K ) N T ψ N T   0 , as   ( N , T ) .
Combining Case 1 and Case 2 together, the information criterion in (18) can consistently select a optimal tuning parameter γ , which yields an oracle estimator for β with probability approaching 1.    □

Appendix C. Testing for Heterogeneity

It is important to determine whether the heterogeneity effect is statistically significant. The hypothesis of no heterogeneity can be represented by
H 0 : β 1 = β 2 = = β N
against the alternative hypothesis
H 1 : β i β j , for   any   i j
Under the null hypothesis, the model is
y i t = μ i + β 1 T x i t + ϵ i t
By applying the fixed-effect transformation, we have
y ˜ i t = β 1 T x ˜ i t + ϵ ˜ i t
where ϵ ˜ i t = ϵ i t 1 T t = 1 T ϵ i t . The regression parameter β 1 can be estimated by OLS, yielding residuals ϵ ^ ¯ i t , and the sum of squared errors S 0 = ϵ ^ ¯ ϵ ^ ¯ , where ϵ ^ ¯ = ( ϵ ^ ¯ 1 T , , ϵ ^ ¯ N T ) T and ϵ ^ ¯ i = ( ϵ ^ ¯ i 1 , , ϵ ^ ¯ i T ) . The likelihood ratio test of H 0 is based on
F = S 0 S 1 ( K ^ ) δ ^ 2
where S 1 ( K ^ ) is the sum of squared errors using the latent structures model 2, and  K ^ is the estimated number of groups. The residual variance δ ^ 2 is given by
δ ^ 2 = 1 N ( T 1 ) ϵ ^ ϵ ^ = 1 N ( T 1 ) S 1 ( K ^ )
However, the asymptotic distribution of F 1 is non-standard, making it challenging to tabulate critical values directly. To address this issue, we propose a residual-based bootstrap procedure to calculate the p-values. The detailed procedures are outlined in Algorithm A1.
Algorithm A1:  Residual bootstrap procedure
  • Step 1. Calculate the likelihood ratio test statistic F based on the original sample { ( y i t , i t ) , i = 1 , , N ; t = 1 , T }. 
  • Step 2. Obtain the usual fixed effect estimators β ^ ¯ 1 without latent structures specification, and obtain the residual u ^ ¯ i t = ϵ ^ ¯ i t ϵ ^ ¯ i · , where ϵ ^ ¯ i · = T 1 t = 1 T ϵ ^ ¯ i t
  • Step 3. Generate iid draws v i t from i = 1 , , N and t = 1 , , T from N ( 0 , 1 ) , and set ϵ i t = ϵ ^ ¯ i t v i t and y i t = x i t β ^ ¯ 1 + ϵ ^ ¯ i · + ϵ i t
  • Step 4. Calculate the testing statistic F ( 1 ) based on the bootstrap resample { ( y i t , x i t ) , i = 1 , , N ; t = 1 , , T }
  • Step 5. Repeat the Steps 3–4 B times and obtain a lis of F denoted by { F ( 1 ) , , F ( B ) } . The p-value is the proportion of F exceeding F ( b ) .  

References

  1. Hsiao, C. Analysis of Panel Data; Number 54; Cambridge University Press: Cambridge, UK, 2014. [Google Scholar]
  2. Bonhomme, S.; Manresa, E. Grouped patterns of heterogeneity in panel data. Econometrica 2015, 83, 1147–1184. [Google Scholar] [CrossRef]
  3. Su, L.; Chen, Q. Testing homogeneity in panel data models with interactive fixed effects. Econom. Theory 2013, 29, 1079–1135. [Google Scholar] [CrossRef]
  4. Phillips, P.C.; Sul, D. Transition modeling and econometric convergence tests. Econometrica 2007, 75, 1771–1855. [Google Scholar] [CrossRef]
  5. Su, L.; Shi, Z.; Phillips, P.C. Identifying latent structures in panel data. Econometrica 2016, 84, 2215–2264. [Google Scholar] [CrossRef]
  6. Ke, Z.T.; Fan, J.; Wu, Y. Homogeneity pursuit. J. Am. Stat. Assoc. 2015, 110, 175–194. [Google Scholar] [CrossRef]
  7. Xiao, D.; Ke, Y.; Li, R. Homogeneity structure learning in large-scale panel data with heavy-tailed errors. J. Mach. Learn. Res. 2021, 22, 1–42. [Google Scholar]
  8. Zhang, Y.; Wang, H.J.; Zhu, Z. Quantile-regression-based clustering for panel data. J. Econom. 2019, 213, 54–67. [Google Scholar] [CrossRef]
  9. Fan, J.; Li, R. Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  10. Tibshirani, R. Regression shrinkage and selection via the lasso. J. R. Stat. Soc. Ser. Stat. Methodol. 1996, 58, 267–288. [Google Scholar] [CrossRef]
  11. Huang, W.; Jin, S.; Su, L. Identifying latent grouped patterns in cointegrated panels. Econom. Theory 2020, 36, 410–456. [Google Scholar] [CrossRef]
  12. Huang, W.; Su, L.; Zhuang, Y. Detecting unobserved heterogeneity in efficient prices via classifier-lasso. J. Bus. Econ. Stat. 2023, 41, 509–522. [Google Scholar] [CrossRef]
  13. Su, L.; Ju, G. Identifying latent grouped patterns in panel data models with interactive fixed effects. J. Econom. 2018, 206, 554–573. [Google Scholar] [CrossRef]
  14. Su, L.; Wang, X.; Jin, S. Sieve estimation of time-varying panel data models with latent structures. J. Bus. Econ. Stat. 2019, 37, 334–349. [Google Scholar] [CrossRef]
  15. Mehrabani, A. Estimation and identification of latent group structures in panel data. J. Econom. 2023, 235, 1464–1482. [Google Scholar] [CrossRef]
  16. Ando, T.; Bai, J. Panel data models with grouped factor structure under unknown group membership. J. Appl. Econom. 2016, 31, 163–191. [Google Scholar] [CrossRef]
  17. Liu, R.; Shang, Z.; Zhang, Y.; Zhou, Q. Identification and estimation in panel models with overspecified number of groups. J. Econom. 2020, 215, 574–590. [Google Scholar] [CrossRef]
  18. Bai, J. Estimating multiple breaks one at a time. Econom. Theory 1997, 13, 315–352. [Google Scholar] [CrossRef]
  19. Wang, W.; Su, L. Identifying latent group structures in nonlinear panels. J. Econom. 2021, 220, 272–295. [Google Scholar] [CrossRef]
  20. Su, L.; Wang, W.; Xu, X. Identifying latent group structures in spatial dynamic panels. J. Econom. 2023, 235, 1955–1980. [Google Scholar] [CrossRef]
  21. Wang, Y.; Phillips, P.C.; Su, L. Panel data models with time-varying latent group structures. J. Econom. 2024, 240, 105685. [Google Scholar] [CrossRef]
  22. Beer, J.C.; Aizenstein, H.J.; Anderson, S.J.; Krafty, R.T. Incorporating prior information with fused sparse group lasso: Application to prediction of clinical measures from neuroimages. Biometrics 2019, 75, 1299–1309. [Google Scholar] [CrossRef] [PubMed]
  23. Li, Y.; Jin, B. Pairwise fusion approach incorporating prior constraint information. Commun. Math. Stat. 2020, 8, 47–62. [Google Scholar] [CrossRef]
  24. Li, F.; Sang, H. Spatial homogeneity pursuit of regression coefficients for large datasets. J. Am. Stat. Assoc. 2019, 114, 1050–1062. [Google Scholar] [CrossRef]
  25. Zhang, X.; Liu, J.; Zhu, Z. Learning coefficient heterogeneity over networks: A distributed spanning-tree-based fused-lasso regression. J. Am. Stat. Assoc. 2024, 119, 485–497. [Google Scholar] [CrossRef]
  26. Wang, W.; Phillips, P.C.; Su, L. Homogeneity pursuit in panel data models: Theory and application. J. Appl. Econom. 2018, 33, 797–815. [Google Scholar] [CrossRef]
  27. Boyd, S.; Parikh, N.; Chu, E. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers; Now Publishers Inc.: Delft, The Netherlands, 2011. [Google Scholar]
  28. Ke, Y.; Li, J.; Zhang, W. Structure identification in panel data analysis. Ann. Stat. 2016, 44, 1193–1233. [Google Scholar] [CrossRef]
  29. Ma, S.; Huang, J. A concave pairwise fusion approach to subgroup analysis. J. Am. Stat. Assoc. 2017, 112, 410–423. [Google Scholar] [CrossRef]
  30. Jeon, J.-J.; Kwon, S.; Choi, H. Homogeneity detection for the high-dimensional generalized linear model. Comput. Stat. Data Anal. 2017, 114, 61–74. [Google Scholar] [CrossRef]
  31. Zhang, C.-H. Nearly unbiased variable selection under minimax concave penalty. Ann. Stat. 2010, 38, 894–942. [Google Scholar] [CrossRef]
  32. Belloni, A.; Chernozhukov, V. Least squares after model selection in high-dimensional sparse models. Bernoulli 2013, 19, 521–547. [Google Scholar] [CrossRef]
  33. Shiu, A.; Lam, P.-L. Electricity consumption and economic growth in china. Energy Policy 2004, 32, 47–54. [Google Scholar] [CrossRef]
  34. Xu, S.-C.; He, Z.-X.; Long, R.-Y. Factors that influence carbon emissions due to energy consumption in china: Decomposition analysis using lmdi. Appl. Energy 2014, 127, 182–193. [Google Scholar] [CrossRef]
  35. Akkemik, K.A.; Göksal, K.; Li, J. Energy consumption and income in chinese provinces: Heterogeneous panel causality analysis. Appl. Energy 2012, 99, 445–454. [Google Scholar] [CrossRef]
  36. Wang, N.; Fu, X.; Wang, S. Economic growth, electricity consumption, and urbanization in china: A tri-variate investigation using panel data modeling from a regional disparity perspective. J. Clean. Prod. 2021, 318, 128529. [Google Scholar] [CrossRef]
  37. Pesaran, M.H. Estimation and inference in large heterogeneous panels with a multifactor error structure. Econometrica 2006, 74, 967–1012. [Google Scholar] [CrossRef]
  38. Miao, K.; Li, K.; Su, L. Panel threshold models with interactive fixed effects. J. Econom. 2020, 219, 137–170. [Google Scholar] [CrossRef]
  39. Tseng, P. Convergence of a block coordinate descent method for nondifferentiable minimization. J. Optim. Theory Appl. 2001, 109, 475–494. [Google Scholar] [CrossRef]
Figure 1. The normalized mutual information across different N and T for DGPs 1–3.
Figure 1. The normalized mutual information across different N and T for DGPs 1–3.
Mathematics 13 01505 g001
Figure 2. The average model estimation errors across different N and T for DGPs 1–3.
Figure 2. The average model estimation errors across different N and T for DGPs 1–3.
Mathematics 13 01505 g002
Table 1. The notations in our working model.
Table 1. The notations in our working model.
NotationIllustration
Observed quantities y i t the univariate response variable
x i t the p × 1 explanatory variables
C the prior constraint information set
Nthe number of all individuals
Tthe number of observations for each individual
Unknown parameters μ i the fixed effect of the ith individual
β i the slope of the ith individual
G the latent group structures
G k the membership of the kth group
α k the common slope of the kth group
Kthe number of latent groups
Table 2. Comparison with other literature in latent structures identification for panel models.
Table 2. Comparison with other literature in latent structures identification for panel models.
MethodsPenalty FormsPrior Information
C-Lasso [5] i = 1 N k = 1 K β i α k No prior information
Panel-CARDS [26] l = 1 L 1 i B l , j B l + 1 p γ 1 ( β i β j ) + An ordered segmentation
l = 1 L i , j B l p γ 2 ( β i β j ) { B 1 , , B L }
PAGFL [15] 1 i < j N w ¨ i j β i β j Adaptive weights w ¨ i j ’s
Our work 1 i < j N p γ ( β i β j 1 ) + I C ( β ) A convex set C
Table 3. The parameter estimation results of α 11 for DGPs 1-3 based on N = 200 .
Table 3. The parameter estimation results of α 11 for DGPs 1-3 based on N = 200 .
DGP No-Prior [5]PriorOracle
T = 10 T = 20 T = 40 T = 10 T = 20 T = 40 T = 10 T = 20 T = 40
1Bias−0.025−0.008−0.002−0.007−0.004−0.0030.010−0.001−0.002
SD0.0880.0420.0290.0840.0410.0290.0540.0370.029
ESE0.0600.0410.0290.0620.0420.0290.0580.0410.029
ECP0.8300.9200.9700.8500.9500.9600.9800.9600.970
2Bias−0.011−0.015−0.0010.008−0.004−0.001−0.004−0.0070.000
SD0.0860.0400.0270.0890.0420.0270.0620.0400.027
ESE0.0600.0420.0290.0610.0420.0290.0580.0410.029
ECP0.7800.9800.9600.8300.9500.9600.9400.9800.970
3Bias0.0050.0030.0030.0180.0050.0040.0050.0040.003
SD0.0780.0420.0310.0770.0430.0320.0600.0420.031
ESE0.0660.0430.0290.0650.0430.0290.0580.0410.029
ECP0.9100.9700.9300.9200.9600.9200.9400.9500.930
Bias is empirical bias; SD is the Monte Carlo standard deviation; ESE is the estimated standard error and ECP is the empirical coverage probability of 95% Wald confidence interval.
Table 4. Classification results among 30 provinces in China.
Table 4. Classification results among 30 provinces in China.
With Prior InformationWithout Prior Information
Group 1Group 2Group 1Group 2Group 3Group 4Group 5
1.1692 3.4770 0.6935 0.8715 0.98371.08793.4351
AnhuiBeijingGuangdongShandongZhejiangAnhuiBeijing
FujianGansuJiangsu FujianGansu
GuangdongGuizhou GuangxiGuizhou
GuangxiHeilongjiang HainanHeilongjiang
HainanJilin HebeiJilin
HebeiJiangxi HenanJiangxi
HenanNingxia HubeiNingxia
HubeiQinghai HunanQinghai
HunanShannxi LiaoningShannxi
JiangsuTianjin Inner MongoliaTianjin
LiaoningChongqing ShanxiChongqing
Inner Mongolia Shanghai
Shandong Sichuan
Shanxi Xinjiang
Shanghai Yunan
Sichuan
Xinjiang
Yunnan
Zhejiang
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Y.; Luo, X.; Liao, M. Incorporating Prior Information in Latent Structures Identification for Panel Data Models. Mathematics 2025, 13, 1505. https://doi.org/10.3390/math13091505

AMA Style

Li Y, Luo X, Liao M. Incorporating Prior Information in Latent Structures Identification for Panel Data Models. Mathematics. 2025; 13(9):1505. https://doi.org/10.3390/math13091505

Chicago/Turabian Style

Li, Yi, Xingxing Luo, and Mengqi Liao. 2025. "Incorporating Prior Information in Latent Structures Identification for Panel Data Models" Mathematics 13, no. 9: 1505. https://doi.org/10.3390/math13091505

APA Style

Li, Y., Luo, X., & Liao, M. (2025). Incorporating Prior Information in Latent Structures Identification for Panel Data Models. Mathematics, 13(9), 1505. https://doi.org/10.3390/math13091505

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop