Next Article in Journal
Impact of Digital Transformation on the Energy Sector: A Review
Previous Article in Journal
Detection of Plausibility and Error Reasons in Finite Element Simulations with Deep Learning Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Model-Robust Estimation of Multiple-Group Structural Equation Models

by
Alexander Robitzsch
1,2
1
IPN—Leibniz Institute for Science and Mathematics Education, Olshausenstraße 62, 24118 Kiel, Germany
2
Centre for International Student Assessment (ZIB), Olshausenstraße 62, 24118 Kiel, Germany
Algorithms 2023, 16(4), 210; https://doi.org/10.3390/a16040210
Submission received: 11 February 2023 / Revised: 2 April 2023 / Accepted: 15 April 2023 / Published: 17 April 2023
(This article belongs to the Special Issue Statistical learning and Its Applications)

Abstract

:
Structural equation models (SEM) are widely used in the social sciences. They model the relationships between latent variables in structural models, while defining the latent variables by observed variables in measurement models. Frequently, it is of interest to compare particular parameters in an SEM as a function of a discrete grouping variable. Multiple-group SEM is employed to compare structural relationships between groups. In this article, estimation approaches for the multiple-group are reviewed. We focus on comparing different estimation strategies in the presence of local model misspecifications (i.e., model errors). In detail, maximum likelihood and weighted least-squares estimation approaches are compared with a newly proposed robust L p loss function and regularized maximum likelihood estimation. The latter methods are referred to as model-robust estimators because they show some resistance to model errors. In particular, we focus on the performance of the different estimators in the presence of unmodelled residual error correlations and measurement noninvariance (i.e., group-specific item intercepts). The performance of the different estimators is compared in two simulation studies and an empirical example. It turned out that the robust loss function approach is computationally much less demanding than regularized maximum likelihood estimation but resulted in similar statistical performance.

1. Introduction

Confirmatory factor analysis (CFA) and structural equation models (SEM) are amongst the most important statistical approaches for analyzing multivariate data in the social sciences [1,2,3,4,5,6,7]. These models relate a multivariate vector x = ( X 1 , , X I ) of observed I variables (also referred to as items) to a vector of latent variables (i.e., factors) η of lower dimension using a linear model. SEMs represent the mean vector μ and the covariance matrix Σ of the random variable x as a function of an unknown parameter vector θ . That is, the mean vector is represented as μ ( θ ) , and the covariance matrix is given by Σ ( θ ) .
SEM and CFA, as particular cases, impose a measurement model that relates the observed variables x to latent variables η
x = ν + Λ η + ϵ .
In addition, we denote the covariance matrix Var ( ϵ ) = Ψ , and η as well as ϵ are multivariate normally distributed random vectors. Note that η and ϵ are assumed to be uncorrelated. In CFA, the multivariate normal (MVN) distribution is represented as η MVN ( α , Φ ) and ϵ MVN ( 0 , Ψ ) . Hence, one can represent the mean and the covariance matrix in CFA as
μ ( θ ) = ν + Λ α and Σ ( θ ) = Λ Φ Λ + Ψ .
In SEM, more structured relationships among the latent variables can be imposed. A matrix B of regression coefficients is specified such that
η = B η + ξ with E ( ξ ) = α and Var ( ξ ) = Φ .
Hence, the mean vector and the covariance matrix are represented in SEM as
μ ( θ ) = ν + Λ ( I B ) 1 α and Σ ( θ ) = Λ ( I B ) 1 Φ [ ( I B ) 1 ] Λ + Ψ ,
where I denotes the identity matrix.
In practice, SEM parsimoniously parametrizes the mean vector and the covariance matrix using a parameter θ as a summary. As is true with all statistical models, these restrictions are unlikely to hold in practice, and model assumptions in SEM are only an approximation of a true data-generating model. In SEM, model deviations (i.e., model errors) in covariances emerge as a difference between a population covariance matrix Σ and a model-implied covariance matrix Σ ( θ ) (see [8,9]). Similarly, there can be model errors in the mean vector leading to a difference between the population mean vector μ and the model-implied mean vector μ ( θ ) . As a consequence, the SEM is misspecified at the population level. Note that the model errors are defined at the population level in infinite sample sizes. With finite samples in real-data applications, the empirical covariance matrix S estimates the population covariance matrix Σ and the mean vector x ¯ estimates the population mean vector μ . In this article, we investigate estimators that possess some kind of resistance against model deviations. That is, the presence of some amount of model errors does not impact the parameter estimate θ . This kind of robustness is referred to as model robustness and follows the principles of robust statistics [10,11,12]. While in classical robust statistics, observations (i.e., cases or subjects) do not follow an imposed statistical model and should be treated as outliers, model errors in SEM occur as residuals in the modeled mean vector and the modeled covariance matrix. That is, an estimator in an SEM should automatically treat large deviations in μ μ ( θ ) and Σ Σ ( θ ) as outliers that should not (substantially) impact the estimated parameter θ . We compare regularized maximum likelihood estimation with robust moment estimation approaches. Robust moment estimation has been scarcely used in SEM. We propose an alternative and more flexible loss function for robust moment estimation and demonstrate that it results in similar statistical performance compared to the computationally much more tedious maximum likelihood estimation.
In this article, we discuss model-robust estimation of SEM when modeling means and covariance matrices in the more general case of multiple groups [4,13]. Model errors in the modeled covariance matrix can emerge to incorrectly specified loadings in the matrix Λ or unmodeled residual error correlations in the matrix Ψ . In the following, we only consider the case of unmodeled residual error correlations. Moreover, model errors in the modeled mean vectors are mainly due to incorrectly specified item intercepts ν . In the multiple-group SEM, this case is often referred to as a violation of measurement invariance [14,15,16]. The investigation of model-robust SEM estimators under such a measurement noninvariance situation is also the focus of this article. For discrete items, measurement noninvariance is also referred to as differential item functioning (DIF; [17,18]), and several approaches for model-robust estimation have been proposed (e.g., [19,20,21]).
The remainder of the article is organized as follows: Different model-robust estimation methods are discussed in Section 2. Section 3 presents results from a simulation study with unmodeled residual error correlations. In Section 4, we focus on standard error estimation methods in a selected number of conditions of the simulation study presented in Section 3Section 5 presents results from a simulation study with unmodeled item intercepts which indicates the presence of violations of measurement invariance (i.e., the presence of DIF). In Section 6, results from an empirical example involving data from the European social survey are offered. Finally, the article closes with a discussion in Section 7.

2. Model-Robust Estimation in Structural Equation Modeling

In Section 2.1, we review the most frequently used estimation methods (i.e., maximum likelihood and weighted least squares estimation) in multiple-group SEM. These methods are known not to be resistant to model errors and are, therefore, referred to as model-non-robust estimation methods. Section 2.2 introduces robust moment estimation of multiple-group SEMs that is robust to the presence of model errors. As an alternative, Section 2.3 discusses regularized maximum likelihood estimation as a robust-estimation method.

2.1. Multiple-Group Structural Equation Modeling

We now describe different estimation approaches for multiple-group SEMs. Note that some identification constraints must be imposed to estimate the covariance structure model (4) (see [2]). For modeling multivariate normally distributed data without any missing values, the empirical mean vector x ¯ and the empirical covariance matrix S are sufficient statistics for estimating μ and Σ . Hence, they are also sufficient statistics for the parameter vector θ .
Now, assume that there are G groups with sample sizes N g and empirical means x ¯ g and covariance matrices S g ( g = 1 , , G ). The population mean vectors are denoted by μ g , and the population covariance matrices are denoted by Σ g . The model-implied mean vector shall be denoted by μ g ( θ ) and the model-implied covariance matrix by Σ g ( θ ) . Note that the parameter vector θ does not have an index g to indicate that there can be common and unique parameters across groups. As a typical application of a CFA, equal factor loadings and item intercepts across groups are imposed (i.e., measurement invariance holds) by assuming the same loading matrix Λ and the same intercept vector ν across groups.
The maximum likelihood (ML) function for the parameter θ is given by (see [2,4])
F ML ( θ ; { x ¯ g } , { S g } ) = g = 1 G N g 2 I log ( 2 π ) + log | Σ g ( θ ) | + tr ( S g Σ g ( θ ) 1 ) + ( x ¯ g μ g ( θ ) ) Σ g ( θ ) 1 ( x ¯ g μ g ( θ ) ) ,
  where { x ¯ g } and { S g } are the sets of the empirical mean vectors and empirical covariance matrices of all groups, respectively. In practice, the model-implied covariance matrix will be misspecified [22,23,24], and θ is a pseudo-true parameter defined as the maximizer of the fitting function F ML in (5).
A more general class of fitting functions is weighted least squares (WLS) estimation [3,4,25]. The parameter vector θ is determined as the minimizer of
F WLS ( θ ; { x ¯ g } , { S g } ) = g = 1 G ( x ¯ g μ g ( θ ) ) W 1 g ( x ¯ g μ g ( θ ) ) + ( S g σ g ( θ ) ) W 2 g ( S g σ g ( θ ) ) ,
where matrices Σ and S have been replaced by vectors σ and S that collect all nonduplicated elements of the matrices in vectors. The weight matrices W 1 g and W 2 g ( g = 1 , , G ) can also depend on parameters that must be estimated. Diagonally weighted least squares (DWLS) estimation involves choosing diagonal weight matrices W g 1 and W g 2 . The unweighted least squares (ULS) estimation is obtained when these matrices are identity matrices. Interestingly, the minimization in (6) can be interpreted as a nonlinear least squares estimation problem with sufficient statistics { x ¯ g } and { S g } as input data [26].
It has been shown that ML estimation can be approximately written as DWLS estimation [27] with particular weight matrices. DWLS can be generally written as
F DWLS ( θ ; { x ¯ g } , { S g } ) = g = 1 G i = 1 I w 1 g i ( x ¯ g i μ g i ( θ ) ) 2 + g = 1 G i = 1 I j = i I w 2 g i j ( s g i j σ g i j ( θ ) ) 2 ,
  where w 1 g i and w 2 g i j are appropriate elements in W 1 g and W 2 g , respectively. In ML estimation, the weights are approximately determined by w 1 g i = 1 / u g i 2 and w 2 g i j = 1 / ( u g i 2 u g j 2 ) , where u g i 2 are sample unique standardized variances with u g i 2 = ψ g i i / σ g i i (see [27]). With smaller residual variances ψ g i i , more trust is put on a mean μ g i or a covariance σ g i j in the fitting function in ML estimation.

2.2. Robust Moment Estimation Using Robust Loss Functions

It is evident that the DWLS fitting function is not robust to outlying observations because the square loss function is used. In model-robust SEM estimation in the context of this article, outlying observations are defined as deviations μ g μ g ( θ ) and Σ g Σ g ( θ ) . In order to allow resistance of the estimator against model deviations, the square loss function in (7) can be replaced with a robust loss function ρ (see [13])
F rob ( θ ; { x ¯ g } , { S g } ) = g = 1 G i = 1 I w 1 g i ρ ( x ¯ g i μ g i ( θ ) ) + g = 1 G i = 1 I j = i I w 2 g i j ρ ( s g i j σ g i j ( θ ) ) .
The robust mean absolute deviation (MAD) loss function ρ ( x ) = | x | was considered in [28,29]. This fitting function is more robust to a few model violations, such as unmodeled item intercepts or unmodeled residual correlations of residuals ε i (see [8]). In this article, we consider the more general L p loss function ρ ( x ) = | x | p with 0 < p 1 can ensure even more model-robust estimates [30]. This estimation approach is referred to as robust moment estimation (RME). The L p loss function with p = 2 is the square loss function ρ ( x ) = x 2 and corresponds to ULS estimation.
The L p loss function has been successively applied in linking methods [13,31], which can be considered as an alternative approach to a joint estimation of SEM for multiple groups. The loss function with p = 0.5 or p = 0.25 has been proposed in the invariance alignment linking approach [32,33,34].
In the minimization of (8), the nondifferentiable loss function ρ is substituted by a differentiable approximation (e.g., ρ ( x ) = | x | is replaced by ρ ˜ ε ( x ) = ( x 2 + ε ) 1 / 2 for a small ε > 0 , such as ε = 10 3 ; see [30,32,35,36]). In practice, it is advisable to use reasonable starting values and to minimize (8) using a sequence of differentiable approximations with decreasing ε values (i.e., subsequently fitting ρ ˜ ε with ε = 10 1 , 10 2 , 10 3 and using the previously obtained result as the initial value for the subsequent minimization problem).

2.2.1. Bias Derivation in the Presence of Model Errors

In this section, we formally study the bias in the estimated parameter θ in the presence of model errors if the robust fitting function F rob in (8) is utilized. General asymptotic results for misspecified models were presented in [6,37].
In the following, we consider the slightly more specific minimization problem
F rob ( θ ; ζ ) = h = 1 H ρ ( ζ h ζ h ( θ ) ) ,
where ζ = ( ζ 1 , , ζ 0 , 0 , 0 H ) is a vector that can contain all group-wise mean vectors m g and covariance matrices S g for g = 1 , , G . Let vech ( S ) denote the nonduplicated elements in a covariance matrix S . Then, the vector ζ contains all sufficient statistics ζ = ( m 1 , , m G , S 1 , , S G ) , where S g = vech ( S g ) for g = 1 , , G .
We study the estimation problem using the optimization function in (9) at the population level. That is, we study the model differences between ζ and ζ ( θ ) . According to [26], we can frame the estimation problem (9) as a nonlinear regression
ζ = ζ ( θ ) + r
for some model error r = ( r 1 , , r H ) . We now study the impact of sufficiently small model errors r . It is assumed that there exists a perfectly fitting SEM. That is, there exists a parameter vector θ 0 such that ζ 0 = ζ ( θ 0 ) = ( ζ 01 , , ζ 0 H ) . This situation corresponds to the absence of model error.
We now replace the minimization problem (9) using a robust loss function with a weighted least squares estimation problem. This approach is ordinarily employed in robust statistics when estimating a robust regression model with iteratively weighted least squares estimation [38,39]. Define Δ θ = θ θ 0 and r = ζ ζ ( θ 0 ) . Then, we can rewrite the minimization problem (9) as
F rob ( Δ θ ; ζ ) = h = 1 H ρ ( ζ 0 h + r h ζ h ( θ ) ) .
The nonlinear function ζ h ( θ ) can be linearized around θ = θ 0 . Hence, we can write
ζ h ( θ ) ζ 0 h + a h Δ θ ,
where a h = ( ζ h ) / ( θ ) evaluated at θ 0 . Moreover, we approximate the robust estimation problem in (11) by a weighted least squares problem [38,40]
F rob ( Δ θ ; ζ ) = h = 1 H ω h 2 ( r h a h Δ θ ) 2
using ω h = ρ ( e h ) / e h for e h = r h a h Δ θ . Then, the change in the θ parameter estimate can be directly expressed as a function of model deviations r using the weighted least squares regression formula [38]. Let Ω be the diagonal matrix containing weights ω h in the diagonal. Let A be the design matrix that contains rows a h for h = 1 , , H . Then, the asymptotic bias in the estimated parameter vector is given by
Δ θ = ( A Ω A ) 1 A Ω r .
If a robust loss function ρ is chosen, gross model errors r h result in small weights ω h . Hence, these model errors should only have a minor impact with respect to the parameter estimate  θ . Note that a similar formula like (14) has been presented in [37] for a general differentiable discrepancy function.

2.2.2. Standard Error Estimation

In this subsection, standard error formulas for RME are derived. Note that the parameter estimate θ ^ is a nonlinear function of the vector of sufficient statistics ζ that contains group-wise mean vectors m g and nonduplicated elements of group-wise covariance matrices S g = vech ( S g ) . The covariance matrix V ζ = Var ( ζ ) is derived under multivariate normality of observations x . The vectors m g and S g are uncorrelated within group g and across groups g = 1 , , G . Hence, V ζ can be written as a block-diagonal matrix of covariance matrices Var ( m g ) and Var ( S g ) . If N g is the sample size of group g, the covariance matrix of the mean vector is estimated by
Var ( m g ) = N g 1 S g .
Let vec ( S ) be the vectorized matrix of the covariance matrix S . Then, there exists a transition matrix K that contains appropriate entries 0, 0.5, and 1 such that (see [3])
S g = vech ( S g ) = K vec ( S g ) .
If ⊗ denotes the Kronecker product, the covariance matrix of S g is estimated by (see [3])
Var ( S g ) = N g 1 K S g S g K .
Hence, V ζ is given as a block-diagonal matrix with block-wise matrix entries defined by (15) and (17).
RME can be viewed as obtaining the parameter estimate θ ^ by minimizing a (possibly robust) discrepancy and (approximated) differentiable function F ( θ , ζ ) . Denote by F θ = ( F ) / ( θ ) the vector of partial derivatives with respect to θ . The parameter estimate is given as the root of the nonlinear equation
F θ ( θ ^ , ζ ) = 0 .
Assume that the population sufficient statistics are denoted by ζ 0 and there exists a pseudo-true parameter θ 0 such that F θ ( θ 0 , ζ 0 ) = 0 . Note that the parameter ζ 0 does not refer to a data-generating parameter but is defined by choosing a particular discrepancy function F. Different pseudo-true parameters will be obtained for different choices of discrepancy functions in misspecified SEMs.
We now derive the covariance matrix of θ ^ by utilizing a Taylor expansion of F θ . Denote by F θ θ and F θ ζ the matrices of second-order partial derivatives of F θ with respect to θ and ζ , respectively. We obtain
F θ ( θ ^ , ζ ) = F θ ( θ 0 , ζ 0 ) + F θ θ ( θ 0 , ζ 0 ) ( θ ^ θ 0 ) + F θ ζ ( θ 0 , ζ 0 ) ( ζ ζ 0 ) .
As the parameter estimate θ ^ is a nonlinear function of ζ , the Taylor expansion (19) provides the approximation
θ ^ θ 0 = F θ θ ( θ 0 , ζ 0 ) 1 F θ ζ ( θ 0 , ζ 0 ) ( ζ ζ 0 ) .
By defining a = F θ θ ( θ , ζ 0 ) 1 F θ ζ ( θ , ζ 0 ) , we get, using the multivariate delta formula [22,41]
Var ( θ ^ ) = A V ζ A
This approach is ordinarily used for differentiable discrepancy functions in the SEM literature [3,7,25,37]. We also apply the Formula (21) for differentiable approximations ρ ˜ ε ( x ) = ( x 2 + ε ) p / 2 of the robust L p loss function ρ ( x ) = | x | p for p 1 and a sufficiently small ε > 0 in RME.
As an alternative to the multivariate delta formula, standard errors can be obtained with resampling approaches, such as bootstrap or jackknife [42]. However, these approaches are computationally more demanding because SEM estimation has to be carried out in each replication sample.

2.3. Regularized Maximum Likelihood Estimation

Regularized estimation of SEMs might be a more direct approach for modeling misspecifications in means and covariances [43,44,45,46,47,48,49,50]. In this case, model errors are represented as outliers in item intercepts (i.e., in the vector ν ) and nondiagonal entries in the residual error covariance (i.e., in the matrix Ψ ).
In regularized estimation, a nonidentified SEM is typically utilized, and only outlying entries in ν and Ψ are estimated differently from zero. To ensure the identifiability of model parameters, regularized estimation relies on sparsity assumptions [51,52,53] in item intercepts and residual correlations. The main idea in regularized ML (RegML) estimation is that a penalty function is subtracted from the likelihood. The penalty function P controls the sparsity in a subset of the estimated parameter vector θ . For example, the subset in θ might refer to nondiagonal residual error correlations or group-specific item intercepts.
We now discuss the choice of the penalty function. For a scalar parameter x, the lasso penalty is a popular penalty function used in regularization [52], and it is defined as
P Lasso ( x , κ ) = κ | x | ,
where κ is a nonnegative regularization parameter that controls the extent of sparsity in the obtained parameter estimate. It is known that the lasso penalty introduces bias in estimated parameters. To circumvent this issue, the smoothly clipped absolute deviation (SCAD; [54]) penalty has been proposed. The SCAD penalty is defined as
P SCAD ( x , κ ) = κ | x | if | x | < κ ( x 2 2 a κ | x | 2 + κ 2 ) ( 2 ( a 1 ) ) 1 if κ | x | a κ ( a + 1 ) κ 2 if | x | > a κ
with a > 2 . In many studies, the recommended value of a = 3.7 (see [54]) has been adopted (e.g., [51,55,56]). The SCAD penalty retains the penalization rate and the induced bias of the lasso for model parameters close to zero, but continuously relaxes the rate of penalization as the absolute value of the model parameters increases. Note that P SCAD has the property of the lasso penalty around zero, but has zero derivatives for x values strongly differing from zero.
We now describe RegML estimation in more detail. In the general multiple-group SEM, the parameter vector θ is extended to include group-specific residual error correlations ψ g (which is the vectorized version of Ψ g ) and group-specific item intercepts ν g ( g = 1 , , G ). By including all effects in the SEM, the parameter θ would not be identified. Therefore, a penalty function is subtracted from the log-likelihood function, and the following function is minimized
F ML ( θ ; { x ¯ g } , { S g } ) N * g = 1 G h P SCAD ( ν g h , κ ν ) N * g = 1 G h P SCAD ( ψ g h , κ ψ )
using a sample size N * that can be equal to g = 1 G N g . Note that, for brevity, we suppress to indicate the dimensionality of the regularized parameters ψ g and ν g . In addition, note that there are two regularization parameters κ ν and κ ψ .
In practice, minimization of (24) for fixed values of the two κ parameters results in a subset of ν g and ψ g parameters that are different from zero, where the remaining parameters have been set to zero. Typically, the two κ regularization parameters are unknown nuisance parameters in (24) that must also be estimated. In practice, the minimization of (24) is carried out on a discrete one- or two-dimensional grid of κ values, and the optimal regularization parameter is selected that minimizes the Bayesian information criterion (BIC). The optimization of the nondifferentiable fitting function can be carried out using gradient descent [52] approaches or by substituting the nondifferentiable optimization functions with differentiable approximating functions [8,35,36,57]. In our experience, the latter approach performs satisfactorily in applications.
RegML has the disadvantage that standard errors are difficult to obtain. Of course, resampling methods, such as bootstrap or other demanding approaches, can be applied [52,58,59]. However, this approach is computationally demanding because the tuning parameter selection must be applied in each bootstrap sample. Therefore, avoiding RegML in situations where computation time matters might be preferable.

3. Simulation Study 1: Unmodeled Residual Error Correlation

In this Simulation Study 1, we compare the estimation approaches RME and RegML in a misspecified single-group two-dimensional factor model. It is of interest which of the estimation methods is resistant to model misspecification. It can be expected that RegML and RME with powers p 1 will provide some resistance to model errors.

3.1. Method

The data-generating model is a two-dimensional factor model involving six manifest variables X 1 , , X 6 , and two latent (factor) variables η 1 and η 2 . The data-generating model is graphically presented in Figure 1. The first three items load on the first factor, while the last three item load on the second factor. Model errors exist by introducing two unmodeled residual correlations of items corresponding to the two factors (i.e., of items X 1 and X 4 , and items X 2 and X 5 ).
The means of all variables were set to equal in the population. However, the mean structure was left saturated in estimation because it is not of interest in this simulation study. The specified factor loadings λ were all set to 0.7, and all error variances were set to 0.5. The latent correlation ϕ 12 of the factors η 1 and η 2 was set to 0.5 in the simulation. The factor variable η , as well as the residuals ϵ , were multivariate normally distributed.
We varied the size of both residual error correlations as 0.5 · δ with δ = 0 , 0.3, and 0.6, indicating moderate-sized and large-sized residual error correlations. The sample size N was chosen to 500, 1000, or 2500.
We estimated the two-dimensional factor model with unmodeled residual error correlations. That is, the analysis models were misspecified (except for RegML). The six factor loadings and the six residual variances were freely estimated. The variances of the latent factors were set to 1. We employed ML, ULS, RegML, and RME, utilizing the L p loss function with powers p = 0.25 , 0.5, and 1. ML and ULS can be considered non-robust SEM estimation approaches, while RegML and RME are considered robust approaches. In total, six different estimates were compared for each simulated dataset in each condition.
For RegML, we only imposed the SCAD penalty on nondiagonal entries in the residual covariance matrix Ψ . RegML was estimated at a grid of regularization parameters between 10 and 0.002 (see replication material on https://osf.io/v8f45 (accessed on 27 March 2023) for specification details). In this simulation study, the tuning parameter a was fixed to 3.7. We chose estimates of the regularization approach using the optimal regularization κ ψ based on the BIC.
We studied the estimation recovery of the factor correlation ϕ 12 , a factor loading λ 11 , and a residual variance ψ 11 . We assessed the bias and root mean square error (RMSE) of a parameter θ ^ of interest. In each of R = 2000 replications in a simulation condition, the estimate θ ^ r ( r = 1 , , R ) was calculated. The bias was estimated by
Bias ( θ ^ ) = 1 R r = 1 R ( θ ^ r θ ) .
The RMSE was estimated by
RMSE ( θ ^ ) = 1 R r = 1 R ( θ ^ r θ ) 2 .
To ease the comparability of the RMSE between different methods across sample sizes, we used a relative RMSE in which we divided the RMSE of a particular method by the RMSE of the best-performing method in a simulation condition. Hence, a relative RMSE of 100 is the reference value for the best-performing method.
The statistical software R [60] was employed for all parts of the simulation. All estimators of the SEM were obtained using the mgsem() function in the sirt [61] package. Replication material can be found at https://osf.io/v8f45 (accessed on 27 March 2023).

3.2. Results

In Table 1, the bias and the relative RMSE for the factor correlation ϕ 12 , the factor loading λ 11 , and the residual variance ψ 11 are presented.
It turned out that in the condition of no unmodeled residual error correlations (i.e., δ = 0 and no model misspecification), no biases occurred for all parameters. However, there were slight efficiency losses with RME for N = 500 , which were particularly pronounced for the powers p = 0.25 and p = 0.5 for the residual variance. The efficiency loss for the factor loading for the robust estimation methods was slightly smaller than for the residual variance. Finally, there was almost no efficiency loss for the robust methods for the factor correlation.
In the conditions with model errors (i.e., in the case of unmodeled residual correlations), ML and ULS were severely biased for the factor correlation. The bias for ML and ULS turned out to be slightly smaller for the factor loading and the residual variance. RegML and RME, using the powers p = 0.25 or p = 0.5 , were unbiased, while RME with p = 1 had a small bias for the latent correlation. Regarding efficiency, RME with p = 0.25 was superior to p = 0.5 .
In general, RME with p = 0.25 performed quite competitively to RegML. Given the fact that RMSE is computationally much less demanding than RegML, one could prefer model-robust SEM estimation in the presence of unmodeled residual error correlations using the L p robust loss function in practical applications.

4. Focused Simulation Study 1A: Computation of Standard Errors

In this Focused Simulation Study 1A, different standard error estimation formulas are compared for selected conditions of Simulation Study 1.

4.1. Method

The same data-generating model of Simulation Study 1 (see Section 3) was utilized in a subset of conditions. Multivariate data was generated based on a two-dimensional factor model with two residual correlations. The correlations were omitted from the analysis model. The size of the residual correlations was chosen as 0.5 · δ with δ equal to 0 or 0.6. The same sample sizes as in Simulation Study 1 were employed (i.e., N = 500 , 1000, and 2500).
The same estimation methods as in Simulation Study 1, except RegML, were used (i.e., ML, ULS, and RME with p = 0.25 , 0.5, and 1). Standard errors were computed for each estimation method based on the multivariate delta formula (DF; see Equation (21)) that is implemented in the sirt::mgsem() function in the R package sirt [61]. Moreover, the resampling methods jackknife (JK) and bootstrap (BS) were used [42,62]. For jackknife, 40 replication zones were employed for determining the standard error of model parameter estimates. For bootstrap, 100 bootstrap samples were used to compute the standard deviation of a parameter estimate across bootstrap samples that were used as the standard error estimate. Moreover, standard errors for ML were also computed based on the observed information (OI) matrix obtained as the Hessian of the log-likelihood function. Note that the OI method is expected to result in biased standard error estimates in misspecified models (i.e., in the case of δ = 0.6 ).
Confidence intervals around model parameter estimates were constructed under a normality assumption of a parameter estimate at a confidence level of 95%. The empirical coverage rate was determined as the proportion that the confidence interval covers the pseudo-true parameter of an estimation method. A corresponding pseudo-true parameter was obtained by applying the respective estimation method to the covariance matrix of observed variables (i.e., Var ( X ) ) in an infinite sample size (i.e., at the population level).
In total, 4000 replications were used for estimating the coverage rate at the confidence level of 95%. A dedicated function for JK and BS standard error computation was written in R [60]. Model estimation and standard error computation based on DF were carried out using the mgsem() function in the R package sirt [61].

4.2. Results

In Table 2, coverage rates at the confidence level of 95% of model parameter estimates as a function of sample size N and the size δ of unmodeled residual error correlations are displayed. Overall, it turned out that DF, JK, and BS standard error estimates resulted in satisfactory coverage rates (i.e., between 92.5% and 97.5%) for a correctly specified (i.e., δ = 0 ) and a misspecified (i.e., δ = 0.6 ) SEM. As expected, standard errors based on the OI resulted in undercoverage for a misspecified SEM. Given the fact that standard error computation based on DF is much less computationally demanding, it can be recommended for usage in RME. Computationally more involved resampling methods, such as JK or BS, are, therefore, not necessarily required.

5. Simulation Study 2: Noninvariant Item Intercepts (DIF)

In this Simulation Study 2, we investigate the impact of unmodelled group-specific item intercepts in a multiple-group one-dimensional factor model. The presence of group-specific item intercepts indicates measurement noninvariance (i.e., in the presence of DIF). This simulation study investigates whether robust estimation approaches can handle the occurrence of DIF.

5.1. Method

The setup of the simulation study closely follows [32]. Data were simulated from a one-dimensional factor model involving five items and three groups. The factor variable η 1 was normally distributed with group means α 1 = 0 , α 2 = 0.3 , and α 3 = 0.8 . The group variances were set to 1, 1.5, and 1.2, respectively. All factor loadings were set to 1, and all measurement error variances were set to 1 in all groups and uncorrelated with each other. The factor variable, as well as the residuals, were normally distributed.
DIF effects were simulated in exactly one of the five items in each group. In the first group, the fourth item intercepts had a DIF effect δ . In the second group, the first item had a DIF effect δ , while the second item had a DIF effect δ in the third group. The DIF effect δ was chosen as either 0 (no DIF, measurement invariance), 0.3 (small DIF), or 0.6 (moderate DIF). The sample size per group was chosen as N = 500 , 1000, or 2000.
A multiple-group one-dimensional SEM was specified as the analysis model. The analysis model assumes invariant item intercepts and factor loadings. In the first group, the factor mean was set to 0, and the factor variance was set to 1 for identification reasons. Note that the data-generating model included some group-specific item intercepts that remained unmodelled in the analysis models (except for RegML). This, in turn, led to misspecified analysis models. ML, ULS, RegML with the SCAD penalty on group-specific item intercepts, and RME with powers p = 0.25 , 0.5, and 1 were utilized. In RegML, the optimal regularization parameter κ ν was chosen based on the BIC.
We investigated the estimation recovery of group means α 2 and α 3 of the second and the third group, respectively. Bias and relative RMSE (see Section 3.1) were again used for assessing the performance of the estimates. In total, 2000 replications were conducted. The models were estimated using the mgsem() function in the R [60] package sirt [61]. Replication material can be found at https://osf.io/v8f45 (accessed on 27 March 2023).

5.2. Results

In Table 3, bias and relative RMSE for the factor means of the second and the third group (i.e., α 2 and α 3 ) are presented.
In the condition of no DIF effects (i.e., δ = 0 ), all different estimators were unbiased. There were minor efficiency losses for robust moment estimation with p = 0.25 and p = 0.5 for the sample size N = 500 . However, in larger samples, there were almost no differences among the different estimators.
In the simulation conditions with small or moderate DIF effects, ML and ULS were substantially biased. Notably, RME with p = 1 also produced a nonnegligible bias. RME with powers p = 0.25 or p = 0.5 performed similarly to RegML concerning bias. RME had only slightly increased efficiency losses. Interestingly, RME p = 0.25 performed better than p = 0.5 in terms of RMSE.
Given the fact that there were only small efficiency losses with RME, the computationally more expensive RegML could perhaps be avoided in practice if the goal is estimating factor group means.

6. Empirical Example: ESS 2005 Data

We now present an empirical example to illustrate the performance of the different robust and non-robust SEM estimation approaches.

6.1. Method

In this empirical example, we use a dataset that was also analyzed in [32,63,64]. The data came from the European social survey (ESS) conducted in the year 2005 (ESS 2005) including subjects from 26 countries. The latent factor variable of tradition and conformity was assessed by four items presented in portrait format, where the scale of the items is such that a high value represents a low level of tradition conformity. The wording of the four items was (see [32]): “It is important for him to be humble and modest. He tries not to draw attention to himself.” (item TR9), “Tradition is important to him. He tries to follow the customs handed down by his religion or family.” (item TR20), “He believes that people should do what they’re told. He thinks people should follow rules at all times, even when no one is watching.” (item CO7), and “It is important for him to always behave properly. He wants to avoid doing anything people would say is wrong.” (item CO16). The full dataset used in [32] was downloaded from https://www.statmodel.com/Alignment.shtml (accessed on 27 March 2023). For this empirical example, a subsample of 33,060 persons from 17 selected countries was included to restrict the range of variability of country factor means. The sample sizes per country ranged between 1358 and 2963, with an average of 1997.1 ( SD = 464.5 ). We only included participants in the sample that had no missing values on all four items.
We specified a one-dimensional factor model with 17 groups (i.e., 17 countries) assuming invariant item parameters (i.e., invariant intercepts, loadings, and residual variances) in the estimation approaches ML, ULS, and RME, with powers p = 0.25 , 0.5, and 1. In RegML, the SCAD penalty was imposed on group-specific item intercepts.
The obtained country means and country standard deviations of the factor variable were linearly transformed for all different estimators such that the total population comprising all persons from all 17 countries had a mean of 0 and a standard deviation of 1.
The analysis was conducted using the mgsem() function from the R [60] package sirt [61]. The processed dataset and replication syntax can be found at https://osf.io/v8f45 (accessed on 27 March 2023).

6.2. Results

In the following, we present the estimates of RegML with the obtained optimal regularization parameter κ ν of 0.02.
In Table 4, the country means for the six different estimators are presented. The maximum absolute difference between country means stemming from the different models ranged between 0.032 and 0.337, with an average of 0.116 ( SD = 0.081 ). This showed considerable variability. Hence, the different robust and non-robust estimators differently weighted DIF effects in the computation of country means.
We also computed the ranks of the 17 countries across the six different estimates. The maximum absolute country rank difference ranged between 0 (country C10, rank 1) and 8 (country C21, rank 2) with an average of 3.9 ( SD = 2.1 ). A large range for country means was observed for country C21 (rank 2). The means for this country ranged between −0.01 (RME) and 0.33 (ML). Note that robust approaches provided similar results, but they substantially differed from RME with p = 1 , ULS, and ML. Similarly, for country C06 (rank 3), there was a range of country means with values between 0.06 and 0.26. However, low dependencies of country means from the model estimator choice were observed for countries C10 (rank1), C09 (rank 13), and C13 (rank 14).
In Figure 2, average absolute differences in country means between different models are displayed. Large absolute differences are displayed in darker red-brown color, while small differences are colored in light yellow-orange. RME with p = 0.25 (i.e., RME0.25) and RME with p = 0.5 (i.e., “RME0.5”) provided similar country means with an average distance of 0.01. Notably, RegML also resulted in similar estimates to RME, with p = 0.25 or p = 0.5 (i.e., average distances of 0.02 or 0.03). In addition, ML and ULS resulted in similar estimates. Interestingly, RME with p = 1 performed slightly differently compared to the other approaches. As in the simulation studies, it did not show full robustness to outlying DIF effects in item intercepts.
In Figure 3, the regularization paths of the estimated country means as a function of the regularization parameter κ ν are displayed. The RegML estimates of the country means are displayed on the left side of the figure, and the ML estimates are displayed on the right side of the figure. Interestingly, the paths are not monotone, and there is some variability in country mean estimates depending on the choice of the regularization parameter.
In Table 5, the estimated country-specific item intercepts for regularized ML estimation are presented. Overall, three countries had three regularized item intercepts, 12 countries had two regularized item intercepts, and two countries had one regularized item intercept. In total, 35 out of 68 item intercepts were regularized (i.e., the country-specific item intercepts were estimated as zero).
The regularization paths for country-specific item intercepts for four selected countries, C01, C09, C13, and C21, are displayed in Figure 4. One immediately recognizes that more and stronger deviations from invariance were observed with smaller values of the regularization parameter κ ν .

7. Discussion

In this article, we have demonstrated that, using model-robust SEM estimation methods, such as regularized maximum likelihood estimation and robust moment estimation, provides some resistance to model errors for modeled means and covariances. Importantly, this robustness property is only guaranteed if model errors are sparsely distributed. That is, most of the model errors are zero (or approximately zero), while only a few model errors are allowed to differ from zero. The plausibility of this assumption should be evaluated for each empirical application that potentially involves model error. Moreover it is noted that robust moment SEM estimation approaches did not result in practically relevant efficiency losses compared to non-robust approaches (i.e., maximum likelihood or unweighted least squares estimation) in the absence of model errors. Hence, robust estimation approaches can be recommended in empirical applications with moderate to large sample sizes (i.e., N 500 ). Importantly, robust moment estimation is computationally much less demanding than regularized estimation. Moreover, it also provides valid standard errors that are much more difficult to obtain with regularized ML estimation.
We applied robust moment estimation and regularized ML estimation for unmodeled item intercepts in Simulation Study 2 in Section 5. In this case, there is essentially no need to utilize the more computationally expensive regularization technique. The crucial assumption is that there are only a few model deviations in the modeled mean vectors and covariance matrices. This property is also referred to as the sparsity assumption [51], which means that the majority of the entries in mean vectors and covariance matrices are correctly specified. The methods discussed in this article will typically fail if the majority of or all the elements in the mean vectors and covariance matrices are misspecified [65,66]. This will likely be the case if the item loadings in the confirmatory factor model vary across groups. This situation is referred to as nonuniform DIF. Model deviations in item loadings lead to model deviations in the covariance matrix that is not as sparse as deviations in the mean structure. Hence, this is a situation in which regularized ML estimation might be preferred over robust moment estimation.
The treatment in this article was based on the multivariate normal distribution utilizing mean vectors and covariance matrices as sufficient statistics. Sometimes, researchers tend to model ordinal items by relying on underlying latent normally distributed variables [67,68]. In this case, thresholds and polychoric correlations replace mean vectors and covariance matrices in the weighted least squares fitting function. The approach based on model-robust fitting function can also be simply transferred to the case of modeling ordinal items. Future research might compare the performance of robust moment estimation with limited information methods relying on thresholds and polychoric correlations with regularized maximum likelihood estimation. Clearly, the computational advantages of model-robust limited information methods would be even more pronounced compared to the case of continuous items.
It should be noted that multivariate normality is not a requirement for obtaining consistent estimates in correctly specified structural equation models [37]. This article investigates the robustness of parameter estimates regarding model misspecification in the mean and the covariance structure. There is a distinct literature that focuses on robust estimation of SEM in the violation of multivariate normality [69,70]. These approaches might result in more efficient estimates in the case of heavy-tailed or contaminated distributions. However, outliers in this article are defined in the sense that some entries of the mean vector and the covariance matrix are misspecified infinite sample sizes (i.e., at the population level).
As pointed out in this article, the multiple-group SEM with unmodeled item intercepts corresponds to the estimation in the violation of measurement invariance (i.e., in the presence of differential item functioning). Linking approaches, such as invariance alignment [32,71] or robust Haberman linking [30], also deal with the estimation of factor means in one- or multidimensional confirmatory factor models. They do so by first estimating model parameters of the factor model separately in each group in the first step. In the second step, group-specific estimated item intercepts and factor loadings are transformed with a robust or non-robust linking function to identify factor means and factor standard deviations. This approach is advantageous if there is a misfit in the mean and the covariance structure. If the covariance structure is correctly specified, a joint robust moment estimation approach of the multiple-group SEM could have higher efficiency. In practical applications, it has been shown that misspecification more frequently occurs in the mean structure than in the covariance structure [72]. Hence, the robust moment estimation approach proposed in this article might be a viable alternative to the frequently employed invariance alignment technique. Future research could thoroughly compare robust moment estimation with invariance alignment or robust Haberman linking.
It has been argued that regularized estimation could also be applied with a fixed tuning parameter [73]. Recent research indicated that model parameter estimation could improve when using a fixed tuning parameter instead of an optimal tuning parameter based on the minimal information criterion [20]. Notably, regularized estimation in SEM is also referred to as penalized estimation [74]. In this framework, the penalty function P ( x ) | x | 1 / 2 is interpreted as a fixed prior distribution that has a density f ( x ) exp ( | x | 1 / 2 ) [74] using the so-called alignment loss function (see [30,32]).
An anonymous reviewer also suggested investigating the Huber loss function [10,38] in addition to the L p power loss function ρ ( x ) = | x | p ( p 1 ). As a disadvantage, the Huber loss function requires an additional tuning parameter. We have some limited empirical experience in using the Huber loss for robust Haberman linking [30]. In this situation, the Huber loss function performed in an inferior way to the L p loss function. This was particularly the case for asymmetrically distributed (i.e., asymmetrically structured) model deviations. Studying a wider range of robust loss functions in multiple-group SEM might be an interesting topic for future research, although we are somewhat skeptical that the Huber loss function will outperform the L p loss function in the conditions of our simulation studies.
Finally, it is always a substantive question of whether model errors are allowed to have outliers or not [75]. That is, should statistical models automatically remove particular means or covariances for estimating the parameter θ ? We attacked such a mechanistic approach to model estimation and argued that model deviations, such as violations of measurement invariance, should not automatically result in model refinement [76]. Hence, we do not think that there is always an optimal loss function in every application [75]. If this were true, the statistical inference would always be based on maximum likelihood estimation. However, we believe that researchers intentionally choose a statistical parameter of interest and apply M-estimation theory [77] for the statistical inference that is not based on correctly specified models. In this sense, SEMs can be estimated with non-robust estimation approaches (such as maximum likelihood or unweighted least squares estimation) using an intentionally misspecified model [78,79]. Appearing model errors can be included as additional uncertainty in standard errors of the estimated θ parameter. In SEM, ref. [9] proposed such an approach by imposing a stochastic model for model errors. We think that this approach should deserve more attention in empirical research.

Funding

This research received no external funding.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The dataset used in the empirical example is available at https://osf.io/v8f45 (accessed on 27 March 2023).

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BICBayesian information criterion
BS bootstrap
CFAconfirmatory factor analysis
DF delta formula
DIFdifferential item functioning
DWLSdiagonally weighted least squares
ESSEuropean social survey
JK jackknife
MADmean absolute deviation
MLmaximum likelihood
MVNmultivariate normal
OI observed information
RegMLregularized maximum likelihood
RMErobust moment estimation
RMSEroot mean square error
SCADsmoothly clipped absolute deviation
SEMstructural equation model
ULSunweighted least squares

References

  1. Bartholomew, D.J.; Knott, M.; Moustaki, I. Latent Variable Models and Factor Analysis: A Unified Approach; Wiley: New York, NY, USA, 2011. [Google Scholar] [CrossRef]
  2. Bollen, K.A. Structural Equations with Latent Variables; John Wiley & Sons: New York, NY, USA, 1989. [Google Scholar] [CrossRef]
  3. Browne, M.W.; Arminger, G. Specification and estimation of mean-and covariance-structure models. In Handbook of Statistical Modeling for the Social and Behavioral Sciences; Arminger, G., Clogg, C.C., Sobel, M.E., Eds.; Springer: Boston, MA, USA, 1995; pp. 185–249. [Google Scholar] [CrossRef]
  4. Jöreskog, K.G.; Olsson, U.H.; Wallentin, F.Y. Multivariate Analysis with LISREL; Springer: Basel, Switzerland, 2016. [Google Scholar] [CrossRef]
  5. Mulaik, S.A. Foundations of Factor Analysis; CRC Press: Boca Raton, FL, USA, 2009. [Google Scholar] [CrossRef]
  6. Shapiro, A. Statistical inference of covariance structures. In Current Topics in the Theory and Application of Latent Variable Models; Edwards, M.C., MacCallum, R.C., Eds.; Routledge: London, UK, 2012; pp. 222–240. [Google Scholar] [CrossRef]
  7. Yuan, K.H.; Bentler, P.M. Structural equation modeling. In Handbook of Statistics, Vol. 26: Psychometrics; Rao, C.R., Sinharay, S., Eds.; Elsevier: Amsterdam, The Netherlands, 2007; Volume 26, pp. 297–358. [Google Scholar] [CrossRef]
  8. Robitzsch, A. Comparing the robustness of the structural after measurement (SAM) approach to structural equation modeling (SEM) against local model misspecifications with alternative estimation approaches. Stats 2022, 5, 631–672. [Google Scholar] [CrossRef]
  9. Wu, H.; Browne, M.W. Quantifying adventitious error in a covariance structure as a random effect. Psychometrika 2015, 80, 571–600. [Google Scholar] [CrossRef] [PubMed]
  10. Huber, P.J.; Ronchetti, E.M. Robust Statistics; Wiley: New York, NY, USA, 2009. [Google Scholar] [CrossRef]
  11. Maronna, R.A.; Martin, R.D.; Yohai, V.J. Robust Statistics: Theory and Methods; Wiley: New York, NY, USA, 2006. [Google Scholar] [CrossRef]
  12. Ronchetti, E. The main contributions of robust statistics to statistical science and a new challenge. Metron 2021, 79, 127–135. [Google Scholar] [CrossRef]
  13. Robitzsch, A. Estimation methods of the multiple-group one-dimensional factor model: Implied identification constraints in the violation of measurement invariance. Axioms 2022, 11, 119. [Google Scholar] [CrossRef]
  14. Leitgöb, H.; Seddig, D.; Asparouhov, T.; Behr, D.; Davidov, E.; De Roover, K.; Jak, S.; Meitinger, K.; Menold, N.; Muthén, B.; et al. Measurement invariance in the social sciences: Historical development, methodological challenges, state of the art, and future perspectives. Soc. Sci. Res. 2023, 110, 102805. [Google Scholar] [CrossRef]
  15. Meredith, W. Measurement invariance, factor analysis and factorial invariance. Psychometrika 1993, 58, 525–543. [Google Scholar] [CrossRef]
  16. Millsap, R.E. Statistical Approaches to Measurement Invariance; Routledge: New York, NY, USA, 2011. [Google Scholar] [CrossRef]
  17. Holland, P.W.; Wainer, H. (Eds.) Differential Item Functioning: Theory and Practice; Lawrence Erlbaum: Hillsdale, NJ, USA, 1993. [Google Scholar] [CrossRef]
  18. Penfield, R.D.; Camilli, G. Differential item functioning and item bias. In Handbook of Statistics, Vol. 26: Psychometrics; Rao, C.R., Sinharay, S., Eds.; Elsevier: Amsterdam, The Netherlands, 2007; pp. 125–167. [Google Scholar] [CrossRef]
  19. Chen, Y.; Li, C.; Xu, G. DIF statistical inference and detection without knowing anchoring items. arXiv 2021, arXiv:2110.11112. [Google Scholar]
  20. Robitzsch, A. Comparing robust linking and regularized estimation for linking two groups in the 1PL and 2PL models in the presence of sparse uniform differential item functioning. Stats 2023, 6, 192–208. [Google Scholar] [CrossRef]
  21. Wang, W.; Liu, Y.; Liu, H. Testing differential item functioning without predefined anchor items using robust regression. J. Educ. Behav. Stat. 2022, 47, 666–692. [Google Scholar] [CrossRef]
  22. Boos, D.D.; Stefanski, L.A. Essential Statistical Inference; Springer: New York, NY, USA, 2013. [Google Scholar] [CrossRef]
  23. Kolenikov, S. Biases of parameter estimates in misspecified structural equation models. Sociol. Methodol. 2011, 41, 119–157. [Google Scholar] [CrossRef]
  24. White, H. Maximum likelihood estimation of misspecified models. Econometrica 1982, 50, 1–25. [Google Scholar] [CrossRef]
  25. Browne, M.W. Generalized least squares estimators in the analysis of covariance structures. S. Afr. Stat. J. 1974, 8, 1–24. [Google Scholar] [CrossRef]
  26. Savalei, V. Understanding robust corrections in structural equation modeling. Struct. Equ. Model. 2014, 21, 149–160. [Google Scholar] [CrossRef]
  27. MacCallum, R.C.; Browne, M.W.; Cai, L. Factor analysis models as approximations. In Factor Analysis at 100; Cudeck, R., MacCallum, R.C., Eds.; Lawrence Erlbaum: Hillsdale, NJ, USA, 2007; pp. 153–175. [Google Scholar] [CrossRef]
  28. Siemsen, E.; Bollen, K.A. Least absolute deviation estimation in structural equation modeling. Sociol. Methods Res. 2007, 36, 227–265. [Google Scholar] [CrossRef]
  29. van Kesteren, E.J.; Oberski, D.L. Flexible extensions to structural equation models using computation graphs. Struct. Equ. Model. 2022, 29, 233–247. [Google Scholar] [CrossRef]
  30. Robitzsch, A. Lp loss functions in invariance alignment and Haberman linking with few or many groups. Stats 2020, 3, 246–283. [Google Scholar] [CrossRef]
  31. Kolen, M.J.; Brennan, R.L. Test Equating, Scaling, and Linking; Springer: New York, NY, USA, 2014. [Google Scholar] [CrossRef]
  32. Asparouhov, T.; Muthén, B. Multiple-group factor analysis alignment. Struct. Equ. Model. 2014, 21, 495–508. [Google Scholar] [CrossRef]
  33. Muthén, B.; Asparouhov, T. IRT studies of many groups: The alignment method. Front. Psychol. 2014, 5, 978. [Google Scholar] [CrossRef]
  34. Pokropek, A.; Lüdtke, O.; Robitzsch, A. An extension of the invariance alignment method for scale linking. Psych. Test Assess. Model. 2020, 62, 303–334. [Google Scholar]
  35. Battauz, M. Regularized estimation of the nominal response model. Multivar. Behav. Res. 2020, 55, 811–824. [Google Scholar] [CrossRef]
  36. Oelker, M.R.; Tutz, G. A uniform framework for the combination of penalties in generalized structured models. Adv. Data Anal. Classif. 2017, 11, 97–120. [Google Scholar] [CrossRef]
  37. Shapiro, A. Statistical inference of moment structures. In Handbook of Latent Variable and Related Models; Lee, S.Y., Ed.; Elsevier: Amsterdam, The Netherlands, 2007; pp. 229–260. [Google Scholar] [CrossRef]
  38. Fox, J.; Weisberg, S. Robust Regression in R: An Appendix to an R Companion to Applied Regression. 2010. Available online: https://bit.ly/3canwcw (accessed on 27 March 2023).
  39. Holland, P.W.; Welsch, R.E. Robust regression using iteratively reweighted least-squares. Commun. Stat. Theory Methods 1977, 6, 813–827. [Google Scholar] [CrossRef]
  40. Chatterjee, S.; Mächler, M. Robust regression: A weighted least squares approach. Commun. Stat. Theory Methods 1997, 26, 1381–1394. [Google Scholar] [CrossRef]
  41. Ver Hoef, J.M. Who invented the delta method? Am. Stat. 2012, 66, 124–127. [Google Scholar] [CrossRef]
  42. Kolenikov, S. Resampling variance estimation for complex survey data. Stata J. 2010, 10, 165–199. [Google Scholar] [CrossRef]
  43. Chen, J. Partially confirmatory approach to factor analysis with Bayesian learning: A LAWBL tutorial. Struct. Equ. Model. 2022, 22, 800–816. [Google Scholar] [CrossRef]
  44. Geminiani, E.; Marra, G.; Moustaki, I. Single- and multiple-group penalized factor analysis: A trust-region algorithm approach with integrated automatic multiple tuning parameter selection. Psychometrika 2021, 86, 65–95. [Google Scholar] [CrossRef]
  45. Hirose, K.; Terada, Y. Sparse and simple structure estimation via prenet penalization. Psychometrika 2022. [Google Scholar] [CrossRef]
  46. Huang, P.H.; Chen, H.; Weng, L.J. A penalized likelihood method for structural equation modeling. Psychometrika 2017, 82, 329–354. [Google Scholar] [CrossRef]
  47. Huang, P.H. lslx: Semi-confirmatory structural equation modeling via penalized likelihood. J. Stat. Softw. 2020, 93, 1–37. [Google Scholar] [CrossRef]
  48. Jacobucci, R.; Grimm, K.J.; McArdle, J.J. Regularized structural equation modeling. Struct. Equ. Model. 2016, 23, 555–566. [Google Scholar] [CrossRef] [PubMed]
  49. Li, X.; Jacobucci, R.; Ammerman, B.A. Tutorial on the use of the regsem package in R. Psych 2021, 3, 579–592. [Google Scholar] [CrossRef]
  50. Scharf, F.; Nestler, S. Should regularization replace simple structure rotation in exploratory factor analysis? Struct. Equ. Model. 2019, 26, 576–590. [Google Scholar] [CrossRef]
  51. Fan, J.; Li, R.; Zhang, C.H.; Zou, H. Statistical Foundations of Data Science; Chapman and Hall/CRC: Boca Raton, FL, USA, 2020. [Google Scholar] [CrossRef]
  52. Hastie, T.; Tibshirani, R.; Wainwright, M. Statistical Learning with Sparsity: The Lasso and Generalizations; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar] [CrossRef]
  53. Friedrich, S.; Groll, A.; Ickstadt, K.; Kneib, T.; Pauly, M.; Rahnenführer, J.; Friede, T. Regularization approaches in clinical biostatistics: A review of methods and their applications. Stat. Methods Med. Res. 2023, 32, 425–440. [Google Scholar] [CrossRef] [PubMed]
  54. Fan, J.; Li, R. Variable selection via nonconcave penalized likelihood and its oracle properties. J. Am. Stat. Assoc. 2001, 96, 1348–1360. [Google Scholar] [CrossRef]
  55. Chen, Y.; Liu, J.; Xu, G.; Ying, Z. Statistical analysis of Q-matrix based diagnostic classification models. J. Am. Stat. Assoc. 2015, 110, 850–866. [Google Scholar] [CrossRef]
  56. Zhang, H.; Li, S.J.; Zhang, H.; Yang, Z.Y.; Ren, Y.Q.; Xia, L.Y.; Liang, Y. Meta-analysis based on nonconvex regularization. Sci. Rep. 2020, 10, 5755. [Google Scholar] [CrossRef]
  57. Tutz, G.; Gertheiss, J. Regularized regression for categorical data. Stat. Model. 2016, 16, 161–200. [Google Scholar] [CrossRef]
  58. Kyung, M.; Gill, J.; Ghosh, M.; Casella, G. Penalized regression, standard errors, and Bayesian lassos. Bayesian Anal. 2010, 5, 369–411. [Google Scholar] [CrossRef]
  59. Minnier, J.; Tian, L.; Cai, T. A perturbation method for inference on regularized regression estimates. J. Am. Stat. Assoc. 2011, 106, 1371–1382. [Google Scholar] [CrossRef]
  60. R Core Team. R: A Language and Environment for Statistical Computing; R Foundation for Statistical Computing: Vienna, Austria, 2022; Available online: https://www.R-project.org/ (accessed on 11 January 2022).
  61. Robitzsch, A. sirt: Supplementary Item Response Theory Models; R Package Version 3.13-128. 2023. Available online: https://github.com/alexanderrobitzsch/sirt (accessed on 2 April 2023).
  62. Efron, B.; Tibshirani, R.J. An Introduction to the Bootstrap; CRC Press: Boca Raton, FL, USA, 1994. [Google Scholar] [CrossRef]
  63. Knoppen, D.; Saris, W. Do we have to combine values in the Schwartz’ human values scale? A comment on the Davidov studies. Surv. Res. Methods 2009, 3, 91–103. [Google Scholar] [CrossRef]
  64. Beierlein, C.; Davidov, E.; Schmidt, P.; Schwartz, S.H.; Rammstedt, B. Testing the discriminant validity of Schwartz’ portrait value questionnaire items—A replication and extension of Knoppen and Saris (2009). Surv. Res. Methods 2012, 6, 25–36. [Google Scholar] [CrossRef]
  65. Muthén, B.; Asparouhov, T. New Methods for the Study of Measurement Invariance with Many Groups. Technical Report. 2013. Available online: https://bit.ly/3nBbr5M (accessed on 4 March 2023).
  66. Muthén, B.; Asparouhov, T. Recent methods for the study of measurement invariance with many groups: Alignment and random effects. Sociol. Methods Res. 2018, 47, 637–664. [Google Scholar] [CrossRef]
  67. Muthén, B. A general structural equation model with dichotomous, ordered categorical, and continuous latent variable indicators. Psychometrika 1984, 49, 115–132. [Google Scholar] [CrossRef]
  68. Robitzsch, A. On the bias in confirmatory factor analysis when treating discrete variables as ordinal instead of continuous. Axioms 2022, 11, 162. [Google Scholar] [CrossRef]
  69. Yuan, K.H.; Bentler, P.M.; Chan, W. Structural equation modeling with heavy tailed distributions. Psychometrika 2004, 69, 421–436. [Google Scholar] [CrossRef]
  70. Yuan, K.H.; Bentler, P.M. Robust procedures in structural equation modeling. In Handbook of Latent Variable and Related Models; Lee, S.Y., Ed.; Elsevier: Amsterdam, The Netherlands, 2007; pp. 367–397. [Google Scholar] [CrossRef]
  71. Pokropek, A.; Davidov, E.; Schmidt, P. A Monte Carlo simulation study to assess the appropriateness of traditional and newer approaches to test for measurement invariance. Struct. Equ. Model. 2019, 26, 724–744. [Google Scholar] [CrossRef]
  72. Rutkowski, L.; Svetina, D. Assessing the hypothesis of measurement invariance in the context of large-scale international surveys. Educ. Psychol. Meas. 2014, 74, 31–57. [Google Scholar] [CrossRef]
  73. Liu, X.; Wallin, G.; Chen, Y.; Moustaki, I. Rotation to sparse loadings using Lp losses and related inference problems. Psychometrika 2023. [Google Scholar] [CrossRef]
  74. Asparouhov, T.; Muthén, B. Penalized Structural Equation Models. Technical Report. 2023. Available online: https://bit.ly/3TlbxdC (accessed on 4 March 2023).
  75. Hennig, C.; Kutlukaya, M. Some thoughts about the design of loss functions. Revstat Stat. J. 2007, 5, 19–39. [Google Scholar] [CrossRef]
  76. Robitzsch, A.; Lüdtke, O. Why full, partial, or approximate measurement invariance are not a prerequisite for meaningful and valid group comparisons. Struct. Equ. Model. 2023. [Google Scholar]
  77. Stefanski, L.A.; Boos, D.D. The calculus of M-estimation. Am. Stat. 2002, 56, 29–38. [Google Scholar] [CrossRef]
  78. Hennig, C. How wrong models become useful-and correct models become dangerous. In Between Data Science and Applied Data Analysis. Studies in Classification, Data Analysis, and Knowledge Organization; Schader, M., Gaul, W., Vichi, M., Eds.; Springer: Berlin, Germany, 2003; pp. 235–243. [Google Scholar] [CrossRef]
  79. Saltelli, A.; Funtowicz, S. When all models are wrong. Issues Sci. Technol. 2014, 30, 79–85. [Google Scholar]
Figure 1. Simulation Study 1: Data-generating model.
Figure 1. Simulation Study 1: Data-generating model.
Algorithms 16 00210 g001
Figure 2. Empirical example: Visualization of average absolute deviations in country means between different estimation methods. Note. ML = maximum likelihood estimate; RegML = regularized maximum likelihood estimate using the Bayesian information criterion for regularization parameter selection; RME = robust moment estimation with power p in the loss function ρ ( x ) = | x | p ; ULS = unweighted least squares estimate.
Figure 2. Empirical example: Visualization of average absolute deviations in country means between different estimation methods. Note. ML = maximum likelihood estimate; RegML = regularized maximum likelihood estimate using the Bayesian information criterion for regularization parameter selection; RME = robust moment estimation with power p in the loss function ρ ( x ) = | x | p ; ULS = unweighted least squares estimate.
Algorithms 16 00210 g002
Figure 3. Empirical example: Regularization paths for estimated country means in regularized maximum likelihood estimation as a function of the regularization parameter κ ν .
Figure 3. Empirical example: Regularization paths for estimated country means in regularized maximum likelihood estimation as a function of the regularization parameter κ ν .
Algorithms 16 00210 g003
Figure 4. Empirical example: Regularization paths for estimated country-specific item intercepts in regularized maximum likelihood estimation as a function of the regularization parameter κ ν .
Figure 4. Empirical example: Regularization paths for estimated country-specific item intercepts in regularized maximum likelihood estimation as a function of the regularization parameter κ ν .
Algorithms 16 00210 g004aAlgorithms 16 00210 g004b
Table 1. Simulation Study 1: Bias and relative root mean square error (RMSE) of model parameter estimates as a function of sample size N and size of residual error correlations δ .
Table 1. Simulation Study 1: Bias and relative root mean square error (RMSE) of model parameter estimates as a function of sample size N and size of residual error correlations δ .
Bias Relative RMSE
RME with p = RME with p =
Par δ NRegML0.250.51ULSML RegML0.250.51ULSML
ϕ 12 0500−0.01 0.00 0.00 0.00 0.00 0.00 102102101101100101
1000 0.00 0.00 0.00 0.00 0.00 0.00 102101101100100100
2500 0.00 0.00 0.00 0.00 0.00 0.00 103100100100100100
0.3500 0.02 0.02 0.02 0.03 0.07 0.07 100101103113162166
1000 0.01 0.01 0.01 0.03 0.07 0.07 100101105122213218
2500 0.01 0.01 0.01 0.03 0.07 0.07 100101108144318328
0.6500 0.01 0.01 0.01 0.03 0.15 0.16 100101102119312336
1000 0.00 0.01 0.01 0.03 0.15 0.16 100100103129429462
2500 0.00 0.01 0.01 0.03 0.15 0.16 100100105152672722
λ 11 0500 0.00 0.00 0.00 0.00 0.00 0.00 101108106104104100
1000 0.00 0.00 0.00 0.00 0.00 0.00 101105104103104100
2500 0.00 0.00 0.00 0.00 0.00 0.00 101104104104104100
0.3500 0.00 0.00 0.00 0.01 0.02 0.01 102104103101109100
1000 0.00 0.00 0.00 0.01 0.02 0.01 100101101102117102
2500 0.00 0.00 0.00 0.01 0.02 0.01 100100101105137108
0.6500 0.00 0.00 0.00 0.01 0.04 0.02 100108107107139138
1000 0.00 0.00 0.00 0.01 0.04 0.02 100107107110166146
2500 0.00 0.00 0.00 0.01 0.04 0.02 100104105111219160
ψ 11 0500 0.00 0.00 0.00 0.00 0.00 0.00 100117113109109100
1000 0.00 0.00 0.00 0.00 0.00 0.00 100109107106107100
2500 0.00 0.00 0.00 0.00 0.00 0.00 101107107107108100
0.3500−0.02−0.01−0.01−0.01−0.03−0.02 107110107105116100
1000−0.01−0.01−0.01−0.01−0.03−0.02 104100100103127102
2500−0.01 0.00−0.01−0.01−0.03−0.01 107100101108155111
0.6500−0.02 0.00−0.01−0.01−0.06−0.03 105102100100139136
1000−0.01 0.00 0.00−0.01−0.06−0.03 102101100104176145
2500 0.00 0.00 0.00−0.01−0.06−0.03 100101102112260177
Note. Par  = model parameter; ML = maximum likelihood estimate; RegML = regularized maximum likelihood estimate using the Bayesian information criterion for regularization parameter selection; RME  = robust moment estimation with power p in the loss function ρ ( x ) = | x | p ; ULS  = unweighted least squares estimate; Biases with absolute values of at least 0.03 are printed in bold. RMSE values of at least 110 are printed in bold.
Table 2. Focused Simulation Study 1A: Coverage rate at confidence level 95% of model parameter estimates as a function of sample size N and size of residual error correlations δ for different standard error estimation methods.
Table 2. Focused Simulation Study 1A: Coverage rate at confidence level 95% of model parameter estimates as a function of sample size N and size of residual error correlations δ for different standard error estimation methods.
RME with p =
0.25 0.5 1 ULS ML
Par δ NDFJKBS DFJKBS DFJKBS DFJKBS OIDFJKBS
ϕ 12 050095.895.195.5 95.594.894.9 95.394.495.0 94.994.294.6 95.095.194.294.6
100095.294.794.9 95.194.694.7 94.894.494.7 94.694.294.3 94.794.794.494.5
250095.494.795.3 95.394.794.9 95.394.694.9 95.294.794.7 95.295.294.794.9
0.650094.894.594.6 94.594.194.5 93.793.093.3 94.794.094.3 93.594.493.794.4
100094.394.094.7 94.393.794.1 93.793.393.1 94.794.094.4 93.694.694.094.1
250095.194.595.0 95.094.394.7 95.094.294.1 95.294.594.8 94.395.294.394.6
λ 11 050095.895.496.0 95.695.195.9 95.294.595.1 94.794.094.3 95.095.193.994.7
100095.294.395.6 95.193.995.1 94.693.794.4 94.493.493.7 94.694.793.893.8
250094.593.894.6 94.493.894.1 94.293.994.3 94.293.894.0 94.694.693.894.1
0.650095.594.996.0 95.694.796.2 95.694.595.3 94.993.894.5 88.596.295.796.4
100095.194.595.5 94.994.595.0 94.794.494.8 94.693.894.4 88.595.694.795.9
250095.194.795.2 95.094.794.9 95.394.594.9 95.294.394.9 89.095.494.595.4
ψ 11 050096.696.297.6 96.495.896.9 95.894.895.4 95.194.394.4 95.095.094.494.9
100096.195.396.7 95.995.195.9 95.594.895.3 95.394.694.9 95.195.194.794.9
250095.094.795.3 95.194.695.1 95.194.694.9 95.094.594.7 95.095.094.394.4
0.650096.195.797.1 96.195.696.9 95.795.295.8 94.994.594.6 88.796.796.596.8
100095.395.095.9 95.495.095.8 95.394.795.3 94.994.294.7 88.695.895.196.4
250095.194.795.0 95.194.695.0 94.994.594.8 95.194.594.7 88.895.294.895.6
Note. Par  = model parameter; ML = maximum likelihood estimate; RegML = regularized maximum likelihood estimate using the Bayesian information criterion for regularization parameter selection; RME  = robust moment estimation with power p in the loss function ρ ( x ) = | x | p ; ULS  = unweighted least squares estimate; DF = standard error estimation based on delta formula; JK = standard error estimation based on jackknife; BS = standard error estimation based on bootstrap; OI = standard error estimation based on observed information.
Table 3. Simulation Study 2: Bias and relative root mean square error (RMSE) of model parameter estimates as a function of sample size N and size of DIF effects δ .
Table 3. Simulation Study 2: Bias and relative root mean square error (RMSE) of model parameter estimates as a function of sample size N and size of DIF effects δ .
Bias Relative RMSE
RME with p = RME with p =
Par δ NRegML0.250.51ULSML RegML0.250.51ULSML
α 2 0500 0.00 0.00 0.00 0.00 0.00 0.00 100103102100100100
1000 0.00 0.00 0.00 0.00 0.00 0.00 100102101100100100
2500 0.00 0.00 0.00 0.00 0.00 0.00 100100100100100100
0.3500−0.02−0.03−0.03−0.07−0.12−0.12 101100100113152152
1000 0.00−0.01−0.02−0.06−0.12−0.12 100103106140231230
2500 0.00−0.01−0.01−0.05−0.12−0.12 100103107169356355
0.6500 0.00−0.01−0.02−0.07−0.24−0.24 100104104129303302
1000 0.00−0.01−0.01−0.06−0.24−0.24 100102103144435433
2500 0.00 0.00−0.01−0.05−0.24−0.24 100101104171698692
α 3 0500 0.01 0.01 0.01 0.01 0.01 0.01 100103102101100100
1000 0.00 0.00 0.00 0.00 0.00 0.00 100102101100100100
2500 0.00 0.00 0.00 0.00 0.00 0.00 100101101100100100
0.3500−0.02−0.03−0.03−0.07−0.12−0.11 100103102115153149
1000 0.00−0.01−0.02−0.06−0.12−0.11 100104107139223216
2500 0.00−0.01−0.01−0.05−0.12−0.12 100103108167344332
0.6500 0.00−0.01−0.01−0.06−0.23−0.21 100103102124286267
1000 0.00 0.00−0.01−0.05−0.23−0.22 100103103139421390
2500 0.00−0.01−0.01−0.05−0.24−0.22 100101104174693641
Note. Par  = model parameter; ML = maximum likelihood estimate; RegML = regularized maximum likelihood estimate using the Bayesian information criterion for regularization parameter selection; RME  = robust moment estimation with power p in the loss function ρ ( x ) = | x | p ; ULS  = unweighted least squares estimate; Biases with absolute values of at least 0.03 are printed in bold. RMSE values of at least 110 are printed in bold.
Table 4. Empirical example: Estimated country means for different estimation methods.
Table 4. Empirical example: Estimated country means for different estimation methods.
RME with p =
RankCNTRegML0.250.51ULSML
1C100.370.370.370.340.340.35
2C210.01−0.010.010.180.270.33
3C060.060.170.190.260.250.26
4C030.330.330.320.250.210.21
5C080.130.150.160.230.230.20
6C120.120.140.150.200.190.16
7C050.150.170.170.150.110.13
8C160.140.050.060.090.060.07
9C010.120.110.100.070.080.07
10C14−0.11−0.10−0.10−0.09−0.05−0.06
11C220.02−0.01−0.02−0.06−0.08−0.07
12C15−0.18−0.17−0.17−0.16−0.01−0.07
13C09−0.21−0.21−0.21−0.21−0.19−0.23
14C13−0.30−0.30−0.30−0.32−0.28−0.30
15C17−0.20−0.19−0.21−0.29−0.36−0.33
16C25−0.17−0.18−0.19−0.29−0.39−0.34
17C24−0.29−0.32−0.34−0.37−0.38−0.39
Note. Rank = country rank based on estimated country mean with ML; CNT = country labels originally used in [32]; ML = maximum likelihood estimate; RegML = regularized maximum likelihood estimate using the Bayesian information criterion for regularization parameter selection; RME = robust moment estimation with power p in the loss function ρ ( x ) = | x | p ; ULS = unweighted least squares estimate.
Table 5. Empirical example: Estimated country-specific item intercepts for regularized maximum likelihood estimation based on BIC regularization parameter selection.
Table 5. Empirical example: Estimated country-specific item intercepts for regularized maximum likelihood estimation based on BIC regularization parameter selection.
CNTTR09TR20CO07CO16
C01·−0.17·−0.17
C03·−0.32·−0.26
C05−0.500.29··
C060.09·0.24·
C08·0.12··
C09·−0.24·0.09
C10−0.470.25··
C12·0.05··
C130.29··−0.47
C140.100.31·−0.42
C150.420.07·−0.22
C16−0.26··−0.13
C17−0.52−0.12··
C21··0.48·
C22−0.14··−0.36
C24·−0.32·−0.17
C25−0.52··−0.31
Note. CNT = country labels originally used in [32]. Regularized item intercepts are indicated with a dot (i.e., with “·”).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Robitzsch, A. Model-Robust Estimation of Multiple-Group Structural Equation Models. Algorithms 2023, 16, 210. https://doi.org/10.3390/a16040210

AMA Style

Robitzsch A. Model-Robust Estimation of Multiple-Group Structural Equation Models. Algorithms. 2023; 16(4):210. https://doi.org/10.3390/a16040210

Chicago/Turabian Style

Robitzsch, Alexander. 2023. "Model-Robust Estimation of Multiple-Group Structural Equation Models" Algorithms 16, no. 4: 210. https://doi.org/10.3390/a16040210

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop