Next Article in Journal
Electric Vehicle Routing Problem with Heterogeneous Energy Replenishment Infrastructures Under Capacity Constraints
Previous Article in Journal
Probabilistic Contingent Planning Based on Hierarchical Task Network for High-Quality Plans
Previous Article in Special Issue
An Automated Framework for Streamlined CFD-Based Design and Optimization of Fixed-Wing UAV Wings
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Computational Aspects of L0 Linking in the Rasch Model

by
Alexander Robitzsch
1,2
1
IPN—Leibniz Institute for Science and Mathematics Education, Olshausenstraße 62, 24118 Kiel, Germany
2
Centre for International Student Assessment (ZIB), Olshausenstraße 62, 24118 Kiel, Germany
Algorithms 2025, 18(4), 213; https://doi.org/10.3390/a18040213
Submission received: 20 February 2025 / Revised: 4 April 2025 / Accepted: 7 April 2025 / Published: 9 April 2025
(This article belongs to the Special Issue Numerical Optimization and Algorithms: 3rd Edition)

Abstract

:
The L 0 linking approach replaces the L 2 loss function in mean–mean linking under the Rasch model with the L 0 loss function. Using the L 0 loss function offers the advantage of potential robustness against fixed differential item functioning effects. However, its nondifferentiability necessitates differentiable approximations to ensure feasible and computationally stable estimation. This article examines alternative specifications of two approximations, each controlled by a tuning parameter ε that determines the approximation error. Results demonstrate that the optimal ε value minimizing the RMSE of the linking parameter estimate depends on the magnitude of DIF effects, the number of items, and the sample size. A data-driven selection of ε outperformed a fixed ε across all conditions in both a numerical illustration and a simulation study.

1. Introduction

Item response theory (IRT) models [1,2,3] are statistical models for multivariate discrete random variables. This article focuses on dichotomous (i.e., binary) random variables and the comparison of two groups through a linking method.
Let X = ( X 1 , , X I ) denote a vector of I N random variables X i { 0 , 1 } , commonly referred to as items or (scored) item responses. A unidimensional IRT model [4] represents the probability distribution P ( X = x ) for x = ( x 1 , , x I ) { 0 , 1 } I ,
P ( X = x ; δ , γ ) = i = 1 I P i ( θ ; γ i ) x i 1 P i ( θ ; γ i ) 1 x i ϕ ( θ ; μ , σ ) d θ ,
where ϕ denotes the density of the normal distribution with mean μ and standard deviation (SD) σ . The distribution parameters of the latent variable θ , often labeled as a trait or ability variable, are collected in the vector  δ = ( μ , σ ) . The vector  γ = ( γ 1 , , γ I ) contains the item parameters associated with the item response functions (IRFs) P i ( θ ; γ i ) = P ( X i = 1 | θ ) for i = 1 , , I . The IRF of the Rasch model [5,6,7] is defined as
P i ( θ ; γ i ) = Ψ θ b i ,
where b i denotes the item difficulty b i , and Ψ ( x ) = ( 1 + exp ( x ) ) 1 is the logistic distribution function. In this formulation, the item parameter vector is given as γ i = ( b i ) .
For a sample of N individuals with independently and identically distributed observations x 1 , , x N from the distribution of the random variable X , the unknown parameters of the IRT model in (1) can be consistently estimated using maximum likelihood estimation (ML; [8,9,10]).
IRT models are commonly used to compare the performance of two groups on a test by examining differences in the latent variable θ within the framework of the IRT model in (1). This article focuses on a generalization of the mean–mean linking method [11] based on the Rasch model. The purpose of a linking method is to estimate the difference between the distributions of θ in the two groups. This difference serves as a summary measure of group performance on the multivariate vector of dichotomous items X .
The linking approach consists of two steps. First, the Rasch model is estimated separately for each group, allowing for potential differential item functioning (DIF), where items may function differently across groups [12,13,14]. Second, differences in item parameters are used to estimate group differences in the θ variable through a linking method [11,15,16].
This article investigates the performance of a generalization of mean–mean (MM; [11]) linking in the presence of fixed uniform DIF [14] in item difficulties. The traditional MM method relies on mean differences in item difficulties, which corresponds to using an  L 2 loss function. MM linking is a widely used linking or equating method, as discussed in popular textbooks on Rasch modeling [6,7,17,18]. In this study, we investigate a generalization of the MM linking method in the Rasch model using the L 0 loss function [19]. Using this robust loss function can essentially remove items with DIF effects from the group comparison directly in the linking method (see also [20,21,22,23,24,25,26,27]). DIF effects can also be treated through the application of robust procedures in MM linking [19,28,29,30,31].
The L 0 loss function, being nondifferentiable, requires differentiable approximations to ensure feasible and computationally stable estimation. This article explores alternative specifications of these approximations, which depend on a tuning parameter ε that controls the approximation error. Previous research relied on fixed ε values based on prior knowledge or simulation studies. Here, the choice of a fixed ε parameter is examined and compared with a data-driven approach for determining ε .
The rest of this article is structured as follows. Section 2 introduces L 0 linking in the Rasch model. Section 3 presents a numerical illustration of alternative specifications of L 0 linking in a simplified setting. In Section 4, findings from a simulation study are reported. An empirical example is provided in Section 5. This article closes with a discussion in Section 6 and a conclusion in Section 7.

2. L 0 Linking in the Rasch Model

This section discusses the L 0 linking approach in the Rasch model. Section 2.1 covers item parameter identification and the ordinary MM linking method. Section 2.2 introduces the L 0 loss function and presents two differentiable approximations. Section 2.3 applies these approximations to define the L 0 linking approach in the Rasch model. Finally, Section 2.4 examines its statistical properties.

2.1. Identified Item Parameters and Mean–Mean Linking

To introduce the L 0 linking method as a replacement for ordinary mean–mean linking in the Rasch model, we first outline the identification of item parameters in a group-specific estimation of the Rasch model when no DIF effects are present. In this case, the identified item parameters in the first group are b ^ i 1 = b i , where b i represents the invariant item parameters across both groups. For identification, the mean of θ in the first group is fixed at 0, while the standard deviation σ remains identifiable within this group.
In the second group, the mean μ and the SD σ can be identified when invariant item difficulties are assumed. Thus, μ and σ represent group differences in θ . The identified item parameters in this group, estimated separately under the assumption of a θ mean of 0 and an estimated SD σ , are given by
b ^ i 2 = b i μ .
Linking methods aim to recover the group parameter μ using the group-specific item parameters obtained from separate Rasch model estimations.
The MM linking method estimates the group mean μ ^ for the second group as
μ ^ = 1 I i = 1 I b ^ i 2 b ^ i 1 .
The linking parameter μ ^ in MM linking is obtained as the minimizer of the squared loss function (i.e., the L 2 loss function; [19])
μ ^ = arg min μ H ( μ ) , with H ( μ ) = 1 I i = 1 I ρ b ^ i 2 b ^ i 1 + μ , where ρ ( x ) = 1 2 x 2 .
Thus, the estimate μ ^ satisfies the estimating equation
1 I i = 1 I ρ b ^ i 2 b ^ i 1 + μ ^ = 0 , where ρ ( x ) = x
and ρ denotes the derivative of ρ with respect to x. Clearly, (6) is equivalent to (4).
This paper investigates the computational aspects of MM linking in (5) when the squared L 2 loss function is replaced with the L 0 loss function, which aims to robustify the linking method in the presence of fixed DIF effects.

2.2. L 0 Loss Function and Differentiable Approximations

In this section, we introduce the L 0 loss function [32,33], which is defined as
ρ ( x ) = 1 ( x 0 ) .
This loss function equals 0 for x = 0 and 1 for x 0 . The L 0 loss function has been widely applied in statistical modeling, particularly for obtaining sparse solutions (see, e.g., [34,35,36,37,38,39]).
The exact L 0 loss function in (7) has the disadvantage of being nondifferentiable at x = 0 , making it difficult to use in numerical optimization. To address this, differentiable approximations of the L 0 loss function have been proposed.
The ratio loss function, as an approximation of the L 0 loss function, is defined as (see [40,41,42,43])
ρ ε ( x ) = x 2 x 2 + ε ,
where ε > 0 is a tuning parameter. We have found that ε = 0.01 performs satisfactorily in applications [19,44].
The linking parameter estimate based on the L 2 loss function can be obtained using the L 0 approximation (8) with a sufficiently large ε value, such as ε = 1000 . In this case, ρ ε ( x ) x 2 / ε , and the resulting estimate closely matches the L 2 estimate, which is the ordinary MM parameter estimate.
An alternative differentiable approximation of the L 0 loss function is based on the Gaussian density function, denoted as the Gaussian function hereafter, and is given by (see [45,46,47])
ρ ε ( x ) = 1 exp x 2 ε ,
where ε > 0 is again a tuning parameter.
Figure 1 displays the ratio and the Gaussian loss functions for ε = 0.001 and ε = 0.01 . It can be seen that the two loss functions exhibit similar behavior around x = 0 , which can be explained by the fact that their first two derivatives coincide, with ρ ε ( 0 ) = 0 and  ρ ε ( 0 ) = 2 / ε . For larger | x | values, the Gaussian loss function grows faster toward the upper asymptote of 1 compared to the ratio loss function.
The ordinary MM parameter estimate can be obtained using the Gaussian L 0 loss function approximation, similar to the ratio function, by selecting a sufficiently large ε value.

2.3. L 0 Linking as a Robust Mean–Mean Linking in the Rasch Model

MM linking can be modified by using the L 0 loss function instead of the L 2 loss function. In the former case, the loss is defined by counting the number of non-vanishing deviations b ^ i 2 from b ^ i 1 + μ , which corresponds to the number of non-vanishing DIF effects. To obtain a stable linking parameter estimate in the presence of sampling errors, the L 0 loss function is replaced by the differentiable approximation ρ ε , and the estimated mean μ ^ is obtained as
μ ^ = arg min μ H ( μ ) with H ( μ ) = 1 I i = 1 I ρ ε b ^ i 2 b ^ i 1 + μ .
This approach can also be referred to as robust MM linking or L 0 linking in the Rasch model, where “robust” refers to the linking parameter estimate being resilient to the presence of fixed DIF effects.
Equivalently to (10), the linking parameter estimate μ ^ solves the estimating equation
H μ = 1 I i = 1 I ρ ε b ^ i 2 b ^ i 1 + μ ^ = 0 .
Equation (10) represents a one-dimensional optimization problem that can be numerically solved using standard optimizers implemented in statistical software.

2.4. Statistical Properties of the Estimated Linking Parameter in L 0 Linking

We now investigate the statistical properties of the linking parameter estimate μ ^ obtained from (10). The difference in estimated item difficulties required in the L 0 linking approach can be written as
b ^ i 2 b ^ i 1 = μ + κ i + u i
with fixed DIF effects κ i and sampling errors u i , where E ( u i ) = 0 and Var ( u i ) = V i . Asymptotic unbiasedness follows from the general properties of ML estimation.
The L 0 linking approach is analyzed by starting from the estimating equation
1 I i = 1 I ρ ε b ^ i 2 b ^ i 1 + μ ^ = 1 I i = 1 I ρ ε μ ^ μ + κ i + u i = 0 .
Now, we apply a Taylor expansion with respect to μ , as in standard M-estimation theory ([48]; see also [25,49]), and obtain
0 = 1 I i = 1 I ρ ε μ ^ μ + κ i + u i 1 I i = 1 I ρ ε κ i + u i + 1 I i = 1 I ρ ε κ i + u i μ ^ μ = 0 .
The bias and variance of μ ^ can be derived from (14) and are given by
Bias ( μ ^ ) = i = 1 I E ρ ε κ i + u i i = 1 I E ρ ε κ i + u i and
Var ( μ ^ ) = i = 1 I Var ρ ε κ i + u i i = 1 I E ρ ε κ i + u i 2 ,
where approximate independence of item parameters across items is assumed in (16). As a summary precision measure of the linking parameter estimate, the mean squared error (MSE) can be determined as
MSE ( μ ^ ) = E μ ^ μ 2 = Bias ( μ ^ ) 2 + Var ( μ ^ ) = i = 1 I E ρ ε κ i + u i 2 + Var ρ ε κ i + u i i = 1 I E ρ ε κ i + u i 2 .
Although it might not be immediately evident from Equations (15)–(17), the bias, variance, and MSE depend on the choice of the tuning parameter ε in the L 0 approximation  ρ ε .
In the next two sections, we compare the two differentiable approximations—the ratio and Gaussian loss functions—regarding their statistical properties in a numerical illustration and a simulation study. In particular, we focus on the choice of the tuning parameter ε .

3. Numerical Illustration

3.1. Method

In this Numerical Illustration, the properties of the estimated linking parameter μ ^ were studied in a simplified setting in which no item responses were simulated. It was assumed that the difference b ^ i 2 b ^ i 1 had a variance of V / N , where N denotes the sample size. Hence, this illustration assumed equal sampling variances of item parameter differences, which might be violated in practice. However, this assumption eases the statistical treatment and aims at yielding clearly interpretable results. Furthermore, we assumed that we had I items, and a proportion π of the items had an unbalanced DIF effect κ , while a proportion 1 π of the items did not show DIF.
From Section 2.4, we know that the linking parameter satisfies the estimating equation
1 I i = 1 I ρ ε b ^ i 2 b ^ i 1 + μ ^ = 1 I i = 1 I ρ ε μ ^ μ + κ i + u i = 0 ,
where Var ( u i ) = V / N . Because all item parameter differences have equal sampling variances, (18) can be simplified to
i = 1 I π ρ ε μ ^ μ + κ + u i + I π + 1 I ρ ε μ ^ μ + u i = 0 ,
assuming that I π is an integer. By using the bias Formula (15), we obtain
Bias ( μ ^ ) = π E ρ ε κ + u i π E ρ ε κ + u i + ( 1 π ) E ρ ε u i ,
where we use E ρ ε u i = 0 because u i has a symmetric distribution. The variance can be computed as (see (16))
Var ( μ ^ ) = π Var ρ ε κ + u i + ( 1 π ) Var ρ ε u i π E ρ ε κ + u i + ( 1 π ) E ρ ε u i 2 ,
The root mean square error (RMSE) as the square root of the MSE can be obtained as
RMSE ( μ ^ ) = MSE ( μ ^ ) = Bias ( μ ^ ) 2 + Var ( μ ^ ) .
The expected values and variances in (20) and (21) can be numerically evaluated.
The bias, SD, and RMSE are evaluated for μ ^ for sample sizes N of 250, 500, and 1000, as well as the number of items I as 10, 20, and 40. We simulated fixed DIF effects κ = 0.4 and κ = 0.8 , representing small and large DIF. The tuning parameter ε in the ratio loss function and Gaussian loss function ρ ε was chosen as 1, 0.95, 0.9, 0.85, 0.8, 0.75, 0.7, 0.65, 0.6, 0.55, 0.5, 0.45, 0.4, 0.35, 0.3, 0.25, 0.2, 0.15, 0.1, 0.095, 0.09, 0.085, 0.08, 0.075, 0.07, 0.065, 0.06, 0.055, 0.05, 0.045, 0.04, 0.035, 0.03, 0.025, 0.02, 0.015, 0.01, 0.009, 0.008, 0.007, 0.006, 0.005, 0.004, 0.003, 0.002, and 0.001. In total, 46 ε values were evaluated in this Numerical Illustration. We only report results for the ratio loss function because the findings for the Gaussian loss function were very similar. The statistical software R (Version 4.4.1; [50]) was used for the analysis in this study. Symbolic derivatives of the two loss functions were computed with the R package Deriv (Version: 4.1.6; [51]). Replication material for this illustration can be retrieved from https://osf.io/un6q4 (accessed on 20 February 2025).

3.2. Results

Figure 2 displays the absolute bias, the SD, and the RMSE of the estimated mean μ ^ for I = 20 items as a function of the DIF effect κ and the sample size N. The tuning parameter ε is displayed on a logarithmic scale on the x-axis. The bias increased with increasing values of ε in the loss function ρ ε . In contrast, the SD decreased with increasing values of ε . The RMSE reflects the bias–variance trade-off of the linking parameter estimate. There was an optimal ε parameter that minimizes the RMSE. This optimal ε parameter decreased with increasing sample size N. Moreover, the optimal ε parameter was larger for κ = 0.8 than for κ = 0.4 .
Table 1 presents the optimal ε values that minimize the RMSE of the estimated mean  μ ^ under a ratio loss function. These values are reported for different DIF effect sizes κ , numbers of items I, and sample sizes N. As the sample size N increased, the optimal  ε generally decreased. This pattern was more pronounced for small item numbers (e.g., I = 10 ), where  ε dropped sharply with increasing N. Larger item numbers I were associated with smaller optimal  ε values, suggesting that when more items were available, the best trade-off under the ratio loss function occurred at a lower ε value. This effect was particularly evident for  ε = 0.4 , where ε decreased substantially as I increased from 10 to 40. When the DIF effect size κ increased from 0.4 to 0.8, the optimal ε values tended to be slightly higher for the same I and N. This suggests that stronger DIF effects required a higher ε to achieve the lowest RMSE, likely due to greater bias introduced by DIF at the smaller effect size κ = 0.4 .
Overall, the results highlight a trade-off between sample size, the number of items, and DIF effect size in determining the optimal correction for minimizing the RMSE.

4. Simulation Study

4.1. Method

The Rasch model was used as the IRF in the data-generating model for two groups. For identification purposes, the mean of the latent variable θ in the first group was fixed at 0, with an SD σ of 1. In the second group, the mean μ was set to 0.3 to represent the difference in θ between the groups, while the SD σ was set to 1.2.
The simulation study was conducted for I = 10 , 20, and 40 items. In the I = 10 condition, base item difficulty values b i were set to −0.314, 0.411, −1.097, −0.542, −1.854, −0.403, −0.895, 0.715, 0.841, and 0.139, resulting in a mean of M = 0.300 and an SD of 0.850. These item difficulties were used in the data generation for the first group. In the second group, the same values were applied, except for the first three items, where item difficulties were adjusted to b i 2 = b i + κ . This introduced fixed DIF in 30% of the items. The DIF effect size κ was set to 0.4 and 0.8, representing small and large DIF conditions, respectively. For item sets larger than 10, the same 10-item parameter set was repeated accordingly. The item parameters are also available at https://osf.io/un6q4 (accessed on 20 February 2025). Note that item parameters remained fixed across all replications within each condition.
Sample sizes of N = 125 , 250, 500, and 1000 per group were selected to reflect small- to large-scale applications of the Rasch model.
A total of 5000 replications were conducted for each of the 24 conditions, corresponding to the combinations of 4 sample sizes (N) × 3 item numbers (I) × 2 DIF effect sizes ( κ ). For each simulated dataset, L 0 linking was performed using various specifications. Both the ratio and Gaussian loss functions ρ ε (see Section 2.2) were applied. The ρ ε loss functions were evaluated for a sequence of 11 ε values: 1, 0.75, 0.50, 0.25, 0.10, 0.075, 0.05, 0.025, 0.01, 0.005, and 0.001.
Additionally, as demonstrated in Section 3, the optimal ε value corresponding to the minimal RMSE is dependent on the sample size N. To account for this, a data-driven approach for selecting the ε value was explored. Let V ¯ denote the average variance of the estimated item difficulty differences b ^ i 2 b ^ i 1 . For a fixed y ( 0 , 1 ) , ε y is chosen such that
ρ ε y V ¯ = y .
In the simulation, the following y values were chosen: 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, 0.7, 0.8, and 0.9.
The bias, standard deviation (SD), and root mean square error (RMSE) of the estimated mean, μ ^ , were calculated for all specifications of the L 0 linking approach. The relative RMSE of μ ^ was defined as the ratio of the RMSE for a given specification to the RMSE of a reference method, multiplied by 100. Across all conditions, the reference method was the ratio loss function with a fixed ε value of 0.01.
All analyses for this simulation study were conducted using the statistical software R (Version 4.4.1; [50]). The Rasch model was fitted with the sirt::xxirt() function from the R package sirt (Version 4.2-106; [52]). The optimization of L 0 linking was performed with the stats::nlminb() function. Replication materials for this simulation study can be assessed at https://osf.io/un6q4 (accessed on 20 February 2025).

4.2. Results

Table 2 presents the absolute bias as a function of the DIF effect size κ , the number of items I, and the sample size N. Overall, the absolute bias generally decreased as κ increased from 0.4 to 0.8. The bias in the estimated μ ^ was highest for the smallest number of items ( I = 10 ) and slightly decreased as the number of items increased. This reduction in bias was more pronounced for smaller sample sizes. As the sample size N increased, the bias systematically decreased, with the lowest bias observed for the largest sample size ( N = 1000 ). For the data-driven ε choices ε 0.2 and ε 0.4 , the bias also decreased as the sample size increased.
The biases for the ratio and Gaussian loss functions were relatively similar, although the bias was slightly smaller under conditions with a larger DIF effect size ( κ = 0.8 ) compared to a smaller DIF effect size ( κ = 0.4 ). Additionally, the bias decreased as ε values in the ρ ε function were reduced. Specifically, the bias was slightly higher for ε = 0.2 compared to ε = 0.4 , particularly at smaller N. These findings suggest that while the choice of the loss function influenced estimation accuracy, its effect diminished as I and N increased.
Table 3 presents the relative RMSE of the estimated mean μ ^ as a function of the DIF effect size κ , the number of items I, and the sample size N. In general, larger ε values performed better in smaller samples and with fewer items. Thus, the optimal ε for minimizing the RMSE depends on κ , I, and N. The data-driven estimates ε 0.2 and ε 0.4 generally outperformed the fixed choice ε = 0.01 based on the loss function ρ ε . Overall, it can be concluded that ε 0.4 was superior to ε 0.2 . Although ε 0.2 produced a lower RMSE than ε 0.4 in many simulation conditions, the differences were typically small. However, in scenarios where ε 0.2 performed worse than ε 0.4 , its RMSE exceeded the reference value of 100 and was substantially higher than that of ε 0.4 . This supports the use of ε 0.4 as a conservative default choice. In most conditions, the Gaussian loss function slightly outperformed the ratio loss function. In large samples ( N = 1000 ), smaller ε values, such as ε = 0.05 or ε = 0.01 , proved effective.
To provide further insight into the absolute bias and relative RMSE, two additional figures are presented below.
Figure 3 presents the absolute bias, SD, and RMSE of the estimated mean μ ^ using the ratio loss function for I = 20 items as a function of the DIF effect size κ and sample size N. The absolute bias increased with larger ε , while the SD decreased, illustrating a bias–variance trade-off in the RMSE. This trade-off results in an optimal ε value that minimizes the RMSE. As shown in Figure 3, the optimal ε was smaller for a larger DIF effect size (i.e.,  κ = 0.8 ) than for κ = 0.4 . Additionally, the optimal ε decreased with increasing sample size N. As expected, the RMSE was lower for larger sample sizes.
Figure 4 presents the relative RMSE of the estimated mean based on data-driven  ε y values for the ratio loss function ρ ε with ε = ε y . The results indicate that an optimal RMSE was achieved for a specific y value, depending on the DIF effect size κ , the number of items I, and the sample size N. Overall, the findings in Figure 4 support the selection of  ε 0.2 and  ε 0.4 , as presented in Table 2 and Table 3.
Following the suggestion of a reviewer, bias and RMSE were regressed onto the simulation factors using a two-way analysis of variance (ANOVA). The factors included sample size N, number of items I, the size of the DIF effect κ , the type of loss function (i.e., ratio vs. Gaussian function), and the chosen ε value in the loss function ρ ε . All simulation factors were treated as categorical factors in the ANOVA. The same item numbers I (e.g., I = 10 , 20 , 40 ) and ε values (i.e., ε = 0.01 , 0.05 , 0.10 , 0.25 ) as reported in Table 2 and Table 3 were used. The main focus of the analysis was the proportion of variance explained by the main effects and two-way interactions of the factors in the ANOVA.
In the two-way ANOVA with bias as the dependent variable, 96.0 % of the variance was explained by the main factors and their two-way interactions. The largest proportion was attributed to sample size N ( 39.8 % ), followed by κ ( 24.8 % ) and ε ( 24.8 % ). The number of items I ( 0.1 % ) and the type of loss function ( 0.2 % ) contributed minimally to the variability in bias. Among the two-way interactions, only the interaction between N and κ ( N × κ , 2.3 % ) and N × ε ( 1.7 % ) accounted for non-negligible proportions of variance.
In the two-way ANOVA with RMSE as the dependent variable, 99.3 % of the variance was explained by the simulation factors and their two-way interactions. As expected, sample size N accounted for the largest proportion of explained variance ( 89.9 % ), followed by number of items I ( 3.2 % ), κ ( 1.0 % ), ε ( 0.5 % ), and the type of loss function ( 0.0 % ). Among the two-way interactions, N × I ( 1.3 % ), N × κ ( 1.0 % ), and N × ε ( 2.1 % ) contributed with non-negligible amounts to the variance.

5. Empirical Example

We now illustrate L 0 linking in the Rasch model using an empirical dataset. The dataDIF dataset from the R package equateIRT (Version 1.0.0; [53,54]) was selected for this purpose. The dataset contains 20 dichotomous items and three groups, each with 1000 subjects. For illustration, only Groups 1 and 2 were linked using the Rasch model.
As in the Simulation Study (see Section 4), the Rasch model was fitted using the sirt::xxirt() function from the R package sirt (Version 4.2-106; [52]). The optimization of L 0 linking was performed using the stats::nlminb() function, with R code specifically written for this paper, available at https://osf.io/un6q4 (accessed on 20 February 2025).
The average standard error of item difficulty differences was 0.126. The mean ability difference between Group 1 and Group 2, obtained from MM linking based on the L 2 loss function, was 0.482. Slight deviations from this estimate were observed with different specifications of the L 0 linking function. For the ratio loss function, the mean estimate based on ε = 0.01 was 0.535. Empirical choices of the ε parameter, specifically ε 0.2 and ε 0.4 , were also explored. For the ratio loss function, the estimates for ε 0.2 = 0.064 and ε 0.4 = 0.024 yielded mean differences of 0.518 and 0.520, respectively. The mean estimate based on the Gaussian loss function was 0.537, with estimates for ε 0.2 = 0.072 and ε 0.4 = 0.031 yielding 0.519 and 0.517, respectively.
DIF effects for the 20 items were also computed based on the estimated mean difference of 0.520, obtained with ε 0.4 and the ratio loss function. Figure 5 displays the DIF effects along with their 95% confidence intervals. Item 1 exhibited a significantly positive DIF effect (0.798, with a standard error of 0.112), while DIF effects for the remaining 19 items did not significantly differ from 0.

6. Discussion

This article examined the computational aspects of L 0 linking in the Rasch model. L 0 linking serves as a robust alternative to MM linking when fixed DIF effects are present. Since the L 0 loss function is not differentiable at x = 0 , the ratio loss function and the Gaussian loss function were used as differentiable approximations. Both approximations rely on a tuning parameter ε , which controls the approximation error. A numerical illustration and a simulation study demonstrated that the optimal  ε value minimizing the RMSE of the linking parameter estimate μ ^ depends on the magnitude of DIF effects, the number of items, and the sample size. A data-driven selection of  ε provided better performance than a fixed  ε across all conditions. Additionally, the Gaussian loss function showed a slight advantage over the ratio loss function, though the differences are likely negligible in practice.
This study aimed to determine group differences in a robust manner, ensuring insensitivity to DIF effects. Detecting DIF items was treated as a nuisance, as the goal was not to identify deviating items but rather to obtain a consistent estimate of the mean difference in ability between the two groups. In contrast, much of the psychometric literature focuses on detecting DIF items [22,27,55,56,57], with group difference estimation being, at best, a by-product of the procedure.
Using the L 0 loss function effectively removes items with large DIF effects from the linking process. As a result, the estimated group difference is based solely on items with little to no DIF. This approach may pose a threat to validity, as it alters the construct being measured by treating items with DIF effects as construct-irrelevant [58,59]. Restricting the item set in this way can change the interpretation of the ability variable [60,61,62,63,64].

7. Conclusions

The L 0 loss function proved effective in linking two groups based on the Rasch model in the presence of differential item functioning. Two differentiable approximations (i.e., the ratio and Gaussian loss functions) of the nondifferentiable L 0 loss function have been proposed. The smoothness of the approximation depends on the tuning parameter ε . A simulation study demonstrated that a data-driven choice of ε , based on the average standard error of item difficulty differences, produced estimates with a lower RMSE compared to estimates based on a fixed ε value, such as ε = 0.01 .

Funding

This research received no external funding.

Data Availability Statement

This article only uses simulated datasets. Replication material for Section 3 and Section 4 can be found at https://osf.io/un6q4 (accessed on 20 February 2025). The dataset dataDIF used in the empirical example in Section 5 is available from the equateIRT package (https://doi.org/10.32614/CRAN.package.equateIRT; accessed on 20 February 2025).

Conflicts of Interest

The author declares no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ANOVAanalysis of variance
DIFdifferential item functioning
IRFitem response function
IRTitem response theory
MLmaximum likelihood
MMmean–mean
MSEmean square error
RMSEroot mean square error
SDstandard deviation

References

  1. Bock, R.D.; Gibbons, R.D. Item Response Theory; Wiley: Hoboken, NJ, USA, 2021. [Google Scholar] [CrossRef]
  2. Chen, Y.; Li, X.; Liu, J.; Ying, Z. Item response theory—A statistical framework for educational and psychological measurement. arXiv 2021, arXiv:2108.08604. [Google Scholar] [CrossRef]
  3. Yen, W.M.; Fitzpatrick, A.R. Item response theory. In Educational Measurement; Brennan, R.L., Ed.; Praeger Publishers: Westport, CT, USA, 2006; pp. 111–154. [Google Scholar]
  4. Van der Linden, W.J. Unidimensional logistic response models. In Handbook of Item Response Theory, Volume 1: Models; van der Linden, W.J., Ed.; CRC Press: Boca Raton, FL, USA, 2016; pp. 11–30. [Google Scholar] [CrossRef]
  5. Rasch, G. Probabilistic Models for Some Intelligence and Attainment Tests; Danish Institute for Educational Research: Copenhagen, Denmark, 1960. [Google Scholar]
  6. Bond, T.; Yan, Z.; Heene, M. Applying the Rasch Model; Routledge: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  7. Debelak, R.; Strobl, C.; Zeigenfuse, M.D. An Introduction to the Rasch Model with Examples in R; CRC Press: Boca Raton, FL, USA, 2022. [Google Scholar] [CrossRef]
  8. Bock, R.D.; Aitkin, M. Marginal maximum likelihood estimation of item parameters: Application of an EM algorithm. Psychometrika 1981, 46, 443–459. [Google Scholar] [CrossRef]
  9. Glas, C.A.W. Maximum-likelihood estimation. In Handbook of Item Response Theory, Volume 2: Statistical Tools; van der Linden, W.J., Ed.; CRC Press: Boca Raton, FL, USA, 2016; pp. 197–216. [Google Scholar] [CrossRef]
  10. Robitzsch, A. A comprehensive simulation study of estimation methods for the Rasch model. Stats 2021, 4, 814–836. [Google Scholar] [CrossRef]
  11. Kolen, M.J.; Brennan, R.L. Test Equating, Scaling, and Linking; Springer: New York, NY, USA, 2014. [Google Scholar] [CrossRef]
  12. Holland, P.W.; Wainer, H. (Eds.) Differential Item Functioning: Theory and Practice; Lawrence Erlbaum: Hillsdale, NJ, USA, 1993. [Google Scholar] [CrossRef]
  13. Millsap, R.E. Statistical Approaches to Measurement Invariance; Routledge: New York, NY, USA, 2011. [Google Scholar] [CrossRef]
  14. Penfield, R.D.; Camilli, G. Differential item functioning and item bias. In Handbook of Statistics, Volume 26: Psychometrics; Rao, C.R., Sinharay, S., Eds.; Elsevier: Amsterdam, The Netherlands, 2007; pp. 125–167. [Google Scholar] [CrossRef]
  15. Lee, W.C.; Lee, G. IRT linking and equating. In The Wiley Handbook of Psychometric Testing: A Multidisciplinary Reference on Survey, Scale and Test; Irwing, P., Booth, T., Hughes, D.J., Eds.; Wiley: New York, NY, USA, 2018; pp. 639–673. [Google Scholar] [CrossRef]
  16. Sansivieri, V.; Wiberg, M.; Matteucci, M. A review of test equating methods with a special focus on IRT-based approaches. Statistica 2017, 77, 329–352. [Google Scholar] [CrossRef]
  17. Andrich, D.; Marais, I. A Course in Rasch Measurement Theory; Springer: New York, NY, USA, 2019. [Google Scholar] [CrossRef]
  18. Lamprianou, I. Applying the Rasch Model in Social Sciences Using R and BlueSky Statistics; Routledge: New York, NY, USA, 2019. [Google Scholar] [CrossRef]
  19. Robitzsch, A. Extensions to mean–geometric mean linking. Mathematics 2025, 13, 35. [Google Scholar] [CrossRef]
  20. Von Davier, M.; Bezirhan, U. A robust method for detecting item misfit in large scale assessments. Educ. Psychol. Meas. 2023, 83, 740–765. [Google Scholar] [CrossRef]
  21. De Boeck, P. Random item IRT models. Psychometrika 2008, 73, 533–559. [Google Scholar] [CrossRef]
  22. Halpin, P.F. Differential item functioning via robust scaling. Psychometrika 2024, 89, 796–821. [Google Scholar] [CrossRef]
  23. He, Y.; Cui, Z.; Fang, Y.; Chen, H. Using a linear regression method to detect outliers in IRT common item equating. Appl. Psychol. Meas. 2013, 37, 522–540. [Google Scholar] [CrossRef]
  24. Magis, D.; De Boeck, P. Identification of differential item functioning in multiple-group settings: A multivariate outlier detection approach. Multivar. Behav. Res. 2011, 46, 733–755. [Google Scholar] [CrossRef]
  25. Robitzsch, A. Robust and nonrobust linking of two groups for the Rasch model with balanced and unbalanced random DIF: A comparative simulation study and the simultaneous assessment of standard errors and linking errors with resampling techniques. Symmetry 2021, 13, 2198. [Google Scholar] [CrossRef]
  26. Robitzsch, A. A comparison of linking methods for two groups for the two-parameter logistic item response model in the presence and absence of random differential item functioning. Foundations 2021, 1, 116–144. [Google Scholar] [CrossRef]
  27. Wang, W.; Liu, Y.; Liu, H. Testing differential item functioning without predefined anchor items using robust regression. J. Educ. Behav. Stat. 2022, 47, 666–692. [Google Scholar] [CrossRef]
  28. Hu, H.; Rogers, W.T.; Vukmirovic, Z. Investigation of IRT-based equating methods in the presence of outlier common items. Appl. Psychol. Meas. 2008, 32, 311–333. [Google Scholar] [CrossRef]
  29. Jurich, D.; Liu, C. Detecting item parameter drift in small sample Rasch equating. Appl. Meas. Educ. 2023, 36, 326–339. [Google Scholar] [CrossRef]
  30. Liu, C.; Jurich, D. Outlier detection using t-test in Rasch IRT equating under NEAT design. Appl. Psychol. Meas. 2023, 47, 34–47. [Google Scholar] [CrossRef] [PubMed]
  31. Manna, V.F.; Gu, L. Different Methods of Adjusting for Form Difficulty Under the Rasch Model: Impact on Consistency of Assessment Results; (Research Report No. RR-19-08); Educational Testing Service: Princeton, NJ, USA, 2019. [Google Scholar] [CrossRef]
  32. Oelker, M.R.; Pößnecker, W.; Tutz, G. Selection and fusion of categorical predictors with L0-type penalties. Stat. Model. 2015, 15, 389–410. [Google Scholar] [CrossRef]
  33. Oelker, M.R.; Tutz, G. A uniform framework for the combination of penalties in generalized structured models. Adv. Data Anal. Classif. 2017, 11, 97–120. [Google Scholar] [CrossRef]
  34. Atamturk, A.; Gómez, A.; Han, S. Sparse and smooth signal estimation: Convexification of l0-formulations. J. Mach. Learn. Res. 2021, 22, 1–43. [Google Scholar]
  35. Dai, S. Variable selection in convex quantile regression: L1-norm or L0-norm regularization? Eur. J. Oper. Res. 2023, 305, 338–355. [Google Scholar] [CrossRef]
  36. Huang, J.; Jiao, Y.; Liu, Y.; Lu, X. A constructive approach to L0 penalized regression. J. Mach. Learn. Res. 2018, 19, 1–37. [Google Scholar]
  37. Panokin, N.V.; Kostin, I.A.; Karlovskiy, A.V.; Nalivaiko, A.Y. Comparison of sparse representation methods for complex data based on the smoothed L0 norm and modified minimum fuel neural network. Appl. Sci. 2025, 15, 1038. [Google Scholar] [CrossRef]
  38. Soubies, E.; Blanc-Féraud, L.; Aubert, G. A continuous exact l0 penalty (CEL0) for least squares regularized problem. SIAM J. Imaging Sci. 2015, 8, 1607–1639. [Google Scholar] [CrossRef]
  39. Yang, Y.; McMahan, C.S.; Wang, Y.B.; Ouyang, Y. Estimation of l0 norm penalized models: A statistical treatment. Comp. Stat. Data Anal. 2024, 192, 107902. [Google Scholar] [CrossRef]
  40. Liu, W.; Li, Z.; Chen, W. Evaluating model robustness using adaptive sparse L0 regularization. arXiv 2024, arXiv:2408.15702. [Google Scholar] [CrossRef]
  41. O’Neill, M.; Burke, K. Variable selection using a smooth information criterion for distributional regression models. Stat. Comput. 2023, 33, 71. [Google Scholar] [CrossRef]
  42. Wang, B.; Wang, L.; Yu, H.; Xin, F. A new regularized reconstruction algorithm based on compressed sensing for the sparse underdetermined problem and applications of one-dimensional and two-dimensional signal recovery. Algorithms 2019, 12, 126. [Google Scholar] [CrossRef]
  43. Xiang, J.; Yue, H.; Yin, X.; Wang, L. A new smoothed l0 regularization approach for sparse signal recovery. Math. Probl. Eng. 2019, 2019, 1978154. [Google Scholar] [CrossRef]
  44. Robitzsch, A. L0 and Lp loss functions in model-robust estimation of structural equation models. Psych 2023, 5, 1122–1139. [Google Scholar] [CrossRef]
  45. Paik, J.W.; Lee, J.H.; Hong, W. An enhanced smoothed L0-norm direction of arrival estimation method using covariance matrix. Sensors 2021, 21, 4403. [Google Scholar] [CrossRef]
  46. Wang, L.; Yin, X.; Yue, H.; Xiang, J. A regularized weighted smoothed L0 norm minimization method for underdetermined blind source separation. Sensors 2018, 18, 4260. [Google Scholar] [CrossRef]
  47. Zhu, J.; Li, X. A smoothed l0-norm and l1-norm regularization algorithm for computed tomography. J. Appl. Math. 2019, 2019, 8398035. [Google Scholar] [CrossRef]
  48. Boos, D.D.; Stefanski, L.A. Essential Statistical Inference; Springer: New York, NY, USA, 2013. [Google Scholar] [CrossRef]
  49. Simakhin, V.A.; Shamanaeva, L.G.; Avdyushina, A.E. Robust parametric estimates of heterogeneous experimental data. Russ. Phys. J. 2021, 63, 1510–1518. [Google Scholar] [CrossRef]
  50. R Core Team. R: A Language and Environment for Statistical Computing; R Core Team: Vienna, Austria, 2024; Available online: https://www.R-project.org (accessed on 15 June 2024).
  51. Clausen, A.; Sokol, S. Deriv: Symbolic Differentiation, 2024. R Package Version 4.1.6. Available online: https://cran.r-project.org/web/packages/Deriv/ (accessed on 13 September 2024).
  52. Robitzsch, A. sirt: Supplementary Item Response Theory Models, 2024. R Package Version 4.2-106. Available online: https://github.com/alexanderrobitzsch/sirt (accessed on 31 December 2024).
  53. Battauz, M. equateIRT: An R package for IRT test equating. J. Stat. Softw. 2015, 68, 1–22. [Google Scholar] [CrossRef]
  54. Battauz, M. equateMultiple: Equating of Multiple Forms, 2024. R Package Version 1.0.0. Available online: https://cran.r-project.org/web/packages/equateMultiple/index.html (accessed on 13 September 2024).
  55. Chen, Y.; Li, C.; Ouyang, J.; Xu, G. DIF statistical inference without knowing anchoring items. Psychometrika 2023, 88, 1097–1122. [Google Scholar] [CrossRef] [PubMed]
  56. Halpin, P.F.; Gilbert, J. Testing whether reported treatment effects are unduly dependent on the specific outcome measure used. arXiv 2024, arXiv:2409.03502. [Google Scholar] [CrossRef]
  57. Strobl, C.; Kopf, J.; Kohler, L.; von Oertzen, T.; Zeileis, A. Anchor point selection: Scale alignment based on an inequality criterion. Appl. Psychol. Meas. 2021, 45, 214–230. [Google Scholar] [CrossRef]
  58. Camilli, G. The case against item bias detection techniques based on internal criteria: Do item bias procedures obscure test fairness issues? In Differential Item Functioning: Theory and Practice; Holland, P.W., Wainer, H., Eds.; Erlbaum: Hillsdale, NJ, USA, 1993; pp. 397–417. [Google Scholar]
  59. Shealy, R.; Stout, W. A model-based standardization approach that separates true bias/DIF from group ability differences and detects test bias/DTF as well as item bias/DIF. Psychometrika 1993, 58, 159–194. [Google Scholar] [CrossRef]
  60. De Los Reyes, A.; Tyrell, F.A.; Watts, A.L.; Asmundson, G.J.G. Conceptual, methodological, and measurement factors that disqualify use of measurement invariance techniques to detect informant discrepancies in youth mental health assessments. Front. Psychol. 2022, 13, 931296. [Google Scholar] [CrossRef]
  61. El Masri, Y.H.; Andrich, D. The trade-off between model fit, invariance, and validity: The case of PISA science assessments. Appl. Meas. Educ. 2020, 33, 174–188. [Google Scholar] [CrossRef]
  62. Funder, D.C.; Gardiner, G. MIsgivings about measurement invariance. Eur. J. Pers. 2024, 38, 889–895. [Google Scholar] [CrossRef]
  63. Welzel, C.; Inglehart, R.F. Misconceptions of measurement equivalence: Time for a paradigm shift. Comp. Political Stud. 2016, 49, 1068–1094. [Google Scholar] [CrossRef]
  64. Zwitser, R.J.; Glaser, S.S.F.; Maris, G. Monitoring countries in a changing world: A new look at DIF in international surveys. Psychometrika 2017, 82, 210–232. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Ratio loss function and Gaussian loss function ρ ε for ε = 0.001 and ε = 0.01 as differentiable approximations of the L 0 loss function.
Figure 1. Ratio loss function and Gaussian loss function ρ ε for ε = 0.001 and ε = 0.01 as differentiable approximations of the L 0 loss function.
Algorithms 18 00213 g001
Figure 2. Numerical Illustration: Absolute bias, standard deviation (SD), and root mean square error (RMSE) of the estimated mean μ ^ using the ratio loss function for I = 20 items as a function of sample size N and DIF effect size κ . The minimum RMSE values are depicted with a triangle. The ε parameter on the x-axis is displayed on a logarithmic scale.
Figure 2. Numerical Illustration: Absolute bias, standard deviation (SD), and root mean square error (RMSE) of the estimated mean μ ^ using the ratio loss function for I = 20 items as a function of sample size N and DIF effect size κ . The minimum RMSE values are depicted with a triangle. The ε parameter on the x-axis is displayed on a logarithmic scale.
Algorithms 18 00213 g002
Figure 3. Simulation Study: Absolute bias, standard deviation (SD), and root mean square error (RMSE) of the estimated mean μ ^ using the ratio loss function for I = 20 items as a function of sample size N and DIF effect size κ . The minimum RMSE values are depicted with a triangle.
Figure 3. Simulation Study: Absolute bias, standard deviation (SD), and root mean square error (RMSE) of the estimated mean μ ^ using the ratio loss function for I = 20 items as a function of sample size N and DIF effect size κ . The minimum RMSE values are depicted with a triangle.
Algorithms 18 00213 g003
Figure 4. Simulation Study: Relative root mean square error (RMSE) of the estimated mean μ ^ using the ratio loss function based on ε = ε y determined through ρ ε y ( V ¯ ) = y as a function of sample size N, the number of items I, and DIF effect size κ . The minimum relative RMSE values are depicted with a triangle. The ratio loss function with ε = 0.01 was chosen as the reference method when computing the relative RMSE.
Figure 4. Simulation Study: Relative root mean square error (RMSE) of the estimated mean μ ^ using the ratio loss function based on ε = ε y determined through ρ ε y ( V ¯ ) = y as a function of sample size N, the number of items I, and DIF effect size κ . The minimum relative RMSE values are depicted with a triangle. The ratio loss function with ε = 0.01 was chosen as the reference method when computing the relative RMSE.
Algorithms 18 00213 g004
Figure 5. Empirical example: DIF parameter estimates (displayed with a black triangle) along with their 95% confidence intervals.
Figure 5. Empirical example: DIF parameter estimates (displayed with a black triangle) along with their 95% confidence intervals.
Algorithms 18 00213 g005
Table 1. Numerical Illustration: optimal ε value yielding the minimum root mean square error (RMSE) of the estimated mean μ ^ using the ratio loss function as a function of the DIF effect size κ , number of items I, and sample size N.
Table 1. Numerical Illustration: optimal ε value yielding the minimum root mean square error (RMSE) of the estimated mean μ ^ using the ratio loss function as a function of the DIF effect size κ , number of items I, and sample size N.
I κ = 0.4 , N = κ = 0.8 , N =
25050010002505001000
100.9000.0650.030 0.1500.1000.090
200.2000.0500.030 0.1000.1000.090
400.1000.0350.025 0.1000.0950.085
Table 2. Simulation Study: Absolute bias of estimated mean μ ^ as a function of the DIF effect size κ , number of items I, and sample size N.
Table 2. Simulation Study: Absolute bias of estimated mean μ ^ as a function of the DIF effect size κ , number of items I, and sample size N.
κ I N Ratio Function, ε = Gaussian Function, ε =
0.250.100.050.01 ε 0.2 ε 0.4 0.250.100.050.01 ε 0.2 ε 0.4
0.4101250.1070.1030.1020.1020.1100.105 0.1070.1010.1000.1010.1120.106
2500.0980.0870.0820.0790.0960.085 0.1010.0860.0800.0780.1010.087
5000.0880.0660.0520.0410.0680.049 0.0970.0670.0460.0330.0750.048
10000.0820.0500.0310.0150.0340.019 0.0960.0570.0260.0100.0350.013
201250.1040.1000.0990.0990.1070.102 0.1040.0990.0960.0960.1090.103
2500.0980.0860.0800.0750.0970.084 0.1020.0840.0750.0690.1020.086
5000.0890.0660.0510.0360.0690.048 0.0970.0680.0440.0280.0760.046
10000.0810.0490.0280.0090.0310.014 0.0950.0560.0230.0030.0320.008
401250.1050.1010.0980.0960.1080.103 0.1060.0990.0960.0960.1100.105
2500.0980.0860.0770.0700.0970.083 0.1010.0840.0710.0630.1010.086
5000.0990.0760.0590.0390.0790.055 0.1070.0780.0510.0290.0870.054
10000.0870.0540.0330.0120.0360.018 0.1000.0620.0270.0060.0360.012
0.8101250.1140.0930.0870.0840.1380.102 0.1050.0780.0740.0740.1520.099
2500.0690.0370.0270.0230.0640.033 0.0660.0210.0150.0150.0670.023
5000.0480.0170.0080.0030.0190.006 0.0470.0040.0000.0010.0090.000
10000.0390.0110.0040.0010.0050.001 0.0400.0020.0000.0000.0000.000
201250.1120.0820.0730.0690.1400.096 0.1030.0630.0550.0540.1550.095
2500.0690.0330.0200.0140.0630.029 0.0650.0150.0070.0050.0650.017
5000.0470.0150.0060.0000.0170.004 0.0460.0030.0020.0030.0070.001
10000.0380.0110.0040.0000.0050.001 0.0390.0020.0000.0000.0000.000
401250.1090.0750.0610.0540.1400.092 0.0990.0520.0410.0360.1560.090
2500.0710.0330.0200.0110.0650.029 0.0670.0150.0060.0050.0670.018
5000.0510.0200.0110.0050.0220.009 0.0500.0070.0030.0030.0110.003
10000.0390.0120.0050.0010.0060.002 0.0400.0020.0010.0010.0010.001
Note. ε 0.2 , ε 0.4  = computed ε values in loss function ρ ε such that ρ ε y ( V ¯ ) = y for y = 0.2 and y = 0.4 , where V ¯ is the square root of the average variance of the item difficulty difference b ^ i 2 b ^ i 1 ; absolute biases larger than 0.03 are printed in bold.
Table 3. Simulation Study: Relative root mean square error (RMSE) of estimated mean μ ^ as a function of the DIF effect size κ , number of items I, and sample size N.
Table 3. Simulation Study: Relative root mean square error (RMSE) of estimated mean μ ^ as a function of the DIF effect size κ , number of items I, and sample size N.
κ I N Ratio Function, ε = Gaussian Function, ε =
0.250.100.050.01 ε 0.2 ε 0.4 0.250.100.050.01 ε 0.2 ε 0.4
0.41012583.590.595.010080.286.3 83.092.999.0105.279.184.0
25083.488.193.310083.789.4 83.289.196.6105.883.288.1
50092.689.091.010089.192.0 95.689.892.3103.890.591.6
1000123.1102.594.810095.596.2 134.4108.194.6102.497.594.0
2012583.288.693.810081.885.1 83.090.998.1105.681.483.6
25085.986.991.310085.787.8 86.487.093.7104.586.486.5
500101.993.491.710094.092.1 106.394.191.0103.296.890.7
1000141.1111.197.710099.295.7 155.9118.495.8100.4100.791.7
4012585.688.993.510085.286.5 85.489.897.1105.985.285.6
25091.589.791.610091.189.9 92.489.392.5104.492.489.3
500120.2105.797.9100107.296.8 126.2107.194.9101.5112.295.7
1000169.0127.9106.9100109.198.3 187.6137.6103.099.7110.894.2
0.81012587.292.696.010085.689.3 87.995.9100.5106.186.888.7
25090.389.492.810089.690.2 91.489.194.6106.291.489.0
50094.887.289.810087.190.7 97.784.087.5105.384.686.6
1000101.587.988.410088.093.7 106.285.686.5102.487.291.9
2012588.790.995.010091.688.7 88.592.297.6104.794.488.2
25094.688.390.910092.988.5 94.885.690.8104.994.785.2
50097.286.387.610086.688.6 99.083.786.1104.184.285.5
1000106.890.590.210090.094.0 109.888.188.7102.488.391.4
4012592.190.794.010098.390.2 90.289.495.3103.8102.788.8
250101.090.891.210098.690.4 100.387.590.1104.0100.187.6
500105.490.389.710090.890.2 106.187.087.9104.987.787.6
1000116.494.893.310093.395.4 118.692.092.2101.891.993.7
Note. ε 0.2 , ε 0.4  = computed ε values in loss function ρ ε such that ρ ε y ( V ¯ ) = y for y = 0.2 and y = 0.4 , where V ¯ is the square root of the average variance of the item difficulty difference b ^ i 2 b ^ i 1 . The ratio loss function with ε = 0.01 was chosen as the reference method when computing the relative RMSE. Cells with the smallest RMSE values are printed with a gray background color.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Robitzsch, A. Computational Aspects of L0 Linking in the Rasch Model. Algorithms 2025, 18, 213. https://doi.org/10.3390/a18040213

AMA Style

Robitzsch A. Computational Aspects of L0 Linking in the Rasch Model. Algorithms. 2025; 18(4):213. https://doi.org/10.3390/a18040213

Chicago/Turabian Style

Robitzsch, Alexander. 2025. "Computational Aspects of L0 Linking in the Rasch Model" Algorithms 18, no. 4: 213. https://doi.org/10.3390/a18040213

APA Style

Robitzsch, A. (2025). Computational Aspects of L0 Linking in the Rasch Model. Algorithms, 18(4), 213. https://doi.org/10.3390/a18040213

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop