Next Article in Journal
Investigating the Effectiveness of Self-Regulated, Pomodoro, and Flowtime Break-Taking Techniques Among Students
Previous Article in Journal
How the Human–Artificial Intelligence (AI) Collaboration Affects Cyberloafing: An AI Identity Perspective
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Reliability Generalization Meta-Analysis of the Antisocial Process Screening Device

1
Department of Psychology, Guangzhou University, Guangzhou 510006, China
2
Department of Psychology, University of Bath, Bath BA2 7AY, UK
3
Department of Applied Social Sciences, The Hong Kong Polytechnic University, Hong Kong 999077, China
4
School of Psychology & Neuroscience, University of Glasgow, Glasgow G12 8QB, UK
5
Saw Swee Hock School of Public Health, National University of Singapore and National University Health System, Singapore 117549, Singapore
6
Lloyd’s Register Foundation Institute for the Public Understanding of Risk, National University of Singapore, Singapore 117602, Singapore
7
School of Medicine and Psychology, Australian National University, Canberra 2600, Australia
*
Author to whom correspondence should be addressed.
Behav. Sci. 2025, 15(7), 860; https://doi.org/10.3390/bs15070860 (registering DOI)
Submission received: 22 March 2025 / Revised: 25 May 2025 / Accepted: 13 June 2025 / Published: 25 June 2025
(This article belongs to the Section Child and Adolescent Psychiatry)

Abstract

Although the Antisocial Process Screening Device (APSD) is one of the most widely used instruments for assessing psychopathic traits, the reliability estimates of the APSD show great heterogeneity across different studies. This investigation evaluated the reliability of the APSD using a reliability generalization meta-analytic technique across 158 studies (N = 75,749). The APSD demonstrated marginal to acceptable coefficient alphas ranging from 0.62 (for the Callous–Unemotional subscale) to 0.79 (for the total scale). Further moderation analysis revealed that the differences in administration formats significantly explained the variance of coefficient alphas for the APSD total and subscales, and the self-report version of the APSD manifested poorer coefficient alphas than other-report versions. The standard deviation of scale scores also partly accounted for the variance of the coefficient alphas. Overall, the APSD was found to be a reliable, practical measurement of psychopathic traits in youth, which can be widely applied in various study settings. Nevertheless, we recommend that parent- and teacher-report versions of the APSD as the preferred administering format of the measure when used for children and juveniles, while the self-report version of the APSD is recommended with caution when assessing youth psychopathy, unless multiple-assess methods are used.

1. Introduction

Psychopathic traits have features including interpersonal antagonism (e.g., superficial charm, egocentricity), distinctive affective dysfunctions (e.g., low empathy, remorselessness), and behavioral disinhibition (e.g., impulsivity, poor planning; Hare & Neumann, 2008; Patrick et al., 2009). Over the last decades, psychopathy as a concept has been extended to include children and adolescents (Frick, 2009; Lynam et al., 2007; Salekin, 2017; Salekin & Lynam, 2010). Consistent empirical evidence has shown that psychopathy in youth is closely correlated with aggression, disruptive behavior disorders, delinquency (e.g., fraud, theft), and greater criminal recidivism (e.g., Bijttebier & Decoene, 2009; Laajasalo et al., 2016; Muñoz & Frick, 2007; Ooi et al., 2017; Poore et al., 2020; Poythress et al., 2006b), and can predict adult psychopathy (Lynam et al., 2007). Precise assessment of children and adolescent psychopathy has important theoretical and clinical implications on understanding the relationship between psychopathic traits and conduct problems (Colins et al., 2018; Salekin, 2016), investigating clinical and predictive utility of externalizing behaviors (Salekin, 2017), and guiding diagnostic classifications of childhood disruptive behavior disorders and developing targeted interventions (Frick, 2022). Several instruments were designed for use in children and adolescents, such as the Antisocial Process Screening Device (APSD; Frick & Hare, 2001), Youth Psychopathic Traits Inventory—Child Version (YPI-CV; van Baardewijk et al., 2008), and the Inventory of Callous–Unemotional Traits (ICU; Frick, 2003) which specifically captures the affective dimension of psychopathy. Among these instruments, the APSD was modeled directly after the Psychopathy Checklist-Revised (PCL-R; Hare, 2003). To date, the APSD (Frick & Hare, 2001; Frick et al., 1994) is one of the most widely used measures to assess psychopathic traits in children and adolescents. Nevertheless, reliability estimates of some of the APSD subscales are questionable (e.g., Dong et al., 2014; Poore et al., 2020; Wang et al., 2015). The present study focused on evaluating the overall reliability of the APSD scores across various test settings using Reliability Generalization (RG) analysis.

1.1. The Antisocial Process Screening Device (APSD)

The APSD is one of the most widely multi-informant measures for assessing psychopathy and antisocial behaviors in children aged 6 to 13 (Johnstone & Cooke, 2004; Kotler & McMahon, 2010; Salekin, 2017). To date, the APSD has been translated into various languages and used in a variety of countries (Eremsoy et al., 2011; Roose et al., 2011; Roşan et al., 2015; Shaffer et al., 2016; Wang et al., 2015). The items were adapted from the PCL-R (Hare, 2003) to be appropriate for children and youths. The APSD was initially developed as a scale rated by parents and teachers regarding target preadolescents (Frick et al., 1994). A self-report version was then proposed and favored in older children and adolescents (Lilienfeld & Andrews, 1996; Lynam et al., 1999), as the self-report scale’s reliability and validity were substantially greater for older children and adolescents than for younger children. The self-report scale is necessary when parents or teachers have little information about the target individual’s psychopathy, particularly in detained youths (Frick et al., 2000). The parent- and self-report ratings of the APSD have been shown to have moderate associations in previous studies of adolescents from different cultural contexts (e.g., Dutch, de Wied et al., 2014; U.S., detained juveniles, Falkenbach et al., 2003; Russian, Grigorenko et al., 2010; U.S., non-referred children, Muñoz & Frick, 2007).
The subscales of the 20-item APSD vary in the existing literature due to inconsistent factor structures which have been reported across different samples. Frick et al. (1994) observed a three-factor structure comprising the original Callous-Emotional (CU; six items; e.g., “Good at keeping promises”, “Is concerned about the feelings of others”), Narcissism (NAR; seven items; e.g., “Uses or cons others”, “Blames others for mistakes”), and Impulsivity (IMP; five items; “Engages in risky or dangerous activities”, “Gets bored easily”) in both clinical and community samples in the United States (Frick et al., 2000). It has been argued that this three-factor structure is ideal for distinguishing the different components of psychopathy (Kotler & McMahon, 2010), and it has been supported by empirical evidence in both clinical and community samples (e.g., Bijttebier & Decoene, 2009; Fite et al., 2009; Goodwin et al., 2015; Laajasalo et al., 2016; Shaffer et al., 2016). Thus, the current study focused on the three-factor model of the ASPD.

1.2. Internal Consistency Reliability of the APSD Scores

Reliability denotes the proportion of observed scores that are true scores (Anastasi & Urbina, 1997; Botella et al., 2010). Reliability indicates the precision of the measurement and, in the case of the APSD scores, can influence the precision of assessing psychopathic traits (American Educational Research Association et al., 2014; Green et al., 2011).
Internal consistency is the most common form of reliability used in the literature (Henson & Thompson, 2002), reflecting the degree to which test items scores capture a single latent construct coherently (Geisinger, 2013; Revelle & Condon, 2017). Low internal consistency may suggest a significant proportion of measurement errors in or variability among the items. Cronbach’s alpha is one of the most commonly used measurements of internal consistency reliability reported for the APSD in previous studies.
The internal consistency reliability of APSD total and subscales have generally been in marginally to moderately acceptable levels. For example, Sellbom and Phillips (2013) obtained poor coefficient alphas for the APSD scores from a college student sample, ranging from 0.42 (for CU) to 0.78 (for total scale). In another study, the coefficient alphas for the subscales also ranged from poor to marginal levels (0.53 for IMP to 0.66 for NAR; Javdani et al., 2011). However, the CU subscale tends to show the weakest reliability scores overall. For example, in one study, a coefficient alpha of only 0.26 was observed for the CU scores (Colins et al., 2017) while a large body of other studies reported coefficient alphas which were under acceptable levels (e.g., Fung et al., 2010; Vitacco et al., 2003). Some studies (e.g., Colins et al., 2012; Deng et al., 2019; Suárez-Alvarez et al., 2018) have suggested that factors such as the small number of items in the subscales, a combination of positively and negatively worded items, inaccurate expression of items, and the affective, behavioral, and cognitive features of juveniles, may contribute to the low CU and IMP coefficient alphas.
Meanwhile, Cronbach’s alpha values for the APSD have varied across test settings and sample characteristics. Coefficient alphas of the APSD total score have been most homogenous for the teacher-report version (APSD-TR; ranging from 0.86 to 0.97; Barrutieta & Prieto-Ursúa, 2015; Eremsoy et al., 2011), followed by the parent-report version (APSD-PR; from 0.73 to 0.89; Eremsoy et al., 2011; Frick et al., 2003), and the other-report version (APSD-OR; 0.73 to 0.88, from psychiatrists or other caregivers; de Wied et al., 2014; Murrie et al., 2004). The coefficient alpha had the greatest variability for the self-report version (APSD-SR; from 0.62 to 0.88; Gillen et al., 2019; Vitacco et al., 2003). In terms of sample characteristics, coefficient alphas of the total scale scores ranged from 0.73 to 0.91 when adolescent samples with a clinical diagnosis were the research targets of studies (e.g., Ooi et al., 2017; Thomson & Centifanti, 2018), from 0.70 to 0.97 when studies used community samples (e.g., Barrutieta & Prieto-Ursúa, 2015; Garland et al., 2009), from 0.68 to 0.88 when studies used detained samples (e.g., Murrie et al., 2004; Vitacco et al., 2003), and from 0.74 to 0.88 when mixed samples were used (e.g., Feilhauer et al., 2012; Javdani et al., 2011). Likewise, sample size may also influence the coefficient alphas, for example, de Wied et al. (2014) reported low coefficient alphas, particularly for the CU (0.54) and IMP (0.30) subscales in a small sample of 74 adolescents. Finally, the alphas of the APSD total scores might be influenced by scale language. Statistical support exists from intervals of alphas in both English (0.62 to 0.92; Frick et al., 2003; Vitacco et al., 2003) and non-English (e.g., Finnish version: 0.67 to 0.76, Laajasalo et al., 2014; Portuguese version: 0.56 to 0.77; Pechorro et al., 2014; Persian version: 0.46 to 0.69; Ebrahimi et al., 2021) versions of the APSD.
The heterogeneity of the reliability estimates of the APSD may limit its predictive utility in research settings (Shaffer et al., 2016), especially considering the low internal consistency reliability. Therefore, a systematic investigation is needed to understand the potential factors that are likely to contribute to this poor reliability, and to provide evidence for whether the APSD is in fact an appropriate tool to be used in certain testing contexts.

1.3. The Current Study

The current study aimed to evaluate the internal consistency reliability (i.e., coefficient alphas) of the APSD and its subscale scores across different test settings using an RG meta-analysis. RG is a meta-analytic methodology which synthesizes reliability estimates across different studies (Vacha-Haase et al., 2002). It offers a more compositive and steadier reliability estimate than any statistic derived from a single research sample. An RG study on the APSD helps assess the degree of variability in reliability coefficients across different types of measures, samples, and settings, and offers the opportunity to advance our understanding of factors that can affect reliability, thus allowing researchers and practitioners to be more informed when using the APSD in assessing juvenile psychopathy.
The current study had several intentions. First, we tested the average coefficient alpha for the APSD total and subscale scores. Based on previous empirical studies (e.g., Crapanzano et al., 2011; Stellwagen & Kerig, 2010; Van Zwieten et al., 2013), we hypothesized that the total scores would exhibit an acceptable mean coefficient alpha while the CU subscale would have a relatively low mean coefficient alpha. Second, we examined the heterogeneity of the coefficient alphas across all studies, and expected to observe significant heterogeneity among them. Third, we investigated whether the heterogeneity of the coefficient alpha estimates could be explained by study or sample characteristics, such as administration formats, language of the measure, or sample age. As the self-report measures have already been shown to have advantages, such as more precisely detecting interpersonal and affective deficits (Lilienfeld & Fowler, 2006), we expected that the APSD-SR would have higher coefficient alpha values than the other-report versions. Additionally, some research used translated versions of the APSD that had limited evidence for its translation quality (e.g., pilot studies or studies that tested the translation such as whether items were understandable and interpretable; Tsang et al., 2017). In these cases, participants may not have correctly or precisely understood the translated APSD items (e.g., Bijttebier & Decoene, 2009; Eremsoy et al., 2011). Therefore, it is reasonable to hypothesize that the alpha values of the APSD would be higher for the English version of the measure than for non-English versions.
For the sample characteristics, older adolescents generally have been shown to have a better understanding of the APSD items than younger individuals. Therefore, we expected that the alpha values would be lower for samples that had a lower mean age than for samples that had a higher mean age. Moreover, a positive relationship between the standard deviation of scores and the coefficient alphas was anticipated, on account of the classical test theory which contends that higher standard deviations of scores will cause larger coefficient alpha values when the standard error of measurement is constant (Crocker & Algina, 1986). Finally, we explored the potential effects of other sample characteristics, including mean scores, gender composition, and standard deviation of participants.

2. Method

2.1. Literature Search and Selection

A literature search was performed using EBSCOhost, Elsevier, Google Scholar, PsycARTICLES, Psychology and Behavioral Sciences Collection, PsycINFO, and Springer, using the terms “Antisocial Process Screening Device”, “APSD”, and “Antisocial Process Screening Device and reliability”. The references of the retrieved studies were also checked. After discarding duplicates, studies were included based on the following three criteria: (a) studies were published in either Chinese or English; (b) they were empirical studies; and (c) they reported a sample-specific reliability estimate. Figure 1 shows the details of the literature selection process. A total of 158 studies which identified 361 coefficient alphas altogether were included in the final analysis. There were 97, 104, 82, and 78 coefficient alphas for the APSD total scale, CU, NAR, and IMP scales, respectively.

2.2. Data Extraction and Coding

To test how the potential study characteristics might have accounted for the variability of reliability, a range of information reported in the studies was extracted and coded. Coded variables at the categorical level were as follows: (a) scale version (i.e., APSD-PR, APSD-SR, APSD-TR, and APSD-OR); (b) country (i.e., United States vs. other countries); (c) scale language (i.e., English vs. non-English); and (d) sample type (i.e., clinical, community, detained, and mixed samples). The coded continuous variables were (e) mean age of target adolescent samples; (f) standard deviation of the age of target adolescent samples; (g) gender composition of the target sample (male%); and (h) sample size. Studies lacking information for these characteristics were labeled as being with missing values.1 A predictive mean matching approach (PMM) was applied to replace missing data in continuous characteristics (i.e., characteristics [e] to [h]) with observations from complete cases which had similar characteristics but required further imputation. The PMM provides an unbiased estimate when sample sizes are small, or when data are seriously skewed (Kleinke, 2017). The variables that were used for matching these cases were from characteristics (e) to (h).
Along with these moderator variables, coefficient alpha values for the APSD total score and subscales were extracted based on three criteria. First, only the coefficient alphas of the 20-item APSD full scale and the 6-item CU, 7-item NAR, and 5-item IMP subscales were extracted to be included in the current analysis. Second, if the study was a longitudinal study, only the coefficient alphas collected at the baseline were extracted, given that the coefficient alpha would likely artificially increase due to the practice effect (American Educational Research Association et al., 2014).
Third, when only one coefficient alpha was reported based on multiple samples, the sample characteristics of the multiple samples would be aggregated for that single alpha value. Supplemental Table S2 presents the details of all the information extracted from the included studies.
The coding process was completed by four trained psychology students. To ensure the accuracy of the coded dataset, 20% of the observations and the corresponding information was selected at random and independently double-checked by two of the authors. Disagreements during the double-checking session were resolved through discussion and consensus.

2.3. Analytical Strategies

All analyses with untransformed coefficient alphas in the current study were carried out using the package metafor 2.4 (Viechtbauer, 2010) and mice 3.11 (Buuren & Groothuis-Oudshoorn, 2010) in R program 3.3.3 for MacOS. Separate meta-analyses were managed for the reliability coefficients obtained from the APSD total scale and for each subscale of each measurement instrument. In all cases, random-effects models were used (Borenstein et al., 2009; López-López et al., 2013; Sánchez-Meca et al., 2013).
Heterogeneity was first estimated among the reliability coefficients in each meta-analysis using the Q statistic and I2 index, which were calculated as the quantification of the overall heterogeneity across studies. A p value for the Q statistic < 0.05 is an indicator of heterogeneity among coefficient alphas (Higgins et al., 2003), while I2 values of approximately 25%, 50%, and 75% refer to low, moderate, and considerable heterogeneity across studies, respectively (Huedo-Medina et al., 2006). To identify possible influential cases or outliers, the leave-one-out diagnoses were conducted. The leave-one-out diagnoses used studentized residuals, Cook’s distances, and changes in Q and I2. The studentized residuals and Cook’s distances were based on a mixed-effect linear meta-regression model without potential moderators. Potential influential cases were identified by inspecting the plots of the studentized residuals and Cook’s distances. We verified the influence of the identified cases by further examining the changes in Q and I2 for models for estimating mean coefficient alphas across scales and models for moderation analysis.
After examining heterogeneity existing across reliability coefficients, moderator analysis using mixed effect models was applied to quantify the degree to which the variability of coefficient alphas could be accounted for by study and sample characteristics. The effects of moderators were tested based on the Knapp and Hartung’s method (Knapp & Hartung, 2003; see also López-López et al., 2013; Viechtbauer et al., 2015). The statistical significance of the association was reflected by p values of QB and t statistics for categorical and continuous moderators. The R2 statistic estimates the proportion of variance which is accounted for by the moderator variables (López-López et al., 2014).
Finally, the publication bias, namely file-drawer problems, was examined by funnel plots and the trim-and-fill method (Duval & Tweedie, 2000). Funnel plots for the APSD total scale and subscales were constructed to allow the threat of missing possible effect sizes to be assessed graphically. Meanwhile, the trim-and-fill method provided a null hypothesis test which assumes the number of missing studies (which refers to coefficient alphas in the current study) is zero, while Begg’s rank test (Begg & Mazumdar, 1994) examined whether the observed coefficient alphas correlated to the corresponding sampling errors. A high Kendall’s tau rank correlation coefficient computed by Begg’s rank test can indicate an asymmetric funnel plot, implying a potential publication bias.

3. Results

3.1. Description of the Data

The literature search identified 361 coefficient alphas from 158 studies, with a total sample size of 75,749. Most coefficient alphas were observed from scores of the APSD-SR total scale and subscales, with a total number (k) of 189 coefficient alphas (52.35%), followed by the scores of the APSD-PR (k = 87, 24.10%). Of the 361 coefficient alphas, only a small number of them were obtained from translated or adapted versions of the APSD (k = 84, 23.27%), and a total of 189 coefficient alphas (52.35%) were from studies conducted in the United States. Finally, 49.58% of the coefficient alphas (k = 179) were reported from community samples, while 24.38% (k = 88) were from detained samples, 11.36% were from mixed samples (k = 41), 13.30% were from clinical samples (k = 48), and 1.34% (k = 5) did not specify the sample type.

3.2. Mean Reliability and Heterogeneity

As shown in Table 1, the overall (weighted) coefficient alphas ranged from 0.62 (CU) to 0.79 (total scale). These alpha coefficients indicated a statistically significant heterogeneity across studies with significant Q values of between 848.965 and 4106.680 (all ps < 0.001), and I2 values exceeding 90%.
A detailed inspection of the data indicated two possible influential studies accounting for this considerable heterogeneity. First, the Barrutieta and Prieto-Ursúa (2015) study was identified as an influential case for the APSD total scores. This research reported the highest coefficient alpha (0.97) among all studies observing coefficient alphas from total scores. This study was the only study that used the teacher/psychologist version of the APSD, while other other-report studies only used the teacher-report version. Results from the leave-one-out diagnoses marked this study as an influential case with a residual term (3.503) from the mixed-effect model, together with a large Cook’s distance (0.094), compared to other observations. Its removal led to reduced heterogeneity QQ = 2683.501) and I2I2% = 1.93) indices.
Another potential influential case was the de Wied et al. (2014) study, which administrated the IMP subscale in a community adolescent sample. This study reported a rather low coefficient alpha of the IMP subscale (i.e., 0.30) and contributed to a great amount of heterogeneity. The leave-one-out diagnoses also indicated that this study could be an outlier and an influential case, reflected by a large residual term from the mixed-effect model (−3.686) and Cook’s distance (0.184). Excluding the de Wied et al. (2014) study led to decreases in QQ = 40.532), I2I2% = 1.03), and tau22 = 0.005 with its removal) values for the IMP (see Table 1).
However, both studies met our inclusion criteria and did not substantially influence the findings in our analyses, including the pooled effect sizes (i.e., mean coefficient alphas). Thus, both studies were retained in the analyses. Sensitivity analyses omitting Barrutieta and Prieto-Ursúa for total score and de Wied et al. (2014) for IMP were conducted, and the results are presented in the Supplemental Materials.

3.3. Moderator Analysis

Table 2 presents the estimated alpha values for APSD based on administration formats. We found that the APSD-TR had the highest mean coefficient alphas (0.90 for total scale, 0.73 for CU, 0.85 for NAR, and 0.77 for IMP), while the APSD-SR had the lowest mean coefficient alphas, ranging from 0.51 for CU to 0.77 for the APSD total scale. The administration format accounted for 43.90% (k = 97) of the total variance of the alpha value of the total scale (p < 0.001), 47.40% (k = 104) for the CU (p < 0.001), 56.70% (k = 82) for the NAR (p < 0.001), and 44.50% (k = 78) for the IMP scales (p < 0.001).
Studies conducted in the United States yielded similar mean coefficient alphas (i.e., 0.79 for total scale, 0.62 for CU, 0.65 for IMP, and 0.74 for NAR) to those conducted in other countries, such as Belgium, China, and Turkey (i.e., 0.78 for total scale, 0.61 for CU, 0.66 for IMP, and 0.73 for NAR). Likewise, there were no significant differences in the total APSD and subscale scores between the original English version and other translated versions (i.e., Chinese, Dutch, Portuguese versions).
As demonstrated in Table 3, only the sample type explained a statistically significant amount of variance in the reliability estimates for the CU (p = 0.044), accounting for 7.10% (k = 104). Although the effects were not statistically significant for the total, IMP, and NAR estimates, clinical samples generally yielded relatively higher mean coefficient alphas for total scale (i.e., 0.82), NAR (i.e., 0.77), and IMP (i.e., 0.67) than non-clinical samples. The detained samples showed relatively lower mean coefficient alphas for the total scale (i.e., 0.78), CU (i.e., 0.56), and NAR (i.e., 0.72), but higher for IMP (i.e., 0.64) compared with mixed samples. Regarding the continuous variables analyzed using meta-regression (see Table 4), the standard deviation of scores showed significant positive relationships with the coefficient alphas for all scales (estimates ranging from 0.023 for the total scale to 0.060 for IMP, all ps < 0.05), except for CU (b = 0.040, p = 0.058). Standard deviation accounted for 20.70%, 15.90%, and 32.50% of the variance of the coefficient alphas for the total scale, NAR, and IMP, respectively. The mean age of the adolescents had a significantly negative association with the coefficient alphas for the APSD subscales, with estimates ranging from −0.013 for IMP (p < 0.001) to −0.005 for the total score (p = 0.009), and accounting for 8.50% (total score), 8.40% (CU), 23.70% (NAR), and 28.70% (IMP) of the variance of the coefficient alphas. The standard deviation of the assessed participants’ age was found to have a weak but significant positive association with the coefficient alpha for IMP with an estimate of 0.024 (p = 0.044), accounting for 5.70% of the variance. None of the other continuous variables significantly explained the variance of the coefficient alphas.

3.4. Explanatory Models

Weighted multiple regression analysis was conducted assuming mixed-effect models for the purpose of identifying the best set of moderator sets which would explain the variance of the coefficient alphas for the APSD total and subscales.
For the APSD total, the explanatory model initially included the administration format, mean of adolescents’ age, and the standard deviation of the scores as three predictors. Although the full model was statically significant (k = 41, F = 3.993, p < 0.001), the estimates of mean age and standard deviation of scores were not significant (p = 0.798 for the mean age and p = 0.444 for the standard deviation of scores). Therefore, only the administration format was retained in the final explanatory model as indicated in the meta-ANOVA results.
As for CU, the administration format, sample type, and mean age of the target adolescent sample were included in the multiple meta-regression analysis. When the administration format was a part of the model, neither of the remaining two moderators or their combinations remained statistically significant. As such, the administration format was the only moderator for this explanatory model.
With regards to NAR, three moderators, including administration format, mean age of target adolescents, and standard deviation of subscale scores, were included in the model. In the subsequent analysis, the regression estimate of the age of target adolescents was not significant. The final model which incorporated the administration format and standard deviation of NAR scores was statistically significant (k = 52, F = 17.881, p < 0.001, accounting for 66.22% of the total variance).
For IMP, the administration format, mean age of target adolescents, standard deviation of adolescents’ age, and the standard deviation of observed scores were included in the initial explanatory model. Only the standard deviation of observed score was significant (p = 0.006). Hence, this is the only moderator retained in the explanatory model, with the results presented in Table 4.
Overall, the results of the model misspecification test in all explanatory models were statistically significant (ps < 0.001), which indicated that there were other possible moderators also affecting the coefficient alpha variabilities of the APSD total and subscale scores.

3.5. Publication Bias

Publication bias in our results was detected by constructing funnel plots, as exhibited in Figure 2, and then trim-and-fill analysis (Duval & Tweedie, 2000) was used to examine the symmetry of the funnel plots. Results demonstrated that no missing study was found to violate the symmetry of the funnel plots with regards to the total score, NAR, or IMP, with acceptances of the null hypotheses (all ps = ns). However, the trim-and-fill method indicated 13 coefficient alphas on the funnel plot for CU (indicated by white dots), along with the rejection of the null hypothesis that no missing coefficient alphas on the right side of the funnel plot would affect its symmetry (p < 0.001). This could be due to very small sample sizes or high standard errors (Shi & Lin, 2019). The result of Begg’s rank test for CU was also statistically significant (Kendall’s τ = −0.208, p = 0.002). Thus, publication bias may be present for the CU subscale.

4. Discussion

The current RG meta-analysis evaluated the coefficient alphas of APSD, and examined the factors that might contribute to the variability of its reliability coefficients. The internal consistency coefficients as indexed by Cronbach’s alpha of the APSD fluctuated across different application contexts.
We found that the APSD total scores had an adequate average coefficient alpha, while the CU and IMP subscale scores had less satisfactory coefficient alphas. The low alpha values of these two subscales could be due in part to the small number of items in these subscales (van Baardewijk et al., 2010; Colins et al., 2012) or insufficient coverage of the symptoms by the limited items.
For the CU subscale, two factors might contribute to the low alpha values. First, the CU subscale has a mixture of positively and negatively worded items, which may artificially create a multidimensionality structure of the scale, thus also contributing to the low coefficient alpha value (Suárez-Alvarez et al., 2018). Second, it might be difficult for participants and observers to detect unemotional traits, which are mostly about an absence of emotions. Absence of emotions may also be mistaken for other factors such as anhedonia, or the subject being withdrawn or shy (Cardinale & Marsh, 2017).
The low coefficient alpha of the IMP subscale, particularly for the self-report version, might be due to some of the symptoms measured by the subscale items being confounded by factors other than impulsiveness. For example, Item 1 (“Blames others for their mistakes”) has strong implications of interpersonal or social outcomes. This item could be influenced by impulsiveness, general risk-seeking in the social domain, or social boldness. On the other hand, items such as Item 13 (“Engages in risky and dangerous behavior”) and Item 17 (“Does not plan ahead”) place the emphasis on the occurrence of the behaviors rather than on the motivation of the individual. In addition, Deng et al. (2019) have suggested that adolescents might not accurately assess whether they are engaging in dangerous activities or not recognize potential risks due to their limited cognitive ability. Adolescent examinees may misread the test items or misjudge their behavioral engagement.

4.1. Moderators of the Coefficient Alphas

We found that differences in the administration formats of the measure significantly explained the heterogeneity in coefficient alphas in the APSD total as well as in its subscales. The self-report version of the APSD presented the lowest average Cronbach’s alpha. Self-reports have been assumed to outperform other versions only when subjects can and do accurately assess their symptoms and behaviors (Van Riezen & Segal, 1988). However, youths may have difficulty identifying and reporting their personal feelings correctly (Clark & Watson, 1991) which may increase the measure error of the self-report versions of the ASPD. On the other hand, as the APSD items focus on adolescent behaviors, observers (e.g., parents, teachers) demonstrate more a reliable assessment of overt behavioral symptoms, while adolescent self-reports show limitations in evaluating their own symptoms due to their incapacity to self-monitor (Frick et al., 2000; Poythress et al., 2006a). These findings underscore the critical role of informant discrepancies in child psychopathology assessment, highlighted by the Attribution Bias Context Model (ABC Model) (De Los Reyes & Kazdin, 2005; De Los Reyes et al., 2015). For example, older children and adolescents typically demonstrate greater self-awareness of their psychological symptoms compared to younger children, leading to improved concordance between self- and observer-reports. Moreover, the magnitude of discrepancy may vary by symptom domain: externalizing problems generally show higher inter-rater agreement due to their observable nature, whereas internalizing problems exhibit greater discrepancies as they rely more heavily on subjective experience. These patterns highlight the need for future research utilizing comprehensive multi-informant, longitudinal designs to elucidate the nature of observed discrepancies in psychopathology assessment.
We also found that sample type had a significant moderation effect on the heterogeneity in the alpha values in the CU subscale. The detained samples had the lowest and most unacceptable mean coefficient alphas. This could be because detained youths often tend to display more fraudulent or defensive response styles to conceal their true thoughts or sentiments (Colins et al., 2014). Detained samples also tend to have a relatively low level of education (Lankester et al., 2025), which limits their comprehension abilities when it comes to the CU items. Furthermore, some CU items might not be applicable to incarcerated juveniles. For example, Item 3 (“Concerned about schoolwork”) and Item 20 (“Keeps few of the same friends”) focus on interpersonal situations in schools and communities, which may be less applicable to detained samples. Finally, all alpha values for the detained samples were based on the self-report version CU subscale (see Supplemental Table S2), which could also be another reason for the low alpha values for detained samples. This was also reflected in the fact that the sample type effect diminished when the scale version was entered into the explanatory model.
As for the continuous variables, the standard deviation of the subscale scores significantly explained the variance of the coefficient alphas in both NAR and IMP subscales. This finding is consistent with the predictions of psychometric theories that the larger the standard deviation of the score, the larger the coefficient alpha (Crocker & Algina, 1986). More specifically, the standard deviation is the square root of the variance, and the alpha is larger if the value of variance is larger according to the formula for the Cronbach’s alpha. Therefore, the standard deviation of the subscale may be a moderator for the variation in each scale’s coefficient alpha.
Furthermore, the mean age of the target adolescents was negatively associated with the coefficient alphas in all three APSD subscales. However, when included in the final explanatory model, this association became insignificant. Nevertheless, Fung et al. (2010) have noted an insignificant correlation between psychopathic traits and age within a large sample of East Asian schoolchildren. More research, particularly longitudinal studies, are needed to further understand this correlation.

4.2. Publication Bias

Publication bias was found in terms of the reported alpha values of the CU subscale, with an underreporting of high alpha values, specifically. In light of this, the mean coefficient alpha of the CU subscale should be regarded as an underestimated value. The possibility for obtaining a high coefficient alpha for the CU subscale was limited because of the reasons mentioned above, which include mixed item wording, the limited number of items, and the CU trait’s low construct stability. Overall, caution is needed when interpreting the results of CU alpha values, and more research into this subscale specifically is needed.

4.3. Recommendations, Limitations, and Future Directions

The average coefficient alphas for the APSD total and subscale scores provide a guide as to the assessment precision of the APSD across different test settings. The results of the explanatory models can also help researchers evaluate the factors that may affect the coefficient alphas of the APSD scores. According to our findings, the APSD has relatively acceptable internal consistency reliability, and both the total and subscale scores of the parent-report and teacher-report versions of the APSD exhibit good alpha reliability. We recommend that the informant-report version of the APSD be preferred when using the APSD among youths, and that multiple-assess methods should be considered (Achenbach et al., 1987; Ooi et al., 2017; Sadeh et al., 2009).
Some limitations should be noted for the current study. We focused primarily on the studies that adopted the three-factor structure of the APSD. Future research could consider other forms of the APSD to test the generalizability of the current findings. Publication bias in the CU subscale’s alpha values also limits the generalization of the results of this study. Finally, the test of model misspecification indicated statistically significant residual heterogeneity, which demonstrated that moderators other than those included in the current study may also affect the variance of the coefficient alphas. Future studies should investigate a wider range of potential moderators.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/bs15070860/s1, Table S1: Model Specification of the Frick et al. (2000) Three-Factor Solution of the APSD; Table S2: Information Extracted from Studies Included in Meta-Analysis.

Author Contributions

Conceptualization, M.-C.W. and Y.S.; methodology, M.-C.W. and Y.S.; formal analysis, J.D., X.Z. and J.L.; data curation, J.D., X.Z. and J.L.; writing—original draft preparation, M.-C.W., J.D., X.Z. and J.L.; writing—review and editing, M.-C.W. and Y.S.; project administration, M.-C.W. and J.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Note

1
Emails were sent to the corresponding authors of articles that had missing values. The absence of reply from some authors yielded a dataset with 88 studies which including missing values.

References

  1. Achenbach, T. M., McConaughy, S. H., & Howell, C. T. (1987). Child/adolescent behavioral and emotional problems: Implications of cross-informant correlations for situational specificity. Psychological Bulletin, 101(2), 213–232. [Google Scholar] [CrossRef] [PubMed]
  2. American Educational Research Association, American Psychological Association & National Council on Measurement in Education. (2014). Standards for educational and psychological testing (2014th ed.). American Educational Research Association (AERA). [Google Scholar]
  3. Anastasi, A., & Urbina, S. (1997). Psychological testing (7th ed.). Prentice Hall. [Google Scholar]
  4. Barrutieta, L. H., & Prieto-Ursúa, M. (2015). Neurophysiological indicators of emotional processing in youth psychopathy. Psicothema, 27(3), 235–240. [Google Scholar] [CrossRef] [PubMed]
  5. Begg, C. B., & Mazumdar, M. (1994). Operating characteristics of a rank correlation test for publication bias. Biometrics, 50(4), 1088–1101. [Google Scholar] [CrossRef] [PubMed]
  6. Bijttebier, P., & Decoene, S. (2009). Assessment of psychopathic traits in children and adolescents: Further validation of the antisocial process screening device and the childhood psychopathy scale. European Journal of Psychological Assessment, 25(3), 157–163. [Google Scholar] [CrossRef]
  7. Borenstein, M., Hedges, L. V., Higgins, J. P. T., & Rothstein, H. R. (2009). Introduction to meta-analysis. Wiley. [Google Scholar] [CrossRef]
  8. Botella, J., Suero, M., & Gambara, H. (2010). Psychometric inferences from a meta-analysis of reliability and internal consistency coefficients. Psychological Methods, 15(4), 386–397. [Google Scholar] [CrossRef]
  9. Buuren, S. V., & Groothuis-Oudshoorn, K. (2010). mice: Multivariate imputation by chained equations in R. Journal of Statistical Software, 45(3), 1–67. [Google Scholar] [CrossRef]
  10. Cardinale, E. M., & Marsh, A. A. (2017). The reliability and validity of the inventory of callous unemotional traits: A meta-analytic review. Assessment, 27(1), 57–71. [Google Scholar] [CrossRef]
  11. Clark, L. A., & Watson, D. (1991). Tripartite model of anxiety and depression: Psychometric evidence and taxonomic implications. Journal of Abnormal Psychology, 100(3), 316–336. [Google Scholar] [CrossRef]
  12. Colins, O. F., Andershed, H., Salekin, R. T., & Fanti, K. A. (2018). Comparing different approaches for subtyping children with conduct problems: Callous-unemotional traits only versus the multidimensional psychopathy construct. Journal of Psychopathology and Behavioral Assessment, 40, 6–15. [Google Scholar] [CrossRef]
  13. Colins, O. F., Bijttebier, P., Broekaert, E., & Andershed, H. (2014). Psychopathic-like traits among detained female adolescents: Reliability and validity of the antisocial process screening device and the youth psychopathic traits inventory. Assessment, 21(2), 195–209. [Google Scholar] [CrossRef]
  14. Colins, O. F., Noom, M., & Vanderplasschen, W. (2012). Youth psychopathic traits inventory-short version: A further test of the internal consistency and criterion validity. Journal of Psychopathology and Behavioral Assessment, 34, 476–486. [Google Scholar] [CrossRef] [PubMed]
  15. Colins, O. F., Van Damme, L., Fanti, K. A., & Andershed, H. (2017). The prospective usefulness of callous–unemotional traits and conduct disorder in predicting treatment engagement among detained girls. European Child & Adolescent Psychiatry, 26, 75–85. [Google Scholar] [CrossRef]
  16. Crapanzano, A. M., Frick, P. J., Childs, K., & Terranova, A. M. (2011). Gender differences in the assessment, stability, and correlates to bullying roles in middle school children. Behavioral Sciences & the Law, 29(5), 677–694. [Google Scholar] [CrossRef]
  17. Crocker, L., & Algina, J. (1986). Introduction to classical and modern test theory. Holt, Rinehart & Winston. [Google Scholar]
  18. De Los Reyes, A., Augenstein, T. M., Wang, M., Thomas, S. A., Drabick, D. A. G., Burgers, D. E., & Rabinowitz, J. (2015). The validity of the multi-informant approach to assessing child and adolescent mental health. Psychological Bulletin, 141(4), 858–900. [Google Scholar] [CrossRef]
  19. De Los Reyes, A., & Kazdin, A. E. (2005). Informant discrepancies in the assessment of childhood psychopathology: A critical review, theoretical framework, and recommendations for further study. Psychological Bulletin, 131(4), 483–509. [Google Scholar] [CrossRef]
  20. Deng, J., Wang, M.-C., Zhang, X., Shou, Y., Gao, Y., & Luo, J. (2019). The inventory of callous unemotional traits: A reliability generalization meta-analysis. Psychological Assessment, 31(6), 765–780. [Google Scholar] [CrossRef]
  21. de Wied, M., van der Baan, H., Raaijmakers, Q., de Ruiter, C., & Meeus, W. (2014). Factor structure and construct validity of the Dutch version of the Antisocial Process Screening Device. Journal of Psychopathology and Behavioral Assessment, 36, 84–92. [Google Scholar] [CrossRef]
  22. Dong, L., Wu, H., & Waldman, I. D. (2014). Measurement and structural invariance of the antisocial process screening device. Psychological Assessment, 26(2), 598–608. [Google Scholar] [CrossRef]
  23. Duval, S., & Tweedie, R. (2000). Trim and fill: A simple funnel-plot-based method of testing and adjusting for publication bias in meta-analysis. Biometrics, 56(2), 455–463. [Google Scholar] [CrossRef]
  24. Ebrahimi, A., Athar, M. E., Darvishi, M., & Colins, O. F. (2021). The Persian self-report version of the Antisocial Process Screening Device (APSD-P): A psychometric evaluation. Frontiers in Psychiatry, 12, 760531. [Google Scholar] [CrossRef]
  25. Eremsoy, C. E., Karanci, A. N., & Berument, S. K. (2011). Psychometric properties of the antisocial process screening device in a non-clinical sample of Turkish children. Procedia-Social and Behavioral Sciences, 30, 2319–2323. [Google Scholar] [CrossRef]
  26. Falkenbach, D. M., Poythress, N. G., & Heide, K. M. (2003). Psychopathic features in a juvenile diversion population: Reliability and predictive validity of two self-report measures. Behavioral Sciences and the Law, 21(6), 787–805. [Google Scholar] [CrossRef] [PubMed]
  27. Feilhauer, J., Cima, M., & Arntz, A. (2012). Assessing callous–unemotional traits across different groups of youths: Further cross-cultural validation of the inventory of callous–unemotional traits. International Journal of Law and Psychiatry, 35(4), 251–262. [Google Scholar] [CrossRef]
  28. Fite, P. J., Greening, L., Stoppelbein, L., & Fabiano, G. A. (2009). Confirmatory factor analysis of the antisocial process screening device with a clinical inpatient population. Assessment, 16(1), 103–114. [Google Scholar] [CrossRef] [PubMed]
  29. Frick, P. J. (2003). The inventory of callous-unemotional traits. The University of New Orleans. [Google Scholar]
  30. Frick, P. J. (2009). Extending the construct of psychopathy to youth: Implications for understanding, diagnosing, and treating antisocial children and adolescents. The Canadian Journal of Psychiatry, 54(12), 803–812. [Google Scholar] [CrossRef]
  31. Frick, P. J. (2022). Some critical considerations in applying the construct of psychopathy to research and classification of childhood disruptive behavior disorders. Clinical Psychology Review, 96, 102188. [Google Scholar] [CrossRef]
  32. Frick, P. J., Bodin, S. D., & Barry, C. T. (2000). Psychopathic traits and conduct problems in community and clinic-referred samples of children: Further development of the psychopathy screening device. Psychological Assessment, 12(4), 382–393. [Google Scholar] [CrossRef]
  33. Frick, P. J., & Hare, R. (2001). The antisocial process screening device (APSD): Technical manual. Multi-Health Systems. [Google Scholar]
  34. Frick, P. J., Kimonis, E. R., Dandreaux, D. M., & Farell, J. (2003). The 4 year stability of psychopathic traits in non-referred youth. Behavioral Sciences & the Law, 21(4), 713–736. [Google Scholar] [CrossRef]
  35. Frick, P. J., O’Brien, B. S., Woorton, J., & McBurnett, K. (1994). Psychopathy and conduct problems in children. Journal of Abnormal Psychology, 103(4), 700–707. [Google Scholar] [CrossRef]
  36. Fung, A., Gao, Y., & Raine, A. (2010). The utility of the child and adolescent psychopathy construct in Hong Kong, China. Journal of Clinical Child & Adolescent Psychology, 39(1), 134–140. [Google Scholar] [CrossRef]
  37. Garland, E. L., Howard, M. O., & Perron, B. E. (2009). Nitrous oxide inhalation among adolescents: Prevalence, correlates, and co-occurrence with volatile solvent inhalation. Journal of Psychoactive Drugs, 41(4), 337–347. [Google Scholar] [CrossRef] [PubMed]
  38. Geisinger, K. F. (2013). Reliability. In K. F. Geisinger (Ed.), APA handbook of testing and assessment in psychology: Vol. 1. Test theory and testing and assessment in industrial and organizational psychology (pp. 21–41). American Psychological Association. [Google Scholar] [CrossRef]
  39. Gillen, C. T. A., MacDougall, E. A. M., Forth, A. E., Barry, C. T., & Salekin, R. T. (2019). Validity of the Youth Psychopathic Traits Inventory–Short Version in justice-involved and at-risk adolescents. Assessment, 26(3), 479–491. [Google Scholar] [CrossRef] [PubMed]
  40. Goodwin, B. E., Sellbom, M., & Salekin, R. T. (2015). Elucidating the construct validity of the antisocial process screening device (APSD) in a sample of young adults. Journal of Psychopathology and Behavioral Assessment, 37, 1–11. [Google Scholar] [CrossRef]
  41. Green, C. E., Chen, C. E., Helms, J. E., & Henze, K. T. (2011). Recent reliability reporting practices in Psychological Assessment: Recognizing the people behind the data. Psychological Assessment, 23(3), 656–669. [Google Scholar] [CrossRef]
  42. Grigorenko, E. L., Geiser, C., Slobodskaya, H. R., & Francis, D. J. (2010). Cross-informant symptoms from CBCL, TRF, and YSR: Trait and method variance in a normative sample of Russian youths. Psychological Assessment, 22(4), 893–911. [Google Scholar] [CrossRef] [PubMed]
  43. Hare, R. D. (2003). Hare psychopathy checklist-revised (PCL-R): Technical manual (2nd ed.). Multi-Health Systems Inc. [Google Scholar]
  44. Hare, R. D., & Neumann, C. (2008). Psychopathy as a clinical and empirical construct. Annual Review of Clinical Psychology, 4, 217–246. [Google Scholar] [CrossRef]
  45. Henson, R. K., & Thompson, B. (2002). Characterizing measurement error in scores across studies: Some recommendations for conducting “reliability generalization” studies. Measurement and Evaluation in Counseling and Development, 35(2), 113–127. [Google Scholar] [CrossRef]
  46. Higgins, J. P. T., Thompson, S. G., Deeks, J. J., & Altman, D. G. (2003). Measuring inconsistency in meta-analyses. British Medical Journal, 327(7414), 557–560. [Google Scholar] [CrossRef]
  47. Huedo-Medina, T. B., Sánchez-Meca, J., Marín-Martínez, F., & Botella, J. (2006). Assessing heterogeneity in meta-analysis: Q statistic or I2 index? Psychological Methods, 11(2), 193–206. [Google Scholar] [CrossRef]
  48. Javdani, S., Sadeh, N., & Verona, E. (2011). Suicidality as a function of impulsivity, callous–unemotional traits, and depressive symptoms in youth. Journal of Abnormal Psychology, 120(2), 400–413. [Google Scholar] [CrossRef]
  49. Johnstone, L., & Cooke, D. J. (2004). Psychopathic-like traits in childhood: Conceptual and measurement concerns. Behavioral Sciences and the Law, 22(1), 103–125. [Google Scholar] [CrossRef]
  50. Kleinke, K. (2017). Multiple imputation under violated distributional assumptions: A systematic evaluation of the assumed robustness of predictive mean matching. Journal of Educational and Behavioral Statistics, 42(4), 371–404. [Google Scholar] [CrossRef]
  51. Knapp, G., & Hartung, J. (2003). Improved tests for a random effects meta-regression with a single covariate. Statistics in Medicine, 22(17), 2693–2710. [Google Scholar] [CrossRef] [PubMed]
  52. Kotler, J. S., & McMahon, R. J. (2010). Assessment of child and adolescent psychopathy. In R. T. Salekin, & D. R. Lynam (Eds.), Handbook of child and adolescent psychopathy. The Guilford Press. [Google Scholar]
  53. Laajasalo, T., Aronen, E. T., Saukkonen, S., Salmi, V., Aaltonen, M., & Kivivuori, J. (2016). To tell or not to tell? Psychopathic traits and response integrity in youth delinquency surveys. Criminal Behaviour and Mental Health, 26(2), 81–93. [Google Scholar] [CrossRef]
  54. Laajasalo, T., Saukkonen, S., Kivivuori, J., Salmi, V., Lipsanen, J., & Aronen, E. T. (2014). Brief report: Self-reported psychopathic-like features among Finnish community youth: Investigation of the factor structure of the antisocial personality screening device. Journal of Adolescence, 37(7), 1185–1188. [Google Scholar] [CrossRef]
  55. Lankester, M., Coles, C., Trotter, A., Scott, S., Downs, J., Dickson, H., & Wickersham, A. (2025). The association between academic achievement and subsequent youth offending: A systematic review and meta-analysis. Journal of Developmental and Life-Course Criminology, 10, 477–500. [Google Scholar] [CrossRef] [PubMed]
  56. Lilienfeld, S. O., & Andrews, B. P. (1996). Development and preliminary validation of a self-report measure of psychopathic personality traits in noncriminal populations. Journal of Personality Assessment, 66(3), 488–524. [Google Scholar] [CrossRef]
  57. Lilienfeld, S. O., & Fowler, K. A. (2006). The self-report assessment of psychopathy: Problems, pitfalls, and promises. In C. J. Patrick (Ed.), Handbook of psychopathy (pp. 107–132). Guilford Press. [Google Scholar]
  58. López-López, J. A., Botella, J., Sánchez-Meca, J., & Marín-Martínez, F. (2013). Alternatives for mixed-effects meta-regression models in the reliability generalization approach: A simulation study. Journal of Educational and Behavioral Statistics, 38(5), 443–469. [Google Scholar] [CrossRef]
  59. López-López, J. A., Marín-Martínez, F., Sánchez-Meca, J., Van den Noortgate, W., & Viechtbauer, W. (2014). Estimation of the predictive power of the model in mixed-effects meta-regression: A simulation study. British Journal of Mathematical and Statistical Psychology, 67(1), 30–48. [Google Scholar] [CrossRef]
  60. Lynam, D. R., Caspi, A., Moffitt, T. E., Loeber, R., & Stouthamer-Loeber, M. (2007). Longitudinal evidence that psychopathy scores in early adolescence predict adult psychopathy. Journal of Abnormal Psychology, 116(1), 155–165. [Google Scholar] [CrossRef]
  61. Lynam, D. R., Whiteside, S., & Jones, S. (1999). Self-reported psychopathy: A validation study. Journal of Personality Assessment, 73(1), 110–132. [Google Scholar] [CrossRef] [PubMed]
  62. Muñoz, L. C., & Frick, P. J. (2007). The reliability, stability, and predictive utility of the self-report version of the antisocial process screening device. Scandinavian Journal of Psychology, 48(4), 299–312. [Google Scholar] [CrossRef]
  63. Murrie, D. C., Cornell, D. G., Kaplan, S., McConville, D., & Levy-Elkon, A. (2004). Psychopathy scores and violence among juvenile offenders: A multi-measure study. Behavioral Sciences & the Law, 22(1), 49–67. [Google Scholar] [CrossRef]
  64. Ooi, Y. P., Glenn, A. L., Ang, R. P., Vanzetti, S., Falcone, T., Gaab, J., & Fung, D. S. S. (2017). Agreement between parent- and self-reports of psychopathic traits and externalizing behaviors in a clinical sample. Child Psychiatry and Human Development, 48, 151–165. [Google Scholar] [CrossRef]
  65. Patrick, C. J., Fowles, D. C., & Krueger, R. F. (2009). Triarchic conceptualization of psychopathy: Developmental origins of disinhibition, boldness, and meanness. Development and Psychopathology, 21(3), 913–938. [Google Scholar] [CrossRef] [PubMed]
  66. Pechorro, P., Gonçalves, R. A., Marôco, J., Gama, A. P., Neves, S., & Nunes, C. (2014). Juvenile delinquency and psychopathic traits: An empirical study with Portuguese adolescents. International Journal of Offender Therapy and Comparative Criminology, 58(2), 174–189. [Google Scholar] [CrossRef]
  67. Poore, H. E., Watts, A. L., Lilienfeld, S. O., & Waldman, I. D. (2020). Construct validity of youth psychopathic traits as assessed by the antisocial process screening device. Psychological Assessment, 32(9), 527–540. [Google Scholar] [CrossRef] [PubMed]
  68. Poythress, N. G., Dembo, R., Wareham, J., & Greenbaum, P. E. (2006a). Construct validity of the youth psychopathic traits inventory (YPI) and the antisocial process screening device (APSD) with justice-involved adolescents. Criminal Justice and Behavior, 33(1), 26–55. [Google Scholar] [CrossRef]
  69. Poythress, N. G., Douglas, K. S., Falkenbach, D., Cruise, K., Lee, Z., Murrie, D. C., & Vitacco, M. (2006b). Internal consistency reliability of the self-report Antisocial Process Screening Device. Assessment, 13(1), 107–113. [Google Scholar] [CrossRef]
  70. Revelle, W., & Condon, D. M. (2017). Reliability. In P. Irwing, T. Booth, & D. J. Hughes (Eds.), The Wiley handbook of psychometric testing: A multidisciplinary reference on survey, scale and test development (Vol. 2, pp. 709–749). Wiley Blackwell. [Google Scholar]
  71. Roose, A., Bijttebier, P., Claes, L., & Lilienfeld, S. O. (2011). Psychopathic traits in adolescence: Associations with the revised reinforcement sensitivity theory systems. Personality and Individual Differences, 50(2), 201–205. [Google Scholar] [CrossRef]
  72. Roşan, A., Frick, P. J., Gottlieb, K. A., & Faşicaru, L. (2015). Callous-unemotional traits and anxiety in a sample of detained adolescents in Romania. Journal of Evidence-Based Psychotherapies, 15(1), 79–95. [Google Scholar]
  73. Sadeh, N., Verona, E., Javdani, S., & Olson, L. (2009). Examining psychopathic tendencies in adolescence from the perspective of personality theory. Aggressive Behavior, 35(5), 399–407. [Google Scholar] [CrossRef] [PubMed]
  74. Salekin, R. T. (2016). Psychopathy in childhood: Toward better informing the DSM–5 and ICD-11 conduct disorder specifiers. Personality Disorders: Theory, Research, and Treatment, 7(2), 180–191. [Google Scholar] [CrossRef]
  75. Salekin, R. T. (2017). Research Review: What do we know about psychopathic traits in children? Journal of Child Psychology and Psychiatry, 58(11), 1180–1200. [Google Scholar] [CrossRef]
  76. Salekin, R. T., & Lynam, D. R. (2010). Handbook of child and adolescent psychopathy. Guilford Press. [Google Scholar]
  77. Sánchez-Meca, J., López-López, J. A., & López-Pina, J. A. (2013). Some recommended statistical analytic practices when reliability generalization (RG) studies are conducted. British Journal of Mathematical and Statistical Psychology, 66(3), 402–425. [Google Scholar] [CrossRef] [PubMed]
  78. Sellbom, M., & Phillips, T. R. (2013). An examination of the triarchic conceptualization of psychopathy in incarcerated and nonincarcerated samples. Journal of Abnormal Psychology, 122(1), 208–214. [Google Scholar] [CrossRef]
  79. Shaffer, C., Gatner, D., Gray, A. L., Douglas, K. S., Viljoen, J. L., Tweed, R., Bhatt, G., Dooley, S., & Gagnon, N. (2016). Incremental and predictive validity of the antisocial process screening device in a community sample of male and female ethnic minority and Caucasian youth. Journal of Abnormal Child Psychology, 44, 1599–1612. [Google Scholar] [CrossRef]
  80. Shi, L., & Lin, L. (2019). The trim-and-fill method for publication bias: Practical guidelines and recommendations based on a large database of meta-analyses. Medicine (Baltimore), 98(23), e15987. [Google Scholar] [CrossRef]
  81. Stellwagen, K., & Kerig, P. K. (2010). Relating callous-unemotional traits to physically restrictive treatment measures among child psychiatric inpatients. Journal of Child & Family Studies, 19, 588–595. [Google Scholar] [CrossRef]
  82. Suárez-Alvarez, J., Pedrosa, I., Lozano, L. M., García-Cueto, E., Cuesta, M., & Muñiz, J. (2018). Using reversed items in Likert scales: A questionable practice. Psicothema, 30(2), 149–158. [Google Scholar] [CrossRef]
  83. Thomson, N. D., & Centifanti, L. C. M. (2018). Proactive and reactive aggression subgroups in typically developing children: The role of executive functioning, psychophysiology, and psychopathy. Child Psychiatry and Human Development, 49, 197–208. [Google Scholar] [CrossRef] [PubMed]
  84. Tsang, S., Royse, C. F., & Terkawi, A. S. (2017). Guidelines for developing, translating, and validating a questionnaire in perioperative and pain medicine. Saudi Journal of Anaesthesia, 11(Suppl. S1), S80–S89. [Google Scholar] [CrossRef] [PubMed]
  85. Vacha-Haase, T., Henson, R. K., & Caruso, J. C. (2002). Reliability generalization: Moving toward improved understanding and use of score reliability. Educational and Psychological Measurement, 62(4), 562–569. [Google Scholar] [CrossRef]
  86. van Baardewijk, Y., Andershed, H., Stegge, H., Nilsson, K. W., Scholte, E., & Vermeiren, R. (2010). Development and tests of short versions of the youth psychopathic traits inventory and the youth psychopathic traits inventory-child version. European Journal of Psychological Assessment, 26(2), 122–128. [Google Scholar] [CrossRef]
  87. van Baardewijk, Y., Stegge, H., Andershed, H., Thomaes, S., Scholte, E., & Vermeiren, R. (2008). Measuring psychopathic traits in children through self-report. The development of the youth psychopathic traits inventory-child version. International Journal of Law and Psychiatry, 31(3), 199–209. [Google Scholar] [CrossRef] [PubMed]
  88. Van Riezen, H., & Segal, M. (1988). Comparative evaluation of rating scales for clinical psychopharmacology. Elsevier. [Google Scholar]
  89. Van Zwieten, A., Meyer, J., Hermens, D. F., Hickie, I. B., Hawes, D. J., Glozier, N., Naismith, S. L., Scott, E. M., Lee, R. S. C., & Guastella, A. J. (2013). Social cognition deficits and psychopathic traits in young people seeking mental health treatment. PLoS ONE, 8, e67753. [Google Scholar] [CrossRef]
  90. Viechtbauer, W. (2010). Conducting meta-analyses in r with the metafor package. Journal of Statistical Software, 36(3), 1–48. [Google Scholar] [CrossRef]
  91. Viechtbauer, W., López-López, J. A., Sánchez-Meca, J., & Marín-Martínez, F. (2015). A comparison of procedures to test for moderators in mixed-effects meta-regression models. Psychological Methods, 20(3), 360–374. [Google Scholar] [CrossRef]
  92. Vitacco, M. J., Rogers, R., & Neumann, C. S. (2003). The antisocial process screening device: An examination of its construct and criterion-related validity. Assessment, 10(2), 143–150. [Google Scholar] [CrossRef]
  93. Wang, M.-C., Deng, Q., Armour, C., Bi, X., & Zeng, H. (2015). The psychometric properties and factor structure of the antisocial process screening device self-report version in chinese adolescents. Journal of Psychopathology and Behavioral Assessment, 37, 553–562. [Google Scholar] [CrossRef]
Figure 1. Flow diagram of study selection process.
Figure 1. Flow diagram of study selection process.
Behavsci 15 00860 g001
Figure 2. Funnel plots of total scale, callous–unemotional subscale, narcissism subscale, and impulsivity subscale. Notes. Black dots = observed studies; white dots = imputed missing studies estimated by the trim-and-fill method.
Figure 2. Funnel plots of total scale, callous–unemotional subscale, narcissism subscale, and impulsivity subscale. Notes. Black dots = observed studies; white dots = imputed missing studies estimated by the trim-and-fill method.
Behavsci 15 00860 g002
Table 1. Mean effect sizes and heterogeneity statistics of the APSD.
Table 1. Mean effect sizes and heterogeneity statistics of the APSD.
kα+95% CITau2QI2%H2
Total Score970.79[0.77, 0.80]0.0034106.680 ***94.6518.68
Total Score a960.78[0.77, 0.80]0.0031423.179 ***92.7213.74
Callous–Unemotional1040.62[0.59, 0.64]0.0121007.952 ***92.2212.86
Narcissism820.73[0.72, 0.75]0.0041379.95 ***94.1617.12
Impulsivity780.65[0.63, 0.67]0.006848.965 ***92.5213.37
Impulsivity b770.66[0.64, 0.68]0.005808.883 ***91.5211.79
Notes. k = Number of coefficient alphas; α+ = Mean coefficient alpha; 95% CI = 95% confidence interval for α+; Q = Heterogeneity statistic, df = (k − 1); Tau2 = Estimated total heterogeneity; I2 = Heterogeneity index. a = Results of impulsivity with Barrutieta and Prieto-Ursúa (2015) study; b = Results of impulsivity without de Wied et al. (2014) study. *** p < 0.001.
Table 2. Comparison of coefficient alphas of the subscales across categorical variables.
Table 2. Comparison of coefficient alphas of the subscales across categorical variables.
Total ScaleCallous–UnemotionalNarcissismImpulsivity
kα+95% CIkα+95% CIkα+95% CIkα+95% CI
Administration Format
APSD-PR150.81[0.78, 0.84]310.62[0.58, 0.65]200.75[0.73, 0.78]210.69[0.66, 0.72]
APSD-SR700.77[0.76, 0.78]330.51[0.48, 0.55]440.70[0.69, 0.71]420.62[0.60, 0.64]
APSD-TR50.90[0.88, 0.92]150.73[0.69, 0.77]90.85[0.83, 0.87]70.77[0.74, 0.81]
APSD-OR70.85[0.79, 0.90]250.68[0.65, 0.71]90.76[0.70, 0.81]80.65[0.56, 0.74]
Country
United States430.79[0.77, 0.81]580[0.59, 0.65]440.74[0.72, 0.76]440.65[0.63, 0.68]
Other540.78[0.77, 0.80]460.61[0.57, 0.65]380.73[0.71, 0.75]340.66[0.62, 0.69]
Scale Language
English500.79[0.77, 0.81]740.62[0.59, 0.65]580.74[0.72, 0.76]580.65[0.63, 0.68]
Non-English370.79[0.77, 0.80]210.59[0.54, 0.65]140.73[0.69, 0.76]120.66[0.61, 0.70]
Sample Type
Clinical Sample60.82[0.77, 0.87]200.64[0.58, 0.70]100.77[0.73, 0.82]120.67[0.63, 0.72]
Community Sample450.78[0.77, 0.80]520.61[0.57, 0.64]420.73[0.70, 0.75]400.66[0.63, 0.69]
Detained Sample320.78[0.75, 0.80]150.56[0.49, 0.63]210.72[0.69, 0.75]200.64[0.61, 0.67]
Mixed Sample130.79[0.77, 0.82]170.67[0.63, 0.71]60.76[0.71, 0.80]50.62[0.52, 0.71]
Notes. k = Number of coefficient alphas for each subgroup of the moderator variable; α+ = Weighted mean coefficient alpha for each subgroup of the moderator variable; APSD-OR refers to the other-report versions of the APSD (i.e., staff-report, psychiatrist-report, and caregiver-report); Other countries refers to countries such as Australia, China, The Netherlands, etc. (see Supplemental Table S2); Non-English language refers to languages such as Chinese, Dutch, Portuguese, etc. (see Supplemental Table S2).
Table 3. Meta-ANOVA by categorical moderators.
Table 3. Meta-ANOVA by categorical moderators.
Total ScoreCallous–UnemotionalNarcissismImpulsivity
R2QBpR2QBpR2QBpR2QBp
Administration Format0.43960.412<0.0010.47471.297<0.0010.56776.635<0.0010.44540.292<0.001
Country0.0050.3270.5670.0010.2870.5920.0060.3520.5530.0010.0200.888
Language0.0000.0010.9790.011.0800.2990.0090.2750.6000.0000.0050.946
Sample Type0.0262.4830.4780.0718.0900.0440.0595.1610.1600.0322.0470.563
Notes. R2 = Proportion of variance explained by the moderator variable; QB = Between-group heterogeneity test; p = p value for the statistical tests of QB. Significant results are in bold.
Table 4. Results of simple meta-regression analysis.
Table 4. Results of simple meta-regression analysis.
kbtpR2QE
Total Score
Age Mean87−0.005−2.6140.0090.0851023.132 ***
Age SD860.0050.7970.4250.0081155.398 ***
Scores Mean47−0.003−1.6920.0910.065567.180 ***
Scores SD460.0233.0240.0020.207448.175 ***
Male % of sample910.000−0.0490.9610.0003011.995 ***
Sample Size970.000−1.2820.2000.0253462.726 ***
Callous–Unemotional
Age Mean95−0.009−2.7560.0060.084740.735 ***
Age SD94−0.003−0.1940.8460.002751.111 ***
Scores Mean640.0050.5880.5560.002572.123 ***
Scores SD610.0401.8990.0580.048550.730 ***
Male % of sample980.0011.5630.1180.021891.213 ***
Sample Size1040.000−1.4400.1500.030963.491 ***
Narcissism
Age Mean73−0.009−4.3360.0000.237672.882 ***
Age SD730.0151.6320.1030.0341172.357 ***
Scores Mean52−0.006−0.9420.3460.024643.070 ***
Scores SD520.0402.9680.0030.159609.404 ***
Male % of sample770.000−0.1390.8890.0001269.825 ***
Sample Size820.0000.6770.4980.0031258.529 ***
Impulsivity
Age Mean70−0.013−4.3030.0000.287438.220 ***
Age SD of Age700.0242.0130.0440.057733.898 ***
Scores Mean51−0.011−1.5750.1150.062478.451 ***
Scores SD490.0604.0680.0000.325254.423 ***
Male % of sample730.0000.2950.7680.002757.510 ***
Sample Size780.0001.0780.2810.015834.698 ***
Notes. k = Number of coefficient alphas for each subgroup of the moderator variable; b = unstandardized regression coefficient; t = significance test of moderator regression coefficient; p = p value of significance test, significant results are in bold; R2 = Total that proportion of variance accounts for; QE = Statistic for test of residual heterogeneity. *** p < 0.001.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, M.-C.; Deng, J.; Zhang, X.; Liang, J.; Shou, Y. A Reliability Generalization Meta-Analysis of the Antisocial Process Screening Device. Behav. Sci. 2025, 15, 860. https://doi.org/10.3390/bs15070860

AMA Style

Wang M-C, Deng J, Zhang X, Liang J, Shou Y. A Reliability Generalization Meta-Analysis of the Antisocial Process Screening Device. Behavioral Sciences. 2025; 15(7):860. https://doi.org/10.3390/bs15070860

Chicago/Turabian Style

Wang, Meng-Cheng, Jiaxin Deng, Xintong Zhang, Jinghui Liang, and Yiyun Shou. 2025. "A Reliability Generalization Meta-Analysis of the Antisocial Process Screening Device" Behavioral Sciences 15, no. 7: 860. https://doi.org/10.3390/bs15070860

APA Style

Wang, M.-C., Deng, J., Zhang, X., Liang, J., & Shou, Y. (2025). A Reliability Generalization Meta-Analysis of the Antisocial Process Screening Device. Behavioral Sciences, 15(7), 860. https://doi.org/10.3390/bs15070860

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop