Next Article in Journal
How Workplace Telepressure Fuels Job Burnout Among Educators: Mediated by Work-Related Rumination and Moderated by Perceived Organizational Support
Previous Article in Journal
A Process Evaluation of the UK Randomised Trial Evaluating ‘iSupport’, an Online e-Health Intervention for Adult Carers of People Living with Dementia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Factor Performance Validity?—Network and Factor Structure of Performance Validity Measures in the Clinical Evaluation of Adult ADHD

by
Emily Raasch
1,2,
Anselm B. M. Fuermaier
1,*,
Johanna Kneidinger
2,
Björn Albrecht
2 and
Hanna Christiansen
2
1
Department of Clinical and Developmental Neuropsychology, University of Groningen, 9712 TS Groningen, The Netherlands
2
Department of Clinical Child and Adolescent Psychology and Psychotherapy, Philipps University Marburg, 35032 Marburg, Germany
*
Author to whom correspondence should be addressed.
Behav. Sci. 2025, 15(8), 1108; https://doi.org/10.3390/bs15081108
Submission received: 9 June 2025 / Revised: 7 August 2025 / Accepted: 12 August 2025 / Published: 15 August 2025

Abstract

Performance validity tests (PVTs) and symptom validity tests (SVTs) are central to evaluating neuropsychological test results in clinical adult ADHD assessments. Although their relationships have been widely examined, the constructs these measures assess remain poorly understood in applied contexts. This study investigates the conceptual similarities and distinctions of performance validity measures, i.e., the Groningen Effort Test (GET), the Medical Symptom Validity Test (MSVT), and the Amsterdam Short-Term Memory (ASTM) test, within a comprehensive diagnostic battery for adult ADHD. The diagnostic battery included symptom self-reports and a continuous performance test (CPT). Network and factor analyses investigated these relationships. A three-factor structure was hypothesized, consisting of (1) performance validity measures, (2) continuous performance measures, and (3) symptom reports (including embedded SVTs). Data from a large clinical referral sample (N = 461) of adults with suspected ADHD were analyzed to explore these constructs. Network analysis revealed that the PVTs did not form a cohesive network with CPT measures. Symptom reports, including embedded SVTs, formed their own cluster, separate from performance-based attention measures. Factor analysis rejected a unified construct of performance validity measures. Regression analysis showed that cognitive deficits, education level, and impulsivity predicted ASTM test performance, whilst the MSVT and GET did not. These findings suggest that PVTs should be interpreted in the context of ADHD assessment, particularly in high-stakes forensic evaluations, where the accuracy of performance evaluation is critical. Future research should explore multidimensional models of performance validity, addressing domain-specific underperformance and individual variability in ADHD evaluations.

1. Introduction

Attention-deficit/hyperactivity disorder (ADHD) is a frequent neurodevelopmental disorder, marked by the three core symptoms of inattention, hyperactivity, and impulsivity (American Psychiatric Association, 2013). ADHD is increasingly understood as a potential lifespan disorder that begins in childhood. Research indicates that 50% to 65% of individuals diagnosed in childhood continue to experience symptoms in adulthood (Faraone et al., 2000; Martel et al., 2016). Unfortunately, many cases go undiagnosed until early or mid-adulthood, making it harder to diagnose early onset when childhood ADHD was not established (Sibley, 2021). Therefore, diagnosing adult ADHD requires a thorough evaluation of both current and early-life symptoms, along with an assessment of impairment (Onandia-Hinchado et al., 2021; Sibley et al., 2018). While the core symptoms of ADHD may persist into adulthood, they often manifest in more subtle and diverse ways. These include ongoing challenges with sustained attention, disorganization, inattention to detail, distractibility, and forgetfulness, with motor hyperactivity typically decreasing over time (Zalsman & Shilton, 2016). If left untreated, these described difficulties can lead to significant impairments across multiple areas of daily functioning, such as academic aspirations (Barkley et al., 2006), occupational issues (Adamou et al., 2013), and difficulties in social–emotional contexts (Klein et al., 2012). Furthermore, comorbidity rates of up to 80% are common, often involving substance use disorders, depression, anxiety, and personality disorders (Adamou et al., 2013; Choi et al., 2022). Given this clinical complexity, it is essential to accurately assess the consistency of information provided by, or about, first-time adult ADHD clients.
To address this need, empirically informed guidelines emphasize a combination of approaches and instruments to ensure a reliable diagnosis. These are designed to enhance accuracy and reduce the risk of false positives and negatives. Such approaches include (structured) diagnostic interviews, following DSM criteria, and standardized self-report rating scales (Sibley, 2021). Additionally, adult ADHD is associated with cognitive difficulties that affect various aspects of attention and executive control, which, while common in ADHD, are not specific to the disorder (Guo et al., 2023; Onandia-Hinchado et al., 2021). Neuropsychological tests, particularly the continuous performance test (CPT), are, however, often used to assess sustained attention in individuals suspected of ADHD (Albrecht et al., 2015; Arrondo et al., 2024). Integrating the CPT with other diagnostic tools enhances the diagnostic process and supports ADHD treatment planning (Fuermaier et al., 2022; Gustafsson & Hansen, 2023).
However, these clinical tools are only useful in diagnosis and treatment if they accurately reflect true symptomatology and cognitive abilities. Arising concerns over the credibility of self-reported symptoms and neuropsychological test performance in adult ADHD evaluations have led to the implementation of performance validity tests (PVTs) and symptom validity tests (SVT; Hirsch et al., 2022; Musso & Gouvier, 2014; Tucha et al., 2015). Performance validity refers to the assessment of an individual’s actual task performance, which reflects their true cognitive abilities. In contrast, symptom validity refers to the extent to which an individual’s self-and observer-reported symptoms accurately reflect their actual experiences (Larrabee, 2012; Van Dyke et al., 2013). Invalid results of PVTs and/or SVTs may suggest that a patient’s test profile is uninterpretable, but it does not provide insight into the underlying cause of this failure (Merten & Merckelbach, 2013). This is especially problematic in adult ADHD, where individuals may report ADHD-like symptoms to gain perceived benefits, such as non-medical use of stimulant medication, excuse-making behaviors (Harrison & Edwards, 2010; Sattler et al., 2025; Sullivan et al., 2007), or advantages in academic or workplace settings (Adamou et al., 2013). Thus, distorted symptom reports and test performance can indeed compromise diagnostic assessments, misguide treatment plans, and reduce adherence (Hoelzle et al., 2019; Roor et al., 2022).
Although SVT research in adult ADHD assessments is less extensive, it is crucial. This is evident, as studies show that common self-report scales often fail to distinguish genuine from simulated ADHD (Tucha et al., 2015). Recently, Ovsiew et al. (2023) reported that 22% of adults undergoing ADHD evaluations failed symptom validity measures. This highlights the urgent need for effective detection methods (Fuermaier et al., 2024). A promising SVT is the Infrequency Index (CII) that is embedded in the Conners’ Adult ADHD Rating Scale (Conners et al., 1999; Suhr et al., 2011).
In parallel, PVT research has shown that these tests, alongside SVTs, assess different but complementary constructs. Hence, both should be measured simultaneously in clinical evaluations (Hirsch et al., 2022; Ovsiew et al., 2023). Drawing from recent large-scale research, estimations of base rates of cognitive underperformance in adult ADHD assessments range from 9 to up to 27% following the currently claimed criterion for underperformance of ≥2 failed stand-alone PVTs (Hirsch et al., 2022; Mascarenhas et al., 2023; Ovsiew et al., 2023; Phillips et al., 2023). Stand-alone PVTs offer the highest classification accuracy for detecting noncredible performance in neuropsychological assessments, as they are specifically designed for this purpose (Hoelzle et al., 2019; Wallace et al., 2019). Examples of such tests include verbal memory tests (e.g., Word Memory Test; P. Green, 2003), non-verbal memory tests (e.g., Test of Memory Malingering; Tombaugh, 1996), and working memory tests (e.g., Rey Dot Counting Test; Boone et al., 2002). Still, the majority of stand-alone PVTs are not specifically developed for the assessment of adults with ADHD. An exception includes the Groningen Effort Test (GET), serving especially as an attention-based PVT (Fuermaier et al., 2021). In general, it is recommended to select a range of PVTs that address different cognitive domains, as individuals may underperform in specific areas but not others (Boone, 2009). To ensure meaningful assessment, the tests within the battery should not be highly correlated, though some degree of overlap is expected (Rosenfeld et al., 2000). While a PVT may appear subjectively difficult, they are designed to be easily achievable, minimizing external influences and ensuring optimal test performance (Lippa, 2018). The underlying concept is that most participants should perform at or near their best level, with those who underperform likely to fail the PVT.
With the increasing use of PVTs and SVTs in clinical assessments, it is essential to clarify their conceptual similarity, particularly in adults with suspected ADHD. Although PVTs vary in design, all aim to identify underperformance in diagnostic evaluations. Therefore, understanding and defining the conceptual structure of PVTs, alongside SVTs, within a comprehensive neuropsychological evaluation is crucial. Analyzing their structure determines whether they measure distinct factors, thereby improving the accuracy of first-time ADHD assessments and providing valid diagnostic information.
While many studies have used prediction and discrimination methods to examine the relationship between PVTs and SVTs (Smart et al., 2008; Whitney et al., 2008), fewer have explored the underlying constructs of these tests (Basso et al., 2024; Nelson et al., 2007; Ruocco et al., 2008; Van Dyke et al., 2013). For instance, Van Dyke et al. (2013) applied a confirmatory factor analysis (CFA) to assess the factor structure of PVTs and SVTs. Their model revealed three factors, consisting of cognitive performance, performance validity, and symptom self-report, with SVTs loading onto the latter. Recently, Basso et al. (2024) also concluded that PVTs and SVTs represent distinct constructs. Overall, both Van Dyke et al. (2013) and Basso et al. (2024) recommended further explorations of validity measures alongside standard diagnostic practices. But neither specifically examined adult ADHD. Given the clinical presentation of adult ADHD in first-time diagnostic evaluations, it is crucial to extend these methods to this population. This includes exploring the underlying network and factor structure of PVTs, alongside SVTs, and other clinical measures.
The present study aimed to clarify the conceptual similarities and distinctions between PVTs, CPTs, and ADHD symptom report scales, including embedded SVTs, addressing a gap in the understanding of the validity framework in adult ADHD assessments. To test this, we employed a combination of network and factor analysis techniques to examine both the interrelationships and factorial structures of three PVT measures among a large set of neuropsychological measures. This analysis was based on data from a large clinical referral sample of adults with suspected ADHD. The comprehensive test battery included ADHD symptom self- and observer reports, two embedded SVTs, a CPT, and three stand-alone PVTs. Similarly to Van Dyke et al. (2013), we examine the presence of a three-factor structure of validity assessments. This structure encompasses (1) performance validity measures, (2) continuous performance measures, and (3) symptom reports. SVTs are expected to load onto the symptom report factor, due to their close relationship with reported ADHD self-reported symptoms. Additionally, we examined potential influential factors of PVT results. For this, we considered a range of demographic (age, sex, education level) and clinical characteristics (ADHD diagnosis, ADHD symptom reports, comorbidity, cognitive deficit). As freestanding PVTs are designed to be easily achievable, we expect that these demographic and clinical factors should have little to no influence on PVT pass/failure rates.

2. Materials and Methods

2.1. Procedure and Participants

All participants were recruited between 2022 and 2024 from the adult ADHD outpatient clinic at the Department of Psychology, Philipps University Marburg, Germany. See Kneidinger et al. (2025; manuscript submitted for publication) for details and previous research on subsets of this dataset. All clients provided written informed consent for their data to be used for research purposes. The local Ethical Review Committee approved the use of routine data from the outpatient clinic.
Individuals were referred for a thorough diagnostic assessment of ADHD due to concerns raised by medical professionals, significant others, or themselves. The diagnostic assessment was based on the empirically informed guidelines for first-time ADHD diagnoses in adults (Sibley, 2021). Clients were assessed by board-certified clinical psychologists based on provided and reported clinical history. The standard diagnostic examination included (1) semi-structured interviews, the diagnostic interview for ADHD in adults (DIVA-5; Kooij et al., 2019), and the structured clinical interview for DSM-IV (SCID-I; Wittchen et al., 1997) for assessing potential comorbidity or other primary disorder explaining symptoms; and (2) the Conners Adult ADHD Scale (CAARS-L; Christiansen et al., 2014; Conners et al., 1999) employed both as self (CAARS-L:S) and observer (CAARS-L:O) questionnaires for assessing ADHD psychopathology. Alongside these clinical tools, clinicians also assessed the impact of ADHD symptoms on various areas of daily functioning, such as academic challenges, workplace difficulties, participation in risky behaviors, and substance abuse issues.
Neuropsychological assessment of ADHD core symptoms was done with the Quantified Behaviour Test (Qb+; Ulberstad, 2012), a continuous performance test (CPT) with actigraphy of head movement that served as objective and normed measures of attention, impulsivity, and motor activity. This was followed by three PVTs, namely the Amsterdam Short-Term Memory Test (ASTM; Schmand & Lindeboom, 2005), the Medical Symptom Validity Test (MSVT; P. Green, 2004), and the Groningen Effort Test (GET; Fuermaier et al., 2021; see below for details). For this study, two embedded SVTs, the Inconsistency Index and Infrequency Index (CII), were calculated separately (Christiansen et al., 2014; Conners et al., 1999; Suhr et al., 2011). These were based on the CAARS-L:S questionnaire, providing additional insight into infrequent and inconsistent response patterns.
Validity test results were considered in the diagnostic process, particularly in cases with ambiguous findings. If an ADHD diagnosis was strongly supported by other assessment measures and the validity concerns were explainable, the diagnosis remained unaffected. However, in cases with inconclusive or inconsistent results, validity findings played a decisive role in diagnostic determination.
Data was collected from 479 clients, which included performance data from PVT, SVT, CPT, and diagnostic assessments. Clients were excluded from the analyses due to incomplete cases (N = 17) and administrative errors (N = 1). The final sample included 461 participants. Descriptive statistics are depicted in Table 1. Overall, individuals’ average age was 33.2 years (SD = 10.4), and more than half of them were women (58.1%, N = 268). The participants’ educational backgrounds varied, including individuals not reported/without a school-leaving certificate (0.9%, N = 4) and general education (basic/secondary school degree; 32.5%, N = 150), with the majority having advanced education (grammar school/university degree; 66.6%, N = 307). Among all adults who underwent ADHD diagnostic assessment, 84.2% (N = 388) received a diagnosis post evaluation. Within this sample, 26.0% (N = 101) of adults reported at least one (-up to three) comorbid mental disorder(s).

2.2. Measures

2.2.1. Amsterdam Short-Term Memory (ASTM) Test

The ASTM test functions as a performance validity test (PVT), detecting negative response bias and assessing insufficient effort or motivation during psychological evaluations (Schmand & Lindeboom, 2005). It is a memory test that uses a forced-choice recognition method, designed to assess short-term memory and attention. The ASTM test involves displaying five semantically related words for eight seconds, which participants are instructed to read aloud and memorize. Afterwards, they complete a simple arithmetic problem. Subsequently, five words are shown again, including three from the original set. Participants must identify the three words they saw previously. The used cut-off values and interpretation guidelines follow the recommendations outlined in the test manual (Schmand & Lindeboom, 2005). Schmand and Lindeboom (2005) reported high internal consistency (Cronbach’s α = 0.90) for various samples. For experimental simulants with neurological impairments, sensitivity for lack of motivation was 91% and specificity was 89%. Patients with neurological disorders, i.e., brain tumors and concussion, rarely have difficulties handling this test, given that they do not have serious cognitive deficits.

2.2.2. Medical Symptom Validity Test (MSVT)

The MSVT is a brief computerized measure of cognitive effort and memory (P. Green, 2004). According to the author, the MSVT was designed to shorten the administration of the test’s parent version, the Word Memory Test (WMT; P. Green, 2003). Like the WMT, the MSVT displays common word pairs over two consecutive trials; Immediate Recall (IR), Delayed Recall (DR), and Consistency Scores (CNS) were used to evaluate performance validity based on the test’s manual criteria. The Genuine Memory Impairment profile was not used, as it is cautioned against for individuals who do not exhibit significant functional decline (Resch et al., 2022). Given the nature of our sample, such a decline was not expected, and its assessment was beyond the scope of the present study. The classification of values for IR, DR, or CNS as PVT failure follows the guidelines outlined in the test manual (P. Green, 2004). The MSVT has shown adequate sensitivity and specificity in detecting invalid performance. For example, sensitivity predicting the WMT was 0.50 to 0.62 (P. Green, 2007).

2.2.3. Groningen Effort Test (GET)

The GET is an attention-based validity assessment, focusing on detecting underperformance (Fuermaier et al., 2021). It is an Embedded Figure Test (EFT), meaning the GET consists of a simple geometric figure embedded in a more complex one (Fuermaier et al., 2016). During the test, participants need to indicate on a computer screen whether the simple target figure is embedded within a complex figure. Both figures are displayed simultaneously, requiring participants to indicate the presence or absence of the embedded figure by pressing a button. The interpretation of index values and error rates for assessing cognitive performance in individuals with ADHD follows the guidelines provided in the test manual (Fuermaier et al., 2021). The GET index score is calculated by combining processing speed and accuracy measures within each block of trials. In addition, the total number of errors is calculated. Both the index score and total number of errors were used in the current study to indicate noncredible performance. Analogue studies indicate that the GET successfully differentiates simulated malingered ADHD from genuine ADHD with high sensitivity (68%) and specificity (91%) when using a suggested cut-off score for total errors (Fuermaier et al., 2021).

2.2.4. Quantified Behaviour Test Plus (Qb+)

The QbTest+ (Qbtech, 2025) is a computer-based continuous performance test (CPT), measuring sustained attention with a one-back working memory task. Essentially, participants are required to recall the same object (red or blue circles and squares) in shape and color. A response key is to be pressed when the current stimulus matches the previous one in both attributes (Dolgin, 2014; Ulberstad, 2012). Participants are instructed to store and compare stimuli in working memory (Lis et al., 2010), with a target-to-nontarget stimulus ratio of 25:75. While performing the CPT, participants sit 1 m from a monitor with an infrared head movement tracking camera and use a handheld responder. The camera captures movement during task execution, at a frequency of 50 samples a second with a spatial resolution of 1/27 mm per infrared camera unit. Normative data have been gathered from 1307 individuals between 6–60 years of age for both versions of the test (QbTest 6–12 and Qb+) with an even age and gender distribution (Ulberstad, 2012). Q-scores are derived for hyperactivity, inattention, and impulsivity. They are interpreted similarly to Z-scores with a mean of 0 and a standard deviation of 1. A Q-score of >1.5 is observed as an atypical result (Hirsch & Christiansen, 2017; Ulberstad, 2012). The Qb+ reports nine parameters that are categorized into activity and CPT measures. Of these, we report four specific CPT measures in total: (1) Reaction Time Q-score (RTQ) in milliseconds, reflecting the average time for correct responses, offering insights into both information processing latency and motor response speed; (2) RT Variation (RTVarQ) is calculated using the standard deviation of the mean of correct response times and is a measure of the participant’s inconsistency in response times; (3) Omission Error score (OmissQ) reflects the total number of missed targets; (4) Commission Error score (ComissQ) is the number of false hits. In the present study, Q-scores that are normed for age and sex were utilized for the listed parameters. Sensitivity and specificity are within the acceptable range (Bellato et al., 2024).

2.2.5. Conners Adult ADHD Rating Scales (CAARS-L:S; CAARS-L:O)

The CAARS-L:S and CAARS-L:O (German version) assess ADHD symptoms in adults aged 18 and older. The long version consists of 66 items rated on a Likert-type scale ranging from 0 (not all all/never) to 3 (very much/very frequently). Due to statistical restrictions, only 42 items were included in the original factor analysis, revealing four factors: inattention/memory problems, hyperactivity/restlessness, impulsivity/emotional lability, and problems with self-concept (Conners et al., 1999). These were further confirmed in the German version (Christiansen et al., 2013; Christiansen et al., 2012). Both rating scales have been considered as reliable and cross-culturally valid tools for assessing current ADHD symptoms in adults (Christiansen et al., 2012). Norm scores are calculated based on sex and age. For the present study, the four core symptoms of inattention, hyperactivity, impulsivity, and self-concept problems were included.
In addition, two embedded SVTs were obtained from the self-rated questionnaires. Firstly, the CAARS-L:S—Inconsistency Index was computed, indicating the probable occurrence of careless responding for any score > 8. The manual suggests inspecting this score first to detect problems with internally consistent responding, which may affect the validity of the CAARS. Secondly, the CII (Suhr et al., 2011) was calculated by summing twelve item scores, which were shown to be infrequently reported by healthy controls and adults with credible ADHD. Any score of the CII > 21 indicates an elevated, noncredible symptom report. Early studies of the CII showed sensitivity estimates ranging from 17% to 19% (Becke et al., 2021; C. Cook et al., 2018; C. M. Cook et al., 2016; Fuermaier et al., 2016) to 86% (E. V. Robinson & Rogers, 2018) and 95% (Walls et al., 2017).

2.3. Statistical Analyses

All analyses were performed using RStudio (Version 4.3.3; RStudio, 2024). Less than 6% of clients (25 out of 461) had missing CAARS-L:S items, accounting for 5.4% of the data. In line with the CAARS-L:S manual, ≤5 missing items is considered acceptable (Christiansen et al., 2014). Missing items were handled using multiple imputation with the mice package (Rubin, 1976, 2004).
To examine the interrelationships among the given variables (PVTs, Qb+, CAARS-L:S/O, including its SVTs), a heatmap analysis of the Pearson correlation matrix was conducted. This step ensured sufficient common variance for factor extraction and helped identify potential multicollinearity. Furthermore, a network analysis was then performed to explore structural connections among the variables. To test the hypothesized factor structure of PVTs in adults with suspected ADHD, confirmatory factor analysis (CFA) was conducted. Due to insufficient support from the CFA for the hypothesized structure, post hoc exploratory factor analyses (EFAs) were subsequently performed to further explore the latent dimensions. Lastly, binary logistic regression models were conducted to examine factors influencing PVT pass and fail outcomes.

2.3.1. Descriptive Statistics and Preliminary Heatmap Correlations

Test performance (PVTs, Qb+) and clinical information (CAARS-L:S/O, including SVTs) are presented descriptively for all individuals. PVT results are shown as positive results, as the interpretation of PVT outcomes is commonly conducted dichotomously (positive indicating failure, negative indicating pass). Furthermore, ADHD symptom reports and Qb+ measures are presented using descriptive statistics and interpreted based on test norms, i.e., showing the proportion of individuals with “below average” self-reports. “Below average” is defined by Guilmette et al. (2020) as a percentile rank (PR) ≤ 8 on tests and a PR ≥ 92 on self-report scales.
To explore the interrelationships and underlying factors among the given variables, we first applied a heatmap analysis to the Pearson correlation matrix. Visualizing the correlations among PVTs, Qb+, and CARRS-L:S/O (including SVTs) in a heatmap was essential for evaluating the suitability of these variables in identifying a potential PVT factor. This preliminary step allowed for the assessment of the strength and direction of intercorrelations, ensuring that the variables shared sufficient common variance required for factor extraction. Additionally, the heatmap helped detect multicollinearity, which could hinder meaningful PVT factor identification. It also confirmed that the relationships between these diverse measures were robust enough for both network and factor analysis.

2.3.2. Network Analysis

A network analysis was conducted to visually and structurally explore the interrelationships of all variables, including PVTs (MSVT, ASTM, GET), Qb+, and CAARS-L:S/O (including SVTs). The network analysis consists of three core analytical steps: (1) estimating the network; (2) assessing node centrality; (3) evaluating accuracy and stability. First, network estimation visually represents the relationships between variables, which can be quantified using partial correlation to measure the strength of these connections. The networks were constructed using the R package “bootnet” (Version 1.6; Epskamp et al., 2018; Epskamp et al., 2012), employing Graphical Gaussian Models with the Least Absolute Shrinkage and Selection Operator (EBIC; Foygel & Drton, 2010). To enhance interpretability and prevent spurious connections, the “Least Absolute Shrinkage and Selection Operator” (LASSO; Tibshirani, 1996) regularization algorithm was employed. Network visualization relied on the Fruchterman–Reingold algorithm (Fruchterman & Reingold, 1991). In the graph produced by the Fruchterman–Reingold algorithm, nodes with stronger connections are positioned closer to each other, and edges between nodes with higher absolute coefficients are depicted with thicker and more saturated colored lines. A detailed description of the analytic procedures is available in Costantini and Epskamp (2017) and Epskamp et al. (2018). As our data was not normally distributed, Spearman correlations were utilized as input (Isvoranu & Epskamp, 2023).
Next, we evaluated the centrality of each item, which indicates how connected an item is to all other items in the network. We used expected influence, a centrality measure that sums the connections for each node. To compute and visualize the expected influence, we employed the centrality, centralityTable, and centralityPlot functions from the “qgraph” package (Epskamp et al., 2012). These analyses identified whether items in the network clustered together in a particular manner.
Lastly, we evaluated the precision of edge weights and the consistency of the node centrality order. The accuracy of edge weights was evaluated using bootstrapping to compute the 95% confidence interval (CI) for each edge. Smaller CIs suggest greater accuracy in the ordering of edges within the network. Node centrality stability was assessed using the correlation stability (CS) coefficient. Based on the simulation approach outlined by Epskamp et al. (2018), CS coefficients above 0.25 indicate moderate stability, while those above 0.5 indicate strong stability. These were performed using the R package “bootnet” (Epskamp et al., 2018; Epskamp et al., 2012).

2.3.3. Confirmatory Factor Analysis

Furthermore, a confirmatory factor analysis (CFA) was employed to assess the expected factor structure of PVTs in adults with suspected ADHD. CFA validates the pre-established hypothesized relationships between observed indicators and latent factors, providing evidence of constructs in theory-based instrument construction and development (Kline, 2023). As our observed variables (20 items) include both count and continuous data (RTQ, RTVarQ, OmissQ, ComissQ), and considering our sample size (<500), we used the Robust Maximum Likelihood Estimator (MLR). The MLR was chosen as it adjusts for nonnormality and is well-suited for managing mixed-type data (Kiliç & Doğan, 2021; Lubke & Muthén, 2004).
To evaluate the overall goodness-of-fit of the model we employed standard indices, including standard chi-square statistics (χ2/df), the robust root mean square error of approximation (RMSEA), the Robust Confirmatory Fit Index (CFI), and the Robust Tucker–Lewis Index (TLI) (Brown, 2015; Hu & Bentler, 1999). Model–data fit is considered acceptable when χ2/df ranges from 2 to 5, the RMSEA is below 0.06, and the CFI and TLI exceed 0.80. A good fit is achieved when χ2/df is close to 1, and both the CFI and TLI exceed 0.95 (Hox et al., 2017). In line with previous research, suggesting the PVT as a latent factor, we specified three models with one (all items), two (Test Performance, Symptom Report), and three (Performance Validity, CPT, Symptom Report) latent factors.

2.3.4. Multiple Logistic Regression Models

To examine factors influencing the likelihood of passing or failing each PVT, a series of binary logistic regression models was conducted. Logistic regression with a binary outcome (pass/fail) was preferred over linear regression, as the goal was to analyze clinically relevant outcomes for PVTs. The analytical approach depended on whether a robust PVT factor structure emerged. If a robust factor structure was identified, a composite PVT factor would be computed by weighting the individual PVTs according to their respective factor scores. Subsequent analyses would then focus on the composite PVT factor. If no robust factor structure emerged, each PVT (ASTM, MSVT, GET) would be analyzed separately. The outcome for each PVT would be binary (pass/failure), based on the cut-off criteria outlined in the respective test manuals (Fuermaier et al., 2021; P. Green, 2004; Schmand & Lindeboom, 2005).
Predictor variables included demographic factors (age, sex, and education level) and clinical data (ADHD diagnosis, ADHD core symptoms, comorbidity, and cognitive deficit). These predictors were selected because they represent key characteristics relevant to adult ADHD and provide insights into individual differences in PVT performance. ADHD core symptoms (inattention, hyperactivity, impulsivity, and self-concept problems) were classified based on the CAARS-L:S/O questionnaires, with clients scoring > 65 (T-scores) considered to exhibit clinically significant symptoms. The cognitive deficit variable was derived from the Qb+ assessments, with a Q-score > 1.5 on any of the four CPT measures (RTQ, RTVarQ, OmissionQ, CommissionQ) considered indicative of a deficit.
In the absence of a factor structure, odds ratios (ORs) for each predictor variable were computed by exponentiating the model coefficients for easier interpretation. Logistic regression analyses were performed separately for each PVT (ASTM, MSVT, and GET) using the “glm” package in R. Model fit was assessed using Nagelkerke’s R2 and the likelihood ratio test to evaluate the explanatory power of the logistic regression models (Cragg & Uhler, 1970; Nagelkerke, 1991).

3. Results

3.1. Descriptive Statistics and Preliminary Correlations

Table 2 presents the descriptive statistics for the PVTs, Qb+, CARRS-L:S/O (including SVTs), along with the proportion of individuals who obtained positive results on each PVT subtest separately. The proportion of conspicuous performance differed by subtest, with 1.7% (MSVT), 10.2% (GET), and 15.2% (ASTM). Inconsistent responding in the CAARS was observed in 23.2% of the individuals, while 23.6% gave infrequent responses (CII). The analysis of the CAARS-L:S/O questionnaires shows that a majority of participants scored PR > 92% on at least one ADHD-related scale, particularly self-reports. Among all CAARSs, the highest number of individuals had elevated scores on “Inattention”, followed by “Self-Concept Problems”, “Impulsivity”, and “Hyperactivity”. Additionally, more individuals exhibited high scores in the self-report scales compared to the observer ratings.
Before analyzing the relationships and factors among the parameters PVTs (MSVT, ASTM, GET), SVTs (CAARS-L:S—Inconsistency, CII), and CAARS-L (CAARS-L:S, CAARS-L:O), all variables were z-transformed using a robust Z-score based on the current data sample. This transformation ensured that each variable had a mean of zero and a standard deviation of one. It was performed to mitigate potential effects of scale discrepancies between variables and to prevent estimation issues caused by differences in response variances. Q-based norm scores (adjusted for age and gender) from the Qb+ test were entered into the analysis.
Figure 1 displays the heatmap of the correlation matrix for each PVT, Qb+ test, and CARRS-L:S/O (including SVTs). ASTM showed weak correlations with MSVT subscales (IR, DR, CNS; r = 0.14 to 0.22). GET variables (Errors, Index) did not correlate significantly with ASTM (r = −0.18 to −0.31) or MSVT (r = −0.08 to −0.25), indicating weak to negligible associations. Correlations between PVT measures and CAARS symptom scales were weak to negligible (r < 0.30). Notably, both the CAARS Infrequency Index (CAARS_CII) and the CAARS Inconsistency scale (CAARS_Incon) showed weak correlations with CAARSs and PVT measures. However, CAARS_CII was moderately correlated with CAARS self-reported scales “Inattention” and “Impulsivity” (r = 0.56 to 0.78). Lastly, Qb+ test measures (RTQ, RTVarQ, CommQ, OmissQ) displayed weak correlations with CAARS symptom scales, SVT, and PVT measures (r < 0.30).

3.2. Network Analysis

The item–network estimation of the 20 PVT, SVT, and Qb+ measures, along with ADHD symptom reports, is presented in Figure 2. The network consists of 75 edges (45 positive, 30 negative), showing multiple connections between the different measures (nodes). The network structure reveals connections between validity measures (PVT: nodes 1–6; SVT: nodes 11–12) with other ADHD assessment domains (QB+: nodes 7–10; CAARS: nodes 13–20), rather than forming a separate PVT subgroup. This indeed goes hand-in-hand with the results of the correlation matrix. SVT nodes (11–12) display strong positive connections with ADHD symptom scales (12–20). Node 11 exhibits stronger connections across impulsivity scales (15, 19), whereas node 12 is more linked to self-concept (16). In contrast, the PVT nodes exhibit distinct patterns of connectivity. Nodes 1–4 show minimal associations with ADHD symptom scales but display some connections with each other. Nodes 5–6 show slightly stronger relationships with Qb+ measures (7–10).
To further explore the centrality of individual nodes within the network, an expected influence analysis was conducted (Figure 3). ADHD reported “self-concept” scales exhibited the highest expected influence. Whilst SVTs displayed moderate expected influence, PVTs demonstrated the lowest expected influence. The edge weight accuracy estimation revealed a moderate confidence interval, indicating that the rank order of edge weights was accurately estimated (Figure 4). Furthermore, the node centrality estimation revealed CS coefficients of 0.751 for the expected influence and strength of the entire sample, indicating that the order of these centrality measures was highly stable. Given the observed network structure and stability, a CFA further assessed the constructs among validity and ADHD measures.

3.3. Confirmatory Factor Analysis

See Table 3 for the item allocation to the respective factor structures. Table 4 presents the fit indices for the one-, two-, and three-factor CFA models. The approximate model fit indices (χ2/df, CFI, RMSEA, and TLI) displayed poor fit for one-, two-, and three-factor models. This suggests that while a structure was present, it differed from the hypothesized factor structure of (1) performance validity measures, (2) continuous performance measures, and (3) symptom reports, including SVTs. For the three-factor model, the chi-square statistic was χ2(1550.4) with a χ2/df ratio of 9.28, exceeding the acceptable range of 2 to 5 (Hox et al., 2017). The RMSEA was 0.136, which is above the commonly accepted threshold of 0.06 for good fit. Additionally, the CFI (0.669) and TLI (0.623) were below the recommended 0.80 threshold for acceptable model fit. Similar patterns of poor fit were observed for the two-factor model (χ2/df = 11.5, RMSEA = 0.154, CFI = 0.570, TLI = 0.517) and the one-factor model (χ2/df = 16.7, RMSEA = 0.19, CFI = 0.34, TLI = 0.27).

3.4. Post Hoc Analysis: Exploring Alternative Model Fits

Due to the poor fit of the pre-specified one-, two-, and three-factor models in the CFA, an exploratory factor analysis (EFA) was subsequently conducted. This analysis aimed to identify potential alternative factor structures underlying the PVTs, CPTs, and symptom reports (including SVTs). Bartlett’s Test of Sphericity was significant χ2(190) = 4427.2, p < 0.001, and the overall Kaiser–Meyer–Olkin (KMO) measure was acceptable at 0.61, supporting factorability (Bartlett, 1954; Kaiser, 1970, 1974). Factors were extracted using maximum likelihood estimation with oblimin rotation, and the number of factors was determined via parallel analysis (Fabrigar et al., 1999). Four factors were retained based on the scree plot and parallel analysis (Figures S1 and S2). Factor 1 included the majority of CAARS subscales (self- and observer ratings) along with the embedded SVT (CII). Factor 2 was defined primarily by Qb+ variables. Factor 3 was characterized by strong loadings from the MSVT subscale. Factor 4 was defined by GET variables.

3.5. Multiple Logistic Regression Models

Since the CFA analysis did not reveal a common performance validity factor, each PVT was analyzed separately in terms of pass/fail likelihood. Multiple logistic regression models were computed to examine the predictors of PVT outcomes. Results are presented in Table 5. Odds ratios (ORs) and statistically significant levels are reported for each predictor.
The logistic regression model predicting the likelihood of passing the ASTM test (PVT 1) was statistically significant, χ2(12) = 26.8, p = 0.008. The model explained 10.7% of the variance in the likelihood of passing the ASTM test (R2 = 0.107). Cognitive deficits were significantly associated with lower odds of passing the ASTM test (β = −1.23, SE = 0.374, OR = 0.29, p < 0.001). Additionally, education was associated with greater odds of passing (β = 0.302, SE = 0.132, OR = 1.35, p < 0.05). Among CAARS-L:O results, the impulsivity ADHD scale was a significant predictor (β = 0.987, SE = 0.360, OR = 2.68, p < 0.01), indicating that higher impulsivity scores were associated with increased odds of passing the ASTM test. Conversely, the impulsivity reported symptom from the CAARS-L:S was negatively associated with passing the test (β = −0.791, SE = 0.402, OR = 0.45, p < 0.05). A further logistic regression was conducted to examine the factors influencing the likelihood of passing the MSVT (PVT 2). The model was not statistically significant, χ2(12) = 18.5, p = 0.101. Finally, a logistic model was used to predict the likelihood of passing the GET (PVT3). The model was not statistically significant, χ2 (12) = 20.5, p = 0.059.

4. Discussion

The primary objective was to explore the shared and unique structural features between PVTs, CPTs, and ADHD symptom scales (including SVTs), using data from a large outpatient sample of adults with suspected ADHD. To address this aim, we initially hypothesized that the three PVTs (ASTM, MSVT, GET) would load onto a common latent construct. However, the results do not support this hypothesis.
Descriptive statistics revealed differences in positive PVT results, with 1.7% (MSVT), 10.2% (GET), and 15.2% (ASTM). This variability broadly resembles findings from prior research (Hirsch & Christiansen, 2018; Mascarenhas et al., 2023) and underscores recommendations to administer multiple PVTs (Boone, 2009). In contrast, SVT elevated scores were more consistent, with 23.2% of individuals showing inconsistent and 23.6% infrequent responses (CII; Fuermaier et al., 2016).
Initial correlational patterns suggested weak relationships between PVTs, ADHD symptom scales (including SVTs), and CPT measures, hinting at structural independence between test modalities. A visual inspection of the network analysis further supported this, revealing four emergent item communities. The ASTM test remained largely independent, with only weak connections to the GET and CPT variables. The MSVT variables formed a cohesive cluster. A similar pattern of internal cohesion and external separation was observed for the GET and CPT variables. As expected, the subscales of the CAARS exhibited strong interrelationships in both self- and observer ratings, closely clustered with SVTs (Conners et al., 1999; Suhr et al., 2011). This pattern is further supported by the finding that ADHD self-concept scales displayed the highest, SVTs moderate, and PVTs the lowest centrality.
Further insights into the PVT structures were gained through factor analytic approaches. As a first approach, we conducted a CFA to test a theoretically driven three-factor model distinguishing (1) performance validity tests, (2) continuous performance measures, and (3) ADHD symptom reports (including SVTs). Two further competing models were tested, namely a one-factor solution and a two-factor model distinguishing test performance (ASTM, MSVT, GET, QB+) from symptom reports (CAARS-L self- and observer ratings, including SVTs). The latter model was theoretically motivated by the assumption that both PVT and CPT measures reflected aspects of test performance, whereas CAARS ratings (including SVTs) capture symptom reports. Contrary to our hypothesized structure and competing model specifications, all tested CFA models (one-, two-, and three-factor) showed poor fit across all indices. Following the poor CFA model fits that falsified the assumption of a common PVT factor, we further explored the latent structure of the measures through an EFA. Unlike the CFA, the EFA supported a four-factor solution, offering clearer differentiation among the test modalities. Notably, the PVTs did not form a single coherent factor but rather loaded onto separate factors, reflecting their distinct test characteristics. This structure complements the item communalities in the network analysis and provides further evidence against a unified PVT construct.
The factor analyses reveal an important distinction, while contradicting previous research by Basso et al. (2024) and Van Dyke et al. (2013). These two studies reported a distinct performance validity factor, primarily based on memory-focused PVTs, which tend to share similar demands (Merten et al., 2007; Messerly et al., 2021). Our battery included a broader range, such as the attention-based GET and performance measures like Qb+, introducing more variability. This diversity likely weakened correlations among tests and reduced the coherence expected in the CFA. In this context, the domain specificity hypothesis proposed by Erdodi (2019) offers a useful theoretical lens. It suggests that PVT outcomes reflect specific cognitive domains and task characteristics, rather than representing a single, unified construct. For example, PVTs assessing sustained attention (e.g., GET) and those assessing verbal memory (e.g., ASTM, MSVT) engage different cognitive systems, which may result in distinct response bias patterns within the same individual. Importantly, PVTs are not intended to measure underlying cognitive ability (Hoelzle et al., 2019; Wallace et al., 2019). Instead, they help contextualize actual test performance and should remain largely independent of the cognitive functions they reference. This perspective explains why correlations among different PVTs are often moderate (Rosenfeld et al., 2000) and supports the widely accepted recommendation to administer multiple PVTs targeting distinct cognitive domains (Boone, 2009). Our findings add empirical support to this view and contradict Leonhard’s (2023a, 2023b) claim of PVT collinearity by showing that each test evaluates distinct cognitive processes. From this viewpoint, the more unified validity structures reported by Basso et al. (2024) and Van Dyke et al. (2013) are less convincing. Their findings challenge the rationale for using multiple PVTs and underestimate the importance of a domain-sensitive approach to performance validity assessment in diverse clinical populations.
The second objective of this study addressed the influence of demographic and clinical factors on PVT performance, further emphasizing the independent functioning of each PVT. Whilst the MSVT and GET followed our expectation and showed no influence of demographic or clinical factors, the ASTM test did, as cognitive deficits, education level, and impulsivity were significant predictors of ASTM performance. Cognitive deficits were linked to lower odds of passing the ASTM test, consistent with prior research (Dandachi-Fitzgerald et al., 2011; Merten et al., 2007). Education level was a positive predictor, aligning with Stulemeijer et al. (2007), who also reported better performance in individuals with higher educational attainment. Impulsivity showed an unexpected, mixed relationship. Higher impulsivity on the CAARS-L:O was associated with passing, whereas impulsivity on the CAARS-L:S had the opposite effect. These findings align with Erdodi’s (2019, 2023) view that performance validity exists along a continuum rather than a strict pass/fail outcome. According to Erdodi (2019), engagement, performance, and cognitive functioning influence where individuals fall on this spectrum. Cognitive deficits, education, and impulsivity may affect how people engage with the ASTM test, leading to varied performance. However, this sensitivity to demographic and clinical factors may increase the risk of false positives. Some individuals could be misclassified as underperforming despite genuine cognitive difficulties or other traits. This limitation may reduce the ASTM test’s usefulness across diverse clinical groups. By contrast, the MSVT and GET appear more robust and less influenced by demographic and clinical factors, thus assessing the actual performance demonstrated during testing rather than being confounded by individuals’ underlying cognitive potential.
More generally, many PVTs have demonstrated low to modest sensitivity at standard cut-offs in ADHD contexts (A. D. Robinson & Calamia, 2025; Suhr et al., 2022). This supports A. D. Robinson and Calamia’s (2025) recommendation to adapt PVT procedures and interpretation specifically for individuals with ADHD. Therefore, PVTs should be interpreted in conjunction with broader clinical data, rather than in isolation. This aligns with the current practice guidelines, which advocate for a comprehensive, multi-method approach to evaluating performance validity (Sherman et al., 2020; Sweet et al., 2021).

4.1. Limitations

Several limitations should be considered when interpreting the findings of this study. The current sample is relatively homogenous in terms of gender (58.1% female) and education (66.6% with at least grammar school). This limits the generalizability of the results. Future analyses involving more diverse samples and subgroup analyses could help clarify which demographic variables influence ADHD diagnoses and validity.
On this note, while the use of a naturalistic clinical referral sample strengthens the study’s ecological validity, it also brings challenges in confidently identifying true ADHD cases. Variability in symptom presentation and the potential for distorted reporting or secondary gain can complicate diagnostic clarity. So, incorporating external validation criteria in future research may help address these concerns (L. W. Green & Glasgow, 2006).
Furthermore, the SVTs included in this study were embedded within ADHD symptom measures, specifically the CAARS self-report scale. While this allows for context-sensitive assessment of symptom validity, the lack of stand-alone SVTs limits the ability to disentangle symptom exaggeration from genuine symptom expression. Future analyses should therefore consider integrating stand-alone SVTs.

4.2. Implications for Future Research

Our findings reinforce the need to move beyond a unidimensional view of performance and potentially symptom validity. Accordingly, future studies should test multidimensional models that account for domain-specific underperformance (e.g., attention- vs. memory-related validity). Recent work by Finley et al. (2023) supports this approach, showing that combining embedded indicators across cognitive domains improves detection of underperformance in ADHD evaluations. These findings underscore the importance of task characteristics and cognitive domain in shaping performance validity.
As a complementary approach, future research may benefit from adopting person-centered analytic strategies to further characterize heterogeneity in performance validity outcomes. This approach could identify subgroups with distinct cognitive profiles, similar to those found by Cerny et al. (2025). In doing so, it may uncover distinct patterns across validity indicators, shedding light on ADHD’s neurocognitive heterogeneity.
Finally, future research should consider Erdodi’s (2019, 2023) conceptualization of performance validity as a continuum. This may be particularly relevant in ADHD, where engagement often varies with task demands. Modelling performance and symptom validity on a continuum could clarify how levels of underperformance relate to cognitive deficits, impulsivity, or education, thus improving the sensitivity of validity assessment in ADHD by capturing individual and task-related variability more accurately.

5. Conclusions

This study highlights the multifaceted nature of PVTs, CPTs, and ADHD symptom reports (including SVTs) in an outpatient adult ADHD population. Our findings suggest that PVTs (ASTM, MSVT, GET) reflect domain-specific processes, with performance influenced by the cognitive domain (e.g., sustained attention versus memory), rather than a unified latent construct. This supports a multidimensional view of performance validity and highlights the role of task-specific and individual factors. Notably, ASTM test performance was sensitive to cognitive deficits, education level, and impulsivity, underscoring the importance of interpreting PVT outcomes within the broader clinical context and along a behavioral continuum. These findings further support the need for ADHD-specific procedures and adjusted cut-offs to improve the accuracy of performance validity assessment in this population. These insights also carry important implications for forensic contexts, emphasizing that performance validity measures offer an additional behavioral indicator essential for reducing misclassifications in legal evaluations.
Future research should model performance validity as a multidimensional, continuous construct. Person-centered approaches may further clarify how cognitive and behavioral profiles shape test performance in ADHD, enhancing the precision and utility of validity assessments.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/bs15081108/s1: Supplementary Figures S1 and S2. Figure S1: Exploratory factor analysis of PVT and CPT measures and ADHD symptom reports (including SVTs) in the full sample (N = 461), Figure S2: Scree plot following the exploratory factor analysis in the full sample (N = 461).

Author Contributions

Conceptualization: E.R., A.B.M.F., J.K., B.A. and H.C.; methodology: E.R., A.B.M.F., J.K., B.A. and H.C.; data curation: E.R. and J.K.; data analysis: E.R.; writing—original draft preparation: E.R.; writing—review and editing: E.R., A.B.M.F., J.K., B.A. and H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Ethics Committee of Faculty of Psychology at Philipps University Marburg (approval code 2024-75k, date of approval 9 December 2024) within the “Coordination of Data Acquisition at Research Clinics for Psychotherapy” project (German abbreviation: KODAP) was granted.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

A.B.M.F is a test author of the Groningen Effort Test (GET) and has a contract with Schuhfried GmbH for the development and evaluation of neuropsychological instruments.

Abbreviations

The following abbreviations are used in this manuscript:
ADHDAttention-deficit/hyperactivity disorder
PVTPerformance validity test
SVTSymptom validity test
GETGroningen Effort Test
MSVTMedical Symptom Validity Test
IR Immediate Recall
DRDelayed Recall
CNSConsistency Scores
ASTMAmsterdam Short-Term Memory
WMTWord Memory Test
CIIInfrequency Index
CPTContinuous performance test
Qb+Quantified Behaviour Test Plus
RTQReaction Time Q-score
RTVarQReaction Time Variation Q-score
OmissQOmission Error score
ComissQCommission Error score
DIVA-5Diagnostic interview for ADHD in adults
SCID-IStructured clinical interview for DSM-IV
CAARSConners Adult ADHD Scale
CAARS-L:S Conners Adult ADHD Scale Self-report
CAARS-L:OConners Adult ADHD Scale Observer report
CFAConfirmatory factor analysis
EFAExploratory factor analysis
GLASSGaussian Models with the Least Absolute Shrinkage and Selection Operator
EBICExtended Bayesian Information Criterion
LASSOLeast Absolute Shrinkage and Selection Operator
CIConfidence interval
MLRRobust Maximum Likelihood Estimator
RMSEARoot mean square error of approximation
CFIRobust Confirmatory Fit Index
TLITucker–Lewis Index
KMO Kaiser–Meyer–Olkin

References

  1. Adamou, M., Arif, M., Asherson, P., Aw, T.-C., Bolea, B., Coghill, D., Guðjónsson, G., Halmøy, A., Hodgkins, P., Müller, U., Pitts, M., Trakoli, A., Williams, N., & Young, S. (2013). Occupational issues of adults with ADHD. BMC Psychiatry, 13(1), 59. [Google Scholar] [CrossRef]
  2. Albrecht, B., Uebel-Von Sandersleben, H., Wiedmann, K., & Rothenberger, A. (2015). ADHD history of the concept: The case of the continuous performance test. Current Developmental Disorders Reports, 2(1), 10–22. [Google Scholar] [CrossRef]
  3. American Psychiatric Association. (2013). Diagnostic and statistical manual of mental disorders. American Psychiatric Association. [Google Scholar] [CrossRef]
  4. Arrondo, G., Mulraney, M., Iturmendi-Sabater, I., Musullulu, H., Gambra, L., Niculcea, T., Banaschewski, T., Simonoff, E., Döpfner, M., & Hinshaw, S. P. (2024). Systematic review and meta-analysis: Clinical utility of continuous performance tests for the identification of attention-deficit/hyperactivity disorder. Journal of the American Academy of Child and Adolescent Psychiatry, 63(2), 154–171. [Google Scholar] [CrossRef] [PubMed]
  5. Barkley, R. A., Fischer, M., Smallish, L., & Fletcher, K. (2006). Young adult outcome of hyperactive children: Adaptive functioning in major life activities. Journal of the American Academy of Child and Adolescent Psychiatry, 45(2), 192–202. [Google Scholar] [CrossRef] [PubMed]
  6. Bartlett, M. S. (1954). A note on the multiplying factors for various χ2 approximations. Journal of the Royal Statistical Society. Series B (Methodological), 16, 296–298. Available online: http://www.jstor.org/stable/2984057 (accessed on 2 March 2025). [CrossRef]
  7. Basso, M. R., Tierney, S. M., Roper, B. L., Whiteside, D. M., Combs, D. R., & Estevis, E. (2024). A tale of two constructs: Confirmatory factor analysis of performance and symptom validity tests. Journal of Clinical and Experimental Neuropsychology, 46(9), 840–847. [Google Scholar] [CrossRef]
  8. Becke, M., Tucha, L., Weisbrod, M., Aschenbrenner, S., Tucha, O., & Fuermaier, A. B. M. (2021). Non-credible symptom report in the clinical evaluation of adult ADHD: Development and initial validation of a new validity index embedded in the Conners’ adult ADHD rating scales. Journal of Neural Transmission, 128(7), 1045–1063. [Google Scholar] [CrossRef]
  9. Bellato, A., Hall, C. L., Groom, M. J., Simonoff, E., Thapar, A., Hollis, C., & Cortese, S. (2024). Practitioner review: Clinical utility of the QbTest for the assessment and diagnosis of ADHD: A systematic review and meta-analysis. Journal of Child Psychology and Psychiatry, 65(6), 845–861. [Google Scholar] [CrossRef] [PubMed]
  10. Boone, K. B. (2009). The need for continuous and comprehensive sampling of effort/response bias during neuropsychological examinations. The Clinical Neuropsychologist, 23(4), 729–741. [Google Scholar] [CrossRef]
  11. Boone, K. B., Lu, P., Back, C., King, C., Lee, A., Philpott, L., Shamieh, E., & Warner-Chacon, K. (2002). Sensitivity and specificity of the Rey Dot Counting Test in patients with suspect effort and various clinical samples. Archives of Clinical Neuropsychology, 17(7), 625–642. [Google Scholar] [CrossRef]
  12. Brown, T. A. (2015). Confirmatory factor analysis for applied research. Guilford publications. [Google Scholar]
  13. Cerny, B. M., Leib, S. I., Robinson, A. D., Ulrich, D. M., Phillips, M. S., Lee, E. J., & Soble, J. R. (2025). Neurocognitive phenotypes among adults with attention-deficit hyperactivity disorder: A latent profile analysis. Neuropsychology, 39(2), 187. [Google Scholar] [CrossRef]
  14. Choi, W.-S., Woo, Y. S., Wang, S.-M., Lim, H. K., & Bahk, W.-M. (2022). The prevalence of psychiatric comorbidities in adult ADHD compared with non-ADHD populations: A systematic literature review. PLoS ONE, 17(11), e0277175. [Google Scholar] [CrossRef]
  15. Christiansen, H., Hirsch, O., Abdel-Hamid, M., & Kis, B. (2014). CAARS. Conners skalen zu aufmerksamkeit. Huber. [Google Scholar]
  16. Christiansen, H., Hirsch, O., Philipsen, A., Oades, R. D., Matthies, S., Hebebrand, J., & Kis, B. (2013). German validation of the conners adult ADHD rating scale-self-report: Confirmation of factor structure in a large sample of participants with ADHD. Journal of Attention Disorders, 17, 690–698. [Google Scholar] [CrossRef] [PubMed]
  17. Christiansen, H., Kis, B., Hirsch, O., Matthies, S., Hebebrand, J., Uekermann, J., & Philipsen, A. (2012). German validation of the Conners Adult ADHD Rating Scales (CAARS) II: Reliability, validity, diagnostic sensitivity and specificity. European Psychiatry, 27, 321–328. [Google Scholar] [CrossRef]
  18. Conners, C. K., Erhardt, D., & Sparrow, E. (1999). Conners’ adult ADHD rating scales (CAARS) (technical manual). Toronto, Multi-Health Systems. [Google Scholar] [CrossRef]
  19. Cook, C., Buelow, M. T., Lee, E., Howell, A., Morgan, B., Patel, K., Bryant, A. M., Menatti, A., & Suhr, J. (2018). Malingered attention deficit/hyperactivity disorder on the conners’ adult ADHD rating scales: Do reasons for malingering matter? Journal of Psychoeducational Assessment, 36(6), 552–561. [Google Scholar] [CrossRef]
  20. Cook, C. M., Bolinger, E., & Suhr, J. (2016). Further validation of the conner’s adult attention deficit/hyperactivity rating scale infrequency index (CII) for detection of non-credible report of attention deficit/hyperactivity disorder symptoms. Archives of Clinical Neuropsychology, 31(4), 358–364. [Google Scholar] [CrossRef] [PubMed]
  21. Costantini, G., & Epskamp, S. (2017). The fused graphical lasso for computing psychological networks. In IMPS 2017 abstracts: Talks (pp. 14–15). University of Zurich. [Google Scholar]
  22. Cragg, J. G., & Uhler, R. S. (1970). The demand for automobiles. The Canadian Journal of Economics/Revue Canadienne d’Economique, 3(3), 386–406. [Google Scholar] [CrossRef]
  23. Dandachi-Fitzgerald, B., Ponds, R. W. H. M., Peters, M. J. V., & Merckelbach, H. (2011). Cognitive underperformance and symptom over-reporting in a mixed psychiatric sample. The Clinical Neuropsychologist, 25(5), 812–828. [Google Scholar] [CrossRef] [PubMed]
  24. Dolgin, E. (2014). FDA clearance paves way for computerized ADHD monitoring. Nature Medicine, 20(5), 454–455. [Google Scholar] [CrossRef]
  25. Epskamp, S., Borsboom, D., & Fried, E. I. (2018). Estimating psychological networks and their accuracy: A tutorial paper. Behavior Research Methods, 50(1), 195–212. [Google Scholar] [CrossRef]
  26. Epskamp, S., Cramer, A. O. J., Waldorp, L. J., Schmittmann, V. D., & Borsboom, D. (2012). qgraph: Network visualizations of relationships in psychometric data. Journal of Statistical Software, 48, 1–18. [Google Scholar] [CrossRef]
  27. Erdodi, L. A. (2019). Aggregating validity indicators: The salience of domain specificity and the indeterminate range in multivariate models of performance validity assessment. Applied Neuropsychology: Adult, 26(2), 155–172. [Google Scholar] [CrossRef]
  28. Erdodi, L. A. (2023). Multivariate models of performance validity: The Erdodi index captures the dual nature of non-credible responding (continuous and categorical). Assessment, 30(5), 1467–1485. [Google Scholar] [CrossRef]
  29. Fabrigar, L. R., Wegener, D. T., MacCallum, R. C., & Strahan, E. J. (1999). Evaluating the use of exploratory factor analysis in psychological research. Psychological Methods, 4(3), 272. [Google Scholar] [CrossRef]
  30. Faraone, S. V., Biederman, J., Feighner, J. A., & Monuteaux, M. C. (2000). Assessing symptoms of attention deficit hyperactivity disorder in children and adults: Which is more valid? Journal of Consulting and Clinical Psychology, 68(5), 830–842. [Google Scholar] [CrossRef]
  31. Finley, J. C. A., Brooks, J. M., Nili, A. N., Oh, A., VanLandingham, H. B., Ovsiew, G. P., Ulrich, D. M., Resch, Z. J., & Soble, J. R. (2023). Multivariate examination of embedded indicators of performance validity for ADHD evaluations: A targeted approach. Applied Neuropsychology: Adult, 32(5), 1254–1267. [Google Scholar] [CrossRef]
  32. Foygel, R., & Drton, M. (2010). Extended Bayesian information criteria for Gaussian graphical models. Advances in Neural Information Processing Systems, 23, 2020–2028. [Google Scholar] [CrossRef]
  33. Fruchterman, T. M., & Reingold, E. M. (1991). Graph drawing by force-directed placement. Software: Practice and Experience, 21(11), 1129–1164. [Google Scholar] [CrossRef]
  34. Fuermaier, A. B. M., Tucha, L., Guo, N., Mette, C., Müller, B. W., Scherbaum, N., & Tucha, O. (2022). It takes time: Vigilance and sustained attention assessment in adults with ADHD. International Journal of Environmental Research and Public Health, 19(9), 5216. [Google Scholar] [CrossRef] [PubMed]
  35. Fuermaier, A. B. M., Tucha, L., Koerts, J., Aschenbrenner, S., & Tucha, O. (2021). Test manual: The Groningen Effort Test (GET)-revised. Schuhfried. [Google Scholar]
  36. Fuermaier, A. B. M., Tucha, L., Koerts, J., Weisbrod, M., Grabemann, M., Zimmermann, M., Mette, C., Aschenbrenner, S., & Tucha, O. (2016). Evaluation of the CAARS infrequency index for the detection of noncredible ADHD symptom report in adulthood. Journal of Psychoeducational Assessment, 34(8), 739–750. [Google Scholar] [CrossRef]
  37. Fuermaier, A. B. M., Tucha, L., Merten, T., Fathollah Gol, M., & Tucha, O. (2024). Symptom validity testing in adults with clinically diagnosed ADHD: Comparison of the Conner’s Adult ADHD Rating Scale (CAARS) and the Self-Report Symptom Inventory (SRSI). Journal of Clinical and Experimental Neuropsychology, 46(8), 693–706. [Google Scholar] [CrossRef]
  38. Green, L. W., & Glasgow, R. E. (2006). Evaluating the relevance, generalization, and applicability of research: Issues in external validation and translation methodology. Evaluation & the Health Professions, 29(1), 126–153. [Google Scholar] [CrossRef]
  39. Green, P. (2003). Green’s word memory test: User’s manual. Green’s Publishing. [Google Scholar]
  40. Green, P. (2004). Green’s medical symptom validity test (MSVT). Green’s Publishing. [Google Scholar]
  41. Green, P. (2007). Spoiled for choice: Making comparisons between forced-choice effort tests. The Guilford Press. [Google Scholar]
  42. Guilmette, T. J., Sweet, J. J., Hebben, N., Koltai, D., Mahone, E. M., Spiegler, B. J., Stucky, K., & Westerveld, M. (2020). American academy of clinical neuropsychology consensus conference statement on uniform labeling of performance test scores. The Clinical Neuropsychologist, 34(3), 437–453. [Google Scholar] [CrossRef]
  43. Guo, N., Fuermaier, A. B. M., Koerts, J., Tucha, O., Scherbaum, N., & Müller, B. W. (2023). Networks of neuropsychological functions in the clinical evaluation of adult ADHD. Assessment, 30(6), 1719–1736. [Google Scholar] [CrossRef]
  44. Gustafsson, U., & Hansen, M. (2023). QbTest in the clinical assessment of attention deficit hyperactivity disorder: A review of the evidence. Mental Health Science, 1(4), 270–281. [Google Scholar] [CrossRef]
  45. Harrison, A. G., & Edwards, M. J. (2010). Symptom exaggeration in post-secondary students: Preliminary base rates in a canadian sample. Applied Neuropsychology, 17(2). [Google Scholar] [CrossRef]
  46. Hirsch, O., & Christiansen, H. (2017). Factorial structure and validity of the quantified behavior test plus (Qb+©). Assessment, 24(8), 1037–1049. [Google Scholar] [CrossRef]
  47. Hirsch, O., & Christiansen, H. (2018). Faking ADHD? Symptom validity testing and its relation to self-reported, observer-reported symptoms, and neuropsychological measures of attention in adults with ADHD. Journal of Attention Disorders, 22(3), 269–280. [Google Scholar] [CrossRef]
  48. Hirsch, O., Fuermaier, A. B. M., Tucha, O., Albrecht, B., Chavanon, M.-L., & Christiansen, H. (2022). Symptom and performance validity in samples of adults at clinical evaluation of ADHD: A replication study using machine learning algorithms. Journal of Clinical and Experimental Neuropsychology, 44(3), 171–184. [Google Scholar] [CrossRef]
  49. Hoelzle, J. B., Ritchie, K. A., Marshall, P. S., Vogt, E. M., & Marra, D. E. (2019). Erroneous conclusions: The impact of failing to identify invalid symptom presentation when conducting adult attention-deficit/hyperactivity disorder (ADHD) research. Psychological Assessment, 31(9), 1174–1179. [Google Scholar] [CrossRef]
  50. Hox, J., Moerbeek, M., & Van de Schoot, R. (2017). Multilevel analysis: Techniques and applications. Routledge. [Google Scholar]
  51. Hu, L., & Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternatives. Structural Equation Modeling: A Multidisciplinary Journal, 6(1), 1–55. [Google Scholar] [CrossRef]
  52. Isvoranu, A.-M., & Epskamp, S. (2023). Which estimation method to choose in network psychometrics? Deriving guidelines for applied researchers. Psychological Methods, 28(4), 925. [Google Scholar] [CrossRef]
  53. Kaiser, H. F. (1970). A second generation little jiffy. Psychometrika, 35(4), 401–415. [Google Scholar] [CrossRef]
  54. Kaiser, H. F. (1974). An index of factorial simplicity. Psychometrika, 39(1), 31–36. [Google Scholar] [CrossRef]
  55. Kiliç, A. F., & Doğan, N. (2021). Comparison of confirmatory factor analysis estimation methods on mixed-format data. International Journal of Assessment Tools in Education, 8(1), 21–37. [Google Scholar] [CrossRef]
  56. Klein, R. G., Mannuzza, S., Olazagasti, M. A. R., Roizen, E., Hutchison, J. A., Lashua, E. C., & Castellanos, F. X. (2012). Clinical and functional outcome of childhood attention-deficit/hyperactivity disorder 33 years later. Archives of General Psychiatry, 69(12), 1295. [Google Scholar] [CrossRef]
  57. Kline, R. B. (2023). Principles and practice of structural equation modeling. Guilford Publications. [Google Scholar]
  58. Kneidinger, J., Fuermaier, A. B. M., Raasch, E., Albrecht, B., & Christiansen, H. (2025). Advancing adult ADHD assessment? The utility of attention based performance validity testing (manuscript submitted for publication). Applied Neuropsychology. under revision. [Google Scholar]
  59. Kooij, J. J. S., Francken, M. H., & Bron, T. I. (2019). Diagnostic interview for ADHD in adults (3rd ed.). DIVA-5. DIVA Foundation. [Google Scholar]
  60. Larrabee, G. J. (2012). Performance validity and symptom validity in neuropsychological assessment. Journal of the International Neuropsychological Society, 18(4), 625–630. [Google Scholar] [CrossRef]
  61. Leonhard, C. (2023a). Review of statistical and methodological issues in the forensic prediction of malingering from validity tests: Part I: Statistical issues. Neuropsychology Review, 33(3), 581–603. [Google Scholar] [CrossRef]
  62. Leonhard, C. (2023b). Review of statistical and methodological issues in the forensic prediction of malingering from validity tests: Part II—Methodological issues. Neuropsychology Review, 33(3), 604–623. [Google Scholar] [CrossRef]
  63. Lippa, S. M. (2018). Performance validity testing in neuropsychology: A clinical guide, critical review, and update on a rapidly evolving literature. The Clinical Neuropsychologist, 32(3), 391–421. [Google Scholar] [CrossRef]
  64. Lis, S., Baer, N., Stein-En-Nosse, C., Gallhofer, B., Sammer, G., & Kirsch, P. (2010). Objective measurement of motor activity during cognitive performance in adults with attention-deficit/hyperactivity disorder. Acta Psychiatrica Scandinavica, 122(4), 285–294. [Google Scholar] [CrossRef]
  65. Lubke, G. H., & Muthén, B. O. (2004). Applying multigroup confirmatory factor models for continuous outcomes to Likert scale data complicates meaningful group comparisons. Structural Equation Modeling, 11(4), 514–534. [Google Scholar] [CrossRef]
  66. Martel, M. M., Levinson, C. A., Langer, J. K., & Nigg, J. T. (2016). A network analysis of developmental change in adhd symptom structure from preschool to adulthood. Clinical Psychological Science, 4(6), 988–1001. [Google Scholar] [CrossRef] [PubMed]
  67. Mascarenhas, M. A., Cocunato, J. L., Armstrong, I. T., Harrison, A. G., & Zakzanis, K. K. (2023). Base rates of non-credible performance in a post-secondary student sample seeking accessibility accommodations. The Clinical Neuropsychologist, 37(8), 1608–1628. [Google Scholar] [CrossRef]
  68. Merten, T., Bossink, L., & Schmand, B. (2007). On the limits of effort testing: Symptom validity tests and severity of neurocognitive symptoms in nonlitigant patients. Journal of Clinical and Experimental Neuropsychology, 29(3), 308–318. [Google Scholar] [CrossRef]
  69. Merten, T., & Merckelbach, H. (2013). Symptom validity testing in somatoform and dissociative disorders: A critical review. Psychological Injury and Law, 6(2), 122–137. [Google Scholar] [CrossRef]
  70. Messerly, J., Soble, J. R., Webber, T. A., Alverson, W. A., Fullen, C., Kraemer, L. D., & Marceaux, J. C. (2021). Evaluation of the classification accuracy of multiple performance validity tests in a mixed clinical sample. Applied Neuropsychology: Adult, 28(6), 727–736. [Google Scholar] [CrossRef] [PubMed]
  71. Musso, M. W., & Gouvier, W. D. (2014). “Why is this so hard?” A review of detection of malingered ADHD in college students. Journal of Attention Disorders, 18(3), 186–201. [Google Scholar] [CrossRef] [PubMed]
  72. Nagelkerke, N. J. (1991). A note on a general definition of the coefficient of determination. Biometrika, 78(3), 691–692. [Google Scholar] [CrossRef]
  73. Nelson, N. W., Sweet, J. J., Berry, D. T. R., Bryant, F. B., & Granacher, R. P. (2007). Response validity in forensic neuropsychology: Exploratory factor analytic evidence of distinct cognitive and psychological constructs. Journal of the International Neuropsychological Society, 13(3), 440–449. [Google Scholar] [CrossRef] [PubMed]
  74. Onandia-Hinchado, I., Pardo-Palenzuela, N., & Diaz-Orueta, U. (2021). Cognitive characterization of adult attention deficit hyperactivity disorder by domains: A systematic review. Journal of Neural Transmission, 128(7), 893–937. [Google Scholar] [CrossRef]
  75. Ovsiew, G. P., Cerny, B. M., Boer, A. B. D., Petry, L. G., Resch, Z. J., Durkin, N. M., & Soble, J. R. (2023). Performance and symptom validity assessment in attention deficit/hyperactivity disorder: Base rates of invalidity, concordance, and relative impact on cognitive performance. The Clinical Neuropsychologist, 37(7), 1498–1515. [Google Scholar] [CrossRef]
  76. Phillips, M. S., Wisinger, A. M., Lapitan-Moore, F. T., Ausloos-Lozano, J. E., Bing-Canar, H., Durkin, N. M., Ovsiew, G. P., Resch, Z. J., Jennette, K. J., & Soble, J. R. (2023). Cross-validation of multiple embedded performance validity indices in the rey auditory verbal learning test and brief visuospatial memory test-revised in an adult attention deficit/hyperactivity disorder clinical sample. Psychological Injury and Law, 16(1), 27–35. [Google Scholar] [CrossRef]
  77. Qbtech. (2025). ADHD tests. Available online: https://www.qbtech.com/adhd-tests (accessed on 25 February 2025).
  78. Resch, Z. J., Rhoads, T., Ovsiew, G. P., & Soble, J. R. (2022). A known-groups validation of the medical symptom validity test and analysis of the genuine memory impairment profile. Assessment, 29(3), 455–466. [Google Scholar] [CrossRef]
  79. Robinson, A. D., & Calamia, M. (2025). Raising the bar: Preliminary investigation of alternative MSVT cutoffs for adults referred for ADHD. Applied Neuropsychology Adult, 16, 1–9. [Google Scholar] [CrossRef] [PubMed]
  80. Robinson, E. V., & Rogers, R. (2018). Detection of feigned ADHD across two domains: The MMPI-2-RF and CAARS for faked symptoms and TOVA for simulated attention deficits. Journal of Psychopathology and Behavioral Assessment, 40(3), 376–385. [Google Scholar] [CrossRef]
  81. Roor, J. J., Dandachi-Fitzgerald, B., Peters, M. J. V., Knoop, H., & Ponds, R. W. H. M. (2022). Performance validity and outcome of cognitive behavior therapy in patients with chronic fatigue syndrome. Journal of the International Neuropsychological Society, 28(5), 473–482. [Google Scholar] [CrossRef]
  82. Rosenfeld, B., Sands, S. A., & Gorp, W. G. V. (2000). Have we forgotten the base rate problem? Methodological issues in the detection of distortion. Archives of Clinical Neuropsychology, 15(4), 349–359. [Google Scholar] [CrossRef]
  83. RStudio. (2024). RStudio: Integrated development for R. In (Version Version 2024.09.0+375) Rstudio, PBC. Available online: https://www.rstudio.com/ (accessed on 1 November 2024).
  84. Rubin, D. B. (1976). Inference and missing data. Biometrika, 63(3), 581–592. [Google Scholar] [CrossRef]
  85. Rubin, D. B. (2004). Multiple imputation for nonresponse in surveys (Vol. 81). John Wiley & Sons. [Google Scholar]
  86. Ruocco, A. C., Swirsky-Sacchetti, T., Chute, D. L., Mandel, S., Platek, S. M., & Zillmer, E. A. (2008). Distinguishing between neuropsychological malingering and exaggerated psychiatric symptoms in a neuropsychological setting. The Clinical Neuropsychologist, 22(3), 547–564. [Google Scholar] [CrossRef]
  87. Sattler, S., Van Veen, F., Hasselhorn, F., El Tabei, L., Fuhr, U., & Mehlkop, G. (2025). Prevalence of legal, prescription, and illegal drugs aiming at cognitive enhancement across sociodemographic groups in germany. Deviant Behavior, 46(3), 253–287. [Google Scholar] [CrossRef]
  88. Schmand, B., & Lindeboom, J. (2005). Amsterdam short-term memory test. APA PsycTests. [Google Scholar]
  89. Sherman, E. M., Slick, D. J., & Iverson, G. L. (2020). Multidimensional malingering criteria for neuropsychological assessment: A 20-year update of the malingered neuropsychological dysfunction criteria. Archives of Clinical Neuropsychology, 35(6), 735–764. [Google Scholar] [CrossRef]
  90. Sibley, M. H. (2021). Empirically-informed guidelines for first-time adult ADHD diagnosis. Journal of Clinical and Experimental Neuropsychology, 43(4), 340–351. [Google Scholar] [CrossRef] [PubMed]
  91. Sibley, M. H., Rohde, L. A., Swanson, J. M., Hechtman, L. T., Molina, B. S. G., Mitchell, J. T., Arnold, L. E., Caye, A., Kennedy, T. M., Roy, A., & Stehli, A. (2018). Late-onset ADHD reconsidered with comprehensive repeated assessments between ages 10 and 25. American Journal of Psychiatry, 175(2), 140–149. [Google Scholar] [CrossRef]
  92. Smart, C. M., Nelson, N. W., Sweet, J. J., Bryant, F. B., Berry, D. T. R., Granacher, R. P., & Heilbronner, R. L. (2008). Use of MMPI-2 to predict cognitive effort: A hierarchically optimal classification tree analysis. Journal of the International Neuropsychological Society, 14(5), 842–852. [Google Scholar] [CrossRef][Green Version]
  93. Stulemeijer, M., Andriessen, T. M. J. C., Brauer, J. M. P., Vos, P. E., & Van Der Werf, S. (2007). Cognitive performance after mild traumatic brain injury: The impact of poor effort on test results and its relation to distress, personality and litigation. Brain Injury, 21(3). [Google Scholar] [CrossRef]
  94. Suhr, J. A., Buelow, M., & Riddle, T. (2011). Development of an infrequency index for the CAARS. Journal of Psychoeducational Assessment, 29(2), 160–170. [Google Scholar] [CrossRef]
  95. Suhr, J. A., Lee, G. J., & Angers, K. (2022). Validity assessment in academic evaluations. In R. W. Schroeder, & P. K. Martin (Eds.), Validity assessment in clinical neuropsychological practice: Evaluating and managing noncredible performance (Vol., pp. 300–318). The Guilford Press. [Google Scholar]
  96. Sullivan, B. K., May, K., & Galbally, L. (2007). Symptom exaggeration by college adults in attention-deficit hyperactivity disorder and learning disorder assessments. Applied Neuropsychology, 14(3), 189–207. [Google Scholar] [CrossRef] [PubMed]
  97. Sweet, J. J., Heilbronner, R. L., Morgan, J. E., Larrabee, G. J., Rohling, M. L., Boone, K. B., Kirkwood, M. W., Schroeder, R. W., & Suhr, J. A. (2021). American Academy of Clinical Neuropsychology (AACN) 2021 consensus statement on validity assessment: Update of the 2009 AACN consensus conference statement on neuropsychological assessment of effort, response bias, and malingering. The Clinical Neuropsychologist, 35(6), 1053–1106. [Google Scholar] [CrossRef] [PubMed]
  98. Tibshirani, R. (1996). Regression shrinkage and selection via the lasso. Journal of the Royal Statistical Society Series B: Statistical Methodology, 58(1), 267–288. [Google Scholar] [CrossRef]
  99. Tombaugh, T. N. (1996). Test of memory malingering, TOMM. Multy-Health Systems. [Google Scholar]
  100. Tucha, L., Fuermaier, A. B. M., Koerts, J., Groen, Y., & Thome, J. (2015). Detection of feigned attention deficit hyperactivity disorder. Journal of Neural Transmission, 122(S1), 123–134. [Google Scholar] [CrossRef]
  101. Ulberstad, F. (2012). QbTest technical manual. QbTech AB. [Google Scholar]
  102. Van Dyke, S. A., Millis, S. R., Axelrod, B. N., & Hanks, R. A. (2013). Assessing effort: Differentiating performance and symptom validity. The Clinical Neuropsychologist, 27(8), 1234–1246. [Google Scholar] [CrossRef]
  103. Wallace, E. R., Garcia-Willingham, N. E., Walls, B. D., Bosch, C. M., Balthrop, K. C., & Berry, D. T. (2019). A meta-analysis of malingering detection measures for attention-deficit/hyperactivity disorder. Psychological Assessment, 31(2), 265–270. [Google Scholar] [CrossRef]
  104. Walls, B. D., Wallace, E. R., Brothers, S. L., & Berry, D. T. (2017). Utility of the conners’ adult ADHD rating scale validity scales in identifying simulated attention-deficit hyperactivity disorder and random responding. Psychological Assessment, 29(12), 1437. [Google Scholar] [CrossRef] [PubMed]
  105. Whitney, K. A., Davis, J. J., Shepard, P. H., & Herman, S. M. (2008). Utility of the Response Bias Scale (RBS) and other MMPI-2 validity scales in predicting TOMM performance. Archives of Clinical Neuropsychology, 23(7–8), 777–786. [Google Scholar] [CrossRef] [PubMed]
  106. Wittchen, H. U., Wunderlich, U., Gruschwitz, S., & Zaudig, M. (1997). Strukturiertes klinisches interview für DSM-IV, Achse I.: Eine deutschsprachige erweiterte bearbeitung der amerikanischen originalversion des SCID-I. Hogrefe. [Google Scholar]
  107. Zalsman, G., & Shilton, T. (2016). Adult ADHD: A new disease? International Journal of Psychiatry in Clinical Practice, 20(2), 70–76. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Heatmap of the correlation matrix for each PVT, SVT, Qb+ test, and self-reported ADHD symptoms (N = 461). Note. ASTM = Amsterdam Short-Term Memory test; MSVT_IR = Medical Symptom Validity Test (Immediate Recall); MSVT_DR = Medical Symptom Validity Test (Delayed Recognition); MSVT_CNS = Medical Symptom Validity Test (Consistency Scores); GET_Index = Groningen Effort Test (Index); GET_errors = Groningen Effort Test (Errors); RTQ = Reaction Time Q-score (Qb+); RTVarQ = Reaction Time Variation Q-score (Qb+); OmissQ = Omission Error Q-score (Qb+); CommQ = Commission Error Q-score (Qb+); RW_CAARS_Inattention = CAARS-L:S (Inattention); RW_CAARS_Hyperactivity = CAARS-L:S (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:S (Impulsivity); RW_CAARS_Self_Concept = CAARS-L:S (Self-Concept); RW_CAARS_O_Inattention = CAARS-L:O (Inattention); RW_CAARS_O_Hyperactivity = CAARS-L:O (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:O (Impulsivity); RW_CAARS_O_Self_Concept = CAARS-L:O (Self-Concept); CAARS_CII = CAARS-L:S (Infrequency Index); CAARS_Incon = CAARS-L:S (Inconsistency Index). Heatmap follows a gradient, where yellow represents a strong positive correlation (r ≈ +1), dark brown/purple represents a strong negative correlation (r ≈ −1), and darker orange shades indicate weaker/no correlation (r ≈ 0).
Figure 1. Heatmap of the correlation matrix for each PVT, SVT, Qb+ test, and self-reported ADHD symptoms (N = 461). Note. ASTM = Amsterdam Short-Term Memory test; MSVT_IR = Medical Symptom Validity Test (Immediate Recall); MSVT_DR = Medical Symptom Validity Test (Delayed Recognition); MSVT_CNS = Medical Symptom Validity Test (Consistency Scores); GET_Index = Groningen Effort Test (Index); GET_errors = Groningen Effort Test (Errors); RTQ = Reaction Time Q-score (Qb+); RTVarQ = Reaction Time Variation Q-score (Qb+); OmissQ = Omission Error Q-score (Qb+); CommQ = Commission Error Q-score (Qb+); RW_CAARS_Inattention = CAARS-L:S (Inattention); RW_CAARS_Hyperactivity = CAARS-L:S (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:S (Impulsivity); RW_CAARS_Self_Concept = CAARS-L:S (Self-Concept); RW_CAARS_O_Inattention = CAARS-L:O (Inattention); RW_CAARS_O_Hyperactivity = CAARS-L:O (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:O (Impulsivity); RW_CAARS_O_Self_Concept = CAARS-L:O (Self-Concept); CAARS_CII = CAARS-L:S (Infrequency Index); CAARS_Incon = CAARS-L:S (Inconsistency Index). Heatmap follows a gradient, where yellow represents a strong positive correlation (r ≈ +1), dark brown/purple represents a strong negative correlation (r ≈ −1), and darker orange shades indicate weaker/no correlation (r ≈ 0).
Behavsci 15 01108 g001
Figure 2. Network analysis of PVT, SVT, and Qb+ measures in relation to ADHD symptoms in the full sample (N = 461). Note: nodes represent the PVT (ASTM, MSVT, GET), QB+ (RTQ, RTVarQ, OmissQ, CommQ), and SVT (CAARS_CII, CAARS_Incon) variables and the CAARS_S/O (Inattention, Hyperactivity, Impulsivity, Self-Concept) scales. Variables are color-scaled in the aforementioned groups. The edges connecting the nodes depict regularized partial Spearman correlations. Thicker and more vividly colored edges indicate stronger absolute correlations. Green edges display positive correlations; red edges display negative correlations. ASTM = Amsterdam Short-Term Memory test; MSVT_IR = Medical Symptom Validity Test (Immediate Recall); MSVT_DR = Medical Symptom Validity Test (Delayed Recognition); MSVT_CNS = Medical Symptom Validity Test (Consistency Scores); GET_Index = Groningen Effort Test (Index); GET_errors = Groningen Effort Test (Errors); RTQ = Reaction Time Q-score (Qb+); RTVarQ = Reaction Time Variation Q-score (Qb+); OmissQ = Omission Error Q-score (Qb+); CommQ = Commission Error Q-score (Qb+); RW_CAARS_Inattention = CAARS-L:S (Inattention); RW_CAARS_Hyperactivity = CAARS-L:S (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:S (Impulsivity); RW_CAARS_Self_Concept = CAARS-L:S (Self-Concept); RW_CAARS_O_Inattention = CAARS-L:O (Inattention); RW_CAARS_O_Hyperactivity = CAARS-L:O (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:O (Impulsivity); RW_CAARS_O_Self_Concept = CAARS-L:O (Self-Concept); CAARS_CII = CAARS-L:S (Infrequency Index); CAARS_Incon = CAARS-L:S (Inconsistency Index).
Figure 2. Network analysis of PVT, SVT, and Qb+ measures in relation to ADHD symptoms in the full sample (N = 461). Note: nodes represent the PVT (ASTM, MSVT, GET), QB+ (RTQ, RTVarQ, OmissQ, CommQ), and SVT (CAARS_CII, CAARS_Incon) variables and the CAARS_S/O (Inattention, Hyperactivity, Impulsivity, Self-Concept) scales. Variables are color-scaled in the aforementioned groups. The edges connecting the nodes depict regularized partial Spearman correlations. Thicker and more vividly colored edges indicate stronger absolute correlations. Green edges display positive correlations; red edges display negative correlations. ASTM = Amsterdam Short-Term Memory test; MSVT_IR = Medical Symptom Validity Test (Immediate Recall); MSVT_DR = Medical Symptom Validity Test (Delayed Recognition); MSVT_CNS = Medical Symptom Validity Test (Consistency Scores); GET_Index = Groningen Effort Test (Index); GET_errors = Groningen Effort Test (Errors); RTQ = Reaction Time Q-score (Qb+); RTVarQ = Reaction Time Variation Q-score (Qb+); OmissQ = Omission Error Q-score (Qb+); CommQ = Commission Error Q-score (Qb+); RW_CAARS_Inattention = CAARS-L:S (Inattention); RW_CAARS_Hyperactivity = CAARS-L:S (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:S (Impulsivity); RW_CAARS_Self_Concept = CAARS-L:S (Self-Concept); RW_CAARS_O_Inattention = CAARS-L:O (Inattention); RW_CAARS_O_Hyperactivity = CAARS-L:O (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:O (Impulsivity); RW_CAARS_O_Self_Concept = CAARS-L:O (Self-Concept); CAARS_CII = CAARS-L:S (Infrequency Index); CAARS_Incon = CAARS-L:S (Inconsistency Index).
Behavsci 15 01108 g002
Figure 3. Node Expected Influence in the full sample (N = 461). Note: higher standardized Z-scores indicate greater expected influence, meaning that nodes with higher scores have stronger and closer connections with other neuropsychological test variables in the network: PVT (ASTM, MSVT, GET), QB+ (RTQ, RTVarQ, OmissQ, CommQ), and SVT (CAARS_CII, CAARS_Incon) variables, and the CAARS_S/O (Inattention, Hyperactivity, Impulsivity, Self-Concept) scales. ASTM = Amsterdam Short-Term Memory test; MSVT_IR = Medical Symptom Validity Test (Immediate Recall); MSVT_DR = Medical Symptom Validity Test (Delayed Recognition); MSVT_CNS = Medical Symptom Validity Test (Consistency Scores); GET_Index = Groningen Effort Test (Index); GET_Errors = Groningen Effort Test (Errors); RTQ = Reaction Time Q-score (Qb+); RTVarQ = Reaction Time Variation Q-score (Qb+); OmissQ = Omission Error Q-score (Qb+); CommQ = Commission Error Q-score (Qb+); RW_CAARS_Inattention = CAARS-L:S (Inattention); RW_CAARS_Hyperactivity = CAARS-L:S (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:S (Impulsivity); RW_CAARS_Self_Concept = CAARS-L:S (Self-Concept); RW_CAARS_O_Inattention = CAARS-L:O (Inattention); RW_CAARS_O_Hyperactivity = CAARS-L:O (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:O (Impulsivity); RW_CAARS_O_Self_Concept = CAARS-L:O (Self-Concept); CAARS_CII = CAARS-L:S (Infrequency Index); CAARS_Incon = CAARS-L:S (Inconsistency Index).
Figure 3. Node Expected Influence in the full sample (N = 461). Note: higher standardized Z-scores indicate greater expected influence, meaning that nodes with higher scores have stronger and closer connections with other neuropsychological test variables in the network: PVT (ASTM, MSVT, GET), QB+ (RTQ, RTVarQ, OmissQ, CommQ), and SVT (CAARS_CII, CAARS_Incon) variables, and the CAARS_S/O (Inattention, Hyperactivity, Impulsivity, Self-Concept) scales. ASTM = Amsterdam Short-Term Memory test; MSVT_IR = Medical Symptom Validity Test (Immediate Recall); MSVT_DR = Medical Symptom Validity Test (Delayed Recognition); MSVT_CNS = Medical Symptom Validity Test (Consistency Scores); GET_Index = Groningen Effort Test (Index); GET_Errors = Groningen Effort Test (Errors); RTQ = Reaction Time Q-score (Qb+); RTVarQ = Reaction Time Variation Q-score (Qb+); OmissQ = Omission Error Q-score (Qb+); CommQ = Commission Error Q-score (Qb+); RW_CAARS_Inattention = CAARS-L:S (Inattention); RW_CAARS_Hyperactivity = CAARS-L:S (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:S (Impulsivity); RW_CAARS_Self_Concept = CAARS-L:S (Self-Concept); RW_CAARS_O_Inattention = CAARS-L:O (Inattention); RW_CAARS_O_Hyperactivity = CAARS-L:O (Hyperactivity); RW_CAARS_Impulsivity = CAARS-L:O (Impulsivity); RW_CAARS_O_Self_Concept = CAARS-L:O (Self-Concept); CAARS_CII = CAARS-L:S (Infrequency Index); CAARS_Incon = CAARS-L:S (Inconsistency Index).
Behavsci 15 01108 g003
Figure 4. Edge weight accuracy estimation in the full sample. Note. Bootstrapped CIs of the estimated edge weights for the estimated network of the full sample. Each horizontal line represents an edge in the network, arranged from the highest to the lowest edge weight. The red line shows the sample edge weights, while the black line represents the bootstrap mean edge weights. The gray area indicates the bootstrapped confidence intervals, and the y-axis labels have been removed to reduce visual clutter.
Figure 4. Edge weight accuracy estimation in the full sample. Note. Bootstrapped CIs of the estimated edge weights for the estimated network of the full sample. Each horizontal line represents an edge in the network, arranged from the highest to the lowest edge weight. The red line shows the sample edge weights, while the black line represents the bootstrap mean edge weights. The gray area indicates the bootstrapped confidence intervals, and the y-axis labels have been removed to reduce visual clutter.
Behavsci 15 01108 g004
Table 1. Sample descriptives and clinical information.
Table 1. Sample descriptives and clinical information.
NMageSDageMinimumageMaximumageADHD Diagnosis (%)
Total sample46133.210.41777388 (84.2%)
Sex
 Females26832.910.52077230 (85.8%)
 Males19333.610.31760158 (81.9%)
Comorbidity 1 15834.611.051877128 (81.0%)
Highest educational level 2
 Unknown337.216.021523 (100%)
 No school degree1----1 (100%)
 Basic school degree7934.913.3177767 (84.8%)
 Secondary school degree7136.510.1226068 (95.8%)
 Grammar school degree17829.58.622064147 (82.6%)
 University degree 12935.49.302065102 (79.1%)
Note. N = number of cases, M = mean, SD = standard deviation. 1 Comorbidity is determined using SCID (SCID-I; Wittchen et al., 1997) reported and classified according to ICD-10 mental disorders (F1–F9), excluding ADHD as a potential diagnosis. 2 Education = no school degree/no formal degree (less than 9 years of schooling); basic school degree (9–10 years of school); secondary school degree with additional vocational training (10–12 years of schooling); grammar school degree with university entrance qualification (usually 12–13 years of schooling); university degree (usually 16–17 years of schooling).
Table 2. Measures of PVTs, SVTs, CPT, and ADHD symptom reports (N = 461).
Table 2. Measures of PVTs, SVTs, CPT, and ADHD symptom reports (N = 461).
Below Average 1
MSDMinimum Maximum Positive Results (%)Female (%)Male (%)
GET
 Index0.093.32−1715141 (30.6%)
 Errors 7.116.8805759 (12.8%)
 Index + Errors 47 (10.2%)
ASTM86.95.03569070 (15.2%)
MSVT 8 (1.7%)
 IR99.42.36801001 (0.3%)
 DR98.93.27751005 (1.1%)
 CNS98.54.22651008 (1.7%)
Qb+ 2
 RTQ0.681.01−24.9 38 (14.2%)51 (26.4%)
 RTVarQ1.081.14−1.84.5 91 (34.0%)65 (33.7%)
 OmissionQ1.31.07−1.53.5 144 (53.7%) 84 (43.5%)
 CommissionQ0.631.23−1.64.7 75 (28.0%)37 (19.2%)
CAARS—Inconsistency5.942.18113107 (23.2%)
CAARS Infrequency Index (CII)17.75.59135109 (23.6%)
CAARS-L:S 3
 Inattention 78.410.63490 240 (89.5%)173 (89.6%)
 Hyperactivity69.914.03490 178 (66.4%)124 (64.2%)
 Impulsivity70.112.33690 204 (76.2%) 111 (57.5%)
 Self-Concept Problems72.012.13590 196 (73.1%) 147 (76.2%)
CAARS-L:O 4
 Inattention71.612.93790 192 (71.6%) 128 (66.3%)
 Hyperactivity64.114.53590 133 (49.6%) 87 (45.1%)
 Impulsivity65.212.73590 147 (54.9%) 89 (46.1%)
 Self-Concept Problems 67.612.23690 169 (63.1%) 120 (62.2%)
Note. M = mean, SD = standard deviation. 1 Below average is defined as PR < 8 on neuropsychological tests (Qb+) and PR > 92 on questionnaire scores. 2 Qb+ = Q-based norm scores (adjusted for age and gender). 3,4 CAARS-L:S; CAARS-L:O = T-scores, adjusted for age and gender. ASTM = Amsterdam Short-Term Memory test; MSVT IR, DR, CNS = Medical Symptom Validity Test (Immediate Recall, Delayed Recognition, Consistency Scores); GET Index, Errors = Groningen Effort Test (Index, Errors); Qb+ = Quantified Behaviour Test; RTQ = Reaction Time Q-score; RTVarQ = Reaction Time Variation Q-score; OmissionQ = Omission Error Q-score; CommissionQ = Commission Error Q-score; CAARS—Inconsistency = CAARS-L:S (Inconsistency Index); CAARS Infrequency Index (CII) = CAARS-L:S (Infrequency Index); CAARS-L:S = Conners Adult ADHD Rating Scale, Longform Self-report (Inattention, Hyperactivity, Impulsivity, Self-Concept Problems); CAARS-L:O = Conners Adult ADHD Rating Scale, Longform Observer report (Inattention, Hyperactivity, Impulsivity, Self-Concept Problems).
Table 3. Item allocation to factor structure (N = 461).
Table 3. Item allocation to factor structure (N = 461).
Measures Conceptual Three-Factor ModelConceptual Two-Factor Model Conceptual One-Factor Model
GET
  IndexPerformance ValidityTest PerformanceOne Factor
  Errors Performance ValidityTest PerformanceOne Factor
ASTM Performance ValidityTest PerformanceOne Factor
MSVT
  IRPerformance ValidityTest PerformanceOne Factor
  DRPerformance ValidityTest PerformanceOne Factor
  CNSPerformance ValidityTest PerformanceOne Factor
Qb+
  RTQContinuous PerformanceTest PerformanceOne Factor
  RTVarQContinuous PerformanceTest PerformanceOne Factor
  OmissionQContinuous PerformanceTest PerformanceOne Factor
  CommissionQ Continuous PerformanceTest PerformanceOne Factor
CAARS—InconsistencySymptom ReportSymptom ReportOne Factor
CAARS Infrequency Index (CII)Symptom ReportSymptom ReportOne Factor
CAARS-L:S
  InattentionSymptom ReportSymptom ReportOne Factor
  HyperactivitySymptom ReportSymptom ReportOne Factor
  ImpulsivitySymptom ReportSymptom ReportOne Factor
  Self-Concept ProblemsSymptom ReportSymptom ReportOne Factor
CAARS-L:O
  InattentionSymptom ReportSymptom ReportOne Factor
  HyperactivitySymptom ReportSymptom ReportOne Factor
  ImpulsivitySymptom ReportSymptom ReportOne Factor
  Self-Concept Problems Symptom ReportSymptom ReportOne Factor
Note. ASTM = Amsterdam Short-Term Memory test; MSVT IR, DR, CNS = Medical Symptom Validity Test (Immediate Recall, Delayed Recognition, Consistency Scores); GET Index, Errors = Groningen Effort Test (Index, Errors); Qb+ = Quantified Behaviour Test; RTQ = Reaction Time Q-score; RTVarQ = Reaction Time Variation Q-score; OmissionQ = Omission Error Q-score; CommissionQ = Commission Error Q-score; CAARS—Inconsistency = CAARS-L:S (Inconsistency Index); CAARS Infrequency Index (CII) = CAARS-L:S (Infrequency Index); CAARS-L:S = Conners Adult ADHD Rating Scale, Longform Self-report (Inattention, Hyperactivity, Impulsivity, Self-Concept Problems); CAARS-L:O = Conners Adult ADHD Rating Scale, Longform Observer report (Inattention, Hyperactivity, Impulsivity, Self-Concept Problems).
Table 4. Measures of fit 1 for the one-, two-, and three-factor confirmatory factor analysis model.
Table 4. Measures of fit 1 for the one-, two-, and three-factor confirmatory factor analysis model.
Goodness of Fit χ2χ2/dfRMSEACFITLI
Three-factor Model1550.49.280.1360.6690.623
Two-factor Model 1945.711.50.1540.5700.517
One-factor Model 2835.916.70.190.340.27
Note. RMSEA = root mean square error of approximation; TLI = Tucker–Lewis index; CFI = comparative fit index. Model–data fit is considered acceptable when χ2/df ranges from 2 to 5, RMSEA is below 0.06, and CFI and TLI estimates exceed 0.80. A fit is considered good when χ2/df is close to 1, and CFI and TLI estimates surpass 0.95 (Hox et al., 2017). 1 The measures of fit are represented as robust variables and z-standardized.
Table 5. Multivariate Logistic Regression models and odds ratios based on descriptive and clinical information to predict the pass/fail PVT outcomes.
Table 5. Multivariate Logistic Regression models and odds ratios based on descriptive and clinical information to predict the pass/fail PVT outcomes.
PredictorsßSEZOR p
ASTM 1
Age−0.0020.013−0.1831.000.855
Sex (Male/Female)0.0020.3050.0061.000.995
ADHD (No/Yes) 0.4020.4310.9331.490.351
Comorbidity 2 (No/Yes) 0.0540.3150.1711.060.865
Cognitive Deficit 3 (No/Yes)−1.230.374−3.280.29<0.001 ***
Education Level 40.3020.1322.301.35<0.05 *
CAARS-L:S
   Inattention0.2110.5130.4121.240.680
   Hyperactivity0.2220.3510.6321.250.528
   Impulsivity −0.7910.402−1.970.45<0.05 *
CAARS-L:O
   Inattention−0.1400.364−0.3830.870.702
   Hyperactivity−0.4990.350−1.440.610.150
   Impulsivity0.9870.3602.782.68<0.01 **
MSVT 1
Age−0.0220.033−0.6610.980.508
Sex (Male/Female)0.8590.7731.112.360.267
ADHD (No/Yes) −17.42849.5−0.0060.000.995
Comorbidity 2 (No/Yes) −0.8120.755−1.080.440.282
Cognitive Deficit 3 (No/Yes)−17.82007.6−0.0090.000.993
Education Level 4 −0.1550.343−0.4520.860.651
CAARS-L:S
   Inattention−17.23516.7−0.0050.000.996
   Hyperactivity−0.6050.909−0.6660.550.506
   Impulsivity 1.350.8731.553.870.121
CAARS-L:O
   Inattention−0.8551.15−0.7410.430.459
   Hyperactivity0.2410.8920.2701.270.787
   Impulsivity0.1010.9520.1061.110.916
GET 1
Age−0.0220.010−2.230.98<0.05 *
Sex (Male/Female)−0.2860.213−1.350.750.178
ADHD (No/Yes) −0.0260.311−0.080.970.933
Comorbidity 2 (No/Yes) −0.1220.215−0.5640.890.573
Cognitive Deficit 3 (No/Yes)−0.2920.217−1.350.750.178
Education Level 4 0.1260.094 1.351.130.178
CAARS-L:S
   Inattention−0.1640.381−0.4310.850.667
   Hyperactivity−0.0790.246−0.3240.920.746
   Impulsivity −0.2750.263−1.040.760.296
CAARS-L:O
   Inattention−0.2720.254−1.070.760.286
   Hyperactivity−0.1260.237−0.5310.880.595
   Impulsivity0.0940.240 0.3931.100.694
Note. 1 PVT measures (ASTM, MSVT, GET) depicted as pass/fail binary variable according to the cut-off criteria described in the test manuals (see; Fuermaier et al., 2021; P. Green, 2004; Schmand & Lindeboom, 2005); ASTM = Amsterdam Short-Term Memory test; MSVT = Medical Symptom Validity Test; GET = Groningen Effort Test. 2 Comorbidity is determined using SCID-I (Wittchen et al., 1997) reported and classified according to ICD-10 mental disorders (F1–F9), excluding ADHD as a potential diagnosis. Depicted as a binary variable indicating the presence/absence of one or more comorbidities. 3 Cognitive Deficit derived from the Qb+ measures (RTQ, RTVarQ, OmissionQ, CommissionQ), any of the four measures Q-score > 1.5; pass/fail binary variable. 4 Data missing from 3 participants, due to unknown education level. Depicted as a continuous variable with the following encoding based on years of schooling: 1 represents no school degree or no formal degree (less than 9 years of schooling), 2 represents a basic school degree (9–10 years of schooling), 3 represents a secondary school degree with additional vocational training (10–12 years of schooling), 4 represents a grammar school degree with university entrance qualification (usually 12–13 years of schooling), and 5 represents a university degree (usually 16–17 years of schooling). Statistically significant at p < 0.05 *, p < 0.01 **, p < 0.001 ***. CAARS-L:S = Conners Adult ADHD Rating Scale, Longform Self-report (Inattention, Hyperactivity, Impulsivity, Self-Concept Problems); CAARS-L:O = Conners Adult ADHD Rating Scale, Longform Observer report (Inattention, Hyperactivity, Impulsivity, Self-Concept Problems).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Raasch, E.; Fuermaier, A.B.M.; Kneidinger, J.; Albrecht, B.; Christiansen, H. Joint Factor Performance Validity?—Network and Factor Structure of Performance Validity Measures in the Clinical Evaluation of Adult ADHD. Behav. Sci. 2025, 15, 1108. https://doi.org/10.3390/bs15081108

AMA Style

Raasch E, Fuermaier ABM, Kneidinger J, Albrecht B, Christiansen H. Joint Factor Performance Validity?—Network and Factor Structure of Performance Validity Measures in the Clinical Evaluation of Adult ADHD. Behavioral Sciences. 2025; 15(8):1108. https://doi.org/10.3390/bs15081108

Chicago/Turabian Style

Raasch, Emily, Anselm B. M. Fuermaier, Johanna Kneidinger, Björn Albrecht, and Hanna Christiansen. 2025. "Joint Factor Performance Validity?—Network and Factor Structure of Performance Validity Measures in the Clinical Evaluation of Adult ADHD" Behavioral Sciences 15, no. 8: 1108. https://doi.org/10.3390/bs15081108

APA Style

Raasch, E., Fuermaier, A. B. M., Kneidinger, J., Albrecht, B., & Christiansen, H. (2025). Joint Factor Performance Validity?—Network and Factor Structure of Performance Validity Measures in the Clinical Evaluation of Adult ADHD. Behavioral Sciences, 15(8), 1108. https://doi.org/10.3390/bs15081108

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop