Next Article in Journal
Cyber-Creativity: A Decalogue of Research Challenges
Previous Article in Journal
The Role of Emotional Intelligence and Frustration Intolerance in the Academic Performance of University Students: A Structural Equation Model
Previous Article in Special Issue
Log File Times as Indicators of Structured Figural Matrix Processing
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Automatic- and Transformer-Based Automatic Item Generation: A Critical Review

Department of Psychology, University of Graz, Universitätsplatz 2, 8010 Graz, Austria
*
Author to whom correspondence should be addressed.
J. Intell. 2025, 13(8), 102; https://doi.org/10.3390/jintelligence13080102
Submission received: 21 May 2025 / Revised: 5 August 2025 / Accepted: 8 August 2025 / Published: 12 August 2025
(This article belongs to the Special Issue Intelligence Testing and Assessment)

Abstract

This article provides a critical review of conceptually different approaches to automatic and transformer-based automatic item generation. Based on a discussion of the current challenges that have arisen due to changes in the use of psychometric tests in recent decades, we outline the requirements that these approaches should ideally fulfill. Subsequently, each approach is examined individually to determine the extent to which it can contribute to meeting the challenges. In doing so, we will focus on the cost savings during the actual item construction phase, the extent to which they may contribute to enhancing test validity, and potential cost savings in the item calibration phase due to either a reduction in the sample size required for item calibration or a reduction in the item loss due to insufficient psychometric characteristics. In addition, the article also aims to outline common recurring themes across these conceptually different approaches and outline areas within each approach that warrant further scientific research.

1. Introduction

The advent of computer-based and internet-based assessment has led to an increase in the use of summative and formative assessments. Combined with changes in best practice recommendations in educational and psychological measurement, this successively resulted in a drastic increase in item construction demands. In this article, we review and discuss several new methods of item construction that have been proposed to meet these challenges. The intention of this article is not to provide a comprehensive review, but to outline the main idea of several conceptually different methods and to discuss the extent to which each method can contribute to the construction of psychometrically sound, valid, and fair cognitive ability tests in a scalable and resource-efficient manner. To this end, we first outline the item construction demands arising from current practices in assessment to set a benchmark for evaluating and comparing various modern item construction methods. In addition, it will discuss current trends within and across different methods and outline gaps in the research literature that require further scientific research.

1.1. Psychometric Costs of Testing on Demand

In the last decade, it has become common to provide test-takers with multiple test-taking opportunities (cf. AERA et al. 2018; Drasgow et al. 2006; Lievens et al. 2005). However, doing so also increases the risk that test-takers attempt to memorize test items and share them with other test-takers who take the psychometric test at a later time point. These prospective test-takers may then memorize the disclosed test items to obtain higher test scores. Several studies indicate that reusing disclosed test items results in a decrease in their item difficulty and item discrimination parameters, and an increase of their local dependencies (e.g., Appelhaus et al. 2023; Joncas et al. 2018; Lievens et al. 2007; Powers 2005; Selvi 2020; Stricker 1984; Veerkamp and Glas 2000; Wagner-Menghin et al. 2013; Wood 2009; Wood et al. 2010; Yang et al. 2018; Zimmermann et al. 2016). If this item disclosure effect is ignored, the person parameter estimates will be biased. The magnitude of the bias depends on the percentage of reused disclosed test items and the magnitude of their item parameter drift (cf. Segall 2004). In addition, the psychometric tests may become more strongly correlated with measures of long-term memory (cf. Lievens et al. 2007). Thus, item disclosure compromises the construct validity of the psychometric test. Unfortunately, it also detrimentally affects the fairness of psychometric tests because test-takers have been shown to differ both in their awareness of the practice of reusing disclosed test items and in their willingness to capitalize on it by memorizing disclosed test items (cf. Powers 2005).
To facilitate testing on demand without compromising validity and fairness, test developers have turned to computerized adaptive testing (CAT: Lang 2011; Segall 2004; van der Linden and Glas 2010), multi-stage testing (MST: Zenisky et al. 2010), and multiple parallel linear on-the-fly test forms (LOFT: Folk and Smith 2002) as an alternative to classic fixed-item linear tests. Research indicates that all three test administration methods are more effective at reducing the risk of item disclosure than classic fixed-item linear tests (cf. Guo et al. 2009; Arendasy and Sommer 2017). Even though CATs have been shown to be more resistant to item disclosure than MSTs and LOFTs, the test security of the latter two methods has been shown to improve as the number of test items in each test form and the number of test forms increase.
In all three test administration methods, test security can be further enhanced by incorporating item- and content-exposure-control algorithms (cf. Georgiadou et al. 2007; Lim and Choi 2024). In addition, using multiple rotating parallel item pools further enhances test security (e.g., Ariel et al. 2006; Zhang and Chang 2005; Zhang et al. 2012). However, all three methods also drastically increase item construction demands (cf. Reckase et al. 2019). Furthermore, the initially constructed item pool needs to be constantly checked and updated (e.g., Ariel et al. 2006; Liu et al. 2019; Sinharay 2017). As a consequence, test developers need to be able to construct a substantial number of psychometrically sound test items with predictable item parameters and content specifications during the entire lifespan of the psychometric test.

1.2. Psychometric Costs of Retesting

Similar to testing on demand, retesting is often conceived as an important aspect of the fair and valid use of psychometric tests by test-takers (cf. AERA et al. 2018; Lievens et al. 2005). Unfortunately, several meta-analyses (cf. Calamia et al. 2012; Hausknecht et al. 2007; Kulik et al. 1984a; Scharfen et al. 2018) indicate that retesting leads to an increase in test-takers’ scores. The effect size of retest-score gains depends on various factors such as the cognitive ability domain measured, test-takers’ general mental ability, the time period between the initial test and retest, the number of retests taken, and the specific kind of retest form used. In general, retest score gains are more pronounced (mean effect sizes: 0.37 to 0.42) when using identical retest forms compared to alternate retest forms (mean effect sizes: 0.22 to 0.23). However, if test items used in the alternate retest form only differ in surface features from those used at the initial test administration, retest score gains are often almost identical to those in identical retest forms (cf. Arendasy and Sommer 2013a, 2017; Matton et al. 2011; Morley et al. 2004). Furthermore, retest effects are larger if the time interval between initial test administration and retest is short. However, they remain notable even with time intervals of several years (cf. Burke 1997). In addition, retest effects tend to increase with the number of retests taken until a plateau is reached. The number of retest sessions needed to reach that plateau differs across cognitive ability domains measured (cf. Calamia et al. 2012; Hausknecht et al. 2007; Kulik et al. 1984a; Scharfen et al. 2018). Furthermore, retest score gains are more pronounced for high-g test-takers than for low-g test-takers (cf. Kulik et al. 1984a).
Several theoretical models have been proposed to explain the observed score gains due to retesting (for a detailed discussion, see Arendasy and Sommer 2017; Lievens et al. 2007). One of the most prominent explanations is the strategy refinement hypothesis. The strategy refinement hypothesis posits that test-takers use various solution strategies that differ in their effectiveness and efficiency (e.g., Bethell-Fox et al. 1984; Fehringer 2020; Heil and Jansen-Osmann 2008; Heil et al. 1998; Jarosz and Wiley 2012; Liu et al. 2023b, 2025; Verguts and De Boeck 2002; Vigneau et al. 2006). The solution strategies are stored in procedural long-term memory together with information on their effectiveness and efficiency given certain task characteristics (e.g., Anderson et al. 1997; Siegler 1996). When solving test items with a lower person-specific difficulty (e.g., person-specific p ≥ .70), test-takers have sufficient cognitive resources left to update information on the effectiveness and efficiency of a solution strategy given certain task characteristics. This allows them to gradually tailor their solution strategy choices to the task demands, which leads to an increase in narrower test-specific cognitive abilities. This hypothesis is consistent with studies indicating that test-takers are able to learn to tailor their solution strategies to task demands during test-taking even with little to no feedback (e.g., Fehringer 2023; Hayes et al. 2015; Heil et al. 1998; Ren et al. 2012; Verguts and De Boeck 2002).
The strategy refinement hypothesis explains several key findings regarding the size of the retest effect. First, since the items of the initial test and an identical retest have more item design features in common, this fosters transfer compared to alternate retest forms, which share fewer item design features with the initial test items (e.g., Anderson et al. 1997; Siegler 1996). Second, the assumption that strategy refinement is a gradual process explains why the number of retests administered has been shown to affect the size of the retest effect. The same explanation also accounts for the finding that high-g test-takers benefit more from retesting than low-g test-takers. High-g test-takers encounter more items with a person-specific lower difficulty while working on a fixed-item linear test than do low-g test-takers. They therefore have more opportunities to learn to refine their solution strategies (Arendasy and Sommer 2017). Third, the hypothesis that strategy refinement leads to a gradual increase in a narrower test-specific cognitive ability is consistent with research findings indicating that retesting neither induces differential item functioning at the item level nor alters the structural relations between the cognitive ability tests and their relation to psychometric g or other higher-order traits (e.g., Arendasy and Sommer 2013a, 2017; Estrada et al. 2015; Reeve and Lam 2005). This is due to strategy refinements leading to a gradual improvement in narrower cognitive abilities. Last but not least, the model also posits that these improvements in narrower cognitive abilities should not generalize to higher-order traits. This final prediction, deduced from the strategy refinement hypothesis, is consistent with research findings indicating that retest score gains do not seem to generalize to higher-order traits (e.g., Arendasy and Sommer 2013a, 2017; Estrada et al. 2015; Krautter et al. 2021; Levacher et al. 2021; Loesche et al. 2015; Matton et al. 2011; Reeve and Lam 2005; Schneider and Sparfeldt 2021a, 2021b; Schneider et al. 2020; te Nijenhuis et al. 2007).
The strategy refinement hypothesis also has implications for how to reduce retest effects. Usually, CATs administer items of higher (p = .50) person-specific difficulty to all test-takers (cf. Arendasy and Sommer 2017; Guo et al. 2009). This should not only prevent strategy refinement during test-taking but also equalize the learning opportunities between high-g and low-g test-takers. As a consequence, high-g test-takers, who take a CAT at the initial test administration session, should no longer benefit more from retesting than their low-g peers. This prediction is consistent with recent research findings indicating that CAT indeed reduces retest effects when administered at the initial test administration session (Arendasy and Sommer 2017). However, the authors also showed that retest effects still exist when test-takers take a fixed-item linear retest form at the initial test administration session, followed by either a CAT or an alternate fixed-item linear test form. This indicates that test-takers may learn to refine their solution strategies in-between test-taking sessions by practicing sample test items, which are commonly used in test preparation.

1.3. Psychometric Costs of Test Preparation

Due to the increased use of psychometric tests, the test preparation industry has become a lucrative business on its own (cf. Buchmann et al. 2010; Powers 2012; Zwick 2002). Prospective test-takers are often offered a variety of different test preparation methods from commercial and/or non-commercial providers. Researchers (e.g., Briggs 2009; Messick 1982; Powers 2012) distinguish between three methods of test preparation: test familiarization, practice-based training, and test coaching.
Test familiarization merely aims to familiarize test-takers with the mechanics of test-taking to reduce construct-irrelevant variance due to test familiarity and test anxiety. This method of test preparation does not intend to improve test-takers’ test scores. Consistent with this aim, research shows that test familiarization increases test-wiseness and reduces test anxiety, but its effect on test-takers’ test scores is negligible (cf. Burns et al. 2008; Powers and Alderman 1983). For this reason, cost-free test familiarization has become a standard practice and is often seen as mandatory by the test-takers (cf. AERA et al. 2018; Denker et al. 2023; Lievens et al. 2005).
Practice-based training, on the other hand, constitutes an informal test-taker-driven test preparation method that aims at improving test-takers’ test scores. During practice-based training, test-takers practice sample-test items with or without feedback to improve their test scores (cf. Briggs 2009; Messick 1982; Powers 2012). Several meta-analyses (cf. Bangert-Drowns et al. 1983; Hausknecht et al. 2007; Kulik et al. 1984a; Messick and Jungeblut 1981) and independent studies (e.g., Burns et al. 2008; Hermes et al. 2019; Hermes et al. 2023; Weppert et al. 2023) indicate that practice-based training improves test-takers’ test scores (Cohen’s d = 0.22 to 0.40). The size of the practice-based training effect varies with the similarity of the sample-test items and the actual test items (cf. Arendasy and Sommer 2013a, 2017; Freund and Holling 2011; Morley et al. 2004), the general mental ability of the test-taker (cf. Kulik et al. 1984a), the amount of time and effort invested into practicing (cf. Bangert-Drowns et al. 1983; Becker 1990; Hermes et al. 2019; Kulik et al. 1984b; Messick and Jungeblut 1981), and the specific cognitive ability domain trained (e.g., Arendasy et al. 2016; Hermes et al. 2019, 2023; Kulik et al. 1984a; Weppert et al. 2023).
Test coaching, on the other hand, is a formal instructor-driven test preparation method that consists of the following components aimed to improve test-takers’ test scores: test familiarization, practice-based training, formal instruction and feedback, and instruction on specific test-taking strategies (cf. Allalouf and Ben-Shakhar 1998; Becker 1990; Briggs 2009; Messick 1982; Powers 2012). Similar to practice-based training, several meta-analyses and large-scale studies (e.g., Becker 1990; Briggs 2009; Bangert-Drowns et al. 1983; Kulik et al. 1984b; Lilly and Montgomery 2011; Messick and Jungeblut 1981; Powers and Rock 1999; Witt 1993; Weppert et al. 2023) indicate that test coaching increases test-takers’ test scores (Cohen’s d = 0.33 to 0.64). The size of the score gains depends on the general mental ability of the test-taker (cf. Kulik et al. 1984b), the time devoted to practicing (cf. Appelrouth et al. 2017; Becker 1990; Messick and Jungeblut 1981), the specific cognitive ability trained (e.g., Becker 1990; Briggs 2009; Bangert-Drowns et al. 1983; Kulik et al. 1984b; Powers and Rock 1999), and the feedback provided (cf. Krautter et al. 2021). The added value of test coaching on its own, compared to practice-based training, is small and has been attributed to the effect of feedback and formal instruction (cf. Appelrouth et al. 2017; Krautter et al. 2021; Levacher et al. 2021; Loesche et al. 2015; Schneider and Sparfeldt 2021a, 2021b) in addition to a more optimal pacing of test preparation (cf. Appelrouth et al. 2017). However, combining test coaching with practice-based training has been shown to lead to a notable increase in test scores.
Thus, both practice-based training and test coaching are effective at improving test scores. Measurement invariance analyses across test-takers differing in the amount and kind of test preparation (e.g., Arendasy et al. 2016; Estrada et al. 2015; Hermes et al. 2019; Levacher et al. 2021; Krautter et al. 2021; Sommer et al. 2025) indicate that individual differences in test preparation neither induce uniform nor non-uniform differential item functioning (DIF) at the item level, nor do they alter the structural relationship between the subtests and the higher-order traits that they have been constructed to measure. Nevertheless, this line of research also clearly shows that score gains do not generalize to higher-order traits. They are confined to the narrower test-specific cognitive abilities trained (cf. Arendasy et al. 2016; Hermes et al. 2019; Sommer et al. 2025). This implies that—similar to retest effects—test preparation-based score gains may reflect an increase in narrower test-specific abilities due to strategy refinement. This conclusion is corroborated by an eye-movement study, which showed that test-takers indeed resorted to more effective and efficient solution strategies after receiving a tutorial on how to solve the test items (Loesche et al. 2015, experiment 4).
Although test preparation does not detrimentally affect measurement invariance and construct validity, it potentially threatens fairness if test-takers differ in their access to test preparation (cf. Borsboom et al. 2008; Buchmann et al. 2010; Sommer et al. 2025). While studies using previous birth cohorts of test-takers (e.g., Buchmann et al. 2010; Park and Becks 2015) indicated that test-takers whose parents are more educated and wealthier had more access to test preparation than their peers, studies conducted with more recent birth cohorts (e.g., Arendasy et al. 2016; Lee et al. 2023a; Sommer et al. 2025) failed to replicate this finding. A possible explanation for this finding is that universities and commercial companies started to provide test-takers with either less costly or even cost-free test preparation opportunities to ensure equal opportunities to perform for all test-takers (cf. AERA et al. 2018; Arendasy et al. 2016; Campion et al. 2019; Hermes et al. 2019; Lee et al. 2023a; Sommer et al. 2025). However, this practice also extends the item construction demands beyond the constructing and maintaining of item pools for assessment.

1.4. Implications for Test Construction

The research outlined above has implications for the construction of test items and sample-test items. Test developers need to be able to construct large item pools that are optimal for their intended use in terms of the shape and distribution of the item parameters, and in terms of the content specifications. Several methods have been proposed to construct blueprints for optimal item pools (e.g., He and Reckase 2014; Reckase 2010; Veldkamp and van der Linden 2010). The drawback of these methods is that they assume that test developers can construct test items with the pre-specified item parameters and content specifications. In the case of cognitive ability tests, content specifications refer to item design features and combinations thereof. Item design features hypothesized to affect the item parameters are called radicals (Irvine 2002), while non-salient item design features (=surface features) that do not affect the item parameters are referred to as incidentals (Irvine 2002). The item pools should not only contain information on the item parameters themselves but also information on the combination of radicals and incidentals used to construct each test item to facilitate the use of item- and content-exposure control algorithms (cf. Georgiadou et al. 2007; Gierl et al. 2022a; Lim and Choi 2024).
To minimize item disclosure, test developers should be able to construct test items whose item parameters not only span the entire range of the person parameter distribution but also mimic its shape and distribution (cf. Reckase 2010; Veldkamp and van der Linden 2010). Furthermore, to construct multiple rotating parallel item pools (e.g., Ariel et al. 2006; Zhang et al. 2012), test developers need to be able to construct multiple instances of all test items with identical radical combinations (and item parameters) but different incidentals (Irvine 2002). These kinds of test items are referred to as item isomorphs (Bejar 2002) or item clones (Glas and van der Linden 2003). Furthermore, to support item- and content-exposure control (e.g., for an overview, see Georgiadou et al. 2007; Gierl et al. 2022a; Lim and Choi 2024), test developers should also be able to construct test items that exhibit statistically identical item parameters, despite using a different combination of radicals, at all segments of the latent trait continuum. These kinds of test items have been referred to as psychometrically matched instances (cf. Arendasy and Sommer 2013a) and are a prerequisite for content exposure control to work as intended.
Item construction demands further increase if the test developers aim to reduce individual differences in access to test preparation by providing cost-free test preparation (e.g., AERA et al. 2018; Arendasy et al. 2016; Campion et al. 2019; Hermes et al. 2019; Lee et al. 2023a; Sommer et al. 2025). According to the standards for educational and psychological assessment (cf. AERA et al. 2018), sample-test items used in test preparation should resemble the actual test items in terms of the item format and item construction principles. This implies that they should be representative of the actual test items in terms of the distribution of their item parameters and also mirror them in terms of their affordances and constraints (Greeno et al. 1993) for construct-related solution strategies. As a result, sample-test items need to be representative for—but not identical to—the actual test items in terms of their radical combinations (Irvine 2002). Furthermore, if test preparation includes feedback and tutorials, these tutorials should also be representative and cover all difficulty levels (cf. Arendasy et al. 2016; Krautter et al. 2021; Sommer et al. 2025). This requires test developers to be able to construct not only multiple item pools intended to be used for assessment, but also one or more item pools consisting of sample-test items that can be used in cost-free practice-based training. Furthermore, since test preparation increases test scores without generalizing to higher-order traits (cf. Section 1.3), test developers need to be able to shift the spectrum of the item difficulties upward to counteract retest and test preparation effects.
The specifications of ideal item pools for assessment and training purposes outlined above are thus rather specific and go beyond merely being able to generate large pools of test items. They can therefore also be used as a benchmark to evaluate and compare conceptually different methods of item construction in terms of their potential to meet these requirements in a cost-efficient and scalable manner.

2. Human-Constructed Test Items (H-IG)

Traditional item construction involves trained human item writers, who are also subject-matter experts in the domain covered by the psychometric test (cf. AERA et al. 2018; Hornke and Habon 1986; Lane et al. 2016). The test development process usually starts with a definition of the latent trait to be measured and the selection of an item format that can be used to measure the latent trait. Next, subject-matter experts outline the cognitive processes hypothesized to be involved in solving the test items, in addition to other content-matter specifications. This is done to provide a more detailed specification of what needs to be achieved during item writing. Once this is done, human item writers construct test items on an item-by-item basis using either their experience and expertise or a blueprint of the item construction rationale (cf. AERA et al. 2018; Hornke and Habon 1986; Lane et al. 2016). Next, the constructed test items are submitted to a panel of independent subject-matter experts for review. This review process covers both subject-matter considerations (e.g., factuality check, etc.) and a fairness review (for details, see AERA et al. 2018; Lane et al. 2016). Once the test items have passed through this review process, they are edited and revised if necessary. In the latter case, the items are again submitted to review. This cycle continues until the test items are deemed to be of sufficient quality by the panel of subject-matter experts. Next, the test items are passed on to field testing to estimate their item parameters, and to evaluate their dimensionality and fairness using descriptive item response theory modeling (cf. De Boeck and Wilson 2004). This brief description of the human item writing process nicely illustrates that item writing is a complex process that requires a considerable amount of financial, time, and human resources. Furthermore, since test items are constructed on an item-to-item basis, this process is hardly scalable. It is therefore not surprising that item construction has been referred to as a critical bottleneck in test construction (cf. Drasgow et al. 2006; Kosh et al. 2019; Wainer 2002; Zickar 2020).
In addition, there is evidence that writing test items requires a lot of expertise and requires continuous training. For instance, Webb et al. (2015) showed that subject-matter experts trained in item writing constructed fewer test items, which were identified as flawed in a subsequent review process, compared to subject-matter experts without any training in item writing. However, a short one-time training in item construction does not seem to suffice (Gupta et al. 2020). Repeated and continuous training is required to construct test items that pass the review process and also exhibit acceptable classical item difficulty and classical item discrimination parameters in subsequent field tests (e.g., Beg et al. 2021; Lee et al. 2024). This result is consistent with research findings indicating that expertise is important in human item writing (e.g., Eleragi et al. 2025; Jozefowicz et al. 2002; Nemec and Welch 2016) and that building expertise through repeated and continuous training and feedback improves the quality of human item writing (cf. Karthikeyan et al. 2019). By contrast, lack of motivation and time pressure reduce the quality of the human-made test items (cf. Karthikeyan et al. 2019). Although—to the best of our knowledge—we currently lack studies examining item writer-related determinants of the quality of human-made items for intelligence tests, it seems plausible that the above-cited results obtained in formative and summative assessment in medical education generalize to the construction of intelligence tests.
Furthermore, there is evidence that human item writers also have difficulties predicting the empirically estimated difficulty of the test items they constructed. Several studies (e.g., Attali et al. 2014; Bejar 1983; Berenbon and McHugh 2023; El Masri et al. 2017; Holmes et al. 2018; Impara and Plake 1998; Rogausch et al. 2010; Sayın and Bulut 2024; Sayın and Gören 2023; Sydorenko 2011; Wonde et al. 2024) indicate that the correlation between empirically estimated item difficulties and item difficulties predicted by subject-matter experts ranges from ≤0.10 to 0.80. One interesting finding from these studies is that the precision of subject-matter experts’ judgement improves with feedback. Subject-matter experts become more accurate at predicting the empirically estimates difficulty the more they learn about the results of actual empirical studies and theoretical models that link item design features to cognitive processing demands. Thus, using theory and research helps human item writers to tailor their item writing more to the demands of the specifications of the blueprint for an ideal item pool and to avoid the use of non-functional distractors that often lead to a misfit of the chosen descriptive IRT model in field testing. However, providing item writers with this information requires conducting research on predictors of the difficulty of test items in the first place. Nonetheless, even with sufficient knowledge in these areas, human item writers still experience difficulties constructing and accurately predicting the difficulty of more difficult test items that exceed their own standing on the latent trait (cf. Hornke and Habon 1986; Sayın and Gören 2023). This line of research nicely illustrates that solely relying on the expertise of human item writers is no viable short-cut to theory-building and empirical studies. Providing detailed specifications of the item construction rationale, as well as theory and research findings on the cognitive processes involved in item solving, helps human item writers to improve the quality of the constructed test items.
Although it is possible to construct psychometrically sound test items using human item writers by combining theory, research, and subject-matter expertise, the item construction process cannot be scaled. This is due to the fact that test items are still constructed and reviewed on an item-to-item basis. This particular feature of human item writing (H-IG) is at odds with the demands for a large number of test items within a short amount of time (cf. Drasgow et al. 2006; Kosh et al. 2019; Wainer 2002; Zickar 2020). The costs associated with the continuous item construction process are often referred to as an explanation for why the anticipated cost saving from moving from fixed-item linear paper-pencil tests to CATs have not been realized thus far (cf. Luecht 2005).
Therefore, researchers and test developers have turned to alternative methods of item construction, which have been shown to be more scalable due to a higher degree of automation during the item construction process (cf. Falcão et al. 2023, 2024; Kosh et al. 2019). The approaches used to achieve this aim are referred to as rule-based automatic item generation (AIG: Arendasy and Sommer 2011; Embretson and Yang 2007; Irvine and Kyllonen 2002) and transformer-based automatic item generation (TB-AIG: Attali et al. 2022; von Davier 2019), respectively. Both methods can be summarized under the umbrella term automatic item generation. The distinction between AIG and TB-AIG resembles the theoretical distinction between the use of symbolic models, on the one hand, and connectionist models, on the other hand, in cognitive modeling research (cf. Klahr and MacWhinney 1997; van der Maas et al. 2021). It also has important consequences on how each method can be used in practice. For ease of reference, we will use the shortcut automatic item generation (AIG) to refer specifically to rule-based automatic item generation in all following sections.

3. Automatic Item Generation (AIG)

Automatic item generation (AIG: Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024; Embretson and Yang 2007; Irvine and Kyllonen 2002) is a collective term for various methods using pre-programmed algorithms to automatically construct large numbers of test items with predictable item parameters. This is done by systematically varying a set of radicals (Irvine 2002) and incidentals (Irvine 2002). The historical roots of AIG can be traced back to the 1970s in the United States (cf. Bejar 2002; Drasgow et al. 2006) and Europe (cf. Arendasy and Sommer 2011). In this article, we use the term automatic item generation in a more narrow, non-technological sense to exclusively refer to rule-based automatic item generation, which uses computer programs programmed from scratch to construct test items using symbolic models. AIG is often incorrectly portrayed as a unified approach (cf. Gierl et al. 2022a; von Davier 2018, 2019). In fact, there are three conceptually different approaches to automatic item generation. The difference concerns (1) the extent to which the selection of radicals is based on a theoretical model (weak- vs. strong theory), (2) the extent to which radicals are varied independently (schema-based vs. element-based), (3) the extent to which the item generation process is automated (human-made vs. semi-automated vs. automated), and (4) the inclusion of a quality control component (single- vs. dual-component item generation), which checks the factual correctness of the scoring key and also aims to reduce the likelihood of differential item functioning (DIF) and construct-irrelevant variance. Differences in these four characteristics between various approaches to AIG not only affect the validity evidence that can be advanced to support test score interpretations, but also the cost savings at various points in the item construction process (cf. Arendasy and Sommer 2011; Kosh et al. 2019).

3.1. Item Model Approach

In the item model approach (e.g., Bejar 2002; Choi and Zhang 2019; LaDuca et al. 1986), the construction of the item generator starts with the selection of one or more templates. These templates (cf. Mislevy and Riconscente 2006) are also referred to as item models (Bejar 2002), schemata (Arendasy and Sommer 2007), item shells (Glas and van der Linden 2003), or parent items (Roid and Haladyna 1982). For ease of communication, we will use the term item model since it has been used more commonly than other terms. Item models constitute the basic building blocks of the item generator and are usually hard coded into the item generator. Each item model consists of fixed and variable elements. The variable elements of an item model can either consist of a list of alternatives for incidentals or a list of alternatives for radicals (Gierl et al. 2012; Glas and van der Linden 2003). All items generated on the basis of a single item model are called instances (Graf et al. 2005) or siblings (Sinharay et al. 2003). Items constructed by only manipulating incidentals are referred to as item isomorphs (Bejar 2002) or item clones (Glas and van der Linden 2003). Item isomorphs are hypothesized to exhibit invariant (statistically identical) item parameters. By contrast, items constructed by manipulating radicals are called item variants (Bejar 2002). Item variants are hypothesized to differ in their item parameters and therefore exhibit more within-item model variation in their item parameters than item isomorphs (cf. Bejar 2002; Geerlings et al. 2013; Glas et al. 2016; Glas and van der Linden 2003; Sinharay et al. 2003; Sinharay and Johnson 2008). In the item model approach, radicals are not varied independently of each other. In fact, each item model is characterized by a specific combination of radicals and radical levels, which are hard coded in the item model.
Item models can be constructed in three different ways, which differ in their theoretical foundation. The first two means rely solely on the subject-matter expertise and intuition of the item-model writers and are therefore referred to as the weak theory approach (Drasgow et al. 2006). This is because even experienced subject-matter experts may not always correctly guess what makes a particular test item more or less difficult and what kind of misconceptions test-takers may hold that lead them to select specific distractors (cf. Attali et al. 2014; Bejar 1983; Berenbon and McHugh 2023; Holmes et al. 2018; Impara and Plake 1998; El Masri et al. 2017; Rogausch et al. 2010; Sayın and Bulut 2024; Sayın and Gören 2023; Sydorenko 2011; Wonde et al. 2024).
Within the weak theory approach, item models can be constructed on the basis of one or more operational test items, which are inspected by subject-matter experts to identify commonalities and differences in order to determine item design features that can be systematically combined and manipulated (cf. Bejar 2002; Fu et al. 2022). Alternatively, item models can be constructed from scratch by asking item model writers to define elements that can be systematically varied within each item model and to propose permissible values for these elements based on their subject-matter expertise (cf. Gierl and Lai 2012).
The third alternative is to construct item models based on pre-existing cognitive processing models for a specific item format (cf. Enright et al. 2002; Graf et al. 2005). The main idea here is to replace the subject-matter expertise and intuition of the item-model writers with an empirically validated cognitive processing model that specifies item design features hypothesized to be linked to cognitive processing demands. The main benefit of this method is that it builds on empirically validated theoretical models, which increase the construct validity of the resulting test items provided that these cognitive processes and the item design features linked to them have been shown to explain the item parameters of the test items (cf. Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024; Embretson and Yang 2007). For this reason, the third approach is often referred to as the strong theory approach (Drasgow et al. 2006).
The item model approach focuses heavily on the item model itself as its central building block. Psychometric modeling often boils down to examining the extent to which item isomorphs exhibit invariant item parameters (e.g., Fu et al. 2022) and examining the extent to which variability between item variants can be adequately modeled by means of expected item family functions (cf. Geerlings et al. 2013; Glas et al. 2016; Sinharay et al. 2003; Sinharay and Johnson 2008; Zorowitz et al. 2024). Validating the radicals by means of explanatory item response theory modeling is often not considered mandatory in the item model approach. While proponents of the strong theory approach acknowledge the benefits of explanatory item response theory modeling, proponents of the weak theory approach (e.g., Gierl et al. 2022b) often argue that calibrating the item models and examining their content validity by means of subject-matter expert ratings are sufficient to provide the necessary validity evidence in high-stakes settings.
The main reason for this focus is that the item model approach aims to reduce item construction costs not only by scaling the item construction process itself but also by shifting the focus of item calibration from the individual test items themselves to the item models. That is, instead of calibrating hundreds of test items, test developers merely need to calibrate the item models and model the variation of the item parameters within each item model by means of expected item family functions. Since each item model reflects a unique combination of radical levels, the number of item models is much lower than the number of item instances that can be generated by an item model. If calibrating item models instead of individual test items works, this can result in notable cost savings. Several simulation studies (e.g., Bejar et al. 2002; Colvin et al. 2016; Glas and van der Linden 2003; Someshwar 2024; Tian and Choi 2023) indicated that although using calibrated item families and expected item family functions instead of the actual item parameters of the item instances introduces a small bias in the person parameter estimate, this effect can be offset by administering a few additional test items. Thus, the approach of calibrating item models instead of individual test items seems feasible and can result in notable cost reduction, provided that the within-family variation is low and can be adequately modeled. Unfortunately, this is not always the case (e.g., Fu et al. 2022; Zorowitz et al. 2024).
The item model approach currently is the most popular approach. This has led some researchers (e.g., Circi et al. 2023; Gierl et al. 2022a; von Davier 2018) to equate the item model approach with the entire field of automatic item generation. Embretson and Yang (2007) explained the popularity of the item model approach with two of its features. On the one hand, the item model approach can be used to construct item models with a stronger or weaker theoretical foundation. On the other hand, item models are convenient to implement in programming terms. Thus far, the item model approach has been successfully used by the Educational Testing Service (ETS) in admission testing (cf. Bejar 2002), the construction of test items for the medical licensing exam in Canada (e.g., Gierl et al. 2012), high-stakes pharmacy exams (e.g., Leslie and Gierl 2023), test items measuring misconceptions in physics (e.g., Wancham et al. 2023), Turkish literature knowledge test items (e.g., Sayin and Gierl 2023), and to construct an open-source cognitive test (Ryoo et al. 2022), among various other examples.
Across different practical applications, the number of item models exhibiting a unique combination of radical levels varied from k = 6 (Ryoo et al. 2022) to k ≈ 60 (e.g., Gierl et al. 2012). Given that the number of item models constitutes an upper limit for the number of different item parameters, the number of psychometrically distinct items is limited by the number of item models implemented into the item generator. This affects the feasibility of constructing multiple item pools that can be used for assessment and practice-based training. In particular, constructing psychometrically matched test items may become more difficult with a lower number of item models. This is unfortunate because a combination of isomorphic and psychometrically matched test items is needed to simultaneously construct item pools for assessment and practice-based training to enhance fairness without compromising validity (cf. Section 1.4).
Some authors (e.g., Gierl et al. 2022a; Glas and van der Linden 2003; Yi et al. 2008) also discuss additional measurement and test security challenges associated with the use of item models. For instance, using item models in CATs requires content coding and content exposure control to ensure that isomorphic test items are not administered within the same test administration session. Doing so would lead to a local stochastic dependency of the item responses and a corresponding bias in the person-parameter estimate (cf. Glas and van der Linden 2003; Segall 2004). Furthermore, if an item model has been disclosed or leaked, all items generated by that item-model are compromised. This can lead to biased person-parameter estimates (cf. Yi et al. 2008). Thus, appropriate technical test- or item model security measures have to be taken to protect the item models from disclosure and organized item theft (for an overview, see Foster 2016).
A final concern relates to the number of test items that have to be discarded after the item calibration phase due to insufficient psychometric characteristics (e.g., misfit of the 2PL model or the 1PL Rasch model to the data). In most applications that examined the dimensionality of the automatically generated test items, the percentage of item loss ranged from 10 to 30 percent. In addition, item models constructed using weak theory often require substantial modifications after the expert reviews and calibration phase to reduce the likelihood that item instances include non-functional distractors or other item-writing flaws (e.g., Falcão et al. 2023, 2024; Gierl and Lai 2012; Lai et al. 2016). Similar to item loss, the number of required revision rounds must be taken into account when evaluating potential cost savings (cf. Kosh et al. 2019).

3.2. Cognitive Design System Approach

The cognitive design system approach (cf. Embretson 1998, 2002, 2016; Embretson and Yang 2007; Gorin 2006) extends the item model approaches based on strong theory. It heavily resorts to cognitive science research to provide a more detailed and precise definition of the latent cognitive ability trait in terms of the cognitive processes hypothesized to be involved in solving the test items. In this sense the, cognitive design system approach is similar to the evidence-centered design system approach proposed by Mislevy and associates (cf. Mislevy et al. 2003). Embretson and associates (cf. Embretson 2002, 2005; Embretson and Yang 2007; Gorin 2006) recommend starting the construction of an item generator with such a precise definition, which is referred to as a cognitive item model. Ideally, this cognitive item model has been implemented as a computational model. Computational models are symbolic and/or connectionist computer programs that not only specify the cognitive processes hypothesized to be involved in solving certain test items but also outline the link between these cognitive processing demands and the item design features that affect them. These item design features serve as radicals (Irvine 2002) in item generation. The behavior of computational models can be compared to human problem solvers as a feasibility check on whether the theoretical model may constitute a possible data-generating mechanism (cf. Farrell and Lewandowsky 2010; Fried 2020; Guest and Martin 2021; Klahr and MacWhinney 1997; Smaldino 2020; van der Maas et al. 2021). The main benefit of implementing verbal theories into a computational model is that this process forces the scientists to reason more precisely about the latent traits that they intend to measure (cf. Farrell and Lewandowsky 2010; Fried 2020; Guest and Martin 2021; Smaldino 2020). Thus, test developers using computational models ultimately operate with a more precise definition of the latent trait to be measured.
The drawback of using computational models to deduce radicals is that such models may not exist in all cognitive ability domains. In such cases, Embretson and associates suggest examining test-takers’ solution strategies using eye-movement analyses, verbal protocols, etc., to deduce hypotheses on cognitive processes involved in problem solving and how these cognitive processes are linked to item design features (for an overview, see Embretson 2016). Next, these hypotheses should be tested by means of explanatory item response theory modeling of existing psychometric test data (e.g., Baldonado et al. 2015; Daniel and Embretson 2010; Draheim et al. 2018; Embretson 1998, 2002, 2023; Embretson and Daniel 2008; Embretson and Gorin 2001; Embretson and Kingston 2018; Gorin 2005; Gorin and Embretson 2006; Ivie and Embretson 2010; Kaller et al. 2011; O’Reilly et al. 2018; Svetina et al. 2011). Evaluating the status of radicals and incidentals by means of explanatory item response theory modeling is also highly recommended in cases in which computational cognitive item models are available. The main argument is that the radicals and corresponding cognitive processes should be able to account for a large percentage of the variance in the item parameters if the cognitive item model constitutes a possible data-generating mechanism (cf. Embretson 1998, 2016; Embretson and Yang 2007). The intention behind combining computational cognitive modeling with explanatory item response theory modeling is to maximize the construct-related variance in the item parameters by systematically manipulating item design features linked to cognitive processing demands, which are central to the definition of the latent cognitive ability trait. However, Embretson and associates note that radicals (Irvine 2002) are often confounded in existing psychometric tests. This complicates the interpretation of the research findings due to multi-collinearity (cf. Fischer 1995). Therefore, the next step consists of the construction of an automatic item generator, in which radicals—and combinations thereof—are more systematically varied, and the use of explanatory item response theory modeling (cf. De Boeck and Wilson 2004) to examine the extent to which the item parameters of the automatically generated test items can be predicted by the radicals. In most studies using the cognitive design system approach, the correlation between the empirically estimated item difficulty parameters and those predicted on the basis of the radicals (Irvine 2002) varied from R = 0.70 to R = 0.90. Thus, the radicals accounted for 49 to 80 percent of the variance in the item difficulty parameters of the respective descriptive item response theory model (e.g., 1PL Rasch model or 2PL model).
In a nutshell, the cognitive design system approach shifts the expenses in terms of financial, time, and human resources to the initial test construction phases. This strategy pays off if it increases the psychometric quality of the automatically generated items in terms of their fit to the chosen descriptive item response theory model (e.g., 1PL Rasch model, 2PL model, etc.) and a lack of differential item functioning (DIF), and in terms of the variance in the item parameters explained by the radicals (cf. Kosh et al. 2019). Several simulation studies (e.g., Doebler 2012; Embretson 1999; Freund et al. 2008; Mislevy et al. 1993; Matteucci et al. 2012; Someshwar 2024) indicated that if the variance in the item parameters accounted for by the radicals is sufficiently high (R2 ≥ 0.80), the effect of using predicted item parameters instead of empirically calibrated item parameters on the precision and bias of the person parameter estimate becomes negligible and can be easily offset by the administration of a few additional test items. Interestingly, Someshwar (2024) recently showed that using item parameters predicted by an explanatory item response theory model such as the Linear Logistic Test Model (LLTM; Fischer 1995) is superior to using expected item family functions in terms of the precision and bias in the person parameter estimates. This indicates that the strong focus on theory in the cognitive design system approach not only has the potential to provide additional validity evidence (cf. AERA et al. 2018) but it can also surpass the potential cost savings of the item model approach. In addition, using explanatory item response theory modeling has also been shown to help explain the intended or unintended variance of the item parameters within item models (cf. Cho et al. 2014; Geerlings et al. 2013; Glas et al. 2016; Sinharay et al. 2003; Sinharay and Johnson 2008). This valuable information can be used to either refine the item models or adjust the item parameters of the automatically generated items accordingly.
Thus far, the cognitive design system approach has been successfully used to automatically generate figural matrices (e.g., Embretson 1998, 2002; Freund et al. 2008; Primi 2002, 2014), analogies (e.g., Blum and Holling 2018; Bejar et al. 2012), number series (e.g., Sun et al. 2019), propositional reasoning items (e.g., Gühne et al. 2020), spatial ability test items (e.g., Fehringer 2020; Ivie and Embretson 2010; Shi et al. 2023), mathematical ability test items (e.g., Daniel and Embretson 2010; Embretson and Kingston 2018; Enright et al. 2002; Holling et al. 2009), perceptual speed test items (e.g., Doebler and Holling 2016), and to semi-automatically generate reading comprehension test items (e.g., Förster and Kuhn 2023; Gorin 2005). Most researchers chose to implement radical combinations by constructing item models due to the ease of programming implementation. The number of item models ranged from 30 to 60 across studies. Since researchers chose to use item models, considerations regarding test security for the cognitive design system approach are almost identical to the ones outlined for the item model approach (cf. Section 3.1, seventh paragraph). Thus, the extent to which the cognitive design system approach can deal with the increased test item demands depends on the number of item models, their distribution along the latent ability continuum, and whether it is possible to generate psychometrically matched item models differing in their radical combinations along the entire latent trait continuum. Like the item model approach, researchers using the cognitive design system approach often reported an item loss of 10 to 30 percent after the item calibration phase due to either a misfit to the 2PL model or differential item functioning. This item loss also has to be taken into account when evaluating cost savings (cf. Kosh et al. 2019).

3.3. Automatic Min-Max Approach

To address the issue of item loss after the item calibration phase, Arendasy and associates (cf. Arendasy 2004; Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024) developed the automatic min-max approach. The automatic min-max approach can be seen as a further development of the cognitive design system approach. Like Embretson and associates (cf. Embretson 1998, 2016; Embretson and Yang 2007; Gorin 2006), the automatic min-max approach combines cognitive science research, individual differences research, and explanatory item response theory modeling in constructing and evaluating an item generator. Like the cognitive design system approach, it is heavily rooted in modern validity theory and cognitive science (cf. Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024).
The automatic min-max approach resorts to individual differences research to provide an initial definition of the latent cognitive ability trait and its location within current models of human intelligence. Next, the cognitive processes involved in solving the test items are specified in more detail by resorting to cognitive processing models that ideally have been implemented as computational models. Like the cognitive design system approach, the automatic min-max approach also emphasises the theoretical and practical value of computational models in constructing a cognitve model that serves as the basis for the selection of the radicals (Irvine 2002). However, it deviates from the cognitive design system approach by proposing that researchers should not solely resort to a single computational model. Instead, Arendasy and associates (cf. Arendasy 2004; Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024) recommended using and comparing multiple computational models. This is because most existing computational models have been primarily tested on a particular item set or a particular item format (for a similar argument, see Yang and Kunda 2023; Yang et al. 2022). Thus, using a single computational model makes it more difficult to separate cognitive processes and corresponding item design features that are central to the core of the definition of the latent trait from those that are more specific to either a certain item format or a certain set of test items (cf. Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024). Ideally, hypotheses deduced in this manner should be empirically tested by means of psychometric experiments. To illustrate this point, we will give two examples. Arendasy and Sommer (2012b) examined item design features that have been hypothesized to moderate the size of gender differences in figural matrices despite not being central to the definition of inductive reasoning. Using several established and experimentally designed figural matrices tests in addition to number series, the authors demonstrated that some non-central item design features used in figural matrices increase the likelihood of gender-related differential item functioning. Furthermore, performance differences in psychometric tests using these item design features did not generalize to a higher-order trait reflecting individual differences in inductive reasoning. Based on these results, the authors concluded that these item design features are not central to the definition of inductive reasoning. Similarly, Arendasy and associates (e.g., Arendasy 2000; Arendasy and Sommer 2010; Arendasy et al. 2010; Arendasy et al. 2010) tested the hypothesis that gender differences in complex three-dimensional mental rotation are in part due to differences in the ability to utilize depth cues using psychometric experiments involving 3D shutter glasses. Unlike the figural matrices reasoning example, differences in the use of depth cues are central to the definition of complex three-dimensional mental rotation according to several computational models. However, one may need to constrain the values of this radical to avoid differential item functioning across gender without compromising construct representation (cf. Arendasy and Sommer 2010). These two examples nicely illustrate that the selection of radicals (Irvine 2002) and the choice of an item format should be well founded empirically and theoretically.
Once the radicals have been selected, the cognitive model is applied to the chosen item format. The manipulation of the radicals in this cognitive item model focuses on the systematic manipulation of radicals (Irvine 2002) that affect cognitive processes central to the definition of the latent trait. By contrast, item design features linked to non-central cognitive processing demands are constrained at a certain value to ensure that they neither induce differential item functioning nor affect the item parameters. This is done to maximize the construct-related variance in the item parameters.
If no prior computational cognitive (item) model is available, the automatic min-max approach proposes using studies on test-takers’ solution strategies in combination with explanatory item response theory modeling to deduce hypotheses on potential radicals (Irvine 2002) and to gradually formulate a cognitive (item) model (cf. Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024). This recommendation is identical to the cognitive design system approach. We briefly outline how this can be done using two examples. Sommer et al. (2009) combined cognitive task analysis and explanatory item response theory modeling to deduce a set of radicals (Irvine 2002) from the test items of an existing tachistoscopic traffic perception test to enable the construction of a new item pool. The hypotheses deduced in this earlier stage were subsequently tested using data obtained with the new semi-automatically constructed test items that were analyzed with the Linear Logistic Test Model (LLTM; Fischer 1995). This approach was chosen because, at that time, no computational models of perceptual speed were available for tachistoscopic tests. Similarly, Arendasy et al. (2024) resorted to computational models for the Tower of London (ToL) and supplemented prior research on cognitive processes involved in ToL problem solving with studies using retrospective verbal reports obtained during scheduling problem solving to deduce a set of radicals (Irvine 2002) that can be used to automatically generate items measuring planning ability using scheduling problems as an item format. Thus, the utilization of the automatic min-max approach is not limited to a strong theory approach. However, the authors emphasize that the strength of validity arguments that can be formed differs between a strong and a weak theory approach.
Similar to the cognitive design system approach (cf. Embretson 1998, 2002, 2016; Embretson and Yang 2007; Gorin 2006), validating the radicals by means of explanatory item response theory models (for an overview, see De Boeck and Wilson 2004) is mandatory in the automatic min-max approach. This is because the radicals constitute the basic building blocks of the item generator in the automatic min-max approach (element-based approach). Thus, instead of implementing item models that exhibit a fixed combination of radicals and radical levels, Arendasy and associates suggested implementing radicals and algorithms on how to combine them. This allows test developers to manipulate the radicals (Irvine 2002) independently of each other. Although the element-based approach is more difficult to implement from a programming perspective, it enhances test security by complicating item theft (cf. Arendasy and Sommer 2011). Despite this advantage, the authors also note that, in some cases, a schema-based approach may be more appropriate. This is the case in algebra word problem solving (cf. Arendasy and Sommer 2007; Arendasy et al. 2006). The main reason for a schema-based approach in the case of algebra word problems is that it mimics the cognitive representation of algebra word problems in the minds of the test-takers.
Up to this point, the automatic min-max approach only differs gradually from the cognitive design system approach. The most important difference between these two approaches resides in the use of a separate quality control component. Arendasy and associates (cf. Arendasy 2004; Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024) argued that researchers should explicitly outline cognitive processes involved in solving the test items that are hypothesized to be construct-unrelated but may be triggered by certain item design features. These item design features are referred to as functional constraints (Greeno et al. 1993). The quality control component aims to minimize construct-unrelated variance in the item parameters and to reduce the likelihood of differential item functioning by suppressing the use of functional constraints during item generation. In order to prevent the risk of construct-underrepresentation, functional constraints need to be empirically validated (for examples, see Arendasy 2004; Arendasy and Sommer 2005, 2010, 2012b, 2013b; Arendasy et al. 2012). Arendasy and associates showed that using a quality control component minimizes item loss due to insufficient psychometric characteristics. Across several cognitive ability domains, the authors report an item loss of ≤1 percent despite the use of more restrictive IRT models such as the 1PL Rasch model (for a replication, see Hines 2017). This not only translated into cost savings (cf. Kosh et al. 2019) but also effectively leads to a more precise definition of the latent trait (cf. Farrell and Lewandowsky 2010; Fried 2020; Guest and Martin 2021; Smaldino 2020).
Once the radicals (Irvine 2002) and functional constraints (Greeno et al. 1993) have been implemented into the item generator, large numbers of test items can be automatically generated. Item generators constructed using the automatic min-max approach have been used to generate number series (e.g., Arendasy and Sommer 2012a), figural matrices and analogies (e.g., Arendasy and Sommer 2005, 2012b; Hines 2017), syllogisms (e.g., Arendasy et al. 2020), algebra word problems (e.g., Arendasy and Sommer 2007; Arendasy et al. 2006), arithmetic fluency items (e.g., Arendasy et al. 2007), computational estimation items (e.g., Arendasy et al. 2020), complex three-dimensional mental rotation items (e.g., Arendasy and Sommer 2010; Arendasy et al. 2010; Arendasy et al. 2011), visualization items (e.g., Arendasy et al. 2020), multi-lingual verbal fluency items (e.g., Arendasy et al. 2012), synonym items (e.g., Arendasy et al. 2020), and planning ability items (e.g., Arendasy et al. 2024). In addition, the automatic min-max approach has also been used to semi-automatically construct reading comprehension items (e.g., Arendasy et al. 2020) and perceptual speed items (e.g., Sommer et al. 2009). In all cases, the Linear Logistic Test Model (LLTM; Fischer 1995) has been shown to fit the data obtained with a roughly representative set of k = 120 to k = 320 test items exhibiting unique radical combinations reasonably well. The correlation between the empirically estimated item parameters and the ones predicted on the basis of the LLTM basic parameter estimates of the radicals (Irvine 2002) varied from R = 0.89 to R = 0.96 (for an overview, see Arendasy et al. 2020; Hines 2017). Thus, cognitive processes linked to the radicals account for 72 to 92 percent of the variance in the item- and person parameter estimates. This finding not only provides empirical evidence on the construct validity of the automatically generated test items but it also indicates that cost saving may be possible.
Several simulation studies (e.g., Doebler 2012; Embretson 1999; Freund et al. 2008; Mislevy et al. 1993; Matteucci et al. 2012; Someshwar 2024) indicated that the effect of using predicted item parameters instead of calibrated item parameters on the precision and bias of the person parameter estimate becomes negligible and can be easily offset by the administration of a few additional test items if the predicted and calibrated item parameters are highly correlated (R ≥ 0.90). Based on these results, using predicted item parameters instead of calibrated item parameters would be possible. Doing so would lead to a drastic reduction in the item calibration costs to develop and update multiple item pools that can be used in CATs and practice-based training. The number of items that can be generated with the item generators also seems to be sufficiently large. Furthermore, there is evidence that the item generators are able to construct psychometrically matched test items along large parts of the latent trait continuum (cf. Arendasy and Sommer 2013a). The high correlation between empirically estimated item parameters and those predicted by the LLTM-basic parameter estimates also enables the automatic generation of test items on-the-fly during the administration of a CAT. This has been referred to as either adaptogenic testing (Arendasy 2004) or on-the-fly item generation (Bejar et al. 2002). To our knowledge, this option has only been realized once as a beta version (Arendasy 2004). In addition to technical considerations (e.g., time needed for automatic item generation, implementation of an item generator, etc.), we currently still lack empirical data and theoretical models that would allow us to estimate whether such an approach would be accepted by test-takers and decision-makers (cf. Arendasy 2004).
Based on the research findings outlined above, the automatic min-max approach seems to largely meet the requirements to keep up with the increased item construction demands outlined in Section 1.4. However, the extent to which this is the case varies across item generators. In addition, as discussed above, not all item generators are able to generate test items fully automatically. For instance, the item generator used to construct reading comprehension test items is only semi-automatic (cf. Arendasy et al. 2020). Human intervention, and in some cases a collaboration between human item writers and transformer networks (cf. Section 4.1), was needed to iteratively construct the reading passages from scratch based on content- and linguistic specifications provided by the item generator. Due to its semi-automatic nature, the costs are naturally higher than for item generators in the other cognitive ability domains that are fully automated.

4. Transformer-Based Automatic Item Generation (TB-AIG)

More recently, several researchers (e.g., Attali et al. 2022; Bulut et al. 2024; Hao et al. 2024; Lee et al. 2023b; von Davier 2018, 2019) proposed including large language models in the toolbox of test developers as a subvariant of AIG. The rapid progress in natural language processing due to the introduction of transformer networks (cf. Vaswani et al. 2017) has enabled large language models to handle a variety of natural language processing tasks (e.g., translation, text summarization, question answering, and writing text) at a level similar to humans. This sparked the interest of professionals working in computer science and psychometrics, who started examining the feasibility of transformer networks in item generation (for reviews: Kurdi et al. 2020; Song et al. 2025; Tan et al. 2024). Reviews on TB-AIG noted that most studies examined the feasibility of using transformer networks in the verbal ability domain as well as the domains of medicine and computer science. Furthermore, most studies merely examined whether TB-AIG items are rated similarly to items constructed by human item writers during the item review process. Few studies included results on the calibration of the TB-AIG items. Thus, at present, there is a lack of data that allows us to examine the extent to which TB-AIG can contribute to meeting the increased item construction demands. What we know thus far is that TB-AIG is feasible and that it reduces the time needed to construct the test items.
However, cost savings during the item construction process are smaller compared to AIG. Advocates of TB-AIG (e.g., Attali et al. 2022; Bulut et al. 2024; Hao et al. 2024; von Davier 2019) recommend keeping human subject-matter experts in the loop during the entire item construction process. This implies that item generation has to be semi-automatic. The reason for this recommendation has to do with the perils and pitfalls arising from the architecture of transformer networks and how they are trained prior to their use. In the next section, we will provide a brief description of transformer networks. Next, we will discuss concerns regarding the use of transformer networks in item construction without human oversight (for an overview, see Hao et al. 2024) before discussing how transformer networks have been used in psychological and educational assessment.

4.1. Introduction to Transformer Networks

Transformer networks are the successors of recurrent neural networks (for a brief overview, see Lee et al. 2023b; von Davier 2019). Their main benefit is that they can process information in parallel and can more accurately capture distant semantic relations between words and text. The architecture of transformer networks consists of a combination of encoder- and decoder blocks (cf. Vaswani et al. 2017). The encoder extracts features from the input, while the decoder uses information received from the encoder in conjunction with its learning database to produce an output.
Both encoder and decoder consist of multiple stacks of units organized into layers. Each encoder unit in the first layer receives the input, which has been converted into tokens (numerical representations of words, etc.) and enriched with position coding. The encoder unit converts the input into a vector, which reflects the meaning of the token in a multidimensional space. These vectors are commonly referred to as word embeddings. To ensure that long-distance relations between tokens are adequately captured, the transformer network uses a self-attention mechanism, which pays more attention to tokens that are jointly more relevant to the meaning of the input. The word embedding vectors go through position-wise feed-forward layers for further transformation by being handed over to the units in subsequent layers. This part of the transformer architecture is similar to neural networks. Although the linear transformations from one layer to another are identical, each layer of the feed-forward network uses different parameters that have been optimized in the pre-training phase (cf. Brown et al. 2020; Vaswani et al. 2017; Yang et al. 2024). In a final step, the optimized embedding is handed over to the decoder.
Similar to the encoder, the decoder also consists of multiple layers of decoder units and self-attention mechanisms. The encoder and decoder are connected to each other via a cross-attention mechanism. Based on the input of the encoder, the decoder predicts the next token as an output on the basis of the probability of that token given the input from the encoder and the tokens in the text corpus it has been trained on. The predicted token is included into the sequence of tokens to create an updated vector, which conditions the next token to be generated by the decoder until the task is finished.
Based on this brief description of transformer networks, transformer networks can be seen as probabilistic models that produce output based on the conditional probability of a token given the input, the training data, and all previously generated tokens that surround it (cf. Sobieszek and Price 2022).
The various transformer networks differ in their focus on the encoder side, the decoder side, or the encoder and decoder side. Decoder models and encoder–decoder models have dominated in recent years (Yang et al. 2024). These two types of transformer networks differ regarding how their model parameters have been optimized. Encoder-decoder models use an unsupervised learning paradigm, in which randomly selected parts of the training data are masked, and the transformer network has to predict the mask token. By contrast, the parameters of decoder transformer networks are optimized by means of an unsupervised learning paradigm, in which they have to predict the next token in the training data given all previously encountered tokens (cf. Yang et al. 2024).
Both unsupervised learning paradigms enable the transformer network to learn the conditional probabilities of tokens from large training data sets, which often consist of billions of text data points. However, the specific learning paradigm affects the kind of transfer tasks on which the models can be expected to excel (e.g., Yang et al. 2024). Prominent examples of decoder models include the Generative Pre-trained Transformer 3 (GPT-3: Brown et al. 2020) and its successor GPT-4 (OpenAI 2023) and LLaMA (Touvron et al. 2023). Examples of encoder–decoder models include the Bidirectional Encoder Representations from Transformers (BERT: Devlin et al. 2018) and the Text-to-text Transfer Transformer (T5: Raffel et al. 2020). The four examples above have been chosen because they were frequently used in item generation studies (cf. Kurdi et al. 2020; Song et al. 2025; Tan et al. 2024). Currently available transformer networks also differ regarding the number of parameters that have been optimized, the tokenizers used, and the scope, quality, diversity, distribution, and specific content and language of the training and validation data (Yang et al. 2024). All of these characteristics are known to affect the performance on subsequent transfer tasks (cf. Yang et al. 2024; Zha et al. 2025). Thus, researchers should choose a transformer network with those characteristics in mind. In practical applications, this often comes down to the need to use different transformer networks for specific sub-tasks during item generation.
Once the model parameters have been optimized, the transformer network can be applied to various transfer tasks by either fine-tuning its parameters through training on that specific transfer task or by in-context learning via prompts.
In the former case, the pre-trained transformer network receives additional training on a specific transfer task (e.g., solving math word problems, text classification, etc.) with feedback (supervised learning). This leads to an update of the parameters of the transformer model. The benefit of this approach is that it improves performance on the transfer task. However, fine-tuning requires large annotated high-quality training examples and sufficient computing power—especially when fine-tuning larger transformer models—to update its parameters (Brown et al. 2020).
In-context learning via prompts, on the other hand, does not alter the parameters of the pre-trained transformer network. Instead, the transformer network is presented with a precise task description, possibly in addition to a few selected examples, and explanations (e.g., Ahn and Yin 2025; Al Faraby et al. 2024; Bozkurt and Sharma 2023; Heston and Khun 2023; Lee et al. 2023c; Li and Zhang 2024; Liu et al. 2023a; Reynolds and McDonell 2021; Sahoo et al. 2024; Schulhoff et al. 2024). This approach is referred to as k-shot learning or few-shot learning, where k is the number of examples provided to the transformer network. The main finding from studies examining prompting techniques is that the quality, structure, and clarity of the prompts, as well as the quality, distribution, and representativeness of the examples, affect the quality of the output. Furthermore, even slight paraphrases of a prompt may sometimes yield qualitatively different output.
This effect is due to how transformer networks work. They generate tokens based on the conditional probability given the tokens in the input (here, the prompts of the user) and the tokens in their training data. Thus, different prompts naturally lead to different conditional probabilities for the output. This also explains the finding that an identical prompt may yield different output depending on the specific transformer network used (for an illustrative example, see Säuberli and Clematide 2024). This is because transformer networks may not only differ in the number of parameters and how they have been optimized; they may also differ in the scope, quality, diversity, distribution, and specific content of their training data.
Furthermore, the research cited above indicates that prompts often need to be revised and optimized iteratively. This process is referred to as either prompt engineering or iterative prompting. Prompt engineering can be facilitated by implementing a collaborative platform that enables human item writers to iteratively test, revise, and refine their prompts based on the output generated by the transformer network to ensure that its final output meets their expectations (cf. Reza et al. 2024).
In a nutshell, transformer networks learn subject-matter domain knowledge and linguistic knowledge by inferring regularities and probabilities between tokens (e.g., words, numbers, etc.) in their training data in terms of conditional probabilities given all tokens that surround them (Sobieszek and Price 2022).

4.2. Perils of Using Transformer-Based Automatic Item Generation in a Fully Automated Manner

Proponents of TB-AIG propose a semi-automatic item generation approach that keeps human subject-matter experts in the loop during the entire test item construction process (e.g., Attali et al. 2022; Bulut et al. 2024; Hao et al. 2024; von Davier 2018, 2019). The main reasons for this proposal are linked to how transformer networks work.
The first concern regarding the use of TB-AIG without human oversight concerns the factual correctness of the output (cf. Bulut et al. 2024; Hao et al. 2024). In constructing cognitive ability test items, it is important that the response(s) scored as the correct answer(s) are indeed factually correct. Research indicates that transformer networks sometimes have difficulties identifying the factually correct answer(s). Two recent meta-analyses (Liu et al. 2024; Waldock et al. 2024) examined whether transformer networks such as GPT and BERT are able to solve United States Medical Licensing Exam (USMLE) items. Liu et al. (2024) summarized k = 45 studies using GPT to solve USMLE items conducted between January 2022 and May 2024. The authors report an overall accuracy rate of 81% [CI: 78%; 84%] for GPT-4 and 58% [CI: 53%; 63%] for GPT-3.5. Waldock et al. (2024) conducted a meta-analysis of k = 32 studies published until September 2023. They report an overall accuracy rate of 51% [CI: 46%; 56%] for GPT-2 and GPT-3 and 64% [CI: 60%; 67%] for ChatCPT (GPT-3.5). Based on these results, newer versions of GPT outperform their precursors. This finding is attributable both to an increase in the number of model parameters and the size of the training data and to the use of previous prompt data in training the successor transformer networks. Furthermore, these meta-analyses indicate that the accuracy of picking the correct answer decreases as the test items become more difficult. However, fine-tuned prompting generally improves accuracy across all difficulty levels. In addition, transformer models have been shown to be inconsistent in their selections of the correct answer alternative when prompted to do so multiple times using the same prompt. This problem was more apparent in older versions than in newer versions of the same transformer network. However, even the more recent version of the transformer networks chose the correct answer consistently in ≤80% of the cases. The above-cited meta-analytic findings were replicated in subsequent studies on USMLE items and in studies using items from other medical or nursing licensing exams (e.g., Bhayana et al. 2023; Funk et al. 2024; Riedel et al. 2023; Su et al. 2024). Similar results were obtained for measures of verbal and numerical reasoning commonly used in educational admission testing in the United States (cf. Abu-Haifa et al. 2024; Hickman et al. 2024), math problems (for an overview, see Lu et al. 2022), and natural science and STEM problems (e.g., Chan et al. 2025; Schulze Balhorn et al. 2024). These results are not surprising because transformer networks have been trained to produce probable output given the constraints of the prompts and the training data. Despite being probable, plausible, and in some cases even convincing, the output does not necessarily have to be factually correct (cf. Sobieszek and Price 2022). This implies that human subject-matter experts need to examine the output and fact-check the content of the test items. This conclusion is consistent with recent research evidence indicating that TB-AIG-generated items often exhibit more than one possible correct answer despite using elaborated state-of-the-art prompts. By contrast, items constructed by means of AIG do not exhibit this problem (e.g., Chung and Kim 2024; Emekli and Karahan 2025). However, the two studies cited above also indicate that human item writers sometimes unintentionally construct items with more than one correct answer if the item exceeds their own standing on the latent trait. This may indicate that, in addition to content reviews by human subject-matter experts, a more rule-based approach may be required to ensure that all constructed test items align with their scoring key.
Copyright and ethical issues are an additional concern (cf. Bulut et al. 2024; Hao et al. 2024). In Section 4.1 (third paragraph), we outlined that the parameters of transformer networks have to be optimized in a pre-training phase. This is done by means of different unsupervised learning paradigms (e.g., masked token prediction or next token prediction) using billions of text data scraped from the internet, various corpora, and private sources. The precise content of these data is often not outlined in detail. This makes it difficult to determine whether the training data contained copyrighted data and the extent to which the output is based on it. Various studies (e.g., Lee et al. 2022; McCoy et al. 2023) indicated that the output of transformer models may indeed contain verbatim and/or paraphrased plagiarism of their training data. Although this does not happen often, it cannot be ruled out. Thus, test developers need to check whether the content of the TB-AIG test items violates copyright claims.
The third concern is also linked to the training data of the transformer networks. Data used to optimize the parameters of the transformer network may include stereotypes about different groups of people. This may lead to the presence of a socio-demographic bias in the output of the transformer network. Various studies (e.g., Balestri 2025; Kamruzzaman et al. 2024; Luca et al. 2025; Ranjan et al. 2024; Thakur 2023) indicated that both closed (commercial) and open-source transformer networks have been trained with data that contain stereotypes and biases against selected groups. Thus, fairness and bias reviews by human subject-matter experts are needed to ensure that the final test items are not biased. However, such fairness reviews may require a sufficiently large and diverse team of subject-matter experts. Belzak et al. (2023) recently showed that both TB-AIG test items and human-written items may contain a bias leading to DIF despite extensive fairness reviews by subject-matter experts. One means to deal with this issue is to use explanatory item response theory modeling (for an overview, see De Boeck and Wilson 2004) to examine whether the source of DIF can be traced down to certain radicals or other item design features (e.g., Arendasy and Sommer 2013b; Rajeb et al. 2024; Yu 1994) in addition to fairness reviews. Results from such studies can be used both in subsequent item generation and in the continuous training of subject-matter experts.
Taken together, these findings indicate that TB-AIG items need to be reviewed and refined by subject-matter experts prior to item calibration. Thus, the advice to keep human subject-matter experts in the loop during the entire item construction process is warranted (cf. Bulut et al. 2024; Hao et al. 2024; Sobieszek and Price 2022). As a consequence, TB-AIG is only feasible in a semi-automatic manner.

4.3. Studies Using Transformer-Based Automatic Item Generation in the Cognitive Ability Domain

Despite these limitations, TB-AIG has the potential to considerably reduce the amount of time and effort in item construction. TB-AIG seems particularly useful in cognitive ability domains, where the aforementioned AIG approaches have so far been difficult to implement in a fully automatic manner. Examples include measures of vocabulary and reading comprehension as well as knowledge tests. Several recent reviews (cf. Kurdi et al. 2020; Song et al. 2025; Tan et al. 2024) indicated that most studies using TB-AIG were indeed conducted in these cognitive ability domains. Furthermore, the reviews indicated that in most articles the TB-AIG items were solely validated by means of subject-matter expert ratings. Examinations of the psychometric characteristics of TB-AIG test items have been rare. If psychometric characteristics were reported at all, researchers often utilized classical test theory instead of item response theory modeling. Thus, at present, there is limited data on the extent to which TB-AIG can contribute to meeting the increased item demands outlined in Section 1.4. In addition, some reviews (cf. Song et al. 2025; Tan et al. 2024) also noted a trend toward the use of a weak theory approach. We agree with all these points. In addition, we noticed another important trend. While some authors incorporate ideas of the item model approach (Bejar et al. 2002) into their prompts, others use ideas inspired by an element-based approach (Arendasy and Sommer 2011). In the next sections, we will outline research conducted within both approaches to TB-AIG in the domains of verbal ability and medical clinical reasoning.

4.3.1. Item Model-Based TB-AIG

Proponents of an item model-based approach to TB-AIG argue that item models (Bejar 2002) facilitate TB-AIG item construction. The various studies differ in how item models have been specified and the role of transformer networks in the item generation process.
The most commonly used means to incorporate item models into TB-AIG is using a test item from an operational test as an item model and requesting the transformer network to construct isomorphic test items. This approach is reminiscent of early work within the item model approach. For instance, Shin and Gierl (2022) selected k = 5 passage-based reading comprehension items from an operationally used test in South Korea. Each item differed regarding the kind of inference required to solve the question following a short passage. The authors used ChatGPT with one-shot prompting to construct several isomorphic items for these k = 5 item models. The TB-AIG-constructed items were rated similar to the original items in terms of the perceived naturalness and attractiveness of the passage and in terms of the perceived attractiveness of the distractors by a panel of pre-service and in-service L2 English teachers. Similarly, Bezirhan and von Davier (2023) examined whether GPT-3 can be used to construct reading comprehension passages for informational and fictional text similar to the ones used in large-scale international studies. To this end, the authors used 1-shot prompting with or without additional specifications to construct isomorphic passages. They evaluated the similarity of the TB-AIG-constructed passages and the original ones in terms of their coherence, appropriateness, and readability for the target audience. The results indicated that providing additional information led to more isomorphic TB-AIG passages. However, the authors also noted that minor revisions and fact-checks were necessary during the passage generation process. Sayin and Gierl (2024) used a similar approach to enable GPT-3.5 to construct sentence comprehension items. This item type requires test-takers to mark one of five sentences that has a different topic than the other four sentences. The authors first designed a prompt for the four thematically matched sentences and requested the generation of k = 20 item isomorphs. The same procedure was used for the sentence with a different topic. The authors asked a panel of N = 3 subject-matter experts to evaluate the usability of TB-AIG items on a four-point single-item scale. All of the evaluated items were rated as either usable or requiring only minor revision. Sayin and Gierl (2024) administered k = 6 selected items in an operational reading comprehension test. Each of the k = 6 items was located at a fixed slot in the six booklets. They calculated classic item difficulty and item discrimination. Although item difficulty was in a narrow range, item discrimination for the test items hypothesized to have statistically identical item parameters varied considerably.
The second option to incorporate the item model approach into TB-AIG is to examine item design features of an existing operational test and to use cluster analysis to categorize existing items into clusters that subsequently serve as item models. This approach was used by Shin and Gierl (2022) to construct fill-in-the-blank reading comprehension items. The authors analyzed k = 2654 items from an operationally used test in South Korea in terms of their syntactic and semantic features. These data were used to distill k = 3 item models using hierarchical cluster analysis. Next, they analyzed k = 100 new human-made passages and categorized them into one of the three item model clusters. This information was used during prompting to determine the part of the text that should be blanked out and to construct the correct answer in addition to a set of distractors. The items were compared with the original human-made test items in terms of their syntactic and semantic item design features.
Kıyak and associates (Kıyak and Kononowicz 2025; Kıyak et al. 2024, 2025) suggested iteratively constructing item models based on a few sample items in the domain of clinical reasoning in medicine. The authors used GPT-Builder to construct a ChatGPT interface, which guides subject-matter experts through the process of constructing an item model commonly used in item model-based AIG. The process mirrors the one described by Gierl and Lai (2012). First, the expert enters a sample item. ChatGPT asks the subject-matter expert which part of the sample item should be varied and how the information in the variable parts relates to the distractors and the correct answer. Next, ChatGPT provides possible values for each variable element. The subject-matter expert can provide feedback on the factual correctness, which prompts ChatGPT to revise its original suggestions. Once the item model has been constructed, Kryak and associates propose implementing the newly constructed item model into an independent AIG software program, which can be used for item generation. The authors constructed various item models in this manner and generated k = 104 clinical reasoning test items. They randomly selected a set of k = 35 items and asked N = 18 experienced medical doctors to indicate the correct answer. Each item was reviewed by three medical doctors. Kıyak et al. (2025) report that 92.6 percent of all expert decisions corresponded with the scoring key proposed by ChatGPT.
As outlined above, most studies using item model TB-AIG utilized a weak theory approach and often merely resorted to judgements from subject-matter experts to validate the test items. Thus, important information on the dimensionality and psychometric characteristics of the TB-AIG items is missing. Furthermore, there seems to be a focus on the construction of item isomorphs (Bejar 2002) rather than item variants (Bejar 2002). This is in parts at odds with the demands of modern assessment practices, which require the construction of item variants and item isomorphs with either predictable or calibrated (=known) item parameters. Since descriptive and explanatory item response theory models were not fitted to the data, it is at present impossible to evaluate the cost savings that can be obtained with this approach during the item calibration phase and the item pool maintenance phase.

4.3.2. Element-Based TB-AIG

Proponents of element-based TB-AIG manipulated radicals (Irvine 2002) independently of each other. Studies using element-based TB-AIG also often exhibit more variation in the extent to which radicals (Irvine 2002) were chosen on the basis of cognitive processing models.
In this section, we first outline research using a weak theory approach to element-based TB-AIG. For instance, Shultz et al. (2025) used ChatGPT with iterative prompt refinement to construct clinical reasoning items for a pharmacy exam. The authors entirely relied on subject-matter expertise during item construction with little to no reference to cognitive processing models that outline radicals and commonly held misconceptions in pharmacy (weak theory approach). Shultz et al. (2025) noted that about 63 percent of their automatically generated TB-AIG items passed the expert panel review because they either required no or only minor revision. Nonetheless, a notable number of these items had to be discarded in a subsequent calibration phase due to deficits in classical item discrimination (e.g., negative item discrimination). Similar results have been reported by Chauhan et al. (2025), who also resorted to a weak theory approach to element-based TB-AIG. The authors used iterative prompt engineering with explanations to construct clinical reasoning items in physiology using GPT-4 on an item-to-item basis. In line with the aforementioned study, the authors reported a notable item loss after the expert panel review for both TB-AIG items and test items written by human item writers. In addition, TB-AIG items exhibited lower item discrimination and were generally easier than test items written by human item writers. The lower discrimination was due to an increased number of non-functional distractors in the TB-AIG-generated test items. Similarly, Lin and Chen (2024) used iterative zero-shot prompting with ChatGPT to construct passage-based reading comprehension items for two passages taken from an operational high-stakes test in China. They administered k = 18 TB-AIG items together with the original k = 10 items constructed for these two passages by human item writers. The authors report an item loss of ≈26% of the TB-AIG test items due to a misfit of the 2PL model. However, the distribution of the 2PL item parameters of the remaining TB-AIG items was similar to items made by human item writers. In addition, the quality of the remaining TB-AIG items was also judged to be similar to the items made by human item writers in a human subject-matter review process. The authors also examined the excluded TB-AIG items and noticed issues in the TB-AIG-constructed distractors as a possible reason for the misfit. Zu et al. (2023) used fine-tuning and iterative prompt engineering with GPT-2 to construct distractors for an operationally used fill-in-the-blank vocabulary test. They also examined different prompt designs and compared their results in terms of relevant linguistic features to a rule-based distractor construction method and to distractors constructed by human item writers. Their results indicated that more fine-tuned prompts yielded distractors most similar to human item writers.
By contrast, other studies resorted to existing cognitive processing models in selecting the radicals (strong theory approach). For instance, Attali et al. (2022) used iterative few-shot prompts with GPT-3 to construct k = 789 passages and test items measuring predefined levels of reading comprehension. That is, the authors analyzed the initial output of GPT-3 in terms of a set of radicals and refined their prompts to command GPT-3 to revise the initial test items according to their specification. This process was repeated until a human subject-matter expert was content with the result. Next, the TB-AIG items were reviewed by a panel of subject-matter experts. Attali et al. (2022) report that they were able to retain k = 454 (57.54%) items after the expert panel review. TB-AIG items that passed the panel review were administered to a few thousand test-takers. Each test-taker only worked on one of the TB-AIG items, which was administered as the last item in an operational high-stakes practice test. Although the authors were unable to examine the fit of descriptive item response theory models (De Boeck and Wilson 2004), their research design at least enabled them to calculate classic item parameters such as item difficulty and discrimination. Both classic test theoretical item statistics varied considerably. Runge et al. (2024) also used iterative prompt refinement to generate listening comprehension scripts and multiple-choice test items. The authors report that 719 (99%) of the TB-AIG test items passed the final expert panel review and were therefore subjected to the item calibration phase. Test-takers worked on two TB-AIG items administered at the end of an operational practice test. Classic test theoretical statistics (here, item difficulty and discrimination) were calculated. Runge et al. (2024) also tried to predict the item parameters on the basis of a set of radicals using multiple regression analysis. The radicals jointly accounted for R2 = 0.10 (item discrimination) to R2 = 29 (item difficulty) of the variance in the item parameters. Last, but not least, Schroeders and Achaa-Amankwaa (2025) used a combination of human item writing and element-based TB-AIG to construct a vocabulary item pool (k = 110), which required test-takers to identify an actual German word presented together with four pseudo-words. Using an incomplete link design, the item pool was calibrated by means of the 2PL model. In addition, multiple logistic regression analysis was used to predict the item difficulty parameter. The systematically manipulated radicals accounted for 57.4% of the variance in the 2PL item difficulty parameters. Similar to the aforementioned studies, the authors noted that severe prompt tuning and subject-matter reviews were required to get ChatGPT to construct the test items.
Research on element-based TB-AIG seems to resort to item calibration and explanatory item response theory modeling more frequently than item model-based TB-AIG. In addition, the authors resort more often to theoretical models to select radicals (Irvine 2002). However, there seems to be considerable variation in the extent to which this is the case. Furthermore, even strong(er)-theory element-based TB-AIG utilized cognitive processing models and/or computational models to a lesser extent than either the cognitive design system approach or the automatic min-max approach. This is unfortunate, since a more theory-based selection of radicals (Irvine 2002) seems to lead to a higher correlation between empirically estimated and predicted item parameters even in the verbal ability domain (cf. Arendasy et al. 2020; Kapoor et al. 2025). Thus, a stronger theoretical focus of the TB-AIG item construction process may be a fruitful area of further research. This may also include further refining the prompts to incorporate sketches, for example, story plots—like the ones commonly used by professional writers and game developers—to systematically manipulate reading comprehension inference demands. Recent research in computer science indicates that such an approach is in principle technically feasible (cf. Wang et al. 2024). Last, but not least, a stronger focus on descriptive and explanatory item response theory modeling (cf. De Boeck and Wilson 2004) and exploring the feasibility of including ideas similar to a quality control component would be desirable to build even stronger validity arguments for TB-AIG-constructed test items.

5. Discussion

The initial phase of switching from fixed-item linear paper-pencil tests to CATs and MSTs was associated with the hope that this would lead to a more flexible diagnostic process in addition to a reduction in costs (Luecht 2005). However, the hoped-for cost reduction has not been fulfilled. Practitioners and researchers soon realized that testing on demand, providing multiple retest opportunities, and equilizing individual differences in test preparation by means of providing cost-free practice-based training come at a cost. To ensure that these practices do not compromise the psychometric characteristics of the cognitive ability test, test developers need to be able to construct large pools of test items that have to be constantly monitored and updated (e.g., Ariel et al. 2006; Lang 2011; Liu et al. 2019; Segall 2004; Sinharay 2017; van der Linden and Glas 2010). Human item writing thus soon turned out to be a critical bottleneck in psychological and educational assessment (cf. Drasgow et al. 2006; Hornke and Habon 1986; Kosh et al. 2019; Luecht 2005; Wainer 2002; Zwick 2002) and has often been blamed for the fact that the hoped-for cost reductions were not achieved (cf. Luecht 2005).

5.1. Capacity of Item Generations Methods to Deal with the Specific Item Construction Demands

Research (e.g., He and Reckase 2014; Reckase 2010; Reckase et al. 2019; Veldkamp and van der Linden 2010) indicates that the distribution of the item parameters of an item pool should mimic the distribution of the person parameters of the target population. Furthermore, the item pool should only contain item variants (Bejar 2002). This is because item isomorphs (Bejar 2002) may introduce local dependencies, which lead to biased person parameter estimates (cf. Segall 2004; Glas and van der Linden 2003). To support item- and content-exposure control (cf. Georgiadou et al. 2007; Gierl et al. 2022a; Lim and Choi 2024), each item pool should contain a fair amount of psychometrically matched instances (Arendasy and Sommer 2013a) that are evenly distributed across the entire continuum of the latent trait. Psychometrically matched instances are test items that are (almost) identical to another test item in terms of their item parameters despite exhibiting a different combination of radicals. Despite the success of this approach in preventing item disclosure and reducing retest effects (e.g., Guo et al. 2009; Arendasy and Sommer 2017), the item construction demands are rather severe. To determine the extent to which the various item construction methods are able to keep pace with these demands, we need to determine how many different item variants and psychometrically matched instances each method is able to produce.
In the item model approach (e.g., Bejar 2002; Choi and Zhang 2019; Gierl et al. 2022b; LaDuca et al. 1986) and the cognitive design model approach (cf. Embretson 1998, 2002, 2016; Embretson and Yang 2007; Gorin 2006), the number of item variants is restricted by the number of item models. As outlined in Section 3.1 and Section 3.2, the number of item models varies considerably, but rarely exceeds 60 item models. Taking into account that, based on their expected item family function, some item models were intended to construct item variants, we would obtain an upper bound of ≤180 item variants in most practical applications. Unfortunately, there is to date little information on the extent to which these two approaches are capable of generating psychometrically matched item instances along the entire latent trait continuum. Given its importance to support the use of item- and content-exposure control algorithms, and to continuously update an existing item pool, this would be a fruitful field for further scientific research. By contrast, research using the automatic min-max approach (cf. Arendasy 2004; Arendasy and Sommer 2011, 2012a; Arendasy et al. 2020; Arendasy et al. 2024) often reports a number of item variants ranging from k = 120 to k = 320. However, the authors stress that these numbers may not be exhaustive but represent a rather representative selection of the number of item variants that can be generated. In addition, Arendasy and Sommer (2013a) report that most item generators were able to construct psychometrically matched instances. However, the number of psychometrically matched instances varied along various points of the latent trait continuum. Regarding TB-AIG, the low number of empirical studies using item calibration to evaluate the TB-AIG items prevents an evaluation of the extent to which this item construction method is able to meet the rather specific item demands in constructing and maintaining an initial item pool. As outlined in Section 4.3, more research on the psychometric characteristics of TB-AIG test items is needed before their contribution can be evaluated.
In addition to item variants (Bejar 2002), item isomorphs (Bejar 2002) also contribute to increasing test security. More specifically, item isomorphs could be used to construct multiple rotating item pools, which have been shown to reduce item exposure and enhance test security (cf. Ariel et al. 2006; Zhang et al. 2012). The various approaches to automatic item generation (AIG) do not seem to differ much in the extent to which they enable the construction of item isomorphs. However, the practical implementation of constructing item isomorphs seems to be more convenient in the case of the item model approach compared to the element-based approach. Tentative evidence indicates that TB-AIG also seems to be capable of constructing item isomorphs. However, more studies are needed in which TB-AIG items are calibrated to determine the extent to which isomorphic test items indeed exhibit (almost) statistically identical item parameters (cf. Section 4.3).

5.2. Capacity of Item Generation Methods to Provide Cost-Free Practice-Based Training

Based on evidence that both practice-based training and test coaching have been shown to be effective at increasing test-takers’ test scores, several researchers (e.g., AERA et al. 2018; Arendasy et al. 2016; Campion et al. 2019; Hermes et al. 2019; Lee et al. 2023a; Sommer et al. 2025) proposed providing cost-free practice-based training in order to equalize individual differences in access to commercial test preparation methods. The standards for educational and psychological assessment (AERA et al. 2018) propose that sample-test items used in practice-based training and test coaching should be representative of the actual test items. However, the practice-based training items should not be identical to operationally used test items in terms of their radical combinations. Furthermore, if sample-test items include feedback and tutorials, they should cover all difficulty levels of the operationally used test items (cf. Arendasy et al. 2016; Krautter et al. 2021; Sommer et al. 2025). However, providing effective feedback requires that test-takers’ cognitive processes and misconceptions are sufficiently well understood, and that reasons for incorrect responses can be inferred from their item responses. Thus, test item construction and sample-test item construction should ideally be based on validated cognitive processing models developed by cognitive scientists and professionals working in the domain of subject-matter didactics. In sum, the above-mentioned requirements for the construction of sample-test items used in practice-based training imply that sample-test item pools should ideally consist of item variants constructed by means of strong(er) theory. The main argument against the use of item isomorphs in practice-based training is that item isomorphs have sometimes been shown to lead to local dependencies between isomorphic items, and more pronounced score gains (cf. Arendasy and Sommer 2013a, 2017; Matton et al. 2011; Morley et al. 2004). Both effects induce a bias in the person parameter estimates, which calls the validity and fairness of the psychometric test into question. Since shorter time intervals between practice-based training and test-taking are common in high-stakes test preparation (cf. Powers 2012), the use of item isomorphs should be avoided. Instead, a separate item pool consisting of item variants (Bejar 2002) and psychometrically matched intances (Arendasy and Sommer 2013a) should be constructed for use in cost-free practice-based training. Unfortunately, this extends the item construction demands beyond the initial construction and maintenance of multiple item pool used for assessment proposes to the simultaneous construction of item pools used for assessment and an additional item pool used in practice-based training (cf. Arendasy et al. 2016; Hermes et al. 2019; Lee et al. 2023a; Sommer et al. 2025). Given that the specifications based on the empirical evidence outlined in Section 1.3 are identical to those outlined in Section 5.1, the evaluation of the extent to which current methods of automatic item generation (AIG) and transformer-based automatic item generation (TB-AIG) are able to meet these requirements is also identical.

5.3. Potential Cost Savings During the Actual Item Construction Phase

Despite differences in the various methods of automatic item generation (AIG)—including transformer-based automatic item generation (TB-AIG)—in terms of how radicals are deduced and manipulated (e.g., weak vs. strong theory approach, schema-based vs. element-based approach), all methods aim to make the item construction process more scalable and cost-efficient. Cost savings during the item construction process itself heavily depend on the degree of automatization that is possible. In our discussion of TB-AIG, we outlined the reasons for the recommendations to keep humans in the loop during the entire item construction process (cf. Attali et al. 2022; Bulut et al. 2024; Hao et al. 2024; von Davier 2019). This implies that semi-automatic TB-AIG requires more resources in terms of time and human manpower compared to fully automatic item generation. Thus, the cost saving in fully automatic AIG is naturally higher than in TB-AIG and semi-automatic AIG. Nevertheless, both semi-automatic item generation and TB-AIG seem to outperform human item writing in terms of the time needed to construct the test items, even when considering that 1 to 43 percent of the TB-AIG items were discarded after the final expert panel review (cf. Attali et al. 2022; Runge et al. 2024).

5.4. Potential Cost Savings in the Item Calibration Phase and the Item Pool Maintenance Phase

The various methods of automatic item generation (AIG) and transformer-based automatic item generation (TB-AIG) also differ in the extent to which they lead to cost savings in the item calibration phase and the item pool maintenance phase. Most studies conducted within the item model approach (e.g., Bejar 2002; Choi and Zhang 2019; Gierl et al. 2022b; LaDuca et al. 1986) and the cognitive design model approach (cf. Embretson 1998, 2002, 2016; Embretson and Yang 2007; Gorin 2006) utilizes the 2PL model to calibrate the test items. This also holds for the few studies using TB-AIG that use item response theory to calibrate the semi-automatically constructed TB-AIG items (e.g., Lin and Chen 2024; Schroeders and Achaa-Amankwaa 2025). By contrast, proponents of the automatic min-max approach (cf. Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024) utilized the 1PL Rasch model for item calibration (cf. Section 3.3). These two descriptive item response theory models differ in the number of item parameters that need to be estimated. While the 2PL model estimates an item difficulty parameter and an item discrimination parameter for each individual test item, the 1PL Rasch model restricts the item discrimination parameters to be equal across all test items (assumption of essential tau equivalent measurement). As a consequence, the item discrimination parameters do not need to be estimated. Several simulation studies (e.g., Belzak 2019; Sahin and Anil 2017; Schroeders and Gnambs 2025; Zimmer et al. 2024; Zwick et al. 1995) indicate that—all other things being equal—the 2PL Rasch model requires larger sample sizes to obtain reliable item parameter estimates and to conduct DIF analyses, than the 1PL Rasch model. Given that sample size is a major cost factor, cost savings may be achieved by automatically generating test items whose data have a higher chance to fit the 1PL Rasch model no worse than the 2PL model without running the risk of construct underrepresentation. To this end, Arendasy and associates (cf. Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024 ; Hines 2017) proposed conducting psychometric experiments to examine possible reasons for item-specific differences in the empirically estimated item discrimination parameters and to evaluate whether these reasons are linked to cognitive processes deemed to be central to the latent cognitive ability trait (for examples, see Arendasy and Sommer 2005, 2010, 2012b, 2013b; Arendasy et al. 2012).
A similar argument can be made with regard to cost savings in DIF analyses. Several studies (e.g., Kara and Dogan 2022; Sahin Kursad and Yalcin 2024; Piromsombat 2014) indicate that ignoring DIF in CAT and MST leads to biased person parameter estimates. Thus, DIF analyses should be mandatory. The various approaches to automatic item generation (AIG)—including transformer-based automatic item generation (TB-AIG)—also differ regarding the way in which DIF items and items exhibiting a misfit to the chosen descriptive item response theory model (here: 1PL model vs. 2PL model) are handled. Within the item model approach (e.g., Bejar 2002; Choi and Zhang 2019; Gierl et al. 2022b; LaDuca et al. 1986), the cognitive design model approach (cf. Embretson 1998, 2002, 2016; Embretson and Yang 2007; Gorin 2006), and TB-AIG (e.g., Lin and Chen 2024; Schroeders and Achaa-Amankwaa 2025), items are often excluded based on technical item selection criteria. This has been reported to lead to an item loss of ≤30 percent during the item calibration phase. By contrast, the automatic min-max approach proposes to either model sources of the misfit and/or DIF in terms of radicals or functional constraints (e.g., Arendasy and Sommer 2013b; Rajeb et al. 2024; Yu 1994), or to conduct psychometric experiments with precise hypotheses on possible causes for an item misfit or DIF deduced from cognitive processing models (cf. Arendasy and Sommer 2005, 2010, 2012b, 2013b; Arendasy et al. 2012; Hines 2017). The results are subsequently used in the construction and evaluation of the quality control component, which aims to minimize the likelihood of DIF due to construct-unrelated cognitive processes. This process has been reported to lead to an item loss of ≤1 percent (for a replication of this finding, see Arendasy et al. 2020; Hines 2017). This indicates that resorting to a strong(er) theory approach may not only lead to an improvement in validity but also to cost savings during the calibration of the initial item pool.
In addition, automatic item generation (AIG) also enables cost savings during the item pool maintenance phase. While traditional human item writing requires that all items need to be calibrated individually, AIG proposes either to calibrate item models or individual radicals as the basic building blocks of the item generators (cf. Section 3.1, Section 3.2 and Section 3.3). Since most studies in TB-AIG did not model the item parameters in terms of their radicals (for an exception, see Runge et al. 2024; Schroeders and Achaa-Amankwaa 2025), the extent to which TB-AIG may lead to cost savings in the item pool maintenance phase remains unclear. Several simulation studies (e.g., Bejar et al. 2002; Colvin et al. 2016; Glas and van der Linden 2003; Someshwar 2024; Tian and Choi 2023) indicate that—provided that predicted and calibrated item parameters are highly correlated (R ≥ 0.90)—using expected item family functions instead of the actually calibrated test item parameters is feasible and only leads to a negligible bias in the person parameter estimates that can be offset by a slight increase in the number of items administered during CAT. Almost identical results have been reported for the use of item parameters predicted on the basis of the radicals themselves (e.g., Doebler 2012; Embretson 1999; Freund et al. 2008; Mislevy et al. 1993; Matteucci et al. 2012; Someshwar 2024). Provided that studies using the cognitive design system approach (R = 0.70 to R = 0.90) and the automatic min-max approach (R = 0.89 to R = 0.96) often yielded a high correlation between predicted and empirically estimated item parameters, the results indicate that cost savings in the item pool maintenance phase may be possible by skipping the item calibration phase and using predicted item parameters instead of actually calibrated item parameters (for details, see Section 3.2 and Section 3.3). Interestingly, strong(er) theory item model approaches, which aim to model the variance of the item parameters within an item model in terms of radicals (cf. Cho et al. 2014; Geerlings et al. 2013; Glas et al. 2016), tend to perform better in this regard than weak(er) theory item model approaches. Thus, resorting to cognitive processing models and computational models in choosing radicals (Irvine 2002) not only provides valuable validity evidence in terms of cognitive processes involved in test item solving (for a review, see AERA et al. 2018) but also leads to cost savings in the item pool maintenance phase. The beneficial effect of strong(er)-theory approaches is also evident in human item writing (cf. Section 2). The more human item writers know about construct-related cognitive processes and how they are triggered by certain item design features, the better they are able to tailor their item writing to the demands of the specifications of the blueprint for an ideal item pool. This recurring finding across different item construction methods points to the potential benefit of a closer cooperation between theoretical psychologists, who are concerned with the development and evaluation of cognitive processing models and computational models, and applied psychometricians, who aim to develop valid and fair cognitive ability tests (for a similar argument, see Smaldino 2020).

5.5. Differences in Test Security Concerns

Without a doubt, the item model approach (e.g., Bejar 2002; Choi and Zhang 2019; LaDuca et al. 1986) is the most popular approach to AIG to date. This has led some researchers (e.g., Circi et al. 2023; Gierl et al. 2022a; von Davier 2018) to incorrectly equate the item model approach with the entire field of AIG. The main reason for this is that item models can be constructed using both weak and strong theoretical foundations (Drasgow et al. 2006). In addition, item models are convenient to implement in programming terms (Embretson and Yang 2007). Furthermore, usability studies indicate that item models are easier to handle for subject-matter experts, who often need to review and revise the item models without having a background in programming (cf. Falcão et al. 2023, 2024; Choi and Zhang 2019; Gierl and Lai 2012; Lai et al. 2016). The ease in handling item models—also partly due to their often hierarchical structure and finite number of item models—surely contributes to the cost savings reported during the actual item construction process and the development of the item generator (cf. Falcão et al. 2023, 2024; Choi and Zhang 2019; Gierl and Lai 2012; Gierl et al. 2022a). However, these benefits also come at a cost. Using item models in CAT and MST requires both content coding and item exposure control at the item- and item model level (cf. Gierl et al. 2022a). This is necessary to ensure that isomorphic test items (Bejar 2002) are not administered within the same test session to prevent a bias in the person parameter estimates (cf. Glas and van der Linden 2003; Segall 2004). Furthermore, the ease of use and the often hierarchical structure of item models make them more vulnerable to organized item theft. Research (cf. Yi et al. 2008) indicates that item theft detrimentally affects the person parameter estimates and, consequentially, the validity and fairness of the psychometric test. If an entire item model has been disclosed or leaked, all items generated by that item model are compromised and need to be removed from operational use. By contrast, in the element-based approach, radicals (Irvine 2002) are manipulated independently of each other. While this may complicate the ease of use of the item generator by non-experts, it also enhances test security. This is because item theft can only occur on an item-to-item basis instead of at the level of item models or item model families. Thus, using item models requires additional technical test- or item-model security measures to protect the item models from item disclosure and organized item theft (for an overview, see Foster 2016). A related argument can be made with regard to the use of open and commercial transformer models in item generation (cf. Bulut et al. 2024; Hao et al. 2024; Section 4.2).

6. Concluding Remarks

Thus, different methods of automatic item generation (AIG; Arendasy 2004; Arendasy and Sommer 2011, 2012a; Arendasy et al. 2024; Embretson and Yang 2007; Irvine and Kyllonen 2002), including transformer-based automatic item generation (TB-AIG: Attali et al. 2022; von Davier 2019), come with unique costs and benefits at various stages of the construction and maintenance phase of multiple item pools needed in assessment as well as cost-free practice-based training. Unfortunately, current cost-benefit studies neither take differences between methods in the costs associated with the construction and maintenance of multiple item pools into account nor do they factor in differences in the costs associated with varying levels of risk of item disclosure and organized item theft (cf. Kosh et al. 2019). An update of such cost-benefit analyses that takes these aspects into account would help test developers choose a particular approach to automatic (including transformer-based automatic) item generation that is more tailored to their current situational constraints and their short-term and long-term needs.

Author Contributions

Conceptualization, M.A. and M.S.; investigation, M.A. and M.S.; writing—original draft preparation, M.S.; writing—review and editing, M.A. and M.S. All authors have read and agreed to the published version of the manuscript.

Funding

Open Access Funding by the University of Graz.

Acknowledgments

The authors would like to thank Tanja Newiger for her help in editing and proofreading the references.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AIGAutomatic item generation
CATComputerized adaptive test
DIFDifferential item functioning
LLTMLinear Logistic Test Model
LOFTLinear on-the-fly test
MSTMulti-stage test
TB-AIGTransformer-based automatic item generation

References

  1. Abu-Haifa, Mohammad, Bara’a Etawi, Huthaifa Alkhatatbeh, and Ayman Ababneh. 2024. Comparative analysis of ChatGPT, GPT-4, and Microsoft Copilot Chatbots for GRE test. International Journal of Learning, Teaching and Educational Research 23: 327–47. [Google Scholar] [CrossRef]
  2. Ahn, Jihyun J., and Wenpeng Yin. 2025. Prompt-Reverse Inconsistency: LLM Self-Inconsistency Beyond Generative Randomness and Prompt Paraphrasing. arXiv arXiv:2504.01282. [Google Scholar]
  3. Al Faraby, Said, Ade Romadhony, and Adiwijaya. 2024. Analysis of llms for educational question classification and generation. Computers and Education: Artificial Intelligence 7: 100298. [Google Scholar] [CrossRef]
  4. Allalouf, Avi, and Gershon Ben-Shakhar. 1998. The effect of coaching on the predictive validity of scholastic aptitude tests. Journal of Educational Measurement 35: 31–47. [Google Scholar] [CrossRef]
  5. American Educational Research Association (AERA), American Psychological Association (APA), and National Council on Measurement in Education (NCME). 2018. Standards for Educational and Psychological Testing. Washington, DC: American Psychological Association. [Google Scholar]
  6. Anderson, John R., Jon M. Fincham, and Scott Douglass. 1997. The role of examples and rules in the acquisition of a cognitive skill. Journal of Experimental Psychology: Learning, Memory and Cognition 23: 932–45. [Google Scholar] [CrossRef]
  7. Appelhaus, Stefan, Susanne Werner, Pascal Grosse, and Juliane E. Kämmer. 2023. Feedback, fairness, and validity: Effects of disclosing and reusing multiple-choice questions in medical schools. Medical Education Online 28: 2143298. [Google Scholar] [CrossRef]
  8. Appelrouth, Jed I., Karen M. Zabrucky, and DeWayne Moore. 2017. Preparing students for college admissions tests. Assessment in Education: Principles, Policy and Practice 24: 78–95. [Google Scholar] [CrossRef]
  9. Arendasy, Martin. 2000. Psychometrischer Vergleich Computergestützter Vorgabeformen bei Raumvorstellungsaufgaben: Stereoskopisch-Dreidimensionale und Herkömmlich-Zweidimensionale Darbietung. Ph.D thesis, Universität Wien, Wien, Austria. [Google Scholar]
  10. Arendasy, Martin. 2004. Automatisierte Itemgenerierung und Psychometrische Qualitätssicherung am Beispiel des Matrizentests GEOM. Lausanne: Peter Lang. [Google Scholar]
  11. Arendasy, Martin, and Markus Sommer. 2005. The effect of different types of perceptual manipulations on the dimensionality of automatically generated figural matrices. Intelligence 33: 307–24. [Google Scholar] [CrossRef]
  12. Arendasy, Martin, and Markus Sommer. 2007. Using psychometric technology in educational assessment: The case of a schema-based isomorphic approach to the automatic generation of quantitative reasoning items. Learning and Individual Differences 17: 366–83. [Google Scholar] [CrossRef]
  13. Arendasy, Martin, and Markus Sommer. 2010. Evaluating the contribution of different item features to the effect size of the gender difference in three-dimensional mental rotation using automatic item generation. Intelligence 38: 574–81. [Google Scholar] [CrossRef]
  14. Arendasy, Martin, and Markus Sommer. 2011. Automatisierte Itemgenerierung: Aktuelle Ansätze, Anwendungen und Forschungen. In Enzyklopädie für Psychologie: Methoden der Psychologischen Diagnostik. Edited by Lutz F. Hornke, Manfred Amelang and Martin Kersting. Göttingen: Hogrefe, pp. 215–80. [Google Scholar]
  15. Arendasy, Martin, and Markus Sommer. 2012a. Using automatic item generation to meet the increasing item demands of high-stakes assessment. Learning and Individual Differences 22: 112–17. [Google Scholar] [CrossRef]
  16. Arendasy, Martin, and Markus Sommer. 2012b. Gender differences in figural matrices: The moderating role of item design features. Intelligence 40: 584–97. [Google Scholar] [CrossRef]
  17. Arendasy, Martin, and Markus Sommer. 2013a. Quantitative differences in retest effects across different methods used to construct alternate test forms. Intelligence 41: 181–92. [Google Scholar] [CrossRef]
  18. Arendasy, Martin, and Markus Sommer. 2013b. Reducing response elimination strategies enhances the construct validity of figural matrices. Intelligence 41: 234–43. [Google Scholar] [CrossRef]
  19. Arendasy, Martin, Markus Sommer, and Andreas Hergovich. 2007. Psychometrische Technologie: Automatische Zwei-Komponenten-Itemgenerierung am Beispiel eines neuen Aufgabentyps zur Messung der Numerischen Flexibilität. Diagnostica 53: 119–30. [Google Scholar] [CrossRef]
  20. Arendasy, Martin, Markus Sommer, and Friedrich Mayr. 2012. Using automatic item generation to simultaneously con-struct German and English versions of a verbal fluency test. Journal of Cross-Cultural Psychology 43: 464–79. [Google Scholar] [CrossRef]
  21. Arendasy, Martin, Markus Sommer, and Georg Gittler. 2010. Combining automatic item generation and experimental designs to investigate the contribution of cognitive components to the gender difference in mental rotation. Intelligence 38: 506–12. [Google Scholar] [CrossRef]
  22. Arendasy, Martin, Markus Sommer, and Georg Gittler. 2020. Manual Intelligence-Struktur-Battery 2 (INSBAT-2). Mödling: SCHUHFRIED GmbH. [Google Scholar]
  23. Arendasy, Martin, Markus Sommer, Andreas Hergovich, and Martina Feldhammer. 2011. Evaluating the impact of depth cue salience in working three-dimensional mental rotation tasks by means of psychometric experiments. Learning and Individual Differences 21: 403–8. [Google Scholar] [CrossRef]
  24. Arendasy, Martin, Markus Sommer, Georg Gittler, and Andreas Hergovich. 2006. Automatic generation of quantitative reasoning items: Pilot study. Journal of Individual Differences 27: 2–14. [Google Scholar] [CrossRef]
  25. Arendasy, Martin E., and Markus Sommer. 2017. Reducing the effect size of the retest effect: Examining different approaches. Intelligence 62: 89–98. [Google Scholar] [CrossRef]
  26. Arendasy, Martin E., Markus Sommer, Karin Gutierrez-Lobos, and Joachim F. Punter. 2016. Do individual differences in test preparation compromise the measurement fairness of admission tests? Intelligence 55: 44–56. [Google Scholar] [CrossRef]
  27. Arendasy, Martin E., Markus Sommer, Reinhard Tschiesner, Martina Feldhammer-Kahr, and Konstantin Umdasch. 2024. Using automatic item generation to construct scheduling problems measuring planning ability. Intelligence 106: 101855. [Google Scholar] [CrossRef]
  28. Ariel, Adelaide, Wim J. van Der Linden, and Bernard P. Veldkamp. 2006. A strategy for optimizing item-pool management. Journal of Educational Measurement 43: 85–96. [Google Scholar] [CrossRef]
  29. Attali, Yigal, Andrew Runge, Geoffrey T. LaFlair, Kevin Yancey, Sarah Goodwin, Yena Park, and Alina A. von Davier. 2022. The interactive reading task: Transformer-based automatic item generation. Frontiers in Artificial Intelligence 5: 903077. [Google Scholar] [CrossRef]
  30. Attali, Yigal, Luis Saldivia, Carol Jackson, Fred Schuppan, and Wilbur Wanamaker. 2014. Estimating Item Difficulty with Comparative Judgments. Princeton: ETS. [Google Scholar]
  31. Baldonado, Angela A., Dubravka Svetina, and Joanna Gorin. 2015. Using necessary information to identify item dependence in passage-based reading comprehension tests. Applied Measurement in Education 28: 202–18. [Google Scholar] [CrossRef]
  32. Balestri, Roberto. 2025. Gender and content bias in Large Language Models: A case study on Google Gemini 2.0 Flash Experimental. Frontiers in Artificial Intelligence 8: 1558696. [Google Scholar] [CrossRef] [PubMed]
  33. Bangert-Drowns, Robert L., James A. Kulik, and Chen-Lin C. Kuli. 1983. Effects of coaching programs on achievement test performance. Review of Educational Research 53: 571–85. [Google Scholar] [CrossRef]
  34. Becker, Betsy J. 1990. Coaching for the Scholastic Aptitude Test: Further synthesis and appraisal. Review of Educational Research 60: 373–417. [Google Scholar] [CrossRef]
  35. Beg, Mirza A., Afifa Tabassum, and Sobia Ali. 2021. Role of faculty development workshop for improving MCQS quality in basic medical sciences. Biomedica 37: 51–55. [Google Scholar] [CrossRef]
  36. Bejar, Isaac I. 1983. Subject matter experts’ assessment of item statistics. Applied Psychological Measurement 7: 303–10. [Google Scholar] [CrossRef]
  37. Bejar, Isaac I. 2002. Generative testing: From conception to implementation. In Item Generation for Test Development. Edited by Sidney H. Irvine and Patrick C. Kyllonen. Mahwah: Lawrence Erlbaum, pp. 199–217. [Google Scholar]
  38. Bejar, Isaac I., René R. Lawless, Mary E. Morley, Michael E. Wagner, Randy E. Bennett, and Javier Revuelta. 2002. A Feasibility Study of On-the-Fly Item Generation in Adaptive Testing (GRE Board Professional Rep. No. 98-12P). Princeton: ETS. [Google Scholar]
  39. Bejar, Isaac I., Roger Chaffin, and Susan Embretson. 2012. Cognitive and Psychometric Analysis of Analogical Problem Solving. Berlin: Springer. [Google Scholar]
  40. Belzak, William C., Ben Naismith, and Jill Burstein. 2023. Ensuring fairness of human-and AI-generated test items. In International Conference on Artificial Intelligence in Education. Cham: Springer Nature Switzerland, pp. 701–7. [Google Scholar]
  41. Belzak, William C. M. 2019. Testing differential item functioning in small samples. Multivariate Behavioral Research 55: 722–47. [Google Scholar] [CrossRef]
  42. Berenbon, Rebecca F., and Bridget C. McHugh. 2023. Do subject matter experts’ judgments of multiple-choice format suitability predict item quality? Educational Measurement: Issues and Practice 42: 13–21. [Google Scholar] [CrossRef]
  43. Bethell-Fox, Charles E., David F. Lohman, and Richard E. Snow. 1984. Adaptive reasoning: Componential and eye movement analysis of geometric analogy performance. Intelligence 8: 205–38. [Google Scholar] [CrossRef]
  44. Bezirhan, Ummugul, and Matthias von Davier. 2023. Automated reading passage generation with OpenAI’s large language model. Computers and Education: Artificial Intelligence 5: 100161. [Google Scholar] [CrossRef]
  45. Bhayana, Rajesh, Satheesh Krishna, and Robert R. Bleakney. 2023. Performance of ChatGPT on a radiology board-style examination: Insights into current strengths and limitations. Radiology 307: e230582. [Google Scholar] [CrossRef]
  46. Blum, Diego, and Heinz Holling. 2018. Automatic generation of figural analogies with the imak package. Frontiers in Psychology 9: 1286. [Google Scholar] [CrossRef] [PubMed]
  47. Borsboom, Denny, Jan-Willem Romeijn, and Jelte M. Wicherts. 2008. Measurement invariance versus selection invariance: Is fair selection possible? Psychological Methods 13: 75–98. [Google Scholar] [CrossRef] [PubMed]
  48. Bozkurt, Aras, and Ramesh C. Sharma. 2023. Generative AI and prompt engineering: The art of whispering to let the genie out of the algorithmic world. Asian Journal of Distance Education 18: 1–7. Available online: https://www.asianjde.com/ojs/index.php/AsianJDE/article/view/749 (accessed on 3 March 2025).
  49. Briggs, Derek C. 2009. Preparation for College Admission Exams (2009 NACAC Discussion Paper). Arlington: National Association for College Admission Counseling. [Google Scholar]
  50. Brown, Tom B., Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared Kaplan, Prafulla Dhariwal, Arvind Neelakantan, Pranav Shyam, Girish Sastry, Amanda Askell, and et al. 2020. Language models are few-shot learners. Advances in Neural Information Processing Systems 33: 1877–901. [Google Scholar]
  51. Buchmann, Claudia, Dennis J. Condron, and Vincent J. Roscigno. 2010. Shadow education, American style: Test preparation, the SAT and college enrollment. Social Forces 89: 435–82. [Google Scholar] [CrossRef]
  52. Bulut, Okan, Maggie Beiting-Parrish, Jodi M. Casabianca, Sharon C. Slater, Hong Jiao, Dan Song, and Polina Morilova. 2024. The rise of artificial intelligence in educational measurement: Opportunities and ethical challenges. arXiv arXiv:2406.18900. [Google Scholar]
  53. Burke, Eugene F. 1997. A short note on the persistence of retest effects on aptitude scores. Journal of Occupational and Organizational Psychology 70: 295–301. [Google Scholar] [CrossRef]
  54. Burns, Gary N., Brian P. Siers, and Neil D. Christiansen. 2008. Effects of providing pre-test information and preparation materials on applicant reactions to selection procedures. International Journal of Selection and Assessment 16: 73–77. [Google Scholar] [CrossRef]
  55. Calamia, Matthew, Kristian Markon, and Daniel Tranel. 2012. Scoring higher the second time around: Meta-analyses of practice effects in neuropsychological assessment. The Clinical Neuropsychologist 26: 543–70. [Google Scholar] [CrossRef] [PubMed]
  56. Campion, Michael C., Emily D. Campion, and Michael A. Campion. 2019. Using practice employment tests to improve recruitment and personnel selection outcomes for organizations and job seekers. Journal of Applied Psychology 104: 1089–102. [Google Scholar] [CrossRef] [PubMed]
  57. Chan, Kuang W., Farhan Ali, Joonhyeong Park, Kah S. B. Sham, Erdalyn Y. T. Tan, Francis W. C. Chong, and Guan K. Sze. 2025. Automatic item generation in various STEM subjects using large language model prompting. Computers and Education: Artificial Intelligence 8: 100344. [Google Scholar] [CrossRef]
  58. Chauhan, Archana, Farah Khaliq, and Kirtana Raqhurama Nayak. 2025. Assessing quality of scenario-based multiple-choice questions in physiology: Faculty-generated vs. ChatGPT-generated questions among phase I medical students. International Journal of Artificial Intelligence in Education, 1–30. [Google Scholar] [CrossRef]
  59. Cho, Sun-Joo, Paul De Boeck, Susan Embretson, and Sophia Rabe-Hesketh. 2014. Additive multilevel item structure models with random residuals: Item modeling for explanation and item generation. Psychometrika 79: 84–104. [Google Scholar] [CrossRef]
  60. Choi, Jaehwa, and Xinxin Zhang. 2019. Computerized item modeling practices using computer adaptive formative assessment automatic item generation system: A tutorial. The Quantitative Methods for Psychology 15: 214–25. [Google Scholar] [CrossRef]
  61. Chung, Jinmin, and Sungyeun Kim. 2024. Comparison of rule-based models and Large Language Models in item and feedback generation. Journal of Science Education 48: 154–69. [Google Scholar] [CrossRef]
  62. Circi, Ruhan, Juanita Hicks, and Emmanuel Sikali. 2023. Automatic item generation: Foundations and machine learning-based approaches for assessments. Frontiers in Education 8: 858273. [Google Scholar] [CrossRef]
  63. Colvin, Kimberly F., Lisa A. Keller, and Frederic Robin. 2016. Effect of imprecise parameter estimation on ability estimation in a multistage test in an automatic item generation context. Journal of Computerized Adaptive Testing 4: 1–18. [Google Scholar] [CrossRef]
  64. Daniel, Robert C., and Susan E. Embretson. 2010. Designing cognitive complexity in mathematical problem-solving items. Applied Psychological Measurement 34: 348–64. [Google Scholar] [CrossRef]
  65. De Boeck, Paul, and Mark Wilson. 2004. Explanatory Item Response Models: A Generalized Linear and Nonlinear Approach. New York: Springer. [Google Scholar]
  66. Denker, Marek, Clara Schütte, Martin Kersting, Daniel Weppert, and Stephan J. Stegt. 2023. How can applicants’ reactions to scholastic aptitude tests be improved? A closer look at specific and general tests. Frontiers in Education 7: 931841. [Google Scholar] [CrossRef]
  67. Devlin, Jacob, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2018. Bert: Pre-training of deep bidirectional transformers for language understanding. arXiv arXiv:1810.04805. [Google Scholar]
  68. Doebler, Anna. 2012. The problem of bias in person parameter estimation in adaptive testing. Applied Psychological Measurement 36: 255–70. [Google Scholar] [CrossRef]
  69. Doebler, Anna, and Heinz Holling. 2016. A processing speed test based on rule-based item generation: An analysis with the Rasch Poisson Counts model. Learning and Individual Differences 52: 121–28. [Google Scholar] [CrossRef]
  70. Draheim, Christopher, Tyler L. Harrison, Susan E. Embretson, and Randall W. Engle. 2018. What item response theory can tell us about the complex span tasks. Psychological Assessment 30: 116–29. [Google Scholar] [CrossRef]
  71. Drasgow, Fritz, Richard M. Luecht, and Randy Bennett. 2006. Technology and testing. In Educational Measurement, 4th ed. Edited by R. L. Brennan. Westport: American Council on Education and Praeger Publishers, pp. 471–55. [Google Scholar]
  72. Eleragi, Ali M. S., Elhadi Miskeen, Kamal Hussein, Assad A. Rezigalla, Masoud I. Adam, Jaber A. Al-Faifi, and Osama A. Mohammed. 2025. Evaluating the multiple-choice questions quality at the College of Medicine, University of Bisha, Saudi Arabia: A three-year experience. BMC Medical Education 25: 233. [Google Scholar] [CrossRef]
  73. El Masri, Yasmine H., Steve Ferrara, Peter W. Foltz, and Jo-Anne Baird. 2017. Predicting item difficulty of science national curriculum tests: The case of key stage 2 assessments. The Curriculum Journal 28: 59–82. [Google Scholar] [CrossRef]
  74. Embretson, Susan. 2023. Understanding examinees’ item responses through cognitive modeling of response accuracy and response times. Large-Scale Assessments in Education 11: 9. [Google Scholar] [CrossRef]
  75. Embretson, Susan E. 1998. A cognitive design system approach to generating valid tests: Application to abstract reasoning. Psychological Methods 3: 300–96. [Google Scholar] [CrossRef]
  76. Embretson, Susan E. 1999. Generating items during testing: Psychometric issues and models. Psychometrika 64: 407–33. [Google Scholar] [CrossRef]
  77. Embretson, Susan E. 2002. Generating abstract reasoning items with cognitive theory. In Generating Items for Cognitive Tests: Theory and Practice. Edited by Sidney Irvine and Patrick Kyllonen. Mahwah: Erlbaum. [Google Scholar]
  78. Embretson, Susan E. 2005. Measuring human intelligence with artificial intelligence. In Cognition and Intelligence. Edited by Robert J. Sternberg and Jean E. Pretz. New York: Cambridge University Press, pp. 251–67. [Google Scholar]
  79. Embretson, Susan E. 2016. Understanding examinees’ responses to items: Implications for measurement. Educational Measurement: Issues and Practice 35: 6–22. [Google Scholar] [CrossRef]
  80. Embretson, Susan E., and Joanna S. Gorin. 2001. Improving construct validity with cognitive psychology principles. Journal of Educational Measurement 38: 343–68. [Google Scholar] [CrossRef]
  81. Embretson, Susan E., and Neal M. Kingston. 2018. Automatic item generation: A more efficient process for developing mathematics achievement items? Journal of Educational Measurement 55: 112–31. [Google Scholar] [CrossRef]
  82. Embretson, Susan E., and Robert C. Daniel. 2008. Understanding and quantifying cognitive complexity level in mathematical problem solving items. Psychology Science 50: 328–44. [Google Scholar]
  83. Embretson, Susan E., and Xiangdong Yang. 2007. Automatic item generation and cognitive psychology. In Handbook of Statistics: Vol 26 Psychometrics. Edited by Calyampudi R. Rao and Sandip Sinharay. North Holland: Elsevier, pp. 747–68. [Google Scholar]
  84. Emekli, Emre, and Betül N. Karahan. 2025. Comparison of automatic item generation methods in the assessment of clinical reasoning skills. Revista Española de Educación Médica 6: 1–12. [Google Scholar] [CrossRef]
  85. Enright, Mary K., Mary Morley, and Kathleen M. Sheehan. 2002. Items by design: The impact of systematic feature variation on item statistical characteristics. Applied Measurement in Education 15: 49–74. [Google Scholar] [CrossRef]
  86. Estrada, Eduardo, Emilio Ferrer, Fransisco J. Abad, Fransisco J. Román, and Roberto Colom. 2015. A general factor of intelligence fails to account for changes in tests’ scores after cognitive practice: A longitudinal multi-group latent variable study. Intelligence 50: 93–99. [Google Scholar] [CrossRef]
  87. Falcão, Filipe, Daniela M. Pereira, Nuno Gonçalves, Andre De Champlain, Patrício Costa, and José M. Pêgo. 2023. A suggestive approach for assessing item quality, usability and validity of Automatic Item Generation. Advances in Health Sciences Education 28: 1441–65. [Google Scholar] [CrossRef]
  88. Falcão, Filipe M. V., Daniela M. Pereira, José M. Pêgo, and Patrício Costa. 2024. Progress is impossible without change: Implementing automatic item generation in medical knowledge progress testing. Education and Information Technologies 29: 4505–30. [Google Scholar] [CrossRef]
  89. Farrell, Simon, and Stephan Lewandowsky. 2010. Computational models as aids to better reasoning in psychology. Current Directions in Psychological Science 19: 329–35. [Google Scholar] [CrossRef]
  90. Fehringer, Benedict C. 2020. Spatial thinking from a different view: Disentangling top-down and bottom-up processes using eye tracking. Open Psychology 2: 138–212. [Google Scholar] [CrossRef]
  91. Fehringer, Benedict C. 2023. Different perspectives on retest effects in the context of spatial thinking: Interplay of behavioral performance, cognitive processing, and cognitive workload. Journal of Intelligence 11: 66. [Google Scholar] [CrossRef] [PubMed]
  92. Fischer, Gerhard H. 1995. The Linear Logistic Test Model. In Rasch Models. Foundations, Recent Developments, and Applications. Edited by G. H. Fischer and I. W. Molenaar. New York: Springer, pp. 157–80. [Google Scholar]
  93. Folk, Valerie G., and Robert L. Smith. 2002. Models for delivery of CBTs. In Computer-Based Testing: Building the Foundation for Future Assessments. Edited by Craig Mills, Maria Potenza, John Fremer and William Ward. Mahwah: Lawrence Erlbaum, pp. 41–66. [Google Scholar]
  94. Foster, David. 2016. Testing technology and its effects on test security. In Technology and Testing: Improving Educational and Psychological Measurement. Edited by Fritz Drasgow. New York: Routledge, pp. 235–55. [Google Scholar]
  95. Förster, Natalie, and Jörg-Tobias Kuhn. 2023. Ice is hot and water is dry: Developing equivalent reading tests using rule-based item design. European Journal of Psychological Assessment 39: 96–105. [Google Scholar] [CrossRef]
  96. Freund, Philipp A., and Heinz Holling. 2011. How to get real smart: Modeling retest and training effects in ability testing using computer-generated figural matrices items. Intelligence 39: 233–43. [Google Scholar] [CrossRef]
  97. Freund, Philipp A., Stefan Hofer, and Heinz Holling. 2008. Explaining and controlling for the psychometric properties of computer-generated figural matrix items. Applied Psychological Measurement 32: 195–210. [Google Scholar] [CrossRef]
  98. Fried, Eiko I. 2020. Lack of theory building and testing impedes progress in the factor and network literature. Psychological Inquiry 31: 271–88. [Google Scholar] [CrossRef]
  99. Fu, Yanyan, Edison M. Choe, Hwanggyu Lim, and Jaehwa Choi. 2022. An evaluation of automatic item generation: A case study of weak theory approach. Educational Measurement: Issues and Practice 41: 10–22. [Google Scholar] [CrossRef]
  100. Funk, Paul F., Cosima C. Hoch, Samuel Knoedler, Leonard Knoedler, Sebastian Cotofana, Giuseppe Sofo, and Michael Alfertshofer. 2024. ChatGPT’s response consistency: A study on repeated queries of medical examination questions. European Journal of Investigation in Health, Psychology and Education 14: 657–68. [Google Scholar] [CrossRef]
  101. Geerlings, Hanneke, Wim J. van der Linden, and Cees A. Glas. 2013. Optimal test design with rule-based item generation. Applied Psychological Measurement 37: 140–61. [Google Scholar] [CrossRef]
  102. Georgiadou, Elissavet, Evangelos Triantafillou, and Anastasios A. Economides. 2007. A review of item exposure control strategies for computerized adaptive testing developed from 1983 to 2005. Journal of Technology, Learning, and Assessment 5. Available online: https://files.eric.ed.gov/fulltext/EJ838610.pdf (accessed on 3 March 2025).
  103. Gierl, Mark J., and Hollis Lai. 2012. The role of item models in automatic item generation. International Journal of Testing 12: 273–98. [Google Scholar] [CrossRef]
  104. Gierl, Mark J., Hollis Lai, and Simon Turner. 2012. Using automatic item generation to create multiple-choice items for assessments in medical education. Medical Education 46: 757–65. [Google Scholar] [CrossRef]
  105. Gierl, Mark J., Jinnie Shin, Tahereh Firoozi, and Hollis Lai. 2022a. Using content coding and automatic item generation to improve test security. Frontiers in Education 7: 853578. [Google Scholar] [CrossRef]
  106. Gierl, Mark J., Kimberly Swygert, Donna Matovinovic, Allison Kulesher, and Hollis Lai. 2022b. Three sources of validation evidence needed to evaluate the quality of generated test items for medical licensure. Teaching and Learning in Medicine 36: 72–82. [Google Scholar] [CrossRef]
  107. Glas, Cees A., Wim J. van der Linden, and Hanneke Geerlings. 2016. Item-family models. In Handbook of Item Response Theory. Edited by Wim J. van der Linden. Boca Raton: Chapman and Hall/CRC, vol. 1, pp. 465–76. [Google Scholar]
  108. Glas, Cees A. W., and Wim J. van der Linden. 2003. Computerized adaptive testing with item cloning. Applied Psychological Measurment 27: 247–61. [Google Scholar] [CrossRef]
  109. Gorin, Joanna S. 2005. Manipulating processing difficulty of reading comprehension questions: The feasibility of verbal item generation. Journal of Educational Measurement 42: 351–73. [Google Scholar] [CrossRef]
  110. Gorin, Joanna S. 2006. Test design with cognition in mind. Educational Measurement: Issues and Practice 25: 21–35. [Google Scholar] [CrossRef]
  111. Gorin, Joanna S., and Susan E. Embretson. 2006. Item diffficulty modeling of paragraph comprehension items. Applied Psychological Measurement 30: 394–411. [Google Scholar] [CrossRef]
  112. Graf, Edith A., Stephen Peterson, Manfred Steffen, and René Lawless. 2005. Psychometric and Cognitive Analysis as a Basis for the Design and Revision of Quantitative Item Models (No. RR-05-25). Princeton: Educational Testing Service. [Google Scholar]
  113. Greeno, James G., Joyce L. Moore, and David R. Smith. 1993. Transfer of situated learning. In Transfer on Trial: Intelligence, Cognition, and Instruction. Edited by Douglas K. Detterman and Robert J. Sternberg. New York: Ablex Publishing, pp. 99–167. [Google Scholar]
  114. Guest, Olivia, and Andrea E. Martin. 2021. How computational modeling can force theory building in psychological science. Perspectives on Psychological Science 16: 789–802. [Google Scholar] [CrossRef] [PubMed]
  115. Gühne, Daniela, Philipp Doebler, David M. Condon, Fang Luo, and Luning Sun. 2020. Validity and reliability of automatically generated propositional reasoning items. European Journal of Psychological Assessment 16: 325–39. [Google Scholar] [CrossRef]
  116. Guo, Jing, Louis Tay, and Fritz Drasgow. 2009. Conspiracy and test compromise: An evaluation of the resistance of test systems to small-scale cheating. International Journal of Testing 9: 283–309. [Google Scholar] [CrossRef]
  117. Gupta, Piyush, Pinky Meena, Amir M. Khan, Rajeev K. Malhotra, and Tejinder Singh. 2020. Effect of faculty training on quality of multiple-choice questions. International Journal of Applied and Basic Medical Research 10: 210–14. [Google Scholar] [CrossRef]
  118. Hao, Jiangang, Alina A. von Davier, Victoria Yaneva, Susan Lottridge, Matthias von Davier, and Deborah J. Harris. 2024. Transforming assessment: The impacts and implications of large language models and generative AI. Educational Measurement: Issues and Practice 43: 16–29. [Google Scholar] [CrossRef]
  119. Hausknecht, John P., Jane A. Halpert, Nicole T. DiPaolo, and Meghan O. Moriarty Gerrard. 2007. Retesting in selection: A meta-analysis of coaching and practice effects for tests of cognitive ability. Journal of Applied Psychology 92: 373–85. [Google Scholar] [CrossRef]
  120. Hayes, Taylor R., Alexander A. Petrov, and Per B. Sederberg. 2015. Do we really become smarter when our fluid intelligence test scores improve? Intelligence 48: 1–14. [Google Scholar] [CrossRef]
  121. He, Wei, and Mark D. Reckase. 2014. Item pool design for an operational variable-length computerized adaptive test. Educational and Psychological Measurement 74: 473–94. [Google Scholar] [CrossRef]
  122. Heil, Martin, and Petra Jansen-Osmann. 2008. Sex differences in mental rotation with polygons of different complexity: Do men utilize holistic processes whereas women prefer piecemeal ones? The Quarterly Journal of Experimental Psychology 61: 683–89. [Google Scholar] [CrossRef]
  123. Heil, Martin, Frank Rösler, Michael Link, and Jasmin Bajric. 1998. What is improved if mental rotation task is repeated: The efficiency of memory access, or the speed of transformation routine? Psychological Research 61: 99–106. [Google Scholar] [CrossRef]
  124. Hermes, Michael, Frank Albers, Jan R. Böhnke, Gerrit Huelmann, Julia Maier, and Dirk Stelling. 2019. Measurement and structural invariance of cognitive ability tests after computer-based training. Computers in Human Behavior 93: 370–78. [Google Scholar] [CrossRef]
  125. Hermes, Michael, Julia Maier, Justin Mittelstädt, Frank Albers, Gerrit Huelmann, and Dirk Stelling. 2023. Computer-based training and repeated test performance: Increasing assessment fairness instead of retest effects. European Journal of Work and Organizational Psychology 32: 450–59. [Google Scholar] [CrossRef]
  126. Heston, Thomas F., and Charya Khun. 2023. Prompt engineering in medical education. International Medical Education 2: 198–205. [Google Scholar] [CrossRef]
  127. Hickman, Luis, Patrick D. Dunlop, and Jasper L. Wolf. 2024. The performance of large language models on quantitative and verbal ability tests: Initial evidence and implications for unproctored high-stakes testing. International Journal of Selection and Assessment 32: 499–511. [Google Scholar] [CrossRef]
  128. Hines, Scott. 2017. The Development and Validation of an Automatic-Item Generation Measure of Cognitive Ability. Ph.D. dissertation, Louisiana Tech University, Ruston, LA, USA. Available online: https://digitalcommons.latech.edu/dissertations/71 (accessed on 3 March 2025).
  129. Holling, Heinz, Jonas P. Bertling, and Nina Zeuch. 2009. Automatic item generation of probability word problems. Studies in Educational Evaluation 35: 71–76. [Google Scholar] [CrossRef]
  130. Holmes, Stephen D., Michelle Meadows, Ian Stockford, and Qingping He. 2018. Investigating the comparability of examination difficulty using comparative judgement and Rasch modelling. International Journal of Testing 18: 366–91. [Google Scholar] [CrossRef]
  131. Hornke, Lutz F., and Michael W. Habon. 1986. Rule-based item bank construction and evaluation within the linear logistic framework. Applied Psychological Measurement 10: 369–80. [Google Scholar] [CrossRef]
  132. Impara, James C., and Barbara S. Plake. 1998. Teachers’ ability to estimate item difficulty: A test of the assumptions in the Angoff standard setting method. Journal of Educational Measurement 35: 69–81. [Google Scholar] [CrossRef]
  133. Irvine, Sidney H. 2002. The foundations of item generation for mass testing. In Item Generation for Test Development. Edited by Sidney H. Irvine and Patrick C. Kyllonnen. Mahwah: Lawrence Erlbaum Associates, pp. 3–34. [Google Scholar]
  134. Irvine, Sidney H., and Patrick C. Kyllonen. 2002. Item Generation for Test Development. Mahwah: Lawrence Erlbaum Associates. [Google Scholar]
  135. Ivie, Jennifer L., and Susan E. Embretson. 2010. Cognitive process modeling of spatial ability: The assembling objects task. Intelligence 38: 324–35. [Google Scholar] [CrossRef]
  136. Jarosz, Andrew F., and Jennifer Wiley. 2012. Why does working memory capacity predict RAPM performance? A possible role of distraction. Intelligence 40: 427–38. [Google Scholar] [CrossRef]
  137. Joncas, Sébastien X., Christina St-Onge, Sylvie Bourque, and Paul Farand. 2018. Re-using questions in classroom-based assessment: An exploratory study at the undergraduate medical education level. Perspectives on Medical Education 7: 373–78. [Google Scholar] [CrossRef]
  138. Jozefowicz, Ralph F., Bruce M. Koeppen, Susan Case, Robert Galbraith, David Swanson, and Robert H. Glew. 2002. The quality of in-house medical school examinations. Academic Medicine 77: 156–61. [Google Scholar] [CrossRef] [PubMed]
  139. Kaller, Christoph P., Benjamin Rahm, Lena Köstering, and Josef M. Unterrainer. 2011. Reviewing the impact of problem structure on planning: A software tool for analyzing tower tasks. Behavioural Brain Research 216: 1–8. [Google Scholar] [CrossRef] [PubMed]
  140. Kamruzzaman, Mahammed, Hieu Nguyen, Nazmul Hassan, and Gene L. Kim. 2024. “A Woman is More Culturally Knowledgeable than A Man?”: The Effect of Personas on Cultural Norm Interpretation in LLMs. arXiv arXiv:2409.11636. [Google Scholar]
  141. Kapoor, Radhika, Sang T. Truong, Nick Haber, Maria A. Ruiz-Primo, and Benjamin W. Domingue. 2025. Prediction of item difficulty for reading comprehension items by creation of annotated item repository. arXiv arXiv:2502.20663. [Google Scholar]
  142. Kara, Basek Erdem, and Nuri Dogan. 2022. The effect of ratio of items indicating differential item functioning on computer adaptive and multi-stage tests. International Journal of Assessment Tools in Education 9: 682–96. [Google Scholar] [CrossRef]
  143. Karthikeyan, Sowmiya, Elizabeth O’Connor, and Wendy Hu. 2019. Barriers and facilitators to writing quality items for medical school assessments–a scoping review. BMC Medical Education 19: 123. [Google Scholar] [CrossRef]
  144. Kıyak, Yavuz S., and Andrzej A. Kononowicz. 2025. Using a hybrid of AI and template-based method in automatic item generation to create multiple-choice questions in medical education: Hybrid AIG. JMIR Formative Research 9: e65726. [Google Scholar] [CrossRef]
  145. Kıyak, Yavuz S., Andrzej A. Kononowicz, and Stanislaw Górski. 2024. Multilingual template-based automatic item generation for medical education supported by generative artificial intelligence models ChatGPT and Claude. Bio-Algorithms and Med-Systems 20: 81–89. [Google Scholar] [CrossRef]
  146. Kıyak, Yavuz S., Emre Emekli, Özlem Coşkun, and Işil İ Budakoğlu. 2025. Keeping humans in the loop efficiently by generating question templates instead of questions using AI: Validity evidence on Hybrid AIG. Medical Teacher 47: 744–47. [Google Scholar] [CrossRef]
  147. Klahr, David, and Brian MacWhinney. 1997. Information Processing. In Cognition, Perception, and Language. Handbook of Child Psychology, 5th ed. Edited by William Damon, Deanna Kuhn and Robert Siegler. Hoboken: John Wiley and Sons, vol. 2, pp. 631–78. [Google Scholar]
  148. Kosh, Audra E., Mary A. Simpson, Lisa Bickel, Mark Kellogg, and Ellie Sanford-Moore. 2019. A cost–benefit analysis of automatic item generation. Educational Measurement: Issues and Practice 38: 48–53. [Google Scholar] [CrossRef]
  149. Krautter, Kai, Jessica Lehmann, Eva Kleinort, Marco Koch, Frank M. Spinath, and Nicolas Becker. 2021. Test preparation in figural matrices tests: Focus on the difficult rules. Frontiers in Psychology 12: 619440. [Google Scholar] [CrossRef] [PubMed]
  150. Kulik, James A., Chen-Lin C. Kulik, and Robert L. Bangert. 1984a. Effects of practice on aptitude and achievement test scores. American Educational Research Journal 21: 435–47. [Google Scholar] [CrossRef]
  151. Kulik, James A., Robert L. Bangert-Drowns, and Chen-Lin C. Kulik. 1984b. Effectiveness of coaching for aptitude tests. Psychological Bulletin 95: 179–88. [Google Scholar] [CrossRef]
  152. Kurdi, Ghader, Jared Leo, Bijan Parsia, Uli Sattler, and Salam Al-Emari. 2020. A systematic review of automatic question generation for educational purposes. International Journal of Artificial Intelligence in Education 30: 121–204. [Google Scholar] [CrossRef]
  153. LaDuca, Anthony, William I. Staples, Bryce Templeton, and Gerald B. Holzman. 1986. Item modelling procedures for constructing content-equivalent multiple-choice questions. Medical Education 20: 53–56. [Google Scholar] [CrossRef]
  154. Lai, Hollis, Mark J. Gierl, Claire Touchie, Debra Pugh, André-Philippe Boulais, and André De Champlain. 2016. Using automatic item generation to improve the quality of MCQ distractors. Teaching and Learning in Medicine 28: 166–73. [Google Scholar] [CrossRef]
  155. Lane, Suzanne, Mark Raymond, and Thomas Haladyna. 2016. Test development process. In Handbook of Test Development. Edited by Suzanne Lane, Mark Raymond and Thomas Haladyna. New York: Routledge, pp. 3–18. [Google Scholar]
  156. Lang, Jonas W. B. 2011. Computer-adaptives Testen. In Enzyklopädie für Psychologie: Verfahren zur Leistungs-, Intelligenz- und Verhaltensdiagnostik. Edited by Lutz F. Hornke, Manfred Amelang and Martin Kersting. Göttingen: Hogrefe, pp. 405–46. [Google Scholar]
  157. Lee, Hye Y., So J. Yune, Sang Y. Lee, Sunju Im, and Bee S. Kam. 2024. The impact of repeated item development training on the prediction of medical faculty members’ item difficulty index. BMC Medical Education 24: 599. [Google Scholar] [CrossRef]
  158. Lee, Jennifer C., Natasha Quadlin, and Denise Ambriz. 2023a. Shadow education, pandemic style: Social class, race, and supplemental education during COVID-19. Research in Social Stratification and Mobility 83: 100755. [Google Scholar] [CrossRef]
  159. Lee, Jooyoung, Thai Le, Jinghui Chen, and Dongwon Lee. 2022. Do language models plagiarize? arXiv arXiv:2203.07618. [Google Scholar]
  160. Lee, Philseok, Shea Fyffe, Mina Son, Zihao Jia, and Ziyu Yao. 2023b. A paradigm shift from “human writing” to “machine generation” in personality test development: An application of state-of-the-art natural language processing. Journal of Business and Psychology 38: 163–90. [Google Scholar] [CrossRef]
  161. Lee, Unggi, Haewon Jung, Younghoon Jeon, Younghoon Sohn, Wonhee Hwang, Jewoong Moon, and Hyeoncheol Kim. 2023c. Few-shot is enough: Exploring ChatGPT prompt engineering method for automatic question generation in English education. Education and Information Technologies 29: 11483–515. [Google Scholar] [CrossRef]
  162. Leslie, Tara, and Mark J. Gierl. 2023. Using automatic item generation to create multiple-choice questions for pharmacy assessment. American Journal of Pharmaceutical Education 87: 100081. [Google Scholar] [CrossRef] [PubMed]
  163. Levacher, Julie, Marco Koch, Johanna Hissbach, Frank M. Spinath, and Nicolas Becker. 2021. You can play the game without knowing the rules-but you’re better off knowing them: The influence of rule knowledge on figural matrices tests. European Journal of Psychological Assessment 38: 15–23. [Google Scholar] [CrossRef]
  164. Li, Kunze, and Yu Zhang. 2024. Planning first, question second: An LLM-guided method for controllable question generation. In Findings of the Association for Computational Linguistics ACL 2024. Bangkok: Association for Computational Linguistics, pp. 4715–29. [Google Scholar]
  165. Lievens, Filip, Charlie L. Reeve, and Eric D. Heggestad. 2007. An examination of psychometric bias due to retesting on cognitive ability tests in selection settings. Journal of Applied Psychology 92: 1672–82. [Google Scholar] [CrossRef]
  166. Lievens, Filip, Tine Buyse, and Paul R. Sackett. 2005. Retest effects in operational selection settings: Development and test of a framework. Personnel Psychology 58: 981–1007. [Google Scholar] [CrossRef]
  167. Lilly, Jane, and Paul Montgomery. 2011. Systematic reviews of the effects of preparatory courses on university entrance examinations in high school-age students. International Journal of Social Welfare 21: 3–12. [Google Scholar] [CrossRef]
  168. Lim, Sangdon, and Seung W. Choi. 2024. Item exposure and utilization control methods for optimal test assembly. Behaviormetrika 51: 125–56. [Google Scholar] [CrossRef]
  169. Lin, Zhiqing, and Huilin Chen. 2024. Investigating the capability of ChatGPT for generating multiple-choice reading comprehension items. System 123: 103344. [Google Scholar] [CrossRef]
  170. Liu, Cheng, Kyung T. Han, and Jun Li. 2019. Compromised item detection for computerized adaptive testing. Frontiers in Psychology 10: 1–16. [Google Scholar] [CrossRef]
  171. Liu, Mingxin, Tsuyoshi Okuhara, XinYi Chang, Ritsuko Shirabe, Yuriko Nishiie, Hiroko Okada, and Takahiro Kiuchi. 2024. Performance of ChatGPT across different versions in medical licensing examinations worldwide: Systematic review and meta-analysis. Journal of Medical Internet Research 26: e60807. [Google Scholar] [CrossRef] [PubMed]
  172. Liu, Pengfei, Weizhe Yuan, Jinlan Fu, Zhengbao Jiang, Hiroaki Hayashi, and Graham Neubig. 2023a. Pre-train, prompt, and predict: A systematic survey of prompting methods in natural language processing. ACM Computing Surveys 55: 1–35. [Google Scholar] [CrossRef]
  173. Liu, Yaohui, Keren He, Kaiwen Man, and Peida Zhan. 2025. Exploring critical eye-tracking metrics for identifying cognitive strategies in Raven’s Advanced Progressive Matrices: A data-driven perspective. Journal of Intelligence 13: 14. [Google Scholar] [CrossRef] [PubMed]
  174. Liu, Yaohui, Peida Zhan, Yanbin Fu, Qipeng Chen, Kaiwen Man, and Yikun Luo. 2023b. Using a multi-strategy eye-tracking psychometric model to measure intelligence and identify cognitive strategy in Raven’s advanced progressive matrices. Intelligence 100: 101782. [Google Scholar] [CrossRef]
  175. Loesche, Patrick, Jennifer Wiley, and Marcus Hasselhorn. 2015. How knowing the rules affects solving the Raven Advanced Progressive Matrices Test. Intelligence 48: 58–75. [Google Scholar] [CrossRef]
  176. Lu, Pan, Liang Qiu, Wenhao Yu, Sean Welleck, and Kai-Wei Chang. 2022. A survey of deep learning for mathematical reasoning. arXiv arXiv:2212.10535. [Google Scholar]
  177. Luca, Massimiliano, Ciro Beneduce, Bruno Lepri, and Jacopo Staiano. 2025. The LLM wears Prada: Analysing gender bias and stereotypes through online shopping data. arXiv arXiv:2504.01951. [Google Scholar]
  178. Luecht, Richard M. 2005. Some useful cost-benefit criteria for evaluating computer-based test delivery models and systems. Association of Test Publishers Journal 7. Available online: http://jattjournal.net/index.php/atp/article/view/48338 (accessed on 3 March 2025).
  179. Matteucci, Mariagiulia, Stefania Mignani, and Bernard P. Veldkamp. 2012. The use of predicted values for item parameters in item response theory models: An application in intelligence tests. Journal of Applied Statistics 39: 2665–83. [Google Scholar] [CrossRef]
  180. Matton, Nadine, Stéphane Vautier, and Éric Raufaste. 2011. Test-specificity of the advantage of retaking cognitive ability tests. International Journal of Selection and Assessment 19: 11–17. [Google Scholar] [CrossRef]
  181. McCoy, R. Thomas, Paul Smolensky, Tal Linzen, Jianfeng Gao, and Asli Celikyilmaz. 2023. How much do language models copy from their training data? evaluating linguistic novelty in text generation using raven. Transactions of the Association for Computational Linguistics 11: 652–70. [Google Scholar] [CrossRef]
  182. Messick, Samuel. 1982. Issues of effectiveness and equity in the coaching controversy: Implications for educational testing and practice. Educational Psychologist 17: 67–91. [Google Scholar] [CrossRef]
  183. Messick, Samuel, and Ann Jungeblut. 1981. Time and method in coaching for the SAT. Psychological Bulletin 89: 191–216. [Google Scholar] [CrossRef]
  184. Mislevy, Robert J., and Michelle M. Riconscente. 2006. Evidence-centered assessment design: Layers, concepts, and terminology. In Handbook of Test Development. Edited by Steven Downing and Thomas Haladyna. Mahwah: Lawrence Erlbaum Associates, pp. 61–90. [Google Scholar]
  185. Mislevy, Robert J., Kathleen M. Sheehan, and Marilyn Wingersky. 1993. How to equate tests with little or no data. Journal of Educational Measurement 30: 55–78. [Google Scholar] [CrossRef]
  186. Mislevy, Robert J., Russell G. Almond, and Janice F. Lukas. 2003. A Brief Introduction to Evidence-Centered Design (Research Report: RR-03-16). Princeton: Educational Testing Service. [Google Scholar]
  187. Morley, Mary E., Brent Bridgeman, and René R. Lawless. 2004. Transfer Between Variants of Quantitative Items (GRE Board Rep. No. 00-06R). Princeton: ETS. [Google Scholar]
  188. Nemec, Eric C., and Beth Welch. 2016. The impact of a faculty development seminar on the quality of multiple-choice questions. Currents in Pharmacy Teaching and Learning 8: 160–63. [Google Scholar] [CrossRef]
  189. OpenAI. 2023. GPT-4 technical report. arXiv arXiv:2303.08774. [Google Scholar]
  190. O’Reilly, Tenaha, Gary Feng, John Sabatini, Zuowei Wang, and Joanna Gorin. 2018. How do people read the passages during a reading comprehension test? The effect of reading purpose on text processing behavior. Educational Assessment 23: 277–95. [Google Scholar] [CrossRef]
  191. Park, Julie J., and Ann H. Becks. 2015. Who benefits from SAT prep? An examination of high school context and race/ethnicity. The Review of Higher Education 39: 1–23. [Google Scholar] [CrossRef]
  192. Piromsombat, Chayut. 2014. Differential Item Functioning in Computerized Adaptive Testing: Can CAT Self-Adjust Enough? (Publication No. 3620715). Doctoral dissertation, University of Minnesota, Minneapolis, MN, USA. [Google Scholar]
  193. Powers, Donald E. 2005. Effects of Pre-Examination Disclosure of Essay Prompts for the GRE Analytical Writing Assessment (Research Report: RR-05–01). Princeton: Educational Testing Service. [Google Scholar]
  194. Powers, Donald E. 2012. Understanding the Impact of Special Preparation for Admissions Tests. In Advancing Human Assessment: The Methodological, Psychological and Policy Contributions of ETS. ETS Research Report Series; Cham: Springer International Publishing. [Google Scholar]
  195. Powers, Donald E., and Donald A. Rock. 1999. Effects of coaching on SAT I: Reasoning scores. Journal of Educational Measurement 36: 93–118. [Google Scholar] [CrossRef]
  196. Powers, Donald E., and Donald L. Alderman. 1983. Effects of test familiarization on SAT performance. Journal of Educational Measurement 20: 71–79. [Google Scholar] [CrossRef]
  197. Primi, Ricardo. 2002. Complexity of geometric inductive reasoning tasks: Contribution to the understanding of fluid intelligence. Intelligence 30: 41–70. [Google Scholar] [CrossRef]
  198. Primi, Ricardo. 2014. Developing a fluid intelligence scale through a combination of Rasch modeling and cognitive psychology. Psychological Assessment 26: 774–88. [Google Scholar] [CrossRef] [PubMed]
  199. Raffel, Colin, Noam Shazeer, Adam Roberts, Katherine Lee, Sharan Narang, Michael Matena, and Peter J. Liu. 2020. Exploring the limits of transfer learning with a unified text-to-text transformer. The Journal of Machine Learning Research 21: 5485–551. [Google Scholar]
  200. Rajeb, Mehdi, Andrew T. Krist, Qingzhou Shi, Daniel O. Oyeniran, Stefanie A. Wind, and Joni M. Lakin. 2024. Mental rotation performance: Contribution of item features to difficulties and functional adaptation. Journal of Intelligence 13: 2. [Google Scholar] [CrossRef]
  201. Ranjan, Rajesh, Shailja Gupta, and Saranyan N. Singh. 2024. Gender Biases in LLMs: Higher intelligence in LLM does not necessarily solve gender bias and stereotyping. arXiv arXiv:2409.19959. [Google Scholar]
  202. Reckase, Mark D. 2010. Designing item pools to optimize the functioning of a computerized adaptive test. Psychological Test and Assessment Modeling 52: 127–41. [Google Scholar]
  203. Reckase, Mark D., Unhee Ju, and Sewon Kim. 2019. How adaptive is an adaptive test: Are all adaptive tests adaptive? Journal of Computerized Adaptive Testing 7: 1–14. [Google Scholar] [CrossRef]
  204. Reeve, Charlie L., and Holly Lam. 2005. The psychometric paradox of practice effects due to retesting: Measurement invariance and stable ability estimates in the face of observed score changes. Intelligence 33: 535–49. [Google Scholar] [CrossRef]
  205. Ren, Xuezhu, Frank Goldhammer, Helfried Moosbrugger, and Karl Schweizer. 2012. How does attention relate to the ability-specific and position-specific components of reasoning measured by APM? Learning and Individual Differences 22: 1–7. [Google Scholar] [CrossRef]
  206. Reynolds, Laria, and Kyle McDonell. 2021. Prompt programming for large language models: Beyond the few-shot paradigm. In Extended Abstracts of the 2021 CHI Conference on Human Factors in Computing Systems. New York: Association for Computing Machinery, pp. 1–7. [Google Scholar]
  207. Reza, Mohi, Ioannis Anastasopoulos, Shreya Bhandari, and Zachary A. Pardos. 2024. PromptHive: Bringing subject matter experts back to the forefront with collaborative prompt engineering for educational content creation. arXiv arXiv:2410.16547. [Google Scholar]
  208. Riedel, Maximilian, Katharina Kaefinger, Antonia Stuehrenberg, Viktoria Ritter, Niklas Amann, Anna Graf, Florian Recker, Evelyn Klein, Marion Kiechle, Fabian Riedel, and et al. 2023. ChatGPT’s performance in German OB/GYN exams—Paving the way for AI-enhanced medical education and clinical practice. Frontiers in Medicine 10: 129661. [Google Scholar] [CrossRef]
  209. Rogausch, Anja, Rainer Hofer, and René Krebs. 2010. Rarely selected distractors in high stakes medical multiple-choice examinations and their recognition by item authors: A simulation and survey. BMC Medical Education 10: 85. [Google Scholar] [CrossRef]
  210. Roid, Gale H., and Thomas M. Haladyna. 1982. Toward a Technology of Test-Item Writing. New York: Academic. [Google Scholar]
  211. Runge, Andrew, Yigal Attali, Geoffrey T. LaFlair, Yena Park, and Jaqueline Church. 2024. A generative AI-driven interactive listening assessment task. Frontiers in Artificial Intelligence 7: 1474019. [Google Scholar] [CrossRef]
  212. Ryoo, Ji H., Sunhee Park, Hongwook Suh, Jaehwa Choi, and Jongkyum Kwon. 2022. Development of a new measure of cognitive ability using automatic item generation and its psychometric properties. SAGE Open 12: 1–13. [Google Scholar] [CrossRef]
  213. Sahin, Alper, and Duygu Anil. 2017. The effects of test length and sample size on item parameters in item response theory. Educational Science: Theory and Practice 17: 321–35. [Google Scholar] [CrossRef]
  214. Sahin Kursad, Merve, and Seher Yalcin. 2024. Effect of differential item functioning on computer adaptive testing under different conditions. Applied Psychological Measurement 48: 303–22. [Google Scholar] [CrossRef] [PubMed]
  215. Sahoo, Pranab, Ayush K. Singh, Sriparna Saha, Vinija Jain, Samrat Mondal, and Aman Chadha. 2024. A systematic survey of prompt engineering in large language models: Techniques and applications. arXiv arXiv:2402.07927. [Google Scholar]
  216. Sayin, Ayfer, and Mark J. Gierl. 2023. Automatic item generation for online measurement and evaluation: Turkish literature items. International Journal of Assessment 10: 218–31. [Google Scholar] [CrossRef]
  217. Sayin, Ayfer, and Marl Gierl. 2024. Using OpenAI GPT to generate reading comprehension items. Educational Measurement: Issues and Practice 43: 5–18. [Google Scholar] [CrossRef]
  218. Sayın, Ayfer, and Okan Bulut. 2024. The difference between estimated and perceived item difficulty: An empirical study. International Journal of Assessment Tools in Education 11: 368–87. [Google Scholar] [CrossRef]
  219. Sayın, Ayfer, and Sebahat Gören. 2023. Comparing estimated and real item difficulty using multi-facet Rasch analysis. Journal of Measurement and Evaluation in Education and Psychology 14: 440–54. [Google Scholar] [CrossRef]
  220. Säuberli, Andreas, and Simon Clematide. 2024. Automatic generation and evaluation of reading comprehension test items with large language models. arXiv arXiv:2404.07720. [Google Scholar]
  221. Scharfen, Jana, Judith M. Peters, and Heinz Holling. 2018. Retest effects in cognitive ability tests: A meta-analysis. Intelligence 67: 44–66. [Google Scholar] [CrossRef]
  222. Schneider, Benedikt, and Jörn R. Sparfeldt. 2021a. How to get better: Taking notes mediates the effect of a video tutorial on number series. Journal of Intelligence 9: 55. [Google Scholar] [CrossRef]
  223. Schneider, Benedikt, and Jörn R. Sparfeldt. 2021b. How to solve number series items: Can watching video tutorials increase test scores? Intelligence 87: 101547. [Google Scholar] [CrossRef]
  224. Schneider, Benedikt, Nicolas Becker, Florian Krieger, Frank M. Spinath, and Jörn R. Sparfeldt. 2020. Teaching the underlying rules of figural matrices in a short video increases test scores. Intelligence 82: 101473. [Google Scholar] [CrossRef]
  225. Schroeders, Ulrich, and Priscilla Achaa-Amankwaa. 2025. Developing NOVA: Next-generation open vocabulary assessment. Unpublished manuscript. [Google Scholar]
  226. Schroeders, Ulrich, and Timo Gnambs. 2025. Sample-size planning in item-response theory: A tutorial. Advances in Methods and Practices in Psychological Science 8: 25152459251314798. [Google Scholar] [CrossRef]
  227. Schulhoff, Sander, Michael Ilie, Nishant Balepur, Konstantine Kahadze, Amanda Liu, Chenglei Si, and Philip Resnik. 2024. The prompt report: A systematic survey of prompting techniques. arXiv arXiv:2406.06608. [Google Scholar]
  228. Schulze Balhorn, Lukas, Jana M. Weber, Stefan Buijsman, Julian R. Hildebrandt, Martina Ziefle, and Artur M. Schweidtmann. 2024. Empirical assessment of ChatGPT’s answering capabilities in natural science and engineering. Scientific Reports 14: 4998. [Google Scholar] [CrossRef]
  229. Segall, Daniel O. 2004. A sharing item response theory model for computerized adaptive testing. Journal of Educational and Behavioral Statistics 29: 439–60. [Google Scholar] [CrossRef]
  230. Selvi, Hüseyin. 2020. Should items and answer keys of small-scale exams be published? Higher Education Studies 10: 107–13. [Google Scholar] [CrossRef]
  231. Shi, Qingzhou, Stefanie A. Wind, and Joni M. Lakin. 2023. Exploring the influence of item characteristics in a spatial reasoning task. Journal of Intelligence 11: 152. [Google Scholar] [CrossRef] [PubMed]
  232. Shin, Jinnie, and Mark J. Gierl. 2022. Generating reading comprehension items using automated processes. International Journal of Testing 22: 289–311. [Google Scholar] [CrossRef]
  233. Shultz, Benjamin, Robert J. DiDomenico, Kristen Goliak, and Jeffrey Mucksavage. 2025. Exploratory assessment of GPT-4’s effectiveness in generating valid exam items in pharmacy education. American Journal of Pharmaceutical Education 89: 101405. [Google Scholar] [CrossRef]
  234. Siegler, Robert S. 1996. Emerging Minds: The Process of Change in Children’s Thinking. New York: Oxford University Press. [Google Scholar]
  235. Sinharay, Sandip. 2017. Which statistic should be used to detect item pre-knowledge when the set of compromised items is known? Applied Psychological Measurement 41: 403–21. [Google Scholar] [CrossRef]
  236. Sinharay, Sandip, and Matthew S. Johnson. 2008. Use of item models in a large-scale admissions test: A case study. International Journal of Testing 8: 209–36. [Google Scholar] [CrossRef]
  237. Sinharay, Sandip, Matthew S. Johnson, and David M. Williamson. 2003. Calibrating item families and summarizing the results using family expected response functions. Journal of Educational and Behavioral Statistics 28: 295–313. [Google Scholar] [CrossRef]
  238. Smaldino, Paul E. 2020. How to build a strong theoretical foundation. Psychological Inquiry 31: 297–301. [Google Scholar] [CrossRef]
  239. Sobieszek, Adam, and Tadeusz Price. 2022. Playing games with AIs: The limits of GPT-3 and similar large language models. Minds and Machines 32: 341–64. [Google Scholar] [CrossRef]
  240. Someshwar, Shonai. 2024. Quality Control and the Impact of Variation and Prediction Errors on Item Family Design. Doctoral dissertation, The University of North Carolina at Greensboro, Greensboro, NC, USA. [Google Scholar]
  241. Sommer, Markus, Margit Herle, Joachim Häusler, and Martin Arendasy. 2009. Von TAVTMB zu ATAVT: Eine Anwendung der Automatisierten Itemgenerierung unter einschränkenden Rahmenbedingungen. In Zweites Österreichisches Symposium für Psychologie im Militär. Edited by Georg Ebner and Günther Fleck. Wien: Schriftreihe der Landesverteidigungsakademie, pp. 27–52. [Google Scholar]
  242. Sommer, Markus, Martin E. Arendasy, Joachim F. Punter, Martina Feldhammer-Kahr, and Anita Rieder. 2025. Does test preparation mediate the effect of parents’ level of educational attainment on medical school admission test performance? Intelligence 108: 101893. [Google Scholar] [CrossRef]
  243. Song, Yishen, Junlei Du, and Qinhua Zheng. 2025. Automatic item generation for educational assessments: A systematic literature review. Interactive Learning Environments, 1–20. [Google Scholar] [CrossRef]
  244. Stricker, Lawrence J. 1984. Test disclosure and retest performance on the SAT. Applied Psychological Measurement 8: 81–87. [Google Scholar] [CrossRef]
  245. Su, Mei-Chin, Li-En Lin, Li-Hwa Lin, and Yu-Chun Chen. 2024. Assessing question characteristic influences on ChatGPT’s performance and response-explanation consistency: Insights from Taiwan’s nursing licensing exam. International Journal of Nursing Studies 153: 104717. [Google Scholar] [CrossRef] [PubMed]
  246. Sun, Luning, Yanan Liu, and Fang Luo. 2019. Automatic generation of number series reasoning items of high difficulty. Frontiers in Psychology 10: 884. [Google Scholar] [CrossRef]
  247. Svetina, Dubravka, Joanna Gorin, and Kikumi K. Tatsuoka. 2011. Defining and comparing the reading comprehension construct: A cognitive-psychometric modeling approach. International Journal of Testing 11: 1–23. [Google Scholar] [CrossRef]
  248. Sydorenko, Tetyana. 2011. Item writer judgments of item difficulty versus actual item difficulty: A case study. Language Assessment Quarterly 8: 34–52. [Google Scholar] [CrossRef]
  249. Tan, Bin, Nour Armoush, Elisabetta Mazzullo, Okan Bulut, and Mark J. Gierl. 2024. A review of automatic item generation techniques leveraging large language models. EdArXiv. [Google Scholar] [CrossRef]
  250. te Nijenhuis, Jan, Annelies E. M. Vianen, and Henk van der Flier. 2007. Score gains on g-loaded tests: No g. Intelligence 35: 283–300. [Google Scholar] [CrossRef]
  251. Thakur, Vishesh. 2023. Unveiling gender bias in terms of profession across LLMs: Analyzing and addressing sociological implications. arXiv arXiv:2307.09162. [Google Scholar]
  252. Tian, Chen, and Jaehwa Choi. 2023. The impact of item model parameter variations on person parameter estimation in computerized adaptive testing with automatically generated items. Applied Psychological Measurement 47: 275–90. [Google Scholar] [CrossRef]
  253. Touvron, Hugo, Thibaut Lavril, Gautier Izacard, Xavier Martinet, Marie-Anne Lachaux, Thimothée Lacroix, and Guillaume Lample. 2023. Llama: Open and efficient foundation language models. arXiv arXiv:2302.13971. [Google Scholar]
  254. van der Linden, Wim J., and Cees A. Glas. 2010. Elements of Adaptive Testing. New York: Springer. [Google Scholar]
  255. van der Maas, Han L., Lukas Snoek, and Claire E. Stevenson. 2021. How much intelligence is there in artificial intelligence? A 2020 update. Intelligence 87: 101548. [Google Scholar] [CrossRef]
  256. Vaswani, Ashish, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N. Gomez, Łukasz U. Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in Neural Information Processing Systems. Edited by Isabelle Guyon, Ulrike von Luxburg, Samy Bengio, Hanna Wallach, Rob Fergus, Vichy SVN Vishwanathan and Roman Garnett. Red Hook: Curran Associates, pp. 5998–6008. [Google Scholar]
  257. Veerkamp, Wim J., and Cees A. W. Glas. 2000. Detection of known items in adaptive testing with a statistical quality control method. Journal of Educational and Behavioral Statistics 25: 373–89. [Google Scholar] [CrossRef]
  258. Veldkamp, Bernard P., and Wim J. van der Linden. 2010. Designing item pools for adaptive testing. In Computerized Adaptive Testing: Theory and Practice. Edited by Wim J. van der Linden and Cees A. W. Glas. New York: Springer, pp. 149–62. [Google Scholar]
  259. Verguts, Tom, and Paul De Boeck. 2002. The induction of solution rules in Raven’s Progressive Matrices. European Journal of Cognitive Psychology 14: 521–47. [Google Scholar] [CrossRef]
  260. Vigneau, François, André F. Caissie, and Douglas A. Bors. 2006. Eye-movement analysis demonstrates strategic influences on intelligence. Intelligence 34: 261–72. [Google Scholar] [CrossRef]
  261. von Davier, Matthias. 2018. Automated item generation with recurrent neural networks. Psychometrika 83: 847–57. [Google Scholar] [CrossRef]
  262. von Davier, Matthias. 2019. Training Optimus prime, M.D.: Generating medical certification items by fine-tuning OpenAI’s gpt2 transformer model. arXiv arXiv:1908.08594. [Google Scholar]
  263. Wagner-Menghin, Michaela, Ingrid Preusche, and Michael Schmidts. 2013. The effects of reusing written test items: A study using the Rasch model. ISRN Education 2013: 585420. [Google Scholar] [CrossRef]
  264. Wainer, Howard. 2002. On the automatic generation of items: Some whens, whys and hows. In Item Generation for Test Development. Edited by Sidney H. Irvine and Patrick C. Kyllonen. Mahwah: Lawrence Erlbaum, pp. 287–316. [Google Scholar]
  265. Waldock, William J., Joe Zhang, Ahmad Guni, Ahmad Nabeel, Ara Darzi, and Hutan Ashrafian. 2024. The accuracy and capability of artificial intelligence solutions in health care examinations and certificates: Systematic review and meta-analysis. Journal of Medical Internet Research 26: e56532. [Google Scholar] [CrossRef]
  266. Wancham, Kittitas, Kamonwan Tangdhanakanond, and Sirichai Kanjanawasee. 2023. Development of the automatic item generation system for the diagnosis of misconceptions about force and laws of motion. Eurasia Journal of Mathematics, Science and Technology Education 19: em2282. [Google Scholar] [CrossRef]
  267. Wang, Yi, Qian Zhou, and David Ledo. 2024. StoryVerse: Towards co-authoring dynamic plot with LLM-based character simulation via narrative planning. Paper presented at 19th International Conference on the Foundations of Digital Games, Worcester, MA, USA, May 21–24. [Google Scholar]
  268. Webb, Emily M., Jonathan S. Phuong, and David M. Naeger. 2015. Does educator training or experience affect the quality of multiple-choice questions? Academic Radiology 22: 1317–22. [Google Scholar] [CrossRef]
  269. Weppert, Daniel, Dorothee Amelung, Malvin Escher, Leander Troll, Martina Kadmon, Lena Listunova, and Jana Montasser. 2023. The impact of preparatory activities on the largest clinical aptitude test for prospective medical students in Germany. Frontiers in Education 8: 1104464. [Google Scholar] [CrossRef]
  270. Witt, Elizabeth A. 1993. Meta-analysis and the effects of coaching for aptitude tests. Paper presented at the Annual Meeting of the American Educational research Association, Atlanta, GA, USA, April 12–16. [Google Scholar]
  271. Wonde, Shewatatek G., Tefera Tadesse, Belay Moges, and Stefan K. Schauber. 2024. Experts’ prediction of item difficulty of multiple-choice questions in the Ethiopian Undergraduate Medicine Licensure Examination. BMC Medical Education 24: 1016. [Google Scholar] [CrossRef]
  272. Wood, Timothy J. 2009. The effect of reused questions on repeat examinees. Advances in Health Sciences Education 14: 465–73. [Google Scholar] [CrossRef]
  273. Wood, Timothy J., Christina St-Onge, André-Philippe Boulais, David E. Blackmore, and Thomas O. Maguire. 2010. Identifying the unauthorized use of examination material. Evaluation and the Health Professions 33: 96–108. [Google Scholar] [CrossRef]
  274. Yang, Eunbae B., Myung A. Lee, and Yoon S. Park. 2018. Effects of test item disclosure on medical licensing examination. Advances in Health Sciences Education 23: 265–74. [Google Scholar] [CrossRef] [PubMed]
  275. Yang, Jingfeng, Hongye Jin, Ruixiang Tang, Xiaotian Han, Qizhang Feng, Haoming Jiang, and Xia Hu. 2024. Harnessing the power of LLMs in practice: A survey on ChatGPT and beyond. ACM Transactions on Knowledge Discovery from Data 18: 1–32. [Google Scholar] [CrossRef]
  276. Yang, Yuan, and Mathilee Kunda. 2023. Computational models of solving Raven’s Progressive Matrices: A comprehensive introduction. arXiv arXiv:2302.04238. [Google Scholar]
  277. Yang, Yuan, Deepayan Sanyal, Joel Michelson, James Ainooson, and Mathilee Kunda. 2022. Automatic item generation of figural analogy problems: A review and outlook. arXiv arXiv:2201.08450. [Google Scholar]
  278. Yi, Qing, Jinming Zhang, and Hua-Hua Chang. 2008. Severity of organized item theft in computerized adaptive testing: A simulation study. Applied Psychological Measurement 32: 543–58. [Google Scholar] [CrossRef]
  279. Yu, Jiayuan. 1994. Homogenity of problem solving strategies and the fitting of linear logistic model. Acta Psychologica Sinica 26: 219–24. [Google Scholar]
  280. Zenisky, April, Ronald K. Hambleton, and Richard M. Luecht. 2010. Multistage testing: Issues, designs, and research. In Elements of Adaptive Testing. Edited by Wim J. van der Linden and Cees A. Glas. New York: Springer, pp. 355–72. [Google Scholar]
  281. Zha, Daochen, Zaid P. Bhat, Kwei-Herng Lai, Fan Yang, Zhimeng Jiang, Saochen Zhong, and Xiaben Hu. 2025. Data-centric artificial intelligence: A survey. ACM Computing Surveys 57: 1–42. [Google Scholar] [CrossRef]
  282. Zhang, Jinming, and Hua-Hua Chang. 2005. The Effectiveness of Enhancing Test Security by Using Multiple Item Pools (ETS RR-05-19). Princeton: ETS. [Google Scholar]
  283. Zhang, Jinming, Hua-Hua Chang, and Qing Yi. 2012. Comparing single-pool and multiple-pool designs regarding test security in computerized testing. Behavior Research Methods 44: 742–52. [Google Scholar] [CrossRef] [PubMed]
  284. Zickar, Michael J. 2020. Measurement development and evaluation. Annual Review of Organizational Psychology and Organizational Behavior 7: 213–32. [Google Scholar] [CrossRef]
  285. Zimmer, Felix, Mirka Henninger, and Rudolf Debelak. 2024. Sample size planning for complex study designs: A tutorial for the mlpwr package. Behavior Research Methods 56: 5246–63. [Google Scholar] [CrossRef] [PubMed]
  286. Zimmermann, Stefan, Dietrich Klusmann, and Wolfgang Hampe. 2016. Are exam questions known in advance? Using local dependence to detect cheating. PLoS ONE 11: e0167545. [Google Scholar] [CrossRef]
  287. Zorowitz, Samuel, Gabriele Chierchia, Sarah-Jayne Blakemore, and Nathaniel D. Daw. 2024. An item response theory analysis of the matrix reasoning item bank (MaRs-IB). Behavior Research Methods 56: 1104–22. [Google Scholar] [CrossRef]
  288. Zu, Jiyun, Ikkyu Choi, and Jiangang Hao. 2023. Automated distractor generation for fill-in-the-blank items using a prompt-based learning approach. Psychological Testing and Assessment Modeling 65: 55–75. Available online: https://www.psychologie-aktuell.com/fileadmin/Redaktion/Journale/ptam_2023-1/PTAM__1-2023_3_kor.pdf (accessed on 3 March 2025).
  289. Zwick, Rebecca. 2002. Is the SAT a ‘wealth test’? Phi Delta Kappan 84: 307–11. [Google Scholar] [CrossRef]
  290. Zwick, Rebecca, Dorothy T. Thayer, and Marilyn Wingersky. 1995. Effect of Rasch calibration on ability and DIF estimation in computer-adaptive tests. Journal of Educational Measurement 32: 341–63. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sommer, M.; Arendasy, M. Automatic- and Transformer-Based Automatic Item Generation: A Critical Review. J. Intell. 2025, 13, 102. https://doi.org/10.3390/jintelligence13080102

AMA Style

Sommer M, Arendasy M. Automatic- and Transformer-Based Automatic Item Generation: A Critical Review. Journal of Intelligence. 2025; 13(8):102. https://doi.org/10.3390/jintelligence13080102

Chicago/Turabian Style

Sommer, Markus, and Martin Arendasy. 2025. "Automatic- and Transformer-Based Automatic Item Generation: A Critical Review" Journal of Intelligence 13, no. 8: 102. https://doi.org/10.3390/jintelligence13080102

APA Style

Sommer, M., & Arendasy, M. (2025). Automatic- and Transformer-Based Automatic Item Generation: A Critical Review. Journal of Intelligence, 13(8), 102. https://doi.org/10.3390/jintelligence13080102

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop