Next Article in Journal
General Ability Level Moderates Cognitive–Achievement Relations for Mathematics
Previous Article in Journal
Log File Times as Indicators of Structured Figural Matrix Processing
Previous Article in Special Issue
Generative AI in Game Design: Enhancing Creativity or Constraining Innovation?
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Scoring German Alternate Uses Items Applying Large Language Models

1
Department of Psychology, University of Graz, 8010 Graz, Austria
2
Department of Psychology, Charlotte Fresenius Hochschule München, 80797 Munich, Germany
3
Department of Psychology, LMU Munich, 80802 Munich, Germany
4
GESIS—Leibniz Institute for the Social Sciences, 68159 Mannheim, Germany
5
Institute of Psychology, University of Münster, 48149 Münster, Germany
6
Hector Research Institute of Education Sciences and Psychology, University of Tübingen, 72072 Tübingen, Germany
7
Institute for Psychology, University of Hildesheim, 31141 Hildesheim, Germany
*
Authors to whom correspondence should be addressed.
J. Intell. 2025, 13(6), 64; https://doi.org/10.3390/jintelligence13060064
Submission received: 7 February 2025 / Revised: 15 May 2025 / Accepted: 20 May 2025 / Published: 29 May 2025
(This article belongs to the Special Issue Generative AI: Reflections on Intelligence and Creativity)

Abstract

:
The alternate uses task (AUT) is the most popular measure when it comes to the assessment of creative potential. Since their implementation, AUT responses have been rated by humans, which is a laborious task and requires considerable resources. Large language models (LLMs) have shown promising performance in automatically scoring AUT responses in English as well as in other languages, but it is not clear which method works best for German data. Therefore, we investigated the performance of different LLMs for the automated scoring of German AUT responses. We compiled German data across five research groups including ~50,000 responses for 15 different alternate uses objects from eight lab and online survey studies (including ~2300 participants) to examine generalizability across datasets and assessment conditions. Following a pre-registered analysis plan, we compared the performance of two fine-tuned, multilingual LLM-based approaches [Cross-Lingual Alternate Uses Scoring (CLAUS) and the Open Creativity Scoring with Artificial Intelligence (OCSAI)] with the Generative Pre-trained Transformer (GPT-4) in scoring (a) the original German AUT responses and (b) the responses translated to English. We found that the LLM-based scorings were substantially correlated with human ratings, with higher relationships for OCSAI followed by GPT-4 and CLAUS. Response translation, however, had no consistent positive effect. We discuss the generalizability of the results across different items and studies and derive recommendations and future directions.

1. Introduction

Responses in, scores out. No recruiting and training of human raters. No waiting for weeks of coding. How does that sound? Researchers interested in creative thinking have long dreamed of a psychometrically sound automated solution to score the creative quality of responses in divergent thinking (DT) tests. While early attempts of automated scoring yielded promising results (for an overview, see Forthmann and Doebler 2022; Paulus and Renzuli 1968), significant advancements have been made only recently with the advent of large language models (LLMs; Goecke et al. 2024; Organisciak et al. 2023; Patterson et al. 2024a; Zielińska et al. 2023). To date, most automated scoring approaches have been developed and tested using English-language data, particularly in prototypical DT tasks such as the alternate uses task (AUT; Guilford 1967; Wallach and Kogan 1965; Wilson et al. 1954). However, recent studies have begun to explore the feasibility of applying automated scoring methods to other languages, either by first translating responses into English or by making use of their increasing multilingual capacities (Forthmann and Doebler 2022; Patterson et al. 2024a; Stevenson et al. 2022; Zielińska et al. 2023). However, little is known about the specific conditions under which automated scoring performs best in languages such as German. Therefore, this work investigates the performance of three LLM-based approaches for creativity scoring of German AUT responses.

1.1. Assessment of Divergent Thinking

Individual differences in creative thinking ability are typically assessed using DT tasks, which require participants to generate creative ideas in response to open-ended questions (Guilford 1950; Torrance 1972). For example, in the highly popular AUT, participants are asked to come up with creative uses for everyday objects (e.g., a brick) within limited time (Saretzki et al. 2024a). Various approaches for scoring the creative quality of DT responses have been proposed, such as coding responses based on a test manual, computing their sample-based infrequency (e.g., uniqueness scoring), or evaluating them by human raters consistent with the Consensual Assessment Technique (Amabile 1982, 1996; Baer et al. 2004). The rater-based method offers several advantages, as it allows raters to simultaneously consider multiple dimensions of creative quality and enables a robust handling of unique responses that may not be covered by test manuals. However, this rater-based approach also comes with notable downsides. First, human judgements are inherently subjective (Mouchiroud and Lubart 2001; Silvia et al. 2008). Rater agreement can be influenced by individual characteristics and rating conditions—for instance, experienced raters tend to achieve higher inter-rater reliability (Ceh et al. 2022), while cognitive workload can lead to greater disagreement (Forthmann et al. 2017b). Additionally, rating consistency depends on the specific rating instructions, which often vary across research labs. Second, human ratings are time-consuming and costly. Studies employing multiple DT tasks easily yield a few thousands of responses, depending on factors such as sample size, task instructions (e.g., focusing on the quality or quantity of responses), and time on task. As a result, ratings of a larger pack of responses can delay the research process and become expensive (Forthmann et al. 2023). Therefore, an effective automated scoring system for DT tasks across different languages would greatly enhance the reproducibility and efficiency of creativity research.

1.2. Automated Scoring of Divergent Thinking Tasks

Early computational approaches to evaluating creativity date back to text-mining methods, which laid the foundations for more advanced techniques incorporating natural language processing. First attempts relied on linguistic variables such as elaboration (i.e., word count) as predictors for the creativity of ideas (Paulus and Renzuli 1968). Later, methods shifted focus toward assessing originality by measuring the semantic distance estimated based on co-occurrences within large text corpora, mathematically represented as vectors in high-dimensional spaces (Landauer et al. 1998)—between a response and the problem cue.
Several studies have demonstrated positive relationships between semantic distance scores and human creativity ratings, but correlations across responses were modest (Beaty and Johnson 2021; Dumas et al. 2021; Yu et al. 2023). For example, semantic distance scores were compared across data from 12 languages and correlations were found to range from r = 0.23 (for Hebrew) to r = 0.52 (for English; German: r = 0.41; Patterson et al. 2023).
Earlier automated scoring methods relying on semantic distance and word counts have recently been surpassed by supervised machine learning and Transformer Language Models (TLMs) for evaluating creative responses (Organisciak et al. 2023). Supervised learning involves training models on labeled datasets, enabling them to make predictions on new, unseen data. These models often leverage decoder-only transformer architectures—such as those used in Generative Pre-trained Transformer (GPT) models—and can be fine-tuned to better align with human evaluations, thereby achieving high correlations with human ratings and outperforming base models (Organisciak et al. 2023; Zielińska et al. 2023). Recent studies report that fine-tuned supervised models can reach response-level correlations of up to 0.70 or even 0.80 with human ratings, with even higher correlations for scores aggregated across responses and when considering latent correlations (DiStefano et al. 2024; Goecke et al. 2024; Luchini et al. 2025b). These correlations lead to the assumption that supervised models may perform similarly to human raters in creativity scoring—maybe even more consistently and certainly much faster.
Even though ratings by fine-tuned models show high correlations with human ratings for held-out data, comparisons of the performance for completely new study data remain scarce. Moreover, while fine-tuned multilingual models are rapidly advancing, empirical evidence on their performance across different languages is still rare. One study found that the Open Creativity Scoring with Artificial Intelligence (OCSAI), based on GPT-3, worked well for Polish AUT data, achieving response-level correlations ranging from r = 0.42 to r = 0.72. Notably, translating responses into English before scoring did not yield any substantial improvements (Zielińska et al. 2023). Similarly, another study showed that a fine-tuned multilingual LLM reliably predicted the originality of stories across 11 languages in the held-out data (r > 0.70), including German (r = 0.75; Luchini et al. 2025a). These results suggest that LLM-based scoring approaches are increasingly capable of handling non-English data with high accuracy. However, empirical evidence remains limited regarding how well these models generalize to new datasets, particularly for German AUT responses.

1.3. The Present Research

While previous studies have demonstrated the potential of LLM-based approaches for the automated scoring of creativity, research on their effectiveness in non-English languages remains limited. In particular, little is known about how well these models perform when scoring prototypical DT tasks like the AUT in languages like German. Additionally, it remains an open question whether translating responses from the target language into English before scoring improves performance.
The present study addresses this existing research gap by systematically examining the performance of three LLM-based approaches for automatically scoring German AUT responses. Specifically, we investigated how different LLM architectures align with human ratings and whether scoring original German responses differs from scoring responses that have been automatically translated into English beforehand.
To this end, we focused on LLMs specifically trained for the creativity context as well as a general-purpose model. We compared two fine-tuned models—the Cross-Lingual Alternate Uses Scoring (CLAUS) and OCSAI—which have been explicitly trained on DT responses, with the base version of GPT-4, a versatile but non-specialized model. The main pre-registered research questions (RQs) were as follows (see pre-registration at: https://aspredicted.org/b8tx-xs3p.pdf, accessed on 13 July 2024):
RQ1: How do creative evaluations by LLMs compare to human raters for German AUT responses?
RQ2: Which LLM method performs best?
RQ3: Are findings generalizable across datasets and items?
We expected substantial positive correlations between LLMs and human ratings of AUT responses (Organisciak et al. 2023) but had no specific expectations regarding the relative performance of specific LLMs under specific language conditions.

2. Materials and Methods

2.1. Transparency and Openness

All accompanying materials, data, and analysis scripts are publicly accessible in an Open Science Framework (OSF) repository: https://osf.io/eaqtb/, accessed on 5 September 2024. All deviations from the protocol are reported below.

2.2. Data Sources

We compiled published as well as unpublished rated German AUT data across five research groups. The eight lab and online studies included 49,391 responses from 2320 participants for 15 different alternate uses objects.1 Studies varied in sample characteristics (e.g., age, educational level, etc.) as well as in certain task characteristics such as task instructions, items, and timing but all studies involved human creativity ratings of AUT responses. Key characteristics of the included studies are presented in Table 1. A more detailed overview of study-specific characteristics—including task instructions, items used, timing, and the scored DT dimensions—is available on the OSF.

2.3. Human Creativity Ratings

In all datasets, individual AUT responses were coded by at least two human raters for creativity on a 5-point scale, which was homogenized for all studies to the range of 1–5. In each study, raters underwent initial training where they were instructed on how to evaluate responses. While the definition of what characterized a creative response varied slightly across studies, it generally emphasized both its novelty (in the sense of being unusual, original, unique, or surprising), and effectiveness (in the sense of being appropriate, useful, clever, interesting, exciting, or humorous). Inter-rater reliabilities ranged from substantial (Study 1: ICC(3,3) = 0.786, 95% CI: [0.779, 0.793]) to almost perfect agreement (Study 8: ICC(3,3) = 0.902, 95% CI: [0.890, 0.913]) at the study level (see Table 1). The exact rater guides used are available as Supplemental Material.

2.4. Large Language Model Creativity Ratings

Three LLM architectures were considered in this study: CLAUS, OCSAI, and GPT-4.2
CLAUS employs a fine-tuned version of the XLM-RoBERTa language model to predict human creativity ratings for AUT responses (Patterson et al. 2024a). It was developed and validated using AUT responses from 12 languages (including German) and is available as part of the Creativity Assessment Platform (CAP; https://cap.ist.psu.edu/claus, accessed on 19 September 2024), an open-access website for the administration and automated scoring of creativity tasks in verbal, visual, science, technology, engineering, and mathematics domains. After uploading the AUT prompt and response data as a .csv file to CLAUS, it returns creativity scores in the range between 0 and 1 (uncreative to highly creative), with a precision of up to 10 decimal places. Multilingual OCSAI (version used: 1.6), developed by Organisciak et al. (2023; https://openscoring.du.edu/scoringllm, accessed on 17 December 2024), employs a GPT-4o-mini based network trained on a dataset of 27,000 human-rated AUT responses. It provides scores from 1 to 5, with higher scores reflecting greater originality. GPT-4, developed by OpenAI, represents a transformer-based LLM trained through self-supervised learning on a large corpus of text. Its architecture allows it to generate context-sensitive responses by modeling text probabilities.
While CLAUS and OCSAI do not require explicit prompting in their scoring process, GPT-4 was prompted in a way that reflected the essential instructions of the human raters across the eight studies, highlighting the relevance of both novelty and effectiveness of responses, but without giving specific example data (i.e., zero-shot). The instructions asked to provide differentiated ratings from 10 to 50 to make them more comparable to the rating range of the other LLMs as well as the average human ratings:
Evaluate how creative the following use for the object %s is: %s.
An idea should be considered creative if it meets the following criteria: On the one hand, a creative idea should be novel (in the sense of being unusual, original, unique, or surprising), and on the other hand, it should also be effective (in the sense of being appropriate, useful, clever, interesting, exciting, or humorous).
Rate each idea on a scale from 10 (not creative at all) to 50 (very creative).
The symbols “%s” indicate placeholders that were automatically replaced by the object and response from the input table for each individual prompt. However, it is important to note that CLAUS does apply an implicit prompt in the background, which is also included in the exported file (e.g., Identify how surprising, creative, unexpected, or interesting the following alternate use for the object %s is: %s.). Similarly, OCSAI employs a structured few-shot text-to-text prompt, as described in Organisciak et al. (2023).
Data collection via GPT-4 was conducted between 29 October and 20 November 2024.3

2.5. Alternate Uses Task Data Preparation

Human ratings of the AUT data were transformed to a five-point scale [from 1 (not at all creative) to 5 (very creative)] if not originally used, scaling linearly between the minimum and maximum score. Invalid responses that were flagged by different exclusion codes across labs were uniformly set to zero. Zeros were not included in the mean score calculation, and if more than 50% of the raters scored a response as incomprehensible, the mean score of that response was set to not applicable and excluded from further analysis. In cases where fewer than half of the raters marked a response as invalid, we retained the response and calculated the mean score based on the valid ratings. This approach followed our pre-registered rule and was chosen to avoid discarding potentially meaningful data based on isolated judgements. Disagreements among raters typically reflected differences in knowledge rather than inattentiveness. For example, one rater might not recognize that a “guiro” is a musical instrument and therefore mark the response as incomprehensible, while others rate it as valid.
GPT-4 ratings were initially rated on a 10–50 scale to allow for finer differentiation of DT creativity ratings, as specified in the prompt above. These scores, as well as the CLAUS scores, were subsequently rescaled to a 1–5 range to ensure comparability with human ratings and the outputs of OCSAI.
For explorations of potential language effects and to compare models with different language capabilities under standardized output conditions, all responses were translated into English using DeepL (https://www.deepl.com, accessed on 12 September 2024). The accuracy of the translation was examined with a systematic cross-check of 1000 randomly drawn responses across datasets. A student assistant majoring in Psychology and English Studies reviewed the translations and categorized them into three categories: (1) “I would translate it exactly the same”, (2) “Translation is inappropriate”, or (3) “Not sure”. This process allowed us to quantify translation accuracy and identify potential issues. In only 6.0% of these responses, a different translation approach was suggested, while in 2.4% of the cases, the translation appeared inappropriate (mostly in cases of spelling errors, ambiguous phrasing, or incomplete responses), indicating overall good translation quality. Nevertheless, DeepL has consistently demonstrated high translation quality in dependent evaluations, particularly for preserving context and idiomatic meaning (e.g., Kamaluddin et al. 2024; Linlin 2024). In our analysis, the original DeepL translations were used without further corrections in order to reflect the aims of an automated scoring pipeline.

2.6. Analysis Strategy

To investigate the relationship between human creativity ratings and LLM scores, we employed an integrative data analysis (IDA; Curran and Hussong 2009), which has multiple advantages compared to synthesizing summary statistics obtained from each dataset (e.g., ecological fallacies are prevented; Kaufmann et al. 2016). IDA is considered the gold standard for meta-analysis when datasets of individual participants are available (Rogozińska et al. 2017; Thomas et al. 2014) and has been suggested as an approach to integratively analyze data from multi-lab replications (van Aert 2022).
We used a one-step estimation approach (Curran and Hussong 2009; van Aert 2022) based on linear mixed models with random intercepts and random slopes. Specifically, the following model was estimated (cf. van Aert 2022):
y i j = θ 0 i + λ 0 j + θ 1 i x i j + λ 1 j x i j + i j ,
with average human rating y i j for response j in dataset i , dataset-specific intercept θ 0 i , item-specific intercept λ 0 j , dataset-specific linear slope θ 1 i , item-specific linear slope λ 1 j , automated score x i j for response j in dataset i , and sampling error i j for response j in dataset i .
To account for variability across both studies and items, the dataset-specific intercepts and slopes ( θ 0 i , θ 1 i ) and the item-specific intercepts and slopes ( λ 0 j , λ 1 j ) were modeled as crossed random effects. We further assume a bivariate normal distribution N ( μ θ , Σ θ ) and a bivariate normal distribution ( μ λ , Σ λ ) for the random effects across studies and items, respectively. The mean vectors μ incorporated the average intercepts and slopes, and the covariance matrices Σ included freely estimated variance and covariance parameters. Covariances between random effects across studies and random effects across items were fixed to zero. All models were estimated by means of restricted maximum likelihood (Bates et al. 2015).
The dependent variable y i j and the independent variable x i j represent the average human rating and the automated score, respectively, divided by their item-specific standard deviation within each dataset. These transformations ensured that the dataset-specific linear slopes θ 1 i and the item-specific slopes λ 1 j could be interpreted analogously to the product-moment correlation coefficient (i.e., a measure for the linear relationship between variables with unit standard deviation). For that reason, we refer to the estimated coefficients θ ^ 1 i and λ ^ 1 j simply as correlations. The employed model accounted for variability across both studies and items, addressing the heterogeneity inherent in the overall dataset. To quantify the heterogeneity in our results, prediction intervals were estimated, which indicate the range of expected correlation coefficients for future studies. To illustrate, a standard deviation of correlations across studies of 0.30 implies that 95% of the population correlations should vary between ± 1.96 × 0.30 = 0.59 units of the typical correlation for each of the studies. Detailed linear mixed model estimates are provided as Supplemental Analyses (see Tables S1–S3).
While linear mixed models were the preferred approach for analyzing associations between human ratings and LLM scores, a different approach was used when examining relationships among LLMs. Since all LLM scores were derived from the same dataset and thus shared identical study- and item-level variability, modeling random effects seemed unnecessary. Instead, Pearson correlation coefficients were computed to assess the agreement between LLMs, as they provide a more direct and interpretable measure of their linear association.
Analyses were performed using R (version 4.4.0, R Core Team) with the lme4 package (Bates et al. 2015) and the lmerTest package (Kuznetsova et al. 2017) for linear mixed models.

3. Results

3.1. Distribution of Creativity Ratings Across Studies

Figure 1 provides an overview of the distributions of creativity ratings obtained from human raters and LLMs across studies. Distributions of LLM ratings are shown alternately for the original German dataset and the translated input.
Human ratings were concentrated in the lower to mid-range, forming a distribution that tends to be right-skewed. CLAUS scores exhibited a narrower and more symmetric distribution in both datasets, resembling a compressed normal distribution. OCSAI scores exhibited a distribution with a noticeable right skew in the German dataset, with specific values being clearly overrepresented, e.g., about 61.5% of scores fell into the 2.x range in the German dataset (37.8% 2.3), and the scores 1, 2, and 3 together accounted for 94.2% in the English dataset. GPT-4 scores exhibited a relatively wide distribution in both datasets, with certain scores being overrepresented, reflecting that, even when asking for a differentiated rating (originally 10–50), GPT-4 mostly still resorted to a less differentiated rating (i.e., 10, 15, 20, 30, 35, 50, which was then transformed to 1–5). Full descriptive statistics for each method are presented in the Supplemental Analyses (see Table S4).

3.2. Correlations Between Human and Large Language Model-Based Creativity Ratings

3.2.1. Correlations Across All Responses

Correlations across the full dataset (individual response level, N = 48,507) between human and LLM creativity scores, derived from linear mixed models, are presented in Table 2, column 1.
For the German dataset, correlations with human ratings ranged from 0.46 (CLAUS) to 0.66 (OCSAI), while GPT-4 achieved a correlation of 0.55. Similar trends were observed for the translated dataset, with correlations ranging from 0.47 (CLAUS) to 0.60 (OCSAI) and GPT-4 exhibiting a slightly lower correlation of 0.54. Regarding prediction intervals, correlations for the German dataset ranged from 0.36 to 0.55 for CLAUS, from 0.48 to 0.84 for OCSAI, and from 0.32 to 0.79 for GPT-4. In the translated dataset, the intervals were generally wider, ranging from 0.24 to 0.71 for CLAUS, from 0.29 to 0.90 for OCSAI, and from 0.34 to 0.74 for GPT-4 (see Table 3).
Inter-model correlations revealed substantial agreement between LLMs. The strongest relationships were found between GPT-4 and its translated counterpart (r = 0.81), and between OCSAI and its translation (r = 0.72). Correlations between CLAUS and OCSAI ranged from r = 0.55 (translated dataset) to r = 0.57 (German dataset), reflecting moderate alignment. In contrast, GPT-4 displayed weaker associations with CLAUS (r = 0.48 and r = 0.56), but high correlations with OCSAI (r = 0.73 and r = 0.65) across the German and translated datasets, respectively.

3.2.2. Correlations per Study

Correlations between human ratings and LLM creativity ratings were examined separately for each study to explore their range across different study designs [see Figure 2 and Appendix A (Table A1)].
For CLAUS, correlations ranged from 0.36 (Study 1) to 0.51 (Study 7) for the German dataset and from 0.32 (Study 7) to 0.67 (Study 8) for the translated input, highlighting variability in its performance. For OCSAI, correlations with human ratings were consistently higher, ranging from 0.52 (Study 7) to 0.79 (Study 6) for the German dataset, and from 0.35 (Study 7) to 0.77 (Study 3) for the translations. GPT-4 exhibited moderate correlations on average but showed noticeable fluctuations across studies. Estimates ranged from 0.34 (Study 7) to 0.70 (Study 6) for the German dataset and from 0.34 (Study 7) to 0.66 (Study 6) for the translations.

3.2.3. Correlations per Item

For a deeper insight into item-specific associations, we analyzed the correlations between human and LLM-based creativity ratings for each of the 15 items (see Figure 3). Some of the items were assessed in two or three studies. Across all models and studies, correlations of these 27 item-study combinations ranged from r = 0.18 [Car tires, Study 7, OCSAI (Translation)] to r = 0.91 (Brick, Study 6, OCSAI). Full descriptive statistics for each item are presented in the Supplemental Analyses (see Table S5).
For CLAUS, correlations in the German dataset exhibited moderate variability, spanning from r = 0.26 (Towel, Study 1) to r = 0.66 (Spoon, Study 4). Similarly, in the translated dataset, correlations ranged from r = 0.21 (Socks, Study 4) to r = 0.81 (Brick, Study 6). Notably, only 14.8% of the items in the German dataset achieved correlations of r ≥ 0.60, compared to 22.2% in the translated dataset. For OCSAI, correlations in the German dataset ranged from r = 0.43 (Car tires, Study 7) to r = 0.91 (Brick, Study 6), with 51.9% of the items achieving correlations of r ≥ 0.65. In the translated dataset, correlations ranged from r = 0.18 (Car tires, Study 7) to r = 0.85 (Pen, Study 6), with 37.0% of items reaching correlations of r ≥ 0.65, reflecting slightly lower consistency compared to the German datasets. For GPT-4, correlations in the German dataset ranged from r = 0.20 (Car tires, Study 7) to r = 0.84 (Pen, Study 6). Of these, 22.2% achieved correlations of r ≥ 0.65, while 64.2% reached r ≥ 0.50. In the translated dataset, correlations ranged from r = 0.26 (Car tires, Study 7) to r = 0.76 (Pen, Study 6). Here, 18.5% of items achieved correlations of r ≥ 0.65, while 58.0% reached r ≥ 0.50, showing a slight decline compared to the German dataset.

3.3. Exploring the Effectiveness of Large Language Model Scorings by Means of Rater Statistics

To further gauge the suitability of LLMs as raters in the context of creativity research, we computed rater statistics at the response-level (equivalent to item statistics) for those five datasets that have three or more raters (Studies 1, 5, 6, 7, 8). We computed rater-total correlations as an index of the discriminatory power of LLM methods compared to the human ratings. Table 4 summarizes the relative frequencies of cases where LLMs were identified as the weakest or best rater, or neither in terms of their rater-total correlation. Detailed results are available on the OSF (Table S6).
LLMs were identified as the weakest rater in 50.0% of cases, fell into the neither/nor category (neither weakest nor best) in 43.3%, and were considered the best rater in 6.7% with a weighted overall mean rater-total statistic of 0.71. CLAUS was identified as the weakest in 80.0% of cases and fell into the neither/nor category in 20.0%, with no instances of being classified as the best. Its mean rater-total correlation was 0.66 for the original German input and 0.69 for the translated input. OCSAI was the weakest in 20.0% of cases for the original German input and in 40.0% for the translated input. It was the best rater in 40.0% of cases for the original input, with no such cases for the translated input. Its mean rater-total correlation was 0.80 for the original input and 0.71 for the translated input. GPT-4 was the weakest in 20.0% of cases for the original input and 60.0% for the translated input, falling into the neither/nor category in the remaining cases (80.0% for the original German dataset and 40.0% for the translated input). Its mean rater-total correlation was 0.74 for the original German and 0.72 for the translated input. Mean rater-total correlation without LLMs was 0.74.

4. Discussion

Recent advances in machine learning are more and more applied to evaluate creativity in human responses and allow for an automated and less laborious scoring of such tasks (Goecke et al. 2024; Wahbeh et al. 2024; Yang et al. 2024; Zielińska et al. 2023). However, systematic comparisons of these scoring approaches remain scarce, particularly in languages other than English. Even fewer studies provide insights into the robustness of automated scoring across different studies and items. This work addresses these gaps by presenting a comprehensive analysis of three current LLM-based creativity scoring methods (i.e., CLAUS, OCSAI, GPT-4) applied to a large dataset of German AUT responses compiled across eight datasets from five research groups. Across the entire dataset of about 50,000 responses, we find that LLM-based models predict human creativity scores with substantial correlations in the range of 0.46 to 0.66. These findings speak directly to our first research question, which focused on the extent to which LLM-based scoring aligns with human creativity judgements.
Our findings add to the evidence that LLM-based predictions are clearly better than those previously obtained with semantic distance measures (Organisciak et al. 2023; Patterson et al. 2023). They are consistent with previous results that observed correlations for OCSAI between 0.42 and 0.72 depending on the item (Zielińska et al. 2023). Higher response-level correlations around 0.80 and higher have only been reported previously when testing the prediction of held-out data within the same English dataset (Luchini et al. 2025a; Organisciak et al. 2023; Patterson et al. 2023), whereas this study tested the autoscoring performance of LLMs in non-English data from independent datasets. In general, these findings suggest that LLM-based scoring can be conducted with German DT data. However, there are a number of nuances that need to be carefully considered when deciding on automated scoring with a specific LLM. In the following, we discuss to what extent findings depend on the LLM, translation of responses, and how findings generalize across studies and items.

4.1. Comparison of Large Language Models

We compared the creativity rating performance of CLAUS, OCSAI, and GPT-4 across our fairly large dataset against human ratings. This is specifically interesting from a conceptual multi-trait-multi-method (MTMM) perspective (cf. Campbell and Fiske 1959). From this perspective, we treat originality as a focal trait and distinguish among different methods of scoring responses—namely, human raters versus algorithmic scoring through distinct LLMs. In line with this framework, high agreement (i.e., correlations) across methodologically distinct methods (e.g., human and LLM-based) when assessing the same trait (=originality) would support convergent validity, while systematic differences between methods would indicate divergence. As stated by Campbell and Fiske (1959, p. 83), reliability and “validity can be seen as regions of a continuum”, and in this sense, “validity is represented in the agreement between two attempts to measure the same trait through maximally different methods”.
At this point, it is important to clarify the structure of our measurement design. We consider rater types (human vs. LLMs) as competing instantiations of one method dimension. In our study design, the single AUT items might be understood as a second method dimension. Validity evidence hinges on the extent to which trait-relevant variance (originality differences) is consistent across both item- and rater-based methodological variations.
We found that all three models predicted human creativity ratings at substantial levels, however, with serious heterogeneity. OCSAI was the top-performing model in most instances. Its performance remained satisfying across the eight included studies and nearly all 15 distinct items. In contrast, CLAUS and GPT-4 showed substantial but relatively smaller correlations with the human ratings, along with greater heterogeneity across items. This means that OCSAI aligns most closely with human creativity ratings, offering stronger evidence for convergent validity, while CLAUS and GPT-4 show weaker and more variable agreement—highlighting method effects across items and models. This finding addressed our second research question, which aimed at determining which LLM method performs best when predicting human creativity judgements.
From a judge response theory perspective (Myszkowski and Storme 2019), individual raters—whether humans or LLMs—can be conceptualized as indicators of an underlying latent construct (e.g., creative performance or originality in our case). This framework supports our interpretation that convergent validity is indicated when different methods yield similar scores across items. Nonetheless, the degree of independence among methods (as emphasized by Campbell and Fiske 1959) remains a critical point of discussion. While our scoring methods are not completely independent (since both humans and LLMs rely on the same participant responses), they do represent distinct procedures, which we argue justifies a cautious interpretation of our results as evidence for convergent validity.
The advantage of OCSAI likely stems from the fact that it combines the power and versatility of GPT-4o mini with additional fine-tuning (Organisciak et al. 2023). Hence, the increased performance beyond that of the GPT-4 base model was to be expected. Nevertheless, the performance of the GPT-4 base model was still remarkable given that it did not rely on any fine-tuning, but instead just a simple and straightforward prompt asking the model to evaluate the creativity of responses based on a short definition. This highlights the base model’s already substantial capacity to approximate human-like creativity assessments, and, given the rapid development of LLMs, it is to be expected that this capability might increase further in the near future.
The results suggested that GPT-4 tended to outperform CLAUS, especially when rating the original German responses; for the translated responses, CLAUS showed higher correlations with the human ratings than GPT-4 in four out of eight studies. This pattern underscores the potential influence of linguistic and methodological nuances in automated scoring approaches (Yang et al. 2024). This is further emphasized through the result that CLAUS performed very well for a number of items (and hence different prompts), but its performance dropped considerably for others. In contrast to that, GPT-4 showed a more consistent performance across items (r ≥ 0.40, except for one item). This variability is further reflected in the prediction intervals, which indicate considerable uncertainty in the expected range of correlations between human and LLM creativity ratings across future studies, with interval widths ranging from 0.10 to 0.31 depending on the model and language condition. Such variability within and across models is central to our third research question, which examined generalizability across datasets and items.
Viewed through the MTMM perspective, these results illustrate why employing multiple methods to assess creativity might remain crucial in the near future. OCSAI’s consistently higher correlations with human ratings suggest that it captures the target trait the best, but from our view, it should be understood as merely one method to approximate what is to be measured in the first place. After all, human ratings are also made up of several “methods”: different raters. Although CLAUS and GPT-4 showed somewhat lower consistency in their performance, the correlations were still substantial and also higher than previous automated scoring approaches using semantic distances (Forthmann et al. 2017b; Forthmann and Doebler 2022). Future research can build on these findings by examining more granular aspects of how different LLMs or fine-tuning strategies influence the ratings of creativity responses.

4.2. Effects of Response Translation

The translation of responses to English had only a limited effect on the correlations between LLM and human creativity ratings. CLAUS tended to benefit from the translation as it slightly increased correlations in five out of eight studies, whereas for OCSAI and GPT-4, translations reduced correlations in the same amount of studies. Taking a closer look at the LLMs themselves and their correlation with their translated counterparts, we observed varying degrees of consistency. GPT-4 exhibited the highest correlation between the original and translated versions, followed by OCSAI and CLAUS. These findings suggest that while GPT-4 demonstrates relatively stable performance across languages, translation may still introduce some variability. Compared to previous studies that reported near-complete interchangeability of OCSAI scores across languages (e.g., Zielińska et al. 2023), our results indicate that multilingual performance depends on the specific model architecture. Notably, CLAUS, despite being developed for creative scoring across multiple languages, showed the lowest cross-linguistic correlation. This suggests that its performance may be more sensitive to the input language than previously assumed (Luchini et al. 2025a; Patterson et al. 2023). These findings highlight that while LLMs offer strong multilingual capabilities, translation effects should still be considered, particularly for models with less explicit multilingual training, and for languages that are currently less represented in cross-language training and less accessible for model fine-tuning.

4.3. Variation Across Items and Studies

A closer inspection of the relationships at the level of single studies or items corroborated the general robustness of our findings but also pointed to notable variability. For example, while some AUT items like Spoon yielded high correlations with human ratings for all LLMs (r > 0.60), other items like Book had low correlations for all models (r < 0.50), and yet other items like Socks implied a large range in correlations across models (r = 0.21 to r = 0.63). The reasons for this variability were not fully explored. Differences across models could be due to specific characteristics of the item, responses, or study. For example, AUT items are known to differ in their semantic richness, and those with more sparse semantic networks (e.g., Pen) elicit more creative responses (Beaty et al. 2022), which may increase variance and thereby enhance correlations. Higher correlations could also be due to a high rate of uncreative responses (e.g., in studies applying “be-fluent” instructions), which typically yield excellent agreement across raters but also LLMs. Finally, item-related variability across models may partly be due to how these concepts are represented in the LLMs or the extent of exposure to them during fine-tuning. For example, CLAUS showed the highest variance in correlations across items, ranging between r = 0.21 and r = 0.81. Study characteristics that might also explain some of the found variability in the items include varying characteristics of the samples themselves (adolescents versus adults), the design choices (e.g., un-proctored versus proctored) and slightly different rating instructions [e.g., creativity rating emphasizing novelty and effectiveness of ideas (Diedrich et al. 2015)]. Since LLM-based creativity scores were always generated in the context of the respective item, including the item as a random effect in our models serves to adjust for overall item-level differences in scoring, without removing meaningful item-response interactions. The considerable variation in item-level performance across models—particularly the observation that some items yield substantially lower correlations—raises important questions about the generalizability of current scoring approaches. A more systematic investigation of such variability, including whether it is driven by semantic features of items, instruction types, or sample characteristics, represents a valuable direction for future research.

4.4. Can Large Language Models Replace Human Raters?

Our findings demonstrate that LLMs are able to predict human creativity ratings to a substantial degree, but also that considerable shares of variance are unaccounted for. Most of our analyses used the average human creativity ratings as the criterion, but human raters also did not agree perfectly—which is also not to be expected. Therefore, we additionally computed rater statistics at the response-level to explore how LLM-ratings compare to individual human raters. These analyses revealed that the discriminatory power of LLMs is very similar to that of human raters, referring to how well a rater or model can distinguish between more versus less creative responses (cf. Silvia 2008). In fact, in 80.0% of cases, OCSAI was equally or even more discerning than the human raters. Similarly, in 80.0%, ratings by GPT-4 were also comparable to those of the human raters. Ratings by CLAUS showed similar discriminatory power as the human raters in 20.0%. These rater statistics transfer the logic of item statistics—such as item-total correlations from classical test theory—to the level of raters, treating each rater (or model) as an “item” whose consistency with the overall judgement can be quantified. While exploratory in nature, this analysis suggested that CLAUS exhibited notably weaker discriminatory ability in this framework, pointing to potential limitations in its capacity to align with human judgements. As a potential qualification of this analysis, the relevant total was determined by a larger number of human raters than LLMs; on the other hand, LLMs evaluated creativity on a more differentiated scale than human raters who only used a five-point scale. Taken together, our findings suggest that ratings by LLMs—especially OCSAI and GPT-4—can approximate the consistency of human raters. This does not necessarily mean that LLM ratings can directly replace human ratings, but at least replace single human raters, as we need to assume that a single LLM is about as unreliable as a single human rater (Ceh et al. 2022). Still, from now on, we may consider teaming up different LLMs or LLM agents with different personas (Huang et al. 2024; Vinchon et al. 2023) to achieve high overall reliability. This reinforces the idea that using multiple raters in tandem with LLMs could strengthen the overall creativity assessment. Importantly, we observed that LLMs have no big issues with orthographically incorrect or even partly incomplete responses, which used to be an important benefit of human raters (e.g., compared to earlier approaches based on semantic distance). Hence, just like human raters, LLMs in this study rated most responses but also usually flagged invalid responses such as those that were entirely incomprehensible.

4.5. Limitations and Future Directions

A central goal of this study was to explore the feasibility of automatic methods to increase both efficiency and objectivity in creativity scoring. In practice, we ran into some issues during LLM-based scoring that need further attention to ensure a fully automatic, objective scoring process. CLAUS and OCSAI both come with graphical user interfaces that support convenient upload of response data; however, both methods could not handle data as large as ours (initially over 60,000 responses). OCSAI can also be accessed via an API which deals better with larger data. Curiously, OCSAI returned consistently lower ratings for some items (i.e., Brick and Knife), maybe related to its fine-tuning process, which can be an issue when correlating across responses from different items, but this was accounted for in our analysis. For GPT-4, the way the model is prompted still introduces a degree of subjectivity. Just as an inadequate prompt may lower an LLM’s automated scoring performance, we expect potential for further increasing its performance using soft prompting optimization or engineering of hard prompts. Hence, systematic analyses of the effect of different prompting methods on automated scoring methods represent an important avenue for future research. Moreover, we observed that scorings are not necessarily stable across time and conditions such as how many responses are scored at once. Therefore, we ended up scoring responses individually via the API (i.e., no batches for the GPT-4 approach, individual prompt per line), which was a costly procedure both with respect to time and money. Notably, commercial LLMs are always subject to change without notice, thereby undermining the reproducibility of scoring. For the future, it would be interesting to also consider other powerful LLMs such as Claude, Gemini, as well as open-source LLMs like LLaMA (Grattafiori et al. 2024; Touvron et al. 2023). When using base model LLMs, verbatim prompts should be disclosed together with model parameters (e.g., temperature) just like it is recommended for human rating instructions.
As a limitation of this study, we focused on a single DT task, the popular AUT (Saretzki et al. 2024b). At the same time, this has been the biggest effort in leveraging German AUT data to date. An increasing number of studies suggests that LLMs perform well also for several other DT tasks such as creative story writing, metaphor production, real-world problem solving, scientific ideation, and even for DT in the visual domain (Cropley and Marrone 2022; DiStefano et al. 2024; Luchini et al. 2025a; Organisciak et al. 2023; Patterson et al. 2024b). So far, most of this evidence relies on LLMs that were fine-tuned to a specific task. Our findings suggest that powerful base models like GPT-4 already outperform traditional semantic distance approaches in assessing creativity. Given the rapid pace of development in LLMs, this gap may narrow quickly, as more advanced models are expected to improve their ability to assess creative performance across multiple tasks and languages without requiring fine-tuning. Hence, future work should examine how automatic scoring performance varies across different types of DT tasks, as using different DT tasks likely supports construct coverage. As another limitation, our analyses only focused on what method maximized the relationship with human raters. This approach does not consider how scoring methods affect validity evidence such as in terms of predicting creative behavior and achievement (Saretzki et al. 2024b), and more reliable measures are not necessarily more valid (Benedek et al. 2013). One can further envision hybrid scoring approaches, where humans remain in the loop to ensure quality standards. Moreover, by reducing the costly human resources in the rating process, LLMs could be tasked to engage in more complex rating procedures such as assessing different quality dimensions (e.g., novelty, effectiveness, remoteness) which then could be combined to more powerful compound scores or studied separately.

5. Conclusions

Scoring DT responses with LLMs is nearly as reliable as human ratings and thus a promising alternative to subjective creativity scoring, with a growing number of open-access resources for researchers to facilitate automated creativity assessment for a range of DT tasks. Our findings demonstrate that specific LLMs work increasingly well for German AUT data usually without the need to translate responses to English prior to scoring. Once reproducible scoring pipelines are established, LLM-based creativity scoring can be a powerful tool to complement and replace human scoring and thereby expedite and standardize creativity research.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jintelligence13060064/s1.

Author Contributions

Conceptualization, J.S., T.K., B.F., B.G., A.-K.J., M.B. and S.W.; Data curation, J.S., T.K.; Formal analysis, J.S., B.F., T.K.; Funding acquisition, M.B., B.G.; Methodology, J.S. and B.F.; Writing—original draft, J.S., M.B. and S.W.; Writing—review & editing, J.S., T.K., B.F., B.G., A.-K.J., M.B. and S.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in whole, or in part, by the Austrian Science Fund (FWF) (https://doi.org/10.55776/P35661). For the purpose of open access, the author has applied a CC BY public copyright license to any Author Accepted Manuscript version arising from this submission.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Institutional Review Board (or Ethics Committee) of all participating affiliations.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

All files for analyses are available on the Open Science Framework: https://osf.io/eaqtb, accessed on 5 September 2024.

Acknowledgments

We gratefully acknowledge the advice and help by Roger Beaty and Peter Organisciak with the scoring of large data sets using CLAUS and OCSAI. Further, we thank all the research assistants for their efforts in rating the responses.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Table A1. Study level correlations of human and LLM creativity ratings using linear mixed models.
Table A1. Study level correlations of human and LLM creativity ratings using linear mixed models.
StudyResponsesCLAUSOCSAIGPT-4
GermanTranslationGermanTranslationGermanTranslation
Across Studies48,5070.460.470.660.600.550.54
111,1510.360.430.570.490.620.51
223190.460.590.750.680.620.57
327330.450.380.620.770.480.58
449170.470.360.650.760.490.59
576080.480.490.670.490.520.46
613,2440.450.620.790.700.700.66
756750.510.320.520.350.340.34
88600.470.670.720.530.660.58
Notes. LLM = large language model. All correlations reported were significant at p < .001.
Table A2. Item level correlations of human and LLM creativity ratings using linear mixed models.
Table A2. Item level correlations of human and LLM creativity ratings using linear mixed models.
IDItem CLAUSOCSAIGPT-4
NGermanTranslationGermanTranslationGermanTranslation
1Buch (Book)18250.360.430.480.460.400.42
1Handtuch (Towel)44930.260.340.560.500.540.42
1Holzlatte (Wooden slat)18030.350.370.580.340.590.47
1Messer (Knife)30300.460.500.620.670.610.58
2Autoreifen (Car tires)10780.320.540.650.510.470.49
2Ziegelstein (Brick)12410.530.790.870.770.720.63
3Glühbirne (Light bulb)5170.300.320.620.600.460.44
3Holzlatte (Wooden slat)9210.440.320.630.620.450.53
3Löffel (Spoon)6360.640.690.760.750.670.68
3Socken (Socks)6590.400.230.590.520.490.43
4Glühbirne (Light bulb)10090.320.300.650.580.470.46
4Holzlatte (Wooden slat)14670.460.300.660.610.460.54
4 Löffel (Spoon)11660.660.680.800.730.690.69
4Socken (Socks)12750.430.210.630.500.500.44
5Eimer (Bucket)17510.530.570.670.640.580.60
5Gürtel (Belt)14760.420.450.580.490.540.48
5Kerze (Candle)12940.440.490.710.670.540.52
5Messer (Knife)13160.580.560.720.670.510.53
5Socke (Sock)17710.330.290.600.570.570.57
6Autoreifen (Car tires)31820.320.560.690.540.560.59
6Konservendose (Tin can)35290.610.490.700.670.590.59
6Stift (Pen)31540.570.760.860.850.840.76
6Ziegelstein (Brick)33790.520.810.910.800.800.72
7Autoreifen (Car tires)19690.370.260.430.180.200.26
7Stift (Pen)17420.630.460.600.490.480.44
7Ziegelstein (Brick)19640.570.510.650.440.450.40
8Messer (Knife)8600.570.630.770.710.650.66
Notes. All correlations reported were significant at p < .001.

Notes

1
Three additional datasets were initially considered in the pre-registration but not included in the analyses. Weiss et al. (2023) was excluded because its ratings focused solely on assessing the appropriateness of responses. Kleinkorres et al. (2021) and Forthmann and Doebler (2022) were excluded because they were part of the German training dataset for the fine-tuned LLMs (see Patterson et al. 2023), which undermines the independence of the test data.
2
We used the GPT-4 model instead of the pre-registered GPT-4o, as GPT-4o was not available for the data volume at the time of data collection.
3
Note: GPT is available both through OpenAI’s chat interface and via a paid Application Programming Interface (API). While the chat interface allows for manual interactions, the API is recommended for processing larger datasets. The R code provided in the OSF Supplemental Material documents the procedure used in this study.

References

  1. Amabile, Teresa M. 1982. Social Psychology of Creativity: A Consensual Assessment Technique. Journal of Personality and Social Psychology 43: 997–1013. [Google Scholar] [CrossRef]
  2. Amabile, Teresa M. 1996. Creativity in Context: Update to the Social Psychology of Creativity. London: Routledge. [Google Scholar]
  3. Baer, John, James C. Kaufman, and Claudia A. Gentile. 2004. Extension of the Consensual Assessment Technique to Nonparallel Creative Products. Creativity Research Journal 16: 113–17. [Google Scholar] [CrossRef]
  4. Bates, Douglas, Martin Mächler, Ben Bolker, and Steve Walker. 2015. Fitting Linear Mixed-Effects Models Using Lme4. Journal of Statistical Software 67: 1–48. [Google Scholar] [CrossRef]
  5. Beaty, Roger E., and Dan R. Johnson. 2021. Automating creativity assessment with SemDis: An open platform for computing semantic distance. Behavior Research Methods 53: 757–80. [Google Scholar] [CrossRef] [PubMed]
  6. Beaty, Roger E., Dan R. Johnson, Daniel C. Zeitlen, and Boris Forthmann. 2022. Semantic Distance and the Alternate Uses Task: Recommendations for Reliable Automated Assessment of Originality. Creativity Research Journal 34: 245–60. [Google Scholar] [CrossRef]
  7. Benedek, Mathias, and Izabela Lebuda. 2024. Managing Your Muse: Exploring Three Levels of Metacognitive Control in Creative Ideation. Creativity Research Journal 13: 1–12. [Google Scholar] [CrossRef]
  8. Benedek, Mathias, Caterina Mühlmann, Emanuel Jauk, and Aljoscha C. Neubauer. 2013. Assessment of Divergent Thinking by Means of the Subjective Top-Scoring Method: Effects of the Number of Top-Ideas and Time-on-Task on Reliability and Validity. Psychology of Aesthetics, Creativity, and the Arts 7: 341–49. [Google Scholar] [CrossRef]
  9. Campbell, Donald T., and Donald W. Fiske. 1959. Convergent and Discriminant Validation by the Multitrait-Multimethod Matrix. Psychological Bulletin 56: 81–105. [Google Scholar] [CrossRef]
  10. Ceh, Simon Majed, Carina Edelmann, Gabriela Hofer, and Mathias Benedek. 2022. Assessing Raters: What Factors predict Discernment in Novice Creativity Raters? The Journal of Creative Behavior 56: 41–54. [Google Scholar] [CrossRef]
  11. Cropley, David H., and Rebecca L. Marrone. 2022. Automated scoring of figural creativity using a convolutional neural network. Psychology of Aesthetics, Creativity, and the Arts 19: 77–86. [Google Scholar] [CrossRef]
  12. Curran, Patrick J., and Andrea M. Hussong. 2009. Integrative Data Analysis: The Simultaneous Analysis of Multiple Data Sets. Psychological Methods 14: 81–100. [Google Scholar] [CrossRef] [PubMed]
  13. Diedrich, Jennifer, Mathias Benedek, Emanuel Jauk, and Aljoscha C. Neubauer. 2015. Are Creative Ideas Novel and Useful? Psychology of Aesthetics, Creativity, and the Arts 9: 35–40. [Google Scholar] [CrossRef]
  14. DiStefano, Paul V., John D. Patterson, and Roger E. Beaty. 2024. Automatic Scoring of Metaphor Creativity with Large Language Models. Creativity Research Journal, 1–15. [Google Scholar] [CrossRef]
  15. Dumas, Denis, Peter Organisciak, Shannon Maio, and Michael Doherty. 2021. Four Text-Mining Methods for Measuring Elaboration. The Journal of Creative Behavior 55: 517–31. [Google Scholar] [CrossRef]
  16. Forthmann, Boris, and Philipp Doebler. 2022. Fifty Years Later and Still Working: Rediscovering Paulus et al.’s (1970) Automated Scoring of Divergent Thinking Tests. Psychology of Aesthetics, Creativity, and the Arts 19: 63–76. [Google Scholar] [CrossRef]
  17. Forthmann, Boris, Benjamin Goecke, and Roger E. Beaty. 2023. Planning Missing Data Designs for Human Ratings in Creativity Research: A Practical Guide. Creativity Research Journal 37: 167–78. [Google Scholar] [CrossRef]
  18. Forthmann, Boris, David Jendrycko, J. Meinecke, and Heinz Holling. 2017a. Tackling Creative Quality in Divergent Thinking: Dissection of Ingredients. [Unpublished manuscript]. Münster: University of Münster. [Google Scholar]
  19. Forthmann, Boris, Heinz Holling, Pinar Çelik, Martin Storme, and Todd Lubart. 2017b. Typing Speed as a Confounding Variable and the Measurement of Quality in Divergent Thinking. Creativity Research Journal 29: 257–69. [Google Scholar] [CrossRef]
  20. Goecke, Benjmain, Paul V. DiStefano, Wolfgang Aschauer, Kurt Haim, Roger Beaty, and Boris Forthmann. 2024. Automated Scoring of Scientific Creativity in German. The Journal of Creative Behavior 58: 321–27. [Google Scholar] [CrossRef]
  21. Grattafiori, Aaron, Abhimanyu Dubey, Abhinav Jauhri, Abhhinav Pandey, Abhishek Kadian, Ahmad Al-Dahle, Aiesha Letman, Akhil Mathur, Alan Schelten, Alex Vaughan, and et al. 2024. The llama 3 herd of models. arXiv arXiv:2407.21783. [Google Scholar]
  22. Guilford, Joy Paul. 1950. Creativity. American Psychologist 5: 444–54. [Google Scholar] [CrossRef]
  23. Guilford, Joy Paul. 1967. The Nature of Human Intelligence. New York: MacGraw Hill. [Google Scholar]
  24. Huang, Yudong, Hongyang Du, Xinyuan Zhang, Dusit Niyato, Jiawen Kang, Zehui Xiong, Shuo Wang, and Tao Huang. 2024. Large Language Models for Networking: Applications, enabling Techniques, and Challenges. IEEE Network 39: 235–42. [Google Scholar] [CrossRef]
  25. Jaggy, Ann-Kathrin, Dan John, Noel Wytopil, Amelie Schönle, Jessika Golle, Benjamin Nagengast, and Ulrich Trautwein. 2025. Talent-Study. Available online: https://osf.io/kbwz7 (accessed on 3 February 2025).
  26. Kamaluddin, Mohamad Ihsan, Moch Wildan Khoeurul Rasyid, Fourus Haznutal Abqoriyyah, and Andang Saehu. 2024. Accuracy analysis of DeepL: Breakthroughs in machine translation technology. Journal of English Education Forum 4: 122–26. [Google Scholar] [CrossRef]
  27. Kaufmann, Esther, Ulf-Dietrich Reips, and Katharina Maag Merki. 2016. Avoiding Methodological Biases in Meta-Analysis. Zeitschrift für Psychologie 224: 157–167. [Google Scholar] [CrossRef]
  28. Kleinkorres, Ruben, Boris Forthmann, and Heinz Holling. 2021. An Experimental Approach to Investigate the Involvement of Cognitive Load in Divergent Thinking. Journal of Intelligence 9: 3. [Google Scholar] [CrossRef] [PubMed]
  29. Knopf, Thomas, and Clemens M. Lechner. 2025. Measuring Innovation Skills with the Behavioral, Emotional and Social Skills Inventory (BESSI): Validating Self-Reports Against Eight Performance Tests of Cognitive Abilities. [Manuscript in Preparation]. Mannheim: GESIS–Leibniz Institute for the Social Sciences. [Google Scholar]
  30. Kuznetsova, Alexandra, Per B. Brockhoff, and Rune H. B. Christensen. 2017. lmerTest Package: Tests in Linear Mixed Effects Models. Journal of Statistical Software 82: 1–26. [Google Scholar] [CrossRef]
  31. Landauer, Thomas K., Peter W. Foltz, and Darrell Laham. 1998. An Introduction to Latent Semantic Analysis. Discourse Processes 25: 259–84. [Google Scholar] [CrossRef]
  32. Lebuda, Izabela, and Mathias Benedek. 2024. Contributions of Metacognition to Creative Performance and Behavior. The Journal of Creative Behavior 59: e652. [Google Scholar] [CrossRef]
  33. Linlin, Li. 2024. Artificial Intelligence Translator DeepL Translation Quality Control. Procedia Computer Science 247: 710–17. [Google Scholar] [CrossRef]
  34. Luchini, Simone A., Ibraheem Muhammad Moosa, John D. Patterson, Dan R. Johnson, Matthijs Baas, Baptiste Barbot, Iana P. Bashmakova, Mathias Benedek, Qunlin Chen, Giovanni E. Corazza, and et al. 2025a. Automated Assessment of Creativity in Multilingual Narratives. Psychology of Aesthetics, Creativity, and the Arts. advance online publication. [Google Scholar] [CrossRef]
  35. Luchini, Simone A., Nadine T. Maliakkal, Paul V. DiStefano, Antonio Laverghetta, John D. Patterson, Roger E. Beaty, and Roni Reiter-Palmon. 2025b. Automatic Scoring of Creative Problem-Solving with Large Language Models: A Comparison of Originality and Quality Ratings. Psychology of Aesthetics, Creativity, and the Arts. advance online publication. [Google Scholar] [CrossRef]
  36. Mouchiroud, Christophe, and Todd Lubart. 2001. Children’s Original Thinking: An Empirical Examination of Alternative Measures derived from Divergent Thinking Tasks. The Journal of Genetic Psychology 162: 382–401. [Google Scholar] [CrossRef]
  37. Myszkowski, Nils, and Martin Storme. 2019. Judge response theory? A call to upgrade our psychometrical account of creativity judgments. Psychology of Aesthetics, Creativity, and the Arts 13: 167–75. [Google Scholar] [CrossRef]
  38. Organisciak, Peter, Selcuk Acar, Denis Dumas, and Kelly Berthiaume. 2023. Beyond Semantic Distance: Automated Scoring of Divergent Thinking Greatly Improves with Large Language Models. Thinking Skills and Creativity 49: 101356. [Google Scholar] [CrossRef]
  39. Patterson, John D., Baptiste Barbot, James Lloyd-Cox, and Roger E. Beaty. 2024a. AuDrA: An Automated Drawing Assessment Platform for Evaluating Creativity. Behavior Research Methods 56: 3619–36. [Google Scholar] [CrossRef] [PubMed]
  40. Patterson, John D., Hannah M. Merseal, Dan R. Johnson, Sergio Agnoli, Matthisj Baas, Brendan S. Baker, Baptiste Barbot, Mathias Benedek, Khatereh Borhani, Qunlin Chen, and et al. 2023. Multilingual Semantic Distance: Automatic Verbal Creativity Assessment in many Languages. Psychology of Aesthetics, Creativity, and the Arts 17: 495. [Google Scholar] [CrossRef]
  41. Patterson, John, Jimmy Pronchick, Ruchi Panchanadikar, Mark Fuge, Janet van Hell, Scarlett Miller, Dan Johnson, and Roger Beaty. 2024b. CAP: The Creativity Assessment Platform for Online Testing and Automated Scoring. [Manuscript submitted for publication]. State College: Department of Psychology, Pennsylvania State University. [Google Scholar]
  42. Paulus, Dieter H., and Joseph S. Renzuli. 1968. Scoring Creativity Tests by Computer. Gifted Child Quarterly 12: 79–83. [Google Scholar] [CrossRef]
  43. Rogozińska, Ewelina, Nadine Marlin, Shakila Thangaratinam, Khalid S. Khan, and Javier Zamora. 2017. Meta-Analysis using Individual Participant Data from Randomised Trials: Opportunities and Limitations created by Access to Raw Data. BMJ Evidence-Based Medicine 22: 157–62. [Google Scholar] [CrossRef] [PubMed]
  44. Saretzki, Janika, Boris Forthmann, and Mathias Benedek. 2024a. A systematic quantitative review of divergent thinking assessments. Psychology of Aesthetics, Creativity and the Arts. advance online publication. [Google Scholar] [CrossRef]
  45. Saretzki, Janika, Rosalie Andrae, Boris Forthmann, and Mathias Benedek. 2024b. Investigation of Response Aggregation Methods in Divergent Thinking Assessments. The Journal of Creative Behavior. [Google Scholar] [CrossRef]
  46. Silvia, Paul J. 2008. Discernment and creativity: How well can people identify their most creative ideas? Psychology of Aesthetics, Creativity, and the Arts 2: 139–46. [Google Scholar] [CrossRef]
  47. Silvia, Paul J., Beate P. Winterstein, John T. Willse, Christopher M. Barona, Joshua T. Cram, Karl I. Hess, Jenna L. Martinez, and Crystal A. Richard. 2008. Assessing Creativity with Divergent Thinking Tasks: Exploring the Reliability and Validity of new Subjective Scoring Methods. Psychology of Aesthetics, Creativity, and the Arts 2: 68–85. [Google Scholar] [CrossRef]
  48. Stevenson, Claire, Iris Smal, Matthijs Baas, Raoul Grasman, and Han van der Maas. 2022. Putting GPT-3’s Creativity to the (Alternative Uses) Test. arXiv arXiv:2206.08932. [Google Scholar]
  49. Thomas, Doneal, Sanyath Radji, and Andrea Benedetti. 2014. Systematic Review of Methods for Individual Patient Data Meta-Analysis with Binary Outcomes. BMC Medical Research Methodology 14: 79. [Google Scholar] [CrossRef] [PubMed]
  50. Torrance, E. Paul. 1972. Predictive Validity of the Torrance Tests of Creative Thinking. The Journal of Creative Behavior 6: 236–52. [Google Scholar] [CrossRef]
  51. Touvron, Hugo, Louis Martin, Kevin Stone, Peter Albert, Amjad Almahairi, Yasmine Babaei, Nikolay Bhargava, Soumya Batra, Prajjwal Bhargava, Shruti Bhosale, and et al. 2023. Llama 2: Open foundation and fine-tuned chat models. arXiv arXiv:2307.09288. [Google Scholar]
  52. van Aert, Robbie C. 2022. Analyzing Data of a Multilab Replication Project With Individual Participant Data Meta-Analysis. Zeitschrift für Psychologie 230: 60–72. [Google Scholar] [CrossRef]
  53. Vinchon, Florent, Todd Lubart, Sabrina Bartolotta, Valentin Gironnay, Marion Botella, Samira Bourgeois-Bougrine, Jean-Marie Burkhardt, Nathalie Bonnardel, Giovanni Emanuele Corazza, Vlad Glăveanu, and et al. 2023. Artificial Intelligence & Creativity: A manifesto for Collaboration. The Journal of Creative Behavior 57: 472–84. [Google Scholar] [CrossRef]
  54. Wahbeh, Helané, Cedric Cannard, Garret Yount, Arnaud Delorme, and Dean Radin. 2024. Creative Self-Belief Responses Versus Manual and Automated Guilford Alternate Use Task Scoring: A Cross-Sectional Study [Preprint]. PsyArXiv. Available online: https://osf.io/preprints/psyarxiv/vjqnu_v1 (accessed on 27 January 2025).
  55. Wallach, Michael A., and Nathan Kogan. 1965. A New Look at the Creativity-Intelligence Distinction. Journal of Personality 33: 348–69. [Google Scholar] [CrossRef]
  56. Weiss, Selina, Benjamin Goecke, and Oliver Wilhelm. 2024. How Much Retrieval Ability Is in Originality? The Journal of Creative Behavior 58: 370–87. [Google Scholar] [CrossRef]
  57. Weiss, Selina, Sally Olderbak, and Oliver Wilhelm. 2023. Conceptualizing and Measuring Ability Emotional Creativity. Psychology of Aesthetics, Creativity, and the Arts 19: 450–65. [Google Scholar] [CrossRef]
  58. Wilson, Robert C., Joy Paul Guilford, Paul R. Christensen, and Donald J. Lewis. 1954. A Factor-Analytic Study of Creative-Thinking Abilities. Psychometrika 19: 297–311. [Google Scholar] [CrossRef]
  59. Yang, Tianchen, Qifan Zhang, Zhaoyang Sun, and Yubo Hou. 2024. Automatic Assessment of Divergent Thinking in Chinese Language with TransDis: A Transformer-Based Language Model Approach. Behavior Research Methods 56: 5798–819. [Google Scholar] [CrossRef] [PubMed]
  60. Yu, Yuhua, Roger E. Beaty, Boris Forthmann, Mark Beeman, John Henry Cruz, and Dan Johnson. 2023. A MAD method to assess idea novelty: Improving validity of automatic scoring using maximum associative distance (MAD). Psychology of Aesthetics, Creativity, and the Arts. advance online publication. [Google Scholar] [CrossRef]
  61. Zielińska, Aleksandra, Peter Organisciak, Denis Dumas, and Maciej Karwowski. 2023. Lost in Translation? Not for Large Language Models: Automated Divergent Thinking Scoring Performance Translates to Non-English Contexts. Thinking Skills and Creativity 50: 101414. [Google Scholar] [CrossRef]
Figure 1. Distribution of human and large language model creativity ratings across studies.
Figure 1. Distribution of human and large language model creativity ratings across studies.
Jintelligence 13 00064 g001
Figure 2. Correlations of creativity ratings by human and LLMs examined with linear mixed models per study. Notes. LLM = large language model. Error bars represent standard error of estimates.
Figure 2. Correlations of creativity ratings by human and LLMs examined with linear mixed models per study. Notes. LLM = large language model. Error bars represent standard error of estimates.
Jintelligence 13 00064 g002
Figure 3. Correlations of human and LLM creativity ratings using LLMs per item. Notes. LLM = large language model. Shaded area reflects overall correlations across studies and LLMs from lowest (0.46, CLAUS) to highest (0.66, OCSAI). Items in bold represent lowest and highest single item/model correlation. German items were ordered alphabetically.
Figure 3. Correlations of human and LLM creativity ratings using LLMs per item. Notes. LLM = large language model. Shaded area reflects overall correlations across studies and LLMs from lowest (0.46, CLAUS) to highest (0.66, OCSAI). Items in bold represent lowest and highest single item/model correlation. German items were ordered alphabetically.
Jintelligence 13 00064 g003
Table 1. Key characteristics of included studies.
Table 1. Key characteristics of included studies.
IDReferenceNItemsResponsesRatersICC [95% CI]
1Weiss et al. 2024328411,59930.786 [0.779, 0.793]
2Knopf and Lechner 20254432244520.810 [0.795, 0.825]
3Jaggy et al. 2025129428352 10.804 [0.744, 0.857]
4Jaggy et al. 2025218450422 10.814 [0.785, 0.909]
5Saretzki et al. 2024b3005769160.855 [0.850, 0.860]
6Lebuda and Benedek 2024425413,24460.813 [0.808, 0.818]
7Benedek and Lebuda 20243173567560.808 [0.800, 0.815]
8Forthmann et al. 2017a160186030.902 [0.890, 0.913]
Total-232027 249,39130-
Notes. The number of participants and responses reflect data prior to the exclusion of 884 invalid responses and 30 participants without valid responses during data processing. ICC values represent item-specific ICC(3,k), except for Studies 3 and 4, which report ICC(1,k) due to the two-rater combination setup. 1 Five raters in different two-rater-combinations (every response has been rated by two raters). 2 Number of distinct items: 15. Note that “Sock” and “Socks” were treated as separate items.
Table 2. Overall correlation matrix of human and LLM creativity ratings.
Table 2. Overall correlation matrix of human and LLM creativity ratings.
1. 12.3.4.5.6.
1. Human Ratings-
2. CLAUS0.46-
3. CLAUS (Translation)0.470.54-
4. OCSAI0.660.570.63-
5. OCSAI (Translation)0.600.450.550.72-
6. GPT-40.550.480.540.730.59-
7. GPT-4 (Translation)0.540.450.560.690.650.81
Notes. LLM = large language model. All correlations reported were significant at p < .001. 1 lme4 Syntax for estimating the model in Equation (1) (i.e., the correlation between human and LLM creativity ratings): HumanRatings ~ LLM + (1 + LLM | Study) + (1 + LLM | Item).
Table 3. Prediction intervals for human-LLM creativity rating correlations.
Table 3. Prediction intervals for human-LLM creativity rating correlations.
LLMEstimateSEPrediction IntervalInterval Range
LLUL
CLAUS0.460.040.360.550.10
CLAUS (Translation)0.470.060.240.710.23
OCSAI0.660.040.480.840.18
OCSAI (Translation)0.600.070.290.900.31
GPT-40.550.050.320.790.24
GPT-4 (Translation)0.540.050.340.740.20
Notes. LLM = large language model.
Table 4. Rater statistics including LLMs.
Table 4. Rater statistics including LLMs.
LLM(s) as Weakest RaterLLM(s) Neither Weakest nor BestLLM(s) as
Best Rater
Mean LLM-Total Correlation
CLAUS80.0%20.0%-0.66
CLAUS (Translation)80.0%20.0%-0.69
OCSAI20.0%40.0%40.0%0.80
OCSAI (Translation)40.0%60.0%-0.71
GPT-420.0%80.0%-0.74
GPT-4 (Translation)60.0%40.0%-0.72
Average50.0%43.3%6.7%0.71
Notes. LLMs = large language models. Relative frequencies always refer to the relevant total of the number of rater statistics including datasets with three or more raters (i.e., Studies 1, 5, 6, 7, and 8).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Saretzki, J.; Knopf, T.; Forthmann, B.; Goecke, B.; Jaggy, A.-K.; Benedek, M.; Weiss, S. Scoring German Alternate Uses Items Applying Large Language Models. J. Intell. 2025, 13, 64. https://doi.org/10.3390/jintelligence13060064

AMA Style

Saretzki J, Knopf T, Forthmann B, Goecke B, Jaggy A-K, Benedek M, Weiss S. Scoring German Alternate Uses Items Applying Large Language Models. Journal of Intelligence. 2025; 13(6):64. https://doi.org/10.3390/jintelligence13060064

Chicago/Turabian Style

Saretzki, Janika, Thomas Knopf, Boris Forthmann, Benjamin Goecke, Ann-Kathrin Jaggy, Mathias Benedek, and Selina Weiss. 2025. "Scoring German Alternate Uses Items Applying Large Language Models" Journal of Intelligence 13, no. 6: 64. https://doi.org/10.3390/jintelligence13060064

APA Style

Saretzki, J., Knopf, T., Forthmann, B., Goecke, B., Jaggy, A.-K., Benedek, M., & Weiss, S. (2025). Scoring German Alternate Uses Items Applying Large Language Models. Journal of Intelligence, 13(6), 64. https://doi.org/10.3390/jintelligence13060064

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop