Next Article in Journal
Syngas Production and Heavy Metals Distribution During the Gasification of Biomass from Phytoremediation Poplar Prunings: A Case Study
Previous Article in Journal
Green Transportation Planning for Smart Cities: Digital Twins and Real-Time Traffic Optimization in Urban Mobility Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Can ChatGPT Replace the Teacher in Assessment? A Review of Research on the Use of Large Language Models in Grading and Providing Feedback

Department of Logic and Cognitive Science, Faculty of Psychology and Cognitive Science, Adam Mickiewicz University, Szamarzewskiego 89, 60-568 Poznań, Poland
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2026, 16(2), 680; https://doi.org/10.3390/app16020680
Submission received: 14 November 2025 / Revised: 1 January 2026 / Accepted: 3 January 2026 / Published: 8 January 2026
(This article belongs to the Special Issue Artificial Intelligence in Education: Latest Advances and Prospects)

Abstract

This article presents a systematic review of empirical research on the use of large language models (LLMs) for automated grading of student work and providing feedback. The study aimed to determine the extent to which generative artificial intelligence models, such as ChatGPT, can replace teachers in the assessment process. The review was conducted in accordance with PRISMA guidelines and predefined inclusion criteria; ultimately, 42 empirical studies were included in the analysis. The results of the review indicate that the effectiveness of LLMs in grading is varied. These models perform well on closed-ended tasks and short-answer questions, often achieving accuracy comparable to human evaluators. However, they struggle with assessing complex, open-ended, or subjective assignments that require in-depth analysis or creativity. The quality of the prompts provided to the model and the use of detailed scoring rubrics significantly influence the accuracy and consistency of the grades generated by LLMs. The findings suggest that LLMs can support teachers by accelerating the grading process and delivering rapid feedback at scale, but they cannot fully replace human judgment. The highest effectiveness is achieved in hybrid assessment systems that combine AI-driven automatic grading with teacher oversight and verification.

1. Introduction

In recent years, there has been a rapid increase in the number of publications addressing the use of generative artificial intelligence (GenAI), including large language models (LLMs), in education. This literature depicts numerous scenarios of positive transformation in educational practices [1]. Personalized and adaptive systems powered by LLMs are expected to tailor educational content and learning pace to individual student needs, in effect acting as an intelligent tutor [2,3,4,5]. Such GenAI-based assistants are starting to provide psychological and emotional support to students, interacting with them directly but also monitoring their condition and alerting staff and relevant adults when necessary, thus potentially relieving overburdened school counseling services and other forms of institutional support [6,7,8]. Technological companies responsible for the GenAI revolution promise the expansion and democratization of access to knowledge. With features like automatic translations and simplifying content’s complexity, and the possibility of presenting the same message in alternative formats nearly immediately, GenAI tools can cater to students having diverse learning requirements [9,10,11].
This assistive role can relieve the teachers and administration in educational institutions as well, aiding with documentation, preparing classes, or providing feedback for students [12,13,14,15]. It enables the rapid generation of various types of questions and exercises, automatic adjustment of their difficulty level, personalization of educational materials, and helping with the automatic assessment of student progress [16,17]. Beyond these convenient enhancements, researchers anticipate that engaging with GenAI could help develop digital competencies and more universal cognitive functions like creativity and analytical thinking. Being literate in digital technology, and GenAI in particular, is thought to help prepare students for a future job market, and simultaneously improve their metacognitive skills. In the classroom, GenAI is said to inspire students by offering multiple perspectives and problem solutions, thus encouraging creative thinking. Teachers are already experimenting with image and text generators in arts and humanities courses to spark imagination and increase student engagement [18,19,20,21,22].
LLMs offer capabilities far beyond earlier rule-based or statistical systems [23]. In contrast to these earlier architectures, research on automated essay scoring highlights that traditional systems were constrained by surface-level features and lacked mechanisms for modeling deeper discourse structures [24]. LLM architectures can analyze large volumes of text, enabling nuanced interpretation of students’ work and authors’ intent, which constitutes a substantive advance over previous technologies. The model not only interprets an assignment but also adapts dynamically to specified evaluative criteria through the attention mechanism, functioning analogously to a teacher highlighting central ideas and disregarding peripheral material. By weighting the relevance of each word in relation to the entire text, the model can distinguish coherent argumentation from fragmented reasoning and, when guided by the instructor’s prompt, generate personalized and contextually appropriate feedback that supports more individualized learning.
But despite these broad expectations and hopes, experimental validation of GenAI usage in education remains limited. Existing work is often theoretical, focusing on speculation, hopes, conceptual analyses (e.g., [14,18]), or preliminary exploratory studies (e.g., [16]) rather than systematic empirical investigations. This abundance of opinions accompanied by a scarcity of results can potentially distort our perception of what are the actual benefits, opportunities, and limitations of GenAI in education.
It could also be argued that many peer-reviewed studies on GenAI educational usage make little reference to the technical benchmarks and performance tests found in model cards. Such benchmarks are created for different purposes—such as engineering metrics and competitive evaluations—so their relevance to classroom practice is only partial. The end result, however, is that educational research often lags behind rapid model developments, assessing earlier versions while industry moves on to reasoning and agentic models. This gap highlights the constant need for systematic reviews of academic research.
This article aims to bridge this gap by systematically reviewing empirical studies focused on one specific aspect of the technology’s impact on education: using GenAI to assess and grade student work. To map empirical studies on this topic, we followed the PRISMA protocol [25,26,27].
Given the rapid pace of technological progress, we believed it was best to keep our review questions broad. More specifically, the review addresses the following research questions:
RQ1: 
In what educational contexts, subjects, and time periods have GenAI models been used for automated grading and feedback, and which concrete models have been deployed?
Information on where and how GenAI is applied will provide us with a better understanding of whether there are any specific fields of education in which GenAI use is already more widespread, and whether some AI tools are more popular than others.
RQ2: 
How are GenAIs operationalized in grading workflows, including their roles (automatic grader, feedback generator, co-grader), the types of student work they assess, and the prompting or other strategies used?
Information on the details of GenAI implementation in the assessment workflows is crucial, as they are likely to be a significant factor impacting the quality and relevance of algorithmic outputs, as well as the degree of its autonomy.
RQ3: 
How does the grading and feedback quality of GenAI compare with human assessment across different tasks, domains, and educational levels?
This is a central question relating to the value of using GenAI in assessment contexts.
RQ4: 
Which technical, task-related, and contextual factors influence the performance and reliability of GenAI-based grading and feedback?
This question recognizes the degree to which factors related to the technical side (e.g., prompting techniques), tasks (e.g., assessment of closed or open-ended student responses), and context (e.g., class subject) impact GenAI helpfulness in assessment contexts.
RQ5: 
What educational and pedagogical implications, opportunities, and risks are reported for integrating GenAI-based grading and feedback into assessment practice?
This question is helpful in recognizing the validity of the expectations and hopes associated with the use of GenAI in education.
Taken together, this review delineates the conditions that shape LLM performance in student grading relative to human raters and documents several recurring weaknesses, including decreased reliability in multilingual contexts and complex tasks that require nuanced judgment. The evidence to date suggests that GenAI is most appropriately deployed within a human-in-the-loop model that safeguards transparency and accuracy. We conclude by identifying significant methodological gaps in the existing scholarship and outlining a research agenda to guide future empirical investigations into the role of GenAI in educational assessment.

2. Methods

2.1. Eligibility Criteria and the Search

We were interested in the studies that did the following:
  • Focused on students across different educational levels, i.e., both elementary and higher education, as well as specific educational contexts like studies on only medical students or only English as a Foreign Language (EFL) students.
  • Involved the use of the LLM type of generative AI, both commercially available (like ChatGPT, Gemini, or Claude) and open sourced or custom LLMs. The usage of GenAI needed to be within the context of student grading (e.g., automatic scoring, response sheets, or descriptive grades). So, for instance, we were not interested in studies that involved using GenAI solely to provide feedback to students without the grading context.
  • Were empirical. Other reviews and purely opinion-based pieces were excluded from our review.
  • Were published in peer-review journals and conference proceedings after the official launch of ChatGPT in late 2022.
We did not exclude studies that did not involve direct comparisons of GenAI and humans, as we expected that a vast variety of empirical designs were being used to explore the usefulness of GenAI in student grading contexts. Also, while quantitative metrics give us a chance to compare different models and GenAIs directly—as do different measures of human vs GenAI agreement—given how quickly such metrics and other benchmarks change, we decided to neither include nor exclude phrases directly linked to outcome metrics.
The full list of search criteria is presented in Table 1. We searched for articles in four databases: Scopus, Web of Science, PubMed, and EBSCO. All the queries were performed in March 2025.

2.2. Screening and Quality Assessment

Our queries identified 1806 articles. After removing duplicates, we screened all the records based on title and abstract to assess their relevance. We then retrieved and assessed the full texts of potentially eligible studies against our inclusion and exclusion criteria.
All the screening steps were performed independently by two reviewers. Discrepancies were resolved through discussion. The number of records identified, screened, assessed for eligibility, and included in the final synthesis is presented in the PRISMA flow diagram (Figure 1).
In the end, 42 papers were included in the review. Their methodological quality was assessed using the Mixed Methods Appraisal Tool [28]. Following its guidelines, no quality score was calculated. Instead, criteria were individually assessed per study. Most studies met at least three criteria relevant to their respective designs. The most common issue was the use of small, convenience-based, or otherwise unrepresentative samples. Additionally, several studies lacked sufficient reporting on analytical procedures. Full details of the quality appraisal are provided in the Supplementary Materials.

3. Results

Detailed information on all studies analyzed in this review is presented in Table 2.

3.1. RQ1: In What Educational Contexts, Subjects, and Time Periods Have GenAI Models Been Used for Automated Grading and Feedback, and Which Concrete Models Have Been Deployed?

The studies analyzed in this review cover 2023–2025, reflecting the dynamic growth of interest in using LLMs for educational assessment. In 2023, the first pilots and case studies were published (5 studies). The year 2024 saw the most significant increase in publications (29 studies), including GPT-4 implementations and experiments with assessment across various subjects. In the first half of 2025 (8 studies), analyses emerged that extended the use of GenAI to assessing creativity, spoken responses, and high-stakes examinations, indicating the rapid maturation of this technology within education.
The studies analyzed in this article focus primarily on higher education—as many as 28 out of 42 publications concern using large language models to assess student work at universities and technical colleges. The remaining studies address education at the secondary school level (4 studies), primary school level (5 studies), and those categorized as K-12 (5 studies), i.e., studies marked as K-12, meaning they simultaneously cover both primary and secondary education.
In terms of academic disciplines and school subjects, the most frequently represented areas in the analyzed studies are the following:
  • Computer science—7 studies (mainly automatic assessment of programming tasks and code);
  • Foreign languages—7 studies (assessment of essays and written assignments in foreign languages);
  • Mathematics—5 studies (calculation tasks, mathematical reasoning);
  • Medicine—5 studies (exam questions, open-ended tasks in medical education);
  • Engineering—3 studies (engineering projects, open-ended tasks);
  • Pedagogy, social sciences, and humanities—single studies (assessment of essays, reflections, creativity).
Among the analyzed publications, the most frequently used model was GPT-4, which appeared in 23 studies. This was followed by GPT-3.5 (20 studies), the classic BERT model (7 studies), and the latest GPT-4o (3 studies). Considering the time needed for the peer-review process, this distribution indicates that researchers primarily focused on the most recent generative language models, while also conducting comparisons with earlier solutions.
Regarding model integration methods, 20 studies utilized API (enabling automated processing of large datasets). In contrast, in 16 cases, the data were entered manually into the ChatGPT interface—a solution widespread in pilot projects and experiments conducted on a smaller scale.
Taken together, most studies from the chosen period (2023–2025) are concentrated in higher education and cluster around computer science, foreign languages, mathematics, and medicine. The most frequently used models were ones from OpenAI, namely GPT-4 and GPT-3.5.

3.2. RQ2: How Are GenAIs Operationalized in Grading Workflows, Including Their Roles, the Types of Student Work They Assess, and the Prompting or Other Strategies Used?

In the analyzed studies, the most frequently assessed types of student tasks by language models were essays (16 studies, e.g., [64,70]) and short written answers (11 studies, e.g., [29,35,37]). Programming assignments appeared less often (5 studies, e.g., [33,49]), as did multiple-choice questions [39,62]. This distribution shows that researchers are particularly eager to test the potential of GenAI where content analysis and the ability to understand argumentation are essential.
In many analyzed studies, particular emphasis was placed on instructing language models before the assessment process. Most often, this involved providing exemplar solutions or detailed scoring rubrics—the model would receive an answer key or a description of the criteria it should use when evaluating assignments. In 32 cases, the prompts included a scoring rubric or a sample correct answer for comparison (e.g., [40,57,69]). Additionally, 8 studies applied holistic assessment criteria, such as overall impression [43,55].
In the included studies, GenAI played an active role in evaluating student work. However, the scope of this role varied. In most studies, artificial intelligence acted as an automatic grader, independently assigning points or grades to student responses with minimal human involvement, e.g., by generating a score for a short exam answer (e.g., [37,61,66]). In some cases, GenAI provided grades and generated feedback: the model assigned a score to the assignment and commented on its strengths and weaknesses, much as a human teacher would (e.g., [33,36,43]). There were also instances in which GenAI supported the teacher by suggesting a grade that the educator could then verify (so-called co-grading, e.g., [49]).
In sum, GenAI was used mainly as either an automatic grader or both a grader and feedback provider, or as a co-grader supporting human teachers, and it was most often applied to essays and short written answers. Additionally, GenAI was typically guided using rubric- or exemplar-based prompts.

3.3. RQ3: How Does the Grading and Feedback Quality of GenAI Compare with Human Assessment Across Different Tasks, Domains, and Educational Levels?

Findings on the alignment between GenAI and human graders were mixed. Several studies reported high agreement for specific tasks, including assessments of original thinking [34], macroeconomic and engineering tasks [32,40], programming assignments [33], in biology [35], in mathematics [68], and across other courses [37]. Notably, some evidence suggests that LLMs can surpass human raters in terms of reliability in particular contexts, like EFL essay scoring [36,69]. Nonetheless, longer essays remain challenging for GenAI [41,69]. In many cases, GenAI performance was comparable to human grading. For example, models from OpenAI and from iFLYTEK achieved similar accuracy scores in assessing Chinese writing [30]. Other studies indicate medium to good accuracy when compared to human assessment, e.g., in higher education exams and in medical education [47,65].
However, some studies also identified instances of GenAI performing worse or facing significant limitations. GenAI tends to perform worse in evaluating open-ended programming and mathematics tasks [48], in recognizing nuanced errors [69], and in assignments requiring subjective judgment [69]. Issues also arise regarding the consistency of grading [68], as well as the tendency for hallucinations [33] or for producing overly generic feedback [53]. More specifically, GPT-3.5 struggled with open-ended coding tasks, especially complex ones, sometimes producing false corrections and failing to distinguish basic programming naming conventions like capitalization [33]. GPT-4 had difficulties in visual programming tasks for children [58] and in calculus problem-solving and grading [68]. And while in many instances, the self-consistency of GenAI was high (e.g., [32,37,69]), in some cases multiple grading attempts yielded varying results, even with the usage of a predefined scoring rubric [55].
There is no one grading tendency identified across all studies. In some cases, GenAI was shown to be more lenient than human raters, assigning higher grades for average to good essays [41], and more positive in its feedback [64]. In others, it was more strict [32]. And finally, in several studies, GenAI displayed a tendency to assign middle scores, avoiding extreme grades [55,57].
LLM-based GenAIs were found to provide more comprehensive feedback than human graders [36,46]. But while larger in volume, such synthetic feedback was sometimes evaluated as too abstract, generic, or even confusing, with instances when the final grade did not align with the generated comments [55].
The analyzed studies yield mixed results when comparing GenAI and human assessment. GenAI performed worse than humans in 11 cases. In 12 instances, its effectiveness was comparable to human ratings. Regarding just the feedback quality, GenAI was worse 5 times, comparable 3 times, and better than humans in 3 cases.
The studies we looked at reveal a mixed picture when it comes to how well GenAI aligns with human graders. For shorter, more straightforward tasks, their performance tends to be similar. However, when it comes to longer, more complex, or subjective assignments, even though the feedback from GenAI can be more detailed, it often lacks the nuance that comes from human grading. In some cases, it felt too generic and did not quite match up with the scores given, or it was influenced by inaccuracies in the model’s understanding.

3.4. RQ4: Which Technical, Task-Related, and Contextual Factors Influence the Performance and Reliability of GenAI-Based Grading and Feedback?

The effectiveness of using GenAI for grading looks to be influenced by the quality of prompts, specific model of LLM being used, language of assessment, types of tasks, and level of education.
The final outcome is especially heavily influenced by the prompt given to the GenAI. Testing prompts beforehand and focusing on their precision, for instance by having detailed scoring rubrics supplemented with sample correct answers, seem to be especially vital for consistent and effective grading [43,48,60]. On the other hand, using specific strategies often recommended by online prompting courses may improve GenAI grading performance but also lead to overfit with simpler sub-problems, as in the case of chain-of-thought and tree-of-thought prompting, as well as with self-consistency checks [44,48].
Interestingly, only a handful of studies allow us to track the progress of subsequent generations of LLM models. The available publications rarely provide a basis for determining whether the newest models actually outperform their predecessors in grading or feedback quality. Initial results indicate that, for instance, GPT-4 generally outperforms GPT-3.5 in the accuracy of short-answer grading; however, this advantage is not yet consistent across all tasks and domains [36,37,54]. Fine-tuning and customizing models can help increase the reliability and precision of GenAI grading [69].
Language is another determinant. GenAI models perform best in English, although they are increasingly capable of handling other languages. Nevertheless, the effectiveness in detecting errors and linguistic nuances can vary depending on the student’s working language [37,64], which highlights the importance of attention when using multilingual models to grade linguistic tasks.
Another critical factor is the nature of the task itself and what the student has written. Artificial intelligence performs very well with short, clearly formulated answers; however, its effectiveness often decreases when it comes to more complex assignments requiring broader context or subjective judgment [37,43,54,69]. Interpreting non-textual works (e.g., drawings produced by hand) or very long and complex texts can also be challenging for GenAI [48].
On the other hand, GenAI demonstrates strong effectiveness in grading short-answer responses [37,60,61,65], in generating clear and well-structured feedback (though, as noted in other articles, this was also raised as a concern against GenAI-based assessment) [29], and in reducing grading bias. In some tasks, automated systems outperform human assessors in detecting language errors and in the speed and repeatability of grading [33,47].
Taken together, the main factors impacting GenAI grading performance in the reviewed studies are prompt and rubric quality, task type and complexity, model choice and its customization, and the language of assessment, with the effectiveness generally highest for short, clearly formulated responses under explicit and detailed criteria.

3.5. RQ5: What Educational and Pedagogical Implications, Opportunities, and Risks Are Reported for Integrating GenAI-Based Grading and Feedback into Assessment Practice?

The reviewed studies highlight several implications of using GenAI in educational assessment, from good practices regarding grading and feedback to broader instructional design and ethical considerations.
A recurrent theme is the potential of GenAI to reduce teacher workload while enabling more personalized and timely feedback, particularly in large-scale settings where manual grading is resource-intensive [45,47,56,58,65,69,70]. GenAI-assisted scoring can improve consistency by mitigating human fatigue and bias [43,64,70]. Moreover, formative assessments supported by LLMs can provide detailed explanations behind scores, helping both students and educators identify areas for improvement [33,44,46].
Beyond grading, GenAI contributes to instructional design. Models have been used to generate adaptive exercises tailored to individual learning needs [59], create exam questions and lesson plans, and expand the scope of what is evaluated to include more nuanced reasoning and argumentation [50,68]. Such generated tasks can help assess not only theoretical knowledge but also practical skills and communication abilities [48,67,68]. Fine-tuning models on discipline-specific literature further aligns feedback with pedagogical frameworks and curricular goals [45]. Such tasks can help assess not only theoretical knowledge but also practical skills and communication abilities [48,67,68].
Despite these opportunities, studies consistently emphasize the necessity of human oversight. Current LLMs struggle with errors, algorithmic biases, and loss of coherence, and they often lack full contextual understanding [32,41,49,57,64,65,68]. Ethical concerns ranging from privacy and transparency to fairness and accountability are particularly salient for GenAI use in educational settings [29,44,62,64,68,69]. Consequently, GenAI is widely regarded as an assistive rather than autonomous tool, with integration strategies emphasizing hybrid approaches and robust institutional safeguards.
In summary, the reviewed papers frame GenAI as an assistive tool that can reduce grading workload and accelerate feedback. However, it also consistently emphasized that human oversight and institutional safeguards are necessary due to errors, bias, limited contextual understanding, and ethical concerns.

4. Discussion

A systematic review of 42 empirical studies on the use of generative artificial intelligence for grading student work reveals promising opportunities and significant limitations of these technologies. The results are highly heterogeneous—no clear, universal trend exists across all the studies. A similar review by [71] found that existing studies on GenAI use in education lack a consistent and systematic approach from the educational perspective. To fill this gap, the authors mentioned above emphasize the key role of increased interdisciplinary, or even transdisciplinary, collaboration that would include educators and education researchers. We agree that only such joint efforts can ensure that the implementation of AI in education is aligned with pedagogical frameworks.
In many cases, GenAI models (especially GPT-4) aligned well with teachers on well-defined tasks: short answers in science and engineering, or structured problems in other domains, where expected responses are clear. Here, GenAI sometimes matched human ratings or even surpassed them in consistency. For example, in [44], GPT-4 achieved high agreement with human assessment in grading short open-answered science-related questions (Quadratic Weighted Kappa up to 0.95, and Macro F1 up to 1.00). Several studies confirmed high correlation or comparable accuracy between GenAI and teachers, suggesting that LLMs can imitate human grading in selected domains fairly well. For instance, in [32], the ICC of GPT-4 feedback across tasks ranged from 0.94 to 0.99, and in [40], GPT-4o’s assessment with human tutors had Spearman’s ρ exceeding 0.6 for most tasks. Also, both OpenAI and iFlytek models assessed Chinese essays comparably to instructors [30]. These results indicate that with clearly defined evaluation criteria, contemporary LLMs can serve as teaching assistants for specific tasks.
At the same time, the obtained results indicate that GenAI is far from being a universal or fully reliable grader. Performance dropped below that of human teachers in numerous situations, especially for open-ended, complex, or subjective assignments, particularly those requiring nuanced judgment, contextual understanding, or creativity. For instance, in [54], LLMs performed worse in such tasks, with the highest F1 score for GPT-4 being only 0.69, compared to 0.74 for shorter student answers. Characteristic minor errors, logical inconsistencies, or contextual mistakes, usually detected by a human teacher, may elude GenAI models, as was the case in [33], where LLM had a Recall metric of 0.57, meaning that much of the generated output was contextually inappropriate. Similar concerns were described in another review article [72]. In one study, GPT-3.5 struggled with assessing code correctness and occasionally invented false corrections. Even GPT-4 had difficulties with children’s visual logic tasks and advanced mathematics. Consistency was another recurrent issue: in some studies, the same model, using the same criteria, produced different grades across attempts. Phenomena such as GenAI hallucinations and overly generic feedback were also observed. These problems have also been noted in other review articles [73]. Such limitations demonstrate that current LLMs cannot fully replace teachers, particularly in situations where in-depth analysis, creativity, or subjective evaluation is required.
It is important to emphasize that there is no uniform pattern of bias in GenAI grading. In some cases, models were slightly more lenient than teachers (e.g., ChatGPT more often awarded somewhat higher grades, as in [70]), while in others, they were stricter (e.g., GPT-4 was harsher than humans when grading short answers in [39]). Some studies indicate that GenAI “avoids extreme grades,” tending to cluster around the middle of the scale (e.g., [65]). This means that the behavior of the models depends on both the model version and the context. Regardless of these differences, about 12 analyzed studies showed comparability of GenAI and human grades without significant deviations. Frequently, the feedback generated by GenAI was more extensive than that of teachers. Still, its quality was sometimes questioned: GenAI sometimes wrote too abstract or generic comments, not always matching the actual assessment.
The analysis also enables us to identify key factors that affect the quality of GenAI grading. One of the most important is prompt quality and the level of detail in the grading rubric. In many studies, prompt engineering and providing example answers or model criteria significantly increased the agreement between GenAI and teacher grades. Well-defined criteria (rubrics) are a condition for stable results. At the same time, more advanced prompting techniques, such as “chain-of-thought,” yielded mixed results—they improved grading in some types of tasks but led to redundancy or repetitiveness in others [44]. Another critical factor is the specific model version: although new models (e.g., GPT-4) usually perform better than earlier ones (e.g., GPT-3.5), this advantage is neither universal nor guaranteed across all tasks (e.g., [42]). How the capabilities of successive, newer models in grading student work are changing, and this requires further exploration. Attention is also drawn to the language of the assessed work—AI performs best in English, with effectiveness declining for other languages, especially those less represented in the training data [64]. GenAI generally works best for short, closed-ended answers where a specific response is expected, but it loses precision on open-ended tasks or those requiring contextual understanding (e.g., [37]). The principle “garbage in, garbage out” is key here, which requires educators to be skilled in “prompt engineering” and aware of the limitations of GenAI.
It can be said, then, that across the reviewed studies, agreement with human grades was strongest when tasks were short, tightly structured, and written in English; when models received detailed rubrics or exemplar answers; and when newer versions such as GPT-4 were used through stable API workflows (as in [32]). As tasks became longer, more open-ended, or non-English, accuracy and consistency often declined—sometimes regardless of model generation—unless prompts contained rich, explicit criteria and a human remained in the loop.
The discussed findings also have broad didactic implications, in agreement with earlier reviews such as [74,75]. GenAI can reduce teachers’ workload by taking over some of the routine grading, which in large student groups means faster feedback and greater process scalability. LLMs can generate detailed comments for each student and support learning personalization. Automation may also reduce certain forms of human bias and fatigue, improving fairness in large-scale assessments. GenAI also enables the development of new forms of assessment, such as generating quizzes or coding tasks tailored to the student’s level. However, nearly all studies emphasize that GenAI should only support the teacher, not replace them. Oversight is necessary to catch errors, address ethical issues such as bias and transparency, and provide the contextual and empathetic judgment that current LLMs cannot replicate. This oversight is especially critical for high-stakes grading contexts and wherever subjective interpretation is required for the assessment.

5. Conclusions

Based on 42 empirical studies published between 2023 and 2025, this review does not support replacing teachers with GenAI in assessment. Under specific conditions—short, well-structured tasks, explicit rubrics or exemplar answers, newer models (e.g., GPT-4), and stable API-based pipelines—LLMs can achieve accuracy and internal consistency comparable to human raters and can scale timely feedback.
These gains are not general. The effectiveness of GenAI-assisted grading strongly depends on the type of task, the quality of prompts and rubrics, the subject domain, and the language used to interact with GenAI. Performance declines for complex, open-ended, or subjective assignments requiring nuanced judgment, and in less-represented languages; issues with between-run consistency, hallucinations, and generic or misaligned feedback persist. No uniform grading bias emerged across the studies (sometimes more lenient, sometimes stricter, with a tendency in several cases to avoid extreme scores). Improvements over earlier model generations are measurable but uneven and task-dependent.
The review highlights that LLMs can support teachers by automating routine grading and generating rapid feedback at scale. Still, they cannot fully replace human judgment, especially in high-stakes or subjective assessments. GenAI-based scoring systems should be deployed as assistive tools, integrated within a hybrid, human-in-the-loop framework that maintains transparency, oversight, fairness, and adaptability to evolving educational needs. Human educators remain indispensable for validating grades and GenAI-produced personalized feedback on students’ progress.
Finally, this evidence base itself needs tightening. Future work should do the following:
  • Adopt common reporting standards and stronger sampling beyond convenience cohorts;
  • Track model generations using preregistered protocols;
  • Expand multilingual and modality-diverse evaluations;
  • Connect grading metrics to downstream learning outcomes and equity impacts.
Studies on the newest full-scaled models like ChatGPT 5, Gemini 2.5 Pro, and Claude Opus 4.1, most of which currently appear only in preprints and therefore fall outside the scope of this review, seem to confirm our recommendations, showing more consistent outputs regarding grading and assessment, albeit still requiring human oversight [76]. However, until the identified gaps narrow and published evidence shows even more robust reliability and quality of automated grading, GenAI can help make assessment faster and richer in feedback, but only where clarity, precision, rubrics, and human judgment meet.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/app16020680/s1.

Author Contributions

Conceptualization, M.J. and M.W.; methodology, M.J. and M.W.; software, M.J.; validation, M.J. and M.W.; formal analysis, M.J. and M.W.; investigation, M.J. and M.W.; resources, M.W.; data curation, M.W.; writing—original draft preparation, M.J. and M.W.; writing—review and editing, M.J. and M.W.; visualization, M.W.; supervision, M.J. and M.W.; project administration, M.J.; funding acquisition, M.J. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Polish National Science Centre (grant no. 2025/09/X/HS6/00038).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

Our thanks go to Aleksandra Wasielewska for invaluable remarks that improved this manuscript.

Conflicts of Interest

The authors have no competing interests to declare that are relevant to the content of this article.

References

  1. Jukiewicz, M. How generative artificial intelligence transforms teaching and influences student wellbeing in future education. Front. Educ. 2025, 10, 1594572. [Google Scholar] [CrossRef]
  2. Gill, S.S.; Xu, M.; Patros, P.; Wu, H.; Kaur, R.; Kaur, K.; Fuller, S.; Singh, M.; Arora, P.; Parlikad, A.K.; et al. Transformative effects of chatgpt on modern education: Emerging era of ai chatbots. Internet Things Cyber-Phys. Syst. 2024, 4, 19–23. [Google Scholar] [CrossRef]
  3. Al Murshidi, G.; Shulgina, G.; Kapuza, A.; Costley, J. How understanding the limitations and risks of using chatgpt can contribute to willingness to use. Smart Learn. Environ. 2024, 11, 36. [Google Scholar] [CrossRef]
  4. Castillo, A.; Silva, G.; Arocutipa, J.; Berrios, H.; Marcos Rodriguez, M.; Reyes, G.; Lopez, H.; Teves, R.; Rivera, H.; Arias-Gonzáles, J. Effect of chat gpt on the digitized learning process of university students. J. Namib. Stud. Hist. Politics Cult. 2023, 33, 1–15. [Google Scholar]
  5. Hung, J.; Chen, J. The benefits, risks and regulation of using chatgpt in chinese academia: A content analysis. Soc. Sci. 2023, 12, 380. [Google Scholar] [CrossRef]
  6. Raile, P. The usefulness of chatgpt for psychotherapists and patients. Humanit. Soc. Sci. Commun. 2024, 11, 47. [Google Scholar] [CrossRef]
  7. De Duro, E.S.; Improta, R.; Stella, M. Introducing counsellme: A dataset of simulated mental health dialogues for comparing llms like haiku, llamantino and chatgpt against humans. Emerg. Trends Drugs Addict. Health 2025, 5, 100170. [Google Scholar] [CrossRef]
  8. Maurya, R.K.; Montesinos, S.; Bogomaz, M.; DeDiego, A.C. Assessing the use of chatgpt as a psychoeducational tool for mental health practice. Couns. Psychother. Res. 2025, 25, 12759. [Google Scholar] [CrossRef]
  9. Memarian, B.; Doleck, T. Chatgpt in education: Methods, potentials, and limitations. Comput. Hum. Behav. Artif. Humans 2023, 1, 100022. [Google Scholar] [CrossRef]
  10. Diab Idris, M.; Feng, X.; Dyo, V. Revolutionizing higher education: Unleashing the potential of large language models for strategic transformation. IEEE Access 2024, 12, 67738–67757. [Google Scholar] [CrossRef]
  11. Hadi Mogavi, R.; Deng, C.; Juho Kim, J.; Zhou, P.; Kwon, Y.D.; Hosny Saleh Metwally, A.; Tlili, A.; Bassanelli, S.; Bucchiarone, A.; Gujar, S.; et al. Chatgpt in education: A blessing or a curse? A qualitative study exploring early adopters’ utilization and perceptions. Comput. Hum. Behav. Artif. Humans 2024, 2, 100027. [Google Scholar] [CrossRef]
  12. Davis, R.O.; Lee, Y.J. Prompt: Chatgpt, create my course, please! Educ. Sci. 2024, 14, 24. [Google Scholar] [CrossRef]
  13. Onal, S.; Kulavuz-Onal, D. A cross-disciplinary examination of the instructional uses of chatgpt in higher education. J. Educ. Technol. Syst. 2024, 52, 301–324. [Google Scholar] [CrossRef]
  14. Li, M. The impact of chatgpt on teaching and learning in higher education: Challenges, opportunities, and future scope. In Encyclopedia of Information Science and Technology, 6th ed.; IGI Global Scientific Publishing: Hershey, PA, USA, 2025; pp. 1–20. [Google Scholar]
  15. Lucas, H.C.; Upperman, J.S.; Robinson, J.R. A systematic review of large language models and their implications in medical education. Med Educ. 2024, 58, 1276–1285. [Google Scholar] [CrossRef] [PubMed]
  16. Jukiewicz, M. The future of grading programming assignments in education: The role of chatgpt in automating the assessment and feedback process. Think. Ski. Creat. 2024, 52, 101522. [Google Scholar] [CrossRef]
  17. Kıyak, Y.S.; Emekli, E. Chatgpt prompts for generating multiple-choice questions in medical education and evidence on their validity: A literature review. Postgrad. Med. J. 2024, 100, 858–865. [Google Scholar] [CrossRef] [PubMed]
  18. Ali, M.M.; Wafik, H.A.; Mahbub, S.; Das, J. Gen Z and generative AI: Shaping the future of learning and creativity. Cogniz. J. Multidiscip. Stud. 2024, 4, 1–18. [Google Scholar] [CrossRef]
  19. Ahmed, A.R. Navigating the integration of generative artificial intelligence in higher education: Opportunities, challenges, and strategies for fostering ethical learning. Adv. Biomed. Health Sci. 2025, 4, 1–2. [Google Scholar] [CrossRef]
  20. Reimer, E.C. Examining the role of generative ai in enhancing social work education: An analysis of curriculum and assessment design. Soc. Sci. 2024, 13, 648. [Google Scholar] [CrossRef]
  21. Tzirides, A.O.O.; Zapata, G.; Kastania, N.P.; Saini, A.K.; Castro, V.; Ismael, S.A.; You, Y.-l.; Santos, T.A.; Searsmith, D.; O’Brien, C.; et al. Combining human and artificial intelligence for enhanced ai literacy in higher education. Comput. Educ. Open 2024, 6, 100184. [Google Scholar] [CrossRef]
  22. Wu, D.; Chen, M.; Chen, X.; Liu, X. Analyzing k-12 ai education: A large language model study of classroom instruction on learning theories, pedagogy, tools, and ai literacy. Comput. Educ. Artif. Intell. 2024, 7, 100295. [Google Scholar] [CrossRef]
  23. Naveed, H.; Khan, A.U.; Qiu, S.; Saqib, M.; Anwar, S.; Usman, M.; Akhtar, N.; Barnes, N.; Mian, A. A comprehensive overview of large language models. ACM Trans. Intell. Syst. Technol. 2025, 16, 1–72. [Google Scholar] [CrossRef]
  24. Ramesh, D.; Sanampudi, S.K. An automated essay scoring systems: A systematic literature review. Artif. Intell. Rev. 2022, 55, 2495–2527. [Google Scholar] [CrossRef] [PubMed]
  25. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The prisma 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
  26. Page, M.J.; Moher, D.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. Prisma 2020 explanation and elaboration: Updated guidance and exemplars for reporting systematic reviews. BMJ 2021, 372, n160. [Google Scholar] [CrossRef] [PubMed]
  27. Rethlefsen, M.L.; Kirtley, S.; Waffenschmidt, S.; Ayala, A.P.; Moher, D.; Page, M.J.; Koffel, J.B.; Blunt, H.; Brigham, T.; Chang, S.; et al. Prisma-s: An extension to the prisma statement for reporting literature searches in systematic reviews. Syst. Rev. 2021, 10, 39. [Google Scholar] [CrossRef] [PubMed]
  28. Hong, Q.N.; Fàbregues, S.; Bartlett, G.; Boardman, F.; Cargo, M.; Dagenais, P.; Gagnon, M.-P.; Griffiths, F.; Nicolau, B.; O’Cathain, A.; et al. The mixed methods appraisal tool (mmat) version 2018 for information professionals and researchers. Educ. Inf. 2018, 34, 285–291. [Google Scholar] [CrossRef]
  29. Grévisse, C. Llm-based automatic short answer grading in undergraduate medical education. BMC Med Educ. 2024, 24, 1060. [Google Scholar] [CrossRef]
  30. Jiang, Z.; Xu, Z.; Pan, Z.; He, J.; Xie, K. Exploring the role of artificial intelligence in facilitating assessment of writing performance in second language learning. Languages 2023, 8, 247. [Google Scholar] [CrossRef]
  31. Li, Z.; Lloyd, S.; Beckman, M.; Passonneau, R.J. Answer-state recurrent relational network (asrrn) for constructed response assessment and feedback grouping. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2023, Singapore, 6–10 December 2023; Association for Computational Linguistics: Stroudsburg, PA, USA, 2023; pp. 3879–3891. [Google Scholar]
  32. Hackl, V.; Müller, A.E.; Granitzer, M.; Sailer, M. Is gpt-4 a reliable rater? evaluating consistency in gpt-4’s text ratings. Front. Educ. 2023, 8, 1272229. [Google Scholar] [CrossRef]
  33. Azaiz, I.; Deckarm, O.; Strickroth, S. Ai-enhanced auto-correction of programming exercises: How effective is gpt-3.5? arXiv 2023, arXiv:2311.10737. [Google Scholar] [CrossRef]
  34. Acar, S.; Dumas, D.; Organisciak, P.; Berthiaume, K. Measuring original thinking in elementary school: Development and validation of a computational psychometric approach. J. Educ. Psychol. 2024, 116, 953–981. [Google Scholar] [CrossRef]
  35. Schleifer, A.G.; Klebanov, B.B.; Ariely, M.; Alexandron, G. Transformer-based hebrew nlp models for short answer scoring in biology. In Proceedings of the 18th Workshop on Innovative Use of NLP for Building Educational Applications (BEA 2023), Toronto, ON, Canada, 13 July 2023; pp. 550–555. [Google Scholar]
  36. Li, J.; Huang, J.; Wu, W.; Whipple, P.B. Evaluating the role of chatgpt in enhancing efl writing assessments in classroom settings: A preliminary investigation. Humanit. Soc. Sci. Commun. 2024, 11, 1268. [Google Scholar] [CrossRef]
  37. Chang, L.-H.; Ginter, F. Automatic short answer grading for finnish with chatgpt. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 26–27 February 2024; Volume 38, pp. 23173–23181. [Google Scholar]
  38. Dong, D. Tapping into the pedagogical potential of infinigochatic: Evidence from iwrite scoring and comments and lu & ai’s linguistic complexity analyzer. Arab. World Engl. J. (AWEJ) Spec. Issue Chatgpt 2024, 124–137. [Google Scholar] [CrossRef]
  39. Thomas, D.R.; Lin, J.; Bhushan, S.; Abboud, R.; Gatz, E.; Gupta, S.; Koedinger, K.R. Learning and ai evaluation of tutors responding to students engaging in negative self-talk. In Proceedings of the Eleventh ACM Conference on Learning @ Scale, Atlanta, GA, USA, 18–20 July 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 481–485. [Google Scholar]
  40. Gao, R.; Guo, X.; Li, X.; Narayanan, A.B.L.; Thomas, N.; Srinivasa, A.R. Towards scalable automated grading: Leveraging large language models for conceptual question evaluation in engineering. In Proceedings of the Machine Learning Research, FM-EduAssess at NeurIPS 2024 Workshop, Vancouver Convention Center, Vancouver, BC, Canada, 15 December 2024. [Google Scholar]
  41. Jackaria, P.M.; Hajan, B.H.; Mastul, A.-R.H. A comparative analysis of the rating of college students’ essays by chatgpt versus human raters. Int. J. Learn. Teach. Educ. Res. 2024, 23, 478–492. [Google Scholar] [CrossRef]
  42. Jin, H.; Kim, Y.; Park, Y.S.; Tilekbay, B.; Son, J.; Kim, J. Using large language models to diagnose math problem-solving skills at scale. In Proceedings of the Eleventh ACM Conference on Learning @ Scale, Atlanta, GA, USA, 18–20 July 2024; pp. 471–475. [Google Scholar]
  43. Li, J.; Jangamreddy, N.K.; Hisamoto, R.; Bhansali, R.; Dyda, A.; Zaphir, L.; Glencross, M. Ai-assisted marking: Functionality and limitations of chatgpt in written assessment evaluation. Australas. J. Educ. Technol. 2024, 40, 56–72. [Google Scholar] [CrossRef]
  44. Cohn, C.; Hutchins, N.; Le, T.; Biswas, G. A chain-of-thought prompting approach with llms for evaluating students’ formative assessment responses in science. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 23182–23190. [Google Scholar]
  45. Zhang, M.; Johnson, M.; Ruan, C. Investigating sampling impacts on an llm-based ai scoring approach: Prediction accuracy and fairness. J. Meas. Eval. Educ. Psychol. 2024, 15, 348–360. [Google Scholar] [CrossRef]
  46. Shamim, M.S.; Zaidi, S.J.A.; Rehman, A. The revival of essay-type questions in medical education: Harnessing artificial intelligence and machine learning. J. Coll. Phys. Surg. Pak. 2024, 34, 595–599. [Google Scholar]
  47. Sreedhar, R.; Chang, L.; Gangopadhyaya, A.; Shiels, P.W.; Loza, J.; Chi, E.; Gabel, E.; Park, Y.S. Comparing scoring consistency of large language models with faculty for formative assessments in medical education. J. Gen. Intern. Med. 2025, 40, 127–134. [Google Scholar] [CrossRef] [PubMed]
  48. Lee, U.; Kim, Y.; Lee, S.; Park, J.; Mun, J.; Lee, E.; Kim, H.; Lim, C.; Yoo, Y.J. Can we use gpt-4 as a mathematics evaluator in education? Exploring the efficacy and limitation of llm-based automatic assessment system for open-ended mathematics question. Int. J. Artif. Intell. Educ. 2024, 35, 1560–1596. [Google Scholar] [CrossRef]
  49. Grandel, S.; Schmidt, D.C.; Leach, K. Applying large language models to enhance the assessment of parallel functional programming assignments. In Proceedings of the 1st International Workshop on Large Language Models for Code, LLM4Code ’24, Lisbon, Portugal, 20 April 2024; Association for Computing Machinery: New York, NY, USA, 2024; pp. 102–110. [Google Scholar]
  50. Martin, P.P.; Graulich, N. Navigating the data frontier in science assessment: Advancing data augmentation strategies for machine learning applications with generative artificial intelligence. Comput. Educ. Artif. Intell. 2024, 7, 100265. [Google Scholar] [CrossRef]
  51. Menezes, T.; Egherman, L.; Garg, N. Ai-grading standup updates to improve project-based learning outcomes. In Proceedings of the 2024 on Innovation and Technology in Computer Science Education V. 1, Milan, Italy, 8–10 July 2024; pp. 17–23. [Google Scholar]
  52. Zhang, C.; Hofmann, F.; Plößl, L.; Gläser-Zikuda, M. Classification of reflective writing: A comparative analysis with shallow machine learning and pre-trained language models. Educ. Inf. Technol. 2024, 29, 21593–21619. [Google Scholar] [CrossRef]
  53. Zhuang, X.; Wu, H.; Shen, X.; Yu, P.; Yi, G.; Chen, X.; Hu, T.; Chen, Y.; Ren, Y.; Zhang, Y.; et al. Toree: Evaluating topic relevance of student essays for chinese primary and middle school education. In Proceedings of the Findings of the Association for Computational Linguistics ACL 2024, Bangkok, Thailand, 11–16 August 2024; Association for Computational Linguistics: Stroudsburg, PA, USA, 2024; pp. 5749–5765. [Google Scholar]
  54. Sonkar, S.; Ni, K.; Tran Lu, L.; Kincaid, K.; Hutchinson, J.S.; Baraniuk, R.G. Automated long answer grading with ricechem dataset. In Proceedings of the International Conference on Artificial Intelligence in Education, Recife, Brazil, 8–12 July 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 163–176. [Google Scholar]
  55. Huang, C.-W.; Coleman, M.; Gachago, D.; Van Belle, J.-P. Using chatgpt to encourage critical ai literacy skills and for assessment in higher education. In Proceedings of the ICT Education, Gauteng, South Africa, 19–21 July 2023; Van Rensburg, H.E., Snyman, D.P., Drevin, L., Drevin, G.R., Eds.; Springer: Cham, Switzerland, 2024; pp. 105–118. [Google Scholar]
  56. Lu, Q.; Yao, Y.; Xiao, L.; Yuan, M.; Wang, J.; Zhu, X. Can chatgpt effectively complement teacher assessment of undergraduate students’ academic writing? Assess. Eval. High. Educ. 2024, 49, 616–633. [Google Scholar] [CrossRef]
  57. Agostini, D.; Picasso, F.; Ballardini, H. Large language models for the assessment of students’ authentic tasks: A replication study in higher education. In Proceedings of the CEUR Workshop Proceedings, Bolzano, IT, USA, 25–28 November 2024; Volume 3879. [Google Scholar]
  58. Tapia-Mandiola, S.; Araya, R. From play to understanding: Large language models in logic and spatial reasoning coloring activities for children. AI 2024, 5, 1870–1892. [Google Scholar] [CrossRef]
  59. Lin, N.; Wu, H.; Zheng, W.; Liao, X.; Jiang, S.; Yang, A.; Xiao, L. Indocl: Benchmarking indonesian language development assessment. In Proceedings of the Findings of the Association for Computational Linguistics: EMNLP 2024, Miami, FL, USA, 12–16 November 2024; pp. 4873–4885. [Google Scholar]
  60. Quah, B.; Zheng, L.; Sng, T.J.H.; Yong, C.W.; Islam, I. Reliability of chatgpt in automated essay scoring for dental undergraduate examinations. BMC Med Educ. 2024, 24, 962. [Google Scholar] [CrossRef] [PubMed]
  61. Henkel, O.; Hills, L.; Boxer, A.; Roberts, B.; Levonian, Z. Can large language models make the grade? An empirical study evaluating llms ability to mark short answer questions in k-12 education. In Proceedings of the Eleventh ACM Conference on Learning @ Scale, Atlanta, GA, USA, 18–20 July 2024; pp. 300–304. [Google Scholar]
  62. Latif, E.; Zhai, X. Fine-tuning chatgpt for automatic scoring. Comput. Educ. Artif. Intell. 2024, 6, 100210. [Google Scholar] [CrossRef]
  63. Xiao, X.; Li, Y.; He, X.; Fang, J.; Yan, Z.; Xie, C. An assessment framework of higher-order thinking skills based on fine-tuned large language models. Expert Syst. Appl. 2025, 272, 126531. [Google Scholar] [CrossRef]
  64. Fokides, E.; Peristeraki, E. Comparing chatgpt’s correction and feedback comments with that of educators in the context of primary students’ short essays written in english and greek. Educ. Inf. Technol. 2025, 30, 2577–2621. [Google Scholar] [CrossRef]
  65. Flodén, J. Grading exams using large language models: A comparison between human and ai grading of exams in higher education using chatgpt. Br. Educ. Res. J. 2025, 51, 201–224. [Google Scholar] [CrossRef]
  66. Gonzalez-Maldonado, D.; Liu, J.; Franklin, D. Evaluating gpt for use in k-12 block based cs instruction using a transpiler and prompt engineering. In Proceedings of the 56th ACM Technical Symposium on Computer Science Education V. 1, SIGCSETS 2025, Pittsburgh, PA, USA, 26 February–1 March 2025; Association for Computing Machinery: New York, NY, USA, 2025; pp. 388–394. [Google Scholar]
  67. Qiu, H.; White, B.; Ding, A.; Costa, R.; Hachem, A.; Ding, W.; Chen, P. Stella: A structured grading system using llms with rag. In Proceedings of the 2024 IEEE International Conference on Big Data (BigData), Washington, DC, USA, 15–18 December 2024; IEEE: New York, NY, USA, 2024; pp. 8154–8163. [Google Scholar]
  68. Gandolfi, A. Gpt-4 in education: Evaluating aptness, reliability, and loss of coherence in solving calculus problems and grading submissions. Int. J. Artif. Intell. Educ. 2025, 35, 367–397. [Google Scholar] [CrossRef]
  69. Yavuz, F.; Çelik, Ö.; Yavaş Çelik, G. Utilizing large language models for efl essay grading: An examination of reliability and validity in rubric-based assessments. Br. J. Educ. Technol. 2025, 56, 150–166. [Google Scholar] [CrossRef]
  70. Manning, J.; Baldwin, J.; Powell, N. Human versus machine: The effectiveness of chatgpt in automated essay scoring. Innov. Educ. Teach. Int. 2025, 62, 1500–1513. [Google Scholar] [CrossRef]
  71. Zhu, M.; Zhang, K. Artificial intelligence for computer science education in higher education: A systematic review of empirical research published in 2003–2023. In Technology, Knowledge and Learning; Springer: Berlin/Heidelberg, Germany, 2025; pp. 1–25. [Google Scholar]
  72. Pereira, A.F.; Mello, R.F. A systematic literature review on large language models applications in computer programming teaching evaluation process. IEEE Access 2025, 13, 113449–113460. [Google Scholar] [CrossRef]
  73. Albadarin, Y.; Saqr, M.; Pope, N.; Tukiainen, M. A systematic literature review of empirical research on chatgpt in education. Discov. Educ. 2024, 3, 60. [Google Scholar] [CrossRef]
  74. Lo, C.K.; Yu, P.L.H.; Xu, S.; Ng, D.T.K.; Jong, M.S.-y. Exploring the application of chatgpt in esl/efl education and related research issues: A systematic review of empirical studies. Smart Learn. Environ. 2024, 11, 50. [Google Scholar] [CrossRef]
  75. Dikilitas, K.; Klippen, M.I.F.; Keles, S. A systematic rapid review of empirical research on students’ use of chatgpt in higher education. Nord. J. Syst. Rev. Educ. 2024, 2, 103–125. [Google Scholar] [CrossRef]
  76. Jukiewicz, M. A systematic comparison of Large Language Models for automated assignment assessment in programming education: Exploring the importance of architecture and vendor. arXiv 2025, arXiv:2509.26483. [Google Scholar] [CrossRef]
Figure 1. PRISMA flow diagram.
Figure 1. PRISMA flow diagram.
Applsci 16 00680 g001
Table 1. Search strings by inclusion and exclusion criteria.
Table 1. Search strings by inclusion and exclusion criteria.
CategorySearch Terms/Filters
Inclusion: AI Technologies“ChatGPT” OR “Claude” OR “Gemini” OR “Copilot” OR “Deepseek” OR “generative AI” OR “generative artificial intelligence” OR “large language model*” OR “LLM*” OR “text-to-text model*” OR “multimodal AI” OR “transformer-based model*”
Inclusion: Education Contexts“adult education” OR “college*” OR “elementary school*” OR “graduate” OR “higher education” OR “high school*” OR “K-12” OR “kindergarten*” OR “middle school*” OR “postgrad*” OR “primary school*” OR “undergrad*” OR “vocational education”
Inclusion: Assessment Contexts“assessment” OR “grading” OR “scoring” OR “feedback” OR “AI grading” OR “automated assessment” OR “automated scoring” OR “automated grading” OR “automated feedback” OR “automatic assessment” OR “automatic scoring” OR “automatic grading” OR “automatic feedback” OR “machine-generated feedback”
Exclusion: Study TypesNOT “bibliometrics analysis” OR “literature review” OR “rapid review” OR “scoping review” OR “systematic review”
Exclusion: Time FrameAND PUBYEAR > 2022
Table 2. Selected LLM grading studies.
Table 2. Selected LLM grading studies.
StudyLevelFieldLLM TypeTask TypeKey Findings
[29]university or collegemedicinegpt4, gemini 1.0 proshort answersGPT-4 assigns significantly lower grades than the teacher, but it rarely incorrectly gives maximum scores to incorrect answers (low false positives). Gemini 1.0 Pro gives grades similar to those given by teachers—their grades did not differ significantly from human scores. Both GPT-4 and Gemini achieve moderate agreement with teacher assessments, but GPT-4 demonstrates high precision in identifying fully correct answers.
[30]secondary schoolforeign languagegpt3.5, gpt4, iFLYTEK, and Baidu CloudessayLLM models (GPT-4, GPT-3.5, iFLYTEK) achieved results very close to human raters in identifying correct and incorrect segments of text. Their accuracy at the T-unit level (syntactic segments) was about 81%, and at the sentence level 71–77%. GPT-4 stood out with the highest precision (fewest false positives), meaning it rarely erroneously marked incorrect answers as correct.
[31]university or collegemathothercalculation tasksEven with advanced prompting (zero/few-shot), GPT-3.5 performs noticeably worse than AsRRN. GPT-3.5 improves when reference examples are provided, but it does not reach the effectiveness of the relational model.
[32]university or collegeeconomicsgpt4short answersGPT-4 grades very consistently and reliably: agreement indices (ICC) for content and style scores range from 0.94 to 0.99, indicating nearly “perfect” consistency between repeated ratings on different sets and at different times.
[33]university or collegecomputer sciencegpt3.5coding tasksThe model correctly classified 73% of solutions as correct or incorrect. Its precision for correct solutions was 80%, and its specificity in detecting incorrect solutions was as high as 88%. However, recall—the ability to identify all correct solutions—was only 57%. In 89–100% of cases, AI-generated feedback was personalized, addressing the specific student’s code. This was much more detailed than traditional e-assessment systems, which usually rely on simple tests.
[34]primary schoolnagpt3otherAI grades were in very high agreement with the ratings of a panel of human experts (r = 0.79–0.91 depending on the task type), meaning that AI can generate scores comparable to subjective human judgments. The model demonstrates high internal consistency of ratings (reliability at H = 0.82–0.89 for different subscales and overall originality), making AI a reliable tool in educational applications.
[35]secondary schoolbiologybertshort answersThe alephBERT-based system clearly outperforms classic CNN models in automatic grading of students’ short biology answers, both in overall and detailed assessment (for each category separately). The model showed surprisingly good performance even in “zero-shot” mode—that is, when grading answers to new questions (about the same biological concepts) that it had not seen during training. This means that AI can be used to automatically grade new tasks, as long as they relate to similar concepts/rubrics.
[36]university or collegeforeign languagegpt4, gpt3.5essayChatGPT 4 achieved higher reliability coefficients for holistic assessment (both G-coefficient and Phi-coefficient) than academic teachers evaluating English as a Foreign Language (EFL) essays. ChatGPT 3.5 had lower reliability than the teachers but was still a supportive tool. In practice, ChatGPT 4 was more consistent and predictable in its scoring than humans. Both ChatGPT 3.5 and 4 generated more feedback, and more relevant feedback, on language, content, and text organization compared to the teachers.
[37]university or collegemixedgpt4, gpt3.5short answersGPT-4 performs well in grading short student answers (QWK ≥ 0.6, indicating good agreement with human scores) in 44% of cases in the one-shot setting, significantly better than GPT-3.5 (21%). The best results occur for simple, narrowly defined questions and short answers; open and longer, more elaborate answers are more challenging. ChatGPT (especially GPT-4) tends to give higher grades than teachers—less often awarding failing grades and more often scoring higher than human examiners.
[38]university or collegeforeign languageInfinigo ChatICessayTexts “polished” by ChatGPT (infinigoChatIC) receive much higher scores in automated grading systems (iWrite) than texts written independently or with classic machine translators. The average score increase was about 17%. AI effectively eliminates grammatical errors and significantly increases the diversity and sophistication of vocabulary
(e.g., more complex and precise expressions). This results in higher scores for both language and technical/structural aspects of the text.
[39]university or collegenagpt4multiple choice tasksGenerative AI (GPT-4) achieved high agreement with human expert ratings when grading open tutor responses in training scenarios (F1 = 0.85 for selecting the best strategy, F1 = 0.83 for justifying the approach). The Kappa coefficient indicates good agreement (0.65–0.69), although the AI scored somewhat more strictly than humans and was more sensitive to answer length (short answers were more often marked incorrect). Overall, AI grading was slightly more demanding than human grading (average scores were about 25% lower when graded by AI).
[40]university or collegeengineeringgpt4oshort answersGPT-4o (ChatGPT) showed strong agreement with teaching assistants (TAs) in grading conceptual questions in mechanical engineering, especially where grading criteria were clear and unambiguous (Spearman’s p exceeded 0.6 in 7 out of 10 tasks, up to 0.94; low RMSE).
[41]k12mixedgpt3.5essayEssay scores assigned by ChatGPT 3.5 show poor agreement with those assigned by experienced teachers. Agreement was low in three out of four criteria (content, organization, language) and only moderate for mechanics (e.g., punctuation, spelling). On average, ChatGPT 3.5 gave slightly higher grades than humans in each category, but the differences were not statistically significant (except for mechanics, where AI was noticeably more “lenient”).
[42]primary schoolmathgpt3.5, gpt4calculation tasksA system based on GPT-3.5 (fine-tuned) achieved higher agreement with teacher ratings than “base” GPT-3.5 and GPT-4, especially for clear, unambiguous criteria.
[43]university or collegecomputer science, medicinegpt4essay, coding tasksChatGPT grading is more consistent and predictable for high-quality work. With weaker or average work, there is greater variability in scores (wider spread), especially in programming tasks.
[44]secondary schoolmixedgpt4short answersGPT-4 using the “chain-of-thought” (CoT) method and few-shot prompting achieved high agreement with human examiners when grading short open-ended answers in science subjects (QWK up to 0.95; Macro F1 up to 1.00). The CoT method allowed AI not only to assign a score but also to generate justification (explanation of the grading decision) in accordance with the criteria (rubrics).
[45]k12mixedbertshort answersLLM models (here: BERT) achieve high agreement with human ratings in short open-ended response tasks in standardized educational tests
(QWK ≥ 0.7 for 11 out of 13 tasks).
[46]university or collegemedicinegpt3.5essayChatGPT 3.5 could grade and classify medical essays according to established criteria, while also generating detailed, individualized feedback for students.
[47]university or collegemedicinegpt3.5essayChatGPT 3.5 achieves comparable grading performance to teachers in formative tasks (including critical analysis of medical literature). Overall agreement between ChatGPT and teachers was 67%, and AUCPR (accuracy index) was 0.69 (range: 0.61–0.76). The reliability of grading (Cronbach’s alpha) was 0.64, considered acceptable for formative assessment.
[48]primary schoolmathgpt4calculation tasksGPT-4, used as an Automated Assessment System (AAS) for open-ended math responses, achieved high agreement with the ratings of three expert teachers for most questions, demonstrating the potential of AI to automate the grading of complex assignments, not just short answers.
[49]university or collegecomputer sciencegpt4coding tasksChatGPT-4, used as part of the semi-automatic GreAIter tool, achieves very high agreement with human examiners (98.2% accuracy, 100% recall, meaning no real errors in code are missed). Despite this high effectiveness, AI can be overly critical (precision 59.3%)—it generates more false alarms (incorrectly detected errors) than human examiners.
[50]university or collegechemistrygpt3.5, bardshort answersGenerative AI (e.g., ChatGPT, Bard) can significantly improve the grading of open-ended student work in natural sciences, especially thanks to data augmentation and more precise mapping of students’ reasoning levels.
[51]university or collegecomputer sciencegpt3short answersAI (GPT-3, after fine-tuning) achieved very high agreement with human ratings in classifying the specificity of statements (97% accuracy) and sufficient accuracy in assessing effort (83%). Zero-shot models (without training) were much less accurate—automatic grading performance increases after fine-tuning on human-graded data.
[52]university or collegeeducationbertessayPretrained models on large datasets, such as BigBird and Longformer, achieved the highest effectiveness in classifying the level of reflection in student assignments (77.2% and 76.3% accuracy, respectively), clearly outperforming traditional “shallow” ML models (e.g., SVM, Random Forest), which did not exceed 60% accuracy.
[53]k12mixedgpt3.5turbo, claude3.5, gpt4turbo, sparkessayGenerative models such as GPT-4 often incorrectly assess the topic relevance of essays and generate feedback that is too general, impractical, or even misleading. This phenomenon results partly from “sycophancy” (excessive leniency) and LLM hallucinations—the models tend to give overly optimistic evaluations.
[54]university or collegechemistrybert, gpt3.5, gpt4otherLanguage models (LLMs, e.g., GPT-4) perform worse in grading long, multi-faceted, factual answers than in grading short responses—even with advanced techniques and rubrics, F1 for GPT-4 was only 0.69 (for short responses, 0.74). Using a detailed rubric (i.e., breaking grading into many small criteria) improves AI’s results by 9% (accuracy) and 15% (F1) compared to classical, general scoring. Rubrics help the model better identify which aspects of the answer are correct or incorrect. GPT-4 achieves the highest results among the tested models (F1 = 0.689, accuracy 70.9%) but still does not perform as well as for short answers and often misses nuances in complex, multi-faceted student answers.
[55]university or collegecomputer scienceotheressayChatGPT should not be used as a standalone grading tool. AI-generated grades are unpredictable, can be inconsistent, and do not always match the criteria used by teachers.
[56]university or collegeeducationotheressayChatGPT 3.5 achieved moderate or good agreement with teacher grades (ICC from 0.6 to 0.8 in various categories), meaning it can effectively score student work based on the same criteria as humans. ChatGPT generated more feedback, often giving general suggestions and summaries, but less frequently providing specific solutions and explanations. AI feedback was less precise, less detailed, and less emotionally supportive, but more “inspiring” for some students. Students implemented teacher suggestions more often (80.2%) than ChatGPT’s (59.9%), and AI feedback was more frequently rejected (40% of cases).
[57]university or collegemixedgpt4o, claude3.5, gemini1.5prootherThe newest models (Claude 3.5 Sonnet, GPT-4o, Qwen2 72B) can be a valuable support in grading complex, open-ended tasks (authentic assignments) in higher education, using rubric-based assessment. GPT-4o and Claude 3.5 Sonnet produce results closest to human grading in terms of score distribution and consistency.
[58]primary schoolmathgpt4ocalculation tasksGPT-4o can effectively support the assessment of tasks requiring logical and spatial reasoning in children, provided that advanced prompting techniques are used. Simple prompting (zero-shot/few-shot) is insufficient—grading effectiveness by LLMs increases significantly when using advanced methods, such as visualization of thought and self-consistency (multiple generations with majority voting). In the best settings, GPT-4o achieved 93–100% correct grades on tasks testing logical and spatial skills, such as coloring objects according to math-logical-spatial rules.
[59]university or collegeforeign languagebert, gpt3.5turboessayChatGPT (GPT-3.5-turbo) performs much worse than dedicated BERT or SIAM models in the task of assessing whether a pair of student essays represents progression (improvement) or regression in language skills. ChatGPT’s effectiveness in this task was significantly lower (accuracy ∼56%) than the best models based on BERT architecture.
[60]university or collegemedicinegpt4essayThe results obtained by ChatGPT strongly correlate with human ratings, especially for questions having clearer scope. AI enables immediate grading of large numbers of assignments and provides individual, detailed, rubric-based feedback—which is hard for a single teacher to deliver with large groups.
[61]k12mixedgpt4, gpt3.5short answersThe GPT-4 model using few-shot prompting achieved human-level agreement (Cohen’s Kappa 0.70), very close to expert level (Kappa 0.75). AI grades were stable regardless of subject (history/science), task difficulty, and student age.
[62]secondary schoolnabert, gpt3.5turbomultiple choice tasksTuned ChatGPT outperforms the state-of-the-art BERT model in both multi-label and multi-class tasks. The average grading performance improvement for ChatGPT was about 9.1% compared to BERT, and for more complex (multi-class) tasks even over 10%. The “base” version of GPT-3.5 is not effective enough—tuning is necessary for subject/class/task-specific data.
[63]university or collegecomputer sciencebertcoding tasksA fine-tuned LLM can automatically assess student work in terms of higher-order cognitive skills (HOTS) without expert involvement, with greater accuracy and scalability than traditional systems.
[64]primary schoolforeign languagegpt4truboessayScores assigned by ChatGPT 3.5 were generally comparable to those given by academic staff. The average exact agreement between AI and humans was 67%. For individual criteria, agreement ranged from 59% to 81%. Using ChatGPT for grading reduced grading time fivefold (from about 7 min to 2 min per assignment), potentially saving hundreds of teacher work hours.
[65]university or collegeengineeringgpt3.5short answersChatGPT 3.5 more often assigns average grades (e.g., C, D, E) and less frequently than humans gives the highest or lowest marks. ChatGPT is consistent—repeated grading of the same work yielded similar results.
[66]k12computer scienceothercoding tasksThe average effectiveness of GPT autograder for simple tasks was 91.5%, and for complex ones only 51.3%. There was a strong correlation between LLM and official autograder scores (r = 0.84), but GPT-4 tended to underestimate scores, i.e., giving students lower results than they actually deserved (prevalence of so-called false negatives).
[67]university or collegebiologygpt4short answersSteLLA achieved “substantial agreement” with human grading—Cohen’s Kappa was 0.67 (compared to 0.83 for two human graders), and raw agreement was 0.84 (versus 0.92 for humans). This means the automated grading system is close to human-level agreement, though not equal to it yet.
[68]university or collegemathgpt4, gpt3.5calculation tasksThe average grade given by GPT-4 for student solutions is very close to grades assigned by human examiners, but detailed accuracy is lower—random errors occur, equivalent solutions can be missed, and calculation mistakes appear.
[69]university or collegeforeign languagegpt4, bardessayThe highest reliability was obtained for the FineTuned ChatGPT version
(ICC = 0.972), with slightly lower reliability for ChatGPT Default (ICC = 0.947) and Bard (ICC = 0.919). The reliability of LLMs exceeded even some human examiners in terms of repeatability. The best match between LLM and human examiners was for “objective” criteria such as grammar and mechanics. Both ChatGPT and Bard almost perfectly matched human scores for grammar and mechanics (punctuation, spelling).
[70]university or collegeengineering, foreign languagegpt4, gpt3.5essayChatGPT (both 3.5 and 4) grades student essays more “generously” than human examiners. The average grades assigned by ChatGPT are higher than those given by teachers, regardless of model version. AI often gives higher grades even if the work quality is not high, leading to a “ceiling effect” (grades clustering near the top end). The agreement between human examiners was high (“good”), while for ChatGPT-3.5 it was only “moderate” and for ChatGPT-4 even “low”.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jukiewicz, M.; Wyrwa, M. Can ChatGPT Replace the Teacher in Assessment? A Review of Research on the Use of Large Language Models in Grading and Providing Feedback. Appl. Sci. 2026, 16, 680. https://doi.org/10.3390/app16020680

AMA Style

Jukiewicz M, Wyrwa M. Can ChatGPT Replace the Teacher in Assessment? A Review of Research on the Use of Large Language Models in Grading and Providing Feedback. Applied Sciences. 2026; 16(2):680. https://doi.org/10.3390/app16020680

Chicago/Turabian Style

Jukiewicz, Marcin, and Michał Wyrwa. 2026. "Can ChatGPT Replace the Teacher in Assessment? A Review of Research on the Use of Large Language Models in Grading and Providing Feedback" Applied Sciences 16, no. 2: 680. https://doi.org/10.3390/app16020680

APA Style

Jukiewicz, M., & Wyrwa, M. (2026). Can ChatGPT Replace the Teacher in Assessment? A Review of Research on the Use of Large Language Models in Grading and Providing Feedback. Applied Sciences, 16(2), 680. https://doi.org/10.3390/app16020680

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop