Next Article in Journal
Federated Learning with Adversarial Optimisation for Secure and Efficient 5G Edge Computing Networks
Previous Article in Journal
Characterizing User Archetypes and Discussions on Social Hypernetworks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

A Meta-Survey of Generative AI in Education: Trends, Challenges, and Research Directions

1
Department of Computer Science, Modeling, Electronics and Systems Engineering (DIMES), University of Calabria, Via P. Bucci, 87036 Rende, Italy
2
School of Computer Science and Engineering, South China University of Technology, Guangzhou 510641, China
3
Pazhou Laboratory, Guangzhou 510640, China
*
Authors to whom correspondence should be addressed.
Big Data Cogn. Comput. 2025, 9(9), 237; https://doi.org/10.3390/bdcc9090237
Submission received: 24 July 2025 / Revised: 4 September 2025 / Accepted: 11 September 2025 / Published: 16 September 2025

Abstract

Education is experiencing a paradigm shift, evolving from traditional learning methods to computer-tool-based education, and now toward the integration of Generative Artificial Intelligence. While classical methods offer structured and standardized learning, they often do not fully address individual learner needs and accessibility. The rise of digital technologies introduced adaptive learning platforms, online classrooms, and interactive educational tools, expanding the reach and flexibility of educational systems. Today, Generative Artificial Intelligence tools are redefining the education landscape by personalized learning experiences, automating content generation, and providing real-time feedback. Intelligent tutoring systems and personalized assessments empower students with customized learning pathways that enhance engagement and academic performance. This paper presents a meta-survey that systematically examines the role of Generative Artificial Intelligence in education, following PRISMA guidelines to analyze trends, frameworks, and research outcomes across a curated body of academic literature. Special attention is given to the emergence of commercial Generative Artificial Intelligence tools, which are increasingly embedded in learning environments. A structured comparison framework and research questions guide the review, offering insights into how Generative Artificial Intelligence technologies are shaping pedagogical practices, influencing assessment, and raising new ethical and technical challenges. The paper also explores future directions, highlighting how Generative Artificial Intelligence is driving the emergence of new learning models.

1. Introduction

The landscape of educational technology is undergoing a profound transformation, driven by rapid advancements in Generative Artificial Intelligence (Gen AI). While the transition from traditional to digital learning models has been underway for decades, the public release of large language models (LLMs) such as ChatGPT (version 4, OpenAI, released March 2023) in late 2022 accelerated research and development at an unprecedented scale. Gen AI tools now demonstrate significant capabilities in generating educational content, supporting personalized learning, and mimicking aspects of human creativity [1]. These technologies are reshaping pedagogy by enabling adaptive tutoring systems, customized content recommendations, and interactive learning environments that respond to the needs of individual students [2,3].
The evolution of education has historically followed technological advances: from classical methods centered on books and lectures [4,5], to digital and e-learning models that expanded access and flexibility, particularly during the COVID-19 pandemic [6,7], and now to Generative AI. Each phase has reshaped pedagogy in distinct ways, but Gen AI represents a qualitatively new stage [8]. Unlike earlier technologies that primarily delivered or organized content, Gen AI actively generates, adapts, and personalizes learning experiences, raising novel opportunities and challenges for research and practice [2].
This rapid growth has been accompanied by a surge in academic publications. A preliminary analysis indicates that more than 40 survey studies on the integration of Gen AI in education have been published since 2022. However, this proliferation has led to a fragmented research landscape. Many reviews focus on specific tools (e.g., chatbots), single educational levels (e.g., higher education) [9], or isolated themes such as ethics and academic integrity [10]. As a result, there is still no consolidated synthesis that compares definitions, frameworks, and applications across contexts. Moreover, few reviews systematically address the pedagogical and institutional impact of commercial Gen AI tools that are increasingly being adopted in academic practice.
To address these gaps, this paper presents a systematic meta-survey of the existing literature on Gen AI in education. By adopting a tertiary review approach—a survey of surveys—we synthesize fragmented findings into a coherent overview of the state of the art. Our methodology is guided by two principles:
  • Comprehensiveness: The review follows the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines [11], a rigorous and reproducible protocol that ensures transparent selection of high-quality studies.
  • Effectiveness: The meta-survey approach aggregates findings from a broad range of secondary studies, enabling generalization and synthesis that make the rapidly expanding research landscape more accessible and actionable.
The structure of the paper is as follows: Section 2 outlines the research objectives and methodology. Section 3 analyzes publication and citation trends from 2022 to April 2025. Section 4 synthesizes the collected literature, focusing on key trends, tools, and frameworks. Section 5 discusses future research directions and emerging learning models enabled by Gen AI. Finally, Section 6 concludes with implications for educators, researchers, and policymakers.

2. Methodology

A systematic analysis of Gen AI applications in education was conducted, covering the period from 2022 to April 2025 (since ChatGpt’s emergence). The study followed the PRISMA guidelines to ensure a comprehensive and structured approach to the literature review and data synthesis, and after completing the record screening and report selection process, we analyzed a total of 40 works. From these, we conducted an additional manual selection, resulting in a systematic analysis literature of 21 papers. The research included the following:
  • Original studies that explore new Gen AI applications in education.
  • Comprehensive surveys and review papers analyzing theoretical frameworks, educational objectives, and case studies that highlight the opportunities and challenges of integrating Gen AI into education.
The undertaken search plan is summarized in Table 1, while Figure 1 depicts the flow-chart of the PRISMA-based selection process.

2.1. Objectives

This review aims to explore the state of the art in Gen AI at the intersection of AI and education. The survey was conducted to analyze current research trends, identify key challenges related to Gen AI, and address the following research questions (RQs):
  • (RQ1): What are current definitions of Gen AI ?
  • (RQ2): What are the current methodologies and frameworks for integrating Gen AI into educational systems?
  • (RQ3): Which educational levels have been the primary focus of research on Gen AI ?
  • (RQ4): What are the primary technical, pedagogical, and ethical challenges associated with the adoption of Gen AI in education?
  • (RQ5): How effective are Gen AI tools in fostering critical thinking and problem-solving skills among students?

2.2. Search Strategy

The search was conducted in February 2025 using a wide range of keywords, and the dataset was subsequently updated with publication and citation information available up to April 2025. To extract the most relevant documents, search queries were conducted using predefined keywords and Boolean operators across multiple digital libraries. The search aimed to identify scientific publications that explore the application and the role of Gen AI in education. The keyword search string was structured around three key concepts: (i) Generative Artificial Intelligence, (ii) Gen AI tools, and (iii) large language models. Based on these terms and their synonyms, the following search string was formulated:
( Generative   Artificial   Intelligence OR   Gen   AI   tools OR   Large   Language   Models ) AND   ( Education OR   Learning OR   Higher Education )
To ensure the identification of relevant results, the search string was applied to article titles with a constraint allowing a maximum gap of two words between key terms. Subsequently, considering the study’s focus on Gen AI research, the following search string was applied to the abstracts of the manuscripts:
( Survey   OR   Review   OR   Literature )

2.3. Eligibility Criteria

To ensure that only highly relevant research is included in this analysis, inclusion and exclusion criteria were established. The articles were eligible for selection if they met all of the following inclusion criteria:
  • If the work is a survey or literature review on the use of Gen AI in education.
  • The work discusses the application of Gen AI in an educational context.
  • If the work focuses on AI-generated content, personalized learning, adaptive assessment, or student support.
  • If the work proposes a solution or framework involving Gen AI in education.
  • If the work is based on Gen AI techniques.
  • If the work addresses subject-specific applications (e.g., programming, engineering, etc.) within the broader context of Gen AI in education, rather than as a domain-specific focus alone.
  • If the work includes a model or provides a formal definition of Gen AI.
The articles were excluded from the selection if they met one of the following exclusion criteria:
  • Studies that discuss Gen AI in general but do not address educational application, such as in healthcare [12,13,14], finance [15,16], and nursing [17,18].
  • Studies published in preprints and the gray literature, including theses, dissertations, technical reports, opinions, discussions, editorials, viewpoints, comments, tutorials, lessons, pilot studies, and presentations.
  • Research on Gen AI that was not written in the English language [19,20].
  • The terms related to Gen AI are mentioned only in the title, abstract, or keywords but are not addressed in the main body of the work [21].
  • The study lacks a clear focus on the role of Gen AI in the field of education.
  • Unpublished works [21,22].
  • Studies focused on AI in education but do not involve Gen AI.
For all studies meeting the eligibility criteria, study metadata (e.g., publication year, venue, methodology, frameworks, and tools) were extracted by the first author and independently verified by the co-authors. Any disagreements regarding data extraction or classification were resolved through discussion until consensus was reached. No automation tools were employed in this process.

2.4. Study Selection

The study selection process followed PRISMA 2020 guidelines, as shown in Figure 1. A total of 130 records were initially retrieved. After removing 33 papers based on predefined technical criteria (e.g., unpublished works, pre-2022 publications, non-English papers, and incomplete content), 97 remained for screening. During the title, abstract, and keyword screening, two authors independently assessed all records, excluding 32 papers that provided only a superficial mention of Gen AI. Disagreements were resolved through discussion with the other co-authors until consensus was reached. A further 23 papers were excluded due to insufficient methodological detail ( n = 8 ) or limited Gen AI contribution beyond superficial mention ( n = 15 ). This left 40 papers for full-text review. In the final quality screening phase, which was conducted collaboratively by the research team, 19 papers were excluded for lack of rigor or relevance, leaving 21 secondary studies.
Out of the 40 eligible articles, 21 reviews met all four methodological quality indicators (see Table 2) and were therefore included in the final synthesis. These 21 reviews form the basis of our in-depth synthesis and are reported in Table 4.

2.5. Risk of Bias Assessment

To ensure the validity of our synthesis, we assessed the methodological quality of the included secondary studies, guided by the principles of the PRISMA 2020 statement [11] Given that this is a meta-survey, a formal application of tools like ROBIS [44] or AMSTAR-2, which are designed for primary systematic reviews, was not directly applicable. Our objective was not to validate the individual data points within the reviews, but to appraise the methodological soundness of the reviews themselves.
We therefore implemented a qualitative assessment using four indicators of review quality: (i) articulation of research questions and eligibility criteria; (ii) use of a systematic and reproducible search strategy; (iii) transparency in reporting study selection and inclusion/exclusion; and (iv) provision of a coherent synthesis of findings aligned with stated objectives.
This review was conducted in accordance with the PRISMA 2020 guidelines [11]. The review was not registered in a database, and no prior protocol was prepared.
This process allowed us to systematically identify potential sources of bias at the review level. During our synthesis, findings from reviews that lacked transparency in these key areas were interpreted with caution and weighted as less robust in our overall analysis. Our evaluation revealed that 12 of the 21 reviews employed transparent search strategies, fewer than half explicitly reported eligibility criteria, and only 7 documented a formal synthesis approach. Table 2 summarizes the results, showing that while most reviews clearly stated research questions and provided coherent syntheses, the reporting of inclusion/exclusion and methodological transparency was inconsistent. This uneven rigor highlights the need for greater methodological standardization in Gen AI educational reviews.

3. Results

A total of 40 articles met the eligibility criteria. This full set is used to depict overall publication and citation trends (Figure 2 and Figure 3). Following the quality assessment described in Section 2, 21 reviews were retained for the in-depth synthesis presented in Section 4 and Section 5.
From January 2022 to April 2025, monthly output in Generative AI in education rose sharply (Figure 2 and Figure 3). Activity in 2022 is sporadic (three surveys in total). In 2023, production becomes more regular (seven in total), and 2024 exhibits a pronounced surge (twenty), accounting for roughly half of all included surveys. The partial window for 2025 (January–April) already contains 10 surveys, indicating continued momentum; because only four months are observed, totals are not directly comparable to full prior years, but the monthly pace is clearly elevated. This upswing temporally coincides with the mainstreaming and rapid uptake of large generative models beginning in late 2022. The descriptive counts use all 40 eligible works; the interpretive synthesis is restricted to the 21 high-quality reviews (Section 2).
Baseline comparator and monthly data construction: To contextualize these dynamics, we established a monthly baseline from a closely related AI+domain AI in Medical Diagnostics/Health and Medicine. A dedicated search was conducted in Scopus and cross-checked in Web of Science using the keywords “medical” AND “Generative AI” (and their controlled vocabulary equivalents). The retrieved records were filtered using the same eligibility criteria applied in our main review: (i) peer-reviewed journal articles or systematic reviews; (ii) exclusion of editorials, conference abstracts, and non-scholarly notes/letters; (iii) English-language publications; and (iv) availability of complete bibliographic metadata. Records were grouped by publication month to construct the baseline time series; items lacking monthly metadata were excluded from the monthly aggregation. Where month-level citation metadata were incomplete, monthly publication counts were used as a transparent proxy, and this limitation is explicitly noted in the figure captions. We selected this comparator because the AI-in-medicine literature is both mature and extensively characterized in prior bibliometric syntheses [45,46], providing a conservative and credible benchmark against which to interpret the acceleration observed in Generative AI in education. The comparison reveals that, while both domains exhibit steady growth over the observed period, the post-2023 rise in Generative AI in education is significantly steeper, consistent with a domain-specific surge in scholarly attention rather than a background effect of general AI research expansion.
Baseline comparison: Compared to Gen AI in Medical Diagnostics, both domains show strong post-2022 growth. However, Gen AI-in-education exhibits a steeper post-2023 acceleration (especially through 2024), indicating a domain-specific surge rather than merely reflecting a general AI upswing. This conclusion holds when baseline 2025 values are treated as partial/omitted in line with source availability.

Geographic Origin of the Reviewed Studies

To provide a broader understanding of the research landscape, we analyzed the geographic origin of the 21 included studies based on the affiliation of the first author (see Table 3).
The distribution highlights a concentration of contributions from Asia (China, India, Malaysia, Qatar, and Bangladesh), Europe (the UK, Portugal, Spain, Denmark, Poland, Sweden, Greece), North America (the USA, California), South America (Colombia), the Middle East (United Arab Emirates), and Africa (Nigeria). This uneven distribution suggests that research on Generative AI in education is heavily concentrated in Asia and Europe, while contributions from Africa and South America remain limited.

4. Literature Review

In the context of analyzing the application of Gen AI in the educational sector, there is a growing academic interest in the transformative potential of these technologies. Numerous studies have examined how Gen AI can enhance the educational experience through the development of adaptive and personalized learning tools.
To organize our meta-survey, we grouped the reviewed studies into three categories: general educational applications of Gen AI, research focused specifically on ChatGPT, and studies examining the use of commercial Gen AI tools in educational settings.

4.1. Educational Applications of Gen AI

This category encompasses papers that examine the general applications, educational outcomes, ethical considerations, and integration strategies of Generative AI across various educational levels and domains. These works provide foundational insights into how Gen AI is reshaping modern education.
The current literature highlights Gen AI transformative potential for education landscape, particularly in higher education, where it supports advancements in personalized learning, adaptive feedback, automated assessment, and the development of virtual instructors and intelligent tutoring systems [23,35,40,41]. Several implementation frameworks such as PAIGE, IDEE, and AI-CRITIQUE have been introduced to ensure ethical and effective deployment [23,24]. A recurring theme in this body of work is the ethical dimension of Gen AI use, including concerns over academic integrity, data privacy, intellectual property, and algorithmic bias. These challenges have prompted calls for institutional policies and clear ethical guidelines [24,30,37].
While positive outcomes are widely reported, including improvements in cognitive, technical, and interpersonal skills [41], scholars also caution against over-reliance on AI technologies. They advocate for educator engagement, critical AI literacy, and balanced integration strategies to avoid undermining independent thinking and pedagogical goals [26,33,36]. Notably, a gap persists in research focusing on K–12 education and teacher perspectives, suggesting a need for broader and more inclusive investigation [25,39].
Finally, bibliometric analyses confirm a sharp increase in Gen AI-related educational research since 2022, driven primarily by the release of ChatGPT [27]. This growth reflects a shift toward interdisciplinary, policy-oriented approaches aimed at ensuring responsible and impactful integration of Gen AI in teaching, learning, and assessment [28,33,40].

4.2. ChatGPT in Educational Practice

This category includes studies that specifically investigate the use of ChatGPT in educational settings, focusing on its pedagogical implications, perceived benefits, and emerging challenges across different educational levels.
The growing body of literature reflects a strong interest in leveraging ChatGPT for educational purposes, particularly in higher education. Studies demonstrate its effectiveness in supporting assignment design, research assistance, and personalized learning pathways [32,38]. ChatGPT’s natural language processing capabilities enhance student engagement, accessibility, and instant feedback, contributing to more adaptive and inclusive learning environments [31,38]. However, significant concerns persist regarding the quality, bias, and factual reliability of its outputs, as well as the risks of academic dishonesty and over-reliance on AI-generated content. These challenges underscore the need for ethical strategies and educator oversight to safeguard academic integrity and ensure meaningful learning [31,34,38].
Research in K–12 contexts, though limited, points to ChatGPT’s potential to improve language accessibility, student motivation, and personalized instruction [29]. successful implementation requires coordinated efforts among educators, policymakers, and stakeholders to address risks such as data privacy, unequal access, and teacher preparedness. Notably, ChatGPT is not merely reshaping learning delivery but is also redefining the educator’s role, shifting from knowledge transmission to facilitation and personalized guidance [31]. Overall, the literature calls for a balanced integration of ChatGPT that aligns with pedagogical goals, supported by ethical frameworks and institutional policies to maximize its benefits and mitigate unintended consequences [34].

4.3. Commercial Gen AI Tools in Education

This category includes papers that review or analyze branded, real-world Gen AI applications (e.g., Bard (Google, 2023), Gemini (Google DeepMind, released 2023), RoomGPT (open-source project, 2023)) and their use in authentic educational settings, especially technical and applied fields.
The paper [42] provides a comprehensive review of state-of-the-art Gen AI models, including GANs, GPT, autoencoders, diffusion models, and transformers, with a focus on their architectures, tasks, and real-world applications. It highlights the rapid rise of commercial tools like ChatGPT, Bard, Stable Diffusion, and Adobe Firefly, highlighting their transformative impact across sectors including education, healthcare, business, and entertainment. In education, the study showcases how Gen AI tools are being used for personalized content creation, tutoring, code generation, and scientific visualization, while also acknowledging challenges such as misinformation, hallucinated outputs, and privacy concerns.
The authors of [43] explore the use of Gen AI in ICT engineering education, with an emphasis on programming and ethical instruction. The study highlights several commercial Gen AI tools actively used in educational settings, including ChatGPT (GPT-3.5, GPT-4), GitHub Copilot, Amazon Code Whisperer, Replit, Android Studio Bot, and OpenAI Codex, among others. These tools support various educational applications, including automatic feedback on code, bug detection, multilingual coding, and intelligent tutoring. The authors include real-world case studies from universities worldwide, showcasing how these tools enhance student learning, programming skills, and engagement, while also acknowledging ethical and practical limitations such as over-reliance, prompt engineering challenges, and bias in large language models.
  • The Dynamic Landscape of Commercial Gen AI Tools:
    While early research (2022–2024) catalogued specific commercial tools such as ChatGPT, Bard, GitHub Copilot, and Adobe Firefly, such inventories risk rapid obsolescence. The commercial Gen AI ecosystem evolves rapidly, with foundation-model vendors frequently releasing new features, subsuming third-party tools, or launching academic-facing access programs. For example, in mid-2025, Perplexity AI partnered with SheerID to provide free access to Perplexity Pro for verified students worldwide [47], fundamentally shifting adoption dynamics and institutional decision-making. Similarly, OpenAI released ChatGPT Edu across the California State University system, offering access to more than 500,000 students and faculty [48].
    Beyond access initiatives, new agentic systems demonstrate capabilities that extend beyond traditional tutoring. For instance, Manus, an AI agent highlighted by researchers, is capable of generating full online course modules in minutes, exemplifying the rise of agentic task execution in education [49]. In parallel, Google’s NotebookLM emerged in 2025 as a study-assistant platform capable of summarizing course readings, integrating with academic workflows, and providing contextualized responses [50,51]. These innovations illustrate how commercial tools are increasingly embedding themselves into institutional platforms and student study habits.
    In light of this volatility, a more durable analytic lens is to abstract from specific brands to capability-based categories. Commercial tools in education can be framed into the following clusters: (i) retrieval-augmented tutoring and document-grounded assistance, (ii) agentic workflow generators, (iii) multimodal and creative generation, and (iv) LMS-integrated study platforms. Mapping risks and governance controls to these categories (rather than to specific brands) provides a framework that remains relevant even as individual tools evolve.
    Recent frameworks, such as the CRAFT model proposed by APRU [52], and adoption studies by Ithaka [53], reinforce this approach by linking capabilities with ethical, cultural, and policy dimensions. Large-scale surveys confirm these shifts: in the UK, the 2025 HEPI survey found that 92% of students now use Gen AI (up from 66% in 2024) and 88% report using it for assessments [54]; globally, more than 86% of students use AI in their studies, with 25% engaging daily [55]. Such patterns indicate that categories of capability and adoption, rather than static tool lists, offer more durable value to both researchers and practitioners.
  • Emerging Categories: Research-Centric, Document-Grounded, and Agentic Systems Beyond early single-chat systems such as ChatGPT, the educational use of Gen AI since late 2024 has expanded into qualitatively distinct categories that require differentiated analysis and oversight. These categories include the following:
    Retrieval-grounded research assistants: Tools such as Perplexity integrate live web search with citation trails, enabling students to generate responses anchored in real-time sources. The educational benefits include improved citation quality and source evaluation practices, while risks involve over-reliance on automatically retrieved content or uncritical trust in surfaced sources. Appropriate controls may include explicit citation rubrics, critical source evaluation exercises, and verification of reference accuracy [47].
    Document-grounded tutors: Platforms like Google’s NotebookLM exemplify study tools that operate over instructor-provided materials (e.g., syllabi, PDFs, or course readings) [50]. Their alignment to curricular objectives allows for personalized scaffolding of reading comprehension and formative assessment. However, risks include bypassing independent reading and undermining student accountability. Case studies in higher education have shown that NotebookLM can support reading compliance and study accountability, but also require careful instructor oversight [51]. Controls therefore include embedding NotebookLM sessions into structured coursework, monitoring reading compliance, and designing assignments that require synthesis beyond the provided documents.
    Agentic systems: Emerging systems such as Manus can plan and execute multi-step educational workflows, including automated generation of complete course modules. These affordances introduce both opportunities (e.g., rapid prototyping of instructional content, automation of routine tasks) and significant risks (e.g., outsourcing entire coursework, opaque decision-making processes). Mitigation measures include process logging, version control trails, oral defenses, and data diaries to verify authorship and learning integrity [49].
    These categories illustrate a transition in educational use of Gen AI: from brand-based discussions to capability-based analysis. Framing tools in terms of what they enable, and how associated risks can be managed, provides a foundation for more resilient adoption strategies. For example, the APRU whitepaper emphasizes the importance of linking AI use with institutional culture, rules, and access policies [52], while the Ithaka report highlights the need for frameworks that remain valid as tools evolve [53]. At the same time, practitioner reports show measurable benefits, such as teacher time savings and reduced workload stress, underscoring why such structured approaches are urgently needed [56].
Table 4 summarizes the main contributions of a selection of academic research in this field, with particular attention to the innovative aspects of each study to help address the research questions of this paper. Analyzed academic works form the dataset for RQs 1–5. The table highlights the innovative contributions of each study, which are systematically synthesized in Section 4.
This section presents the results organized according to the research questions (RQ1–RQ5). Each subsection highlights how the analyzed studies contribute to answering these questions.

4.4. RQ1: What Are the Current Definitions of Generative AI?

With the rapid advancement of artificial intelligence technologies particularly following the release of ChatGPT, Gen AI has become a key focus of research in the educational domain. However, definitions of Gen AI vary considerably across the literature, reflecting differences in technical perspectives, application contexts, and disciplinary orientations.
The reviewed literature offers a variety of definitions for Gen AI, reflecting both technical perspectives and educational applications. These definitions typically fall into two thematic categories: (1) Gen AI as autonomous content creation technology, and (2) Gen AI as a form of human-like creativity.
  • Generative AI as Content Creation Technology
    Gen AI is commonly described as a class of AI systems capable of producing original content by learning from large datasets. These definitions highlight the underlying architectures such as LLMs (e.g., GPT), GANs, VAEs, and NLP models. For instance, Santos and his colleagues [43] and Bengesi [42] refer to Gen AI tools as content-producing models that enhance tasks like feedback, code generation, and personalized learning. Similarly, McGrath [33] and Daniel [41] emphasize the autonomous nature of these tools, while Xia [40] describes the ability of Gen AI to generate synthetic data across multiple modalities. In addition, Ahmed [23] and Mittal [26] define Generative AI as models capable of producing novel content or generating new material from existing sources, while Ogunleye [27] highlights its use in LLM-powered systems for content generation, tutoring, and feedback.
  • Generative AI as Human-like Creativity
    The studies discuss Gen AI in terms of its ability to produce human-like content across various media formats. For example, Yusuf [25] describes Generative AI as AI models that can create human-like content across multiple modalities. These discussions often emphasize the use of advanced AI techniques to generate content that mimics human creativity and cognition. Several studies focus on the application of Gen AI in educational contexts, particularly in K–12 education, highlighting its potential to generate content that enhances learning experiences [29]. Other authors explore how Gen AI models analyze existing digital media to simulate human-level creativity, enabling the generation of text, images, and other forms of content [35]. Additionally, there is a significant focus on Gen AI capabilities in language interpretation, summarization, and prediction, which has led to considerable interest and discussion regarding its implications for educational practices [36].
    Our synthesis reveals inconsistency in definitions across reviews, ranging from purely technical descriptions to broader pedagogical framings (see Table 5). This conceptual fragmentation complicates comparative analysis and underscores the need for a more standardized definition of Gen AI in education.

4.5. RQ2: What Are the Current Methodologies and Frameworks for Integrating Gen AI into Educational Systems?

Although the rapid rise of Gen AI has prompted widespread experimentation in educational settings, the academic literature still offers a limited number of structured methodologies or frameworks for its integration. Among the studies reviewed, only a few propose concrete models or evaluative tools to guide the implementation of Gen AI in educational environments.
Ahmed and his colleagues [23] provide one of the most comprehensive contributions by reviewing four notable frameworks: PAIGE, AI-CRITIQUE, DATS, and IDEE. These frameworks collectively address personalization, ethical considerations, assessment, feedback loops, and instructional design. They aim to guide the responsible deployment of Gen AI tools such as EduChat [57] and CodeHelp [58] within educational curricula.
Ref. [24] proposes a conceptual framework grounded in AI literacy, human oversight, adaptive regulation, and responsible governance. Their model stresses ethical integration of large language models in higher education, calling attention to transparency, inclusivity, and safeguards against plagiarism and bias.
Ref. [41] introduces the UTAUT-EG model, an adaptation of the Unified Theory of Acceptance and Use of Technology (UTAUT) with an ethical governance focus. This framework connects the use of Gen AI tools to the development of cognitive, technical, and interpersonal academic skills, while promoting principles of transparency, accountability, and responsible adoption.
Ref. [37] applies the Analytical Hierarchy Process (AHP) to prioritize ethical concerns surrounding ChatGPT use in academia. The authors proposed a decision-making framework assisting educational institutions in choosing governance strategies, revealing a preference for restriction over legislation as a more adaptable policy approach.
Existing approaches primarily focus on integrating Gen AI into specific educational contexts, addressing particular challenges, and optimizing learning in certain fields or courses. The absence of a standardized framework that leverages the full potential of Gen AI across various educational settings represents a significant challenge.
Existing approaches primarily focus on integrating Gen AI into specific educational contexts, addressing particular challenges, and optimizing learning in certain fields or courses (see Table 6). The absence of a standardized framework that leverages the full potential of Gen AI across various educational settings represents a significant challenge.
Notably, only 7 of the 21 reviews employed a formal synthesis method, and fewer than half reported explicit eligibility criteria. This uneven methodological rigor highlights a lack of standardization, limiting comparability across studies.

4.6. RQ3: Which Educational Levels Have Been the Primary Focus of Research on Gen AI?

Referring to RQ3, our primary focus was on the academic context, and the main educational level targeted in Gen AI research is higher education, while K–12 education remains significantly underrepresented. Most of the selected papers focus on how Gen AI is being integrated, discussed, or evaluated within universities and other tertiary institutions.
Ogunleye and his colleagues [27] and Jensen [30], highlight the explosive growth of Gen AI-related publications since 2022, particularly within the context of higher education. These works emphasize themes such as academic integrity, student engagement, automated assessment, and intelligent tutoring systems, all situated in university environments.
Other authors [32,38] explore Gen AI applications, especially ChatGPT, for research support, assignment generation, and student-centered learning, reinforcing the dominance of higher education as the main setting. McGrath [33] confirms that the majority of the academic discourse surrounding Gen AI appears in the higher education space, including fields beyond traditional computer science, indicating a strong institutional interest in adopting AI for teaching and learning.
On the other hand, K–12 education receives much less attention. Zhang [29] highlights the potential of tools like ChatGPT for personalized instruction, language development, and learner motivation, but also points out that policy constraints, teacher training gaps, and ethical concerns have delayed the broader adoption of Gen AI in schools.
Bengesi [42] and Santos [43] focus on the use of commercial Gen AI tools such as ChatGPT, Bard, Adobe Firefly, GitHub Copilot, Replit, and Amazon CodeWhisperer. While their studies span multiple sectors, they primarily highlight educational use cases in higher education, especially in ICT, computer science, and engineering. For instance, Santos and his colleagues [43] present real-world university-level examples where these tools are used for code generation, debugging, intelligent tutoring, and ethical instruction. Similarly, Bengesi [42] emphasizes how commercial tools support content creation, tutoring, and scientific visualization in tertiary settings.
While higher education dominates the current literature, K–12 remains systematically underrepresented, with only one review addressing this level in depth. This gap points to an urgent need for comparative research across educational stages.

4.7. RQ4: What Are the Main Challenges of Integrating Gen AI into Education?

Referring to RQ4, the integration of Gen AI into the education sector introduces a complex array of challenges. A consistent concern is the reliability and accuracy of Gen AI outputs, with several studies [31,38,42] reporting that Gen AI tools such as ChatGPT and Bard are prone to hallucinated, biased, or misleading responses. These issues raise significant concerns about trust, particularly in academic settings where factual correctness and critical thinking are essential.
Educators also face uncertainty in adapting pedagogy to AI-driven learning environments. As noted by Jensen [30] and McGrath [33], Gen AI alters traditional educational processes such as student engagement, content creation, and assessment practices. This is further compounded by a lack of institutional readiness and structured support, which may lead to over-reliance on AI tools and a decline in deeper learning and creativity, as emphasized in [26,32].
Ethical risks represent another major challenge, especially in relation to plagiarism, academic integrity, data privacy, and AI bias. Qadhi [24] and Bukar [37] stress the importance of ethical governance and human oversight, proposing frameworks to ensure responsible implementation and address regulatory gaps.
Moreover, several studies [35,41] raise broader concerns regarding equitable access and AI literacy. These studies emphasize that without inclusive training, accessible design, and strong policy frameworks, Gen AI risks exacerbating existing educational inequalities.
In addition to general concerns of bias and misinformation, the integration of Gen AI into education creates potential conflicts between innovation and academic integrity. While these tools can stimulate creativity, provide personalized learning pathways, and reduce administrative burdens, their unregulated use may undermine the authenticity of student work, blur boundaries of plagiarism, and complicate assessment practices. Moreover, disparities in AI literacy and access to digital infrastructure, particularly in low-resource educational systems, risk widening existing inequalities, as educators and learners without adequate training may misuse or underutilize Gen AI. These gaps highlight the urgent need for AI literacy programs, clear institutional policies, and governance frameworks that balance pedagogical innovation with safeguards for integrity, equity, and accountability. Addressing these dimensions is essential for ensuring that Gen AI adoption enhances rather than undermines educational quality and inclusiveness (see Table 7).
Although ethical risks such as plagiarism and bias are widely acknowledged, very few reviews connect these issues to governance or institutional frameworks. This indicates a gap between identifying challenges and proposing actionable solutions.

4.8. RQ5: How Effective Are Gen AI Tools in Fostering Critical Thinking and Problem-Solving Skills Among Students?

Referring to RQ5, Gen AI has demonstrated transformative potential in various educational contexts. Specifically, Gen AI tools are increasingly recognized for their capacity to enhance students’ critical thinking and problem-solving abilities. Several papers discuss the potential of Gen AI in supporting higher-order thinking through personalized feedback, adaptive learning pathways, and intelligent tutoring systems.
Daniel [41] reports significant improvements in cognitive skills, including critical thinking, metacognition, and problem-solving, especially when Gen AI tools are implemented with ethical guidance and teacher participation. Similarly, studies by Ali and Xia [38,40] highlight the ability of Gen AI tools to support self-regulated learning and reflective thinking by providing instant feedback and offering alternative perspectives during learning tasks.
However, multiple studies also raise concerns about over-reliance on Gen AI, which may weaken students’ independent reasoning. Baig and Mittal [26,32] argue that students may begin to delegate cognitive tasks to AI systems, limiting their engagement in deep problem-solving processes unless educators purposefully design learning activities that promote active learning and critical analysis.
Jensen [30] adds that many current implementations of Gen AI tend to reinforce existing educational models rather than fundamentally transform how students approach inquiry, creativity, and autonomous learning. Furthermore, despite the optimism surrounding the potential of Gen AI, there is currently limited empirical evidence in most of the reviewed literature to conclusively assess its impact on critical thinking, particularly in comparison to traditional pedagogies.
While our analysis focuses on outcomes reported in academic studies (see Table 8), it is important to note that a wide range of commercial Gen AI tools have recently been introduced into educational environments. However, their effectiveness in fostering critical thinking and problem-solving remains under-examined. More empirical research is needed to evaluate the long-term impact of these tools on learning outcomes across diverse educational contexts.
Evidence of Gen AI’s effectiveness in fostering critical thinking and problem-solving remains limited and often anecdotal, underscoring the absence of large-scale empirical validation.
It is important to acknowledge the limitations of the current literature. First, most studies are conceptual or review-based, with a lack of empirical evidence evaluating Gen AI’s effectiveness in authentic classroom environments. Second, there is a clear imbalance in focus: the majority of research centers on higher education, while K–12 education remains underrepresented despite its critical role in shaping early AI literacy. Third, nearly all included studies are published in English, leading to the near absence of perspectives from non-English-speaking regions, which limits the global generalizability of findings. Addressing these limitations requires a more inclusive research agenda, incorporating diverse educational levels, cultural contexts, and methodological rigor.

5. Future Trends

Generative AI is not only transforming the educational landscape but also catalyzing the development of new learning models that extend beyond traditional classroom structures. To better understand these emerging directions, we asked the following question: What recent trends suggest how Gen AI will drive the development of new learning models in education?
Recent research suggests that the field is moving beyond single-chat interfaces and toward capability-based systems. A growing trend is the rise of agentic AI systems, capable of multi-step planning and task execution. Early platforms such as Manus illustrate how these systems can generate entire projects or learning modules with minimal input. While these capabilities hold promise for productivity and innovation, they also raise concerns about coursework automation, requiring new assessment strategies such as process logging, version trails, or oral defenses [49].
Another major development is the use of document-grounded and retrieval-augmented AI tutors. Tools such as NotebookLM allow students to query and generate study guides directly from course readings, while Perplexity integrates live web search with citation trails [50]. These tools shift pedagogy toward reading accountability and source evaluation, but they also require redesigned assignments and policies to ensure independent learning and academic integrity [51].
Institutional adoption is also accelerating. In 2025, several universities and ed-tech providers launched large-scale student access programs, such as Perplexity’s free academic licenses and OpenAI’s partnership with California State University [47,48]. This signals a transition from isolated experimentation to structured, institution-wide integration of Gen AI in curricula and assessment.
Future models are also expected to emphasizehuman–AI co-creation, where learners collaborate with AI agents to generate content, solve problems, and engage in entrepreneurial projects [53]. This development requires redefining the role of educators as AI facilitators and literacy coaches, emphasizing skills such as prompt design, critical evaluation, and ethical navigation [52].
Finally, governance frameworks are emerging as a critical trend. Moving beyond general concerns of bias or privacy, recent reports call for capability–risk–control approaches, for example, mapping the risks of retrieval grounding, document grounding, and agentic automation to appropriate institutional safeguards. This requires not only technical solutions (e.g., detection systems for academic dishonesty [59]) but also institutional readiness, inclusive policymaking, and frameworks that balance innovation with educational integrity.
Collectively, these trends suggest that by 2025–2026, Gen AI will be understood not as a single application but as an evolving ecosystem of capabilities. This transition carries profound implications for pedagogy and governance, requiring the ongoing adaptation of educational models, assessment practices, and institutional frameworks.

6. Conclusions

Generative AI has significant potential to transform education by enabling personalized learning experiences and producing interactive content tailored to individual learners. This meta-survey provides a comprehensive synthesis of current research on the integration of Gen AI in education, drawing on systematic literature reviews, scoping studies, and thematic analyses across diverse contexts. Using the PRISMA methodology, 21 surveys were examined and their contributions analyzed. The findings indicate that, although research on Gen AI has proliferated rapidly particularly since the emergence of ChatGPT, important gaps remain.
The analysis also highlights the emerging role of commercial Gen AI tools, which are increasingly being incorporated into teaching and learning practices across multiple disciplines. Despite their expanding presence, relatively few studies have explored these tools in depth, with more focused discussions only beginning to appear in 2024. Much of the existing literature continues to emphasize broader AI themes or isolated applications. Nevertheless, commercial tools are expected to become more widely adopted and more systematically studied in the near future, reflecting their growing educational impact.
Overall, the evidence suggests that Gen AI represents more than an enhancement to traditional instruction; it is a transformative force with the capacity to reshape educational models. By identifying current trends, gaps, and opportunities, this meta-survey establishes a foundation for further research and offers insights that may inform the integration of Gen AI into future educational practices.

Author Contributions

Conceptualization, S.B. and G.F.; formal analysis, S.B. and G.F. and F.P., writing—original draft preparation, S.B.; writing—review and editing, S.B. and G.F. and F.P. and M.C.; supervision, G.F. and F.P.; funding acquisition, G.F. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Olga, A.; Saini, A.; Zapata, G.; Searsmith, D.; Cope, B.; Kalantzis, M.; Kastania, N.P. Generative AI: Implications and applications for education. arXiv 2023, arXiv:2305.07605. [Google Scholar] [CrossRef]
  2. Fortino, G.; Mangione, F.; Pupo, F. Intersezione tra intelligenza artificiale generativa e educazione: Un’ipotesi. J. Educ. Cult. Psychol. Stud. ECPS J. 2025, 30, 25–52. [Google Scholar]
  3. Kalota, F.A. A Primer on Generative Artificial Intelligence. Educ. Sci. 2024, 14, 172. [Google Scholar] [CrossRef]
  4. Suleri, J.; Suleri, A. Comparing virtual learning, classical classroom learning and blended learning. Eur. J. Sustain. Dev. Res. 2019, 3, 1–8. [Google Scholar] [CrossRef] [PubMed][Green Version]
  5. Harjana, M.N.S.; Saputra, H.Y.; Tho, C.A. A Review of the Potential Use of Mixed Reality Learning Methods in Comparison to Traditional Learning Methods. Procedia Comput. Sci. 2023, 227, 734–742. [Google Scholar] [CrossRef]
  6. Cardona-Acevedo, S.; Agudelo-Ceballos, E.; Cumpa Vásquez, J.T.; Martínez Rojas, E.; Valencia-Arias, A.; Jimenez Garcia, J.A.; Benjumea-Arias, M.L. E-learning technologies at the secondary education level: Literature review. Front. Educ. 2025, 10, 1539763. [Google Scholar] [CrossRef]
  7. Rose, S. Medical student education in the time of COVID-19. JAMA 2020, 323, 2131–2132. [Google Scholar] [CrossRef] [PubMed]
  8. Ng, D.T.K.; Chan, E.K.C.; Lo, C.K. Opportunities, challenges and school strategies for integrating generative AI in education. Comput. Educ. Artif. Intell. 2025, 8, 100373. [Google Scholar] [CrossRef]
  9. Bashir, S.; Lapshun, A.L. E-learning future trends in higher education in the 2020s and beyond. Cogent Educ. 2025, 12, 2445331. [Google Scholar] [CrossRef]
  10. Hafeez, M.; Akhter, Y. Effects of blended learning in comparison of traditional learning to provide safer learning environment—A comparative review. Int. J. Educ. Res. Soc. Sci. 2021, 2, 1604–1615. [Google Scholar] [CrossRef]
  11. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Moher, D. Updating guidance for reporting systematic reviews: Development of the PRISMA 2020 statement. J. Clin. Epidemiol. 2021, 134, 103–112. [Google Scholar] [CrossRef]
  12. Heinke, A.; Radgoudarzi, N.; Huang, B.B.; Baxter, S.L. A review of ophthalmology education in the era of generative artificial intelligence. Asia-Pac. J. Ophthalmol. 2024, 13, 100089. [Google Scholar]
  13. Shoja, M.M.; Van de Ridder, J.M.; Rajput, V. The emerging role of generative artificial intelligence in medical education, research, and practice. Cureus 2023, 15, e40883. [Google Scholar] [CrossRef]
  14. Tangadulrat, P.; Sono, S.; Tangtrakulwanich, B. Using ChatGPT for clinical practice and medical education: Cross-sectional survey of medical students’ and physicians’ perceptions. JMIR Med. Educ. 2023, 9, e50658. [Google Scholar] [CrossRef]
  15. Chen, B.; Wu, Z.; Zhao, R. From fiction to fact: The growing role of Gen AI in business and finance. J. Chin. Econ. Bus. Stud. 2023, 21, 471–496. [Google Scholar] [CrossRef]
  16. Desai, A.P.; Mallya, G.S.; Luqman, M.; Ravi, T.; Kota, N.; Yadav, P. Opportunities and Challenges of Generative-AI in Finance. arXiv 2024, arXiv:2410.15653. [Google Scholar] [CrossRef]
  17. Gunawan, J.; Aungsuroch, Y.; Montayre, J. ChatGPT integration within nursing education and its implications for nursing students: A systematic review and text network analysis. Nurse Educ. Today 2024, 141, 106323. [Google Scholar] [CrossRef]
  18. Athilingam, P.; He, H.G. ChatGPT in nursing education: Opportunities and challenges. Teach. Learn. Nurs. 2024, 19, 97–101. [Google Scholar]
  19. García-Peñalvo, F.J.; Llorens-Largo, F.; Vidal, J. La nueva realidad de la educación ante los avances de la inteligencia artificial generativa. RIED-Rev. Iberoam. Educ. Distancia 2024, 27, 9–39. [Google Scholar] [CrossRef]
  20. Muster, B. L’intégration de l’IA dans l’enseignement au secondaire II. Master’s Thesis, Haute École Pédagogique BEJUNE, Bienne, Switzerland, 2024. [Google Scholar]
  21. Batsaikhan, B.Z.; Correia, A.P. The effects of Generative Artificial Intelligence on Intelligent Tutoring Systems in higher education: A systematic review. Stud. Technol. Enhanc. Learn. 2024, 4, 1–19. [Google Scholar] [CrossRef]
  22. Lee, A.V.Y. Staying ahead with generative artificial intelligence for learning: Navigating challenges and opportunities with 5Ts and 3Rs. Asia Pac. J. Educ. 2024, 44, 81–93. [Google Scholar] [CrossRef]
  23. Ahmed, Z.; Shanto, S.S.; Rime, M.H.K.; Morol, M.K.; Fahad, N.; Hossen, M.J.; Abdullah-Al-Jubair, M.A. The Gen AI Landscape in Education: Mapping the Terrain of Opportunities, Challenges and Student Perception. IEEE Access 2024, 12, 147023–147050. [Google Scholar] [CrossRef]
  24. Qadhi, S.M.; Alduais, A.; Chaaban, Y.; Khraisheh, M. Gen AI, Research Ethics, and Higher Education Research: Insights from a Scientometric Analysis. Information 2024, 15, 325. [Google Scholar]
  25. Yusuf, A.; Pervin, N.; Román-González, M.; Noor, N.M. Gen AI in education and research: A systematic mapping review. Rev. Educ. 2024, 12, e3489. [Google Scholar]
  26. Mittal, U.; Sai, S.; Chamola, V. A comprehensive review on Gen AI for education. IEEE Access 2024, 12, 142733–142759. [Google Scholar] [CrossRef]
  27. Ogunleye, B.; Zakariyyah, K.I.; Ajao, O.; Olayinka, O.; Sharma, H. A Systematic Review of Gen AI for Teaching and Learning Practice. Educ. Sci. 2024, 14, 636. [Google Scholar] [CrossRef]
  28. Batista, J.; Mesquita, A.; Carnaz, G. Gen AI and higher education: Trends, challenges, and future directions from a systematic literature review. Information 2024, 15, 676. [Google Scholar] [CrossRef]
  29. Zhang, P.; Tur, G. A systematic review of ChatGPT use in K-12 education. Eur. J. Educ. 2024, 59, e12599. [Google Scholar]
  30. Jensen, L.X.; Buhl, A.; Sharma, A.; Bearman, M. Gen AI and higher education: A review of claims from the first months of ChatGPT. High. Educ. 2024, 89, 1145–1161. [Google Scholar] [CrossRef]
  31. Bettayeb, A.M.; Abu Talib, M.; Sobhe Altayasinah, A.Z.; Dakalbab, F. Exploring the impact of ChatGPT: Conversational AI in education. In Frontiers in Education; Frontiers Media SA: Lausanne, Switzerland, 2024; Volume 9, p. 1379796. [Google Scholar]
  32. Baig, M.I.; Yadegaridehkordi, E. ChatGPT in the higher education: A systematic literature review and research challenges. Int. J. Educ. Res. 2024, 127, 102411. [Google Scholar] [CrossRef]
  33. McGrath, C.; Farazouli, A.; Cerratto-Pargman, T. Gen AI chatbots in higher education: A review of an emerging research area. High. Educ. 2024, 89, 1533–1549. [Google Scholar] [CrossRef]
  34. İpek, Z.H.; Gözüm, A.I.C.; Papadakis, S.; Kallogiannakis, M. Educational Applications of the ChatGPT AI System: A Systematic Review Research. Educ. Process Int. J. 2023, 12, 26–55. [Google Scholar]
  35. Bahroun, Z.; Anane, C.; Ahmed, V.; Zacca, A. Transforming education: A comprehensive review of generative artificial intelligence in educational settings through bibliometric and content analysis. Sustainability 2023, 15, 12983. [Google Scholar] [CrossRef]
  36. Zhang, X.; Zhang, P.; Shen, Y.; Liu, M.; Wang, Q.; Gašević, D.; Fan, Y. A Systematic Literature Review of Empirical Research on Applying Generative Artificial Intelligence in Education. Front. Digit. Educ. 2024, 1, 223–245. [Google Scholar] [CrossRef]
  37. Bukar, U.A.; Sayeed, M.S.; Razak, S.F.A.; Yogarayan, S.; Sneesl, R. Decision-making framework for the utilization of generative artificial intelligence in education: A case study of ChatGPT. IEEE Access 2024, 12, 95368–95389. [Google Scholar] [CrossRef]
  38. Ali, D.; Fatemi, Y.; Boskabadi, E.; Nikfar, M.; Ugwuoke, J.; Ali, H. ChatGPT in Teaching and Learning: A Systematic Review. Educ. Sci. 2024, 14, 643. [Google Scholar] [CrossRef]
  39. Ariza, J.; Restrepo, M.B.; Hernández, C.H. Generative AI in Engineering and Computing Education: A Scoping Review of Empirical Studies and Educational Practices. IEEE Access 2025, 13, 30789–30810. [Google Scholar] [CrossRef]
  40. Xia, Q.; Weng, X.; Ouyang, F.; Lin, T.J.; Chiu, T.K. A scoping review on how generative artificial intelligence transforms assessment in higher education. Int. J. Educ. Technol. High. Educ. 2024, 21, 40. [Google Scholar] [CrossRef]
  41. Daniel, K.; Msambwa, M.M.; Wen, Z. Can Generative AI Revolutionise Academic Skills Development in Higher Education? A Systematic Literature Review. Eur. J. Educ. 2025, 60, e70036. [Google Scholar] [CrossRef]
  42. Bengesi, S.; El-Sayed, H.; Sarker, M.K.; Houkpati, Y.; Irungu, J.; Oladunni, T. Advancements in Generative AI: A Comprehensive Review of GANs, GPT, Autoencoders, Diffusion Model, and Transformers. IEEE Access 2024, 12, 69812–69837. [Google Scholar] [CrossRef]
  43. Santos, P.; Urgel, K.; Moreno, V. Generative Artificial Intelligence in teaching and learning of ICT engineering education: A literature review and illustrative scenarios. In Proceedings of the 2024 47th MIPRO ICT and Electronics Convention (MIPRO), Opatija, Croatia, 20–24 May 2024; pp. 1338–1343. [Google Scholar]
  44. Whiting, P.; Savović, J.; Higgins, J.P.T.; Caldwell, D.M.; Reeves, B.C.; Shea, B.; Davies, P.; Kleijnen, J.; Churchill, R. ROBIS: A new tool to assess risk of bias in systematic reviews was developed. J. Clin. Epidemiol. 2016, 69, 225–234. [Google Scholar] [CrossRef] [PubMed]
  45. Dalky, A.; Altawalbih, M.; Alshanik, F.; Khasawneh, R.A.; Tawalbeh, R.; Al-Dekah, A.M.; Alrawashdeh, A.; Quran, T.O.; ALBashtawy, M. Global research trends, hotspots, impacts, and emergence of artificial intelligence and machine learning in health and medicine: A 25-year bibliometric analysis. Healthcare 2025, 13, 892. [Google Scholar] [CrossRef]
  46. Lin, M.; Lin, L.; Lin, L.; Lin, Z.; Yan, X. A bibliometric analysis of the advance of artificial intelligence in medicine. Front. Med. 2025, 12, 1504428. [Google Scholar] [CrossRef]
  47. Nuñez, M. Perplexity Offers Free AI Tools to Students Worldwide in Partnership with SheerID. 2025. Available online: https://venturebeat.com/ai/perplexity-offers-free-ai-tools-to-students-worldwide-in-partnership-with-sheerid/ (accessed on 23 August 2025).
  48. Reuters. OpenAI Targets Higher Education in US with ChatGPT Rollout at California State University. 2025. Available online: https://www.reuters.com/technology/openai-targets-higher-education-us-with-chatgpt-rollout-california-state-2025-02-04/ (accessed on 23 August 2025).
  49. Future of Being Human Institute. AI Agent Creates Online Course in Minutes. 2025. Available online: https://www.futureofbeinghuman.com/p/ai-agent-creates-online-course-in-minutes (accessed on 23 August 2025).
  50. Geeky Gadgets. How to Use NotebookLM in 2025. 2025. Available online: https://www.geeky-gadgets.com/how-to-use-notebooklm-in-2025/ (accessed on 23 August 2025).
  51. Florida Gulf Coast University. Exploring NotebookLM for Digital Learning. 2025. Available online: https://www.fgcu.edu/digitallearning/digital-learning-blog/02-24-2025-notebooklm (accessed on 23 August 2025).
  52. APRU. Generative AI in Higher Education: Current Practices and Ways Forward. Whitepaper, Association of Pacific Rim Universities. 2025. Available online: https://www.apru.org/wp-content/uploads/2025/01/APRU-Generative-AI-in-Higher-Education-Whitepaper_Jan-2025.pdf (accessed on 23 August 2025).
  53. Baytas, C.; Ruediger, D. Making AI Generative for Higher Education: Adoption and Challenges among Instructors and Researchers. ITHAKA S+R 2025, 55. [Google Scholar]
  54. Freeman, J. Student Generative AI Survey 2025; Higher Education Policy Institute: London, UK, 2025. [Google Scholar]
  55. Digital Education Council. AI in Education Statistics 2025. 2025. Available online: https://anara.com/blog/ai-in-education-statistics (accessed on 23 August 2025).
  56. Ferlazzo, L. AI Can Save Teachers Time and Stress. Here’s How. 2025. Available online: https://www.edweek.org/technology/opinion-ai-can-save-teachers-time-and-stress-heres-how/2025/06 (accessed on 23 August 2025).
  57. Dan, Y.; Lei, Z.; Gu, Y.; Li, Y.; Yin, J.; Lin, J.; Qiu, X. Educhat: A large-scale language model-based chatbot system for intelligent education. arXiv 2023, arXiv:2308.02773. [Google Scholar]
  58. Liffiton, M.; Sheese, B.E.; Savelka, J.; Denny, P. Codehelp: Using large language models with guardrails for scalable support in programming classes. In Proceedings of the 23rd Koli Calling International Conference on Computing Education Research, Koli, Finland, 13–18 November 2023; pp. 1–11. [Google Scholar]
  59. Akpan, I.J.; Kobara, Y.M.; Owolabi, J.; Akpan, A.A.; Offodile, O.F. Conversational and generative artificial intelligence and human–chatbot interaction in education and research. Int. Trans. Oper. Res. 2025, 32, 1251–1281. [Google Scholar] [CrossRef]
Figure 1. Flow-chart of the literature review selection process according to the PRISMA guidelines.
Figure 1. Flow-chart of the literature review selection process according to the PRISMA guidelines.
Bdcc 09 00237 g001
Figure 2. Monthly publications for Generative AI in education with a monthly baseline from Gen AI in Medical Diagnostics (January 2022–April 2025).
Figure 2. Monthly publications for Generative AI in education with a monthly baseline from Gen AI in Medical Diagnostics (January 2022–April 2025).
Bdcc 09 00237 g002
Figure 3. Citation counts for Generative AI in education with a baseline at the monthly level from Gen AI in Medical Diagnostics (January 2022–April 2025).
Figure 3. Citation counts for Generative AI in education with a baseline at the monthly level from Gen AI in Medical Diagnostics (January 2022–April 2025).
Bdcc 09 00237 g003
Table 1. Search plan.
Table 1. Search plan.
SourceCriteria
DatabaseIEEEXplore, MDPI, Springer, Elsevier, Scopus, Google Scholar
Date of Publication2022–2025
KeywordsGenerative Artificial Intelligence/Education
Gen AI tools/ Education
LLMs/Education
LanguageEnglish
Type of PublicationSurvey, Review, or SLR
Inclusion CriteriaPapers published after 2022 (after ChatGPT’s emergence)
Papers must be written in English.
Papers that are peer-reviewed and published in journals or conference proceedings.
Comprehensive studies on the use of Gen AI in educational contexts.
Studies emphasizing Gen AI as a central component of innovative educational solutions.
Exclusion CriteriaPapers before 2022 (before ChatGPT’s emergence)
Papers not written in English.
Papers not peer-reviewed.
Contributions with minimal relevance or limited depth regarding educational applications of Gen AI.
Papers where “Gen AI” is mentioned without proper context or detailed analysis.
Table 2. Quality assessment of included reviews (based on four indicators of methodological rigor).
Table 2. Quality assessment of included reviews (based on four indicators of methodological rigor).
StudyClear RQs/EligibilityTransparent Search StrategyInclusion/ Exclusion ReportingCoherent Synthesis
[23]YesYesYesYes
[24]YesYesNoYes
[25]YesYesYesYes
[26]YesYesNoYes
[27]YesYesYesYes
[28]YesYesYesYes
[29]YesYesYesYes
[30]YesNoNoYes
[31]YesYesYesYes
[32]YesYesNoYes
[33]YesNoNoYes
[34]YesYesYesYes
[35]YesYesYesYes
[36]YesYesNoYes
[37]YesYesYesYes
[38]YesYesYesYes
[39]YesYesNoYes
[40]YesYesYesYes
[41]YesYesYesYes
[42]YesYesNoYes
[43]YesNoNoYes
Table 3. Geographic origin of the 21 included studies (first author affiliation).
Table 3. Geographic origin of the 21 included studies (first author affiliation).
Region/CountryNumber of Studies
China4
United Arab Emirates2
USA (incl. California)2
UK1
Portugal1
Spain1
Denmark1
Poland1
Sweden1
Greece1
India1
Bangladesh1
Qatar1
Nigeria1
Malaysia1
Colombia1
England1
Total21
Table 4. Analyzed academic works.
Table 4. Analyzed academic works.
TitleCit.MethodologyPub VenueDatabasePrismaUse CaseFrameworkGAI ToolsReview Objective
[23]23Topical Review
(80 Papers)
IEEE Access, 2024IeeeXploreYesEducationNoNoExamines the impact of Gen AI on academic, highlighting the opportunities and challenges of AI, and exploring proposed frameworks.
[24]29Survey
(24 Papers)
Information, 2024MDPINoHigher EducationYesLLMsProposes a framework for ethical integration of AI in HE, emphasizing adaptive regulations, AI literacy, and responsible AI governance.
[25]34Systematic Mapping Review
(407 Papers)
Review of Education, 2024Wiley Online LibraryNoEducationNoChatGPTThe study highlights GAI educational impact, emphasizing gaps in K-12 (i.e., kindergarten through 12th grade in the U.S. system, corresponding to primary and secondary education) integration research.
[26]67Survey
(26 Papers)
IEEE Access, 2024IeeeXploreNoEducationNoNoGen AI enhances education through personalized learning, collaboration, and ethical practices.
[27]66Systematic Review
(355 Papers)
Education Sciences, 2024MDPIYesHigher EducationNoChatGPT-4, GeminiExplores the current state, trends, and research gaps in Gen AI for teaching and learning in HE.
[28]36Systematic Literature Review
(37 Papers)
Information, 2024MDPIYesHigher EducationNoChatGPT, GPT-3.5, DALL-EAnalyzes empirical research on GAI in HE, categorizes its application areas, and proposes future research directions and policy implications.
[29]155Systematic Review
(13 Papers)
European Journal of Education, 2024Wiley Online LibraryYesK-12NoChatGPTExplores ChatGPT’s impact on K-12 education, highlighting its potential and the challenges, emphasizing the need for structured guidelines and collaborative efforts for effective integration.
[30]35Review
(45 Papers)
Higher Education, 2024SpringerNoHigher EducationNoChatGPTExamines the impact of Gen AI on HE, highlighting its potential benefits and challenges in academic integrity and assessment.
[31]44Systematic Review
(70 Papers)
Frontiers in Education, 2024FrontiersYesEducationNoChatGPTExamines the benefits and challenges of ChatGPT, its impact on student engagement and learning outcomes, ethical considerations, and its effects on educators.
[32]43Systematic Literature Review
(57 Papers)
International Journal of Educational Research, 2024ElsevierNoHigher EducationNoChatGPTExamines ChatGPT’s adoption in HE by analyzing trends, user intentions, applications for various stakeholders, research limitations, and future directions.
[33]29Review
(23 Papers)
Higher Education, 2024SpringerNoHigher EducationNoLLMsExplores the use of GAI chatbots in HE reveals limited empirical research, minimal theoretical grounding, and polarized discourse.
[34]205Systematic Review
(40 Papers)
International Journal, 2023CeeolYesEducationNoChatGPTExplores ChatGPT’s transformative role in education, and its key applications with potential challenges.
[35]572Review
(217 Papers)
Sustainability, 2023MDPIYesGeneral education, medical, engineering, HENoChatGPT
GPT-4
Explores the transformative impact of GAI in education, highlighting its potential benefits across disciplines, and addressing ethical concerns like academic integrity, biases, and responsible use.
[36]6Systematic Literature Review
(48 Papers)
Frontiers of Digital, 2024SpringerNoEducationNoNoExamines studies on GAI in education and its role in learning support, instructional design, assessment, and feedback then addressing ethical concerns, AI literacy.
[37]27Research Article
(36 Papers)
IEEE Access, 2024IeeeXploreNoHigher EducationYesChatGPTExplores the use of ChatGPT in education and proposes a decision-making framework to guide policymakers and educational institutions in ensuring its responsible and ethical implementation.
[38]53Systematic Review
(112 Papers)
Education, 2024MDPIYesEducationNoChatGPTExplores benefits and challenges of ChatGPT in education, focusing on engagement, accessibility, ethical concerns, and academic integrity.
[39]2Scoping Review
(24 Papers)
IEEE Access, 2025IeeeXploreYesEngineering and Computing EducationNoDALL·E, GitHub Copilot, RunwayMLExamines the implications of integrating GAI in engineering and computing education across K-12 to tertiary levels, revealing its benefits and challenges.
[40]89Scoping Review
(32 Papers)
International Journal of Educational Technology in Higher Education, 2024SpringerYesHigher EducationNoChatGPTExamines the potential of GAI in personalized feedback and self-regulated learning, challenges in academic integrity, and the need for teacher training, innovative pedagogy, and revised assessment policies.
[41]1Scoping Review
(32 Papers)
Journal of Education, 2025Wiley Online LibraryYesHigher EducationYesNoAssesses the impact of GAI on academic skill development in higher education and proposes strategies for its ethical and effective integration.
[42]215Topical Review
(20 Papers)
IEEe Access, 2024IEEEXploreNoEducation, business, healthcare, mediaNoChatGPT, Bard, Adobe Firefly, RoomGPT, RunwayML, DALL·E, JukeboxProvides a comprehensive technical and applied review of Generative AI models and tools, including their educational potential and societal impact.
[43]9Literature Review
(26 Papers)
2024 47th MIPRO ICT and Electronics Convention (MIPRO), 2024IEEEXploreNoEngineering EducationNoChatGPT, GitHub Copilot, Amazon CodeWhisperer, OpenAI Codex, Replit, Android Studio Bot, Tabnine, DeepCodeThe review discusses how GAI supports programming and ethical competencies in ICT education, and identifies key tools and real-world teaching scenarios.
Table 5. Definitions of Generative AI across selected studies.
Table 5. Definitions of Generative AI across selected studies.
StudyDefinition of Gen AICategory
[23]Machine learning models that can produce novel, original content without being explicitly programmed.Content Creation
[25]AI models that create human-like content across multiple media formats.Human-Like Creativity
[26]Specialized AI that generates new content from previous materials.Content Creation
[27]LLM-powered systems performing complex tasks like content generation, tutoring, and feedback.Content Creation
[33]AI based on or self-supervised models in generating original outputs (text, images).Content Creation
[35]Framework using existing media to simulate human creativity.Human-Like Creativity
[29]Subset of AI tools generating content via advanced techniques rather than human creation.Human-Like Creativity
[40]Subset of AI generating new content including synthetic data.Content Creation
[42]Models generating data resembling training data (GANs, GPT, etc.).Content Creation
[41]AI systems generating text, images, and music from learned patterns.Content Creation
[36]AI with human-like capabilities in language generation, summarization, and prediction.Human-Like Creativity
[43]Models using LLMs and NLP for content generation in engineering education.Content Creation
Table 6. Frameworks for integrating Gen AI into education.
Table 6. Frameworks for integrating Gen AI into education.
StudyMethodologyFocus Area
[23]PAIGE, AI-CRITIQUE, DATS, IDEE
-
Personalization;
-
Ethics;
-
Instructional design;
-
Adaptive feedback
[24]Ethical Integration Framework
-
AI literacy;
-
Human oversight;
-
Responsible governance;
-
Academic integrity
[41]UTAUT-EG (Unified Theory of Acceptance and Use of Technology—Ethical Governance)
-
Academic skill development;
-
Ethical AI adoption;
-
Transparency
[37]Analytical Hierarchy Process (AHP) Decision Framework
-
Ranking ethical risks;
-
Governance strategy selection;
-
Policy design
Table 7. Challenges of integrating Gen AI into education.
Table 7. Challenges of integrating Gen AI into education.
Challenge CategoryStudiesKey Issues
Reliability and Accuracy[31,38,42]Gen AI outputs are prone to hallucinated, biased, or misleading responses. This undermines trust in academic settings where accuracy and critical thinking are essential.
Pedagogical Adaptation[30,33]Educators face uncertainty in adapting pedagogy to AI-driven environments. AI alters engagement, content creation, and assessment, often without institutional support.
Institutional Readiness[26,32]Lack of structured support and teacher preparation leads to over-reliance on Gen AI tools, potentially weakening student creativity and deep learning.
Ethical Risks[24,37]Challenges include plagiarism, academic integrity, data privacy, and algorithmic bias. The need for ethical governance and human oversight is emphasized.
Equity and AI Literacy[35,41]Without inclusive training, equitable access, and policy frameworks, Gen AI may worsen educational inequalities and exclude underrepresented learners.
Table 8. Reported impacts of Gen AI on critical thinking and problem-solving.
Table 8. Reported impacts of Gen AI on critical thinking and problem-solving.
StudyReported EffectFocus AreaRemarks
[41]Positive impact on cognitive, technical, and interpersonal skillsCritical thinking, metacognition, problem-solvingGen AI enhances learning when paired with ethical guidance and teacher support
[38]Supports self-regulated learning and reflective thinkingInstant feedback, adaptive pathwaysEnhances autonomous learning processes
[40]Encourages reflective learning and self-regulationPersonalized feedback, learning supportGen AI offers multiple learning perspectives
[32]Concern about over-reliance on AIRisk to student independenceStudents may delegate thinking tasks to AI without proper instructional design
[26]Cautions about loss of deep learningOveruse of Gen AI toolsEmphasizes the need for guided implementation to prevent cognitive disengagement
[30]Reinforces existing models, lacks transformative impactPedagogical inertiaGen AI supports rather than redefines learning practices
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bouguettaya, S.; Pupo, F.; Chen, M.; Fortino, G. A Meta-Survey of Generative AI in Education: Trends, Challenges, and Research Directions. Big Data Cogn. Comput. 2025, 9, 237. https://doi.org/10.3390/bdcc9090237

AMA Style

Bouguettaya S, Pupo F, Chen M, Fortino G. A Meta-Survey of Generative AI in Education: Trends, Challenges, and Research Directions. Big Data and Cognitive Computing. 2025; 9(9):237. https://doi.org/10.3390/bdcc9090237

Chicago/Turabian Style

Bouguettaya, Sirine, Francesco Pupo, Min Chen, and Giancarlo Fortino. 2025. "A Meta-Survey of Generative AI in Education: Trends, Challenges, and Research Directions" Big Data and Cognitive Computing 9, no. 9: 237. https://doi.org/10.3390/bdcc9090237

APA Style

Bouguettaya, S., Pupo, F., Chen, M., & Fortino, G. (2025). A Meta-Survey of Generative AI in Education: Trends, Challenges, and Research Directions. Big Data and Cognitive Computing, 9(9), 237. https://doi.org/10.3390/bdcc9090237

Article Metrics

Back to TopTop