Next Article in Journal
Creative and Metacognitive Strategies in Anti-Bullying Programs: A Systematic Review
Previous Article in Journal
Too Loud to Ignore: Knowledge, Attitudes, and Practices Regarding Hearing Protection in Student Musicians and Ensemble Directors
Previous Article in Special Issue
Investigating How a Technology-Enhanced, Systems Thinking-Oriented Engineering Course Influences Students’ Attitudes Towards Design and Technology
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Toward a Coherent AI Literacy Pathway in Technology Education: Bibliometric Synthesis and Cross-Sectional Assessment

Department for Physics and Technology, Faculty of Education, University of Ljubljana, Kardeljeva Ploscad 16, 1000 Ljubljana, Slovenia
*
Author to whom correspondence should be addressed.
Educ. Sci. 2025, 15(11), 1455; https://doi.org/10.3390/educsci15111455 (registering DOI)
Submission received: 3 October 2025 / Revised: 25 October 2025 / Accepted: 30 October 2025 / Published: 1 November 2025
(This article belongs to the Special Issue Technology-Enhanced Education for Engineering Students)

Abstract

Rapid advances in artificial intelligence (AI) are reshaping curricula and work, yet technology and engineering education lack a coherent, critical AI literacy pathway. In this study, we (1) mapped dominant themes and intellectual bases and (2) compared AI literacy between secondary technical students and pre-service technology and engineering teachers to inform curriculum design. Moreover, we conducted a Web of Science bibliometric analysis (2015–2025) and derived a four-pillar framework (Foundational Knowledge, Critical Appraisal, Participatory Design, and Pedagogical Integration) of themes consolidated around GenAI/LLMs and ethics, with strong growth (1259 documents, 587 sources). Phase 2 was a cross-sectional field study (n = 145; secondary n = 77, higher education n = 68) using the AI literacy test. ANOVA showed higher total scores for pre-service teachers than secondary technical students (p = 0.02) and a sex effect favoring males (p = 0.01), with no interaction. MANCOVA found no multivariate group differences across 14 competencies, but univariate advantages for pre-service technology teachers were found in understanding intelligence (p = 0.002) and programmability (p = 0.045); critical AI literacy composites did not differ by group, while males outperformed females in interdisciplinarity and ethics. We conclude that structured, performance-based curricula aligned to the framework—emphasizing data practices, ethics/governance, and human–AI design—are needed in both sectors, alongside measures to close gender gaps.

1. Introduction

Artificial intelligence (AI) is increasingly becoming a key technology that is fundamentally changing various areas of society, from science and industry to people’s daily quality of life. Rapid advances in AI are transforming disciplines, the economy, and jobs, and challenging our established notions of human labor and the role of machines (UNESCO, 2009). Given the far-reaching impact of AI on society and the labor market, many analysts warn that AI will create new jobs while changing existing occupations, requiring a reorientation of workforce competencies (Shen & Zhang, 2024). The continued proliferation of AI has further put its benefits and risks under the spotlight, where nearly 40% of all jobs worldwide will be affected in the coming years in most scenarios by replacing some and complementing others (Pizzinelli et al., 2023). Thus, AI, as it is now, is likely to increase overall inequality (Pizzinelli et al., 2023). Moreover, the rapid development of AI could extend much further into professions, as AI not only complements humans in traditional technological solutions but can also replace some cognitive tasks (Cazzaniga et al., 2024). Due to the potential of AI for higher productivity and a greater emphasis on decision-making, empathy, and the creativity of people, human guidance and control might be critical (Stolpe & Hallström, 2024). Thus, AI literacy acquired and balanced through education can be seen as a key driver for working with and managing AI-driven solutions. This includes technical understanding, ethical awareness, and the ability to analytically evaluate the performance of AI. Therefore, educational institutions must provide future professionals with sound technological, pedagogical, and content knowledge as early as the secondary school level, while promoting an understanding of the social and ethical dimensions of AI use (Guzik et al., 2024; Rütti-Joy et al., 2023).
In recent years, there have been several projects and research both in this country and globally about how digital technologies and AI solutions can be most effectively introduced into education. For example, the national GEN-UI (Generative AI in Education) project aims at the comprehensive, effective, and efficient use of generative AI (GenAI) in the learning process at various levels of education (https://gen-ui.si/, accessed on 15 September 2025). The main objective of the GEN-UI project is to (1) investigate the importance of GenAI in terms of perception and changes in learning and teaching from the perspective of various stakeholders, taking into account the didactic, legal, financial, and ethical aspects of the use of GenAI in education in Slovenia; (2) analyze the status and needs of educational institutions for the effective and efficient use of GenAI in the educational process in Slovenia; (3) develop guidelines for effective and efficient use of GenAI for modern learning and teaching of teachers and students; (4) produce sample teaching scenarios in the field of the use of GenAI in selected educational institutions; and (5) develop guidelines for the effective and efficient use of GenAI for and in education.
Teachers and future teachers will play a central role in the meaningful integration of AI into education, especially in technical education, orienting the cadre toward technical and scientific professions. In addition to pedagogical strategies, teachers will need to develop a basic understanding and skills in the use and evaluation of AI tools to effectively teach content using these technologies meaningfully (Univerzitetna založba UM, 2022). Indeed, their AI competencies will largely shape students’ AI literacy, as teachers’ integration of AI content enables students to learn the basics of the technology and familiarize themselves with its meaningful ethical dimensions (Univerzitetna založba UM, 2022). This is also the strategic direction: teacher education must address the new competencies required in the age of AI because teachers are catalysts for change in the learning environment. Unfortunately, research shows that many teachers lack clear guidelines for teaching AI content, which reduces their confidence in integrating these topics (Lumanlan et al., 2025). This confirms the need to systematically develop the competencies of future and current teachers in the field of AI. This is particularly important in technical education, where students will work as future engineers and technicians in professions where AI is already part of everyday life.
The central motivation for our study stems from the challenges described above. We aim to investigate the current state of AI literacy among two key groups in education: students in secondary technical schools and students who will be the future teachers of technical subjects. These are two groups that will directly shape the future of technical and scientific professions: high school students as future professionals and users of AI in the field, and student teachers as communicators of AI knowledge to future generations. It is important to understand to what extent these students and future teachers are already familiar with AI, and what their attitudes, knowledge, and possible gaps in understanding are—in short, how familiar they are with AI. This will determine how well they will integrate AI into their professional work and how AI will affect their societal role and employability. Therefore, it is scientifically and practically relevant to investigate the level of AI literacy in future teachers of technical subjects and in students at technical schools. The results of such an investigation can contribute to the improvement of study programs and curricula.

1.1. Critical AI Literacy and Teacher Education

Critical AI literacy can be understood as the component of AI literacy that emphasizes questioning, interpreting, and reflecting on the assumptions, processes, and implications of AI (Touretzky et al., 2022). Moreover, critical AI literacy is about recognizing how AI systems are shaped by human decisions, identifying the ethical or societal dimensions of these systems, and critically evaluating their trustworthiness and potential biases (Touretzky et al., 2022). Critical AI literacy typically includes (1) the knowledge of how AI works and where it fails; (2) skills for analysis, verification, and critique; (3) ethical/civic orientations (fairness, accountability, and transparency); and (4) agency (Velander et al., 2024).
In Long and Magerko’s often cited definition of AI literacy, they emphasize that people with AI literacy should be able to “critically evaluate AI technologies, communicate and collaborate effectively with AI, and use AI as a tool” (Long & Magerko, 2020, p. 2) in everyday life. Critical AI literacy focuses on the first aspect, critical evaluation of AI, by focusing on audit, ethics, and reflective practice (Veldhuis et al., 2025). Hornberger et al.’s (2023) work refers to Long and Magerko (2020), for example, and states that an essential component of AI literacy is understanding how the development of AI is guided by human decisions and can perpetuate biases if these decisions are not carefully scrutinized. Across both secondary and higher education, there is a growing consensus that AI literacy must be critical, combining technical understanding of AI with critical thinking about AI’s role in society (Khuder et al., 2024; Velander et al., 2024).
The rapid integration of AI across social, educational, and professional domains has rendered AI literacy a fundamental component of contemporary digital competence. Prior studies have emphasized that AI literacy extends beyond mere familiarity with AI tools and encompasses an understanding of core algorithmic principles, data practices, and socio-ethical dimensions (Ng et al., 2021; Schüller, 2022). However, research to date has been characterized by fragmentation: competency models differ considerably across disciplines and educational levels, producing inconsistent guidelines for curriculum design (Chee et al., 2024). Moreover, existing reviews reveal a dearth of coherent learning pathways that span K-12 through higher education into workforce training, leaving practitioners and policymakers without a unified framework for lifelong AI literacy development (Zhang et al., 2025).
The key thematic clusters of critical AI literacy, as suggested by the literature, include technical foundations, AI’s strengths and weaknesses, data/algorithmic reasoning, ethics and societal impact, privacy and governance, agency and design/creation, classroom application and assessment, and interdisciplinarity.
First, technical foundations and data/algorithmic reasoning establish the necessary cognitive groundwork for learners. This cluster includes understanding core AI models (e.g., supervised vs. unsupervised learning, neural networks), data pipelines (collection, cleaning, and feature engineering), and algorithmic interpretability (Chee et al., 2024; Ng et al., 2021). Proficiency here empowers students to deconstruct AI systems, evaluate data quality, and grasp the mathematical and statistical principles underpinning model training and performance metrics (Schüller, 2022).
Second, AI’s strengths and weaknesses, along with ethics and societal impact and privacy/governance, form a critical–reflective dimension. Learners examine AI’s capabilities, such as pattern recognition, automation, and scalability, against its limitations, such as brittleness, bias, and “hallucinations” in large language models (Johri, 2020). Ethical inquiry covers fairness, accountability, transparency, and the social consequences of AI deployment (e.g., labor displacement, surveillance, and power asymmetries) (Atenas et al., 2023; Gartner & Krašna, 2023), while governance literacy addresses data protection regulations (e.g., GDPR), institutional policies, and public sector AI frameworks (Bing & Leong, 2025; Filgueiras, 2023).
Third, agency and design/creation and interdisciplinarity cultivate a creative–proactive posture. Under human-centered and Participatory Design principles, learners co-construct AI artifacts through prototyping, prompt engineering, and iterative error analysis (Johri, 2020; Shiri, 2024). An interdisciplinary approach integrates perspectives from computer science, engineering, sociology, philosophy, and policy studies to foster systems-level reasoning and address “wicked” socio-technical challenges (Chen et al., 2020; Zhang et al., 2025).
Finally, classroom application and assessment prescribe Pedagogical Integration strategies and evaluation tools. Innovative teaching methods, such as constructivist, project-based learning with authentic datasets (Kim et al., 2025), peer-supported collaborative tasks (Joseph et al., 2024), and case-based discussions, should be paired with performance-based assessments, reflective portfolios, and validated AI literacy scales to measure competence across cognitive, critical, and creative dimensions (Lintner, 2024; Schleiss et al., 2023; Walter, 2024).
To reorganize these themes into a coherent, actionable conceptual framework tailored for technology and engineering education, we organize AI literacy into four interrelated pillars:
  • Foundational Knowledge
    • Encompasses procedural and declarative mastery of AI algorithms, data lifecycle management, and model evaluation.
    • Key outcomes: explain backpropagation, design a simple classifier, and critique dataset biases.
  • Critical Appraisal
    • Fosters reflective understanding of AI’s promises and perils, ethical principles, and governance contexts.
    • Key outcomes: identify unfair outcomes in models, debate policy scenarios, and propose mitigation strategies.
  • Participatory Design
    • Empowers learners as co-designers of AI systems through prototyping, prompt engineering, and interdisciplinary collaboration.
    • Key outcomes: develop a conversational agent via iterative testing, conduct stakeholder co-design workshops, and assess the socio-technical impacts.
  • Pedagogical Integration
    • Guides educators in embedding AI literacy across curricula, leveraging active learning and robust assessment.
    • Key outcomes: structure a K-12 pathway from basic concepts to workforce competencies, apply formative feedback on AI projects, and adapt modules across disciplines.
These pillars are mutually reinforcing: Foundational Knowledge enables Critical Appraisal; critical insights inform responsible design; design practices generate new pedagogical approaches; and pedagogy perpetuates the cycle by nurturing deeper technical and reflective skills. By mapping specific competencies, learning activities, and assessment metrics onto each pillar, this framework offers technology and engineering programs a strategic roadmap to develop graduates who not only operate AI tools but also critically shape AI’s role in society.
Within a theoretical framework, our research builds on existing models and standards for the integration of AI in education. Among the most established is the American AI4K12 initiative, which has identified the so-called “five big ideas” (BIs) of AI as the core of the curriculum: BI1: perception; BI2: representation and reflection; BI3: learning; BI4: natural interaction; and BI5: social impact (Touretzky et al., 2022). These concepts serve as guidelines for what students should learn about AI at different levels of school education. International organizations such as the Organization for Economic Co-operation and Development (OECD) and the United Nations Educational, Scientific, and Cultural Organization (UNESCO) have also begun to issue recommendations and frameworks for the development of AI competencies. For example, in 2023, UNESCO developed a competency framework for AI literacy for students and teachers to support countries in integrating AI into their education systems. UNESCO recommends that such frameworks should be seamlessly integrated into all levels of education, along with the provision of appropriate infrastructure and ethical guidelines for the use of AI (UNESCO, 2009). Such standards and guidelines emphasize that the integration of AI in the classroom should be ethical, safe, and human-centered, while effectively developing the necessary knowledge and skills in learners. Our study sheds light on how much progress has been made in implementing these guidelines for the previously mentioned target groups and provides a starting point for further curriculum development and teacher training in the field of AI.
Building on these gaps, this study proposes a comprehensive competency framework that delineates 14 AI literacy competencies, synthesized from empirical studies and established frameworks (Chee et al., 2024; Long & Magerko, 2020; Shiri, 2024). Visual representation of how 14 AI literacy competences can be linked to four conceptual pillars is shown in Figure 1. The visualization supports rapid gap analysis and sequencing: educators can cluster outcomes by pillar, choose appropriate learning activities (e.g., CRIT for ethics debates; DES for prototyping), and scaffold progression from basic recognition and data literacy to pedagogical application. For curriculum designers, the pillar tags translate directly into course modules, assessment rubrics, and cross-course alignment, enabling coherent coverage without duplication. This structure also facilitates accreditation reporting and iterative program improvement by making the intended learning outcomes traceable to pillar-specific evidences.
It has proven crucial that we demonstrate how these competencies must be configured and prioritized differently for distinct learner cohorts, ranging from K-12 students (emphasizing basic AI concepts and ethical awareness) to higher education learners (focusing on data–algorithmic reasoning and problem-solving) and workforce professionals (centering on AI tool utilization, error detection, and decision-making) (Chee et al., 2024). In addition, we articulate an implied developmental pathway that sequences learning objectives across educational stages, thereby addressing the longstanding absence of a structured, longitudinal approach to AI literacy (Chee et al., 2024).

1.2. Objectives and Research Questions of the Current Study

The purpose of this study was two-fold: (1) to map and critically characterize the dominant themes and intellectual bases of AI literacy in technology and engineering education and trace their evolution over time; (2) to empirically compare AI literacy between students in a technology teacher education program and secondary technical school students—examining overall levels, pinpointing which competence areas differ with statistical significance and associated effect sizes, and assessing differences in AI literacy—to generate evidence for targeted curriculum design and pedagogical interventions in technology and engineering education. Accordingly, we formulated the following research questions (RQ1–RQ4):
  • RQ1: What are the dominant themes and intellectual bases of AI literacy in education, and how have these themes evolved over time?
  • RQ2: What are the differences in the level of AI literacy measured by the total score between students in a technology teacher education program and secondary technical school students?
  • RQ3: In which AI literacy competencies are there statistically significant differences between students in a technology teacher education program and secondary technical school students, and what is the effect?
  • RQ4: What are the differences in critical AI literacy between students in a technology teacher education program and secondary technical school students?
Table 1 demonstrates that our measurement strategy is theoretically anchored and coverage-balanced across the core knowledge and critical dimensions of AI literacy. By aligning each bibliometric theme with Long and Magerko’s (2020) competencies as operationalized in the Hornberger et al. (2023) AI literacy test, the table shows strong construct coverage of (1) Representation and Reasoning and Learning (AI4K12 BI2–BI3) through items on knowledge representations, decision-making, supervised/unsupervised learning, and the machine learning (ML) pipeline; and (2) the societal impact strand (BI5) via ethics, human oversight, and legal challenges. Coverage of natural interaction (BI4) is present, but deliberately minimal (e.g., recognizing a chatbot), while perception (BI1) is absent, which is a direct consequence of Hornberger et al.’s (2023) decision not to develop items for robotics-specific competencies (action and reaction, sensors) and to exclude “Imagine the Future of AI” from the test blueprint. These instrument design choices are documented in their test development rationale and item-to-competency allocation, and they explain the mapping gaps in Participatory Design/Creation and privacy/governance depth (only one legal/policy item) that we treat as curricular implications rather than measurement targets in this study. Importantly, Hornberger et al. (2023) validate a unidimensional structure and report good reliability for a 30-item + 1 sorting task instrument, which justifies the use of a total AI literacy score in our between-group comparisons, while our table clarifies its content coverage across technical, critical, and societal facets.
The proposed framework offers actionable guidance for educators, curriculum designers, and policymakers to shift the pedagogical emphasis from transactional tool use toward critical, strategic, and ethical engagement with AI technologies. In practice, K-12 curricula should integrate project-based modules in which students employ constructivist datasets to explore AI mechanics and impacts (Kim et al., 2025), while higher education programs must embed interdisciplinary case studies that foster data literacy, algorithmic transparency, and governance considerations (Atenas et al., 2023; Filgueiras, 2023). For workforce training, organizations ought to adopt performance-based assessments and reflective portfolios that measure AI competence in real-world tasks (Joseph et al., 2024; Lintner, 2024). At the policy level, ministries of education and accreditation bodies should formalize a lifelong AI literacy continuum—spanning teacher professional development in intelligent TPACK (Velander et al., 2023), standardized assessment scales (Lintner, 2024), and cross-sector governance frameworks—to ensure equitable access and sustained skill development from early schooling through career progression (Filgueiras, 2023; Schüller, 2022). As Rupnik and Avsec (2025) noted, the beginning and end of schooling are very stressful for students, so increasing AI literacy could help reduce discomfort and increase the sense of successful completion, thereby reducing dropout rates and enabling high school students to make more informed decisions about further education.

2. Materials and Methods

In this study, we used a two-phase sequential quantitative design: Phase 1: a bibliometric evidence synthesis to derive and structure the conceptual framework for (critical) AI literacy; Phase 2: a non-experimental, cross-sectional comparative field study in authentic classroom settings to validate the framework Via the Hornberger et al. AI literacy test and compare cohorts.

2.1. Characteristics of Teacher Education and Secondary Technical School Study Programs

2.1.1. Faculty of Education

Information and communication technology (ICT) is systematically integrated into the study programs at the Faculty of Education, University of Ljubljana (n.d.). The dual-subject Mathematics–Computer Science program has ICT at its core: students deepen their knowledge of programming, algorithms, and data technologies while developing pedagogical and didactic competencies for high-quality teaching of mathematics and computer science in primary and secondary schools (Faculty of Education, University of Ljubljana, n.d.). In the Mathematics–Technology and Physics–Technology programs, ICT is an important supporting element, as students learn CAD, work with microcomputers and robotics, and use digital tools for experiments and technical drawing (Faculty of Education, University of Ljubljana, n.d.). ICT is also used in the Biology–Chemistry program for experimental work, measurement systems, and simulations, and in the Chemistry–Home Economics and Biology–Home Economics programs for planning, monitoring, and evaluating practical activities with specialized software (Faculty of Education, University of Ljubljana, n.d.).

2.1.2. Koper Technical Secondary School

The programs at Koper Technical Secondary School consistently incorporate ICT and strengthen digital skills and critical literacy regarding artificial intelligence (Secondary Technical School Koper, n.d.). The technical high school (4-year program) builds the foundations of computational thinking through computer science and mechanics courses and prepares students for further study in technical fields. The computer technician program (2-year vocational-technical) provides an in-depth understanding of computer systems, programming, networks, and IT infrastructure maintenance, and encourages systematic thinking about the operation of complex AI systems (Secondary Technical School Koper, n.d.). The three-year computer science program emphasizes the practical use of ICT in professional situations, from hardware and software maintenance to information management, with a clear focus on digital literacy and critical thinking (Secondary Technical School Koper, n.d.). The mechatronics technician program (2-year PTI) combines mechanical engineering, electrical engineering, and computer science with the use of controllers (PLC), microcomputers, and sensors in automation; the mechatronics operator program (3-year) trains students to work directly with automated systems and basic programming (Secondary Technical School Koper, n.d.). The Mechanical Technician program (4 years) combines classic engineering content with CAD/CAM, CNC machine programming, and simulations, and promotes digital innovation and the thoughtful introduction of AI into industrial environments (Secondary Technical School Koper, n.d.).

2.1.3. Škofja Loka School Center

Škofja Loka School Center offers a range of programs with a strong emphasis on ICT, developing the digital competencies and critical AI literacy of future technical staff (School Centre Škofja Loka, n.d.). The technical high school (4-year program) provides a broad scientific and technical foundation, including an introduction to programming and computational thinking, and encourages critical evaluation of technologies such as artificial intelligence (School Centre Škofja Loka, n.d.). Mechanical engineering offers in-depth knowledge of computer-aided design, CNC machine operation, and automated systems using specialized software (School Centre Škofja Loka, n.d.). The metalworking program develops precise use of CNC technologies, algorithmic thinking, and problem solving; programs for the maintenance and assembly of mechanical systems introduce computer-aided procedures for assembly, diagnostics, and servicing (School Centre Škofja Loka, n.d.). The mechanical installation technician program uses digital tools for technical documentation and management of modern smart energy systems (School Centre Škofja Loka, n.d.).

2.2. Participants and Field Data Collection

In a cross-sectional research design, the online test was administered in class under proctored conditions with the principal investigator present; students completed it on personal mobile devices (smartphones/tablets, e.g., iPads). Administration was device-agnostic, single-session, and standardized across classes. We collected data from secondary technical school students (n = 77, 53.1%) and university students (n = 68, 46.9%). Secondary technical school students were from two high schools, while university students belonged to one university. Data was collected in the 2024/2025 school/academic year. In total, 145 students were engaged in this study, of whom 58 were females (40%) and 87 were males (60%). Secondary school students were on average 17.9 years old (SD = 1.26), while their counterparts from the university were 21.24 years old on average (SD = 2.07).
The Hornberger et al. (2023) AI literacy test (grounded in the work of Long and Magerko (2020)) was administered in class under proctored conditions with the principal investigator present. Students completed the online test on personal mobile devices (smartphones/tablets, e.g., iPads) in a session length of 35–45 min; identical instructions across classrooms were provided. No incentives were offered.
This study was conducted in accordance with the Declaration of Helsinki and was reviewed and approved by the Ethics Commission of the Faculty of Education of the University of Ljubljana (approval code: 7/2025). The AI literacy test was delivered to students online via the 1KA portal (https://www.1ka.si/d/sl, accessed on 15 September 2025), with the main examiner present in the ICT-equipped classroom.

2.3. Research Methods

For this study, we used bibliometrics together with empirical testing as the primary method.

2.3.1. Bibliometrics

Firstly, a bibliometric method was used. We queried the Web of Science Core Collection (WoS CC) for records in English published from 2015 to 2025 (inclusive). Document types included articles, reviews, proceedings articles, and early access retained articles. The search strategy (WoS/TS field) is shown in Appendix A (Table A1).
We exported full records + cited references (CSV) on the final search date. Screening included (1) inclusion: empirical or review work on AI/algorithmic/data literacy (including critical/ethical/governance aspects) in education contexts relevant to technology and engineering, teacher education, and TVET/technical secondary; (2) exclusion: non-educational AI articles; purely technical CS without a literacy/competence focus; editorials, notes, and news; non-English articles; retracted items; and duplicates.
Data curation and cleaning consisted of (1) deduplication by DOI ± exact/near-title match; keeping WoS “early access” merged with final versions; (2) standardization: lowercase; strip punctuation; harmonizing US/UK variants; unifying synonyms (e.g., AI <-> artificial intelligence; pre-service <-> preservice; and TVET <-> technical and vocational education and training); (3) the stop list for overly general terms in co-word analyses: artificial intelligence; AI; education; student; teacher; higher education; case study; literature review; and (4) stemming/lemmatization for keywords where supported.
As a primary tool and for reproducibility, the R/bibliometrix tool was used with the visualization tool Biblioshiny 5.0 (Aria & Cuccurullo, 2017) for performance indicators, co-word networks, thematic maps, and co-occurrence/co-citation visuals.
We later mapped emergent themes to the Hornberger et al. (2023) AI literacy test (aligned to the work of Long and Magerko (2020)) to provide evidence of content validity in our measurement section.
Network analyses and parameters can be delivered as (1) co-word (conceptual structure) and co-occurrence/co-citation: source = author keywords (DE); min keyword frequency = 3; normalization = association strength; and counting = full; (2) co-citation (intellectual base): unit = cited references; threshold = ≥10 citations (reference-level) and ≥20 (source/journal-level); and normalization = association strength; show top 50–100 nodes by weight.
Planned outputs can be (1) basic descriptive and performance summary (annual output; top venues/authors/countries); (2) thematic map + evolution plot (main text); (3) co-citation maps; and (4) coverage check linking high-centrality themes to instrument content (Results/Discussion).

2.3.2. AI Literacy Test

For the purpose of this study, we used an AI literacy test developed by Hornberger et al. (2023), which is grounded on a widely accepted AI literacy framework developed by Long and Magerko (2020). The test items were systematically created and also refined in a cognitive interview and expert reviews. The test was administered several times, and items were also thoroughly examined using response theory, which provides validity evidence. Statistical checks confirmed that the test primarily measures one factor, where 14 competencies from Long and Magerko’s framework (2020) are successfully covered. The test is flexible enough to be adapted for both high school and university settings. It captures the core knowledge of AI—encompassing its nature, technologies, societal/ethical impacts, and data-related considerations—in a psychometrically rigorous manner, making it a justifiable choice for measuring AI literacy across these student populations (Hornberger et al., 2023). As a result, our final instrument covers the following 14 competencies: (1) recognizing AI, (2) understanding intelligence, (3) interdisciplinarity, (4) general vs. narrow, (5) AI’s strengths and weaknesses, (6) representations, (7) decision-making, (8) machine learning steps (i.e., the iterative process of training and testing AI), (9) the human role in AI, (10) data literacy, (11) learning from data, (12) critically interpreting data, (13) ethics, and (14) programmability. The final test consists of 31 items, where 30 of them were structured as multiple-choice items with one correct answer plus three distractors. The dichotomous scoring model was used, where responses are either universally correct with a value of 1 or incorrect with an outcome of 0. At one item, students were asked to sort the five process steps in supervised learning into the correct order, where only the correct order was graded with 1 point. The maximum possible score on the test was 31 and calculated as the sum of all dichotomous responses. Competency scores were calculated by summing the dichotomous responses (coded as 0 or 1) for all items related to each competency. The sum was then divided by the total number of items within each competency, resulting in a proportion ranging from 0 to 1, where higher values indicate a higher level of competency.
The AI literacy test used in our study achieved moderate internal consistency (Cronbach’s α = 0.76), indicating that items consistently measure the same underlying constructs, namely, basic knowledge, conceptual understanding, critical thinking skills around AI, and attitude/awareness, especially around ethics and AI’s impact on society (Tabachnick & Fidell, 2013).
AI4K12 tagging was performed post hoc by mapping each Hornberger et al. (2023) item’s Long and Magerko (2020) competency to the AI4K12 big ideas (see Table 1). Item-to-competency allocations and the removal of Item 07 follow the work of Hornberger et al. (2023). Robotics/sensing competencies were not included in item generation, which explains the absence of BI1 (perception).

2.4. Field Study Data Analysis

The data was analyzed using IBM SPSS software (v.25). The Cronbach alpha coefficient was used to support the reliability of constructs. In addition, descriptive statistics were used to summarize and describe the main characteristics of a dataset, such as the mean and standard deviations of the dependent variable, while multiple analysis of covariance was used to identify and confirm significant differences in AI literacy amongst students from different school settings. A Shapiro–Wilk test was used to check whether the data were taken from a normally distributed sample. An effect size partial eta-squared value (η2) was used to measure the strength of the relationship between variables.

3. Results

3.1. Bibliometrics Analysis of the AI Literacy Conceptual Framework

Our collection comprises 1259 documents published across a substantial 587 unique sources. This indicates a moderately sized yet diverse body of literature. The high number of distinct sources suggests that the research topic is not confined to a few specialized journals but is interdisciplinary or widely published across various fields. The timespan reveals an extremely recent collection, with an average age of just 0.768 years. The inclusion of “2025” in the timespan likely indicates that our WOS download was very recent and captured “early access” or “in press” articles, or perhaps the data pull was configured to project slightly into the future to include forthcoming publications. This strong emphasis on very recent publications means the analysis is focused on the cutting edge of the field, capturing the very latest developments and trends.
A remarkable 4810 unique authors contributed to these 1259 documents. This signifies a vast network of researchers actively engaged in this domain, indicating a vibrant and highly populated research community. The presence of 885 Keywords Plus (system-generated) and a much larger 3463 authors’ keywords (author-provided) suggests a rich and diverse thematic landscape. The higher number of author keywords usually points to a more granular and specific representation of the topics being explored, allowing for a detailed thematic analysis.
The collection covers a broad yet current range of research within its defined scope, involving a significant global community of scholars. The recency of the publications suggests a focus on emerging themes and breakthroughs.
An annual growth rate of 125.47% indicates an extraordinarily high annual growth rate. A rate of 125.47% suggests an explosive growth in publications within this field over the specified timespan. This is a strong indicator of a rapidly emerging, highly active, or perhaps newly recognized research area that has gained significant traction very recently. This is a domain that is experiencing a boom in scientific output.
The average of 4.37 co-authors per document is quite high, and only 174 documents (approx. 13.8%) are single-authored. This clearly indicates that research in this field is predominantly collaborative. It is not a domain where lone scholars typically work; teamwork and shared expertise are the norm. More than a quarter of our documents feature international collaboration, 26.13%. This is a healthy percentage, demonstrating that the research community is not only collaborative within national boundaries but also engages across borders, fostering diverse perspectives and potentially leveraging global expertise and resources.
The field is experiencing exponential growth, driven by a highly collaborative and internationally connected research community. Productivity is robust, characterized by a rapid increase in original research articles and systematic reviews.
An average of 10.46 citations per document is a very good figure, especially when considering the extremely young average age of the documents (0.768 years). Publications typically accumulate citations over time. For documents that are, on average, less than a year old, achieving over 10 citations per document is impressive. This suggests that the research in this collection is not only current but is also being quickly recognized, read, and cited by other researchers, indicating its immediate relevance and influence.
Despite the extreme recency of the collection, the publications are already demonstrating a significant and immediate impact, rapidly gaining traction, and influencing subsequent research. This suggests that the work being carried out is highly relevant and valuable to the wider scientific community. Key findings from the analysis are as follows:
  • A “Booming” Field: The most striking finding is the explosive growth rate, combined with the very young average age of documents and their already significant citation impact. This strongly suggests the analysis of a research area that has either recently emerged, experienced a critical breakthrough, or is attracting a massive surge of interest and investment.
  • Highly Collaborative Environment: The high average co-authorship and significant international collaboration point to a field where networked research and shared expertise are paramount.
  • Timely and Influential Research: The studies, despite their newness, are quickly becoming foundational, indicating the immediate utility and relevance of the work.
The provided plot (see Figure 2) vividly illustrates the rapid evolution of the research landscape concerning AI in education from 2017 to a projected 2025. The Sankey diagram shows the dynamic flow and transformation of topics, while the strategic maps provide a detailed snapshot of thematic positioning based on their centrality (relevance/interconnectedness) and density (development/cohesiveness) within each period.
The most striking and dominant trend across all periods is the explosive emergence and subsequent entrenchment of ChatGPT and Large Language Models (LLMs). From being virtually absent in 2017–2022, they quickly became foundational and driving forces in subsequent periods. “Artificial intelligence” itself, along with “AI literacy” and “ethics” (initially), consistently appear as central or motor themes, indicating their enduring importance as overarching concepts and critical research areas. The research trajectory clearly shifts from a general understanding of AI to its specific integration, challenges, and impact within educational settings, encompassing teaching, learning, and professional development. Over time, the field demonstrates a move toward more rigorous quantitative methodologies, with “structural equation modeling” emerging as a foundational theme. As the core themes mature, the research landscape diversifies significantly, with many specialized topics emerging in the “Niche Themes” quadrant.
Our analysis provides an interesting view into the evolving landscape of AI in various domains, particularly education, over recent years (see Figure 2). The analysis covers a rapidly evolving field, as indicated by the distinct short time periods (2023–2023, 2024–2024, and 2025–2025), which highlights the accelerating pace of research and publication. The use of the authors’ keywords field for clustering provides a conceptual understanding of topic co-occurrence and evolution. The topic evolution map provides a macroscopic view of how research themes have emerged, persisted, and shifted over time. The increasing number of nodes and the complexity of connections from the period 2017–2022 to 2025–2025 illustrate a rapidly expanding and diversifying research landscape.
This plot in Figure 2 vividly illustrates the dynamic nature of research themes:
  • Foundational phase, 2017–2022: The initial period shows a broad interest in fundamental AI concepts (“artificial intelligence,” “machine learning”), their application in education (“AI in education”), and critical skills (“critical AI literacy,” “digital literacy,” and “computational thinking”). Ethical considerations (“ethics”) are already a motor theme, indicating early awareness. Sector-specific applications such as “healthcare,” “k-12 education,” and “medical education” appear as distinct, albeit less interconnected, areas.
  • Rapid expansion and specificity, 2023–2023: This period shows a significant increase in thematic diversity, likely driven by recent advancements in AI. “Artificial intelligence” and “AI in education” remain central hubs, spawning connections to more specific areas. New themes emerge, including “academic integrity,” “assessment,” “bias,” “pre-service teachers,” “literature review,” and “perception.” This suggests a shift toward understanding the *implications* of AI on educational practices and specific stakeholders. “Computational thinking” and “deep learning” also gain prominence.
  • Deepening and diversification, 2024–2024: The network continues to expand. “Artificial intelligence” maintains its pivotal role, linking to areas such as “academic writing,” “computing education,” “educational technology,” “learning,” “professional development,” and “teaching.” “AI in education” and “digital literacy” continue to connect to these new pedagogical and technological integration themes. This phase highlights a move toward practical implementation and the development of educational strategies.
  • Integration, impact, and future focus, 2025–2025: This period presents a highly fragmented yet interconnected web of topics. “Artificial intelligence” and “AI in education” (and their closely related concepts such as “critical AI literacy,” “digital literacy,” and “ethics”) act as central anchors, connecting to a wide array of themes related to (a) human impact: “anxiety,” “feedback,” “perception”; (b) educational outcomes: “competencies,” “skills,” “scale development,” “sustainable development,” and “quality education”; (c) specific contexts: “early childhood education,” “medical education,” and “teacher training”; and (d) methodology/implementation: “structural equation modeling” and “teacher training.”
The thickness of arrows leading from “artificial intelligence” and “AI in education” to many new and diverse topics in 2025–2025 signifies their sustained and growing influence as overarching domains that encompass a broadening spectrum of research questions.
We also analyzed the strategic maps, which allowed us to understand the positioning of topics based on their relevance (centrality) and development (density) within each period. These maps are organized into four quadrants as follows:
  • Motor Themes (Top Right): High centrality, High density. Well-developed and central to the field, they drive research.
  • Niche Themes (Top Left): Low centrality, High density. Specialized areas, internally well-developed but less connected to the broader field.
  • Basic Themes (Bottom Right): High centrality, Low density. Foundational concepts, cross-cutting but less internally developed (often because they are broadly accepted prerequisites).
  • Emerging or Declining Themes (Bottom Left): Low centrality, Low density. Peripheral topics, either new and gaining traction or fading, are shown in Appendix B (see Strategic maps Figure A1, Figure A2, Figure A3 and Figure A4).
Across the period studied, the thematic cartography reveals consolidation around GenAI with LLMs, notably ChatGPT, emerging in 2023 as a durable, field-defining basic theme and remaining foundational through 2025, while the broader “chatbots” strand peaks as a motor theme in 2023–2024 before drifting toward the basic theme as discourse and practice coalesce around LLM-centric affordances and risks (Bettayeb et al., 2024; Boscardin et al., 2023; Eysenbach, 2023; Haroud & Saqri, 2025; Jensen et al., 2024). Foundational constructs—AI, AI literacy, and critical AI literacy—likewise stabilize in the basic quadrant by 2025, reflecting their anchoring role in curricula and competence frameworks and the growing effort to instrument and standardize assessment through scale development and validation (Lintner, 2024; Salhab, 2024; Schüller, 2022; Walter, 2024; Wilby & Esson, 2023). Applications show differentiated trajectories: medical students and medical education shift from motor (2023–2024) themes to basic (2025) themes as scoping and rapid reviews consolidate use cases (virtual patients, decision tutoring, and content generation), but also note limited outcomes evidence and calls for standardized competencies (Boscardin et al., 2023; J. Lee et al., 2021; Hale et al., 2024; Rincón et al., 2025; Sun et al., 2023). In parallel, K-12 AI education moves from the emerging theme to the motor theme by 2024 and sustains high centrality into 2025 as teacher readiness, AI literacy interventions, and design-based pedagogies scale up despite persistent capacity and policy gaps (I. A. Lee et al., 2021; Ng et al., 2022; Relmasira et al., 2023; Velander et al., 2023). Methodologically, the center of gravity moves rightward with an uptick in synthesis and rigor: scoping reviews, systematic reviews, and rapid reviews proliferate across domains, and structural equation modeling (e.g., PLS-SEM) features in studies interrogating human factors and adoption concerns, together signaling maturation of evidence practices by 2025 (Ahmad et al., 2023; Bettayeb et al., 2024; Hale et al., 2024; Hwang et al., 2022; Ji et al., 2022; J. Lee et al., 2021; Lintner, 2024; Sun et al., 2023; Tahiru, 2021). Ethical and societal strands remain salient yet stratify: digital literacy recedes from a broad basic theme anchor toward a more niche position as discourse narrows to AI-specific literacies and governance frameworks (Haroud & Saqri, 2025; Schüller, 2022), acceptance holds in the basic theme amid pragmatic integration in higher education, whereas anxiety occupies a locus nearer the motor theme, animated by concerns over bias, academic integrity, privacy, and skill atrophy documented across sectors (Ahmad et al., 2023; Bettayeb et al., 2024; Haroud & Saqri, 2025; Jensen et al., 2024; Wilby & Esson, 2023). Finally, interaction/skills themes, especially human–AI collaboration, stabilize as a niche with credible potential to move rightward, as emerging design knowledge on student–AI orchestration and teacher–AI role complementarity accumulates but still lacks large-scale empirical validation (Ji et al., 2022; Kim et al., 2025). Collectively, these movements portray a field converging on core literacies and LLM-mediated applications, deepening methodological synthesis and negotiating human-centric concerns, while application domains (medical and K-12) and collaboration paradigms supply the principal engines of near-term centrality.
A data-driven reading of theme dynamics suggests clear priorities. Researchers should concentrate on emergent frontiers—prompt engineering, AI regulation, and human–AI collaboration—while attending to demographic- and context-specific questions and ethical implications. Apparent declines warrant interpretation: many topics have not vanished but have been absorbed into broader, application-first conversations. Stable/basic themes remain essential scaffolding for cumulative inquiry; work that refines AI literacy or advances robust applications of ChatGPT/LLMs will retain high relevance. Niche clusters highlight under-connected sub-fields in which synthesis, theory-building, or bridge studies could unlock broader impact. Finally, the shift from “chatbots” to LLMs as basic themes, together with the prominence of prompt engineering and collaboration, points toward a near-term emphasis on optimizing human–AI interaction and augmenting (rather than replacing) human expertise across diverse educational settings.
We can sum up the key insights and provide guidance for data-driven interpretation as follows:
  • The most striking finding is the rapid and profound impact of GenAI, epitomized by “ChatGPT” and “Large Language Models.” These terms moved almost instantaneously from non-existence (before 2023) to becoming central “basic themes” by 2023–2023 and onward. This indicates they are not merely new topics but fundamental shifts that redefine the research landscape, necessitating a re-evaluation of established practices and theories. Researchers must integrate GenAI tools and their implications into their studies. This includes exploring “prompt engineering” (implied by “engineering” in 2025–2025), addressing “academic integrity” (emerging in 2023–2023), and considering new forms of “digital AI literacy.”
  • A shift in the focus to human elements is detected in recent periods. The field is steadily moving beyond purely technological discussions of AI toward a more nuanced understanding of its integration into human learning environments. Early dominance of core AI concepts (e.g., “machine learning” and “deep learning”) as distinct basic themes gives way to their integration, while “basic” and “emerging” quadrants increasingly feature terms such as “competencies,” “teacher training,” “student well-being,” “engagement,” “motivation,” and “teacher perception.” Future research should prioritize the human experience of AI, focusing on how AI impacts learners, educators, and the educational process itself. This includes developing evidence-based pedagogical designs and understanding the psychological and social implications of AI.
  • The increasing presence of advanced research methodologies such as “factor analysis,” “systematic literature review” (as motor themes), and “structural equation modeling” (as a basic theme) indicates a maturing field. Researchers are moving toward more rigorous and robust quantitative and synthesis methods. Adopting and innovating with advanced research methods will be crucial for producing high-quality, impactful research. Methodological sophistication can lead to more reliable and generalizable findings.
  • While “AI in education” remains a broad domain, there is growing specialization into specific educational contexts (“K-12 education,” “early childhood education,” “higher education,” “medical education,” and “nursing education”) and specialized literacies (“critical AI literacy,” “digital AI literacy,” and “communication AI literacy”). Geographical specificities (“Saudi Arabia”) also emerge, suggesting tailored research needs. Researchers should consider the unique challenges and opportunities of AI in diverse educational settings and cultural contexts. Tailored solutions and context-specific research are likely to gain prominence.
  • “Ethics” was an early motor theme and remains central. “Bias” and “academic integrity” emerged quickly, and “AI regulation” is emerging. “Anxiety” and “student well-being” are also starting to be featured. This reflects a growing awareness of the potential risks and negative impacts of AI. Research on responsible AI, including ethical frameworks, fairness, transparency, and the psychological impact of AI, will be paramount. Developing guidelines and policies for AI use in education will be a key area.
In conclusion, the bibliometric analysis clearly shows a dynamic field experiencing rapid growth, driven significantly by the advent of GenAI. The research focus is evolving from general technical exploration to specific pedagogical applications, human-centric impacts, and rigorous methodological approaches. Researchers should strategically align their work with these emerging trends, paying particular attention to the practical, ethical, and human dimensions of AI integration in education.
The visualization in Figure 3 provides valuable insights into the thematic structure and key concepts within the Web of Science (WoS) dataset, based on authors’ keywords.
The network is highly centralized and dense, particularly around the terms “artificial intelligence” and “ethics, critical AI literacy.” These nodes are the largest in size, indicating their high frequency and significant co-occurrence with other keywords. Their central position signifies that they are the foundational and most overarching themes of the research collection. The network broadly follows a hub-and-spoke model, with “artificial intelligence” acting as the primary central hub, from which numerous connections radiate outward to various related concepts and communities. “Ethics, critical AI literacy” serves as a crucial secondary hub, especially within its own community.
The use of different colors (red, purple, blue, and green) clearly delineates distinct research communities or thematic clusters. This indicates that while “artificial intelligence” is the overarching subject, the research explored through these keywords branches into several focused sub-domains.
The Walktrap algorithm (Donthu et al., 2021) has successfully identified four prominent communities, each representing a specific thematic focus:
  • Red Community: Core AI in Education and Ethical Implications (Central Hub). Key Terms: “artificial intelligence” (largest), “ethics, critical AI literacy” (largest), “large language models,” “education technology,” “pre-service teachers,” “innovation,” “sustainable development,” “collaboration,” “analysis,” “literature review,” “research,” “structural equation modeling,” “computational thinking,” “privacy,” “communication,” “natural language processing,” “health literacy,” “transformation,” “skills,” “creativity,” “competencies,” “chatbots,” “human–computer interaction,” “educational assessment,” “higher education,” “curriculum development,” “human-centered AI,” “responsible AI,” “digital literacy,” “digital citizenship,” “digital transformation,” and “learning analytics.” This is the dominant and most comprehensive community, reflecting the multifaceted discourse around artificial intelligence, especially in the context of education. The prominence of “ethics, critical AI literacy” alongside “artificial intelligence” highlights a strong and mature focus on the societal, ethical, and responsible integration of AI. The inclusion of “large language models” and “chatbots” points to research on cutting-edge AI technologies. Terms such as “pre-service teachers” and “higher education” indicate a focus on educator training and tertiary education. This community also encompasses broader research aspects such as sustainability, collaboration, and various research methodologies.
  • Purple Community: Pedagogical Applications and Educational Stages (Left). Key Terms: “AI in education” (bridge term), “teacher training,” “university students,” “teacher education,” “early childhood education,” “pedagogy,” “assessment”, “academic writing,” “K-12 education,” “feedback,” and “bias.” This community zeroes in on the practical application and implications of AI across different educational stages and settings. It emphasizes the “how-to” and “who” of AI integration, focusing on specific learner groups (university, K-12, and early childhood) and educational processes (teacher training, pedagogy, assessment, academic writing, and feedback). The presence of “bias” indicates critical scrutiny of AI’s fairness and equity in educational contexts. “AI in education” serves as a crucial bridge node, connecting this practical cluster back to the main AI theme.
  • Blue Community: AI in Medical and Healthcare Education (Bottom Right). Key Terms: “machine learning,” “deep learning,” “curriculum,” “medical education,” “healthcare,” “medical students,” “survey,” “perception,” “e-learning,” and “systematic review.” This distinct cluster demonstrates a specialized research niche focusing on the application of specific AI sub-fields, such as “machine learning” and “deep learning,” within the “medical education” and “healthcare” domains. It explores how these technologies are integrated into curricula for “medical students” and the broader healthcare field. Terms such as “perception” and “survey” suggest studies on attitudes and understanding within this professional group, while “systematic review” points to a methodological approach.
  • Green Community: Technology Acceptance and Attitudes (Far Right). Key Terms: “technology,” “attitude,” “acceptance,” and “educational.” This smaller, yet significant, community investigates the human dimension of technology adoption. It focuses on the “acceptance” and “attitude” toward new technologies, like AI, within “educational” settings. This suggests research exploring factors influencing the willingness of stakeholders (e.g., students, teachers) to adopt and utilize AI tools.
The largest nodes are inherently the most relevant as they represent the most frequently co-occurring and central themes:
  • “Artificial intelligence”: This term is the absolute core of the entire dataset. All other concepts and communities revolve around it, affirming its status as the primary subject of the collected research.
  • “Ethics, critical AI literacy”: The equally large size and central position of this node within the red cluster highlight that the discourse is not merely about AI technology itself, but profoundly about its responsible development, deployment, and the necessity for users to critically understand its implications. This signifies a mature and ethically aware research field.
  • “Large language models”: The prominence of LLMs shows that the most recent advancements in AI are actively being discussed and researched within this academic domain.
  • “AI in education”: This node acts as a critical link, solidifying the application context of “artificial intelligence” within the “education” field, and connecting the central theoretical/ethical discussions to more practical pedagogical inquiries.
  • “Machine learning”/“deep learning”: These terms indicate that research is also focused on the specific underlying AI techniques, particularly within applied domains such as medical education.
We can conclude that this “Authors’ Keywords network” reveals a dynamic and multifaceted research landscape centered on artificial intelligence. While the core interest lies in AI’s application and implications within education, there is a strong emphasis on ethical considerations and critical understanding. Furthermore, the network highlights specialized applications in fields such as medical education and an overarching concern with technology acceptance. This comprehensive map can guide researchers in identifying emerging trends, potential collaborators, and key gaps in the existing literature.
Insights from Phase 1 guided our expectations for Phase 2. The bibliometric maps show the field consolidating around two intertwined strands—foundational/cognitive knowledge (representation, learning, and related computational constructs) and critical/ethical engagement—within an LLM-centric landscape that foregrounds pre-service teacher contexts and responsible, human-centered AI (anticipating our cohort comparison).
Consistent with the Hornberger-based instrument’s content coverage (robust in BI2–BI3, substantial in BI5), we therefore expected (RQ2) higher total AI-literacy scores among pre-service technology teachers, given their greater exposure to structured coursework in core AI concepts, and (RQ3) group differences concentrated in computationally oriented competencies (e.g., understanding intelligence, programmability).
At the same time, because Phase 1 situates ethics/criticality as a central, widely diffused theme—and prior work characterizes these critical competencies as transversal and often developed through general digital/media literacies—we anticipated limited between-group divergence in critical AI literacy (RQ4).
Finally, the unidimensional structure and reliability of the instrument support the use of a total score for RQ2 while our theme-to-competency map clarifies how specific subscale contrasts in RQ3 instantiate the conceptual emphases identified in Phase 1.

3.2. AI Literacy Level and Differences

Using the published item-to-competency allocation of the Hornberger et al. (2023) test, our instrument covers BI2 (representation and reasoning) and BI3 (learning) robustly (≥17 items combined), offers substantial BI5 (societal impact) coverage (≥7 items across ethics, legal, and human-in-the-loop), includes minimal BI4 coverage (natural interaction) (chatbots), and, by design, no direct BI1 (perception) coverage because robotics/sensing competencies were not included in item development.
Table 2 shows results from descriptive statistics. Firstly, the total score of AI literacy is shown, followed by the scores of each competency from the work of Long and Magerko (2020). In total AI literacy, students from a technology teacher education program, on average, outperformed secondary technical school students (12.40 and 11.97, respectively). To compare students’ AI literacy measured in our study with results from the study of AI literacy in higher education settings, our average results of technology teacher education students (M = 12.40) are much lower than engineering and natural science students (M = 20.04), but comparable to social science students (M = 13.65) (Hornberger et al., 2023). It seems that for our higher education students, AI literacy is much closer to their counterparts in the United Kingdom and the United States of America (Hornberger et al., 2025).
In a very recent international study by Hornberger et al. (2025), the authors indicated that German higher education students scored higher in comparison with UK and USA students (0.38, −0.12, and −0.24, respectively), while, based on our study, higher education students in Slovenia engaged in teacher education programs are comparable to UK students (−0.015). For comparison, we used item response theory (IRT) scores since they provide a more exact measurement.
The results of our study confirm some of the findings of a recent study by Licardo et al. (2025), in which almost half of high school students engaged in the study believe that their knowledge of AI is very poor or negligible, while the percentage among university students is above half of all those who engaged in the study. It was also found that more than 80% of both high school and university students have not received any organized education or training in the use of AI (Licardo et al., 2025). The survey also found that the school or faculty environment does not sufficiently encourage the development and use of AI. This is the opinion of almost half of the students, while the percentage of students who need more support is slightly lower (Licardo et al., 2025). Therefore, as the authors point out, the knowledge, extent of use, and impact of AI on learning are still unclear or insufficiently articulated (Licardo et al., 2025).
To identify differences between groups of students, we first checked the assumption of normal data distribution. The Shapiro–Wilk test of normality was used for AI literacy in total and for AI literacy competencies. The test revealed that a dataset across different study groups comes from a normal distribution (p > 0.05). Since the normality assumption was met, parametric tests were performed to reveal the differences between the groups involved in this study. Next, we ran a 2 × 2 between-subject factorial analysis of variance (ANOVA) where we had two independent variables, group membership and sex, while the AI literacy total score depended on the variables. When analyzing between-subject effects, we first conducted Levene’s test, which confirmed that dependence on the measure met the assumption of homogeneity of variance (p = 0.343 > 0.05). A test of subjects’ effects revealed significant differences in group membership (p = 0.02, partial η2 = 0.054) and sex effects (p = 0.01, partial η2 = 0.074). The joint effect of group and sex was not significant (p = 0.61 > 0.05). As shown in Figure 4, males from both educational settings outperformed their female counterparts, regardless of their group. The effect size of differences can be regarded as small to medium (Cohen et al., 2013).

3.3. AI Literacy Competencies and Between-Group Differences

In order to answer the second research question, we ran a MANCOVA with 14 competencies of AI literacy as the dependent variable, while group membership was independently controlled by sex as the covariate. Box’s test was conducted to determine whether the covariance matrices from the two samples were statistically equivalent and confirmed non-significance (p = 0.18 > 0.05). When we conducted the MANCOVA to examine the effect of group membership on 14 AI literacy competencies, the results showed a non-significant multivariate effect of group membership controlled by sex, and Wilks’ lambda = 0.86, F(14, 129) = 1.41, p = 0.15 > 0.05, indicating that technology teacher education students and secondary technical school students did not differ on the AI literacy scales where multivariant effects were considered. Excluding multivariate effects, differences were found only on the programmability scale (p = 0.045, small effect size η2 = 0.03) and in understanding intelligence (p = 0.002, medium effect size η2 = 0.07). Moreover, males outperformed females on the interdisciplinarity, data literacy, and ethics scales (p = 0.001, η2 = 0.07; p = 0.011, η2 = 0.05; p = 0.018, η2 = 0.04, respectively), where the effect size could be estimated as small to medium (Cohen et al., 2013).
Furthermore, our findings can be supported by the results of the study by Licardo et al. (2025), in which possible differences in students could be attributed to a better understanding of the operating principles behind these technologies among the university students. Both groups of students use AI mainly for task preparation, problem-solving, and data and information searching, while university students also use AI for translation of learning material and proofreading of their textual assignments (Licardo et al., 2025). It seems that university students are more aware of the complexity of critically evaluating information generated by such technologies.

3.4. Critical AI Literacy and Between-Group Differences

Of Hornberger et al.’s (2023) list of competencies, those most directly related to the critical AI competency align with idea #5, the societal impact of the five big ideas framework (Touretzky et al., 2022), which explains that AI can impact society both positively and negatively. The competencies are as follows:
  • AI’s Strengths and Weaknesses: This competency relates to how AI may outperform or fall short of human abilities, thereby shaping real-world outcomes (positive or negative);
  • Human Role in AI: This competency emphasizes designers’ and users’ influence on AI systems, underscoring how human decisions about data, goals, and oversight affect society;
  • Ethics: This competency addresses issues of fairness, bias, privacy, and accountability in AI, all of which are key considerations for societal impact;
  • Interdisciplinarity: This competency can show how AI’s societal effects are examined from multiple angles (e.g., legal, medical, and ethical).
To answer the third research question, we conducted a MANCOVA with four competencies of critical AI literacy as the dependent variable, while group membership was independently controlled by sex as the covariate. Box’s test was conducted on whether the covariance matrices from the two samples are statistically equivalent and confirmed non-significance (p = 0.31 > 0.05). When we conducted a MANCOVA to examine the effect of group membership on four AI literacy competencies, the results showed a non-significant multivariate effect of group membership controlled by sex, where Wilks’ lambda = 0.96, F(4, 139) = 1.19, p = 0.314 > 0.05, indicating that technology teacher education students and secondary technical school students do not differ on the critical AI literacy scales. Moreover, males outperformed females on the Interdisciplinarity and Ethics scales (p = 0.001, η2 = 0.07; p = 0.018, η2 = 0.04, respectively), where the effect size could be estimated as small to medium (Cohen et al., 2013).
To further support critical AI literacy in our study, authors from a very recent study (Licardo et al., 2025) found that in terms of ethical considerations, secondary school students emphasize privacy as their main concern when using AI, while university students consider the spreading of misinformation as the most pressing ethical challenge. In addition, university students may show higher critical awareness regarding the risk of excessive use of AI technologies compared with secondary school students, but the differences cannot be attributed to statistically proven phenomena. Both groups of students, with their knowledge and experience, agree that AI technologies increase efficacy and automation, which results in greater productivity of their tasks and assignments in the specific educational context (Licardo et al., 2025).
According to recent literature published by Velander et al. (2024), Veldhuis et al. (2025), and Yim (2024), no significant differences (p > 0.05) were found between groups based on critical AI literacy, which can be attributed to several overlapping factors that they identify in their analyses of AI literacy frameworks and practices:
  • Both groups of students may have similarly limited formal AI learning opportunities, leading to comparable basic knowledge or awareness. Veldhuis et al. (2025) and Rihtaršič (2018) note that many learners—whether in secondary or higher education—often draw on the same informal resources (e.g., social media or popular science articles) to develop a basic understanding of AI concepts.
  • Critical AI literacy, as conceptualized in both studies, is concerned with overarching skills for questioning and criticizing AI (e.g., recognizing biases, discussing ethical trade-offs). These transversal skills can be acquired as part of general digital literacy or media literacy. Consequently, Velander et al. (2024) report that even individuals with different educational backgrounds can achieve a similar level of critical understanding if they have a common foundation in digital or media literacy.
  • Critical AI literacy is not a static field. New AI tools or controversies (such as large-scale language models, facial recognition, or algorithmic policy decisions) may emerge, and both groups learn about them simultaneously through widely available media. As a result, the knowledge gap or differences between education levels may not be as pronounced as one would expect when a new AI-related topic permeates public discourse and informal learning channels (Yim, 2024).
Moreover, the literature suggests that a combination of shared informal exposures, the broad nature of critical AI competencies, and the limitations of measurement tools often result in no significant group differences when assessing critical AI literacy (Velander et al., 2024; Veldhuis et al., 2025; Yim, 2024). Despite differences in formal educational pathways, learners may converge on a similar overall capacity or show high variability within groups that nullifies the average difference across them (Khuder et al., 2024).
Table 3 succinctly aggregates inferential results aligned with RQ2–RQ4, emphasizing patterns over descriptive detail. Pre-service technology teacher-education students scored higher on the total AI-literacy measure (small–medium effect size), with competency-level advantages concentrated in Understanding Intelligence and Programmability. Multivariate tests showed no overall group effect across all competencies and no between-group differences on the critical-AI composite, indicating that criticality may function as a transversal competence. Finally, small–medium sex effects emerged on the total score and on Interdisciplinarity, Data Literacy, and Ethics, suggesting targeted pedagogical supports may be warranted to mitigate subgroup disparities.
Citations for the values reported in the table: group main effects and total-score means (M = 12.40 vs. 11.97; p = 0.02; η2 = 0.054), and sex effect (p = 0.01; η2 = 0.074); competency-level univariate effects—Understanding Intelligence (p = 0.002; η2 = 0.07) and Programmability (p = 0.045; η2 = 0.03), plus sex differences on Interdisciplinarity, Data literacy, and Ethics (p = 0.001/0.011/0.018; η2 = 0.07/0.05/0.04).

4. Discussion

The discussion section is divided into five main themes, which together frame the interpretation of the results. This section begins with the prevalent themes and intellectual foundations of (critical) AI literacy in technology and engineering education, followed by analyses of students’ AI literacy and competencies in teacher education and secondary technical schools. Particular focus is placed on critical AI literacy before concluding this study with limitations and directions for future work.

4.1. Dominant Themes and Intellectual Bases of (Critical) AI Literacy in Technology and Engineering Education

Across 2017–2025, the thematic evolution map (Figure 2) and the strategic maps in Appendix B (Figure A1, Figure A2, Figure A3 and Figure A4) show a decisive shift from broad, conceptual treatments of “AI in education” toward application-rich, methodologically stronger work anchored in GenAI (LLMs/ChatGPT), AI/critical AI literacy, and human–AI interaction/governance (Tzirides et al., 2024). Beginning in 2023, LLMs emerge abruptly and stabilize as a basic/foundational theme, while “chatbots” peak as a motor theme in 2023–2024 before subsuming under an LLM-centric umbrella. By 2025, AI literacy and critical AI literacy consolidate in the basic quadrant, reflecting curricular embedding and growth in instrument development/validation; meanwhile, the broader “digital literacy” strand drifts toward a niche, as discourse narrows to AI-specific literacies and policy frameworks. These migrations coincide with a maturation of methods—growth in systematic/scoping reviews and structural equation modeling—and the rise in well-defined application arenas (notably K–12 and medical/health professions education). Together, the maps portray a field converging on core literacies and LLM-mediated practices while negotiating ethics, integrity, privacy, and bias as durable concerns.
The co-occurrence network (Figure 2) clarifies this organization into four Walktrap communities: a red hub around AI and ethics/critical AI literacy (linking to LLMs, pre-service teachers, assessment, and responsible/human-centered AI); a purple community on pedagogical application across educational stages (teacher education, K–12, assessment, and academic writing/feedback); a blue cluster emphasizing medical/health education (ML/DL, curricula, surveys, and systematic reviews); and a green cluster on technology acceptance/attitudes. The position and size of nodes underscore the centrality of ethics/criticality and the practical salience of pre-service teacher contexts, which anticipates our cohort comparison.
Our measurement strategy aligns with the following cartographies: the coverage is robust for AI4K12 BI2–BI3 (representation/reasoning; learning) and substantial for BI5 (societal impact), but minimal for BI4 (natural interaction) and, by design, absent for BI1 (perception/robotics), as already suggested by Licardo et al. (2025). This explains why competencies linked to human–AI interaction/robotics, visible in the maps as emerging or niche, may be underrepresented in our detected differences. Therefore, we interpret several “non-differences” cautiously as measurement coverage effects, not just cohort equivalence.
Critical AI literacy already appears as a very important topic in the first period; however, the development of the topic is relatively low. In periods 2 and 3 of 2023/24, we did not specifically follow it, but it reappeared in the last period, where its importance increased, while the level of development remained low. This finding further supports and justifies the relevance of our study, as Velander et al. (2023) and Veldhuis et al. (2025) point out that critical AI can be crucial for the successful implementation of AI in education (see strategic maps).
Recently, particular attention has been paid to thinking skills and metacognition, especially to promote personalized learning with the help of AI (Kim et al., 2025; Velander et al., 2023; Yim, 2024). Critical literacy is very close to the main cluster (AI) on the network, indicating its additional importance with the centroids of ethics, ChatGPT, AI literacy, and education. Critical AI literacy is directly linked to digital literacy, prompt engineering, teaching, and learning, which are still in the main cluster and the entire core community of clusters.

4.2. AI Literacy in Students in a Technology Teacher Education Program and Secondary Technical School Students

Since our study was the first of its kind to examine the AI literacy of high school students and compare them with students studying education, direct studies in the literature are rare or incomplete (Licardo et al., 2025). Therefore, we present a discussion and answers to research questions based solely on our results obtained from the AI literacy test, and we also discuss them from the perspective of recent empirical quantitative–qualitative research and the results of the state of GPT in education from kindergarten to university studies in Slovenia (Licardo et al., 2025).
Consistent with Figure 2’s emphasis on pre-service teachers and higher education settings, our ANOVA indicates a small-to-moderate overall advantage for technology teacher education students in total AI literacy scores (M = 12.40 vs. 11.97), with group membership significant (p = 0.02, partial η2 = 0.054) and sex effects significant (males > females; p = 0.01, partial η2 = 0.074), but with no interaction effects (p = 0.61). Figure 4 visualizes these estimated marginal means. Benchmarked against multinational evidence using IRT scores, our teacher education cohort aligns closely with recent UK (Hornberger et al., 2025) and Slovenian (Licardo et al., 2025) results, and below STEM-intensive cohorts, highlighting heterogeneous curricular provision and the diffusion of AI-related knowledge Via informal channels.
Our students demonstrate a comparable level to US students, a lower level than UK students, and a significantly lower level than students in Germany, as reported in the Hornberger et al. study (2025). These results are to be expected, as unlike German students, our students have not received any training in AI literacy (Hornberger et al., 2025; Licardo et al., 2025). Moreover, in addition to not having organized formal education, students at Slovenian universities also lack faculty licenses and hardware, and have insufficiently trained staff, which is why students are largely dependent on their own initiative to educate themselves through various social networks, self-study, etc., as reported by an extensive study by Licardo et al. (2025). Similarly, this also applies to high school students, as stated by Licardo et al. (2025), who in some respects even exceed their teachers in terms of learning engagement in the field of AI. AI is mostly used for various teaching preparations, automation, and acceleration of technical and administrative procedures in learning (preparation of seminars, teaching materials, etc.). They use AI mainly for various learning preparations, automation, and acceleration of technical and administrative procedures in learning (preparing seminars, worksheets, searching for sources, mathematical calculations, etc.), which can lead to additional biases and misunderstandings of the subject matter, and consequently to insufficient knowledge (Darvishi et al., 2024). Furthermore, Slovenian high school students perceive the issue of ethical use of AI at a lower level than university students, and as a result, the use of AI leads to undesirable or inappropriate prompts, unlike university students (Licardo et al., 2025).
Pedagogically, this pattern suggests both cohorts have meaningful exposure to AI tools yet lack systematic, assessed coursework commensurate with the basic or motor themes in the literature (LLMs, ethics/governance, and assessment). Therefore, we recommend course-embedded AI modules in both programs—e.g., dataset audits, model critique, explainability exercises, and integrity/privacy casework—aligned with the consolidated themes in Figure 2 and the strategic maps (Appendix B). Such alignment should convert modest mean gaps into substantive gains on assessed learning outcomes rather than opportunistic tool use.

4.3. AI Literacy Competencies in Students in a Technology Teacher Education Program and Secondary Technical School Students

At the competency level, our MANCOVA found no significant multivariate group effect (Wilks’ λ = 0.86, F(14,129) = 1.41, p = 0.15), but two univariate advantages were found for teacher education students: understanding intelligence (p = 0.002, η2 = 0.07) and programmability (p = 0.045, η2 = 0.03). These map to our Foundational Knowledge and Participatory Design pillars and resonate with Figure 2’s red community, where foundational concepts co-locate with practical uptake of generative tools in pre-service contexts. Sex-linked differences favored males on interdisciplinarity, data literacy, and ethics scales (p = 0.001, η2 = 0.07; p = 0.011, η2 = 0.05; p = 0.018, η2 = 0.04). We justify the (non)differences in accordance with the results of Licardo et al. (2025).
The selective between-group advantages align with the basic/motor themes that dominate 2023–2025 (Figure 2), suggesting that even limited, structured exposure in teacher education can produce targeted gains, while broader equivalence reflects shared informal learning. The gender gaps arise precisely in domains the maps treat as central (ethics/criticality; data practices), indicating a need for equitable participation structures and assessment transparency (e.g., rotating roles in data audits and ethics cases, criterion-referenced rubrics for critique tasks). Given our instrument’s limited BI4/absent BI1 coverage, we anticipate that additional divergence could surface if human–AI interaction and robotics/perception tasks were added to the competency blueprint (T. K. F. Chiu & Sanusi, 2024).
Teacher education programs should extend “programmability” toward human-centered AI design studios (prompt engineering + error analysis + stakeholder critique), while secondary technical tracks should strengthen conceptual models of intelligence and hands-on data literacy across the ML pipeline (T. K. F. Chiu & Sanusi, 2024). Both should integrate performance-based assessments that mirror the field’s motor/basic motifs and provide objective measures as suggested by T. Chiu et al. (2024).

4.4. Critical AI Literacy in Students in a Technology Teacher Education Program and Secondary Technical School Students

Restricting the construct to four critical AI literacy competencies (AI’s strengths/weaknesses, human role, ethics, and interdisciplinarity), the MANCOVA yielded no reliable group difference (Wilks’ λ = 0.96, F(4,139) = 1.19, p = 0.314), although males again outperformed females on the interdisciplinarity and ethics scales (p = 0.001, η2 = 0.07; p = 0.018, η2 = 0.04). We interpret the null group effect as the convergence of two forces visible in Figure 1 and Figure 2 and Appendix B: (1) the centrality/diffusion of ethics/criticality in the literature and public discourse, which promotes informal learning that narrows cohort gaps; (2) instrument emphasis on broad societal/ethical reasoning (AI4K12 BI5) with limited depth on policy/governance mechanics and no robotics/perception, potentially compressing variance. These interpretations are consistent with contemporaneous survey evidence conducted by Hornberger et al. (2025) (e.g., privacy vs. misinformation concerns across school levels; higher awareness of over-reliance risks among university students), even when such differences do not rise to statistical significance in our test battery.
In parallel, a recent study by Licardo et al. (2025) confirms that both high school and college students recognize authenticity of content and privacy protection as the most important ethical challenges in evaluating the usefulness of AI. Among high school students, privacy and false information rank highest, whereas plagiarism and anthropomorphizing are considered less critical; among university students, false information and privacy are rated first, while plagiarism holds medium importance (Licardo et al., 2025). Importantly, access to knowledge about responsible AI use remains uneven: high school students often rely on online advertisements, social networks, and peers, which reduces verifiability and encourages uncritical use, while university students draw mostly on individual subjects and peers, though without a systematic understanding of technological limitations (Licardo et al., 2025). On this basis, the monograph proposes measures that prioritize human judgment, namely the integration of clear ethical and legal guidelines into curricula, the strengthening of critical thinking, the verification of sources, and the development of digital literacy as a foundation for the safe and responsible use of AI in secondary and higher education (Licardo et al., 2025).
To turn widespread awareness into assessed competence, both cohorts need explicitly scaffolded and graded critical AI learning aligned to the maps’ dominant motifs: (1) algorithmic accountability labs (bias testing, harm modeling, and mitigation trade-offs), which were also proposed by Rismani et al. (2024); (2) policy/governance clinics (privacy, academic integrity, and auditability), which were also proposed by Miao and Holmes (2023); and (3) human–AI orchestration protocols (division of labor, oversight thresholds, and error escalation), which were also proposed by the NIST (National Institute of Standards and Technology, 2023). Embedding these into course sequences and practica should raise measured performance on the very constructs that define the field’s present phase (T. Chiu et al., 2024; Ng et al., 2024).
Here we explicitly tie the four-pillar framework to our data: Evidence from Phase 2 aligns most strongly with Foundational Knowledge, where pre-service technology teachers outperformed secondary technical students on the total AI-literacy score and, more pointedly, on cognitively oriented competencies—understanding intelligence and programmability—thereby validating the pillar’s emphasis on algorithmic and representational reasoning (RQ2–RQ3).
By contrast, Critical Appraisal appears more transversal: the absence of between-group differences in our critical AI composites suggests that ethical/impact understandings diffuse through shared informal exposure and public discourse (RQ4), refining our model to treat criticality as a cross-cutting competency rather than one tightly coupled to program level.
With respect to Participatory Design, the framework is challenged by instrument coverage limits (e.g., minimal design/creation and governance depth), indicating a measurement–curriculum gap and motivating future assessment development aligned to this pillar.
Finally, Pedagogical Integration is underscored by small-to-moderate sex effects (e.g., male advantages on interdisciplinarity and ethics), which our framework interprets as targets for programmatic supports (assessment, mentoring, and inclusive project roles) to ensure equitable outcomes—thereby operationalizing this pillar as a lever for closing observed disparities.
Enhanced AI literacy among preservice teachers strengthens their ability to discern where AI genuinely adds pedagogical value, leading to more principled selection of tools and clearer boundaries between AI and adjacent technologies. It improves assessment design and feedback practices by enabling candidates to interpret data outputs, check model limitations, and translate insights into equitable instructional decisions. Greater literacy also supports inclusive teaching: candidates are better prepared to anticipate bias, safeguard privacy, and scaffold AI-supported tasks for diverse learners. In practicum settings, AI-competent candidates can streamline planning (e.g., resource forecasting, formative analytics) and redirect saved time toward high-impact interactions. Programs benefit institutionally through richer, interdisciplinary coursework that aligns with contemporary school needs and fosters partnerships with districts around responsible innovation. Finally, graduates enter the workforce as reflective adopters—capable of explaining AI to students and families—thereby raising school capacity while mitigating hype-driven or unsafe use, what confirms the findings of Gabrovšek and Rihtaršič (2025).
To address observed gender differences in AI–interdisciplinary competence among preservice teachers, we organized targeted interventions around three documented gap generators. The table summarizes objectives and high-leverage interventions for teacher education program implementation (Table 4).

4.5. Limitations of the Study and Future Work

Although our study yielded insightful findings, we can still report some limitations that might improve this study in the future. Our first objective was to conduct a bibliometric analysis to identify dominant themes and key challenges in the field of research; however, when conducting the analysis, we used only WoS data. The inclusion of Scopus data and data from other databases may enlarge our knowledge of the phenomena we studied. Our bibliometric insights are constrained by database coverage and language bias (e.g., English-language and WoS indexing), which may underrepresent regional or practitioner literature. Results are sensitive to search strings, inclusion thresholds, and clustering parameters, so alternative specifications could yield different structures. Citation-based indicators favor established topics and are affected by indexing time lags, limiting sensitivity to very recent LLM-era work. Consequently, generalizability is bounded to the sampled corpus, time window, and operational choices.
Next, the present study was only carried out on a sample of students from one larger secondary school and one faculty of education, which may limit its representativeness. The male population in the secondary school was predominant, while female students were overrepresented in the Faculty of Education. This uneven sex distribution within the sample may affect the results and make it difficult to generalize to the wider population. Thus, for future research, we also plan to test AI literacy in a general upper secondary school, a human service technical secondary school, and the inclusion of the second faculty of education, which educates pre-service technology teachers.
Our empirical sample is modest and institutionally bounded (n = 145) to Slovenian contexts—secondary technical schools and a single university—limiting representativeness beyond this geographic and programmatic setting; uneven sex composition further constrains inference to broader populations. Regarding test generalizability, although the AI literacy instrument shows acceptable reliability and broad coverage, its blueprint omits BI1 (perception/robotics) and has minimal BI4 (human–AI interaction), which may understate competencies emphasized in robotics-heavy or interaction-centric curricula (Gabrovšek & Rihtaršič, 2025). Consequently, findings should be extrapolated with caution to other countries, school types, and AI curricula that weigh these domains differently.
Although the test provided a good insight into the current state of AI literacy, there is an opportunity to include additional self-reported instruments to obtain an even more comprehensive picture. Therefore, for future research, we plan to add parallel questionnaires (e.g., on self-efficacy, self-regulation, ethical literacy, etc.) and to conduct a focus group. This would capture additional variables and deepen the understanding of factors that may not be adequately captured by the baseline questionnaire.

5. Conclusions

Our bibliometric and science mapping analyses show a field that is undergoing explosive growth and rapid consolidation: a very young, interdisciplinary, and highly collaborative literature set (over 1200 articles in almost 600 locations) has reorganized itself around GenAI, especially LLMs/ChatGPT, since 2023, rapidly evolving from peripheral to basic/foundational themes alongside AI/critical AI literacy and ethics. “Chatbots” have emerged as a short-lived engine topic before being absorbed into LLM-centric practice; application domains (K-12 and medicine/health professions) and human-centered topics (integrity, privacy, bias, and fear) gain centrality as methodologies mature (systematic/oversight reviews, SEM). AI literacy and critical AI literacy now act as foundational curricular anchor points alongside ethics, governance, and human–AI collaboration. This shift is reflected in parallel developments across higher education: calls for institutional AI policies that link pedagogy, governance, and operations, continued emphasis on data ethics as a transversal skill for critical inquiry, and sectoral syntheses in the health professions and engineering that highlight integration opportunities and integrity risks posed by GenAI.
The co-occurrence network confirms that ethics/criticism is an important node linking foundational constructs to pedagogical implementation and acceptance. Overall, the intellectual basis of the field now privileges LLM-mediated teaching/learning, control, and human–AI orchestration, while niche fronts (e.g., prompt engineering, collaboration protocols) signal short-term curricular relevance. Methodological maturation (scale development, SEM, and design-based approaches) is becoming increasingly visible and can be utilized for curriculum design and assessment. Therefore, curriculum design should focus on LLM use with accountability, standardized skills assessment, and explicitly link foundational topics (AI/critical literacy, ethics) to authentic tasks in the classroom.
Consistent with leading pedagogical guidelines, future-ready AI literacy must go beyond tool familiarization to promote student–AI collaboration, error analysis, and ethical decision-making in authentic tasks at the K–12 and higher education levels.
Our two-cohort comparison shows a small advantage for student teachers in GenAI literacy focused on understanding intelligence and programmability, while multivariate differences in critical AI literacy are not statistically reliable; males outperform females in interdisciplinarity, ethics, and data literacy, which is consistent with regional evidence of persistent gender readiness gaps. Critical AI competency, operationalized as strengths/weaknesses, human role, ethics, and interdisciplinarity, shows no cohort differences, with male advantages persisting in interdisciplinarity and ethics. We interpret these patterns as convergence driven by shared informal experiences and limited formal provision, reinforced by instrument coverage (robust BI2 (representation)–BI3 (learning), partial BI5 (societal impact); minimal BI4 (human–AI interaction); and no BI1 (perception)).
Together with recent studies showing that reliance and dependence on GenAI can affect learning outcomes, sometimes even negatively, these findings emphasize the need to replace incidental experiences with structured, performance-based learning that combines technical practice with governance and integrity safeguards. Therefore, we recommend targeted modules in both cohorts on dataset curation and audits, model critique, bias/harm analysis, and human–AI workflow design, supported by constructivist datasets and validated assessment tools, and embedded in institutional frameworks that ensure transparency, accountability, and equity in the use of GenAI. Such an orientation directly addresses the prevalent issues in the field and can build an equitable, critical AI literacy in secondary and higher education.
Therefore, targeted curricula should (i) deepen teachers’ knowledge that agents are programmable in human-centered design studios (prompting, error analysis, explainability, and supervision); (ii) strengthen secondary students’ conceptual models of intelligence and end-to-end data/ML practices; and (iii) link assessed critical AI work (bias testing, privacy/integrity cases, governance clinics, and collaboration protocols) with equity strategies to close gender gaps. Future research should extend measurement to BI1 (perception)/BI4 (human–AI interaction) and use longitudinal, performance-based assessments to track learning gains.
To enhance the practical contribution of this study, we conceptualize the proposed four-pillar framework as a policy scaffold that can be implemented across educational governance tiers, from the institutional (school) to the ministerial (national) level: at the institutional level, adopt pillar-aligned course maps with staged learning outcomes (Foundational Knowledge → Critical Appraisal → Participatory Design → Pedagogical Integration), embed performance-based assessments (dataset audits, bias tests, error-analysis portfolios), and fund targeted teacher personal development on LLM-mediated pedagogy and assessment integrity; at the national level, issue AI literacy standards that (i) sequence competencies across K-12 and teacher education, (ii) mandate governance/ethics coverage (privacy, academic integrity, transparency) alongside data practices, and (iii) require monitoring of equity indicators (e.g., gender gaps in interdisciplinarity/ethics) with corrective supports (mentoring, inclusive project roles, accessible toolkits). Practically, ministries and institutions can start with a 12–18-month rollout: map existing syllabi to the pillars; add two assessed tasks per course (one data-pipeline audit, one human–AI co-design brief); provide short micro-credentials for staff; and align school quality assurance cycles to track AI literacy outcomes and integrity incidents. This policy-to-practice chain ties the paper’s empirical gaps (e.g., uneven competencies, gender effects) to concrete curricular levers, ensuring coherent, measurable AI literacy progression across the system.
Our two-phase study shows a rapidly consolidating literature around GenAI/LLMs, AI/critical AI literacy, and ethics, which we operationalized in our assessment and cohort comparison. Empirically, technology teacher-education students scored modestly higher on total AI literacy than secondary technical students (group p = 0.02) with small–medium sex effects, and between-group differences concentrated in understanding intelligence and programmability; by contrast, no significant group differences emerged on critical-AI composites.
Looking ahead, future research should broaden the sample beyond a single university and one secondary-school site, incorporate additional databases beyond WoS, and extend measurement to under-represented areas (e.g., BI1 perception/robotics; richer BI4 interaction/participatory design).
Methodologically, mixed-method designs (parallel questionnaires on self-efficacy/regulation/ethics and focus groups) and longitudinal/intervention studies can test how targeted curricular changes shift specific competencies and close observed subgroup gaps.

Author Contributions

Conceptualization, S.A. and D.R.; methodology, S.A. and D.R.; validation, S.A. and D.R.; formal analysis, S.A. and D.R.; investigation, S.A. and D.R.; resources, S.A. and D.R.; data curation, S.A.; writing—original draft preparation, S.A. and D.R.; writing—review and editing, S.A. and D.R.; visualization, S.A. and D.R.; supervision, S.A.; project administration, D.R.; funding acquisition, S.A. All authors have read and agreed to the published version of the manuscript.

Funding

The authors acknowledge the financial support of the Slovenian Research Agency under the project “Developing the Twenty-First-Century Skills Needed for Sustainable Development and Quality Education in the Era of Rapid Technology-Enhanced Changes In The Economic, Social, and Natural Environments (grant no. J5-4573)” and the research core funding “Strategies for Education for Sustainable Development applying Innovative Student-Centered Educational Approaches (ID: P5-0451)” also funded by the Slovenian Research Agency. The authors would also like to thank to financial support funded by the Republic of Slovenia, the Ministry of Higher Education, Science and Innovation, and the European Union—NextGenerationEU (ID—NRP: 3350-24-3502). The study was also funded by the JR MLADI 2025–2026—Public Call for Funding of Educational Programs for Children and Youth to Strengthen Digital Competences and Promote Science and Technology Careers (Ref. No. 430-20/2024-3150), financed by the Ministry of Digital Transformation of the Republic of Slovenia.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and was reviewed and approved by the Ethics Commission of the Faculty of Education of the University of Ljubljana (Approval code: 7/2025; approval date: 10 March 2025).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The data used in this study are available on request from the corresponding author. All data have been anonymized but are not publicly available because of the privacy issues related to the qualitative nature of it.

Acknowledgments

The authors thank the participating students. Declaration of GenAI and AI-assisted technologies in the writing process: While preparing this work, the authors used DeepL Translator and ChatGPT 5 (OpenAI, 2025) software to correct, proofread, and embellish a language that is not their native language. After using these tools, the authors reviewed and edited the content as needed and take full responsibility for the content of the publication.

Conflicts of Interest

The authors declare no conflicts of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Appendix A

Table A1. Search strategy for locating literature on AI/LLM-related literacy, critical/ethical/societal dimensions, and educational/teacher-training contexts.
Table A1. Search strategy for locating literature on AI/LLM-related literacy, critical/ethical/societal dimensions, and educational/teacher-training contexts.
ComponentDescriptionQuery (TS = …)
AI conceptsTerms for artificial intelligence and related generative models("artificial intelligence" OR AI OR "generative AI" OR "large language model*" OR LLM*) NEAR/3 (literac* OR competenc* OR fluenc* OR knowledge OR skill* OR "algorithmic literacy" OR "data literacy" OR "critical data literacy")
Critical/ethical/societal focusTerms capturing critical, ethical, societal and governance aspects(critical OR ethic* OR societal OR civic OR governance OR privacy OR fairness OR transparency OR accountability OR "AI literacy" OR "AI’s Strengths and Weaknesses" OR "Human Role" OR Interdisciplinarity OR "system* thinking")
Educational contextTerms for education, teacher training, technical/vocational settings, and students(educat* OR "technology education" OR "technical education" OR engineering OR TVET OR "technical high school" OR "secondary technical school" OR "teacher education" OR preservice OR "pre-service" OR "teacher training" OR "technology teacher" OR "engineering teacher*" OR student*)
Combined algorithmFull search combining A, B, and C (all must be present)TS = (("artificial intelligence" OR AI OR "generative AI" OR "large language model*" OR LLM*) NEAR/3 (literac* OR competenc* OR fluenc* OR knowledge OR skill* OR "algorithmic literacy" OR "data literacy" OR "critical data literacy")) AND TS = (critical OR ethic* OR societal OR civic OR governance OR privacy OR fairness OR transparency OR accountability OR "AI literacy" OR "AI’s Strengths and Weaknesses" OR "Human Role" OR Interdisciplinarity OR "system* thinking") AND TS = (educat* OR "technology education" OR "technical education" OR engineering OR TVET OR "technical high school" OR "secondary technical school" OR "teacher education" OR preservice OR "pre-service" OR "teacher training" OR "technology teacher" OR "engineering teacher*" OR student*)

Appendix B

Figure A1. Strategic map 1.
Figure A1. Strategic map 1.
Education 15 01455 g0a1
Figure A2. Strategic map 2.
Figure A2. Strategic map 2.
Education 15 01455 g0a2
Figure A3. Strategic map 3.
Figure A3. Strategic map 3.
Education 15 01455 g0a3
Figure A4. Strategic map 4.
Figure A4. Strategic map 4.
Education 15 01455 g0a4

References

  1. Ahmad, S., Han, H., Alam, M. M., Rehmat, M. K., Irshad, M., Arraño-Muñoz, M., & Ariza-Montes, A. (2023). Impact of artificial intelligence on human loss in decision-making, laziness, and safety in education. Humanities and Social Sciences Communications, 10, 1–14. [Google Scholar] [CrossRef]
  2. Aria, M., & Cuccurullo, C. (2017). bibliometrix: An R-tool for comprehensive science mapping analysis. Journal of Informetrics, 11(4), 959–975. [Google Scholar] [CrossRef]
  3. Atenas, J., Havemann, L., & Timmermann, C. (2023). Reframing data ethics in research methods education: A pathway to critical data literacy. International Journal of Educational Technology in Higher Education, 20, 11. [Google Scholar] [CrossRef]
  4. Bettayeb, A. M., Talib, M. A., Altayasinah, A. Z. S., & Dakalbab, F. (2024). Exploring the impact of ChatGPT: Conversational AI in education. In Frontiers in education. Frontiers Media SA. [Google Scholar] [CrossRef]
  5. Bing, Z. J., & Leong, W. Y. (2025). Ethical design of AI for education and learning systems. ASM Science Journal, 20, 1–9. [Google Scholar] [CrossRef]
  6. Boscardin, C., Gin, B. C., Golde, P. B., & Hauer, K. (2023). ChatGPT and generative artificial intelligence for medical education: Potential impact and opportunity. Academic Medicine, 99, 22–27. [Google Scholar] [CrossRef] [PubMed]
  7. Cazzaniga, M., Jaumotte, F., Li, L., Melina, G., Panton, A., Pizzinelli, C., Rockall, E., & Tavares, M. (2024). Gen-AI: Artificial intelligence and the future of work (IMF Staff Discussion Note SDN/2024/001). International Monetary Fund. [Google Scholar] [CrossRef]
  8. Chee, H., Ahn, S., & Lee, J. (2024). A competency framework for AI literacy: Variations by different learner groups and an implied learning pathway. British Journal of Educational Technology, 56(5), 2146–2182. [Google Scholar] [CrossRef]
  9. Chen, L., Chen, P., & Lin, Z. (2020). Artificial intelligence in education: A review. IEEE Access, 8, 75264–75278. [Google Scholar] [CrossRef]
  10. Chiu, T., Chen, Y., Yau, K., Chai, C., Meng, H., King, I., Wong, S., & Yam, Y. (2024). Developing and validating measures for AI literacy tests: From self-reported to objective measures. Computers and Education: Artificial Intelligence, 7, 100282. [Google Scholar] [CrossRef]
  11. Chiu, T. K. F., & Sanusi, I. T. (2024). Define, foster, and assess student and teacher AI literacy and competency for all: Current status and future research direction. Computers and Education: Open, 7, 100189. [Google Scholar] [CrossRef]
  12. Cohen, J., Cohen, P., West, S. G., & Aiken, L. S. (2013). Applied multiple regression/correlation analysis for the behavioral sciences (3rd ed.). Routledge. [Google Scholar] [CrossRef]
  13. Darvishi, A., Khosravi, H., Sadiq, S., Gašević, D., & Siemens, G. (2024). Impact of AI assistance on student agency. Computers and Education, 210, 104967. [Google Scholar] [CrossRef]
  14. Donthu, N., Kumar, S., Mukherjee, D., Pandey, N., & Lim, W. M. (2021). How to conduct a bibliometric analysis: An overview and guidelines. Journal of Business Research, 133, 285–296. [Google Scholar] [CrossRef]
  15. Eysenbach, G. (2023). The role of ChatGPT, generative language models, and artificial intelligence in medical education: A conversation with ChatGPT and a call for papers. JMIR Medical Education, 9(1), e46885. [Google Scholar] [CrossRef] [PubMed]
  16. Faculty of Education, University of Ljubljana. (n.d.). Giving word to knowledge. Available online: https://www.pef.uni-lj.si/ (accessed on 16 September 2025).
  17. Filgueiras, F. (2023). Artificial intelligence and education governance. Education, Citizenship and Social Justice, 19, 349–361. [Google Scholar] [CrossRef]
  18. Gabrovšek, R., & Rihtaršič, D. (2025). Custom Generative Artificial Intelligence Tutors in Action: An Experimental Evaluation of Prompt Strategies in STEM Education. Sustainability, 17(21), 9508. [Google Scholar] [CrossRef]
  19. Gartner, S., & Krašna, M. (2023). Ethics of artificial intelligence in education. Journal of Elementary Education, 16(2), 221–235. [Google Scholar] [CrossRef]
  20. Guzik, A., Tomczak, M. T., & Gawrycka, M. (2024). What is the future of digital education in the higher education sector? An overview of trends with example applications at Gdańsk Tech, Poland. Global Journal of Engineering Education, 26, 95–100. Available online: https://mostwiedzy.pl/pl/publication/download/1/what-is-the-future-of-digital-education-in-the-higher-education-sector-an-overview-of-trends-with-ex_93670.pdf (accessed on 15 September 2025).
  21. Hale, J. D., Alexander, S., Wright, S. T., & Gilliland, K. (2024). Generative AI in undergraduate medical education: A rapid review. Journal of Medical Education and Curricular Development, 11, 23821205241266697. [Google Scholar] [CrossRef]
  22. Haroud, S., & Saqri, N. (2025). Generative AI in higher education: Teachers’ and students’ perspectives on support, replacement, and digital literacy. Education Sciences, 15(4), 396. [Google Scholar] [CrossRef]
  23. Hornberger, M., Bewersdorff, A., & Nerdel, C. (2023). What do university students know about artificial intelligence? Development and validation of an AI literacy test. Computers and Education: Artificial Intelligence, 5, 100151. [Google Scholar] [CrossRef]
  24. Hornberger, M., Bewersdorff, A., Schiff, D. S., & Nerdel, C. (2025). A multinational assessment of AI literacy among university students in Germany, the UK, and the US. Computers and Human Behavior: Artificial Humans, 4, 100167. [Google Scholar] [CrossRef]
  25. Hwang, G., Tang, K., & Tu, Y. (2022). How artificial intelligence (AI) supports nursing education: Profiling the roles, applications, and trends of AI in nursing education research (1993–2020). Interactive Learning Environments, 32, 373–392. [Google Scholar] [CrossRef]
  26. Jensen, L. X., Buhl, A., Sharma, A., & Bearman, M. (2024). Generative AI and higher education: A review of claims from the first months of ChatGPT. Higher Education, 89(4), 1145–1161. [Google Scholar] [CrossRef]
  27. Ji, H., Han, I., & Ko, Y. (2022). A systematic review of conversational AI in language education: Focusing on the collaboration with human teachers. Journal of Research on Technology in Education, 55, 48–63. [Google Scholar] [CrossRef]
  28. Johri, A. (2020). Artificial intelligence and engineering education. Journal of Engineering Education, 109, 20326. [Google Scholar] [CrossRef]
  29. Joseph, G. V., Athira, P., Thomas, M. A., Jose, D., Roy, T. V., & Prasad, M. (2024). Impact of digital literacy, use of AI tools and peer collaboration on AI-assisted learning: Perceptions of the university students. Digital Education Review, 45, 43–49. [Google Scholar] [CrossRef]
  30. Khuder, B., Ou, W., Franzetti, S., & Negretti, R. (2024). Conceptualising and cultivating critical GAI literacy in doctoral academic writing. Journal of Second Language Writing, 66, 100987. [Google Scholar] [CrossRef]
  31. Kim, S., Kim, T., & Kim, K. (2025). Development and effectiveness verification of AI education data sets based on constructivist learning principles for enhancing AI literacy. Scientific Reports, 15, 10725. [Google Scholar] [CrossRef]
  32. Lee, I. A., Ali, S., Zhang, H., DiPaola, D., & Breazeal, C. (2021, March 13–20). Developing middle school students’ AI literacy. 52nd ACM technical symposium on computer science education (pp. 191–197), Virtual Event. [Google Scholar] [CrossRef]
  33. Lee, J., Wu, A. S., Li, D., & Kulasegaram, K. (2021). Artificial intelligence in undergraduate medical education: A scoping review. Academic Medicine, 96, S62–S70. [Google Scholar] [CrossRef] [PubMed]
  34. Licardo, M., Kranjec, E., Lipovec, A., Dolenc, K., Arcet, B., Flogie, A., Plavčak, D., Ivanuš Grmek, M., Bednjički Rošer, B., Sraka Petek, B., & Laure, M. (2025). Generativna umetna inteligenca v izobraževanju: Analiza stanja v primarnem, sekundarnem in terciarnem izobraževanju. Univerzitetna založba Univerze v Mariboru. Available online: https://press.um.si/index.php/ump/catalog/view/950/1409/5110 (accessed on 15 September 2025).
  35. Lintner, T. (2024). A systematic review of AI literacy scales. NPJ Science of Learning, 9, 50. [Google Scholar] [CrossRef]
  36. Long, D., & Magerko, B. (2020). What is AI literacy? Competencies and design considerations. In proceedings of the 2020 CHI conference on human factors in computing systems (pp. 1–16). ACM. [Google Scholar] [CrossRef]
  37. Lumanlan, J. S., Ayson, M. R. I., Bautista, M. J. V., Dizon, J. C., & Ylarde, C. M. L. (2025). AI literacy among pre-service teachers: Inputs towards a relevant teacher education curriculum. International Journal of Multidisciplinary: Education and Research Innovation, 3(1), 242–250. Available online: https://philarchive.org/go.pl?id=LUMALA&proxyId=&u=https%3A%2F%2Fphilpapers.org%2Farchive%2FLUMALA.pdf (accessed on 16 September 2025).
  38. Miao, F., & Holmes, W. (2023). Guidance for generative AI in education and research. UNESCO. [Google Scholar] [CrossRef]
  39. National Institute of Standards and Technology. (2023). Artificial intelligence risk management framework (AI RMF 1.0) (NIST AI 600-1). U.S. Department of Commerce. [CrossRef]
  40. Ng, D. T. K., Leung, J. K. L., Chu, K. W. S., & Qiao, M. S. (2021). AI literacy: Definition, teaching, evaluation and ethical issues. Proceedings of the Association for Information Science and Technology, 58, 504–509. [Google Scholar] [CrossRef]
  41. Ng, D. T. K., Luo, W., Chan, H., & Chu, S. K. W. (2022). Using digital story writing as a pedagogy to develop AI literacy among primary students. Computers and Education: Artificial Intelligence, 3, 100054. [Google Scholar] [CrossRef]
  42. Ng, D. T. K., Wu, W., Leung, J. K. L., Chiu, T. K. F., & Chu, S. K. W. (2024). Design and validation of the AI literacy questionnaire: The affective, behavioural, cognitive, and ethical approach. British Journal of Educational Technology, 55, 1082–1104. [Google Scholar] [CrossRef]
  43. Pizzinelli, C., Panton, A., Mendes Tavares, M., Cazzaniga, M., & Li, L. (2023). Labor market exposure to AI: Cross-country differences and distributional implications (Working paper No. 2023/216). International Monetary Fund. Available online: https://www.imf.org/-/media/Files/Publications/WP/2023/English/wpiea2023216-print-pdf.ashx (accessed on 17 September 2025).
  44. Relmasira, S., Lai, Y. C., & Donaldson, J. (2023). Fostering AI literacy in elementary science, technology, engineering, art, and mathematics (STEAM) education in the age of generative AI. Sustainability, 15(18), 13595. [Google Scholar] [CrossRef]
  45. Rihtaršič, D. (2018). Using an Arduino-based low-cost DAQ in science teacher training. World Transactions on Engineering and Technology Education, 16(4), 380–385. Available online: http://www.wiete.com.au/journals/WTE&TE/Pages/Vol.16,%20No.4%20(2018)/10-Rihtarsic-D.pdf (accessed on 15 September 2025).
  46. Rincón, E. H. H., Jiménez, D., Aguilar, L. A. C., Flórez, J. M. P., Tapia, Á. E. R., & Peñuela, C. L. J. (2025). Mapping the use of artificial intelligence in medical education: A scoping review. BMC Medical Education, 25, 526. [Google Scholar] [CrossRef] [PubMed]
  47. Rismani, S., Dobbe, R., & Moon, A. (2024). From silos to systems: Process-oriented hazard analysis for AI systems. arXiv, arXiv:2410.22526. [Google Scholar] [CrossRef]
  48. Rupnik, D., & Avsec, S. (2025). Student agency as an enabler in cultivating sustainable competencies for people-oriented technical professions. Education Sciences, 15, 469. [Google Scholar] [CrossRef]
  49. Rütti-Joy, O., Winder, G., & Biedermann, H. (2023). Building AI literacy for sustainable teacher education. Zeitschrift für Hochschulentwicklung, 18(4), 175–189. [Google Scholar] [CrossRef]
  50. Salhab, R. (2024). AI literacy across curriculum design: Investigating college instructor’s perspectives. Online Learning, 28(2), 4426. [Google Scholar] [CrossRef]
  51. Schleiss, J., Laupichler, M. C., Raupach, T., & Stober, S. (2023). AI course design planning framework: Developing domain-specific AI education courses. Education Sciences, 13, 954. [Google Scholar] [CrossRef]
  52. School Centre Škofja Loka. (n.d.). In the centre of knowledge. Available online: https://www.scsl.si/ (accessed on 15 September 2025).
  53. Schüller, K. (2022). Data and AI literacy for everyone. Statistical Journal of the IAOS, 38, 477–490. [Google Scholar] [CrossRef]
  54. Secondary Technical School Koper. (n.d.). STŠ koper. Available online: https://www.sts.si/ (accessed on 15 September 2025).
  55. Shen, Y., & Zhang, X. (2024). The impact of artificial intelligence on employment: The role of virtual agglomeration. Humanities and Social Sciences Communications, 11(1), 12. [Google Scholar] [CrossRef]
  56. Shiri, A. (2024). Artificial intelligence literacy: A proposed faceted taxonomy. Digital Library Perspectives, 40, 681–699. [Google Scholar] [CrossRef]
  57. Stolpe, K., & Hallström, J. (2024). Artificial intelligence literacy for technology education. Computers and Education: Open, 6, 100176. [Google Scholar] [CrossRef]
  58. Sun, L., Yin, C., Xu, Q., & Zhao, W. (2023). Artificial intelligence for healthcare and medical education: A systematic review. American Journal of Translational Research, 15(7), 4820–4828. Available online: https://pubmed.ncbi.nlm.nih.gov/37560249 (accessed on 15 September 2025). [PubMed]
  59. Tabachnick, B. G., & Fidell, L. S. (2013). Using multivariate statistics (6th ed.). Pearson Education. Available online: http://ndl.ethernet.edu.et/bitstream/123456789/27657/1/Barbara%20G.%20Tabachnick_2013.pdf (accessed on 15 September 2025).
  60. Tahiru, F. (2021). AI in education: A systematic literature review. Journal of Cases on Information Technology, 23, 1–20. [Google Scholar] [CrossRef]
  61. Touretzky, D., Gardner-McCune, C., & Seehorn, D. (2022). Machine learning and the five big ideas in AI. International Journal of Artificial Intelligence in Education, 33(2), 233–266. [Google Scholar] [CrossRef]
  62. Tzirides, A. O., Zapata, G., Kastania, N. P., Saini, A. K., Castro, V., Ismael, S. A., You, Y., Afonso dos Santos, T., Searsmith, D., O’Brien, C., Cope, B., & Kalantzis, M. (2024). Combining human and artificial intelligence for enhanced AI literacy in higher education. Computers and Education: Open, 6, 100184. [Google Scholar] [CrossRef]
  63. UNESCO. (2009). What you need to know about UNESCO’s new AI competency frameworks for students and teachers. Available online: https://www.unesco.org/en/articles/what-you-need-know-about-unescos-new-ai-competency-frameworks-students-and-teachers (accessed on 15 September 2025).
  64. Univerzitetna založba UM. (2022). Sodobne perspektive družbe: Umetna inteligenca na stičišču znanosti. Univerzitetna založba Univerze v Mariboru. [Google Scholar] [CrossRef]
  65. Velander, J., Otero, N., & Milrad, M. (2024). What is critical (about) AI literacy? Exploring conceptualizations present in AI literacy discourse. In A. Buch, Y. Lindberg, & T. Cerratto Pargman (Eds.), Framing futures in postdigital education. Springer. [Google Scholar] [CrossRef]
  66. Velander, J., Taiye, M. A., Otero, N., & Milrad, M. (2023). Artificial intelligence in K-12 education: Eliciting and reflecting on Swedish teachers’ understanding of AI and its implications for teaching and learning. Education and Information Technologies, 29, 4085–4105. [Google Scholar] [CrossRef]
  67. Veldhuis, A., Lo, P. Y., Kenny, S., & Antle, A. N. (2025). Critical artificial intelligence literacy: A scoping review and framework synthesis. International Journal of Child-Computer Interaction, 43, 100741. [Google Scholar] [CrossRef]
  68. Walter, Y. (2024). Embracing the future of artificial intelligence in the classroom: The relevance of AI literacy, prompt engineering, and critical thinking in modern education. International Journal of Educational Technology in Higher Education, 21, 1–29. [Google Scholar] [CrossRef]
  69. Wilby, R., & Esson, J. (2023). AI literacy in geographic education and research: Capabilities, caveats, and criticality. The Geographical Journal, 190, e12548. [Google Scholar] [CrossRef]
  70. Yim, I. H. Y. (2024). A critical review of teaching and learning AI literacy: Developing an intelligence-based AI literacy framework for primary school education. Computers and Education: Artificial Intelligence, 7, 100205. [Google Scholar] [CrossRef]
  71. Zhang, S., Prasad, P. G., & Schroeder, N. L. (2025). Learning about AI: A systematic review of reviews on AI literacy. Journal of Educational Computing Research, 63(5), 1292–1322. [Google Scholar] [CrossRef]
Figure 1. Mapping AI competences onto four conceptual pillars, where FND—Cognitive–technical foundation, CRIT—critical–reflective, DES—design and creation, and PED—pedagogical. Color boxes represent an alignment with corresponding pillar.
Figure 1. Mapping AI competences onto four conceptual pillars, where FND—Cognitive–technical foundation, CRIT—critical–reflective, DES—design and creation, and PED—pedagogical. Color boxes represent an alignment with corresponding pillar.
Education 15 01455 g001
Figure 2. Thematic evolution map of authors’ keywords from 2017 to 2025.
Figure 2. Thematic evolution map of authors’ keywords from 2017 to 2025.
Education 15 01455 g002
Figure 3. Co-occurrence network of authors’ keywords.
Figure 3. Co-occurrence network of authors’ keywords.
Education 15 01455 g003
Figure 4. Estimated marginal means of the total score of AI literacy (n = 145).
Figure 4. Estimated marginal means of the total score of AI literacy (n = 145).
Education 15 01455 g004
Table 1. Mapping bibliometric themes → AI competency with conceptual framework pillars and AI4K12 tags (Big Idea BI2–BI5), where FND—Cognitive–technical foundation, CRIT—critical–reflective, DES—design and creation, and PED—pedagogical.
Table 1. Mapping bibliometric themes → AI competency with conceptual framework pillars and AI4K12 tags (Big Idea BI2–BI5), where FND—Cognitive–technical foundation, CRIT—critical–reflective, DES—design and creation, and PED—pedagogical.
Bibliometric ThemeAI CompetencyPillarAI4K12
Technical FoundationsRepresentations; Decision-Making; General vs. Narrow; Understanding Intelligence; and ProgrammabilityFNDBI2, BI3
Data and Algorithmic ReasoningData Literacy; Learning from Data; Critically Interpreting Data; and ML StepsFNDBI2, BI3
AI’s Strengths and WeaknessesAI’s Strengths and WeaknessesCRITBI2, BI5
Ethics and Societal ImpactEthicsCRITBI5
Privacy and GovernanceEthics (legal/policy facet)CRITBI5
Human Oversight and AgencyHuman Role in AICRITBI5
Interdisciplinarity and Systems ThinkingInterdisciplinarityCRIT/DESBI5
Recognition of AI and ApplicationsRecognizing AIFNDBI4 (chatbots), BI5 (applications in society)
Participatory Design/CreationProgrammabilityDESBI4, BI5
Classroom Application and Assessment (TPACK)-PED(not an AI4K12 Big Idea; aligns indirectly with BI5 through responsible use)
Table 2. Students’ AI literacy in total and across the competency of AI expressed with mean (M) and standard deviation (SD) (n = 145).
Table 2. Students’ AI literacy in total and across the competency of AI expressed with mean (M) and standard deviation (SD) (n = 145).
AI LiteracyHigher Education (n = 68)Secondary Technical School (n = 77)
FemaleMaleTotalFemaleMaleTotal
MSDMSDMSDMSDMSDMSD
Total AI11.663.2215.213.7012.403.599.752.8712.094.5911.974.54
Recognizing AI0.220.250.280.320.240.250.250.280.300.310.290.30
Understanding intelligence0.760.230.850.280.770.250.660.270.580.300.590.29
Interdisciplinarity0.190.290.530.410.260.350.250.280.390.330.380.33
General vs. narrow0.260.280.400.340.290.300.370.250.300.370.310.37
AI’s strengths and weaknesses0.230.250.320.340.250.270.120.250.240.330.230.33
Representations0.240.250.320.370.260.270.250.270.250.290.250.28
Decision-making0.300.290.410.260.320.290.250.160.380.320.370.31
Machine learning steps0.310.280.310.320.310.290.250.310.190.210.190.21
Human role in AI0.390.330.540.410.420.350.370.450.340.340.340.34
Data literacy0.150.340.430.510.210.400.000.000.280.440.250.44
Learning from data0.500.300.540.360.510.310.370.250.470.350.460.35
Critically interpreting data0.710.450.860.350.740.440.500.570.720.440.710.45
Ethics0.370.240.520.230.410.240.300.110.470.230.450.23
Programmability0.860.340.790.420.840.370.250.500.690.460.660.46
Table 3. Condensed Results for RQ2–RQ4: Key Group and Sex Effects.
Table 3. Condensed Results for RQ2–RQ4: Key Group and Sex Effects.
RQOutcome/ScaleComparisonGroup Means (M)Test/Resultp-ValueEffect Size (η2)Interpretation
RQ2Total AI literacy scoreTeacher Ed (n = 68) vs. Secondary Technical (n = 77)12.40 vs. 11.97ANOVA0.020.054Teacher Ed > Secondary (small–medium)
RQ3Understanding IntelligenceTeacher Ed vs. Secondary TechnicalUnivariate (post-MANCOVA)0.0020.07Teacher Ed advantage (medium)
RQ3ProgrammabilityTeacher Ed vs. Secondary TechnicalUnivariate (post-MANCOVA)0.0450.03Teacher Ed advantage (small)
RQ3Multivariate (14 competencies)Teacher Ed vs. Secondary TechnicalMANCOVA (Wilks’ λ = 0.86, F(14, 129) = 1.41)0.15 No overall multivariate group effect
RQ4Critical AI literacy (4 comps)Teacher Ed vs. Secondary TechnicalMANCOVA (Wilks’ λ = 0.96, F(4, 139) = 1.19)0.314 No group difference
RQ2Sex effect (Total score)Males vs. Females (both groups)ANOVA0.010.074Males > Females (small–medium)
RQ3Sex effects (Interdisciplinarity, Data literacy, Ethics)Males vs. Females (both groups)Univariate (critical & related comps)0.001/0.011/0.0180.07/0.05/00.04Males higher; small–medium
Note. Teacher Ed = pre-service technology teacher education; Secondary Technical = secondary technical school students. Effect size is η2. MANCOVA results are presented with Wilks’ λ and F statistics. Dashes (—) indicate not applicable or not reported.
Table 4. The bridging strategy (S—sense; M—model; D—decide; and A—act).
Table 4. The bridging strategy (S—sense; M—model; D—decide; and A—act).
Gap (What Likely Drove It)Objective (What We Want)High-Leverage Interventions
Knowledge/salience: Limited exposure to AI use-cases outside computer science; examples not anchored in schoolingMake AI uses visible, frequent, and relevant to K-12 practiceWeekly 5 min “AI-in-the-wild” spotlights; misconception repair after mini-quizzes; explicit AI vs. adjacent tech contrasts
Self-efficacy & stereotype threat: Lower confidence in “tech” tasks despite similar prior abilityRaise confidence and belonging without singling out learners5 min values affirmation + utility-value writing; structured pair roles (Driver/Navigator; Analyst/Skeptic); choice of task contexts
Transfer: Difficulty mapping AI concepts across disciplines (e.g., route optimization → classroom logistics)Build ability to map and justify AI across subjects using a stable frameUse a single S–M–D–A template across all labs; weekly near-transfer exit items; jigsaw to re-map a lab to a new subject
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rupnik, D.; Avsec, S. Toward a Coherent AI Literacy Pathway in Technology Education: Bibliometric Synthesis and Cross-Sectional Assessment. Educ. Sci. 2025, 15, 1455. https://doi.org/10.3390/educsci15111455

AMA Style

Rupnik D, Avsec S. Toward a Coherent AI Literacy Pathway in Technology Education: Bibliometric Synthesis and Cross-Sectional Assessment. Education Sciences. 2025; 15(11):1455. https://doi.org/10.3390/educsci15111455

Chicago/Turabian Style

Rupnik, Denis, and Stanislav Avsec. 2025. "Toward a Coherent AI Literacy Pathway in Technology Education: Bibliometric Synthesis and Cross-Sectional Assessment" Education Sciences 15, no. 11: 1455. https://doi.org/10.3390/educsci15111455

APA Style

Rupnik, D., & Avsec, S. (2025). Toward a Coherent AI Literacy Pathway in Technology Education: Bibliometric Synthesis and Cross-Sectional Assessment. Education Sciences, 15(11), 1455. https://doi.org/10.3390/educsci15111455

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop