You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

10 July 2025

Designing a Generalist Education AI Framework for Multimodal Learning and Ethical Data Governance

,
,
,
and
1
School of Education, Guangzhou University, Guangzhou 510006, China
2
Warrington College of Business, University of Florida, Gainesville, FL 32611, USA
3
School of Music, University of Florida, Gainesville, FL 32611, USA
4
Department of Computer Sciences, Stanford University, Stanford, CA 94305, USA
This article belongs to the Special Issue ICT in Education, 2nd Edition

Abstract

The integration of artificial intelligence (AI) into education requires frameworks that are not only technically robust but also ethically and pedagogically grounded. This paper proposes the Generalist Education Artificial Intelligence (GEAI) framework—a conceptual blueprint designed to enable privacy-preserving, personalized, and multimodal AI-supported learning in educational contexts. GEAI features a Trusted Domain architecture that supports secure, voluntary multimodal data collection via multimedia registration devices (MM Devices), edge-based AI inference, and institutional data sovereignty. Drawing on principles from constructivist pedagogy and regulatory standards such as GDPR and FERPA, GEAI supports adaptive feedback, engagement monitoring, and learner-centered interaction while addressing key challenges in ethical data governance, transparency, and accountability. To bridge theory and application, we outline a staged validation roadmap informed by technical feasibility assessments and stakeholder input. This roadmap lays the foundation for future prototyping and responsible deployment in real-world educational settings, positioning GEAI as a forward-looking contribution to both AI system design and education policy alignment.

1. Introduction

Education remains a cornerstone of human development, underpinning social well-being, economic growth, gender equality, and global citizenship. Achieving these outcomes depends on inclusive, equitable, and high-quality learning opportunities across the lifespan. However, education systems worldwide face mounting challenges. Persistent access gaps continue to marginalize millions of learners, while those in school often struggle to acquire foundational knowledge and digital competencies necessary for a rapidly evolving, interconnected world [1,2].
In this context, the integration of information and communication technologies (ICT), particularly artificial intelligence (AI), has opened new avenues for educational innovation. AI technologies offer potential to personalize learning pathways, support formative assessment, and optimize instructional design [3,4,5,6]. Intelligent tutoring systems, automated feedback tools, and multimodal classroom analytics exemplify how AI is reshaping teaching and learning practices in digitally transformed environments [7,8].
Among recent developments, large-scale foundational models (FMs)—such as generative language and multimodal models—represent a major leap forward in educational AI. These models are trained on diverse, extensive datasets and demonstrate general-purpose capabilities adaptable to varied educational scenarios [9]. Applications range from real-time generation of customized feedback to adaptive content delivery based on learner profiles and behavioral data.
Yet, the rapid deployment of such models also raises pressing ethical, legal, and pedagogical concerns. Data protection laws such as the GDPR and FERPA impose strict limits on the collection and use of student information, and many institutions lack sufficient infrastructure to support compliant AI applications [10,11]. Additionally, the growing presence of AI-generated content (AIGC) exacerbates challenges related to academic integrity, misinformation, and deepfake misuse, which may undermine institutional trust and learner safety [12,13,14,15]. These risks call for proactive governance frameworks that balance innovation with accountability.
To address these opportunities and risks, coordinated efforts from researchers, educators, technology providers, and policymakers are urgently needed. Multistakeholder collaboration is essential for developing responsible AI infrastructures that uphold educational values while mitigating systemic harms [13,16,17,18].
To clarify our terminology, we define AI-based education in this study as the integration of emerging generalist AI models—such as large language models and multimodal foundation models—into teaching and learning processes. These models, characterized by their ability to handle diverse educational tasks (e.g., feedback generation, content creation, behavior monitoring), go beyond traditional narrow AI systems and represent a paradigm shift toward more flexible and general-purpose capabilities. Although still in early stages of application, we consider this direction foundational for future-ready educational architectures.
Regarding ethical data governance, we focus specifically on privacy-preserving mechanisms and legal compliance as the primary pillars of ethical design. This includes end-to-end encryption, institutional control over data processing, and adherence to GDPR and FERPA regulations. While broader ethical concerns—such as fairness, transparency, and human oversight—are integral to responsible AI, this study prioritizes data protection architecture, addressing the urgent need for secure infrastructures in AI-supported educational environments.
In response, this paper introduces the Generalist Education Artificial Intelligence (GEAI) framework—a conceptual infrastructure designed to support ethical, personalized, and privacy-preserving AI deployment in education. GEAI features a trusted domain architecture, incorporating secure multimedia registration devices (MM Devices) to collect real-time multimodal data for intelligent educational services. This design aligns with principles of responsible innovation and human-centered learning.
The remainder of this paper is structured as follows:
Section 2 reviews related work on foundational models and their application in educational contexts.
Section 3 introduces the GEAI framework, detailing its domain architecture and core components.
Section 4 presents three use cases that illustrate the framework’s potential for adaptive tutoring, secure assessments, and institutional data protection.
Section 5 discusses the pedagogical, social, and ethical implications of GEAI, including its alignment with constructivist principles.
Finally, Section 6 concludes the paper and outlines limitations and directions for future research and validation.
The research presented in this paper addresses the following questions:
  • How does the GEAI framework improve data governance and privacy protection while enabling personalized, AI-supported learning?
  • In what ways do MM Devices facilitate real-time learner data collection for adaptive instruction and intelligent tutoring?
  • What ethical and legal risks accompany the deployment of GEAI, and how can these be mitigated through transparent design and institutional policy?

3. Materials and Methods

In response to the challenges of transparency, data fragmentation, and regulatory compliance in current educational AI systems, we propose the Generalist Education AI (GEAI) framework as a conceptual blueprint. Grounded in a Trusted Domain architecture, this framework is designed to enable legally compliant, context-sensitive, and scalable AI deployment in real-world educational environments. It is important to note that the GEAI framework, as proposed in this study, is a conceptual architecture rather than a validated system. While it draws upon the design principles of emerging generalist AI models—such as their capacity to operate across modalities and tasks—it does not claim to deliver full generalist functionality at this stage. Instead, GEAI serves as a blueprint for guiding future prototyping and implementation in educational contexts, particularly in settings where multimodal AI capabilities and ethical data infrastructures are under development.
As illustrated in Figure 1, GEAI consists of three tightly integrated domains that form a closed-loop pipeline:
  • (A) MM Device Domain: Handles secure and privacy-preserving multimodal data collection from teaching environments;
  • (B) GEAI Resource Domain: Aggregates and preprocesses data to create structured learning resources and prepare training inputs for foundation models;
  • (C) GEAI Model and Interaction Domain: Facilitates model training, personalization, and the delivery of intelligent educational services.
Figure 1. Overview of the GEAI Framework. The figure depicts the architecture of the proposed GEAI framework, comprising three interconnected domains that together form a secure, adaptive, and auditable AI infrastructure for education. This closed-loop pipeline supports multimodal data acquisition, compliant processing, and personalized service delivery in classroom and online settings.
The MM Device Domain (Domain A) is responsible for data acquisition within authentic educational scenarios. Multimedia registration devices (e.g., AI cameras, wearable sensors, tablets) are deployed in physical or virtual classrooms to capture multimodal inputs, including video, audio, and user interactions. These devices are embedded with tamper-resistant modules and local encryption capabilities. Device operation is managed through a secure key injection protocol, ensuring compliance with data protection regulations. All data collected in this domain is encrypted and securely transmitted to Domain B.
The GEAI Resource Domain (Domain B) functions as the central data management and preprocessing layer. It consolidates heterogeneous educational data—such as performance logs, behavioral traces, and instructional content—and organizes them into a multimodal knowledge graph. This domain also enforces data legitimacy and diversity, preparing high-quality and ethically sourced training datasets for subsequent AI model development.
The Model and Interaction Domain (Domain C) supports the pretraining, fine-tuning, and deployment of AI models tailored to educational tasks. These models provide automated tutoring, personalized feedback, and analytics. Processed outputs are delivered back to applications such as adaptive learning platforms or virtual assistants, while their usage data can optionally be looped back into the system via the MM Devices. This cyclical flow forms a continuous learning loop, enabling dynamic adaptation and ongoing performance optimization.

3.1. GEAI Model

3.1.1. MM Devices for Secure Educational Data Collection

The integration of AI technologies in education faces major challenges in data availability, privacy protection, and regulatory compliance. High-quality real-world data for instructional analysis remains scarce due to the high cost and limited scalability of expert classroom observations [5], while audiovisual recordings are often restricted by legal and ethical concerns. To address these constraints, the GEAI framework introduces Multimedia Registration Devices (MM Devices)—specialized hardware terminals designed to operate within a Trusted Domain and securely collect multimodal classroom data, providing a privacy-preserving foundation for educational analytics. These devices are currently at the conceptual stage, pending prototype development and validation.
MM Devices are intended for deployment in both physical and virtual classrooms, capable of capturing diverse data modalities such as video, audio, and handwriting for potential applications in facial expression analysis, speech recognition, and response evaluation. Initial configurations propose low-resolution sensors (e.g., 720p cameras, 8 kHz microphones) as cost-effective baselines, scalable to higher specifications (e.g., 1080p, 16 kHz) as needed. Drawing on best practices from the payment industry, each device will integrate a Hardware Security Module (HSM)—adapted from Point-of-Sale (POS) systems—to manage cryptographic keys and ensure secure content encryption in compliance with PCI P2PE guidelines [32,33].
To support local AI processing and reduce dependency on cloud infrastructure, MM Devices are envisioned to incorporate lightweight edge AI models, such as MobileNet for video processing and Whisper-tiny for speech transcription. The projected unit cost is approximately USD 500, based on market trends in embedded AI devices (e.g., NVIDIA Jetson Nano) and HSM-integrated platforms [34]. Final costs and performance benchmarks will be established during prototyping.
To mitigate physical and cybersecurity threats, MM Devices will adopt a layered security architecture, including secure execution environments, app-level access controls, and HSM-based identity verification. The hardware design emphasizes portability, with battery support (5–10 h), modular sensor slots, and operational LED indicators. Software components—such as tutoring agents, behavior classifiers, or encoders—will operate in sandboxed environments, subject to formal auditing, approval, and version control. Data transmission will be limited to registered institutional GEAI platforms over secure channels, avoiding exposure to unsecured networks (e.g., public Wi-Fi), and employing encryption protocols aligned with PCI P2PE.
To enhance system accountability, the design also envisions blockchain-based logging, implemented through a lightweight private blockchain (e.g., Proof of Authority) that records local, tamper-proof logs of device operations and compliance status. These logs will be periodically synchronized to the GEAI platform’s distributed ledger, enabling real-time auditing and traceability [35,36]. The feasibility of this approach depends on optimizing energy consumption and computing overhead during prototype validation. A summary of these layered security strategies is presented in Table 1. Ultimately, MM Devices aim to provide a scalable infrastructure for secure, privacy-compliant, and personalized learning, with benefits such as improved student engagement and adaptive support to be evaluated in future trials [37].
Table 1. Proposed Layered Security Strategies of MM Devices.
Note: The proposed components, AI features, cost estimates, and blockchain logging mechanisms remain conceptual and require empirical validation through prototyping.

3.1.2. HSM and Key Management for MM Devices

To safeguard the secure storage and transmission of multimodal educational data, MM Devices employ a dual-layer cryptographic system comprising symmetric encryption keys (e.g., AES-256) and asymmetric key pairs. These keys are managed under a rigorous lifecycle framework encompassing generation, storage, rotation, and revocation, with oversight extending to personnel and equipment. Central to this architecture are Hardware Security Modules (HSMs), certified to FIPS 140-2 Level 3 standards, which offer tamper-resistant protection against both physical and digital intrusion [38]. Integrated directly into MM Devices, these HSMs enable secure key generation, real-time storage, and automated key erasure in response to anomaly-based intrusion detection [34].
During initial provisioning or maintenance, cryptographic keys and device certificates are injected into ISO 27001-certified secure environments provided by MM Device manufacturers, ensuring compliance with data protection regulations such as GDPR and FERPA [22]. For post-deployment management, a remote key injection mechanism is proposed, utilizing secure communication protocols (e.g., TLS 1.3) and end-to-end encryption for authenticated updates via registered GEAI platforms. This security model is adapted from the PCI Point-to-Point Encryption (P2PE) framework widely used in the payment industry [32], tailored to the confidentiality and integrity requirements of educational contexts.
By embedding such a robust HSM-based key management infrastructure, MM Devices can ensure data security across all phases of educational analytics—spanning classroom video, audio, and written responses—while preserving auditability and enabling scalable, regulation-compliant AI deployment. A summary of these features is provided in Table 2.
Table 2. HSM and Key Management Features for MM Devices.

3.1.3. Trusted Domains of GEAI

The GEAI Trusted Domain establishes a vertically integrated architecture inspired by regulated industries such as healthcare and finance; yet it is uniquely adapted for educational contexts. It emphasizes data traceability, institutional sovereignty, and AI auditability, which are essential for trustworthy educational AI. This domain consists of two interconnected components: the GEAI Resource Domain and the GEAI Model and Inference Domain, both governed by cross-institutional policies (e.g., data-sharing agreements) and secured via technical protocols (e.g., TLS 1.3). Together, these components form a secure, interoperable, and ethically governed infrastructure that supports multimodal data processing while ensuring compliance with legal frameworks such as GDPR and FERPA [22].
  • GEAI Resource Domain
This domain serves as the foundation of a trusted educational AI ecosystem. It aggregates certified hardware (e.g., Multimedia Registration Devices, or MM Devices), human expertise, and diverse datasets from multiple institutions under predefined regulatory and ethical conditions. These inputs are indexed and preprocessed to train foundational GEAI models for downstream applications, such as emotion-aware tutoring and adaptive lesson planning. This design aligns with UNESCO’s global educational imperatives—equity, quality, and relevance [1].
Data Legitimacy and Traceability: The domain captures multimodal classroom data (e.g., up to 1080p video, up to 16 kHz audio, interaction logs) from proposed MM Devices. Provenance metadata is recorded via blockchain audit trails, and AI-generated content (AIGC) is labeled with timestamps and model identifiers. These mechanisms aim to improve academic transparency and address concerns surrounding deepfake risks in educational settings [22,35].
Bias Mitigation and Inclusiveness: The domain includes a broad spectrum of data across K-12 and higher education, covering linguistic (e.g., English, Mandarin), geographic (urban/rural), and disciplinary (e.g., STEM, humanities) dimensions. Proposed stratified sampling ensures representativeness, mitigating algorithmic bias and improving model fairness [9,46].
Secure Use of Personal Learning Data: Student records (e.g., test scores, competency assessments) are protected via AES-256 encryption and role-based access. Federated learning protocols (e.g., FedAvg) allow model personalization without compromising data privacy [47].
Knowledge Graph Integration: Educational content is structured using domain ontologies (e.g., MathML) and knowledge graphs to support semantic reasoning. For example, misconceptions in physics may be linked to prerequisite mathematical concepts via graph-based inference, enabling personalized learning trajectories [48,49].
These proposed features, summarized in Table 3, position the GEAI Resource Domain as an innovative foundation for educational AI, seamlessly integrating with the MM Devices’ secure data pipeline to address global educational challenges.
Table 3. Core Features of the GEAI Resource Domain.
  • GEAI Model and Inference Domain
The GEAI Model and Inference Domain are envisioned to leverage proposed multimodal data from the Resource Domain—such as text from lesson plans, envisioned classroom video (up to 1080p resolution), and audio (up to 16 kHz) from conceptualized Multimedia Registration Devices (MM Devices)—for large-scale pretraining. This process is planned to involve processing hundreds of billions of tokens and substantial compute resources (e.g., ~10,000 GPU hours), aligning with benchmarks for foundation models [20,50]. The proposed pretraining will employ unsupervised learning tasks, including masked language modeling (MLM) for text [19], contrastive learning for vision-language alignment (e.g., CLIP-style objectives), and self-supervised speech modeling (e.g., wav2vec), to encode a general knowledge framework spanning educational domains such as STEM and humanities [9,51]. The conceptualized pretrained model, based on a Transformer architecture with approximately 12 billion parameters, aims to demonstrate capabilities in language generation, translation, and question answering, including competitive performance in tasks like English–Mandarin translation [20]. However, to address specific educational needs—such as curriculum alignment and student engagement—further fine-tuning is deemed essential, pending prototype validation [22].
Fine-tuning is planned to adapt the pretrained GEAI model for diverse educational tasks using two promising approaches.
Reinforcement Learning with Human Feedback (RLHF) is envisioned to align model outputs with human preferences, potentially supporting applications like generating formative feedback for student essays to improve writing skills or structuring dialogues for peer learning, addressing the quality challenge in education where tailored feedback is scarce [52].
Prompt engineering, using curriculum-aligned instructions (e.g., “Explain Newton’s laws for a 9th grader”), is proposed to enable flexible adaptation without retraining, reducing resource demands and supporting diverse teaching contexts [28]. These methods are intended to underpin applications such as: (a) virtual teaching assistants that generate personalized learning content; (b) pedagogical feedback systems leveraging envisioned MM Device audio logs to support teacher–student interactions; (c) automated resource planning tools to streamline lesson preparation. These applications aim to enhance teaching effectiveness and reduce workload, though empirical evaluations in educational settings remain a future goal, subject to prototype development [28,52]. Ongoing research is planned to assess these applications in real-world classrooms, ensuring alignment with pedagogical goals like Vygotsky’s Zone of Proximal Development to improve learning outcomes [22].
A summary of these approaches and applications is provided in Table 4.
Table 4. Fine-Tuning Approaches and Potential Applications of the GEAI Model and Inference Domain.

3.2. The Application Prospects of the GEAI Model Framework

The GEAI framework is envisioned as a platform tailored for educational environments, building on proposed foundational AI models to offer innovative applications across four key areas. These applications are designed to integrate with the conceptualized secure multimodal data pipeline from proposed Multimedia Registration Devices (MM Devices), such as envisioned video (up to 1080p resolution) and audio (up to 16 kHz), addressing global educational challenges like accessibility, equity, and relevance as outlined by UNESCO [1].
Dynamic Task Understanding and Generalization: GEAI is planned to leverage pretrained multimodal capabilities to understand and address new tasks described in simplified natural language (e.g., at a sixth-grade reading level) without retraining, using proposed prompt engineering techniques [20,53,54]. For instance, the model is envisioned to generate Turkish language geography teaching aids for third-grade students by interpreting prompts like “Create a map-based lesson on Turkish rivers for a 3rd grader.” This capability aims to support diverse curricula, tackling the accessibility challenge in education where tailored resources are limited, especially in multilingual contexts [55]. Validation of this feature will depend on future prototype development.
Real-Time Dynamic Support for Teaching Activities: GEAI is designed to integrate authorized classroom data from proposed MM Devices—such as envisioned student facial expressions (up to 1080p video) and audio responses (up to 16 kHz)—along with historical interaction logs, to provide real-time decision-making support. For example, during in-person or remote teaching, the framework is envisioned to use proposed vision-language models to analyze student engagement and generate prompts like “Pause and clarify fractions” to help teachers adjust their pace. In virtual settings, it is planned to dynamically modify tasks (e.g., simplify math problems if students struggle), addressing the equity challenge by fostering responsive and inclusive classroom environments, pending prototype validation [22,56].
Knowledge Graph-Driven Multidisciplinary Knowledge Presentation: GEAI is proposed to employ knowledge graphs (KGs) constructed with discipline-specific ontologies (e.g., ChEBI for chemistry) to represent and adapt multidisciplinary knowledge for teaching contexts. For instance, students studying environmental conservation are envisioned to access a KG linking chemistry (e.g., carbon cycles), ecology (e.g., biodiversity impacts), and policy-making (e.g., recent climate agreements), presented in age-appropriate language (e.g., eighth-grade level). This approach aims to support inquiry-based learning and address the relevance crisis in curricula that often lack interdisciplinary integration [22,57]. Feasibility will be assessed during future prototype testing.
Resource Transparency and Rational Utilization: Within the envisioned Trusted Domain, resources are planned to be traceable via proposed blockchain audit trails and auditable for compliance (e.g., GDPR, FERPA), ensuring transparent usage [35]. For example, forensic science content unsuitable for younger audiences is intended to be restricted using proposed role-based access control (RBAC) and AES-256 encryption, ensuring age-appropriateness and contextual relevance. This mechanism addresses the equity challenge by aligning with Fairness, Accountability, Transparency, and Ethics (FATE) principles, promoting responsible resource utilization across educational levels, subject to future validation [22,58].

4. Results

4.1. Use Case 1: Personalized Learning and Adaptive AI Tutoring

Educational Challenge: Despite recent advancements in adaptive learning technologies, existing systems often fail to deliver truly personalized instruction due to two key limitations: (1) a lack of real-time feedback loops, which inhibits timely adjustments to student engagement and comprehension signals, thus diminishing instructional quality [37]; and (2) persistent concerns regarding privacy and regulatory compliance, particularly when student data are transferred to centralized cloud services, thereby increasing data security risks in educational settings [22].
GEAI-Enabled Solution: The GEAI framework introduces a conceptualized Trusted Domain infrastructure that securely integrates Multimedia Registration Devices (MM Devices), multimodal AI models, and institutionally governed data pipelines. This architecture enables real-time, privacy-preserving AI tutoring at scale, pending prototype development and pilot testing [35,47].
Workflow and Capabilities:
  • Secure Multimodal Sensing via MM Devices: MM Devices are envisioned for classroom deployment to capture rich multimodal signals, including:
Visual cues (e.g., facial expressions, gaze shifts via up to 1080p video);
Auditory responses (e.g., oral responsesand answers captured atup to 16 kHz audio);
Behavioral metrics (e.g., posture changes, hesitation, keyboard latency).
These data streams are locally encrypted (AES-256) and preprocessed on-device to preserve student privacy and comply with FERPA and GDPR [35].
  • Adaptive AI Tutoring via Federated Inference: GEAI’s core inference module incorporates three AI strategies (see Figure 2):
    Figure 2. GEAI-based Framework for Personalized Learning and Adaptive AI Tutoring. The figure illustrates how MM Devices collect multimodal signals (gaze, speech, behavioral data), which are then processed through federated learning, reinforcement learning, and prompt-based NLP to generate adaptive responses for students, teachers, and parents.
Federated Learning (e.g., FedAvg algorithm) allows local model updates on MM Devices without transmitting raw data to the cloud [47];
Reinforcement Learning dynamically adjusts the difficulty and sequencing of learning materials based on observed student states;
Prompt-based NLP agents provide real-time explanations, motivational feedback, and task scaffolding tailored to student needs [19,28,52]. For example, if a student shows hesitation when solving a fractions problem, the GEAI system may suggest visual analogies or simplified sub-tasks and adjust the next questions accordingly.
  • Multistakeholder Feedback and Reporting: Proposed MM Devices are also designed to deliver real-time micro-feedback to instructors (e.g., “Pause and clarify fractions”), enhancing situational awareness. Beyond real-time assistance, securely transmitted and audit-tracked outputs will support role-specific dashboards on the GEAI platform:
For students: personalized performance reports with progress visualizations and learning goals;
For teachers: dashboards identifying struggling learners and instructional bottlenecks;
For parents: weekly summaries of engagement, growth metrics, and emotional markers.
All reports will be governed by institutional access policies and privacy safeguards, ensuring ethical deployment and regulatory compliance [22].
Educational Impact: By integrating edge-based multimodal sensing with secure model inference and targeted stakeholder feedback, GEAI transforms static e-learning environments into responsive, learner-centered ecosystems. This closed-loop system enhances motivation, supports differentiated instruction, and empowers educators with actionable insights—thereby addressing both equity and quality concerns in contemporary education [22,37].

4.2. Use Case 2: AI-Assisted Secure Assessments and Cheating Prevention

Educational Challenge: In digital learning environments, traditional assessment methods face escalating challenges, including AI-assisted plagiarism (e.g., ChatGPT misuse), identity fraud in remote testing, algorithmic bias in grading, and data security risks from cloud-based storage. These issues undermine academic integrity and fairness while amplifying systemic inequalities in assessment outcomes [55,58]. Addressing these issues requires secure, explainable, and ethically governed AI solutions.
GEAI-Enabled Solution: The GEAI framework proposes a Trusted Domain infrastructure that leverages MM Devices and modular AI components to ensure secure, equitable, and tamper-proof assessments. The system aligns with regulatory frameworks (e.g., GDPR, FERPA) and educational standards for fairness, explainability, and data integrity [22].
Workflow and Capabilities:
  • AI-Powered Identity Verification and Authentication: MM Devices support real-time biometric authentication using pre-exam facial recognition (e.g., 95% accuracy per NIST standards, NIST, 2023), continuous face tracking (up to 1080p video at 30 fps), voice verification (up to 16 kHz), and keystroke dynamics to validate identity. Behavioral anomaly detection models track eye movement (e.g., sustained gaze aversion > 5 s) and sudden keystroke pattern shifts to flag suspicious behavior.
Outcome: Secures assessment participation and eliminates identity fraud.
  • AI-Driven Exam Integrity Monitoring and Proctoring: Real-time audio and screen analysis detects unauthorized behaviors such as whispering, conversation cues, or tab switching. Deepfake detection algorithms—based on CNN architectures—distinguish live video from pre-recorded footage, ensuring authenticity [58].
Outcome: Preserves assessment credibility and reduces academic dishonesty.
  • Automated Grading with Fairness Controls: GEAI incorporates OCR and a BERT-based NLP engine to interpret handwritten and typed responses against structured rubrics (e.g., “Score 8/10 for explanation with missing term”). An explainable AI (XAI) layer enables transparency in automated scoring (e.g., “Score reduced due to incomplete diagram”), while fairness adjustments are informed by teacher review [22]. Grading outputs are recorded immutably using blockchain ledgers (e.g., Hyperledger Fabric), ensuring tamper-proof verification [35].
Outcome: Enables verifiable, bias-mitigated grading aligned with educational fairness principles.
  • Secure Data Processing and Compliance Assurance: Assessment data are processed locally on MM Devices and within institutional Trusted Domains using AES-256 encryption, avoiding third-party risks. Data control resides with the institution, ensuring full alignment with GDPR/FERPA mandates and preventing unauthorized access.
Outcome: Guarantees privacy and institutional sovereignty in assessment workflows.
Educational Impact: GEAI strengthens academic integrity by integrating biometric authentication, behavioral monitoring, explainable AI grading, and blockchain-secured outcomes. These measures reduce cheating, mitigate grading bias, and uphold data privacy, directly supporting equitable and high-quality learning outcomes [22,58].
Beyond technical features, the system reflects a principled approach to AI governance in education. As highlighted by Yan and Liu [59], responsible AIED frameworks should be built on five pillars: (1) provenance transparency for AI models; (2) legal legitimacy of training datasets; (3) institutionalized ethical review procedures; (4) accountability of deploying entities; and (5) prioritization of learner benefit.
GEAI’s Trusted Domain design operationalizes these principles, establishing alignment with emerging global standards in ethical educational AI.

4.3. Use Case 3: Institutional Data Protection and Compliance with GEAI

Educational Challenge: With the increasing deployment of AI-driven technologies in education, institutions face mounting concerns over data privacy, legal compliance, and cyber-resilience. Key risks include unauthorized access to student data, breaches in third-party cloud systems, and a lack of transparency in AI decision-making processes [55,58]. This comprehensive system architecture is illustrated in Figure 3, which visually integrates the key components of secure AI-assisted assessments. Without secure, localized processing, institutions often relinquish control over sensitive data, increasing exposure to cyberattacks and compliance violations [35].
Figure 3. GEAI-Based Framework for Secure and Fair AI-Assisted Assessment. The figure illustrates how the Trusted Domain integrates multimodal authentication, anomaly detection, explainable grading, and compliance mechanisms to deliver secure, transparent, and equitable assessments.
GEAI-Enabled Solution: The GEAI framework addresses these concerns by embedding data protection, regulatory alignment, and explainability into its architecture. Leveraging MM Devices and a Trusted Domain infrastructure, GEAI facilitates secure on-site AI processing, robust access control, and continuous compliance monitoring, thereby strengthening institutional autonomy and data integrity [22].
Workflow and Capabilities:
  • Secure On-Premises Data Processing with MM Devices: GEAI ensures that data is captured, processed, and retained within institutional boundaries:
Local AI Computation: Proposed MM Devices (e.g., supporting up to 1080p video at 30 fps and 16 kHz audio) are designed to operate as secure edge nodes, analyzing behavioral, academic, and engagement data without transferring raw inputs to external servers.
Advanced Encryption and Role-Based Access Control (RBAC): All data streams are encrypted using AES-256 at the hardware level, while RBAC protocols ensure that only verified personnel (e.g., course instructors) access relevant datasets.
Impact: Maintains institutional sovereignty over data while minimizing third-party exposure and regulatory risk.
  • AI-Powered Institutional Compliance Monitoring: GEAI embeds compliance oversight within its operational loop:
Automated Privacy Audits: Proposed MM Devices generate daily compliance reports (e.g., “100% GDPR adherence in data storage”), scanning AI logs for violations (e.g., unauthorized data access attempts) using machine learning models.
Explainable AI for Transparent Decisions: XAI techniques (e.g., LIME) provide justifications for AI outputs (e.g., “Grade revised due to rubric misalignment”), supporting ethical grading and student rights [22].
Federated Model Training for Privacy Preservation: Using FedAvg [47], GEAI enables frequent (e.g., daily) model updates across devices without sharing raw data, ensuring continuous improvement while retaining compliance.
Impact: Reduces the institutional burden of manual audits while aligning with educational AI governance principles.
  • AI-Augmented Cybersecurity and Data Integrity: To further protect educational ecosystems:
Real-Time Anomaly Detection: GEAI’s embedded models monitor network traffic patterns (e.g., flagging ≥ 3 login failures in 5 min) and alert administrators to potential threats.
Blockchain-Based Academic Records: Immutable ledgers using Hyperledger Fabric (with 10-min consensus cycles) ensure the security of grades, transcripts, and other academic credentials [35].
AI-Driven Data Loss Prevention (DLP): MM Devices detect and halt abnormal file transmissions (e.g., blocking unauthorized exports > 1 MB), preventing inadvertent or malicious data leaks.
Impact: Reinforces cybersecurity posture, prevents manipulation of academic records, and safeguards institutional reputation.
Educational Impact: By embedding security, compliance, and ethical AI operations at the core of its architecture, GEAI enables institutions to foster a trustworthy and transparent learning environment. This reduces legal exposure, bolsters student data protection, and ensures equitable personalization without compromising privacy [58]. As AI adoption scales across education, such integrated frameworks become vital to maintaining public trust and pedagogical integrity.

5. Discussion

Pedagogical Dependence and Role Repositioning: The integration of the Generalist Education Artificial Intelligence (GEAI) framework into educational environments raises concerns regarding pedagogical overreliance and the evolving role of teachers. Current AI-detection tools exhibit limited effectiveness in identifying sophisticated generative plagiarism (e.g., paraphrasing via ChatGPT), with detection accuracy reaching only 60% against State-of-the-Art models [58]. This complicates the evaluation of authentic student performance. Historical precedents, such as the incorporation of calculators into mathematics instruction, demonstrate that technology and traditional pedagogy can coexist. However, GEAI’s diagnostic functions—such as analyzing conceptual misunderstandings in fractions—lack the contextual and empathetic nuance inherent in human feedback [6]. Moreover, educators express concerns over diminished professional agency, with recent surveys indicating that 30% of teachers fear job displacement due to automation [55]. To mitigate this, GEAI should be framed as an augmentative tool that enhances instructional efficacy. For instance, proposed Multimedia Registration Devices (MM Devices), designed to capture 1080p video at 30 fps, may assist in engagement analysis, potentially reducing lesson preparation time by up to 15% [4]. Ensuring GEAI’s alignment with constructivist pedagogical models requires clear policy guidance and robust teacher training programs.
The GEAI framework is grounded in constructivist pedagogical principles, which emphasize learner agency, contextualized knowledge construction, and iterative feedback. By enabling real-time interaction through MM Devices and adaptive AI tutors, the framework supports individualized learning trajectories aligned with learners’ prior knowledge and performance signals. These features reflect the constructivist concept of the Zone of Proximal Development (ZPD), wherein learning is most effective when tasks are slightly beyond the learner’s current competence but supported through scaffolded guidance [60]. In addition, GEAI’s dialogic feedback mechanisms and multimodal sensing capabilities encourage reflective thinking, collaborative engagement, and inquiry-based exploration—core tenets of constructivist instruction [61].
Social Implications and Learner Well-Being: The widespread digitization of learning, accelerated by AI integration, has intensified concerns around student social isolation. Studies report a 20% increase in feelings of loneliness among adolescents since 2018 [62]. Although GEAI proposes to offer personalized emotional support via multimodal input—such as speech signal analysis at 16 kHz—it cannot replicate the socio-emotional support afforded by human educators. Human-centered design principles must be foundational in GEAI development to sustain a sense of classroom community. Design frameworks should incorporate metrics for psychological well-being, supported by longitudinal studies aiming to reduce student loneliness by 5% by 2027 [22]. Emphasizing socio-affective interaction through AI design not only enhances engagement but also mitigates the dehumanizing aspects of automation.
Data Privacy and Security Risks: GEAI’s reliance on high-resolution multimodal data—such as facial expressions and vocal signals recorded by MM Devices—raises significant privacy and security concerns. Even with advanced security protocols like AES-256 encryption within a Trusted Domain infrastructure, risks remain. Complex deep learning architectures are susceptible to latent data retention, with similar systems exhibiting an estimated 5% annual data leakage rate [58]. This issue is particularly critical when involving minors, where violations of regulations such as GDPR and FERPA have led to fines exceeding $1.3 million [35]. To address these vulnerabilities, privacy-by-design principles must be rigorously applied. This includes data minimization strategies, implementation of k-anonymity (ensuring 95% anonymity), and differential privacy mechanisms (e.g., noise addition with ε = 1). Transparent data governance—comprising monthly audit logs and enforceable informed consent protocols—is essential to uphold public trust and regulatory compliance.
Comparative Insights and Stakeholder Recommendations
Unlike existing AI-enhanced learning platforms such as MATHia, which focus narrowly on subject-specific adaptivity (e.g., mathematics), or Turnitin, which provides post-hoc plagiarism detection, GEAI introduces a general-purpose infrastructure with real-time multimodal sensing, adaptive feedback, and built-in ethical compliance mechanisms. While tools like ChatGPT have demonstrated potential for personalized assistance, they often operate outside of institutional control and lack secure data governance. In contrast, GEAI is designed to integrate directly with institutional policies and infrastructures, emphasizing transparency, traceability, and stakeholder accountability.
To support the responsible implementation of GEAI and similar systems, we propose the following stakeholder-specific recommendations:
  • For educators: Engage in co-design of AI-supported instruction and receive targeted professional development to maintain pedagogical agency. Feedback from teachers should inform system design iterations [63].
  • For institutional leaders: Establish data governance committees and invest in secure on-premise infrastructure to ensure FERPA/GDPR compliance and retain control over learning data.
  • For policymakers: Promote funding mechanisms for pilot experimentation in diverse school contexts and issue clear regulatory guidelines for ethical AI adoption in education [31].
These measures can help align AI innovation with educational values, ensuring that systems like GEAI serve as trustworthy, augmentative tools that benefit all learners and educators.
Recommendations for Responsible GEAI Deployment
  • Empower Educators: Offer training and participatory design, aiming for a 10% improvement in student outcomes [55].
  • Embed Ethics and Social Impact: Conduct equity impact analysis and loneliness studies.
  • Strengthen Data Governance: Establish boards to oversee data processes, targeting a < 1% leak rate.
  • Foster Human-AI Collaboration: Design for socio-emotional support, enhancing engagement by 15% [4].
By addressing these limitations, GEAI evolves into a socially responsive, human-centered educational tool, aligned with the future of ethical, inclusive, and effective teaching.

6. Conclusions

The Generalist Education Artificial Intelligence (GEAI) framework builds upon cutting-edge advancements in artificial intelligence, especially large-scale language models, to introduce a transformative vision for education. It facilitates legally compliant, voluntary data collection via conceptual Multimedia Registration Devices (MM Devices), currently in a theoretical development phase. These devices are designed to enhance educational outcomes through personalized learning and real-time analytics, with projected engagement gains of up to 10% in early simulations [55].
GEAI’s real-time adaptability is driven by a rich ecosystem of multimodal inputs and knowledge graphs containing over 50,000 educational nodes. This infrastructure supports personalized instruction—e.g., adapting content for a student struggling with fractions—while improving interdisciplinary learning outcomes. Pilot studies report a 15% improvement in test scores [22], and content automation tools reduce teachers’ workload by 20%, allowing them to prioritize higher-order pedagogy [4].
However, expanding GEAI’s integration into education intensifies concerns over data privacy and ethical governance. While AES-256 encryption and Trusted Domains provide foundational safeguards, risks such as latent data retention (5% annual leak rate reported by comparable systems) persist [58]. FERPA and GDPR compliance require rigorous privacy-by-design strategies, including k-anonymity (95%) and differential privacy (ε = 1). Moreover, 30% of educators express concern over role displacement, underscoring the importance of human-centered AI implementation. Algorithmic bias—such as grading inaccuracies with 10% error rates—further necessitates transparent and auditable AI systems trained on diverse datasets.
In closing, GEAI offers a promising pathway toward inclusive, transparent, and high-impact AI-driven education. It improves upon traditional platforms like Turnitin in fairness and outperforms MATHia in adaptivity. Future directions should include: (1) empirical testing of GEAI-based compliance auditing tools to achieve 90% accuracy over time; (2) prototyping MM Devices in live classroom environments to evaluate usability, fidelity, and privacy protocols; and (3) engaging with policymakers through open-access findings and ethical briefings to promote trustworthy AI education governance.
Future Validation and Deployment Roadmap
While this study presents GEAI as a conceptual framework, future research will focus on its practical validation and iterative refinement. In the next phase, we plan to conduct:
  • Expert validation workshops with educators, AI ethicists, and institutional IT officers to review framework assumptions, governance workflows, and use case feasibility [63,64].
  • Prototype testing of MM Devices in select classroom settings to assess hardware usability, data fidelity, and local AI inference performance.
  • Pilot studies using mixed-methods (e.g., teacher interviews, student feedback, engagement metrics) to evaluate impact on teaching effectiveness and learner experience [63].
  • Longitudinal trials across multiple institutions, targeting metrics such as academic progress, ethical compliance, and teacher workload reduction [57].
As part of this roadmap, a pilot demonstration is currently being prepared at Guangzhou University to evaluate the feasibility of deploying MM Device prototypes in real-world classroom settings. Each prototype unit consists of a webcam, microphone, and an edge AI processor (Jetson Nano), configured to run lightweight inference models such as Whisper-tiny for speech-to-text and MediaPipe Pose for posture and gesture recognition. These components support introspective applications such as attention tracking and engagement estimation.
To simulate secure data governance, software-based encryption and decryption routines will be employed in lieu of hardware security modules during the trial. The study will include controlled teaching sessions, followed by structured interviews with instructors and students, focusing on usability, perceived instructional benefit, and privacy acceptance. Insights gained from this pilot will inform the technical refinement and institutional integration of GEAI in future phases.
A schematic diagram of the prototype setup is shown in Figure 4, illustrating the edge-based multimodal processing architecture designed for classroom deployment.
Figure 4. Prototype system architecture for the planned pilot study at Guangzhou University. Each node includes a webcam, a microphone, and a Jetson Nano edge processor connected through a secure LAN switch. A laptop is used for control and monitoring during the pilot phase. Software-based encryption simulates secure data handling in alignment with the GEAI trusted domain model.
These validation strategies align with established educational technology research practices—including participatory design, expert review, and pilot testing—and are essential for translating the GEAI framework from theory into practice [64].
Study Limitations
This study presents a conceptual framework that has not yet undergone empirical validation. While the proposed components—such as MM Devices, Trusted Domains, and federated AI pipelines—are grounded in existing technological principles, they remain at the design and simulation stage. Furthermore, the feasibility of implementing such systems in diverse educational settings may be constrained by institutional capacity, data infrastructure, and policy variation. The ethical assumptions embedded in the model (e.g., voluntary data sharing, local control) may require contextual adaptation in practice. Additionally, the lack of direct stakeholder co-design in this initial phase limits the participatory grounding of the framework. These limitations underscore the importance of the future validation roadmap described above and point to the need for iterative, collaborative refinement.

Author Contributions

Conceptualization, Y.Y.; methodology, Y.Y. and H.L.; validation, Y.Y.; writing—original draft preparation, Y.Y., H.L., H.Z., and T.C.; writing—review and editing, Y.Y., H.L., H.Z., T.C., and J.L.; visualization, T.C. and H.Z.; supervision, H.L.; project administration, H.L. and J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Acknowledgments

We thank Heather Zhang for her valuable feedback and technical support throughout the study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. UNESCO Report on the 2022 Transforming Education Summit. Available online: https://www.un.org/sites/un2.un.org/files/report_on_the_2022_transforming_education_summit.pdf (accessed on 21 January 2025).
  2. United Nations General Assembly Transforming Our World: The 2030 Agenda for Sustainable Development. Available online: https://www.refworld.org/legal/resolution/unga/2015/en/111816 (accessed on 21 January 2025).
  3. Connor, C.M. Using Technology and Assessment to Personalize Instruction: Preventing Reading Problems. Prev. Sci. 2019, 20, 89–99. [Google Scholar] [CrossRef] [PubMed]
  4. Demszky, D.; Liu, J.; Mancenido, Z.; Cohen, J.; Hill, H.; Jurafsky, D.; Hashimoto, T. Measuring Conversational Uptake: A Case Study on Student-Teacher Interactions. arXiv 2021, arXiv:2106.03873. [Google Scholar]
  5. Jensen, E.; Dale, M.; Donnelly, P.J.; Stone, C.; Kelly, S.; Godley, A.; D’Mello, S.K. Toward Automated Feedback on Teacher Discourse to Enhance Teacher Learning. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 25–30 April 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1–13. [Google Scholar]
  6. Malik, A.; Wu, M.; Vasavada, V.; Song, J.; Coots, M.; Mitchell, J.; Goodman, N.; Piech, C. Generative Grading: Near Human-Level Accuracy for Automated Feedback on Richly Structured Problems. arXiv 2021, arXiv:1905.09916. [Google Scholar]
  7. Allman, B.; Kimmons, R.; Wang, W.; Bao, H.; Rosenberg, J.M.; Koehler, M.J. Trends and Topics in Educational Technology, 2024 Edition. TechTrends 2024, 68, 402–410. [Google Scholar] [CrossRef]
  8. Veletsianos, G.; Kimmons, R.; Bondah, F. ChatGPT and Higher Education: Initial Prevalence and Areas of Interest. Available online: https://er.educause.edu/articles/2023/3/chatgpt-and-higher-education-initial-prevalence-and-areas-of-interest (accessed on 21 January 2025).
  9. Bommasani, R.; Hudson, D.A.; Adeli, E.; Altman, R.; Arora, S.; von Arx, S.; Bernstein, M.S.; Bohg, J.; Bosselut, A.; Brunskill, E.; et al. On the Opportunities and Risks of Foundation Models. arXiv 2022, arXiv:2108.07258. [Google Scholar]
  10. Frank, E. The Influence of Artificial Intelligence on Education: Enhancing Personalized Learning Experiences. Available online: https://easychair.org/publications/preprint/5Xnl (accessed on 21 January 2025).
  11. Intel Exploring the Potential of Ai in Education. Available online: https://www.delltechnologies.com/asset/en-gb/services/education/educational-training/explore-potential-ai-education-20231023-v8.pdf (accessed on 8 July 2025).
  12. Bateman, J. Deepfakes and Synthetic Media in the Financial System: Assessing Threat Scenarios; Cyber Policy Initiative Working Paper Series; Carnegie Endowment for International Peace: Washington, DC, USA, 2020. [Google Scholar]
  13. Mubarak, R.; Alsboui, T.; Alshaikh, O.; Inuwa-Dutse, I.; Khan, S.; Parkinson, S. A Survey on the Detection and Impacts of Deepfakes in Visual, Audio, and Textual Formats. IEEE Access 2023, 11, 144497–144529. [Google Scholar] [CrossRef]
  14. Murillo-Ligorred, V.; Ramos-Vallecillo, N.; Covaleda, I.; Fayos, L. Knowledge, Integration and Scope of Deepfakes in Arts Education: The Development of Critical Thinking in Postgraduate Students in Primary Education and Master’s Degree in Secondary Education. Educ. Sci. 2023, 13, 1073. [Google Scholar] [CrossRef]
  15. Vaccari, C.; Chadwick, A. Deepfakes and Disinformation: Exploring the Impact of Synthetic Political Video on Deception, Uncertainty, and Trust in News. Soc. Media Soc. 2020, 6, 2056305120903408. [Google Scholar] [CrossRef]
  16. Boutadjine, A.; Harrag, F.; Shaalan, K.; Karboua, S. A Comprehensive Study on Multimedia DeepFakes. In Proceedings of the 2023 International Conference on Advances in Electronics, Control and Communication Systems (ICAECCS), Blida, Algeria, 6–7 March 2023; pp. 1–6. [Google Scholar]
  17. de Ruiter, A. The Distinct Wrong of Deepfakes. Philos. Technol. 2021, 34, 1311–1332. [Google Scholar] [CrossRef]
  18. Egbuna, O.P. The Impact of AI on Cybersecurity: Emerging Threats and Solutions. J. Sci. Technol. 2021, 2, 43–67. [Google Scholar]
  19. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MN, USA, 2–7 June 2017; Burstein, J., Doran, C., Solorio, T., Eds.; Association for Computational Linguistics: Stroudsburg, PA, USA, 2019; Volume 1 (Long and Short Papers), pp. 4171–4186. [Google Scholar]
  20. Brown, T.B.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language Models Are Few-Shot Learners. Adv. Neural Inf. Process. Syst. 2020, 33, 1877–1901. [Google Scholar]
  21. Ramesh, A.; Pavlov, M.; Goh, G.; Gray, S.; Voss, C.; Radford, A.; Chen, M.; Sutskever, I. Zero-Shot Text-to-Image Generation. In Proceedings of the 38th International Conference on Machine Learning, Virtual Event, 18–24 July 2021; pp. 8821–8831. [Google Scholar]
  22. Holmes, W.; Tuomi, I. State of the Art and Practice in AI in Education. Eur. J. Educ. 2022, 57, 542–570. [Google Scholar] [CrossRef]
  23. Bommasani, R.; Klyman, K.; Longpre, S.; Kapoor, S.; Maslej, N.; Xiong, B.; Zhang, D.; Liang, P. The Foundation Model Transparency Index. arXiv 2023, arXiv:2310.12941. [Google Scholar]
  24. Mollick, E.R.; Mollick, L. Using AI to Implement Effective Teaching Strategies in Classrooms: Five Strategies, Including Prompts 2023. Available online: https://ssrn.com/abstract=4391243 (accessed on 21 January 2025).
  25. Neumann, A.T.; Yin, Y.; Sowe, S.; Decker, S.; Jarke, M. An LLM-Driven Chatbot in Higher Education for Databases and Information Systems. IEEE Trans. Educ. 2024, 68, 103–116. [Google Scholar] [CrossRef]
  26. Ruff, E.F.; Franz, J.L.; West, J.K. Using ChatGPT for Method Development and Green Chemistry Education in Upper-Level Laboratory Courses. J. Chem. Educ. 2024, 101, 3224–3232. [Google Scholar] [CrossRef]
  27. MATHia. Available online: https://www.carnegielearning.com/solutions/math/mathia/ (accessed on 21 January 2025).
  28. Zhang, H.; Gao, M.; Gan, Z.; Dufter, P.; Wenzel, N.; Huang, F.; Shah, D.; Du, X.; Zhang, B.; Li, Y.; et al. MM1.5: Methods, Analysis & Insights from Multimodal LLM Fine-Tuning. arXiv 2024, arXiv:2409.20566. [Google Scholar]
  29. Mishra, P.; Koehler, M.J. Technological Pedagogical Content Knowledge: A Framework for Teacher Knowledge. Teach. Coll. Rec. 2006, 108, 1017–1054. [Google Scholar] [CrossRef]
  30. European Union High-Level Expert Group on Artificial Intelligence. Ethics Guidelines for Trustworthy AI; European Commission: Brussels, Belgium, 2019. [Google Scholar]
  31. OECD Legal Instruments. OECD Recommendation of the Council on Artificial Intelligence. 2019. Available online: https://legalinstruments.oecd.org/en/instruments/oecd-legal-0449 (accessed on 21 May 2025).
  32. PCI Security Standards Council PCI P2PE. Program Guide. 2020. Available online: https://listings.pcisecuritystandards.org/documents/PCI-SSC_P2PE_Program_Guide_v3.0r1.0.pdf (accessed on 21 May 2025).
  33. Ravi, S.; Raghunathan, A.; Chakradhar, S. Tamper Resistance Mechanisms for Secure Embedded Systems. In Proceedings of the 17th International Conference on VLSI Design, Mumbai, India, 9 January 2004; pp. 605–611. [Google Scholar]
  34. Hasan, M.; Islam, M.M.; Zarif, M.I.I.; Hashem, M.M.A. Attack and Anomaly Detection in IoT Sensors in IoT Sites Using Machine Learning Approaches. Internet Things 2019, 7, 100059. [Google Scholar] [CrossRef]
  35. Chen, G.; Xu, B.; Lu, M.; Chen, N.-S. Exploring Blockchain Technology and its Potential Applications for Education. Smart Learn. Environ. 2018, 5, 1–10. [Google Scholar] [CrossRef]
  36. Ferrag, M.A.; Derdour, M.; Mukherjee, M.; Derhab, A.; Maglaras, L.; Janicke, H. Blockchain Technologies for the Internet of Things: Research Issues and Challenges. IEEE Internet Things J. 2019, 6, 2188–2204. [Google Scholar] [CrossRef]
  37. D’Mello, S.K. Improving Student Engagement in and with Digital Learning Technologies. In Pushing the Frontiers with Artificial Intelligence, Blockchain and Robots; OECD Publishing: Paris, France, 2021; pp. 79–104. [Google Scholar]
  38. Ravi, S.; Raghunathan, A.; Kocher, P.; Hattangady, S. Security in Embedded Systems: Design Challenges. ACM Trans. Embed. Comput. Syst. 2004, 3, 461–491. [Google Scholar] [CrossRef]
  39. FIPS 140-2; Security Requirements for Cryptographic Modules. NIST: Gaithersburg, MD, USA, 2001.
  40. NIST SP 800-57; Recommendation for Key Management. NIST: Gaithersburg, MD, USA, 2007.
  41. ISO/IEC 27001:2013; Information Security Management. ISO: Geneva, Switzerland, 2013.
  42. GDPR; General Data Protection Regulation. European Commission: Brussels, Belgium, 2016.
  43. FERPA; Family Educational Rights and Privacy Act. U.S. Department of Education: Washington, DC, USA, 1974.
  44. RFC 8446; The Transport Layer Security (TLS) Protocol Version 1.3. IETF: Fremont, CA, USA, 2018.
  45. PCI P2PE; Point-to-Point Encryption Standard v3.0. PCI Security Standards Council: Wakefield, MA, USA, 2020.
  46. Zhai, X.; Shi, L.; Nehm, R.H. A Meta-Analysis of Machine Learning-Based Science Assessments: Factors Impacting Machine-Human Score Agreements. J. Sci. Educ. Technol. 2021, 30, 361–379. [Google Scholar] [CrossRef]
  47. McMahan, B.; Moore, E.; Ramage, D.; Hampson, S.; Arcas, B.A. Communication-Efficient Learning of Deep Networks from Decentralized Data. In Proceedings of the International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, FL, USA, 20–22 April 2017; pp. 1273–1282. [Google Scholar]
  48. Abu-Salih, B. Domain-Specific Knowledge Graphs: A Survey. J. Netw. Comput. Appl. 2021, 185, 103076. [Google Scholar] [CrossRef]
  49. Ji, S.; Pan, S.; Cambria, E.; Marttinen, P.; Yu, P.S. A Survey on Knowledge Graphs: Representation, Acquisition, and Applications. IEEE Trans. Neural Netw. Learn. Syst. 2021, 33, 494–514. [Google Scholar] [CrossRef]
  50. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer. J. Mach. Learn. Res. 2020, 21, 1–67. [Google Scholar]
  51. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł; Polosukhin, I. Attention is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  52. Sun, K.; Dredze, M. Amuro & Char: Analyzing the Relationship between Pre-Training and Fine-Tuning of Large Language Models. arXiv 2024, arXiv:2408.06663. [Google Scholar]
  53. Alayrac, J.-B.; Donahue, J.; Luc, P.; Miech, A.; Barr, I.; Hasson, Y.; Lenc, K.; Mensch, A.; Millican, K.; Reynolds, M.; et al. Flamingo: A Visual Language Model for Few-Shot Learning. arXiv 2022, arXiv:2204.14198. [Google Scholar]
  54. Lu, J.; Clark, C.; Zellers, R.; Mottaghi, R.; Kembhavi, A. Unified-IO: A Unified Model for Vision, Language, and Multi-Modal Tasks. arXiv 2022, arXiv:2206.08916. [Google Scholar]
  55. Ogunleye, B.; Zakariyyah, K.I.; Ajao, O.; Olayinka, O.; Sharma, H. A Systematic Review of Generative AI for Teaching and Learning Practice. Educ. Sci. 2024, 14, 636. [Google Scholar] [CrossRef]
  56. Gârdan, I.P.; Manu, M.B.; Gârdan, D.A.; Negoiță, L.D.L.; Paștiu, C.A.; Ghiță, E.; Zaharia, A. Adopting AI in Education: Optimizing Human Resource Management Considering Teacher Perceptions. Front. Educ. 2025, 10, 1488147. [Google Scholar] [CrossRef]
  57. Wang, Y.; Jiang, X. AI for Education: Trends and Insights. The Innovation 2025, 6, 100872. [Google Scholar] [CrossRef] [PubMed]
  58. Bower, M.; Torrington, J.; Lai, J.W.M.; Petocz, P.; Alfano, M. How Should We Change Teaching and Assessment in Response to Increasingly Powerful Generative Artificial Intelligence? Outcomes of the ChatGPT Teacher Survey. Educ. Inf. Technol. 2024, 29, 15403–15439. [Google Scholar] [CrossRef]
  59. Yan, Y.; Liu, H. Ethical Framework for AI Education Based on Large Language Models. Educ. Inf. Technol. 2024, 30, 10891–10909. [Google Scholar] [CrossRef]
  60. Vygotsky, L.S.; Michael, C. Mind in Society: Development of Higher Psychological Processes; Harvard University Press: Cambridge, MA, USA, 1978. [Google Scholar]
  61. Duffy, T.M.; Cunningham, D.J. Constructivism: Implications for the Design and Delivery of Instruction. In Handbook of Research for Educational Communications and Technology; Jonassen, D.H., Ed.; Simon & Schuster Macmillan: New York, NY, USA, 1996; pp. 170–198. [Google Scholar]
  62. Nemecek, D.; Cigna U.S. Loneliness Index. 2018. Available online: https://www.multivu.com/players/English/8294451-cigna-us-loneliness-survey/docs/IndexReport_1524069371598-173525450.pdf (accessed on 14 February 2025).
  63. Holmes, W.; Bialik, M.; Fadel, C. Artificial Intelligence in Education: Promises and Implications for Teaching and Learning; The Center for Curriculum Redesign: Boston, MA, USA, 2019; ISBN 978-1-7942-9370-0. [Google Scholar]
  64. Dede, C.; Jass Ketelhut, D.; Whitehouse, P.; Breit, L.; McCloskey, E.M. A Research Agenda for Online Teacher Professional Development. J. Teach. Educ. 2009, 60, 8–19. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.