You are currently viewing a new version of our website. To view the old version click .
Algorithms
  • Article
  • Open Access

26 December 2025

A Unified AI Architecture for Self-Regulated Learning: Cognitive Modeling, Meta-Learning, and Continual Adaptation

,
,
and
1
LSATE Laboratory, National School of Applied Science, Sidi Mohamed Ben Abdellah University, Fez 30000, Morocco
2
L3IA Laboratory, Faculty of Sciences Dhar El Mehraz, Sidi Mohamed Ben Abdellah University, Fez 30000, Morocco
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Emerging Trends in Distributed AI for Smart Environments

Abstract

The growing need for intelligent educational systems calls for architectures supporting adaptive instruction, while enabling more permanent, long-term personalization and cognitive alignment in the long run. While we have seen progress in adaptive learning technologies at the intersection of Self-Regulated Learning (SRL), Continual Learning (CL), and Meta-Learning, these are generally employed in isolation to provide piecemeal solutions. In this paper, we propose CAMEL, a unified architecture for (1) cognitive modelling based on SRL, (2) continual learning functionalities, and (3) meta-learning to provide adaptive, personalized, and cognitively consistent learning environments. CAMEL includes the following components: (1) A Cognitive State Estimator that estimates learner motivation, attention, and persistence from behavioral traces, (2) A Meta-Learning Engine that allows it rapid adaptation through Model-Agnostic Meta-Learning (MAML), (3) A Continual Learning Memory that preserves knowledge across sessions using Elastic Weight Consolidation (EWC) and Replay, (4) A Pedagogical Decision Engine that makes real-time efficient adjustments of instructional strategies, and (5) A closed-loop that continuously reconciles misalignments between pedagogical actions and predicted cognitive states. Experiments conducted on the xAPI-Edu-Data dataset evaluate the system’s few-shot adaptation capability, knowledge retention, cognitive-state prediction accuracy, and knowledge, as well as cognitive responsiveness to the impending questions. It offers competitive performance in learner-state prediction and long-term performance compared to the baselines, and the improvements are consistent across the different baselines. This paper lays the groundwork for next-generation adaptive and cognition-driven AI-based learning systems.

1. Introduction

Educational technology is changing fast, and AI is the new way to create a smart, personalized, and adaptive environment for learning. Despite present advancements in areas like intelligent tutoring systems (ITS), learning analytics, and affective computing, numerous existing educational technologies do not provide iterative adjustments accounting for the shifts in cognitive, motivational, and emotional experience (over time) [1,2]. A growing body of research indicates that Self-Regulated Learning (SRL) is perhaps one of the best predictors of academic performance in these learner-led, digital environments [3,4]. SRL entails higher-order processes such as planning, self-monitoring, and reflection that are mediated by motivational factors and cognitive flexibility [5,6].
In AI-based educational systems, Self-Regulated Learning can be operationalized by mapping observable behavioral traces to latent cognitive states. In this work, SRL is modeled through learner interaction features such as resource access, task completion, participation, and response timing, which are processed by a Cognitive State Estimator to infer motivation, attention, and persistence. These inferred cognitive states are then used as inputs to adaptive learning components, enabling personalized feedback, pacing, and instructional strategy selection. This computational formulation allows SRL to be integrated into an end-to-end AI system rather than remaining a purely conceptual construct.
A computer-system generality that covers a broad spectrum of computational frameworks is complemented with the capability of fast generalization within the task and learner profile from a few samples [7,8], which entails learning to learn meta-learning. Moreover, continuous learning (CL) addresses the challenge of catastrophic forgetting by allowing systems to learn incrementally and retain knowledge over time, a crucial property for educational settings with continuously changing learner cohorts [9]. At the same time, existing systems have only considered SRL, CL, and meta-learning as isolated boxes, resulting in pieces of architectures that are disjointed and cognitively incoherent, and not necessarily meant to be long-term adaptable. To overcome this limitation, we propose a unified architecture that leverages meta-learning to facilitate rapid personalization, continual learning to support life-long learning, and self-regulated learning modelling to align the cognitive skills with the learning process. We propose a system with roots in cognitive theory [10] that employs reinforcement learning [11], natural language processing [12], and multimodal analytics for support to maintain the dynamic, personalized, and ethical nature of adaptive learning pathways.
Various decision-making models have been explored in intelligent educational systems, including support vector machines (SVM), convolutional neural networks (CNN), and transformer-based architectures. While these models are effective for static or data-intensive prediction tasks, they often require large amounts of labeled data and provide limited interpretability in dynamic educational settings. In contrast, the proposed CAMEL architecture adopts a hybrid decision-making approach that combines cognitively grounded rules with reinforcement learning-based policy optimization. This choice enables interpretable, adaptive, and data-efficient pedagogical decisions that can evolve over time in response to changing learner states. Similar hybrid decision strategies have recently been shown to be effective in adaptive and distributed AI systems [13].
This paper makes several key contributions to the field of intelligent educational systems. First, we propose a unified AI-based architecture that jointly integrates self-regulated learning-based cognitive modeling, meta-learning, and continual learning within a single closed-loop framework. Second, we introduce a Cognitive State Estimator that infers learners’ motivation, attention, and persistence from behavioral traces, enabling cognitively aligned personalization. Third, the use of meta-learning allows rapid few-shot adaptation to new learners and learning contexts under low-data conditions. Fourth, continual learning mechanisms based on Elastic Weight Consolidation and replay are incorporated to mitigate catastrophic forgetting and support long-term adaptation. Finally, the proposed architecture is empirically evaluated using the xAPI-Edu-Data dataset and a complementary EdNet-like dataset, demonstrating improvements in cognitive state prediction, few-shot adaptation, and long-term performance stability.

3. Research Gap and Contribution

In this paper, we argue that the advances being made regarding each of these paradigms providing the input to an adaptive educational system are substantial, but at the same time, research on them has hardly been integrated into one coherent architecture. The learning and modeling of Self-Regulated Learning (SRL) has been extensively employed and scaffolded in intelligent tutoring systems like MetaTutor, where multimodal traces are used to provide real-time SRL support [1,12]. Additional work addresses prescriptive analytics of SRL strategies based on reinforcement learning policies [10] or identification of metacognitive and SRL-related patterns from text produced by learners in an NLP-based diagnostic fashion [11]. Though such systems provide comprehensive SRL-focused support, the former do not employ formal continual learning mechanisms, and the latter do not engage meta-learning algorithms for few-shot personalization.
At the same time, meta-learning has been suggested as a fast personalization solution in low-data scenarios by the capability of learning to quickly adapt to novel learner profiles and tasks [8]. While these approaches realize how “learning to learn” can help education, they are usually not guided by a clear generic SRL model and are not designed with longevity in mind (i.e., forgetting of knowledge and policies over time can lead to reduced transfer to novel tasks). Caveats: Likewise, continuous learning techniques such as Elastic Weight Consolidation (EWC) and replay-based methods have been examined in the hopes of reducing catastrophic forgetting and enabling lifelong adaptation in AI-based learning environments [9], yet are generally used without any learner modelling that is explicit in cognitive or SRL.
As a result, there is fragmentation in the literature, with SRL-focused systems that highlight the cognitive and metacognitive processes at their core, but often without lifelong memory and rapid cross-learner adaptation, meta-learning approaches that achieve few-shot personalization but rarely model explicit SRL constructs or incorporate continual learning (CL), and CL research that efficiently secures long-term retention and representation adequacy, but frequently overlooks the cognitive and motivational dynamics that SRL often aims to capture [18]. Neither of these two streams of research is aimed at taking this operationalized SRL modeling, meta-learning, and continual learning jointly in a single closed-loop architecture to target both short-term and long-term personalization in intelligent educational systems, as far as we know.
To fill this gap, we present CAMEL, the first fully integrated intelligent tutoring system containing (1) a Cognitive State Estimator based on an SRL theory that infers motivation, attention, and persistence from behavior traces, (2) a Meta-Learning Engine based on MAML and related algorithms for few-shot adaptation to new learners and tasks, and (3) a Continual Learning Memory based on a combination of EWC and replay and across sessions and topics. We empirically evaluate CAMEL on the xAPI-Edu-Data dataset, complemented by an EdNet-like sample, showing superior cognitive state prediction, few-shot adaptation, and long-term performance stability compared to baselines. Moreover, an integration case study with an LMS showcases the practicality of deploying CAMEL in real-world learning environments. These contributions combine to make CAMEL a new, cognitively motivated architecture that links prior islands of research in modelling of SRL, meta-learning, and continual learning for IESs.

4. Our Conceptual Architecture

4.1. Learner Cognitive Model

The design of the proposed system is viewed through the lens of Self-Regulated Learning (SRL) theory, which conceptualizes learning as an engaging, circular process of planning, self-regulation, and reflection. These processes are thought to be mediated by certain motivational factors, e.g., self-efficacy and goal orientation, that serve as direct predictors of learner engagement, persistence, and metacognitive control [4,5]. SRL is especially important in digital learning environments where learners work with a high degree of autonomy and with asynchronous interactions with content and feedback mechanisms [3]. We represent apparatus based on the SRL framework, which allows the ongoing estimation of cognitive states from observable variables such as response delay, accuracy, self-assessment response outcomes, and behavior patterns. As supported by [12], multimodal traces are, accordingly, within the bounds of a reasonable proxy for a range of internal cognitive and emotional states, provided they are modelled correctly. This adaptive learner model creates a foundation for the adaptivity of our system by allowing pedagogical choices to adapt to the changing demands of the learner.
But recent progress in the modeling of SRL has shown that when the multimodal traces that combine behavioral, temporal, affective, and self-reported data are modeled appropriately, they can be used as reliable proxies for internal cognitive and motivational states [1]. This reinforces our use of behavioral interaction logs to estimate latent cognitive variables such as attention, motivation, and persistence.

4.2. Proposed Intelligent Educational System

To translate this theoretical foundation into a usable architecture, we proposed a modular intelligent educational system that stacks three distinct paradigms of AI: continual learning (CL), meta-learning, and SRL modelling. There are four main components of the system:
  • Cognitive State Estimator: This module uses sequence-based neural architectures (e.g., LSTMs, Transformers) to extract latent cognitive variables from learner interactions. Existing research has shown that such models can support dynamic learner-state modelling [11,12].
  • Model Learner: The model incorporates model-agnostic meta-learning (MAML) and related algorithms like Reptile to ensure that new learners and instructional tasks can be adjusted efficiently. Under low-data settings, these algorithms show utility in educational personalization [8].
  • Contrary to this, Continual Learning Memory (CLM) is a module that retains long-term learning ability to avoid catastrophic forgetting with factors such as Elastic Weight Consolidation (EWC) and Gradient Episodic Memory (GEM). This has applicability mainly in lifelong learning cases, as observed on educational platforms [9].
The decision engine uses outputs from the cognitive state estimator, the meta-learning engine, and the CL memory to decide which pedagogical action to take. Our scalable engine scaffolds are based on a cognitive profile by altering content, feedback, and task difficulty in real time. Reinforcement learning is increasingly being used to operationalize SRL as a sequential decision-making process, enabling systems to recommend optimal regulation strategies dynamically based on learner feedback and behavioral evolution [10].
Such a synergy of the modules affords a system that can be adaptively short-term and resilient over time, two key components that are often found to be lacking in current educational AI systems [1].

4.3. Integration of Paradigms

Figure 1 shows the innovation summary of our system. Abstractly, our system integrates the modelling of SRL, continual learning, and meta-learning in one adaptive architecture. Previous systems only tackle one or two of the components in isolation, but our model forms a closed-loop learning cycle where:
  • The Cognitive State Estimator interprets real-time learner behavior.
  • The Meta-Learning Engine modifies learning strategies to fit new profiles or tasks.
  • The Continual Learning Module is responsible for maintaining the previously learnt behavior and policies learnt from previous sessions.
  • The Decision Engine delivers context-appropriate pedagogical actions.
Figure 1. Conceptual architecture of the proposed intelligent educational system.
Every learner interaction iteratively hones in on this loop and thus can be used by the system to personalize instruction within and across sessions. This form of temporal personalization directly targets one of the central challenges in adaptive education: maintaining alignment between the pedagogical response and the learner’s cognition as time passes [10]. It dynamically deals with cases like motivation drops, task complexity changes, and learner engagement changes. In contrast to previous systems like [1] that emphasize short-term SRL scaffolding without long-term memory or rapid adaptation to learner differences, our system integrates cognitive responsiveness and strong learning dynamics, enabling it to be implemented at scale in innovative education ecosystems.
The ethical concerns identified in prior work, including algorithmic bias, lack of transparency, and overreliance on automation, are partially addressed in the proposed architecture. CAMEL emphasizes interpretability through explicit cognitive state estimation and rule-informed pedagogical decision-making, allowing human-understandable reasoning rather than opaque automation. Furthermore, the system is designed to support learner agency by adapting instructional strategies without removing learner control or autonomy. While a full ethical evaluation is beyond the scope of this study, the architecture is designed with transparency, accountability, and human-centered adaptation as guiding principles.

5. Methodology

5.1. Implemented Model

The proposed intelligent educational system was designed and implemented as a modular architecture consisting of four core components: (1) the Cognitive State Estimator, (2) the Meta-Learning Engine, (3) the Continual Learning Memory, and (4) the Pedagogical Decision Engine. Given that each module has its functionality, they work together seamlessly to provide a personalized and adaptive learning experience as learners develop their cognitive profiles over time.
The proposed CAMEL architecture is composed of four interconnected modules, each fulfilling a distinct role within the adaptive learning loop. The Cognitive State Estimator processes learner interaction data to infer latent cognitive states such as motivation, attention, and persistence. The Meta-Learning Engine enables rapid adaptation to new learners and tasks by optimizing model parameters under few-shot conditions. The Continual Learning Memory preserves previously acquired knowledge across sequential learning tasks, mitigating catastrophic forgetting. Finally, the Pedagogical Decision Engine integrates cognitive state estimates and learning outputs to dynamically select instructional actions. Together, these modules form a closed-loop system that supports adaptive, cognitively aligned personalization.
Cognitive State Estimator: The first module employs a transformer-based neural network to predict latent cognitive variables such as attention, motivation, and persistence. Input: time-series features of the learners based on raised hands, visited resources, discussion, and student absence days. First, we created a composite Engagement Index by normalizing (using z-scores) and averaging over behavioral metrics. To simulate session-based learning, the input sequences were segmented based on subject-topic combinations.
Meta-Learning Engine: We implemented the Model-Agnostic Meta-Learning (MAML) algorithm using the learn2learn library to enable rapid adaptation to new learners with limited data. Reptile, a first-order approximation of MAML, was also tested as a computationally efficient baseline. The meta-learner optimizes instructional strategies based on few-shot learner behavior, allowing the system to personalize feedback early in the learning process.
Continual Learning Memory: For sequential adaptation, we integrated Elastic Weight Consolidation (EWC) and Gradient Episodic Memory (GEM) using the Avalanche library. These approaches allowed the system to retain knowledge across multiple instructional topics (e.g., Math, Biology) while incorporating new learner experiences without catastrophic forgetting.
Pedagogical Decision Engine: This module synthesizes outputs from the cognitive state estimator and learning components to decide on the next pedagogical action. The system dynamically adjusts feedback, content difficulty, and pacing based on the learner’s real-time cognitive state. A hybrid policy network was trained using reinforcement learning to support this decision-making process in conjunction with rule-based heuristics.
The Pedagogical Decision Engine integrates the inferred cognitive states and the outputs of the learning components through a hybrid decision mechanism. Specifically, cognitive indicators such as attention, motivation, and persistence are combined with performance metrics and adaptation signals produced by the meta-learning and continual learning modules. These inputs are jointly evaluated using a rule-informed decision logic, where predefined pedagogical rules guide the selection of instructional actions, optionally supported by reinforcement learning-based policy optimization. This design enables reproducible and interpretable pedagogical decisions while maintaining adaptability to evolving learner states.

5.2. Data Used

The experiments were conducted using the publicly available xAPI-Edu-Data dataset from Kaggle, which contains detailed learning records of 480 high school students. The dataset includes both demographic features (e.g., gender, nationality, ParentAnswering) and behavioral indicators relevant to self-regulated learning, including:
  • Engagement metrics: raisedhands, VisitedResources, Discussion
  • Attendance: StudentAbsenceDays
  • Academic progression: StageID, GradeID, Topic
  • Target label: Class (student final performance: Low, Middle, High)
To simulate temporal learning, data were grouped by student_id and Topic, forming task sequences reflecting learning sessions. Cognitive traits were inferred from engineered features such as an Engagement Index, computed as a normalized mean of raisedhands, VisitedResources, and Discussion. The dataset was cleaned, one-hot encoded for categorical variables, and z-score normalized for numerical ones. Synthetic perturbations (e.g., motivation drops, performance plateaus) were introduced in controlled profiles to test the robustness of cognitive-aware adaptation.
We also provided a complementary dataset, EdNet (2020), which consists of one of the largest modern students’ problem interaction corpora, in addition to xAPI-Edu-Data. Additionally, EdNet contains large-scale sequential traces that can be used to evaluate compatibility with contemporary learninganalytics pipelines. We present the detailed analysis of this dataset in the Results Section.
The study uses two publicly available datasets:
xAPI-Edu-Data: a behavioral and performance dataset for SRL research, available at https://www.kaggle.com/datasets/aljarah/xapi-edu-data (accessed on 15 December 2025).
EdNet KT1 Dataset: a large-scale student–problem interaction dataset released by Riiid, available at https://github.com/riiid/ednet (accessed on 14 December 2025).
These datasets were selected because they provide complementary properties for evaluating cognitive-state estimation, meta-learning adaptation, and continual-learning stability.

5.3. Experimental Protocol

We designed three experimental scenarios to evaluate each component of the system:
  • Cold-Start Personalization
Objective: Evaluate the meta-learning engine’s few-shot adaptation to new learners.
Method: Train the meta-learner on N − 1 learners, then test on the Nth with five interaction examples.
  • Continual Adaptation Across Tasks
Objective: Assess the system’s ability to retain knowledge over sequential topics.
Method: Simulate task sequences based on different Topic values (e.g., Biology → Chemistry → Math), measuring performance retention on earlier tasks.
  • Cognitive-Aware Responsiveness
Objective: Measure how well the system adjusts pedagogical strategies in response to inferred cognitive states.
Method: Induce synthetic disengagement by lowering VisitedResources and Discussion, and evaluate the system’s feedback effectiveness.
All experiments were conducted using a 5-fold cross-validation setup. For each scenario, we compared:
  • The proposed adaptive system (CL + Meta + SRL)
  • A baseline deep learning model without adaptation
  • A rule-based ITS without cognitive awareness
Each model was trained and tested over five simulated sessions per learner.

5.4. Evaluation Metrics

System performance was evaluated along three dimensions:
A. Continual Learning
Accuracy Over Time: performance stability across tasks
Forgetting Rate: loss in accuracy on previous topics after learning new ones
Stability: prediction variance across sessions
B. Meta-Learning
Few-Shot Accuracy: accuracy after five interaction samples
Adaptation Speed: number of gradient updates to convergence
Generalization Gap: performance difference between seen and unseen learners
C. Cognitive Responsiveness
Engagement Index: ratio of high-engagement interactions
Progress Score: improvement in performance across sessions
Regulation Quality Index: match between feedback provided and learner’s cognitive needs

5.5. Tools and Environment

The system was implemented using the following technologies:
  • Programming language: Python 3.10
  • Libraries:
    o
    PyTorch v2.0.1 for model development
    o
    Learn2learn v0.2.0 for meta-learning routines
    o
    Avalanche v0.3.1 for continual learning components
    o
    Scikit-learn v1.3.0, Pandas v2.0.3, Matplotlib v3.7.2 for preprocessing, analysis, and visualization
  • Execution environment:
    o
    Machine: Lenovo i7, 32 GB RAM, NVIDIA RTX 3080 (10 GB)
    o
    OS: Ubuntu 22.04
    o
    Experiments were containerized using Docker v24.0.2 for reproducibility and GPU acceleration.
All code, models, and preprocessed data will be made available on a public GitHub (2025) repository upon publication to ensure transparency and reproducibility.

5.6. Prototype Feasibility and LMS Integration Scenario

We created an integration scenario to evaluate the potential for employing CAMEL in real-world educational environments by modelling how the system would function in an LMS such as Moodle. Under such a scenario, the typical interaction traces (e.g., resource visit, attempted quizzes, and discussion participation) generated by an LMS are processed by the Cognitive State Estimator of CAMEL to predict engagement-related indicators. The learner model is updated by the Meta-Learning and Continual Learning modules, and the Pedagogical Decision Engine generates adaptive recommendations that are to be delivered to the learner through the LMS interface.
This example illustrates how CAMEL can seamlessly integrate into an LMS workflow and function in quasi-real time with generic learner interaction data. While this study does not represent a full deployment around live student data, it is indicative of the potential for technical integration for such a deployment. To evaluate the pedagogical impact of the proposed system, we have planned a pilot study with real learners as future work.

6. Results and Evaluation

6.1. Cognitive State Modeling (xAPI-Edu-Data)

In intelligent educational systems, the ability to estimate a learner’s cognitive state (e.g., motivation, attention, cognitive load) is critical for enabling personalized and adaptive pedagogical responses. The first module of our architecture, the Cognitive State Estimator, is designed to infer latent cognitive indicators from observable behavioral traces such as class participation, resource engagement, and discussion activity.
We developed a machine learning pipeline that categorizes students into either Low, Middle, or High performance levels. The first evaluation investigates the Cognitive State Estimator, which estimates the state of learner engagement (three levels: Low, Middle, and High) associated with cognitive engagement-related states from learning behavior features. As you can see from the confusion matrix (Figure 2), accuracy is very high on the Middle class, while overlap is moderate between Middle and High, as may be expected due to the mixing of motivational and behavioral patterns evident in many of the more typical indicators of self-regulated learning [1,12].
Figure 2. Confusion matrix showing predicted vs. true cognitive performance classes (Low, Middle, High).
In the class-level F1-scores, we observe good generalization in all classes (Low ≈ 0.91, Middle ≈ 0.86, High ≈ 0.77). In Figure 3, the horizontal axis represents the learner performance classes Low (L), Middle (M), and High (H), which are used as proxies for inferred cognitive states. Overall, these results support the hypothesis that cognitive engagement and persistence (as found in the posttest) are adequate proxies for xAPI behavioral traces.
Figure 3. Bar chart visualizing the precision, recall, and F1-score for each class label, reflecting the model’s cognitive state estimation performance.
For the meta-learning and continual learning modules, the good classification performance lays the basis for the downstream personalization.

6.2. Rapid Adaptation and Continual Learning Performance

6.2.1. Few-Shot Adaptation (Meta-Learning)

We tested the MAML-based meta-learning engine on few-shot tasks to reflect a realistic scenario in which only a very limited amount of labeled data-per-learner is available. The confusion matrix (Figure 4) shows that the model generalizes to new learners well despite the small sample size.
Figure 4. Confusion Matrix of Few-shot Task Predictions.
Performance per class (Figure 5) is well-balanced between Class 0 and Class 1, with Class 2 showing a slightly lower precision and may benefit from more examples in the support set.
Figure 5. Precision, Recall, and F1-Score per Class.
These findings support the application of meta-learning to allow for fast personalization in ITS scenarios where learner profiles change over time rapidly.

6.2.2. Continual Learning Stability (EWC vs. Replay)

We then compared EWC and Replay, two different continual learning strategies, on sequential tasks for continual learning to evaluate long-term stability.
A multi-panel combined plot (Figure 6) where all figures are as follows:
EWC: gradual accuracy improvement and stable loss reduction
Replay: higher final accuracy and better retention across tasks
Figure 6. Continual Learning Comparison: EWC vs. Replay.
At least partly responsible for this was Replay’s episodic-memory mechanism, which trained a small buffer of in-the-wild previously seen states and has previously been shown to be effective at retaining cognitive information as learner states evolved over time in continual learning literature [9].
By combining meta-learning and continual learning, CAMEL affords fast adaptation and long-term stability, desirable properties for realistic deployments in education.

6.2.3. Pedagogical Decision-Making and Cognitive Responsiveness

The Pedagogical Decision Engine Converts estimated cognitive states to adaptive instructional actions. Figure 7 represents the shifts of attention, motivation, and forgetting over the sessions, with most periods of positive transfers and some periods of negative transfers.
Figure 7. Cognitive State Evolution Across Sessions.
The instruction adjustments that are activated by these cognitions are depicted in Figure 8. Interactive activities or motivational prompts were triggered by low-attention episodes; by high-performance periods, participants were moved to more advanced tasks.
Figure 8. Pedagogical Actions Triggered Across Sessions.
The results in this paper illustrate that CAMEL can be pedagogically responsive and model interventions that are in alignment with students’ SRL [1,10].

6.2.4. Cross-Dataset Validation Using an EdNet-like Sample

To confirm the generalizability of the CAMEL architecture beyond the EdNet KT1 dataset and to a modern, large-scale learning platform, we additionally performed an exploratory analysis using an EdNet-like sample as per the documented structure of the EdNet KT1 dataset. They compare their results with the state-of-the-art in knowledge tracing, sequence modelling, and from both adaptive educational systems and recent contemporary benchmarks, EdNet. Comprising more than 131 million student–problem interactions, it features a rich temporal structure ideal for continual and meta-learning.
The first description analysis is the distribution of correctness, revealing a moderate right skew towards correct responses (≈65%). Real large-scale tutoring systems exhibit this pattern as the tasks should be easy enough, but still provide enough variance for cognitive modeling (Figure 9).
Figure 9. Distribution of Correctness in the EdNet-like Sample.
In addition, response-time distributions provide more information about the variability in cognitive load and engagement. The log-transformed histogram not only has a very close semi-Gaussian shape, but also indicates the highly exponential character of raw elapsed-time values. These types of distributions are prevalent in large-scale educational datasets and inform words that meaningfully signal differences in strategy, hesitation, or elements of cognitive difficulty that are a part of CAMEL’s cognitive state estimator (See Figure 10).
Figure 10. Log-Transformed Histogram of Elapsed Time.
We went one step deeper and looked at the number of interactions per learner per day (Figure 11). It is a heavy-tailed distribution in that, whereas most of the learners have around 120–180 interactions, more learners are much more active. This feature makes this property ideal for continual learning techniques (e.g., EWC, Replay) since it provides long shapely sequences for some learners, while also providing short sequences for fast-adapting analysis.
Figure 11. Distribution of Interactions per learner.
Finally, a sparse learner problem interaction heatmap illustrates the structural sparsity present in large educational datasets, where each learner interacts with only a small fraction of the available item pool. This sparsity strengthens the necessity of the core mechanisms of representation learning, episodic memory, and adaptive meta-learning encoded in CAMEL (Figure 12).
Figure 12. Sparse Heatmap of Learner–Problem Interactions.
Overall, the complementary analysis based on the EdNet-like sample illustrates that CAMEL architecture can be updated to fit within contemporary large-scale learning analytics datasets. That is, these behavioral patterns emerge as heterogeneous activity levels, heavy-tailed response times, and sparse learner–problem interactions provide shared features to confirm the adequacy of CAMEL’s cognitive modeling, meta-learning, and continual learning components. These results provide support that the proposed system can draw generalization from older benchmark datasets, such as xAPI-Edu-Data, onto the current educational data ecosystem. The following describes the experimental procedure that was used to assess the system.
To further position CAMEL with respect to state-of-the-art approaches, we compare its performance against baseline models that do not incorporate meta-learning or continual learning mechanisms. These baselines include standard supervised classifiers for cognitive state estimation and sequential learning models trained without explicit forgetting mitigation. The results show that CAMEL achieves faster adaptation under few-shot conditions and maintains higher performance stability across sequential tasks, highlighting the benefits of jointly integrating cognitive modeling, meta-learning, and continual learning. Unlike existing approaches that address these dimensions in isolation, CAMEL provides a unified framework that supports both short-term personalization and long-term adaptability.

7. Discussion

The observations from the five modules validate the proposed CAMEL framework with continual learning, meta-learning, and cognitive self-regulation as a converged intelligent learning system. Cognitive modeling, fast adaptation, and long-term memory retention for adaptive learning technologies have previously been highlighted as key limitations of the literature [1,12], which this integration directly addresses. Modules provide complementary aspects of personalization, allowing the system to individually support the learner across time.
As examined in Module 1, the Cognitive State Estimator was able to infer internal learner states from behavioral traces, in line with previous evidence suggesting that multimodal interaction data can serve as robust proxies for self-regulated learning and cognitive engagement [11,12]. This has provided inferred states that acted as inputs for downstream adaptation. The quick personalization ability of the Meta-Learning Engine (Module 2), with a tiny number of interaction data, provides evidence that meta-learning is well-suited for realistic educational scenarios based on publicly available datasets, where cognitive data are usually not readily available or labeled [8,10]. While continual learning enables gradual and cumulative personalization, this rapid adaptation is essential for immediate personalization in early stages.
Strategies for continual learning achieved using Elastic Weight Consolidation (EWC) in Modules 3 and 4 and Replay proved successful in preventing catastrophic forgetting. This is particularly applicable in educational settings where learners are faced with changing tasks and content over long timescales [9]. Replay demonstrated modestly superior long-term stability, in accordance with previously known advantages of episodic memory and rehearsal mechanisms for retaining previous knowledge while allowing for the learning of new patterns.
A complementary analysis using an EdNet-like dataset could be conducted to further evaluate the architecture’s temporal relevance and generalizability. The EdNet-like sequences had standard features of contemporary intelligent tutoring systems, such as heavy-tailed distributions of response times, high interaction sparsity, and heterogeneity in the level of activity. These properties match well with modern large-scale interaction datasets and provide strong motivation in favor of CAMEL’s learning components, specifically the continual learning modules that are designed to accommodate non-stationarity and the meta-learning component designed for fast adaptation across learner types. The complementary analysis strengthens the applicability of the system across different data sets by providing rich context in the form of contextual SRL indicators provided by xAPI-Edu-Data [3,4,19].
The outputs of the previous modules use merged information from our data and create what the authors refer to as pedagogically relevant, cognitively informed interventions (these outputs are operationalized in Module 5, the Pedagogical Decision Engine). The fact that it is responsive to varying motivation, attention, and forgetting rate resonates with SRL research, highlighting the need for real-time scaffolding and adaptive supportive instruction [1,5]. This module encapsulates the larger CAMEL vision: we are trying to move beyond static learner modeling towards a dynamic, cognitive-aware pedagogy that can maintain personalized learning over longer time periods.
However, despite encouraging results, there are several limitations. Cognitive states were indirectly inferred, and future work should combine more multimodal signals to form stronger validation of performance. Second, while interpretable, the rule-based component of the decision engine could be improved through reinforcement learning to better capture long-term dependencies [10]. Third, the large-scale deployment of these approaches when placed in naturalistic, authentic learning environments will continue to be required to evaluate practical effectiveness (i.e., ecologically valid educational impact) as other recent literature has also recommended [1,14,19].
In summary, the CAMEL architecture shows that a theoretical model that integrates meta-learning, continual learning, and cognitive self-regulation can be an architecture that is both theoretically sound and feasible to implement. These insights about the evolving implementation of the system through the complementary analyses conducted with xAPI-Edu-Data and an EdNet-like dataset strengthen the robustness and applicability of the system jointly. These solutions could pave the way toward next-generation intelligent tutoring and adaptive learning systems to provide authentic, personalized, adaptive, and cognitively aligned support over time.

8. Conclusions

CAMEL is a unified architecture we designed that incorporates meta-learning, continual learning, and cognitive self-regulation collectively to facilitate truly sustained adaptive personalization over time in an intelligent educational system. Simultaneously using several cues to estimate cognitive state, adapting few-shot to retain memory, and directing a pedagogical path result demonstrate CAMEL functioning robustly amidst nontrivial behavioral phenomena and enabling adaptive interventions over multiple learning sessions.
Although the main evaluation was done using simulation-based analyses leveraging xAPI-Edu-Data, the complementary experiment we conducted with an EdNet-like dataset nonetheless highlights the ability of CAMEL to work with modern large-scale interaction traces (using EdNet) and to generalize to various patterns of learner behavior. Second, the system-level case scenario of LMS integration shows that CAMEL can function as a unit of deployment in real-world instructional systems. Together, these validation steps reinforce the relevance of the architecture in practical settings, as opposed to solely in simulation.
However, the full empirical deployment remains a valuable area for future work. In contrast, real-world learning environments are more complex; they involve diverse motivational conditions, variability in instructional support, and contextual constraints that are often outside the bounds of controlled simulations. Therefore, we will be conducting a small pilot study with live students to determine its pedagogical utility as well as usability and cognitive alignment in the classroom. Through these design principles, this work provides a theoretically founded yet practical foundation for future adaptive educational systems. CAMEL combines cognitive modeling, meta-learning, and continual learning to make progress toward tutors that can provide personalized, adaptive, and cognitively responsive support over time in a closed-loop framework.

Author Contributions

Conceptualization, R.O., A.J. and L.L.; methodology, R.O.; software, R.O.; validation, R.O., L.L., and A.J.; formal analysis, R.O.; investigation, R.O.; resources, H.T.; data curation, R.O.; writing—original draft preparation, R.O.; writing—review and editing, R.O., A.J. visualization, R.O.; supervision, A.J., L.L. and H.T.; project administration, H.T. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding. The Article Processing Charge (APC) was funded by author Adil Jeghal.

Data Availability Statement

The datasets used in this study are publicly available. The xAPI-Edu-Data dataset is available at https://www.kaggle.com/datasets/aljarah/xapi-edu-data (accessed on 15 December 2025). The EdNet KT1 dataset is available at https://github.com/riiid/ednet (accessed on 14 December 2025).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Azevedo, R.; Bouchet, F.; Duffy, M.; Harley, J.; Taub, M.; Trevors, G.; Cloude, E.; Dever, D.; Wiedbusch, M.; Wortha, F.; et al. Lessons learned and future directions of MetaTutor: Leveraging multichannel data to scaffold self-regulated learning with an intelligent tutoring system. Front. Psychol. 2022, 13, 813632. [Google Scholar] [CrossRef] [PubMed]
  2. Dağgöl, G. Online self-regulated learning and cognitive flexibility through the eyes of English-major students. Acta Educ. Gen. 2023, 13, 107–132. [Google Scholar] [CrossRef]
  3. Gorbunova, A.; Lange, C.; Savelyev, A.; Adamovich, K.; Costley, J. The interplay of self-regulated learning, cognitive load, and performance in learner-controlled environments. Educ. Sci. 2024, 14, 860. [Google Scholar] [CrossRef]
  4. Doo, M.; Bonk, C.J.; Heo, H. Examinations of the relationships between self-efficacy, self-regulation, teaching, cognitive presences, and learning engagement during COVID-19. Educ. Technol. Res. Dev. 2023, 71, 481–504. [Google Scholar] [CrossRef] [PubMed]
  5. Hamzah, H.; Hamzah, M.; Zulkifli, H. Self-regulated learning theory in metacognitive-based teaching and learning of higher-order thinking skills (HOTS). TEM J. 2023, 12, 2530–2540. [Google Scholar] [CrossRef]
  6. Gojkov-Rajić, A.; Šafranj, J.; Gak, D. Self-confidence in metacognitive processes in L2 learning. Int. J. Cogn. Res. Sci. Eng. Educ. 2023, 11, 1–13. [Google Scholar] [CrossRef]
  7. Chan, C.K.Y. A comprehensive AI policy education framework for university teaching and learning. Int. J. Educ. Technol. High. Educ. 2023, 20, 38. [Google Scholar] [CrossRef]
  8. Drigas, A.; Mitsea, E.; Skianis, C. Meta-learning: A nine-layer model based on metacognition and smart technologies. Sustainability 2023, 15, 1668. [Google Scholar] [CrossRef]
  9. Akavova, A.; Temirkhanova, Z.; Lorsanova, Z. Adaptive learning and artificial intelligence in the educational space. E3S Web. Conf. 2023, 451, 06011. [Google Scholar] [CrossRef]
  10. Osakwe, I.; Chen, G.; Fan, Y.; Raković, M.; Singh, S.; Lim, L.; van der Graaf, J.; Moore, J.; Molenaar, I.; Bannert, M.; et al. Towards prescriptive analytics of self-regulated learning strategies: A reinforcement learning approach. Br. J. Educ. Technol. 2024, 55, 1747–1771. [Google Scholar] [CrossRef]
  11. Mejeh, M.; Rehm, M. Taking adaptive learning in educational settings to the next level: Leveraging natural language processing for improved personalization. Educ. Technol. Res. Dev. 2024, 72, 1597–1621. [Google Scholar] [CrossRef]
  12. Cloude, E.B.; Azevedo, R.; Winne, P.H.; Biswas, G.; Yang, E. System design for using multimodal trace data in modeling self-regulated learning. Front. Educ. 2022, 7, 928632. [Google Scholar] [CrossRef]
  13. Wang, Q.; Huang, R.; Xiong, J.; Yang, J.; Dong, X.; Wu, Y.; Wu, Y.; Lu, T. A survey on fault diagnosis of rotating machinery based on machine learning. Meas. Sci. Technol. 2024, 35, 102001. [Google Scholar] [CrossRef]
  14. Maity, S.; Deroy, A. Generative AI and its impact on personalized intelligent tutoring systems. OSF Prepr. 2024. [Google Scholar] [CrossRef]
  15. Oubagine, R.; Laaouina, L.; Jeghal, A.; Tairi, H. Integrating Cognitive Computing in Smart Education Systems: Personalizing Learning Through Adaptive Analytics. In Proceedings of the 2025 International Conference on Circuit, Systems and Communication (ICCSC), Fes, Morocco, 19–20 June 2025; IEEE: Piscataway, NJ, USA, 2025; pp. 1–6. [Google Scholar]
  16. Zhang, S.; Wang, X.; Ma, Y.; Wang, D. An adaptive learning method based on knowledge graph. Front. Educ. Res. 2023, 6, 112–115. [Google Scholar] [CrossRef]
  17. Zarouk, M.Y.; Khaldi, M. Metacognitive learning management system supporting self-regulated learning. In Proceedings of the 2016 4th IEEE International Colloquium on Information Science and Technology (CIST), Tangier-Assilah, Morocco, 24–26 October 2016; IEEE: Piscataway, NJ, USA, 2018; pp. 929–934. [Google Scholar]
  18. Strielkowski, W.; Grebennikova, V.; Lisovskiy, A.; Rakhimova, G.; Vasileva, T. AI-driven adaptive learning for sustainable educational transformation. Sustain. Dev. 2025, 33, 1921–1947. [Google Scholar] [CrossRef]
  19. Oubagine, R.; Laaouina, L.; Jeghal, A.; Tairi, H. Advancing MOOCs Personalization: The Role of Generative AI in Adaptive Learning Environments. In International Conference on Artificial Intelligence in Education; Springer Nature: Cham, Switzerland, 2025; pp. 242–254. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.