Next Article in Journal
Stress Development in Droplet Impact Analysis of Rain Erosion Damage on Wind Turbine Blades: A Review of Liquid-to-Solid Contact Conditions
Previous Article in Journal
A Multimodal Bone Stick Matching Approach Based on Large-Scale Pre-Trained Models and Dynamic Cross-Modal Feature Fusion
Previous Article in Special Issue
Augmented Reality for Learning Algorithms: Evaluation of Its Impact on Students’ Emotions Using Artificial Intelligence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Systematic Review

Machine Learning and Generative AI in Learning Analytics for Higher Education: A Systematic Review of Models, Trends, and Challenges

by
Miguel Ángel Rodríguez-Ortiz
1,2,
Pedro C. Santana-Mancilla
2 and
Luis E. Anido-Rifón
1,*
1
atlanTTic Research Center, School of Telecommunications Engineering, University of Vigo, 36310 Vigo, Spain
2
School of Telematics, Universidad de Colima, Colima 28040, Mexico
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(15), 8679; https://doi.org/10.3390/app15158679
Submission received: 19 June 2025 / Revised: 28 July 2025 / Accepted: 4 August 2025 / Published: 5 August 2025

Abstract

Featured Application

This review supports the design of hybrid learning analytics systems that combine ML and GenAI to enable early risk detection and personalized feedback in higher education.

Abstract

This systematic review examines how machine learning (ML) and generative AI (GenAI) have been integrated into learning analytics (LA) in higher education (2018–2025). Following PRISMA 2020, we screened 9590 records and included 101 English-language, peer-reviewed empirical studies that applied ML or GenAI within LA contexts. Records came from 12 databases (last search 15 March 2025), and the results were synthesized via thematic clustering. ML approaches dominate LA tasks, such as engagement prediction, dropout-risk modelling, and academic-performance forecasting, whereas GenAI—mainly transformer models like GPT-4 and BERT—is emerging in real-time feedback, adaptive learning, and sentiment analysis. Studies spanned world regions. Most ML papers (n = 75) examined engagement or dropout, while GenAI papers (n = 26) focused on adaptive feedback and sentiment analysis. No formal risk-of-bias assessment was conducted due to heterogeneity. While ML methods are well-established, GenAI applications remain experimental and face challenges related to transparency, pedagogical grounding, and implementation feasibility. This review offers a comparative synthesis of paradigms and outlines future directions for responsible, inclusive, theory-informed AI use in education.

1. Introduction

Amid the growing integration of AI in education, machine learning (ML) and generative AI (GenAI) have gained prominence for their ability to personalize learning, predict outcomes, and support institutional decision-making [1,2,3]. While higher education fosters critical thinking and innovation [4], it continues to face challenges such as dropout, inequality, and the demand for adaptive environments [5]. In response, learning analytics (LA), enhanced by ML, is increasingly used to promote equity and student success through data-driven practices [6,7].
Defined by Siemens [8] as the analysis of data about learners and their contexts to optimize learning, LA has embraced ML techniques—supervised, unsupervised, and semi-supervised—to forecast performance, detect disengagement, and enable timely interventions [9,10,11,12]. Recent advances in GenAI, especially large language models (LLMs), are expanding LA’s capabilities through automated feedback, intelligent tutoring, and interaction analysis [13,14,15], although empirical applications remain limited and underexplored in terms of pedagogy and ethics. Some studies caution that GenAI’s “black-box” nature and lack of pedagogical grounding may undermine trust and educational validity, contrasting with perspectives that emphasize its potential to enhance learner engagement and scalability.
Existing reviews on ML in LA [16,17,18,19] rarely address GenAI or offer an integrated synthesis of both approaches. To fill this gap, this review analyzes empirical studies from 2018 to 2025, including recent developments from major conferences.
The study addresses two research questions:
  • RQ1: How are ML and GenAI applied in LA within higher education?
  • RQ2: What benefits arise from their integration in this context?
The results indicate that, while ML remains dominant in predictive analytics, GenAI is emerging in real-time feedback and adaptive learning but still faces critical challenges around transparency and pedagogical validity. By examining methodological trends, educational applications, and broader implications, this review offers a timely perspective to guide the responsible adoption of AI in higher education.

2. Materials and Methods

2.1. Study Eligibility Criteria

This review adheres to PRISMA 2020 guidelines [20] to ensure methodological transparency and rigor. A completed PRISMA 2020 checklist is provided as Supplementary Materials. It synthesizes empirical studies on the use of ML and GenAI in learning analytics (LA) within higher education, applying stream-specific inclusion and exclusion criteria as follows:
  • Inclusion Criteria (both streams);
    Language: Articles must be published in English;
    Accessibility: Full-text availability is required;
    Study type: Only empirical studies with a clearly stated research question;
    Context: The study must explicitly address a learning analytics objective in higher education.
  • For LA + ML (2018–2023);
    Must apply Machine Learning techniques within LA;
    Must report empirical results, such as predictive performance or analytics-based outcomes.
  • For LA + GenAI (2020–2025);
    Must apply generative AI models in LA;
    Applications must involve generation, adaptation, or feedback, aligned with LA goals.
  • Exclusion Criteria.
    Non-empirical works;
    Preprints and non-peer-reviewed documents;
    Studies lacking either an LA objective or the use of ML/GenAI;
    Conference papers were not formally included in the final dataset to ensure consistent peer-review standards and methodological rigor across all selected studies. This decision aimed to maintain a homogeneous level of academic scrutiny, focusing exclusively on peer-reviewed journal articles.
This stream-specific distinction ensures a consistent comparative analysis and captures the emergence of GenAI as a distinct methodological paradigm within LA.

2.2. Data Sources

The literature search was conducted in two phases to reflect the evolution of the research questions and the emergence of generative AI (GenAI) in education.
In the first phase, covering ML-based learning analytics (LA + ML) from 2018 to 2023, 11 scholarly databases were consulted to construct a comprehensive and systematic corpus: ACM Digital Library, IEEE Xplore, Emerald, ERIC, ProQuest, Sage Journals, Web of Science, ScienceDirect, Wiley Online Library, Taylor & Francis, and Scopus.
In a subsequent phase, an updated search strategy was implemented to include studies on generative AI in learning analytics (LA + GenAI), published between 2020 and early 2025. Given the novelty and rapid development of GenAI applications, the search focused on sources with high coverage of emerging AI research: ACM Digital Library, IEEE Xplore, SpringerLink (EC-TEL), and Scopus. This approach responded to the more limited but rapidly evolving body of GenAI literature in education, which tends to be concentrated in fewer venues.
While the formal PRISMA-based corpus was limited to peer-reviewed journal articles, an exploratory review of flagship conference proceedings, including LAK, L@S, and EC-TEL, was conducted for the full period (2018–2025) to identify recent trends not yet indexed. Selected contributions, particularly from 2024 and 2025, are discussed in the Related Work section to enrich the contextual interpretation.

2.3. Search Strategy

Search queries combined three conceptual pillars—learning analytics (including social learning analytics), machine learning or GenAI, and higher education. Synonyms and Boolean operators ensured comprehensive yet focused retrieval. The search spanned 2018–2025 (2018–2023 for ML in LA). Full queries are detailed in the Supplementary Materials (Table S1) and available on Zenodo (https://doi.org/10.5281/zenodo.15233231). Of the 9590 records retrieved, 8263 were excluded after deduplication and screening (see PRISMA flowchart, Figure 1).

2.4. Data Extraction and Collection Process

All references were managed using RefWorks ProQuest. After deduplication, two parallel screening processes were conducted. For LA + ML, 1344 records were screened, 954 excluded, and 372 assessed for eligibility, yielding 75 empirical studies. For LA + GenAI, 429 records were screened, 351 excluded due to irrelevance or lack of empirical content, and 68 assessed, resulting in 26 included studies. Figure 1 summarizes the inclusion and exclusion process.
Screening was conducted using a structured coding protocol aligned with the review’s research questions. All screening stages (titles/abstracts and full texts) were performed by a single reviewer. Each record was evaluated against the predefined eligibility criteria, with particular emphasis on the presence of an explicit research question and the study’s relevance to addressing the review objectives.

2.5. Quality Appraisal

To assess the methodological quality of the included studies, we applied the Mixed Methods Appraisal Tool (MMAT, 2018 version) [21]. Each study was classified according to MMAT study type (qualitative, quantitative, mixed methods) and assessed using its corresponding checklist. Screening criteria (S1 and S2) and methodological dimensions (e.g., sampling, measurement, data analysis) were evaluated and summarized. The complete dataset with MMAT ratings, checklist responses, study types, and reviewer notes for all 101 studies is openly available on Zenodo (https://doi.org/10.5281/zenodo.16416487). This open dataset enhances transparency and supports reproducibility of the review process.

2.6. Final Dataset

The final review includes 101 empirical studies: 75 focused on traditional ML applications in learning analytics and 26 integrating GenAI models, such as GPT, BERT, or FLAVA, into LA contexts. This dual corpus enables analysis of both the evolution of ML and the rise of GenAI within the LA research landscape.

2.7. Declaration of GenAI Use

During the study, ChatGPT-4.5 (OpenAI) was used to improve language clarity and to assist in refining the Python 3.10 code. All AI-generated output was reviewed and edited by the authors, who take full responsibility for the final content.

2.8. Data Availability

The structured dataset (metadata extracted from 101 studies, including AI models, contexts, and techniques) is publicly available via Zenodo (https://doi.org/10.5281/zenodo.16416465) to ensure transparency and reproducibility.

2.9. Ethical Considerations

No human or animal subjects were involved; ethical approval was not required.

2.10. Registration and Protocol

This review was not prospectively registered, and no separate protocol was prepared. Consequently, no protocol amendments apply.

3. Results

To structure the results, the corpus was divided into two groups—LA&ML (traditional machine learning) and LA&GenAI (generative AI)—based on each study’s core methodology. Within these groups, unsupervised clustering was applied to identify thematic patterns using three categorical variables: AI models, application, and educational context, encoded via one-hot encoding. Dimensionality reduction and visualization were performed with principal component analysis (PCA), and the optimal number of clusters was determined using silhouette scores. The analysis, conducted in Python via Google Colab, informed the thematic organization of Section 3.2 and Section 3.3, enabling a data-driven presentation of the findings across both paradigms.

3.1. Temporal and Geographical Distribution of Publications

This section analyzes the evolution of learning analytics (LA) research using machine learning (ML) and generative AI (GenAI) in higher education between 2018 and 2025, focusing on temporal trends and geographical spread.
The annual trends (Figure 2) show consistent growth in ML-based LA studies, with a peak in 2022. GenAI studies first appeared in 2022 and accelerated in 2024, coinciding with the widespread release and adoption of large language models (LLMs) such as GPT-4. This pattern reflects two coexisting trajectories: the consolidation of traditional ML approaches—classification, regression, and ensemble models applied to LMS data—and the experimental rise of GenAI applications for feedback generation, engagement modeling, and content personalization. While ML remains foundational, GenAI is rapidly gaining traction, though it is still in the early stages of methodological standardization and empirical validation.
Geographical distribution (Figure 3) reveals notable disparities. The United States leads with 17 studies (10 ML, 7 GenAI), followed by Australia, China, and Germany, which show relatively balanced activity across both paradigms. However, GenAI research remains highly concentrated in high-income regions, with limited representation from Latin America, Sub-Saharan Africa, and parts of Asia. These gaps raise concerns about equity, contextual relevance, and disparities in access to GenAI infrastructure for research and implementation in education.
The coexistence of ML and GenAI suggests a transitional phase in LA research. ML offers robustness and interpretability, while GenAI brings adaptability and personalization. Hybrid approaches could combine their strengths to enhance flexibility and inclusiveness.
Three key observations emerge: (1) GenAI adoption aligns with the release of tools like GPT-4, highlighting the role of accessibility; (2) its concentration in high-income regions may deepen epistemic inequalities; and (3) the parallel rise of both paradigms invites integrated, transparent, and context-aware frameworks.

3.2. Learning Analytics with Traditional Models

To analyze the application of traditional machine learning (ML) in learning analytics (LA), we clustered 75 peer-reviewed studies from 2018 to 2023. Using three categorical variables—AI models, application type, and educational context—encoded and reduced via principal component analysis (PCA), we identified thematic groupings centered on engagement prediction, dropout modeling, academic performance forecasting, and feedback systems across online, blended, and face-to-face settings.
Figure 4 presents the top 10 ML models by context. Random forest, support vector machine (SVM), and decision tree are most prevalent, particularly in online and MOOC environments, due to their robustness, interpretability, and compatibility with structured behavioral data. Logistic regression, naive Bayes, and artificial neural networks also appear frequently, indicating a reliance on established methods.

3.2.1. Engagement Prediction in Online Learning with Traditional ML Models

The analysis of student engagement in online and hybrid environments has progressed through the integration of ML within LA. This cluster, comprising 11 studies, focuses on predicting engagement, academic performance, and dropout risk using structured and unstructured data sources.
Most studies rely on classical ML algorithms, such as random forest, decision tree, SVM, and logistic regression, often enhanced with ensemble techniques like AdaBoost and Bagging. These models are widely used in MOOCs and online courses to support an early detection of at-risk behavior. However, few studies, including those addressing engagement and behavioral prediction [22,23,24,25,26], assess whether such predictions lead to real-time, actionable interventions.
The choice of algorithm often reflects the nature of the data: SVM and artificial neural networks (ANN) are used for fine-grained engagement prediction [24], while decision tree and logistic regression are suited to interpret behavioral traces such as video interaction logs [25]. Feature engineering, clustering, and regression remain central analytical techniques [27].
A notable methodological trend is the use of natural language processing (NLP) to analyze forum content and course reviews. Word embeddings—Doc2Vec, Word2Vec, FastText, and GloVe—support sentiment classification and behavioral inference [28,29]. Deep learning models, such as LSTM and GRU, enhance performance in text-based engagement tasks, with GloVe consistently yielding superior accuracy [28].
Among the most comprehensive approaches is the work of Onan [28], which combines traditional classifiers, ensemble techniques, and deep architectures like CNN, GRU, LSTM, and attention-based RNN. The model achieves 95.80% accuracy using LSTM with GloVe embeddings, demonstrating the synergy between affective analytics and attention mechanisms—an intersection between traditional ML and GenAI.
Emerging studies also explore spatiotemporal learning behaviors. Du et al. [30] show that consistent study patterns correlate with academic success. Moubayed et al. [31] apply clustering to blended learning profiles, while Lahza et al. [32] investigate strategy use in learner sourcing platforms, linking it to instructional design quality.
Key applications include:
  • Dropout detection and risk assessment [22,25];
  • Engagement forecasting and behavioral profiling [24,26,27];
  • Sentiment analysis of learner feedback [28];
  • Modeling learning routines and platform interactions [30,31,32].
Challenges and opportunities: Multimodal integration remains limited, with most studies focusing on a single data type. Longitudinal analyses are rare, limiting insights into lasting engagement patterns. Although XAI is occasionally applied [23,26], most models lack transparency, reducing their practical utility in educational settings.
Recommendations for future research:
  • Combine behavioral, textual, and emotional data sources;
  • Conduct longitudinal studies on sustained engagement;
  • Enhance model explainability through tailored XAI tools;
  • Link predictions to adaptive instructional strategies.
In summary, progressing toward multimodal and learner-centric LA systems is essential for fully leveraging ML in higher education contexts.

3.2.2. Dropout Prediction in Digital Education with Traditional ML Models

Student attrition remains a persistent concern in digital higher education, often preceded by disengagement—behavioral, cognitive, or emotional. Predicting these early indicators is crucial for intervention. Predictive learning analytics (PLA) and machine learning (ML) are widely used to identify at-risk students before dropout. This section synthesizes findings from 23 peer-reviewed studies applying AI models across online, blended, and MOOC environments.
Random forest and SVM are the most used models, with each in nine online learning studies. Decision tree, naive Bayes, and logistic regression remain favored for their simplicity and interpretability. Ensemble methods such as random forest and boosting reduce overfitting and handle feature interactions effectively [33,34,35]. In MOOCs, random forest reached F1-scores above 0.88 based on daily progress [34]. Hybrid ensembles combining classification and regression enabled early risk detection before midterms [36]. During the pandemic, extra trees and logistic regression surpassed 90% specificity in asynchronous learning [37].
Neural models like LSTM and GRU outperform traditional classifiers. LSTM exceeded CNNs and MLPs from week six in STEM courses [38]; GRU achieved 98% accuracy in remedial English [39]. IOHMM and sequential logistic regression predicted dropout by week eight, outperforming SVM and logistic models [40].
Interpretable models like decision trees and logistic regression remain valuable for deployment [41,42], often as base learners in ensembles [43]. Semi-supervised models with SHAP explanations combine diverse data types [44]. Dashboards like OU analyse supported instructor monitoring and improved student outcomes [45]. Behavioral profiling through LMS logs and motivational indicators guided personalization [46,47]. Ensemble models also supported post-pandemic readiness [48], and SHAP facilitated fairness-aware predictions [49]. Self-reports on motivation were included in some models [50]. Sentiment analysis with SVM and decision trees reached over 90% accuracy [51]. Random matrix theory and community detection uncovered engagement-success links in programming education [52]. SMOTE addressed class imbalance [39]. Hybrid ensembles also explored fuzzy logic [53]. Generative models remain rare [54], and model generalizability is still limited [55].
Key applications include:
  • Early warning systems in initial weeks [35,38];
  • Instructor dashboards for real-time monitoring [45];
  • Adaptive feedback based on predictions [35];
  • Behavioral profiling using LMS and motivation data [46,47];
  • Post-pandemic readiness via ensembles [48];
  • Fairness-aware interventions using SHAP [49].
Challenges and opportunities: Generative AI is underused [38,54]. LMS logs dominate, while multimodal inputs are scarce [47]. Few systems operate in real time or are deployed in institutions [35,45]. Explainability tools lack integration into actionable dashboards [49]. Generalizability is limited due to narrow datasets [55].
Recommendations for future research:
  • Apply generative models for feedback and simulation;
  • Incorporate multimodal sources (e.g., forums, self-reports);
  • Deploy real-time tools in institutional settings;
  • Embed explainability in instructor-facing systems;
  • Expand validation across contexts and institutions.

3.2.3. Academic Performance Prediction in Face-to-Face Classrooms

Predicting academic performance in face-to-face higher education settings remains a core objective within learning analytics (LA), primarily aiming at the early identification of students at risk. This chapter synthesizes findings from 26 studies utilizing traditional machine learning (ML) models in classroom contexts.
The most frequently employed ML algorithms include random forest (RF), support vector machines (SVM), decision trees (DT), and logistic regression, each favored due to their effectiveness with structured academic data. Specifically, RF has reliably detected underperformance through historical grades [56]. However, its predictive accuracy is notably reduced in small or homogenous student groups [57]. Conversely, SVM demonstrated strong performance with smaller datasets [58,59], though its applicability to larger classrooms remains inadequately explored. Decision trees are valued for their interpretability in guiding interventions [60] but often face challenges with imbalanced datasets [61]. Meanwhile, artificial neural networks (ANNs) effectively model complex multimodal datasets [62,63], though their inherent opacity complicates practical deployment.
Supervised learning methodologies—primarily classification and regression—dominated the reviewed studies, largely based on academic records and limited learning management system (LMS) logs. Despite the inclusion of innovative approaches such as natural language processing (NLP) for feedback analysis [62,64] and sensor-based physical interaction data [65], face-to-face contexts continue to present significant barriers due to insufficient digital interaction data for adaptive and longitudinal modeling.
Advanced analytical techniques, including concept inventories [66], gesture analytics [67], and LMS-integrated predictive models [68], contributed positively to predictive accuracy but were often limited by small sample sizes, affecting generalizability. Additionally, early assessments [69,70], demographic indicators (e.g., socioeconomic status, language proficiency, background) [71,72], and institutional data emerged as reliable predictors. Curriculum personalization emerged as a notable application. Recommendation systems were proposed based on academic pathways [73,74], aiding program design.
Key applications include:
  • Performance forecasting enabling timely academic interventions [57,62,63,65,75];
  • Dropout prediction identifying risk patterns for preventive actions [56,60,61,76,77,78,79];
  • Personalized feedback via NLP [64];
  • Curriculum personalization through predictive analytics informing tailored educational pathways [73,74,80,81].
Challenges and opportunities: Despite significant progress, key limitations persist. Generalizability remains constrained by small or imbalanced datasets and institutional variability [57,58,67]. Face-to-face contexts lack continuous digital traces, limiting adaptive and multimodal analytics. Moreover, critical variables such as motivation, emotion, and socio-economic status are often excluded despite their predictive value [60,66,71,75,76]. Addressing these issues demands not only methodological refinement but also ethical vigilance. Incorporating sensitive data raises concerns about privacy, consent, and bias. Future research must adopt privacy-preserving practices, transparent data governance, and inclusive frameworks to ensure that predictive systems are equitable, respectful of student rights, and contextually appropriate across diverse educational environments.
Recommendations for future research:
  • Enhance multimodal analytics by incorporating sensor-based data (e.g., physical interactions) alongside emotional, motivational, and socio-economic variables;
  • Explore advanced methodologies such as reinforcement learning and generative AI to develop dynamic, adaptive curricular recommendations;
  • Prioritize explainable AI (XAI) development to enhance interpretability and facilitate educator acceptance, rigorously validating effectiveness in authentic educational environments.

3.2.4. Feedback and Performance Modeling in Hybrid Learning with ML

Machine learning is essential in hybrid learning contexts, especially for modeling academic performance and automating feedback. Artificial neural networks (ANN), random forest (RF), and support vector machines (SVM) dominate due to their effectiveness with varied educational data. ANN frequently models complex behaviors like cognitive engagement, though it faces interpretability challenges [82,83,84]. SVM excels at classifying high-dimensional textual data, identifying cognitively relevant content [84,85]. RF balances predictive accuracy and interpretability, analyzing variables such as peer interactions and digital distractions [83,86].
Recent use of transformer-based models like BERT has advanced analysis of emotional and behavioral signals in collaborative settings, enabling real-time personalized feedback and co-regulated learning [87].
Multimodal datasets typically combine structured data (grades, surveys) with unstructured inputs (forum posts, interaction logs, keystroke dynamics). Natural language processing (NLP) is crucial for preprocessing text, classifying engagement, and assessing discourse relevance [83,84]. Keystroke analytics reveal writing fluency and cognitive processes, while LMS logs highlight self-regulated learning patterns [88,89]. Hierarchical regression isolates cognitive, affective, and behavioral influences on student performance [90].
Explainable AI (XAI) techniques like LIME and SHAP have begun enhancing transparency and informing pedagogical decisions, yet broader integration is underdeveloped [91,92].
Key applications include:
  • Cognitive engagement via discourse analysis [84];
  • Analyzing distraction and peer interactions using NLP-classified discussions [83,89];
  • Real-time, emotion-sensitive feedback leveraging transformer-based models [87];
  • Early identification of academic risks using RF with academic and behavioral data [93,94,95];
  • Instructional design optimization through behavioral predictors of satisfaction and performance [96].
Challenges and opportunities: Despite progress, interpretability and adaptive responsiveness remain challenges. Consistent use and pedagogical value of XAI techniques require further integration and validation [85,89,90]. Current multimodal integration mostly focuses on textual or structured inputs, limiting personalization and scalability. Incorporating audio, video, and physiological signals within diverse, multi-institutional datasets could significantly enhance analytical robustness [81,92,93]. Moreover, personalized interventions based on predictive analytics need rigorous longitudinal evaluation to confirm their educational impact [94].
Recommendations for future research:
  • Integrate richer multimodal data (audio, video, physiological signals) to enhance analytical capabilities;
  • Systematically apply advanced XAI techniques for increased model interpretability;
  • Develop and validate adaptive, real-time feedback systems through robust longitudinal and experimental research.
Emerging technologies such as generative AI, multimodal analytics, and advanced XAI offer new opportunities to overcome current limitations in feedback and performance modeling. Their potential lies in enhancing accuracy, personalization, and pedagogical value, but this requires rigorous validation, integration into real hybrid learning contexts, and alignment with instructional goals. Future systems must evolve toward real-time, interpretable, and adaptive support that scales to diverse learner needs.

3.3. Expanding Learning Analytics Through Generative AI

The integration of generative artificial intelligence into learning analytics represents a major methodological shift in higher education. To explore this landscape, we clustered 26 GenAI-focused studies into two groups. The first cluster (n = 14) centers on large language models (LLMs) like GPT-4, GPT-3.5, BERT, and FLAVA, applied primarily in higher education. These studies focus on engagement analytics, adaptive learning, and ethical bias detection, often using multimodal inputs and transformer-based architectures. The second cluster (n = 12) features diverse approaches combining GenAI with emotion classifiers, GANs, or custom copilots, deployed across online, hybrid, and in-person contexts.
Figure 5 presents the most used GenAI models and their distribution across learning environments. GPT-4 dominates in both frequency and application scope, followed by BERT, GPT-3.5, and hybrid models (e.g., ChatGPT with clustering). Despite advances in multimodal GenAI, implementations remain largely text-based, with emphasis on automated feedback and engagement modeling.
These trends reflect innovation but also surface critical concerns around equity, transparency, and practical feasibility in varied educational settings.

3.3.1. Generative AI Applications for Engagement, Feedback, and Adaptation

The growing integration of generative AI (GenAI) into learning analytics (LA) underscores a methodological shift toward transformer-based models, primarily GPT-3.5, GPT-4, BERT, FLAVA, and Whisper. Frequently, these models operate within hybrid frameworks combining supervised classifiers (e.g., SVM, random forest, logistic regression) and unsupervised techniques, such as clustering and topic modeling. Among the 14 reviewed studies, GPT and BERT variants dominate, occasionally supported by recurrent neural networks (RNNs) or long short-term memory (LSTM) networks. Classical methods like decision trees and naive Bayes are occasionally used for simpler classification tasks [97]. However, the general absence of lightweight models raises scalability and equity concerns, particularly for low-resource educational contexts.
Quantitative analysis reveals GPT-4 as the most widely used model, followed by BERT and other GPT variants. Diverse methodologies, including neural networks and clustering algorithms, highlight varied applications, ranging from engagement analytics and adaptive feedback to personalized learning experiences. Large language models (LLMs) facilitate real-time reflective scaffolding [98], the detection of inclusive interactions [99], personalized feedback [100], and improved learner engagement in classroom integration contexts [101]. Some longitudinal research indicates that interaction frequency with GenAI tools enhances learner autonomy via increased social presence [102], yet rigorous empirical evidence on instructional effectiveness or scalability remains sparse.
Mixed-method approaches, integrating digital learning traces, student-generated texts, GenAI interactions, and multimodal data (audio, video, gesture), are prevalent. NLP techniques, especially sentiment analysis and topic modeling using transformers, effectively analyze textual data from learners and AI-generated content [103,104]. Clustering, combined with GPT/BERT embeddings, supports learner profiling and targeted feedback strategies [105]. Sequential models, including RNNs and LSTMs, aid the analysis of trajectories related to self-regulated learning and help-seeking behaviors [101,106].
While explainable AI (XAI) is gaining attention, its application remains limited. Fahl [107] uniquely integrates GPT-4 with semantic knowledge graphs to enhance explainability. However, broader transparency and accountability issues persist, amplifying ethical concerns such as hallucination and trustworthiness.
Key applications include:
  • Formative feedback and automated assessment [100,108];
  • Adaptive prompts supporting self-regulated learning [98];
  • Visual analytics via GenAI-enhanced dashboards [107];
  • Equity-aware collaborative learning tools [99];
  • Detection of help-seeking behaviors [109];
  • Adaptive content delivery in gamified or flipped learning contexts [101].
Challenges and opportunities: Although GenAI offers considerable promise, significant limitations remain. The predominant opacity of GenAI systems hinders interpretability, with rare exceptions employing explainability frameworks [107]. Additionally, the frequent misalignment with self-regulated learning (SRL) principles risks fostering metacognitive passivity, as discussed in existing critiques [110]. Real-time adaptability is another concern, as many GenAI tools provide delayed or static feedback, inadequately addressing immediate learner disengagement—though real-time adaptive scaffolding offers potential solutions [98]. The underutilization of multimodal inputs, such as audio, video, or physiological data, limits the comprehensive understanding of learner behavior and context [99,103]. Lastly, weak theoretical grounding and limited alignment with established instructional frameworks diminish practical pedagogical effectiveness, highlighting the need for evidence-centered design approaches [106].
Recommendations for future research:
  • Incorporate explainable AI techniques extensively to enhance transparency and trustworthiness;
  • Expand multimodal GenAI applications to capture affective, cognitive, and embodied learning experiences;
  • Employ ontological models to systematically structure learning progression and knowledge monitoring [107];
  • Develop reflective GenAI agents supporting deep engagement and co-regulated learning dynamics.
Addressing these recommendations, along with methodological, ethical, and institutional considerations, will significantly advance the sustainable and effective integration of GenAI within higher education.

3.3.2. Technical Approaches and Emerging Trends in GenAI for Learning Analytics

The integration of generative AI (GenAI) into learning analytics (LA) is significantly reshaping methodological frameworks in higher education. Across twelve studies [111,112,113], large language models (LLMs), including GPT-4, GPT-4o, DistilBERT, and Gemini 1.5 Pro, have emerged as central tools for tasks such as feedback generation, affective state detection, and instructional scaffolding. Despite broad usage, most implementations remain preliminary, lacking comprehensive validation in real-world educational environments. Furthermore, the influence of instructional context on model effectiveness and learning outcomes remains underexamined.
GPT-4 variants dominate, appearing in at least eight studies, often integrated with emotion detection tools (HSEmotion), facial recognition (MTCNN), or generative adversarial networks (GANs) for enhanced multimodal analysis [114]. DistilBERT specifically excels in detecting confusion within discourse data [112]. Applications leveraging ChatGPT and GitHub Copilot have effectively captured learner strategies in coding activities [113]. Additionally, collaborative interactions supported by GPT-4 have been linked to improved hint quality and heightened critical thinking [115], while CustomGPT offers tailored insights by analyzing interactions with digital educational resources [116].
Application domains are concentrated mainly around engagement analytics, automated feedback, and risk detection. GPT-driven dashboards and conversational agents provide adaptive support and multimodal instructional scaffolding in real-time interactions [117,118]. Nevertheless, the recurrent usage of basic LLM pipelines—primarily prompt engineering combined with simple classifiers—highlights limitations in methodological innovation and adaptation to diverse instructional contexts.
The optimization of LLM outputs typically involves fine-tuning and strategic prompt design. For instance, GPT-4o has been customized to interpret natural language queries and generate educationally relevant SQL queries and visualizations [118]. Similarly, DistilBERT has successfully facilitated efficient confusion classification in MOOCs [112], while GAN-based techniques have enriched otherwise sparse datasets [114].
Some studies prioritize explainability. For instance, NLP and ChatGPT have been utilized to uncover patterns in collaboration and bias in peer-generated feedback [119]. Furthermore, GitHub Copilot interaction logs have been employed to track cognitive transitions during learning activities [113]. However, instances of feedback inaccuracies or reinforcement of misconceptions indicate the crucial role of human oversight [120].
Hybrid methodologies that combine clustering, regression, and sequential analysis with NLP techniques frequently analyze multimodal data, including LMS logs, reading behaviors, keystrokes, and emotional inputs [111,121]. Although explainable AI methods such as integrated gradients and anchors have been incorporated, their practical impact on instructional decision-making remains inadequately assessed [92,112].
Key applications include:
  • Automated formative feedback in STEM disciplines through GPT4 and ChatGPT [113,115,122];
  • Real-time confusion detection in online learning environments using DistilBERT [112];
  • Multilingual engagement analytics leveraging bilingual prompts and chat logs [120,121];
  • Bias detection and fairness evaluation in peer feedback through explainable GenAI models [119].
Challenges and opportunities: Despite promising advances, significant challenges persist. Many GenAI implementations focus predominantly on retrospective evaluation rather than proactive, real-time instructional utility. Although explainability methods are increasingly employed, their pedagogical grounding and effectiveness in supporting actual learning processes remain unclear. Ethical concerns related to inaccurate or hallucinated LLM outputs are acknowledged but insufficiently addressed through rigorous validation or expert verification [120]. Additionally, evidence supporting sustained cognitive or motivational impacts from GenAI-facilitated interactions is limited, emphasizing a critical need for longitudinal studies. Furthermore, engagement analytics tools frequently demonstrate limited transferability across educational contexts and disciplines, indicating a pressing need for scalable and adaptable solutions.
Recommendations for Future Research:
  • Develop multimodal, theory-driven GenAI models integrating gaze, speech, emotions, and learner behaviors aligned explicitly with frameworks like self-regulated learning (SRL) and feedback literacy;
  • Promote participatory design processes involving educators and students to define meaningful explanations, safety measures, and effective revision practices;
  • Expand the evaluation criteria beyond accuracy to emphasize educational utility, fairness, epistemic validity, and ethical rigor;
  • Foster scalability and transferability through adaptive models capable of continuous learning and refinement across diverse instructional contexts.
In summary, addressing foundational ethical and methodological barriers is essential to unlock the full pedagogical potential of GenAI in learning analytics.

4. Contextual Analysis of ML and GenAI in LA

The integration of artificial intelligence into learning analytics has advanced rapidly, driven by growing interest in traditional machine learning (ML) and the emerging potential of generative AI (GenAI). While ML has supported prediction and classification in education, GenAI enables real-time interaction, feedback, and adaptive content. Yet, the literature highlights persistent gaps in theoretical grounding, teacher involvement, and empirical validation. This section contrasts contributions from both domains to uncover the key trends, challenges, and future directions.

4.1. Generative AI in Learning Analytics

Recent studies emphasize GenAI’s potential to transform LA from passive prediction to real-time personalization. Borah et al. [13] proposed a framework that adapts learning paths and feedback based on cognitive and emotional profiles, integrating multimodal data, contextual rules, and conversational interfaces for co-evolving learner–instructor interaction.
Complementary analyses from the LAK and EC-TEL proceedings reinforce this trend. Huang et al. [123] found RoBERTa the most accurate and explainable model for peer feedback classification, while Pishtari et al. [124] reported significant gains in instructional design quality through GenAI-generated feedback.
Yan et al. [15] mapped GenAI’s integration across the LA cycle—from data augmentation to intervention—highlighting synthetic data and agent-based tools as enablers of engagement. Khosravi et al. [14] addressed ethical and pedagogical considerations, advocating codesign with educators, transparency, and models that safeguard student agency.
Qu and Yang [125] surveyed LLM applications such as ChatGPT in language and medical education, noting limited empirical validation and weak integration into instructional practice.
Despite these developments, most GenAI applications prioritize technical innovation over pedagogical grounding. Few incorporate learning theories or evaluate long-term outcomes, underscoring the need for more critically aligned, education-centered approaches.

4.2. Traditional Machine Learning in Learning Analytics

Traditional machine learning techniques have played a pivotal role in the development of learning analytics, particularly in areas such as prediction, classification, clustering, and recommendation. PeñaAyala’s taxonomy [126] is frequently cited for its comprehensive organization of these functions, focusing on learner performance, engagement, and dropout analysis.
Zawacki-Richter et al. [127] found that most ML-based LA systems focus on performance prediction and early warning but often lack pedagogical alignment, limiting their educational relevance. Similarly, Renz and Hilbig [128] showed that commercial EdTech tools prioritize algorithmic efficiency over interpretability, reducing their utility for instructors.
Regionally, Salas-Pilco and Yang [129] documented promising ML applications in Latin America, such as identifying at-risk students, but the highlighted challenges included limited infrastructure, institutional support, and the need for culturally responsive models. Glandorf et al. [130] revealed that dropout prediction varies by demographic group, while Poellhuber et al. [131] improved predictive efficiency by clustering course structures in Moodle before modeling.
Buitrago-Ropero et al. [132] reframed learner data as sociopedagogical artifacts—“data, action, and service”—arguing for a shift from purely computational to sociotechnical perspectives. Baek and Doleck [133] compared educational data mining and LA, noting shared methods but differing goals: EDM focuses on algorithmic development, while LA emphasizes educational impact—though both often lack theoretical and teacher integration.
Lastly, Ley et al. [134] and Aguilar-Esteva et al. [135] called for human-centered and equity-oriented LA, promoting models that are transparent to educators and responsive to sociocultural and sustainability concerns.

4.3. Critical Synthesis and Research Gaps

Despite technological advances in both traditional ML- and GenAI-based learning analytics (LA), several persistent challenges limit their educational impact. A major concern is the lack of pedagogical integration: many systems operate independently of learning theories or instructional strategies, reducing their capacity to support meaningful learning [131]. Additionally, both ML and GenAI tools often function as opaque black boxes, hindering interpretability and diminishing educator trust in system outputs [126].
Educator involvement also remains limited. Teachers are frequently positioned as passive end users rather than active codesigners, which compromises contextual relevance and adoption [132]. Moreover, LA systems show insufficient adaptation to local constraints, including infrastructure, culture, and language, particularly in underrepresented or resource-constrained educational settings [127]. These systemic limitations underscore the need for more inclusive and transparent LA design approaches.
To address these gaps, future research should prioritize the development of hybrid AI systems that combine the predictive strengths of ML with the interactive potential of GenAI. Such systems could better balance accuracy with personalization. Additionally, codesign practices involving both educators and learners are critical to ensure that LA tools align with real-world instructional needs. Enhancing model explainability is also essential, particularly for translating algorithmic complexity into actionable insights for educators.
Theory-driven design remains an underdeveloped area. Embedding constructs such as self-regulated learning, motivation, and feedback theory into system logic could enhance pedagogical alignment. Finally, advancing culturally responsive LA is imperative, particularly in settings with limited resources and diverse learner populations. Integrating these considerations will be key to developing LA systems that are not only innovative but also equitable, interpretable, and educationally meaningful.
These insights inform the discussion in Section 5, where we outline the broader implications for research and practice in learning analytics.

5. Discussion and Implications

This section critically synthesizes the findings of the review, addressing the two research questions and outlining their implications for educational practice, research, and institutional policy. The discussion is organized into four parts: current implementation (RQ1), potential benefits (RQ2), cross-cutting challenges, and future recommendations.

5.1. Current Implementation of ML and GenAI in Higher Education (RQ1)

The integration of machine learning and generative AI into learning analytics follows two methodological paths: the consolidation of traditional ML models and the exploratory adoption of GenAI systems.
ML-based LA systems are well-established for predicting academic performance (e.g., random forest, SVM, logistic regression), detecting dropout (e.g., LSTM, GRU, decision trees), modeling engagement (clustering, sentiment analysis), and optimizing feedback using k-means or regression. These tools operate mainly on structured LMS data—clickstreams, assessments, participation logs—and are valued for accuracy and simplicity, though they often lack alignment with classroom practice.
GenAI-based LA systems are emerging in personalized feedback (LLMs like GPT-4, BERT), emotional and cognitive engagement modeling (sentiment/discourse analysis), SRL support (tutoring agents), and real-time content scaffolding. These implementations are typically experimental, text-focused, and integrated through prompt engineering or fine-tuning. While studies—such as RoBERTa for peer feedback [123] and GenAI-enhanced instructional design [124]—show promise, most lack empirical validation, scalability, or connection to learning theory. Additionally, multimodal inputs (e.g., gaze, audio, gesture) remain underutilized.
ML-based LA systems are mature, data-driven, and prediction-focused; GenAI systems are flexible, interaction-oriented, and still experimental. Despite their complementarity, both paradigms face common limitations: fragmented implementation, weak pedagogical integration, and minimal educator involvement.

5.2. Potential Benefits of Integrating ML and GenAI into Learning Analytics (RQ2)

The integration of machine learning and generative AI into learning analytics in higher education presents benefits across pedagogical, institutional, and research domains.
Pedagogically, ML enables early risk detection, allowing timely interventions, while adaptive systems personalize feedback, pacing, and resources. GenAI tools enhance formative feedback by generating scalable, context-aware responses to student submissions. Additionally, conversational agents support metacognition and self-regulated learning (SRL) through the scaffolding of reflection, planning, and help-seeking behaviors.
Institutionally, predictive dashboards aid retention monitoring and academic advising, while LA-informed instructional design improves curriculum responsiveness. Scalable intervention systems, such as real-time alerts and adaptive prompts, enhance support in large or asynchronous courses. Cluster-based modeling of LMS data has also improved generalizability in early-warning systems, particularly when courses are grouped by structural similarity [131].
For research and innovation, multimodal learner modeling—integrating behavioral, affective, textual, and biometric data—offers a richer view of the learning process. Advances in explainable AI (XAI), including SHAP, LIME, and knowledge graphs, contribute to transparency and increase educator trust. Furthermore, hybrid LA systems that combine the predictive power of ML with the generative capabilities of GenAI enable more personalized and dynamic learning environments.

5.3. Cross-Cutting Challenges and Gaps

Despite advances in ML and GenAI for learning analytics (LA), several structural limitations persist that hinder their educational effectiveness.
A key cross-cutting gap lies in the limited grounding of LA systems—particularly those powered by GenAI—in established learning theories. For instance, several tools in the corpus provided personalized feedback using large language models, such as the explainable feedback dashboard by Afzaal et al. [91] and the fine-tuned GPT-4 system for hint revision explored by Singh et al. [115]. While both systems offer scalability and automation, they rarely aim to develop students’ feedback literacy—a construct encompassing the capacity to interpret, act on, and benefit from feedback [136,137]. Moreover, the feedback remains largely unidirectional, with little scaffolding for student agency or dialogic interaction [138].
Similarly, only a subset of studies explicitly aligned with the dimensions of self-regulated learning (SRL), such as metacognitive monitoring, goal-setting, and reflection [139,140]. For example, while Dai et al. [100] focused on evaluating the quality of GPT-generated feedback for open-ended writing tasks, they did not examine how such feedback supports SRL processes or learner engagement. In contrast, Li et al. [98] integrated GenAI to deliver adaptive scaffolds based on real-time analytics of SRL processes, showing improved metacognitive strategies—yet still reported variability in learner compliance and limited generalizability.
These gaps suggest that, while ML and GenAI can technically model learner behavior, their educational impact remains limited without deliberate integration into pedagogical frameworks, such as SRL, feedback literacy, and formative assessment [141]. Additionally, many ML and GenAI models remain opaque, functioning as black boxes that impede interpretability and educator trust. While explainable AI (XAI) techniques, such as RoBERTa with LIME for peer feedback classification [123], have shown promise, they are still underused.
Educators are rarely involved as co-designers, resulting in systems poorly aligned with classroom realities. GenAI implementations are often tested in artificial or highly controlled environments. Even when benefits like improved instructional design are observed [124], few studies assess long-term impact in authentic educational settings.
Geographic disparities also persist. Research and deployment efforts are concentrated in high-income regions, limiting global representativeness and inclusivity. Latin America and other underrepresented regions remain largely absent from the discourse [129].
Finally, multimodal and real-time data, such as emotional, behavioral, or biometric streams, are rarely integrated into adaptive feedback loops, which constrains the responsiveness and depth of learner modeling.
Table 1 summarizes these cross-cutting challenges and outlines the recommended research directions to inform the development of pedagogically grounded, explainable, and context-sensitive LA systems.

5.4. Implications and Recommendations

5.4.1. Implications of Generative AI for Engagement, Feedback, and Adaptation

Instructors should be actively involved in the co-design of LA systems to ensure alignment with pedagogical goals. Hybrid models that combine ML’s predictive strength with GenAI’s adaptive feedback capabilities offer promise for more responsive learning environments. Embedding explainability into system interfaces can enhance transparency, trust, and interpretability for both educators and learners.

5.4.2. Research

Future studies should move beyond predictive accuracy and assess actual learning outcomes through longitudinal and mixed methods designs. Aligning model architectures with learning theories, such as SRL, motivation, and engagement, can improve pedagogical relevance. Additionally, the instructional impact of GenAI tools on reflection, autonomy, and knowledge construction warrants deeper exploration.

5.4.3. Institutions and Policy

Institutions must establish ethical frameworks and data governance policies to guide the responsible use of LA and GenAI. Bridging digital gaps in under-resourced contexts is essential to prevent widening educational inequities. Professional development in AI and data literacy should be prioritized for educators, advisors, and policymakers to support informed implementation.

5.5. Looking Forward: Toward Human-Centered, Hybrid Learning Analytics

This review identifies a methodological shift in learning analytics (LA), from retrospective analysis to adaptive, generative, and participatory systems. While ML contributes structure and predictive rigor, GenAI adds adaptability and interactive potential. Their convergence—when guided by educational theory and ethical principles—can foster LA tools that are transparent, equitable, and pedagogically aligned.
Achieving this vision requires moving beyond technical prototypes toward empirical validation, contextual deployment, and active stakeholder collaboration. Future systems must be co-designed, responsive to real educational settings, and focused on empowering both educators and learners.
Ultimately, the future of LA lies not only in algorithmic advancement but in designing systems that serve learning with purpose, transparency, and equity at their core.

6. Study Limitations

Achieving the transformative potential of ML and GenAI in education requires moving beyond technical prototypes toward empirical validation, contextual deployment, and active stakeholder collaboration. Future systems must be co-designed, responsive to authentic educational settings, and focused on empowering both educators and learners.
First, all screening stages (titles/abstracts and full texts) were performed by a single reviewer. While eligibility criteria were clearly defined, this introduces potential selection bias.
Second, although a structured quality appraisal was conducted using the Mixed Methods Appraisal Tool (MMAT), the high methodological heterogeneity across studies limited the possibility of assigning comparable quality scores or excluding low-quality papers. Instead, the appraisal was used descriptively to inform interpretations and increase transparency.
Third, restricting the search to English-language publications may have excluded relevant studies from non-English-speaking regions, potentially limiting the cultural and geographical representativeness of the synthesis. Expanding future reviews to include multilingual sources would enhance global inclusivity.
Finally, although the main corpus focused on peer-reviewed journal articles indexed in major databases, an exploratory scan of LAK, L@S, and EC-TEL proceedings (2018–2025) was conducted. While these records were not formally included in the PRISMA dataset, selected studies were referenced to triangulate findings and highlight emerging developments.
Acknowledging these limitations is key to contextualizing the scope of this review and guiding future research toward greater methodological rigor, inclusiveness, and practical relevance.

7. Conclusions

This review offers a critical synthesis of how machine learning (ML) and generative AI (GenAI) are shaping learning analytics (LA) in higher education. Based on 101 empirical studies (2018–2025), along with insights from recent conferences, it highlights both methodological advances and persistent challenges across academic, technological, and pedagogical dimensions.
Traditional ML models remain central to LA, particularly for performance prediction, dropout detection, and engagement analysis. While robust and interpretable, these models often rely on retrospective, mono-modal data and show limited integration with pedagogical practices. In contrast, GenAI systems—particularly those using large language models (LLMs) like GPT-4—offer promising innovations in personalized feedback and affective modeling, yet remain largely experimental, with minimal real-world validation or theoretical grounding.
Recent efforts to enhance explainability, leverage clustering, and support instructional design mark progress, but critical gaps persist. These include limited use of multimodal and real-time data, weak ethical oversight, minimal educator involvement, and a lack of evidence for sustained impact on learning.
Rather than a paradigm shift, the current trends reveal a transitional phase in which ML and GenAI approaches coexist. This duality calls for hybrid, human-centered systems that combine predictive accuracy with adaptive feedback, grounded in learning theory and responsive to educational context.
Future progress requires a shift from technical performance to pedagogical impact—from opaque experimentation to scalable, explainable, and inclusive implementations. Advancing LA in line with Sustainable Development Goals 4.3 and 4.4 depends not solely on technological innovation, but on institutional support, educator engagement, and critical reflection.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/app15158679/s1, https://doi.org/10.5281/zenodo.15233231: Table S1: Summary of Queries Used in Databases; https://doi.org/10.5281/zenodo.16422602: PRISMA Checklist. Available and https://doi.org/10.5281/zenodo.16416487: MMAT Quality Appraisal at Zenodo.

Author Contributions

Conceptualization, L.E.A.-R. and P.C.S.-M.; methodology, P.C.S.-M.; validation, P.C.S.-M.; formal analysis, M.Á.R.-O.; investigation, M.Á.R.-O., L.E.A.-R. and P.C.S.-M.; data curation, M.Á.R.-O.; writing—original draft preparation, M.Á.R.-O.; writing—review and editing, L.E.A.-R. and P.C.S.-M.; supervision, L.E.A.-R.; project administration, L.E.A.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This publication has been partially funded by the project R+D+I PID2023-147396OB-I00, funded by MCIN/AEI/10.13039/501100011033 and by ERDF, EU.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset supporting this study is openly available at Zenodo: https://doi.org/10.5281/zenodo.16416465.

Acknowledgments

During the preparation of this manuscript, the authors used ChatGPT-4.5 (OpenAI) to improve language clarity and ChatGPT o4-mini-high to assist with Python code for data analysis. The authors have reviewed and edited the output and take full responsibility for the content of this publication. This research was partially supported by institutional resources.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AHCAgglomerative Hierarchical Clustering
ANNArtificial Neural Network
BERTBidirectional Encoder Representations from Transformers
DTDecision Tree
GenAIGenerative Artificial Intelligence
GPTGenerative Pre-trained Transformer
GRUGated Recurrent Unit
LALearning Analytics
LIMELocal Interpretable Model-Agnostic Explanations
LLMLarge Language Model
LMSLearning Management System
LSTMLong Short-Term Memory
MLMachine Learning
MOOCMassive Open Online Course
NLPNatural Language Processing
PLAPredictive Learning Analytics
RFRandom Forest
SHAPSHapley Additive exPlanations
SRLSelf-Regulated Learning
SVMSupport Vector Machine
XAIExplainable Artificial Intelligence

References

  1. Adoption of Data Analytics in Higher Education Learning and Teaching. In Advances in Analytics for Learning and Teaching; Ifenthaler, D., Gibson, D., Eds.; Springer International Publishing: Cham, Switzerland, 2020; ISBN 978-3-030-47391-4. [Google Scholar]
  2. Sghir, N.; Adadi, A.; Lahmer, M. Recent Advances in Predictive Learning Analytics: A Decade Systematic Review (2012–2022). Educ. Inf. Technol. 2023, 28, 8299–8333. [Google Scholar] [CrossRef]
  3. Umer, R.; Susnjak, T.; Mathrani, A.; Suriadi, L. Current Stance on Predictive Analytics in Higher Education: Opportunities, Challenges and Future Directions. Interact. Learn. Environ. 2021, 31, 3503–3528. [Google Scholar] [CrossRef]
  4. United Nations Educational, Scientific and Cultural Organization (UNESCO). Educación 2030: Declaración de Incheon y Marco de Acción para la Realización del Objetivo de Desarrollo Sostenible 4; United Nations Educational, Scientific and Cultural Organization (UNESCO): Paris, France, 2016. [Google Scholar]
  5. Organisation for Economic Co-operation and Development (OECD). Education at a Glance 2022: OECD Indicators. Paris, France, 2022. Available online: https://www.oecd.org/en/publications/education-at-a-glance-2022_3197152b-en.html (accessed on 3 August 2025). ISBN 978-92-64-58258-3.
  6. Castro, R. Blended Learning in Higher Education: Trends and Capabilities. Educ. Inf. Technol. 2019, 24, 2523–2546. [Google Scholar] [CrossRef]
  7. Di Pietro, G. The Determinants of University Dropout in Italy: A Bivariate Probability Model with Sample Selection. Appl. Econ. Lett. 2004, 11, 187–191. [Google Scholar] [CrossRef]
  8. Siemens, G. Learning Analytics: The Emergence of a Discipline. Am. Behav. Sci. 2013, 57, 1380–1400. [Google Scholar] [CrossRef]
  9. El Naqa, I.; Murphy, M.J. What Is Machine Learning? In Machine Learning in Radiation Oncology; El Naqa, I., Li, R., Murphy, M.J., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 3–11. ISBN 978-3-319-18304-6. [Google Scholar]
  10. Alalawi, K.; Athauda, R.; Chiong, R. Contextualizing the Current State of Research on the Use of Machine Learning for Student Performance Prediction: A Systematic Literature Review. Eng. Rep. 2023, 5, e12699. [Google Scholar] [CrossRef]
  11. Albreiki, B.; Zaki, N.; Alashwal, H. A Systematic Literature Review of Student’ Performance Prediction Using Machine Learning Techniques. Educ. Sci. 2021, 11, 552. [Google Scholar] [CrossRef]
  12. Namoun, A.; Alshanqiti, A. Predicting Student Performance Using Data Mining and Learning Analytics Techniques: A Systematic Literature Review. Appl. Sci. 2021, 11, 237. [Google Scholar] [CrossRef]
  13. Borah, A.R.; Nischith, T.N.; Gupta, S. Improved Learning Based on GenAI. In Proceedings of the 2024 2nd International Conference on Intelligent Data Communication Technologies and Internet of Things (IDCIoT), Bengaluru, India, 4–6 January 2024; pp. 1527–1532. [Google Scholar]
  14. Khosravi, H.; Viberg, O.; Kovanovic, V.; Ferguson, R. Generative AI and Learning Analytics. J. Learn. Anal. 2023, 10, 1–6. [Google Scholar] [CrossRef]
  15. Yan, L.; Martinez-Maldonado, R.; Gasevic, D. Generative Artificial Intelligence in Learning Analytics: Contextualising Opportunities and Challenges through the Learning Analytics Cycle. In Proceedings of the 14th Learning Analytics and Knowledge Conference, Kyoto, Japan, 18–22 March 2024; pp. 101–111. [Google Scholar]
  16. Moonsamy, D.; Naicker, N.; Adeliyi, T.T.; Ogunsakin, R.E. A Meta-Analysis of Educational Data Mining for Predicting Students Performance in Programming. Int. J. Adv. Comput. Sci. Appl. 2021, 12, 97–104. [Google Scholar] [CrossRef]
  17. Moreno-Marcos, P.M.; Alario-Hoyos, C.; Munoz-Merino, P.J.; Kloos, C.D. Prediction in MOOCs: A Review and Future Research Directions. IEEE Trans. Learn. Technol. 2019, 12, 384–401. [Google Scholar] [CrossRef]
  18. Topali, P.; Chounta, I.; Martínez-Monés, A.; Dimitriadis, Y. Delving into Instructor-led Feedback Interventions Informed by Learning Analytics in Massive Open Online Courses. J. Comput. Assist. Learn. 2023, 39, 1039–1060. [Google Scholar] [CrossRef]
  19. Zhu, M.; Sari, A.R.; Lee, M.M. Trends and Issues in MOOC Learning Analytics Empirical Research: A Systematic Literature Review (2011–2021). Educ. Inf. Technol. 2022, 27, 10135–10160. [Google Scholar] [CrossRef]
  20. Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 Statement: An Updated Guideline for Reporting Systematic Reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
  21. Hong, Q.N.; Fàbregues, S.; Bartlett, G.; Boardman, F.; Cargo, M.; Dagenais, P.; Gagnon, M.-P.; Griffiths, F.; Nicolau, B.; O’Cathain, A.; et al. The Mixed Methods Appraisal Tool (MMAT) Version 2018 for Information Professionals and Researchers. Educ. Inf. 2018, 34, 285–291. [Google Scholar] [CrossRef]
  22. Abou Gamie, E.; Abou El-Seoud, S.; Salama, M.A. Comparative Analysis for Boosting Classifiers in the Context of Higher Education. Int. J. Emerg. Technol. Learn. IJET 2020, 15, 16. [Google Scholar] [CrossRef]
  23. Arslan, O.; Xing, W.; Inan, F.A.; Du, H. Understanding Topic Duration in Twitter Learning Communities Using Data Mining. J. Comput. Assist. Learn. 2022, 38, 513–525. [Google Scholar] [CrossRef]
  24. Ayouni, S.; Hajjej, F.; Maddeh, M.; Al-Otaibi, S. A New ML-Based Approach to Enhance Student Engagement in Online Environment. PLoS ONE 2021, 16, e0258788. [Google Scholar] [CrossRef]
  25. Lemay, D.J.; Doleck, T. Predicting Completion of Massive Open Online Course (MOOC) Assignments from Video Viewing Behavior. Interact. Learn. Environ. 2022, 30, 1782–1793. [Google Scholar] [CrossRef]
  26. Moreno-Marcos, P.M.; Muñoz-Merino, P.J.; Alario-Hoyos, C.; Estévez-Ayres, I.; Delgado Kloos, C. Analysing the Predictive Power for Anticipating Assignment Grades in a Massive Open Online Course. Behav. Inf. Technol. 2018, 37, 1021–1036. [Google Scholar] [CrossRef]
  27. Cristea, A.I.; Alamri, A.; Alshehri, M.; Pereira, F.D.; Toda, A.M.; de Oliveira, E.H.T.; Stewart, C. The Engage Taxonomy: SDT-Based Measurable Engagement Indicators for MOOCs and Their Evaluation. User Model. User Adapt. Interact. 2023, 34, 323–374. [Google Scholar] [CrossRef]
  28. Onan, A. Sentiment Analysis on Massive Open Online Course Evaluations: A Text Mining and Deep Learning Approach. Comput. Appl. Eng. Educ. 2021, 29, 572–589. [Google Scholar] [CrossRef]
  29. Onan, A.; Toçoğlu, M.A. Weighted Word Embeddings and Clustering-based Identification of Question Topics in MOOC Discussion Forum Posts. Comput. Appl. Eng. Educ. 2021, 29, 675–689. [Google Scholar] [CrossRef]
  30. Du, X.; Zhang, M.; Shelton, B.E.; Hung, J.-L. Learning Anytime, Anywhere: A Spatio-Temporal Analysis for Online Learning. Interact. Learn. Environ. 2022, 30, 34–48. [Google Scholar] [CrossRef]
  31. Moubayed, A.; Injadat, M.; Shami, A.; Lutfiyya, H. Student Engagement Level in E-Learning Environment: Clustering Using K-Means 2020. Am. J. Distance Educ. 2020, 34, 137–156. [Google Scholar] [CrossRef]
  32. Lahza, H.; Khosravi, H.; Demartini, G. Analytics of Learning Tactics and Strategies in an Online Learnersourcing Environment. J. Comput. Assist. Learn. 2023, 39, 94–112. [Google Scholar] [CrossRef]
  33. Akçapınar, G.; Altun, A.; Aşkar, P. Using Learning Analytics to Develop Early-Warning System for at-Risk Students. Int. J. Educ. Technol. High. Educ. 2019, 16, 40. [Google Scholar] [CrossRef]
  34. Dass, S.; Gary, K.; Cunningham, J. Predicting Student Dropout in Self-Paced MOOC Course Using Random Forest Model. Information 2021, 12, 476. [Google Scholar] [CrossRef]
  35. Gupta, A.; Garg, D.; Kumar, P. An Ensembling Model for Early Identification of At-risk Students in Higher Education. Comput. Appl. Eng. Educ. 2022, 30, 589–608. [Google Scholar] [CrossRef]
  36. Kostopoulos, G.; Kotsiantis, S.; Pierrakeas, C.; Koutsonikos, G.; Gravvanis, G.A. Forecasting Students’ Success in an Open University. Int. J. Learn. Technol. 2018, 13, 26. [Google Scholar] [CrossRef]
  37. Karalar, H.; Kapucu, C.; Gürüler, H. Predicting Students at Risk of Academic Failure Using Ensemble Model during Pandemic in a Distance Learning System. Int. J. Educ. Technol. High. Educ. 2021, 18, 63. [Google Scholar] [CrossRef]
  38. Yu, C.-C.; Wu, Y. (Leon) Early Warning System for Online STEM Learning—A Slimmer Approach Using Recurrent Neural Networks. Sustainability 2021, 13, 12461. [Google Scholar] [CrossRef]
  39. Al-Sulami, A.; Al-Masre, M.; Al-Malki, N. Predicting At-Risk Students’ Performance Based on LMS Activity Using Deep Learning. Int. J. Adv. Comput. Sci. Appl. 2023, 14, 1210–1220. [Google Scholar] [CrossRef]
  40. Mubarak, A.A.; Cao, H.; Zhang, W. Prediction of Students’ Early Dropout Based on Their Interaction Logs in Online Learning Environment. Interact. Learn. Environ. 2022, 30, 1414–1433. [Google Scholar] [CrossRef]
  41. Bayazit, A.; Apaydin, N.; Gonullu, I. Predicting At-Risk Students in an Online Flipped Anatomy Course Using Learning Analytics. Educ. Sci. 2022, 12, 581. [Google Scholar] [CrossRef]
  42. Esteban, A.; Romero, C.; Zafra, A. Assignments as Influential Factor to Improve the Prediction of Student Performance in Online Courses. Appl. Sci. 2021, 11, 10145. [Google Scholar] [CrossRef]
  43. Talamás-Carvajal, J.A.; Ceballos, H.G. A Stacking Ensemble Machine Learning Method for Early Identification of Students at Risk of Dropout. Educ. Inf. Technol. 2023, 28, 12169–12189. [Google Scholar] [CrossRef]
  44. Karlos, S.; Kostopoulos, G.; Kotsiantis, S. Predicting and Interpreting Students’ Grades in Distance Higher Education through a Semi-Regression Method. Appl. Sci. 2020, 10, 8413. [Google Scholar] [CrossRef]
  45. Herodotou, C.; Hlosta, M.; Boroowa, A.; Rienties, B.; Zdrahal, Z.; Mangafa, C. Empowering Online Teachers through Predictive Learning Analytics. Br. J. Educ. Technol. 2019, 50, 3064–3079. [Google Scholar] [CrossRef]
  46. Sousa-Vieira, M.E.; López-Ardao, J.C.; Fernández-Veiga, M.; Rodríguez-Rubio, R.F. Study of the Impact of Social Learning and Gamification Methodologies on Learning Results in Higher Education. Comput. Appl. Eng. Educ. 2023, 31, 131–153. [Google Scholar] [CrossRef]
  47. Zamecnik, A.; Kovanović, V.; Joksimović, S.; Liu, L. Exploring Non-Traditional Learner Motivations and Characteristics in Online Learning: A Learner Profile Study. Comput. Educ. Artif. Intell. 2022, 3, 100051. [Google Scholar] [CrossRef]
  48. Asad, R.; Altaf, S.; Ahmad, S.; Mahmoud, H.; Huda, S.; Iqbal, S. Machine Learning-Based Hybrid Ensemble Model Achieving Precision Education for Online Education Amid the Lockdown Period of COVID-19 Pandemic in Pakistan. Sustainability 2023, 15, 5431. [Google Scholar] [CrossRef]
  49. Deho, O.B.; Joksimovic, S.; Li, J.; Zhan, C.; Liu, J.; Liu, L. Should Learning Analytics Models Include Sensitive Attributes? Explaining the Why. IEEE Trans. Learn. Technol. 2023, 16, 560–572. [Google Scholar] [CrossRef]
  50. Imhof, C.; Comsa, I.-S.; Hlosta, M.; Parsaeifard, B.; Moser, I.; Bergamin, P. Prediction of Dilatory Behavior in eLearning: A Comparison of Multiple Machine Learning Models. IEEE Trans. Learn. Technol. 2022, 16, 648–663. [Google Scholar] [CrossRef]
  51. Dake, D.K.; Gyimah, E. Using Sentiment Analysis to Evaluate Qualitative Students’ Responses. Educ. Inf. Technol. 2023, 28, 4629–4647. [Google Scholar] [CrossRef]
  52. Mai, T.T.; Bezbradica, M.; Crane, M. Learning Behaviours Data in Programming Education: Community Analysis and Outcome Prediction with Cleaned Data. Future Gener. Comput. Syst. 2022, 127, 42–55. [Google Scholar] [CrossRef]
  53. Tsiakmaki, M.; Kostopoulos, G.; Kotsiantis, S.; Ragos, O. Fuzzy-Based Active Learning for Predicting Student Academic Performance Using autoML: A Step-Wise Approach. J. Comput. High. Educ. 2021, 33, 635–667. [Google Scholar] [CrossRef]
  54. Flor, M.; Andrews-Todd, J. Towards Automatic Annotation of Collaborative Problem-solving Skills in Technology-enhanced Environments. J. Comput. Assist. Learn. 2022, 38, 1434–1447. [Google Scholar] [CrossRef]
  55. Cui, Y.; Chen, F.; Shiri, A. Scale up Predictive Models for Early Detection of At-Risk Students: A Feasibility Study. Inf. Learn. Sci. 2020, 121, 97–116. [Google Scholar] [CrossRef]
  56. Polyzou, A.; Karypis, G. Feature Extraction for Next-Term Prediction of Poor Student Performance. IEEE Trans. Learn. Technol. 2019, 12, 237–248. [Google Scholar] [CrossRef]
  57. Wakelam, E.; Jefferies, A.; Davey, N.; Sun, Y. The Potential for Student Performance Prediction in Small Cohorts with Minimal Available Attributes. Br. J. Educ. Technol. 2020, 51, 347–370. [Google Scholar] [CrossRef]
  58. Abu Zohair, L.M. Prediction of Student’s Performance by Modelling Small Dataset Size. Int. J. Educ. Technol. High. Educ. 2019, 16, 27. [Google Scholar] [CrossRef]
  59. Damuluri, S.; Islam, K.; Ahmadi, P.; Qureshi, N.S. Analyzing Navigational Data and Predicting Student Grades Using Support Vector Machine. Emerg. Sci. J. 2020, 4, 243–252. [Google Scholar] [CrossRef]
  60. Nuankaew, P. Dropout Situation of Business Computer Students, University of Phayao. Int. J. Emerg. Technol. Learn. IJET 2019, 14, 115. [Google Scholar] [CrossRef]
  61. Azcona, D.; Hsiao, I.-H.; Smeaton, A.F. Detecting Students-at-Risk in Computer Programming Classes with Learning Analytics from Students’ Digital Footprints. User Model. User Adapt. Interact. 2019, 29, 759–788. [Google Scholar] [CrossRef]
  62. Wu, J.-Y.; Hsiao, Y.-C.; Nian, M.-W. Using Supervised Machine Learning on Large-Scale Online Forums to Classify Course-Related Facebook Messages in Predicting Learning Achievement within the Personal Learning Environment. Interact. Learn. Environ. 2018, 28, 65–80. [Google Scholar] [CrossRef]
  63. Crivei, L.M.; Ionescu, V.-S.; Czibula, G. An Analysis of Supervised Learning Methods for Predicting Students’ Performance in Academic Environments 2019. ICIC Express Lett 2019, 13, 181–189. [Google Scholar]
  64. Raković, M.; Winne, P.H.; Marzouk, Z.; Chang, D. Automatic Identification of Knowledge-transforming Content in Argument Essays Developed from Multiple Sources. J. Comput. Assist. Learn. 2021, 37, 903–924. [Google Scholar] [CrossRef]
  65. Spikol, D.; Ruffaldi, E.; Dabisias, G.; Cukurova, M. Supervised Machine Learning in Multimodal Learning Analytics for Estimating Success in Project-based Learning. J. Comput. Assist. Learn. 2018, 34, 366–377. [Google Scholar] [CrossRef]
  66. Bertolini, R.; Finch, S.J.; Nehm, R.H. Testing the Impact of Novel Assessment Sources and Machine Learning Methods on Predictive Outcome Modeling in Undergraduate Biology. J. Sci. Educ. Technol. 2021, 30, 193–209. [Google Scholar] [CrossRef]
  67. Ding, X.; Larson, E.C.; Doyle, A.; Donahoo, K.; Rajgopal, R.; Bing, E. EduAware: Using Tablet-Based Navigation Gestures to Predict Learning Module Performance. Interact. Learn. Environ. 2021, 29, 720–732. [Google Scholar] [CrossRef]
  68. Okike, E.U.; Mogorosi, M. Educational Data Mining for Monitoring and Improving Academic Performance at University Levels. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 570–581. [Google Scholar] [CrossRef]
  69. Everaert, P.; Opdecam, E.; van der Heijden, H. Predicting First-Year University Progression Using Early Warning Signals from Accounting Education: A Machine Learning Approach. Account. Educ. 2022, 33, 1–26. [Google Scholar] [CrossRef]
  70. Nkhoma, C.; Dang-Pham, D.; Hoang, A.-P.; Nkhoma, M.; Le-Hoai, T.; Thomas, S. Learning Analytics Techniques and Visualisation with Textual Data for Determining Causes of Academic Failure. Behav. Inf. Technol. 2020, 39, 808–823. [Google Scholar] [CrossRef]
  71. Li, F.; Lu, Y.; Ma, Q.; Gao, J.; Wang, Z.; Bai, L. SPOC Online Video Learning Clustering Analysis: Identifying Learners’ Group Behavior Characteristics. Comput. Appl. Eng. Educ. 2023, 31, 1059–1077. [Google Scholar] [CrossRef]
  72. Verdú, M.J.; Regueras, L.M.; de Castro, J.P.; Verdú, E. Clustering of LMS Use Strategies with Autoencoders. Appl. Sci. 2023, 13, 7334. [Google Scholar] [CrossRef]
  73. Alsayed, A.O.; Rahim, M.S.M.; AlBidewi, I.; Hussain, M.; Jabeen, S.H.; Alromema, N.; Hussain, S.; Jibril, M.L. Selection of the Right Undergraduate Major by Students Using Supervised Learning Techniques. Appl. Sci. 2021, 11, 10639. [Google Scholar] [CrossRef]
  74. Mai, T.L.; Chung, M.T.; Le, V.T.; Thoai, N. From Transcripts to Insights for Recommending the Curriculum to University Students. SN Comput. Sci. 2020, 1, 323. [Google Scholar] [CrossRef]
  75. Zabriskie, C.; Yang, J.; DeVore, S.; Stewart, J. Using Machine Learning to Predict Physics Course Outcomes. Phys. Rev. Phys. Educ. Res. 2019, 15, 020120. [Google Scholar] [CrossRef]
  76. Fateh Allah, A.G. Using Machine Learning To Support Students’ Academic Decisions. J. Theor. Appl. Inf. Technol. 2020, 98, 3778–3796. [Google Scholar]
  77. Hassan, H.; Ahmad, N.B.; Anuar, S. Improved Students’ Performance Prediction for Multi-Class Imbalanced Problems Using Hybrid and Ensemble Approach in Educational Data Mining. J. Phys. Conf. Ser. 2020, 1529, 052041. [Google Scholar] [CrossRef]
  78. Hussain, S.; Ayoub, M.; Jilani, G.; Yu, Y.; Khan, A.; Wahid, J.A.; Butt, M.F.A.; Yang, G.; Moller, D.P.F.; Weiyan, H. Aspect2Labels: A Novelistic Decision Support System for Higher Educational Institutions by Using Multi-Layer Topic Modelling Approach. Expert Syst. Appl. 2022, 209, 118119. [Google Scholar] [CrossRef]
  79. Nasa-Ngium, P.; Nuankaew, W.S.; Nuankaew, P. Analyzing and Tracking Student Educational Program Interests on Social Media with Chatbots Platform and Text Analytics. Int. J. Interact. Mob. Technol. IJIM 2023, 17, 4–21. [Google Scholar] [CrossRef]
  80. Liu, Z.; Yang, C.; Rüdian, S.; Liu, S.; Zhao, L.; Wang, T. Temporal Emotion-Aspect Modeling for Discovering What Students Are Concerned about in Online Course Forums. Interact. Learn. Environ. 2019, 27, 598–627. [Google Scholar] [CrossRef]
  81. Maldonado, E.; Seehusen, V. Data Mining Student Choices: A New Approach to Business Curriculum Planning. J. Educ. Bus. 2018, 93, 196–203. [Google Scholar] [CrossRef]
  82. Guerrero-Higueras, Á.M.; DeCastro-García, N.; Rodriguez-Lera, F.J.; Matellán, V.; Conde, M.Á. Predicting Academic Success through Students’ Interaction with Version Control Systems. Open Comput. Sci. 2019, 9, 243–251. [Google Scholar] [CrossRef]
  83. Liao, C.-H.; Wu, J.-Y. Deploying Multimodal Learning Analytics Models to Explore the Impact of Digital Distraction and Peer Learning on Student Performance. Comput. Educ. 2022, 190, 104599. [Google Scholar] [CrossRef]
  84. Wu, J.-Y. Learning Analytics on Structured and Unstructured Heterogeneous Data Sources: Perspectives from Procrastination, Help-Seeking, and Machine-Learning Defined Cognitive Engagement. Comput. Educ. 2021, 163, 104066. [Google Scholar] [CrossRef]
  85. Cagliero, L.; Canale, L.; Farinetti, L.; Baralis, E.; Venuto, E. Predicting Student Academic Performance by Means of Associative Classification. Appl. Sci. 2021, 11, 1420. [Google Scholar] [CrossRef]
  86. Sassirekha, M.S.; Vijayalakshmi, S. Predicting the Academic Progression in Student’s Standpoint Using Machine Learning. Automatika 2022, 63, 605–617. [Google Scholar] [CrossRef]
  87. Zheng, L.; Zhong, L.; Niu, J. Effects of Personalised Feedback Approach on Knowledge Building, Emotions, Co-Regulated Behavioural Patterns and Cognitive Load in Online Collaborative Learning. Assess. Eval. High. Educ. 2022, 47, 109–125. [Google Scholar] [CrossRef]
  88. Talebinamvar, M.; Zarrabi, F. Clustering Students’ Writing Behaviors Using Keystroke Logging: A Learning Analytic Approach in EFL Writing. Lang. Test. Asia 2022, 12, 6. [Google Scholar] [CrossRef]
  89. Araka, E.; Oboko, R.; Maina, E.; Gitonga, R. Using Educational Data Mining Techniques to Identify Profiles in Self-Regulated Learning: An Empirical Evaluation. Int. Rev. Res. Open Distrib. Learn. 2022, 23, 131–162. [Google Scholar] [CrossRef]
  90. Choi, Y.; Kim, J. Learning Analytics for Diagnosing Cognitive Load in E-Learning Using Bayesian Network Analysis. Sustainability 2021, 13, 10149. [Google Scholar] [CrossRef]
  91. Afzaal, M.; Nouri, J.; Zia, A.; Papapetrou, P.; Fors, U.; Wu, Y.; Li, X.; Weegar, R. Explainable AI for Data-Driven Feedback and Intelligent Action Recommendations to Support Students Self-Regulation. Front. Artif. Intell. 2021, 4, 723447. [Google Scholar] [CrossRef]
  92. Ramaswami, G.; Susnjak, T.; Mathrani, A. Supporting Students’ Academic Performance Using Explainable Machine Learning with Automated Prescriptive Analytics. Big Data Cogn. Comput. 2022, 6, 105. [Google Scholar] [CrossRef]
  93. Dirin, A.; Saballe, C.A. Machine Learning Models to Predict Students’ Study Path Selection. Int. J. Interact. Mob. Technol. IJIM 2022, 16, 158–183. [Google Scholar] [CrossRef]
  94. Hung, H.-C.; Liu, I.-F.; Liang, C.-T.; Su, Y.-S. Applying Educational Data Mining to Explore Students’ Learning Patterns in the Flipped Learning Approach for Coding Education. Symmetry 2020, 12, 213. [Google Scholar] [CrossRef]
  95. Qushem, U.B.; Oyelere, S.S.; Akçapınar, G.; Kaliisa, R.; Laakso, M.-J. Unleashing the Power of Predictive Analytics to Identify At-Risk Students in Computer Science. Technol. Knowl. Learn. 2023, 29, 1385–1400. [Google Scholar] [CrossRef]
  96. Almasri, A.R.; Yahaya, N.A.; Abu-Naser, S.S. Instructor Performance Modeling For Predicting Student Satisfaction Using Machine Learning—Preliminary Results. J. Theor. Appl. Inf. Technol. 2022, 100, 5481–5496. [Google Scholar]
  97. Samadi, M.A.; Jaquay, S.; Lin, Y.; Tajik, E.; Park, S.; Nixon, N. Minds and Machines Unite: Deciphering Social and Cognitive Dynamics in Collaborative Problem Solving with AI. In Proceedings of the 14th Learning Analytics and Knowledge Conference, Kyoto, Japan, 18–22 March 2024; pp. 885–891. [Google Scholar]
  98. Li, T.; Nath, D.; Cheng, Y.; Fan, Y.; Li, X.; Raković, M.; Khosravi, H.; Swiecki, Z.; Tsai, Y.-S.; Gašević, D. Turning Real-Time Analytics into Adaptive Scaffolds for Self-Regulated Learning Using Generative Artificial Intelligence. In Proceedings of the 15th International Learning Analytics and Knowledge Conference, Dublin, Ireland, 3–7 March 2025; pp. 667–679. [Google Scholar]
  99. Lewis, A. Multimodal Large Language Models for Inclusive Collaboration Learning Tasks. In Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies: Student Research Workshop, Seattle, WA, USA, 10–15 July 2022; pp. 202–210. [Google Scholar]
  100. Dai, W.; Tsai, Y.-S.; Lin, J.; Aldino, A.; Jin, H.; Li, T.; Gašević, D.; Chen, G. Assessing the Proficiency of Large Language Models in Automatic Feedback Generation: An Evaluation Study. Comput. Educ. Artif. Intell. 2024, 7, 100299. [Google Scholar] [CrossRef]
  101. Civit, M.; Escalona, M.J.; Cuadrado, F.; Reyes-de-Cozar, S. Class Integration of ChatGPT and Learning Analytics for Higher Education. Expert Syst. 2024, 41, e13703. [Google Scholar] [CrossRef]
  102. Xie, Z.; Wu, X.; Xie, Y. Can Interaction with Generative Artificial Intelligence Enhance Learning Autonomy? A Longitudinal Study from Comparative Perspectives of Virtual Companionship and Knowledge Acquisition Preferences. J. Comput. Assist. Learn. 2024, 40, 2369–2384. [Google Scholar] [CrossRef]
  103. Milesi, M.E.; Alfredo, R.; Echeverria, V.; Yan, L.; Zhao, L.; Tsai, Y.-S.; Martinez-Maldonado, R. “It’s Really Enjoyable to See Me Solve the Problem like a Hero”: GenAI-Enhanced Data Comics as a Learning Analytics Tool. In Proceedings of the Extended Abstracts of the CHI Conference on Human Factors in Computing Systems, Honolulu, HI, USA, 11–16 May 2024; pp. 1–7. [Google Scholar]
  104. Ilagan, J.B.R.; Ilagan, J.R.S.; Rodrigo, M.M.T. Ethical Education Data Mining Framework for Analyzing and Evaluating Large Language Model-Based Conversational Intelligent Tutoring Systems for Management and Entrepreneurship Courses. In Proceedings of Ninth International Congress on Information and Communication Technology, London, UK, 19–22 February 2024; Yang, X.-S., Sherratt, S., Dey, N., Joshi, A., Eds.; Lecture Notes in Networks and Systems. Springer Nature: Singapore, 2024; Volume 1011, pp. 61–71, ISBN 978-981-9745-80-7. [Google Scholar]
  105. Nguyen, A.; Ilesanmi, F.; Dang, B.; Vuorenmaa, E.; Järvelä, S. Hybrid Intelligence in Academic Writing: Examining Self-Regulated Learning Patterns in an AI-Assisted Writing Task. In Frontiers in Artificial Intelligence and Applications; Lorig, F., Tucker, J., Dahlgren Lindström, A., Dignum, F., Murukannaiah, P., Theodorou, A., Yolum, P., Eds.; IOS Press: Amsterdam, The Netherlands, 2024; ISBN 978-1-64368-522-9. [Google Scholar]
  106. Cheng, Y.; Lyons, K.; Chen, G.; Gašević, D.; Swiecki, Z. Evidence-Centered Assessment for Writing with Generative AI. In Proceedings of the 14th Learning Analytics and Knowledge Conference, Kyoto, Japan, 18–22 March 2024; pp. 178–188. [Google Scholar]
  107. Fahl, W. GraphWiseLearn: Personalized Learning Through Semantified TEL, Leveraging QA-Enhanced LLM-Generated Content. In The Semantic Web: ESWC 2024 Satellite Events; Meroño Peñuela, A., Corcho, O., Groth, P., Simperl, E., Tamma, V., Nuzzolese, A.G., Poveda-Villalón, M., Sabou, M., Presutti, V., Celino, I., Eds.; Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2025; Volume 15345, pp. 74–83. ISBN 978-3-031-78954-0. [Google Scholar]
  108. Pozdniakov, S.; Brazil, J.; Abdi, S.; Bakharia, A.; Sadiq, S.; Gašević, D.; Denny, P.; Khosravi, H. Large Language Models Meet User Interfaces: The Case of Provisioning Feedback. Comput. Educ. Artif. Intell. 2024, 7, 100289. [Google Scholar] [CrossRef]
  109. Chen, A.; Xiang, M.; Zhou, J.; Jia, J.; Shang, J.; Li, X.; Gašević, D.; Fan, Y. Unpacking Help-Seeking Process through Multimodal Learning Analytics: A Comparative Study of ChatGPT vs Human Expert. Comput. Educ. 2025, 226, 105198. [Google Scholar] [CrossRef]
  110. Fan, Y.; Tang, L.; Le, H.; Shen, K.; Tan, S.; Zhao, Y.; Shen, Y.; Li, X.; Gašević, D. Beware of Metacognitive Laziness: Effects of Generative Artificial Intelligence on Learning Motivation, Processes, and Performance. Br. J. Educ. Technol. 2025, 56, 489–530. [Google Scholar] [CrossRef]
  111. Cohn, C.; Snyder, C.; Fonteles, J.H.; TS, A.; Montenegro, J.; Biswas, G. A Multimodal Approach to Support Teacher, Researcher and AI Collaboration in STEM+C Learning Environments. Br. J. Educ. Technol. 2025, 56, 595–620. [Google Scholar] [CrossRef]
  112. Hu, Y.; Giacaman, N.; Donald, C. Enhancing Trust in Generative AI: Investigating Explainability of LLMs to Analyse Confusion in MOOC Discussions. In Proceedings of the Joint Proceedings of LAK 2024 Workshops, Kyoto, Japan, 18–22 March 2024; Volume 3667, pp. 195–204. [Google Scholar]
  113. Valle Torre, M. Learning Sequence Analytics for Support in Learning Tasks. CEUR Workshop Proc. CEUR-WSorg 2023, 3539, 57–62. [Google Scholar]
  114. Zhang, L.; Lin, J.; Sabatini, J.; Borchers, C.; Weitekamp, D.; Cao, M.; Hollander, J.; Hu, X.; Graesser, A.C. Data Augmentation for Sparse Multidimensional Learning Performance Data Using Generative AI. IEEE Trans. Learn. Technol. 2025, 18, 145–164. [Google Scholar] [CrossRef]
  115. Singh, A.; Brooks, C.; Wang, X.; Li, W.; Kim, J.; Wilson, D. Bridging Learnersourcing and AI: Exploring the Dynamics of Student-AI Collaborative Feedback Generation. In Proceedings of the 14th Learning Analytics and Knowledge Conference, Kyoto, Japan, 18–22 March 2024; pp. 742–748. [Google Scholar]
  116. Opanasenko, Y.; Bardone, E.; Pedaste, M.; Siiman, L.A. Sequence Analysis-Enhanced AI: Transforming Interactive E-Book Data into Educational Insights for Teachers. Educ. Sci. 2024, 15, 28. [Google Scholar] [CrossRef]
  117. Jin, Y.; Yang, K.; Yan, L.; Echeverria, V.; Zhao, L.; Alfredo, R.; Milesi, M.; Fan, J.X.; Li, X.; Gasevic, D.; et al. Chatting with a Learning Analytics Dashboard: The Role of Generative AI Literacy on Learner Interaction with Conventional and Scaffolding Chatbots. In Proceedings of the 15th International Learning Analytics and Knowledge Conference, Dublin, Ireland, 3–7 March 2025; pp. 579–590. [Google Scholar]
  118. Wang, Z.; Lin, W.; Hu, X. Self-Service Teacher-Facing Learning Analytics Dashboard with Large Language Models. In Proceedings of the 15th International Learning Analytics and Knowledge Conference, Dublin, Ireland, 3–7 March 2025; pp. 824–830. [Google Scholar]
  119. Volkmann, N. EduBot Unleashed—Elevating Digital Competence in Online Collaborative Learning. In Proceedings of the 2024 21st International Conference on Information Technology Based Higher Education and Training (ITHET), Paris, France, 6–8 November 2024; pp. 1–9. [Google Scholar]
  120. Misiejuk, K.; Kaliisa, R.; Scianna, J. Augmenting Assessment with AI Coding of Online Student Discourse: A Question of Reliability. Comput. Educ. Artif. Intell. 2024, 6, 100216. [Google Scholar] [CrossRef]
  121. Woollaston, S.; Flanagan, B.; Ocheja, P.; Toyokawa, Y.; Ogata, H. ARCHIE: Exploring Language Learner Behaviors in LLM Chatbot-Supported Active Reading Log Data with Epistemic Network Analysis. In Proceedings of the 15th International Learning Analytics and Knowledge Conference, Dublin, Ireland, 3–7 March 2025; pp. 642–654. [Google Scholar]
  122. Ouyang, F.; Guo, M.; Zhang, N.; Bai, X.; Jiao, P. Comparing the Effects of Instructor Manual Feedback and ChatGPT Intelligent Feedback on Collaborative Programming in China’s Higher Education. IEEE Trans. Learn. Technol. 2024, 17, 2173–2185. [Google Scholar] [CrossRef]
  123. Huang, K.; Ferreira Mello, R.; Pereira Junior, C.; Rodrigues, L.; Baars, M.; Viberg, O. That’s What RoBERTa Said: Explainable Classification of Peer Feedback. In Proceedings of the 15th International Learning Analytics and Knowledge Conference, Dublin, Ireland, 3–7 March 2025; pp. 880–886. [Google Scholar]
  124. Pishtari, G.; Sarmiento-Márquez, E.M.; Rodríguez-Triana, M.J.; Wagner, M.; Ley, T. Evaluating the Impact and Usability of an AI-Driven Feedback System for Learning Design. In Responsive and Sustainable Educational Futures; Viberg, O., Jivet, I., Muñoz-Merino, P.J., Perifanou, M., Papathoma, T., Eds.; Lecture Notes in Computer Science; Springer Nature: Cham, Switzerland, 2023; Volume 14200, pp. 324–338. ISBN 978-3-031-42681-0. [Google Scholar]
  125. Qu, T.; Yang, Z. Overview of Artificial Intelligence Applications in Educational Research. In Proceedings of the 2024 International Symposium on Artificial Intelligence for Education, Xian, China, 6–8 September 2024; pp. 101–108. [Google Scholar]
  126. Peña-Ayala, A. Learning Analytics: A Glance of Evolution, Status, and Trends According to a Proposed Taxonomy. WIREs Data Min. Knowl. Discov. 2018, 8, e1243. [Google Scholar] [CrossRef]
  127. Zawacki-Richter, O.; Marín, V.I.; Bond, M.; Gouverneur, F. Systematic Review of Research on Artificial Intelligence Applications in Higher Education—Where Are the Educators? Int. J. Educ. Technol. High. Educ. 2019, 16, 39. [Google Scholar] [CrossRef]
  128. Renz, A.; Hilbig, R. Prerequisites for Artificial Intelligence in Further Education: Identification of Drivers, Barriers, and Business Models of Educational Technology Companies. Int. J. Educ. Technol. High. Educ. 2020, 17, 14. [Google Scholar] [CrossRef]
  129. Salas-Pilco, S.Z.; Yang, Y. Artificial Intelligence Applications in Latin American Higher Education: A Systematic Review. Int. J. Educ. Technol. High. Educ. 2022, 19, 21. [Google Scholar] [CrossRef]
  130. Glandorf, D.; Lee, H.R.; Orona, G.A.; Pumptow, M.; Yu, R.; Fischer, C. Temporal and Between-Group Variability in College Dropout Prediction. In Proceedings of the 14th Learning Analytics and Knowledge Conference, Kyoto, Japan, 18–22 March 2024; pp. 486–497. [Google Scholar]
  131. Poellhuber, L.-V.; Poellhuber, B.; Desmarais, M.; Leger, C.; Roy, N.; Manh-Chien Vu, M. Cluster-Based Performance of Student Dropout Prediction as a Solution for Large Scale Models in a Moodle LMS. In Proceedings of the LAK23: 13th International Learning Analytics and Knowledge Conference, Arlington, TX, USA, 13–17 March 2023; pp. 592–598. [Google Scholar]
  132. Buitrago-Ropero, M.E.; Ramírez-Montoya, M.S.; Laverde, A.C. Digital Footprints (2005–2019): A Systematic Mapping of Studies in Education. Interact. Learn. Environ. 2023, 31, 876–889. [Google Scholar] [CrossRef]
  133. Baek, C.; Doleck, T. Educational Data Mining versus Learning Analytics: A Review of Publications from 2015 to 2019. Interact. Learn. Environ. 2021, 31, 3828–3850. [Google Scholar] [CrossRef]
  134. Ley, T.; Tammets, K.; Pishtari, G.; Chejara, P.; Kasepalu, R.; Khalil, M.; Saar, M.; Tuvi, I.; Väljataga, T.; Wasson, B. Towards a Partnership of Teachers and Intelligent Learning Technology: A Systematic Literature Review of Model-based Learning Analytics. J. Comput. Assist. Learn. 2023, 39, 1397–1417. [Google Scholar] [CrossRef]
  135. Aguilar-Esteva, V.; Acosta-Banda, A.; Carreño Aguilera, R.; Patiño Ortiz, M. Sustainable Social Development through the Use of Artificial Intelligence and Data Science in Education during the COVID Emergency: A Systematic Review Using PRISMA. Sustainability 2023, 15, 6498. [Google Scholar] [CrossRef]
  136. Carless, D.; Boud, D. The Development of Student Feedback Literacy: Enabling Uptake of Feedback. Assess. Eval. High. Educ. 2018, 43, 1315–1325. [Google Scholar] [CrossRef]
  137. Molloy, E.; Boud, D.; Henderson, M. Developing a Learning-Centred Framework for Feedback Literacy. Assess. Eval. High. Educ. 2020, 45, 527–540. [Google Scholar] [CrossRef]
  138. Winstone, N.E.; Nash, R.A.; Parker, M.; Rowntree, J. Supporting Learners’ Agentic Engagement With Feedback: A Systematic Review and a Taxonomy of Recipience Processes. Educ. Psychol. 2017, 52, 17–37. [Google Scholar] [CrossRef]
  139. Zimmerman, B.J. Becoming a Self-Regulated Learner: An Overview. Theory Pract. 2002, 41, 64–70. [Google Scholar] [CrossRef]
  140. Panadero, E. A Review of Self-Regulated Learning: Six Models and Four Directions for Research. Front. Psychol. 2017, 8, 422. [Google Scholar] [CrossRef]
  141. Black, P.; Wiliam, D. Developing the Theory of Formative Assessment. Educ. Assess. Eval. Account. 2009, 21, 5–31. [Google Scholar] [CrossRef]
Figure 1. PRISMA flowchart for study selection (2018–2025), detailing inclusion of 101 studies from an initial pool of 9590 records.
Figure 1. PRISMA flowchart for study selection (2018–2025), detailing inclusion of 101 studies from an initial pool of 9590 records.
Applsci 15 08679 g001
Figure 2. Temporal trends in learning analytics studies using traditional machine learning (left) and generative AI (right) in higher education between 2018 and 2025.
Figure 2. Temporal trends in learning analytics studies using traditional machine learning (left) and generative AI (right) in higher education between 2018 and 2025.
Applsci 15 08679 g002
Figure 3. Geographical distribution of learning analytics studies in higher education (2018–2025) by AI model type.
Figure 3. Geographical distribution of learning analytics studies in higher education (2018–2025) by AI model type.
Applsci 15 08679 g003
Figure 4. Top 10 AI models applied in learning analytics using traditional machine learning techniques, categorized by educational context.
Figure 4. Top 10 AI models applied in learning analytics using traditional machine learning techniques, categorized by educational context.
Applsci 15 08679 g004
Figure 5. Top generative AI models used in learning analytics research.
Figure 5. Top generative AI models used in learning analytics research.
Applsci 15 08679 g005
Table 1. Challenges and future directions in ML and GenAI for learning Analytics.
Table 1. Challenges and future directions in ML and GenAI for learning Analytics.
ChallengeDescriptionFuture Direction
Model opacityDeep models lack transparency, limiting educational use.Advance Explainable AI (XAI) in LA.
Post hoc modelingRetrospective models restrict real-time interventions.Develop online/incremental modeling and real-time feedback systems.
Modality-agnostic designMethods are reused without context adaptation across modalities.Create modality-aware models (e.g., sensor, LMS, multimodal).
Geographic biasResearch is concentrated in high-income regions.Broaden global datasets and include non-Western case studies.
Limited institutional integrationReal-world deployment and constraints are rarely addressed.Study adoption, co-design with educators, and report implementations.
Narrow evaluation metricsAccuracy often outweighs pedagogical value in evaluations.Integrate educational impact and instructional alignment.
Ethical blind spotsFairness, consent, and bias remain underexplored.Apply ethical audits and promote fairness-aware ML and data literacy.
Underdeveloped GenAI useGenAI lacks pedagogical grounding and deep integration into LA.Explore adaptive, generative, and co-creative GenAI applications.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rodríguez-Ortiz, M.Á.; Santana-Mancilla, P.C.; Anido-Rifón, L.E. Machine Learning and Generative AI in Learning Analytics for Higher Education: A Systematic Review of Models, Trends, and Challenges. Appl. Sci. 2025, 15, 8679. https://doi.org/10.3390/app15158679

AMA Style

Rodríguez-Ortiz MÁ, Santana-Mancilla PC, Anido-Rifón LE. Machine Learning and Generative AI in Learning Analytics for Higher Education: A Systematic Review of Models, Trends, and Challenges. Applied Sciences. 2025; 15(15):8679. https://doi.org/10.3390/app15158679

Chicago/Turabian Style

Rodríguez-Ortiz, Miguel Ángel, Pedro C. Santana-Mancilla, and Luis E. Anido-Rifón. 2025. "Machine Learning and Generative AI in Learning Analytics for Higher Education: A Systematic Review of Models, Trends, and Challenges" Applied Sciences 15, no. 15: 8679. https://doi.org/10.3390/app15158679

APA Style

Rodríguez-Ortiz, M. Á., Santana-Mancilla, P. C., & Anido-Rifón, L. E. (2025). Machine Learning and Generative AI in Learning Analytics for Higher Education: A Systematic Review of Models, Trends, and Challenges. Applied Sciences, 15(15), 8679. https://doi.org/10.3390/app15158679

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop