Next Article in Journal
MH-WMG: A Multi-Head Wavelet-Based MobileNet with Gated Linear Attention for Power Grid Fault Diagnosis
Next Article in Special Issue
Early Prediction of Student Performance Using an Activation Ensemble Deep Neural Network Model
Previous Article in Journal
Velocity-Based Approaches More Accurately Estimated the One-Repetition Maximum (1RM) After Four Weeks of Training Compared to Baseline and Group-Adjusted 1RM Approaches
Previous Article in Special Issue
Predicting Student Dropout from Day One: XGBoost-Based Early Warning System Using Pre-Enrollment Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Intelligent System for Student Performance Prediction: An Educational Data Mining Approach Using Metaheuristic-Optimized LightGBM with SHAP-Based Learning Analytics

by
Abdalhmid Abukader
,
Ahmad Alzubi
and
Oluwatayomi Rereloluwa Adegboye
*
Business Administration Department, Institute of Graduate Research and Studies, University of Mediterranean Karpasia, Mersin 10, Northern Cyprus, Lefkosa 99010, Turkey
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(20), 10875; https://doi.org/10.3390/app152010875
Submission received: 2 September 2025 / Revised: 26 September 2025 / Accepted: 26 September 2025 / Published: 10 October 2025

Abstract

Educational data mining (EDM) plays a crucial role in developing intelligent early warning systems that enable timely interventions to improve student outcomes. This study presents a novel approach to student performance prediction by integrating metaheuristic hyperparameter optimization with explainable artificial intelligence for enhanced learning analytics. While Light Gradient Boosting Machine (LightGBM) demonstrates efficiency in educational prediction tasks, achieving optimal performance requires sophisticated hyperparameter tuning, particularly for complex educational datasets where accuracy, interpretability, and actionable insights are paramount. This research addressed these challenges by implementing and evaluating five nature-inspired metaheuristic algorithms: Fox Algorithm (FOX), Giant Trevally Optimizer (GTO), Particle Swarm Optimization (PSO), Sand Cat Swarm Optimization (SCSO), and Salp Swarm Algorithm (SSA) for automated hyperparameter optimization. Using rigorous experimental methodology with 5-fold cross-validation and 20 independent runs, we assessed predictive performance through comprehensive metrics including Coefficient of Determination (R2), Root Mean Squared Error (RMSE), Mean Squared Error (MSE), Relative Absolute Error (RAE), and Mean Error (ME). Results demonstrate that metaheuristic optimization significantly enhances educational prediction accuracy, with SCSO-LightGBM achieving superior performance with R2 of 0.941. SHapley Additive exPlanations (SHAP) analysis provides crucial interpretability, identifying Attendance, Hours Studied, Previous Scores, and Parental Involvement as dominant predictive factors, offering evidence-based insights for educational stakeholders. The proposed SCSO-LightGBM framework establishes an intelligent, interpretable system that supports data-driven decision-making in educational environments, enabling proactive interventions to enhance student success.

1. Introduction

Historically, economic expansion was primarily attributed to the accumulation of physical capital, with labor treated merely as an inherent factor of production. Productivity was expected to rise through investments in tangible assets such as machinery and equipment [1]. In contrast, the modern growth paradigm unequivocally recognizes the central role of human capital in driving economic development. Numerous studies have demonstrated a strong correlation between human capital accumulation and long-term economic progress [2,3,4]. Within empirical research, years of schooling are frequently employed as a proxy for measuring human capital [1]. Individuals with higher levels of education gain a comparative advantage in understanding and integrating new technologies and ideas into production processes. Consequently, policymakers in many countries seek to invest in education as a strategy to stimulate economic growth [5]. However, the impact of such policies varies across contexts because of disparities in educational quality. Student academic performance is a key indicator of educational quality, serving as a direct measure of how effectively the institution supports learning and development [6]. Student performance refers to a student’s level of achievement in their educational pursuits. It is commonly evaluated through grades, examination scores, and overall scholastic accomplishments. Predicting student performance has attracted growing attention in education due to its practical value. Early predictions allow instructors to promptly identify underperforming students and intervene before further decline, thereby increasing the likelihood of success [7]. Applications of student performance prediction have proven useful in identifying at-risk students and forecasting potential dropout rates. At the administrative level, such techniques support the efficient allocation of resources, ensuring that funding, staff, and specialized courses are directed to students who need them most. From the learner’s perspective, predictive analytics facilitate individualized learning by tailoring instruction to personal strengths, weaknesses, and learning styles, thereby improving engagement and achievement [8]. Beyond the classroom, performance prediction also informs educational policy. By analyzing aggregate trends, policymakers can optimize investments in teacher training, infrastructure, technology, and pedagogical reforms, ensuring more effective and equitable educational systems.
Educational Data Mining (EDM) is an emerging discipline dedicated to developing methods for analyzing the rapidly growing volumes of data produced by educational institutions and applying these methods to better understand students and their learning environments [9]. Machine learning has found application in several domains [10,11,12]. In practice, EDM often relies on machine learning methods to examine student records, identify trends, and forecast academic outcomes [13]. As shown in the literature, a wide range of machine learning algorithms have been successfully applied for this purpose, including Decision Trees (DT) [14], Random Forests (RF) [15], Logistic Regression [16], Linear Regression [17], Naïve Bayes (NB) [18], Support Vector Machines (SVM) [19], K-Nearest Neighbors (KNN) [20], Artificial Neural Networks (ANN) [21], Bayesian Networks (BN), Discriminant Analysis (DA) [22], and Principal Component Analysis (PCA) [23]. While highly effective, previous studies showed that the performance of machine learning models is highly sensitive to the choice of hyperparameters [24]. Inadequate or suboptimal hyperparameter settings often lead to poor convergence, overfitting, or entrapment in local optima during training, thereby limiting the model’s generalization ability. This limitation can be addressed by integrating metaheuristic algorithms (MAs) that systematically tune the hyperparameters of a given model to enhance its accuracy. By mimicking natural processes, MAs provide robust exploration and exploitation capabilities that directly address the above limitations of standalone ML models [25].
The present study seeks to enhance student performance prediction by integrating metaheuristic optimization with LightGBM, a gradient boosting framework known for efficient data processing, reduced memory consumption, and rapid training [26]. Specifically, five swarm-based optimization algorithms—FOX, GTO, PSO, SCSO, and SSA—were employed to fine-tune the hyperparameters of LightGBM, addressing the limitations of conventional machine learning models such as poor convergence, overfitting, and sensitivity to parameter selection. Despite the promising predictive capabilities of ML models, most existing studies have provided limited justification for their outputs, with little attention paid to feature-level interpretability in the educational domain. To address this gap, this study incorporated Explainable Artificial Intelligence (XAI), with SHAP serving as the interpretability layer. SHAP systematically decomposes each prediction into the contribution of individual input variables, thereby improving transparency, fostering trust, and supporting evidence-based decision-making within educational institutions.
Accordingly, this study compared five LightGBM models optimized with swarm-based algorithms. The major contributions of this study are as follows:
  • Development of hybrid predictive models that integrate LightGBM with FOX, GTO, PSO, SCSO, and SSA for optimized hyperparameter tuning.
  • Application of SHAP for interpretability, enabling detailed analysis of feature importance and its influence on prediction outcomes.
  • Empirical comparison of five swarm-based optimizers on student performance data, highlighting their relative strengths and weaknesses.
  • Provision of a reliable and transparent framework that combines high predictive accuracy with interpretability, thereby supporting learners, educators, and policymakers.
The remainder of this paper is structured as follows. Section 2 presents a comprehensive review of the existing literature on student performance prediction, machine learning models, and metaheuristic optimization techniques. Section 3 details the research methodology, including descriptions of the FOX, GTO, PSO, SCSO, and SSA algorithms, the LightGBM model, and the proposed optimization framework, as well as the dataset employed in this study. Section 4 reports and discusses the experimental results, covering model evaluation metrics and predictive performance analysis. Finally, Section 5 concludes the paper by summarizing the main findings, highlighting practical implications, and outlining directions for future research.

2. Literature Review

Researchers have increasingly explored machine learning models combined with optimization methods to better predict student performance. These studies aim to help educators understand and support students more effectively. Xu and Kim [27] developed a combination prediction model for student performance by integrating Decision Tree (DT), Support Vector Regression (SVR), and Backpropagation Neural Network (BP). The contribution of each base model was weighted using an Ant Colony Optimization (ACO) algorithm. The hybrid model showed superior performance over standalone models. Fang et al. [28] introduced a hybrid MDBO-BP-Adaboost framework, where a modified Dung Beetle Optimization Algorithm was used to enhance the learning capacity of a backpropagation neural network combined with Adaboost. Experimental comparisons with baseline machine learning models demonstrated that this approach achieved superior prediction accuracy on a student performance dataset. Cheng et al. [29] investigated student performance prediction by comparing several machine learning classifiers, including Random Forest, Decision Tree, K-Nearest Neighbors, Multi Layer Perceptron, and eXtreme Gradient Boosting (XGBoost). The use of the Support Vector Machine with the Synthetic Minority Oversampling Technique (SVM-SMOTE) technique improved performance on imbalanced data, with XGBoost emerging as the most effective model. To further enhance results, metaheuristic algorithms were integrated with XGBoost, achieving significant improvements across accuracy, precision, recall, and F1-score. Among these, the Enhanced Artificial Ecosystem-Based Optimization-based XGBoost hybrid achieved the highest accuracy. Emima and Amalarethinam [30] developed a hybrid Enhanced Teacher–Learner-Based Optimization with the Particle Swarm Optimization (ETLBO-PSO) model algorithm to fine-tune a Machine Learning model for student performance prediction. The model was tested on a secondary education dataset in Mathematics using XGBoost, Light Gradient-Boosting Machine (LightGBM), and Categorical Boosting (CatBoost) classifiers, where it achieved a robust F1-score of 82.43% while reducing redundancy and efficiency loss. Li et al. [31] developed a LightGBM with Genetic Algorithm (GA) Model for an academic performance prediction model for college students. The model achieved high accuracy and low prediction error. From the observed results, predicted grades closely matched actual student performance, demonstrating the model’s reliability.
Apriyadi et al. [32] proposed a student performance prediction model by optimizing SVR with metaheuristic algorithms. Both PSO and GA were applied to tune SVR hyperparameters, resulting in PSVR and GSVR models. Experimental results showed that the PSVR model achieved the lowest RMSE (1.608), outperforming several traditional models such as Naïve Bayes, Decision Tree, and Random Forest. Kamal et al. [33] proposed a metaheuristic and machine learning-based approach for classifying and predicting student performance. Features were first selected using a Relief algorithm, and classifiers including Back Propagation Neural Network (BPNN), Random Forest, and Naïve Bayes were applied. Among these, BPNN achieved the highest accuracy in both classification and prediction tasks. The study demonstrates the effectiveness of combining feature selection with machine learning models for reliable student performance forecasting.
Xu [34] introduced a hybrid IDA-SVR model to predict student performance, incorporating an Improved Duel Algorithm (IDA) for hyperparameter optimization. Unlike traditional approaches, the study emphasized student behavioral data and compared IDA-SVR with DT, ANN, standard SVR, and PSO-SVR. Results showed that IDA-SVR achieved the lowest MSE (0.0089), significantly outperforming competing models. The findings highlight IDA’s ability to avoid local optima and accelerate convergence, making it a strong tool for enhancing prediction accuracy. Song [35] integrated the K-Nearest Neighbor Classification (KNNC) model with two bio-inspired optimization techniques, namely the Honey Badger Algorithm (HBA) and the Arithmetic Optimization Algorithm (AOA). Their approach aimed to improve forecasting precision and reliability to support better educational outcomes. The proposed KNHB model showed superior performance in categorizing final grades with an accuracy of 0.921 and a precision of 0.92. Similarly, it achieved strong results for first-period grade prediction. Ma [36] proposed a Random Forest Classifier (RFC) model enhanced with specifically Electric Charged Particles Optimization (ECPO) and Artificial Rabbits Optimization (ARO), for predicting student performance. The study analyzed input variables to determine their impact on academic outcomes, helping to identify key areas for educational improvement. The optimized RFEC model demonstrated superior predictive accuracy across 4424 students, aligning closely with actual performance measurements. Ali et al. [37] developed a hybrid Principal Component Analysis with Cuckoo Search Neural Network (PCACSNN) to improve student performance prediction and overcome overfitting issues in large datasets. The model combines ANN with the PCACSNN algorithm to enhance convergence and prediction accuracy. Experimental results on the UCI Student Performance datasets for Mathematics and Portuguese showed superior performance compared to ANN, BPNN, and CSBP. Punitha and Devaki [38] proposed a deep learning-based predictive approach for student performance that integrates feature selection with the Modified Red Deer Algorithm (MRDA) and a Deep Ensemble Network (DEnsNet). The MRDA algorithm was used to optimize both feature selection and network parameters, ensuring improved prediction accuracy. Experimental results showed that the proposed model outperformed several comparative models with accuracy improvements ranging from 3% to 9.91% across two datasets. Hai and Wang [39] proposed a hybrid Multilayer Perceptron Classification (MLPC) approach for student performance prediction. The study integrated ensemble techniques and metaheuristic optimizers, specifically the Pelican Optimization Algorithm (POA) and Crystal Structure Algorithm (CSA), to improve model accuracy. Experimental results showed that the hybrid model achieved a 95.78% success rate, outperforming other approaches.
The existing body of research highlights the increasing effectiveness of integrating machine learning with metaheuristic optimization techniques for student performance prediction. Nonetheless, most prior studies placed a predominant emphasis on predictive accuracy, often neglecting rigorous interpretability diagnostics, and frequently restricted their analyses to a limited set of optimization algorithms. Such constraints reduce both the comprehensiveness and the practical applicability of their findings. To address these gaps, the present study conducted a systematic comparative evaluation of five swarm-intelligence-based optimization algorithms, FOX, GTO, PSO, SCSO, and SSA, in enhancing the predictive capability of the LightGBM model. Beyond accuracy, we incorporated SHAP analysis to provide post hoc interpretability, offering insight into the relative importance, variance, and interaction of features driving student outcomes. This dual emphasis on performance and interpretability contributes to the development of predictive analytics frameworks that are not only accurate and robust but also transparent and reliable.

3. Methodology

3.1. Fox Optimization Algorithm (FOX)

The Fox Optimization Algorithm (FOX) is a bio-inspired metaheuristic that emulates the predatory behavior of foxes in snow-covered environments [40]. Foxes rely on auditory cues to detect prey beneath the snow and execute pouncing maneuvers with high precision. The algorithm abstracts this behavior into a computational framework for solving continuous numerical optimization problems. Let the population of candidate solutions (agents) be represented by a matrix X R N × D , where N represents the number of search agents, D denotes the dimensionality of the problem, x i , j represent the value of the j -th variable of the i -th agent as defined in Equation (1).
X = x 1 , 1 x 1 , D x N , 1 x N , D
Each row X i = x i , 1 , x i , 2 , , x i , D represents a solution vector in the D -dimensional search space. Initial positions are generated randomly within the feasible bounds as given in Equation (2):
x i , j = l b j + r i , j u b j l b j
where l b j and u b j denote the lower and upper bounds of the j -th dimension. r i , j is a random number drawn from a uniform distribution over the interval ( 0 , 1 ) . To balance exploration and exploitation, a control parameter a [ 0 , 2 ] is introduced, which decreases linearly over iterations as expressed in Equation (3):
a = 2 1 t T
where t is the current iteration, T is the maximum number of iterations. Additionally, a random parameter r which is drawn from a uniform distribution over the interval ( 0 , 1 ) determines the phase of operation. In the exploration phase, the fox detects prey through sound propagation under snow. The propagation distance is computed in Equation (4):
d s ( t ) = v s T s ( t )
where T s ( t ) is a vector of random sound propagation times within interval [ 0 , 1 ] D , v s is the effective speed of sound in the medium, d s ( t ) represent the total sound propagation distance. The speed of sound is estimated based on the best-known position as expressed in Equation (5):
v s = X best   ( t ) T s ( t )
The fitness of each agent is evaluated using the objective function f ( ) . The best solution found so far is denoted X best   , with corresponding fitness f best   . The fox–prey distance is approximated as half the round-trip propagation distance, which is computed according to Equation (6):
d fox - prey   ( t ) = 0.5 d s ( t )
The pouncing motion is modeled as a parabolic jump influenced by gravity as expressed in Equation (7):
J ( t ) = 1 2 g t 2
where g , which is the gravitational acceleration, is set to 9.81   m / s 2 , t = 1 D j = 1 D   T s ( j ) represent the average sound propagation time across dimensions. A random variable p which is drawn from U ( 0 , 1 ) determines the update rule. If p > 0.18 , the rule in Equation (8) applies:
X i ( t + 1 ) = d fox - prey   ( t ) J ( t ) c 1
Otherwise, the Equation (9) update rule is applied to the current individual:
X i ( t + 1 ) = d fox - prey   ( t ) J ( t ) c 2
where c 1 is a random number within [ 0 , 0.18 ] , c 2 is a random number within [ 0.18 , 1 ] , and both variables are learning coefficients reflecting pounce success probability. During exploitation, a time-controlled random walk refines the search near to the best solution. The position update is expressed in Equation (10):
X i ( t + 1 ) = X best   ( t ) r 1 t m i n a
where r 1 is a random vector within [ 0 , 1 ] D , t min   is the minimum average time across agents.

3.2. Giant Trevally Optimizer (GTO)

The Giant Trevally Optimizer (GTO) is inspired by the hunting behavior of Caranx ignobilis, a marine predator known for leaping from water to capture airborne prey [41]. This behavior involves long-range movement (exploration), area selection, and precise aerial attacks (exploitation). The population matrix as given in Equations (1) and (2), and the best solution is denoted X best   . The extensive Search via Lévy Flight simulates long-range foraging (exploration) using Lévy flights as expressed in Equation (11):
X i ( t + 1 ) = X best   ( t ) r 2 + ( u b l b ) r 3 + l b L e v y ( D )
where r 2 and r 3 are uniform random variables chosen within [0, 1], and Levy ( D ) is a Lévy-distributed random vector. The Lévy step is generated as expressed in Equation (12):
L e v y ( D ) = s t e p u · σ | v | 1 / β ,   step   = 0.01 ,   β = 1.5
With σ , which is the scaling parameter defined in Equation (13):
σ = Γ ( 1 + β ) s i n ( π β / 2 ) Γ 1 + β 2 β 2 ( β 1 ) / 2 1 / β
u , v N ( 0 , 1 ) represent the standard and normal variables. The optimizer models the predator’s ability to identify prey-rich zones as given in Equation (14):
X i ( t + 1 ) = X best   ( t ) A r 4 + M ( t ) X i ( t ) r 5
where A [ 0.3 , 0.4 ] is the area selection coefficient, r 4 and r 5 are uniformly distributed random numbers within [0, 1], and M ( t ) , which denotes the mean position of the swarm, is calculated as follows in Equation (15):
M ( t ) = 1 N i = 1 N   X i t
The Exploitation phase mimics the leap from water, accounting for visual distortion due to refraction, using Snell’s Law, which is calculated as given in Equation (16):
s i n θ 1 = η 1 η 2 s i n θ 2
where, θ 1 is the angle of incidence in water, θ 2 which is chosen from the range [ 0 ,   360 ] is the refracted angle in air, η 1 representing the refractive index of air is set to 1.00029 . η 2 which is the refractive index of water is set to 1.33 . The visual distortion of prey is computed as follows, as given in Equation (17):
V = s i n θ 1 d i ,   d i = X best   ( t ) X i ( t )
Finally, the position update is expressed in Equation (18):
X i ( t + 1 ) = L + V + H
The launching vector L is calculated as follows in Equation (19):
L = X i ( t ) s i n θ 2 f X i ( t )
The Adaptive slope H is computed as follows in Equation (20):
H = r 6 2 2 t T
r 6 represents a random number within [0, 1]. As t tends to T and H tends to zero, the optimizer shifts focus from exploration to exploitation.

3.3. Particle Swarm Optimization (PSO)

PSO is a population-based stochastic optimization method inspired by social dynamics in bird flocks and fish schools. Particles adjust their trajectories based on personal and global experience [42]. Each particle i   ( i = 1 , 2 , , N ) has the following: Position vector X i = x i , 1 , , x i , D , Velocity vector V i = v i , 1 , , v i , D , Personal best P i = p i , 1 , , p i , D , and Global best G = g 1 , , g D . At iteration t , updates are as given in Equations (21) and (22):
v i , j ( t + 1 ) = ω v i , j ( t ) + c 3 r 7 p i , j x i , j ( t ) + c 4 r 8 g j x i , j ( t )
x i , j ( t + 1 ) = x i , j ( t ) + v i , j ( t + 1 )
where ω is the inertia weight, c 3 and c 4 are cognitive and social acceleration coefficients, r 7 and r 8 are uniformly distributed random numbers within [0, 1]. Variable ω is computed as follows in Equation (23):
ω = ω m a x t T ω m a x ω m i n
where ω m a x is set to 0.9, and ω m i n is set to 0.4. Higher ω favors exploration; lower ω enhances exploitation.

3.4. Sand Cat Swarm Optimization (SCSO)

SCSO models the foraging behavior of sand cats (Felis margarita), which use acute hearing to detect subterranean prey [43]. Two phases, prey detection (exploration) and predation (exploitation), are emulated. Population X R N × D initialized as expressed in Equations (1) and (2). During the prey detection phase (Exploration). The sensitivity magnitude S M = 2 . Global sensitivity parameter is updated according to Equation (24):
r G = S M 1 t T
The random control parameter is calculated as follows in Equation (25):
R = 2 r G r 9 r G
r 9 are uniformly distributed random numbers within [0, 1]. The Individual sensitivity span r is calculated as follows in Equation (26):
r = r G r 10
r 10 are uniformly distributed random numbers within [0, 1]. Finally, the exploration update rule is expressed in Equation (27):
X i ( t + 1 ) = r X b ( t ) r 11 X i ( t )
r 11 is a uniformly distributed random number within [0, 1], the best solution is denoted by X b , and X i ( t ) represents the current solution. For the prey attack (Exploitation), the distance to the prey and sand cat is calculated as given in Equation (28):
X r n d = r 12 X b t X i t
r 12 is an abstract uniformly distributed random number within [0, 1]. The exploitation update is expressed in Equation (29):
X i ( t + 1 ) = X b ( t ) r X r n d c o s ( α )
The random angle α is chosen within [0, 2 π ]. The Adaptive Switching Mechanism to transition between phases is governed by | R | , and the update is done as follows in Equation (30):
X i ( t + 1 ) = r X b ( t ) r 11 X i ( t ) ,   if   | R | > 1   ( Exploration )   X b ( t ) r X r n d c o s ( α ) ,   if   | R | 1   ( Exploitation )  
This ensures dynamic balance between global search and local refinement.

3.5. Salp Swarm Algorithm (SSA)

SSA, proposed by Mirjalili et al. [44], mimics the swarming and foraging behavior of salp marine invertebrates that move in chains to efficiently navigate ocean currents and locate food sources. The swarm is represented by matrix X R N × D , where N is the number of salps, and D is the dimensionality as expressed in Equations (1) and (2). Food source F = F 1 , F 2 , , F D represents the target (optimal solution). The first salp (leader) guides the chain toward the food source as given in Equation (31):
x 1 , j ( t + 1 ) = F j + c 3 c 4 u b j l b j + l b j ,   if   c 5 0.5 F j c 3 c 4 u b j l b j + l b j ,   if   c 5 < 0.5
c 3 the convergence coefficient (decreases over time) as expressed in Equation (32):
c 3 = 2 e ( 4 t / T ) 2
c 4 , and c 5 are uniformly distributed random numbers within [0, 1]. Large c 3 promotes exploration; small c 3 enables exploitation. Remaining salps follow the one ahead using a Newtonian motion model as given in Equation (33):
x i , j ( t + 1 ) = 1 2 x i , j ( t ) + x i 1 , j ( t ) ,   i 2
This ensures smooth, coordinated movement toward the leader and ultimately the food source.

3.6. Light Gradient Boosting Machine (LightGBM)

The Light Gradient Boosting Machine (LightGBM) is a boosting algorithm based on decision trees, for applications in regression, classification, and feature selection [45]. It is characterized by high computational efficiency, reduced memory usage, and improved predictive accuracy. The fundamental principle of LightGBM involves iteratively enhancing model performance through the sequential addition of weak learners. Let F t 1 ( x ) denote the learned model from the previous iteration. Given a loss function L y ,   F t 1 ( x ) , the objective in the current iteration is to identify a weak learner h t ( x ) that minimizes the overall loss when added to the existing model. This optimization can be formally expressed as given in Equation (34):
h t ( x ) = a r g   m i n h H     L y ,   F t 1 ( x ) + h t ( x )
To facilitate optimization, the loss function is approximated using the negative gradient (also known as the functional gradient) evaluated at the current model output. This residual, denoted as r t i , represents the pseudo-residuals for each instance x i and is computed in Equation (35):
r t i = L y ,   F t 1 x i F t 1 x i
The weak learner h t ( x ) is then trained to minimize the squared error between these pseudo-residuals and its predictions according to Equation (36):
h t ( x ) = a r g   m i n h H     r t i h t ( x ) 2
Upon obtaining h t ( x ) , the model is updated to form the new strong learner as described in Equation (37):
F t ( x ) = F t 1 ( x ) + h t ( x )
This iterative process continues until a predefined stopping criterion is met, resulting in a final predictive model with enhanced accuracy and robustness.

3.7. Proposed Model Optimization Framework

LightGBM demonstrates strong efficiency for prediction tasks but suffers from sensitivity to hyperparameter settings [46,47,48]. Parameters such as max_depth, n_estimators, and learning_rate strongly influence the model’s ability to balance bias, variance, and computational efficiency. Without systematic tuning, LightGBM often converges to suboptimal solutions, leading to reduced predictive accuracy and poor generalization. To address this, a metaheuristic-based optimization framework is proposed, integrating swarm intelligence algorithms with LightGBM to automatically search for near-optimal hyperparameters. The framework integrates five swarm intelligence algorithms, FOX, GTO, PSO, SCSO, and SSA, to search for the optimal combination of LightGBM hyperparameters systematically. These algorithms mimic natural collective behaviors to explore the parameter space efficiently, avoiding local optima and converging toward globally superior solutions. The key hyperparameters optimized in this framework include the following:
  • max_depth: Controls the maximum depth of individual decision trees. A well-tuned depth prevents overfitting by limiting model complexity. The search range is set to ( 3 , 15 ) .
  • learning_rate: Regulates the contribution of each tree during boosting. A value too low results in slow convergence, while a value too high may lead to overshooting the optimum. The range ( 0.01 , 0.5 ) is explored.
  • no_estimators determines the number of boosting rounds in LightGBM. It directly influences model performance, generalization, and training. The range ( 10 , 150 ) is explored.
The proposed optimization workflow is structured as follows:
  • Step 1: Preprocess the educational dataset by normalizing input features, encoding categorical variables.
  • Step 2: Partition the dataset into training and testing.
  • Step 3: Initialize the LightGBM model with default hyperparameters and define the search bounds for each tunable parameter as specified above.
  • Step 4: Employ each metaheuristic algorithm (FOX, GTO, PSO, SCSO, SSA) to optimize the hyperparameter configuration. Optimizers apply different population update mechanisms as defined in Section 3.1, Section 3.2, Section 3.3, Section 3.4, Section 3.5; steps of each optimizer are iterated in their original literature. The algorithm iteratively updates candidate solutions based on fitness evaluation using the MSE score as the objective function. In this study, the objective function is defined as the mean of the MSE over a 2-fold cross-validation, computed on the training dataset for each candidate hyperparameter set. This approach ensures a robust assessment of model performance while maintaining computational efficiency during the search process.
  • Step 5: At each iteration, compute the fitness of the current hyperparameter set, update the global best and individual best positions, and assess convergence criteria. The process continues until the maximum number of iterations is reached or convergence is achieved.
  • Step 6: Retrieve the optimal hyperparameter set identified by each optimizer. Evaluate the final model on the test set using multiple performance metrics.
  • Step 7: Perform SHAP analysis to interpret model predictions, identify key influencing features, and validate the results against pedagogical insights.
This proposed framework not only enhances the predictive power of LightGBM but also provides a reproducible and scalable methodology for hyperparameter tuning in an educational dataset. By leveraging swarm intelligence, the approach ensures robust, data-driven model calibration, making it highly suitable for real-world academic performance prediction systems. The selection of the five metaheuristic optimizers FOX, GTO, PSO, SCSO, and SSA was guided by three criteria: (1) representation of distinct biological inspiration paradigms, (2) demonstrated empirical efficacy particularly exceptional performance in recent machine learning hyperparameter optimization literature [49,50,51,52,53], and (3) algorithmic diversity in search strategy. PSO, inspired by avian flocking, serves as the established baseline for social swarm intelligence. SSA, modeling marine invertebrate chains, introduces a leader–follower exploitation mechanism distinct from PSO’s velocity-based updates. SCSO, emulating subterranean prey detection in felines, features a unique sensitivity-based switching mechanism between exploration and exploitation. Finally, FOX, simulating auditory prey localization in canids, employs a gravity-influenced pounce model for exploitation. The framework is illustrated in Figure 1.

3.8. Data

The dataset utilized in this study was sourced from Kaggle [54] and contains a comprehensive set of variables reflecting diverse factors influencing student academic performance, with the primary outcome being the exam score, with 6607 data samples before data clean up. It comprises 16 predictor variables after the data clean-up process, encompassing academic, behavioral, socio-economic variables; namely the variables are hours_studied, attendance, parental_involvement, access_to_resources, extracurricular_activities, sleep_hours, previous_scores, motivation_level, internet_access, tutoring_sessions, family_income, school_type, peer_influence, physical_activity, learning_disabilities, and gender. Prior to using LightGBM models to predict, a systematic data cleaning procedure were employed to ensure data quality and modeling reliability. Columns containing missing values were excluded, and duplicate records were removed to maintain data integrity. Categorical variables were transformed into numerical form using label encoding, assigning unique integer values to each category. To mitigate the influence of outliers, the Z-score method was applied, thereby retaining data points within a defined normal range. The dataset was subsequently partitioned into training (80%) and testing (20%) subsets using a random split. Feature normalization was performed using Min-Max scaling, which linearly transforms each feature to a standard range [0, 1] according to Equation (38):
X scaled   = X X m i n X m a x X m i n
where X is the original feature value, and X m i n and X m a x are the minimum and maximum values of the feature in the training set, respectively. This transformation was applied independently to each feature using parameters derived from the training data and subsequently used to scale the test set, preventing data leakage. The target variable is exam score, while other variables are predictors. Table 1 provides a summary statistics table for the 16 variables from the dataset, specifically showing the distribution of variables related to student performance. It presents standard statistical measures for each variable across 4313 observations, after removing columns with empty values, encoding categorical variables into numerical values, and handling outliers without scaling.

3.9. Model Evaluation Metrics

To assess the performance of the predictive models, several regression evaluation metrics were employed: the R 2 , RMSE, MSE, ME, and RAE. These metrics collectively provide insights into model accuracy and error magnitude, enabling robust comparison across different modeling approaches. The R 2 measures the proportion of variance in the dependent variable (exam score) that is predictable from the independent variables. It is defined as expressed in Equation (39):
R 2 = 1 i = 1 n     y i y ˆ i 2 i = 1 n     y i y 2
where y i is the observed value, y ˆ i is the predicted value, and y is the mean of the observed values. R 2 ranges from to 1, with values closer to 1 indicating a better fit. A value of 1 implies perfect prediction, while values below 0 suggest the model performs worse than a simple mean predictor. The MSE quantifies the average squared difference between predicted and actual values, penalizing larger errors more heavily due to the squaring operation as defined in Equation (40):
M S E = 1 n i = 1 n   y i y ˆ i 2
where n is the number of observations. MSE is sensitive to outliers and provides a measure of overall model precision in squared units of the target variable. The RMSE is the square root of MSE and expresses prediction error in the same units as the target variable. It is calculated as expressed in Equation (41):
R M S E = 1 n i = 1 n     y i y ˆ i 2
Lower RMSE values indicate higher accuracy and tighter clustering of predictions around the true values. The ME captures the largest absolute residual across all predictions as expressed in Equation (42):
M E = m a x i = 1 , , n   y i y ˆ i
This metric highlights the worst-case prediction error and is useful for identifying potential model failure cases or sensitivity to extreme values. Finally, the RAE normalizes the total absolute prediction error by the total absolute deviation from the mean according to Equation (43):
R A E = i = 1 n     y i y ˆ i i = 1 n     y i y
RAE provides a relative measure of model performance, where values less than 1 indicate that the model performs better than a naive baseline (predicting the mean), and smaller values reflect superior predictive accuracy. Together, these metrics offer a multi-faceted evaluation of model performance, balancing interpretability, sensitivity to error magnitude, and comparative benchmarking against baseline models.

3.10. Proposed Comparative Optimization Framework

In this study, the LightGBM model was optimized using five metaheuristic algorithms: FOX, GTO, PSO, and SCSO. The hyperparameter settings for each optimizer are detailed in Table 2. The search boundaries of parameters tuned by optimizers are as detailed in Section 3.7. For the baseline LightGBM model, the following fixed parameters are employed: n_estimators is set to 150, learning_rate is set to 0.3, and max_depth is set to 15. The baseline Support Vector Regression (SVR) model is configured with a radial basis function (RBF) kernel, a regularization parameter C of 100, and a kernel coefficient γ of 0.1, which are commonly used settings for achieving stable performance in regression tasks. The baseline model’s parameter configurations are selected from widely used parameters reported in the established literature [55,56,57]. All metaheuristic optimizers are configured with a population size of 30 and a maximum of 100 iterations to ensure a balanced trade-off between computational efficiency and convergence to optimal solutions. The optimization process is guided by minimizing the MSE evaluated via 2-fold cross-validation within each run. The overall experimental framework consists of two phases: first, a 5-fold cross-validation test is done to ensure robust model evaluation and reduce bias in performance estimation; second, the entire process is repeated over 20 independent runs to account for the stochastic nature of the metaheuristic algorithms and to enhance the reliability of the results. This rigorous evaluation protocol enables a comprehensive comparison of the optimization techniques in terms of both predictive accuracy and generalization capability.

4. Experiment and Discussion

4.1. Student Performance Prediction

4.1.1. Cross-Validation Analysis

To obtain unbiased estimates of model generalization, we employed a cross-validation strategy. In each split of the outer 5-fold CV, the training partition was used for hyperparameter optimization. After selecting the optimal hyperparameters, the model was evaluated once on the corresponding outer held-out fold. Performance metrics were then aggregated across the five outer folds. Notably, no cross-validation was applied to the held-out folds, ensuring that the data used for hyperparameter optimization were entirely separate from those used for evaluation. For the results, training-fold CV performance is reported in Table 3, while outer held-out fold performance is presented in Table 4.
Table 3 and Table 4 present the training and held-out performance outcomes of the metaheuristic-optimized LightGBM models and baseline models using 5-fold cross-validation. The training performance metrics presented in Table 3 illustrate that the metaheuristic-optimized LightGBM variants substantially outperformed the baseline LightGBM and SVR models in terms of predictive accuracy and error minimization. Among the optimized models, FOX-LightGBM attained the highest average R2 score of 0.9693, indicating superior fit to the training data, followed by SCSO-LightGBM (0.9609), SSA-LightGBM (0.9606), PSO-LightGBM (0.9598), and GTO-LightGBM (0.9581). This enhanced accuracy is corroborated by lower error metrics: the optimized models achieved RMSE values ranging from 0.0361 to 0.0423, significantly lower than the baseline LightGBM (0.0578) and SVR (0.0523). Similarly, MSE values for the optimizers were lower compared to the baseline models. FOX-LightGBM also recorded the lowest RAE at 0.0695, underscoring its effectiveness in calibrating hyperparameters to capture the underlying data patterns. However, FOX-LightGBM exhibited greater variability across folds, as evidenced by its higher standard deviations in R2, RMSE, and MSE, relative to more stable performers like SCSO-LightGBM and PSO-LightGBM, which demonstrated tighter deviations. Notably, SVR achieved the lowest maximum error (ME) at 0.1008, suggesting reduced peak deviations in training predictions, though this did not translate to overall superiority given its higher aggregated errors.
The held-out set results in Table 4 provide critical insights into model generalization, revealing shifts in relative performance that highlight the optimizers’ robustness to unseen data. Here, SCSO-LightGBM emerged as the top performer with the highest average R2 (0.9412), followed closely by SSA-LightGBM (0.9409) and PSO-LightGBM (0.9406), while GTO-LightGBM (0.9343) and FOX-LightGBM (0.9331) lagged slightly behind. These rankings are mirrored in the error metrics: SCSO-LightGBM achieved the lowest RMSE (0.0502) and MSE (2.522 × 10−3), with SSA-LightGBM and PSO-LightGBM exhibiting comparable values (RMSE: 0.0503 and 0.0505, respectively). In contrast, the baselines yielded higher errors, with LightGBM at RMSE 0.0630 and SVR at 0.0565. PSO-LightGBM recorded the lowest ME (0.1620), indicating minimal systematic bias, whereas SVR and unoptimized LightGBM displayed larger biases (0.2117 and 0.2071, respectively). RAE values further affirm the optimizers’ advantages, with SCSO-LightGBM at the lowest (0.0968), compared to 0.1215 for LightGBM and 0.1088 for SVR. Standard deviations on the test set were generally low across models, but SCSO-LightGBM and PSO-LightGBM maintained particularly tight variances underscoring their consistency. Comparing training and test performances reveals the optimizers’ capacity for balanced generalization, with modest degradations in R2 and proportional increases in RMSE and MSE, indicative of effective hyperparameter tuning that mitigates overfitting. FOX-LightGBM, despite its training dominance, experienced the most pronounced relative drop in R2 increase in RMSE, suggesting a propensity for overfitting due to aggressive parameter optimization. Collectively, these results substantiate the efficacy of metaheuristic algorithms in enhancing LightGBM’s hyperparameter optimization, with SCSO-LightGBM standing out as the most robust variant. It integrates superior accuracy, error minimization, generalization, and stability, making it a promising approach for the student performance prediction task.

4.1.2. Independent Run Analysis

During independent run analysis, the data split is fixed throughout the experiment; the sole intended source of variability is the metaheuristic optimizer’s stochastic initialization and update dynamics. To ensure the robustness and reliability of the performance evaluation, each metaheuristic-optimized LightGBM model was subjected to 20 independent runs using the same hyperparameter settings and experimental configuration from the previous experiment. The results from 20 independent runs, for training as detailed in Table 5, demonstrate the efficacy of metaheuristic-optimized LightGBM models in enhancing predictive performance on the training set. FOX-LightGBM achieved the highest average R 2 of 0.9682, reflecting superior hyperparameter optimization, followed by GTO-LightGBM (0.9610), SCSO-LightGBM (0.9609), SSA-LightGBM (0.9608), and PSO-LightGBM (0.9590). These values substantially outperform the baselines, with unoptimized LightGBM and SVR recording R 2 scores of 0.9229 and 0.9371, respectively. Error metrics further underscore the optimizers’ advantages: FOX-LightGBM yielded the lowest average RMSE and MSE, compared to the unoptimized LightGBM and SVR. The relative absolute error (RAE) was minimized by FOX-LightGBM to 0.0703, outperforming unoptimized LightGBM (0.1106) and SVR (0.0999). The best-case performance of FOX-LightGBM, with an R 2 of 0.9828, RMSE of 0.0273, and MSE of 7.450 × 10−4, highlights its capability to achieve exceptional fits. However, its higher standard deviations indicate greater variability compared to SCSO-LightGBM and SSA-LightGBM, which exhibited greater stability. Notably, SVR achieved the lowest maximum error (ME) at 0.1005, suggesting fewer extreme deviations, though its overall error metrics remained less competitive.
Test set results in Table 6 reveal the generalization capabilities of the models. SSA-LightGBM and SCSO-LightGBM led with average R 2 values of 0.9388 and 0.9386, respectively, followed by PSO-LightGBM (0.9364), while FOX-LightGBM (0.9332) and GTO-LightGBM (0.9316) trailed slightly. Error metrics show that SSA-LightGBM achieved the lowest RMSE of 0.0506 and MSE of 2.559 × 10−3, with SCSO-LightGBM close behind with RMSE of 0.0507, and MSE of 2.570 × 10−3. PSO-LightGBM remained competitive, whereas FOX-LightGBM and GTO-LightGBM exhibited higher errors. The baselines underperformed, with unoptimized LightGBM achieving the lowest R 2 of 0.9032 and the highest RMSE of 0.0636, while SVR outperformed LightGBM but lagged behind the optimizers. RAE values further confirm the optimizers’ superiority, with SSA-LightGBM (0.0995) and SCSO-LightGBM (0.0998) achieving the lowest, compared to 0.1252 for unoptimized LightGBM and 0.1141 for SVR. The best-case test performance highlights SCSO-LightGBM’s edge, with an R 2 of 0.9406 and MSE of 2.486 × 10−3. SCSO-LightGBM also recorded the lowest ME (0.1610), indicating minimal systematic bias, while unoptimized LightGBM and SVR exhibited larger biases. Comparing training and test results reveals distinct generalization patterns. FOX-LightGBM, despite its training dominance, experienced a notable performance drop. The standard deviations across 20 runs highlight the reliability of SCSOLightGBM and SSA-LightGBM, which exhibited the lowest variability. FOX-LightGBM and GTO-LightGBM showed higher variability, indicating less consistent performance. The near-zero standard deviations for the baselines reflect their deterministic nature but do not compensate for their inferior accuracy. In conclusion, while all metaheuristic optimizations enhanced LightGBM’s performance over the baselines, SCSO-LightGBM and SSA-LightGBM emerged as the most effective, combining superior predictive accuracy, robust generalization, and high stability across 20 independent runs.
The scatter plots in Figure 2 illustrate the relationship between the actual and predicted exam scores for both training and testing sets, using the best test outcome with the corresponding training result from the 20 independent trials for each model. FOX-LightGBM achieved the highest training but exhibited a slightly lower testing R 2 , suggesting stronger fitting on the training data but reduced generalization compared to other optimized variants. GTO-LightGBM followed a similar pattern with a training R 2 of 0.957 and a testing R 2 of 0.938, reflecting adequate performance but a modestly larger train-test gap. PSO-LightGBM balanced training and testing outcomes well, while SSA-LightGBM performed, demonstrating consistent predictive capacity across both phases. Notably, SCSO-LightGBM achieved the highest testing accuracy R 2 of 0.941 with a matching training R 2 of 0.961, making it the best generalizing model. The alignment of its predicted points along the fitted line in both training and testing highlights minimal deviation and reduced bias compared with other optimizers. In contrast, the baseline models underperformed: unoptimized LightGBM yielded the lowest test R 2 (0.903), with points more widely dispersed around the fitted line, while SVR achieved moderate accuracy but still lagged behind the optimized LightGBM models. Collectively, these plots confirm that while FOX-LightGBM dominated training, SCSO-LightGBM and SSA-LightGBM achieved superior test generalization, with SCSO-LightGBM emerging as the most reliable and effective method overall.
Figure 3 presents the prediction trends of each model over the entire dataset, partitioned into training and testing phases. The black markers represent the actual values, while the colored markers indicate the model predictions. The dark red curves show the absolute prediction error, providing an intuitive view of both accuracy and stability across samples. The plot shows the absolute error between the best test result of 20 independent runs, including the corresponding training output. For the optimized LightGBM models, the prediction lines closely follow the actual values across both training and testing sets, confirming that the metaheuristic-tuned models successfully captured the underlying data distribution. Among them, SCSO-LightGBM and SSA-LightGBM achieved the most stable error bands, with the red error line remaining nearly flat and centered around zero across samples, which reflects high robustness and minimal bias. PSO-LightGBM also exhibited a consistently narrow error band, demonstrating reliable prediction quality. FOX-LightGBM, despite strong training accuracy, displayed slightly larger fluctuations in the testing region, suggesting reduced stability compared with SCSO-LightGBM and SSA-LightGBM models. Similarly, GTO-LightGBM performed reasonably well but showed slightly higher deviations in its error distribution. The baseline models demonstrated weaker predictive capacity. Unoptimized LightGBM presented the widest dispersion between actual and predicted values, accompanied by a visibly higher error band, particularly in the testing phase. SVR performed better than unoptimized LightGBM, with reduced deviation, but its error curve remained more variable compared to the metaheuristic-optimized LightGBM models. The consistently low and stable error distributions in SCSO-LightGBM and SSA-LightGBM confirm their superior generalization and reliability. These results underscore that, while all metaheuristic optimizations improve LightGBM compared to the baselines, SCSO-LightGBM exhibits the most accurate and stable performance across both training and testing samples, making it the best overall model in this study.
The SHAP violin plot in Figure 4 provides a detailed view of how individual feature values influenced the model predictions in SCSO-LightGBM. The SCSO-LightGBM model is used to understand the most contributive features during the LightGBM prediction process due to its robust performance. Each distribution shows the spread of SHAP values, with the color indicating whether the feature value is high (red) or low (blue). This allows for an examination not only of the importance of features but also of the direction and magnitude of their influence. Attendance exerted the strongest largest positive influence: higher attendance (red points) was consistently associated with positive SHAP values, increasing the predicted exam score, while lower attendance (blue points) shifted predictions downward. Hours Studied showed a similar pattern, with more study hours leading to higher predicted scores, underscoring the strong contribution of academic effort to performance outcomes. Previous Scores also had a positive influence, though with a narrower distribution, reflecting their role as a stable baseline indicator of performance. Moderately influential features such as Parental Involvement, Access to Resources, and Tutoring Sessions exhibited both positive and negative SHAP contributions, suggesting that their impact varies across students depending on context. Features including Family Income, Peer Influence, and Motivation Level produced smaller yet consistent effects, often reinforcing or moderating the impact of the primary academic factors. By contrast, features such as Sleep Hours, School Type, Gender, Learning Disabilities, and Internet Access displayed SHAP values centered closely around zero, indicating negligible influence on model outputs. This further validates that the model prioritized behavioral and academic engagement variables over demographic or structural factors in predicting exam performance. Taken together, the violin plot emphasizes that academic engagement (attendance, study hours, previous scores) and supportive contexts (resources, parental involvement) drive the largest shifts in predictions, whereas demographic variables contribute minimally. This reinforces the interpretability of SCSO-LightGBM, confirming its alignment with intuitive educational predictors while providing nuanced insights into how features interact to shape performance outcomes.
The SHAP decision plot in Figure 5 provides a detailed visualization of how individual features contribute to the cumulative prediction process in the SCSO-LightGBM model. Each line represents a single sample’s path from the model’s base value (average prediction) toward the final output probability, with contributions from features added sequentially. Features displayed at the top, such as Attendance and Hours Studied, exert the largest shifts in prediction, consistent with the importance rankings shown in the SHAP violin plot. Specifically, high attendance and increased study hours consistently push predictions upward (toward higher exam scores), while low values in these features shift the predictions downward. Previous Scores, Access to Resources, and Parental Involvement also exert moderate influence, though their impact is more variable across samples, as indicated by the spread of lines. Features such as Tutoring Sessions, Family Income, and Peer Influence provide smaller but still noticeable shifts in prediction, reflecting their secondary role in determining outcomes. The remaining features, including Physical Activity, Sleep Hours, School Type, Gender, Learning Disabilities, and Internet Access, contributed minimally to decision paths, consistent with their low SHAP importance scores. Conclusively, the decision plot illustrates the interpretability strength of the SCSO-LightGBM model: predictions are largely driven by academic effort (study hours, previous scores), behavioral engagement (attendance, parental involvement), and resource availability. Importantly, the convergence of most decision paths around the correct prediction zone indicates that the model’s decisions are not only accurate but also stable across different samples, thereby validating the reliability of SCSO-LightGBM for educational outcome prediction. The SHAP decision plot (Figure 5) further highlights interaction effects, where Parental Involvement and Access to Resources often amplify the positive contributions of Attendance and Previous Scores, while attenuating their negative impacts when these core features are low. Subgroup-level analysis confirms this pattern: among students; SHAP gains are driven primarily by Attendance and Hours Studied.

5. Conclusions

This study evaluated the efficacy of metaheuristic hyperparameter optimization in enhancing the performance of LightGBM for student performance prediction. Five swarm-based optimization algorithms FOX, GTO, PSO, SCSO, and SSA were benchmarked against baseline models. The experimental design employed 5-fold cross-validation, with 20 independent runs to mitigate stochastic variability, while performance was assessed using multiple metrics (R2, RMSE, MSE, RAE, and ME). Results demonstrated that metaheuristic optimization consistently outperformed the baseline models. SCSO-LightGBM exhibited superior generalization on test datasets. In contrast, unoptimized LightGBM and SVR consistently underperformed across all metrics, reinforcing the value of systematic hyperparameter tuning. SHAP further validated the models’ reliability. For the best-performing models, the most influential predictors were Attendance, Hours Studied, and Previous Scores, with secondary contributions from Access to Resources and Parental Involvement. These findings, consistent across mean importance, decision, and violin plots, align with educational domain knowledge, providing credible and actionable insights for policy development.
It is important to acknowledge that while the findings offer valuable insights. The findings of this study are constrained by reliance on a single dataset, which inherently limits the generalizability of the results across diverse domains, noise regimes, and feature distributions. Furthermore, the computational overhead associated with the deployment of metaheuristic optimizers constitutes a significant practical limitation, particularly in resource-constrained or latency-sensitive applications. Additionally, the comparative evaluation of metaheuristic algorithms was conducted under a fixed iteration, which may not fully capture their convergence behavior or relative efficiency under dynamic or adaptive stopping criteria. Future research should prioritize the development and evaluation of multi-objective optimization frameworks that jointly optimize predictive accuracy, training efficiency, model complexity, and fairness-aware metrics such as bias mitigation. To enhance scalability and practical utility, multi-fidelity search strategies incorporating early stopping mechanisms will be investigated. Robustness evaluations under realistic distributional shifts, adversarial noise, and outlier contamination are also critical to assess real-world applicability. Subsequent studies will extend the current analysis by benchmarking metaheuristic optimizers against established non-metaheuristic baselines, including grid search, random search, and Bayesian optimization. This comparative framework will elucidate the trade-offs inherent in different optimization paradigms. Additionally, future experiments will systematically vary the number of optimization iterations to evaluate convergence properties.

Author Contributions

A.A. (Abdalhmid Abukader): Methodology, Formal Analysis, Original Draft, A.A. (Ahmad Alzubi): Supervision, Resources, Editing. O.R.A.: Conceptualization, Supervision, Resources, Editing, Methodology, Original Draft. All authors have read and agreed to the published version of the manuscript.

Funding

Authors declare no funding was received.

Data Availability Statement

The data obtained through the experiments are available upon request from the corresponding author.

Conflicts of Interest

The authors declare that there are no conflicts of interest.

References

  1. Saksiriruthai, S. Human Capital as a Determinant of Long-Term Economic Growth. In Research Anthology on Preparing School Administrators to Lead Quality Education Programs; IGI Global Scientific Publishing: Hershey, PA, USA, 2021; pp. 1518–1533. [Google Scholar] [CrossRef]
  2. Baumann, C.; Hamin. The role of culture, competitiveness and economic performance in explaining academic performance: A global market analysis for international student segmentation. J. Mark. High. Educ. 2011, 21, 181–201. [Google Scholar] [CrossRef]
  3. Barro, R.J.; McCleary, R.M. Religion and Economic Growth across Countries. Am. Sociol. Rev. 2003, 68, 760–781. [Google Scholar] [CrossRef]
  4. Barro, R.J. Determinants of Economic Growth: A Cross-Country Empirical Study; National Bureau of Economic Research: Cambridge, MA, USA, 1996; p. 5698. [Google Scholar] [CrossRef]
  5. Sultana, T.; Dey, S.R.; Tareque, M. Exploring the linkage between human capital and economic growth: A look at 141 developing and developed countries. Econ. Syst. 2022, 46, 101017. [Google Scholar] [CrossRef]
  6. Ahmed, W.; Wani, M.A.; Plawiak, P.; Meshoul, S.; Mahmoud, A.; Hammad, M. Machine learning-based academic performance prediction with explainability for enhanced decision-making in educational institutions. Sci. Rep. 2025, 15, 26879. [Google Scholar] [CrossRef]
  7. Vaarma, M.; Li, H. Predicting student dropouts with machine learning: An empirical study in Finnish higher education. Technol. Soc. 2024, 76, 102474. [Google Scholar] [CrossRef]
  8. Dwivedi, D.N.; Mahanty, G.; Dwivedi, V.N. The Role of Predictive Analytics in Personalizing Education: Tailoring Learning Paths for Individual Student Success. In Enhancing Education with Intelligent Systems and Data-Driven Instruction; IGI Global Scientific Publishing: Hershey, PA, USA, 2024; pp. 44–59. [Google Scholar] [CrossRef]
  9. Pelima, L.R.; Sukmana, Y.; Rosmansyah, Y. Predicting University Student Graduation Using Academic Performance and Machine Learning: A Systematic Literature Review. IEEE Access 2024, 12, 23451–23465. [Google Scholar] [CrossRef]
  10. Ahmed, A.A.; Sayed, S.; Abdoulhalik, A.; Moutari, S.; Oyedele, L. Applications of machine learning to water resources management: A review of present status and future opportunities. J. Clean. Prod. 2024, 441, 140715. [Google Scholar] [CrossRef]
  11. Ke, W.; Chan, K.-H. A Multilayer CARU Framework to Obtain Probability Distribution for Paragraph-Based Sentiment Analysis. Appl. Sci. 2021, 11, 11344. [Google Scholar] [CrossRef]
  12. Xing, Z.; Lam, C.-T.; Yuan, X.; Im, S.-K.; Machado, P. MMQW: Multi-Modal Quantum Watermarking Scheme. IEEE Trans. Inf. Forensics Secur. 2024, 19, 5181–5195. [Google Scholar] [CrossRef]
  13. Alnasyan, B.; Basheri, M.; Alassafi, M. The power of Deep Learning techniques for predicting student performance in Virtual Learning Environments: A systematic literature review. Comput. Educ. Artif. Intell. 2024, 6, 100231. [Google Scholar] [CrossRef]
  14. Hasan, R.; Palaniappan, S.; Raziff, A.R.A.; Mahmood, S.; Sarker, K.U. Student Academic Performance Prediction by using Decision Tree Algorithm. In Proceedings of the 2018 4th International Conference on Computer and Information Sciences (ICCOINS), Kuala Lumpur, Malaysia, 13–14 August 2018; pp. 1–5. [Google Scholar]
  15. Alamri, H.L.; S. Almuslim, R.; S. Alotibi, M.; K. Alkadi, D.; Ullah Khan, I.; Aslam, N. Predicting Student Academic Performance using Support Vector Machine and Random Forest. In Proceedings of the 2020 3rd International Conference on Education Technology Management, London, UK, 17–19 December 2020; Association for Computing Machinery: New York, NY, USA, 2021; pp. 100–107. [Google Scholar]
  16. Salah Hashim, A.; Akeel Awadh, W.; Khalaf Hamoud, A. Student Performance Prediction Model based on Supervised Machine Learning Algorithms. IOP Conf. Ser. Mater. Sci. Eng. 2020, 928, 032019. [Google Scholar] [CrossRef]
  17. Sravani, B.; Bala, M.M. Prediction of Student Performance Using Linear Regression. In Proceedings of the 2020 International Conference for Emerging Technology (INCET), Belgaum, India, 5–7 June 2020; pp. 1–5. [Google Scholar]
  18. Tripathi, A.; Yadav, S.; Rajan, R. Naive Bayes Classification Model for the Student Performance Prediction. In Proceedings of the 2019 2nd International Conference on Intelligent Computing, Instrumentation and Control Technologies (ICICICT), Kannur, India, 5–6 July 2019; Volume 1, pp. 1548–1553. [Google Scholar]
  19. Huang, C.; Zhou, J.; Chen, J.; Yang, J.; Clawson, K.; Peng, Y. A feature weighted support vector machine and artificial neural network algorithm for academic course performance prediction. Neural Comput. Appl. 2023, 35, 11517–11529. [Google Scholar] [CrossRef]
  20. Ahmed, S.T.; Al-Hamdani, R.; Croock, M.S. Enhancement of student performance prediction using modified K-nearest neighbor. TELKOMNIKA (Telecommun. Comput. Electron. Control) 2020, 18, 1777–1783. [Google Scholar] [CrossRef]
  21. Baashar, Y.; Alkawsi, G.; Mustafa, A.; Alkahtani, A.A.; Alsariera, Y.A.; Ali, A.Q.; Hashim, W.; Tiong, S.K. Toward Predicting Student’s Academic Performance Using Artificial Neural Networks (ANNs). Appl. Sci. 2022, 12, 1289. [Google Scholar] [CrossRef]
  22. Parrales-Bravo, F.; Caicedo-Quiroz, R.; Barzola-Monteses, J.; Cevallos-Torres, L. Applying Bayesian Networks to Predict and Understand the Student Academic Performance. In Proceedings of the 2024 Second International Conference on Advanced Computing & Communication Technologies (ICACCTech), Sonipat, India, 16–17 November 2024; pp. 738–743. [Google Scholar]
  23. Yang, S.J.H.; Lu, O.H.T.; Huang, A.Y.Q.; Huang, J.C.H.; Ogata, H.; Lin, A.J.Q. Predicting Students’ Academic Performance Using Multiple Linear Regression and Principal Component Analysis. J. Inf. Process. 2018, 26, 170–176. [Google Scholar] [CrossRef]
  24. Wilson, A.; Anwar, M.R. The Future of Adaptive Machine Learning Algorithms in High-Dimensional Data Processing. Int. Trans. Artif. Intell. 2024, 3, 97–107. [Google Scholar] [CrossRef]
  25. Yaqoob, A.; Verma, N.K.; Aziz, R.M. Metaheuristic Algorithms and Their Applications in Different Fields. In Metaheuristics for Machine Learning; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2024; pp. 1–35. [Google Scholar] [CrossRef]
  26. Ke, G.; Meng, Q.; Finley, T.; Wang, T.; Chen, W.; Ma, W.; Ye, Q.; Liu, T.-Y. LightGBM: A Highly Efficient Gradient Boosting Decision Tree. In Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: New York, NY, USA, 2017; Available online: https://proceedings.neurips.cc/paper_files/paper/2017/file/6449f44a102fde848669bdd9eb6b76fa-Paper.pdf (accessed on 29 August 2025).
  27. Xu, H.; Kim, M. Combination prediction method of students’ performance based on ant colony algorithm. PLoS ONE 2024, 19, e0300010. [Google Scholar] [CrossRef] [PubMed]
  28. Fang, R.; Zhou, T.; Yu, B.; Li, Z.; Ma, L.; Luo, T.; Zhang, Y.; Liu, X. Prediction model of middle school student performance based on MBSO and MDBO-BP-Adaboost method. Front. Big Data 2025, 7, 1518939. [Google Scholar] [CrossRef] [PubMed]
  29. Cheng, B.; Liu, Y.; Jia, Y. Evaluation of students’ performance during the academic period using the XG-Boost Classifier-Enhanced AEO hybrid model. Expert Syst. Appl. 2024, 238, 122136. [Google Scholar] [CrossRef]
  30. Lowast, A.E.; Amalarethinam, D.I.G. A Hybrid Model of Enhanced Teacher Learner Based Optimization (ETLBO) with Particle Swarm Optimization (PSO) Algorithm for Predicting Academic Student Performance. INDJST 2025, 18, 772–783. [Google Scholar] [CrossRef]
  31. Li, G.; Cui, J.; Fu, H.; Sun, Y. Light GBM and GA Based Algorithm for Predicting and Improving Students’ Performance in Higher Education in the Context of Big Data. In Proceedings of the 2024 7th International Conference on Education, Network and Information Technology (ICENIT), Dalian, China, 16–18 August 2024; pp. 1–5. [Google Scholar] [CrossRef]
  32. Apriyadi, M.R.; Ermatita; Rini, D.P. Hyperparameter Optimization of Support Vector Regression Algorithm using Metaheuristic Algorithm for Student Performance Prediction. Int. J. Adv. Comput. Sci. Appl. 2023, 14, 144–150. [Google Scholar] [CrossRef]
  33. Kamal, M.; Chakrabarti, S.; Ramirez-Asis, E.; Asís-López, M.; Allauca-Castillo, W.; Kumar, T.; Sanchez, D.T.; Rahmani, A.W. Metaheuristics Method for Classification and Prediction of Student Performance Using Machine Learning Predictors. Math. Probl. Eng. 2022, 2022, 2581951. [Google Scholar] [CrossRef]
  34. Xu, H. Prediction of Students’ Performance Based on the Hybrid IDA-SVR Model. Complexity 2022, 2022, 1845571. [Google Scholar] [CrossRef]
  35. Song, X. Student performance prediction employing k-Nearest Neighbor Classification model and meta-heuristic algorithms. Multiscale Multidiscip. Model. Exp. Des. 2024, 7, 4397–4412. [Google Scholar] [CrossRef]
  36. Ma, C. Improving the Prediction of Student Performance by Integrating a Random Forest Classifier with Meta-Heuristic Optimization Algorithms. | EBSCOhost. Available online: https://openurl.ebsco.com/contentitem/doi:10.14569%2Fijacsa.2024.01506106?sid=ebsco:plink:crawler&id=ebsco:doi:10.14569%2Fijacsa.2024.01506106 (accessed on 30 August 2025).
  37. Ali, M.; Liaquat, M.D.; Atta, M.N.; Khan, A.; Lashari, S.A.; Ramli, D.A. Improving Student Performance Prediction Using a PCA-based Cuckoo Search Neural Network Algorithm. Procedia Comput. Sci. 2023, 225, 4598–4610. [Google Scholar] [CrossRef]
  38. Punitha, S.; Devaki, K. A high ranking-based ensemble network for student’s performance prediction using improved meta-heuristic-aided feature selection and adaptive GAN for recommender system. Kybernetes 2024. [Google Scholar] [CrossRef]
  39. Hai, Q.; Wang, C. Optimizing Student Performance Prediction: A Data Mining Approach with MLPC Model and Metaheuristic Algorithm. | EBSCOhost. Available online: https://openurl.ebsco.com/contentitem/doi:10.14569%2Fijacsa.2024.0150407?sid=ebsco:plink:crawler&id=ebsco:doi:10.14569%2Fijacsa.2024.0150407 (accessed on 30 August 2025).
  40. Mohammed, H.; Rashid, T. FOX: A FOX-inspired optimization algorithm. Appl. Intell. 2023, 53, 1030–1050. [Google Scholar] [CrossRef]
  41. Sadeeq, H.T.; Abdulazeez, A.M. Giant Trevally Optimizer (GTO): A Novel Metaheuristic Algorithm for Global Optimization and Challenging Engineering Problems. IEEE Access 2022, 10, 121615–121640. [Google Scholar] [CrossRef]
  42. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the Proceedings of ICNN’95—International Conference on Neural Networks, Perth, WA, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  43. Seyyedabbasi, A.; Kiani, F. Sand Cat swarm optimization: A nature-inspired algorithm to solve global optimization problems. Eng. Comput. 2022, 36, 2627–2651. [Google Scholar] [CrossRef]
  44. Mirjalili, S.; Gandomi, A.H.; Mirjalili, S.Z.; Saremi, S.; Faris, H.; Mirjalili, S.M. Salp Swarm Algorithm: A bio-inspired optimizer for engineering design problems. Adv. Eng. Softw. 2017, 114, 163–191. [Google Scholar] [CrossRef]
  45. Wang, J.; Chi, J.; Ding, Y.; Yao, H.; Guo, Q. Based on PCA and SSA-LightGBM oil-immersed transformer fault diagnosis method. PLoS ONE 2025, 20, e0314481. [Google Scholar] [CrossRef] [PubMed]
  46. Florek, P.; Zagdański, A. Benchmarking state-of-the-art gradient boosting algorithms for classification. arXiv 2023, arXiv:2305.17094. [Google Scholar] [CrossRef]
  47. Yang, H.; Chen, Z.; Yang, H.; Tian, M. Predicting Coronary Heart Disease Using an Improved LightGBM Model: Performance Analysis and Comparison. IEEE Access 2023, 11, 23366–23380. [Google Scholar] [CrossRef]
  48. Alshari, H.; Saleh, A.; Odabas, A. Comparison of Gradient Boosting Decision Tree Algorithms for CPU Performance. Erciyes Tip Derg. 2021, 157–168. [Google Scholar]
  49. Farhan, M.; Chen, T.; Rao, A.; Shahid, M.I.; Xiao, Q.; Salam, H.A.; Ma, F. An experimental study of knock analysis of HCNG fueled SI engine by different methods and prediction of knock intensity by particle swarm optimization-support vector machine. Energy 2024, 309, 133165. [Google Scholar] [CrossRef]
  50. Sun, C.; Qin, W.; Yun, Z. A State-of-Health Estimation Method for Lithium Batteries Based on Fennec Fox Optimization Algorithm–Mixed Extreme Learning Machine. Batteries 2024, 10, 87. [Google Scholar] [CrossRef]
  51. Tan, C.; Sun, M.; Liu, W.; Tan, W.; Zhang, X.; Zhu, C.; Li, D. An Adaptive Layering Dual-Parameter Regularization Inversion Method for an Improved Giant Trevally Optimizer Algorithm. IEEE Access 2024, 12, 160761–160775. [Google Scholar] [CrossRef]
  52. Punitha, A.; Ramani, P.; Ezhilarasi, P.; Sridhar, S. Dynamically stabilized recurrent neural network optimized with intensified sand cat swarm optimization for intrusion detection in wireless sensor network. Comput. Secur. 2025, 148, 104094. [Google Scholar] [CrossRef]
  53. Yang, Z.; Jiang, Y.; Yeh, W.-C. Self-learning salp swarm algorithm for global optimization and its application in multi-layer perceptron model training. Sci. Rep. 2024, 14, 27401. [Google Scholar] [CrossRef]
  54. Student Performance Factors. Available online: https://www.kaggle.com/datasets/lainguyn123/student-performance-factors (accessed on 17 August 2025).
  55. Wang, T.; Zhang, M.; Li, Z. Explainable machine learning links erosion damage to environmental factors on Gansu rammed earth Great Wall. npj Herit. Sci. 2025, 13, 366. [Google Scholar] [CrossRef]
  56. Ahmed, A.H.A.; Jin, W.; Ali, M.A.H. Prediction of compressive strength of recycled concrete using gradient boosting models. Ain Shams Eng. J. 2024, 15, 102975. [Google Scholar] [CrossRef]
  57. Haryadi, D.; Hakim, A.R.; Atmaja, D.M.U.; Yutia, S.N. Implementation of Support Vector Regression for Polkadot Cryptocurrency Price Prediction. JOIV Int. J. Inform. Vis. 2022, 6, 201–207. [Google Scholar] [CrossRef]
Figure 1. Proposed framework optimizer tuned LightGBM.
Figure 1. Proposed framework optimizer tuned LightGBM.
Applsci 15 10875 g001
Figure 2. Actual vs. predicted plots of compared models.
Figure 2. Actual vs. predicted plots of compared models.
Applsci 15 10875 g002aApplsci 15 10875 g002b
Figure 3. Absolute error plots of training and test results of compared models.
Figure 3. Absolute error plots of training and test results of compared models.
Applsci 15 10875 g003
Figure 4. SHAP violin plot of SCSO-LightGBM model.
Figure 4. SHAP violin plot of SCSO-LightGBM model.
Applsci 15 10875 g004
Figure 5. SHAP decision plot of SCSO-LightGBM model.
Figure 5. SHAP decision plot of SCSO-LightGBM model.
Applsci 15 10875 g005
Table 1. Summary of features.
Table 1. Summary of features.
Hours_StudiedAttendanceParental_InvolvementAccess_to_ResourcesExtracurricular_ActivitiesSleep_HoursPrevious_ScoresMotivation_Level
count43134313431343134313431343134313
mean19.9479.961.090.90.67.0275.010.91
std5.2411.450.690.70.491.214.340.69
min8600005500
25%16701006630
50%20801117751
75%24902118871
max3110022191002
Internet_AccessTutoring_SessionsFamily_IncomeSchool_TypePeer_InfluencePhysical_ActivityLearning_DisabilitiesGenderExam_Score
count431343134313431343134313431343134313
mean01.290.790.312.9600.4267.14
std00.970.740.460.890.9800.493.11
min0000010060
25%0100020065
50%0110130067
75%0211240169
max0321250175
Table 2. Optimization algorithm settings.
Table 2. Optimization algorithm settings.
OptimizerParameter Settings
FOX a = [ 0 , 2 ]
GTO A = [ 0.3 , 0.4 ] , H = [ 2 , 0 ]
PSO ω m a x = 0.9, ω m i n = 0.4
SCSO α = [0, 2 π ], C = 0.35 ,   S M = 2
SSA c 3 = [ 2 / e , 2 ]
Table 3. Training performance of LightGBM models using 5-fold cross-validation.
Table 3. Training performance of LightGBM models using 5-fold cross-validation.
FOX-LightGBMGTO-LightGBMPSO-LightGBMSCSO-LightGBMSSA-LightGBMLightGBMSVR
R2AVG0.96930.95810.95980.96090.96060.92240.9363
STD7.212 × 10−35.647 × 10−31.114 × 10−35.173 × 10−48.283 × 10−48.680 × 10−44.054 × 10−4
RMSEAVG3.606 × 10−24.234 × 10−24.155 × 10−24.100 × 10−24.113 × 10−25.775 × 10−25.231 × 10−2
STD4.198 × 10−32.884 × 10−34.934 × 10−43.221 × 10−44.349 × 10−43.102 × 10−42.069 × 10−4
MSEAVG1.318 × 10−31.801 × 10−31.727 × 10−31.681 × 10−31.692 × 10−33.335 × 10−32.737 × 10−3
STD3.073 × 10−42.392 × 10−44.124 × 10−52.650 × 10−53.595 × 10−53.581 × 10−52.145 × 10−5
MEAVG1.365 × 10−11.467 × 10−11.436 × 10−11.396 × 10−11.440 × 10−11.943 × 10−11.008 × 10−1
STD1.515 × 10−27.976 × 10−38.224 × 10−37.349 × 10−37.337 × 10−31.046 × 10−25.541 × 10−4
RAEAVG6.947 × 10−28.156 × 10−28.005 × 10−27.900 × 10−27.925 × 10−21.113 × 10−11.008 × 10−1
STD8.158 × 10−35.286 × 10−37.716 × 10−47.940 × 10−41.060 × 10−38.750 × 10−46.880 × 10−4
Table 4. Held-out fold performance of LightGBM models using 5-fold cross-validation.
Table 4. Held-out fold performance of LightGBM models using 5-fold cross-validation.
FOX-LightGBMGTO-LightGBMPSO-LightGBMSCSO-LightGBMSSA-LightGBMLightGBMSVR
R2AVG0.93310.93430.94060.94120.94090.90740.9257
STD4.131 × 10−34.544 × 10−31.238 × 10−32.060 × 10−31.963 × 10−31.672 × 10−32.589 × 10−3
RMSEAVG5.354 × 10−25.306 × 10−25.049 × 10−25.021 × 10−25.034 × 10−26.303 × 10−25.647 × 10−2
STD1.549 × 10−32.054 × 10−36.341 × 10−41.011 × 10−35.022 × 10−49.757 × 10−41.193 × 10−3
MSEAVG2.869 × 10−32.820 × 10−32.550 × 10−32.522 × 10−32.534 × 10−33.974 × 10−33.190 × 10−3
STD1.679 × 10−42.177 × 10−46.447 × 10−51.031 × 10−45.030 × 10−51.236 × 10−41.353 × 10−4
MEAVG1.755 × 10−11.694 × 10−11.620 × 10−11.637 × 10−11.638 × 10−12.071 × 10−12.117 × 10−1
STD1.543 × 10−21.391 × 10−26.862 × 10−38.939 × 10−31.586 × 10−22.015 × 10−22.909 × 10−2
RAEAVG1.032 × 10−11.022 × 10−19.731 × 10−29.677 × 10−29.701 × 10−21.215 × 10−11.088 × 10−1
STD3.456 × 10−33.861 × 10−32.182 × 10−32.745 × 10−31.314 × 10−33.084 × 10−33.314 × 10−3
Table 5. Training results from 20 independent runs using optimized LightGBM models.
Table 5. Training results from 20 independent runs using optimized LightGBM models.
FOX-LightGBMGTO-LightGBMPSO-LightGBMSCSO-LightGBMSSA-LightGBMLightGBMSVR
R2AVG0.96820.96100.95900.96090.96080.92290.9371
STD8.981 × 10−38.017 × 10−32.782 × 10−35.743 × 10−48.170 × 10−42.220 × 10−160
Best0.98280.97320.96330.96180.96220.92290.9371
RMSEAVG3.667 × 10−24.085 × 10−24.210 × 10−24.111 × 10−24.111 × 10−25.772 × 10−25.217 × 10−2
STD5.348 × 10−34.342 × 10−31.398 × 10−33.012 × 10−43.012 × 10−46.939 × 10−182.776 × 10−17
Best2.730 × 10−23.402 × 10−23.986 × 10−24.065 × 10−24.065 × 10−25.772 × 10−25.217 × 10−2
MSEAVG1.373 × 10−31.688 × 10−31.774 × 10−31.690 × 10−31.697 × 10−33.332 × 10−32.722 × 10−3
STD3.884 × 10−43.467 × 10−41.203 × 10−42.472 × 10−53.530 × 10−58.674 × 10−194.337 × 10−19
Best7.450 × 10−41.158 × 10−31.589 × 10−31.653 × 10−31.636 × 10−33.332 × 10−32.722 × 10−3
MEAVG1.340 × 10−11.483 × 10−11.521 × 10−11.444 × 10−11.432 × 10−11.980 × 10−11.005 × 10−1
STD1.588 × 10−21.516 × 10−21.253 × 10−27.535 × 10−38.347 × 10−304.163 × 10−17
Best1.006 × 10−11.248 × 10−11.323 × 10−11.343 × 10−11.331 × 10−11.980 × 10−11.005 × 10−1
RAEAVG7.028 × 10−27.829 × 10−28.069 × 10−27.880 × 10−27.895 × 10−21.106 × 10−19.999 × 10−2
STD1.025 × 10−28.323 × 10−32.679 × 10−35.773 × 10−48.205 × 10−42.776 × 10−174.163 × 10−17
Best5.233 × 10−26.521 × 10−27.640 × 10−27.791 × 10−27.751 × 10−21.106 × 10−19.999 × 10−2
Table 6. Test results from 20 independent runs using optimized LightGBM models.
Table 6. Test results from 20 independent runs using optimized LightGBM models.
FOX-LightGBMGTO-LightGBMPSO-LightGBMSCSO-LightGBMSSA-LightGBMLightGBMSVR
R2AVG0.93320.93160.93640.93860.93880.90320.9196
STD3.275 × 10−33.517 × 10−32.868 × 10−31.296 × 10−36.900 × 10−44.441 × 10−162.220 × 10−16
Best0.93910.93820.94040.94060.93990.90320.9196
RMSEAVG5.283 × 10−25.348 × 10−25.156 × 10−25.069 × 10−25.058 × 10−26.363 × 10−25.799 × 10−2
STD1.294 × 10−31.366 × 10−31.152 × 10−35.344 × 10−42.851 × 10−41.388 × 10−170
Best5.048 × 10−25.086 × 10−24.995 × 10−24.986 × 10−25.015 × 10−26.363 × 10−25.799 × 10−2
MSEAVG2.793 × 10−32.862 × 10−32.660 × 10−32.570 × 10−32.559 × 10−34.048 × 10−33.363 × 10−3
STD1.370 × 10−41.471 × 10−41.199 × 10−45.428 × 10−52.898 × 10−500
Best2.548 × 10−32.586 × 10−32.495 × 10−32.486 × 10−32.515 × 10−34.048 × 10−33.363 × 10−3
MEAVG1.743 × 10−11.816 × 10−11.758 × 10−11.610 × 10−11.657 × 10−12.201 × 10−12.037 × 10−1
STD9.157 × 10−31.867 × 10−21.419 × 10−21.162 × 10−21.029 × 10−22.776 × 10−175.551 × 10−17
Best1.502 × 10−11.608 × 10−11.510 × 10−11.508 × 10−11.476 × 10−12.201 × 10−12.037 × 10−1
RAEAVG1.040 × 10−11.053 × 10−11.015 × 10−19.976 × 10−29.955 × 10−21.252 × 10−11.141 × 10−1
STD2.546 × 10−32.688 × 10−32.268 × 10−31.052 × 10−35.610 × 10−406.939 × 10−17
Best9.935 × 10−21.001 × 10−19.831 × 10−29.814 × 10−29.870 × 10−21.252 × 10−11.141 × 10−1
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Abukader, A.; Alzubi, A.; Adegboye, O.R. Intelligent System for Student Performance Prediction: An Educational Data Mining Approach Using Metaheuristic-Optimized LightGBM with SHAP-Based Learning Analytics. Appl. Sci. 2025, 15, 10875. https://doi.org/10.3390/app152010875

AMA Style

Abukader A, Alzubi A, Adegboye OR. Intelligent System for Student Performance Prediction: An Educational Data Mining Approach Using Metaheuristic-Optimized LightGBM with SHAP-Based Learning Analytics. Applied Sciences. 2025; 15(20):10875. https://doi.org/10.3390/app152010875

Chicago/Turabian Style

Abukader, Abdalhmid, Ahmad Alzubi, and Oluwatayomi Rereloluwa Adegboye. 2025. "Intelligent System for Student Performance Prediction: An Educational Data Mining Approach Using Metaheuristic-Optimized LightGBM with SHAP-Based Learning Analytics" Applied Sciences 15, no. 20: 10875. https://doi.org/10.3390/app152010875

APA Style

Abukader, A., Alzubi, A., & Adegboye, O. R. (2025). Intelligent System for Student Performance Prediction: An Educational Data Mining Approach Using Metaheuristic-Optimized LightGBM with SHAP-Based Learning Analytics. Applied Sciences, 15(20), 10875. https://doi.org/10.3390/app152010875

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop