1. Introduction
The education sector is changing as artificial intelligence (AI) is increasingly integrated into higher education, aligning with the global agenda of the United Nations’ Sustainable Development Goals (SDGs). Universities are now expected not only to deliver disciplinary knowledge but also to develop students’ cognitive, technological, and sustainability-related competencies that prepare them to address complex global challenges [
1,
2]. Mathematics and other quantitatively intensive subjects are central to this agenda, as they foster problem-solving, modelling, and evidence-based decision-making skills that directly support SDG4 (Quality Education), SDG9 (Industry, Innovation, and Infrastructure), and SDG13 (Climate Action) [
3,
4]. Contemporary mathematics education research further highlights that advanced problem-solving and analytical thinking are essential for success in modern, technology-enhanced learning environments [
5,
6].
From a cognitive science perspective, higher-order cognitive processes play a crucial role in advanced mathematics learning. Working memory (WM) enables the temporary storage and manipulation of information required for complex tasks and has been consistently identified as a strong predictor of mathematics achievement across educational levels [
7,
8]. Metacognition (MET), encompassing planning, monitoring, and self-evaluation of learning, supports students in regulating strategies and adapting their approaches, thereby enhancing reasoning and academic performance [
9,
10]. Reasoning ability (REA) represents a core mathematical competence, particularly in STEM contexts, facilitating abstract thinking, conceptual understanding, and knowledge transfer [
2,
6]. WM, MET, and REA can potentially form a cognitive foundation that guides how students engage with complex mathematical ideas in digital and AI-supported learning environments.
AI-supported learning (AI-SL) systems are now widely adopted in universities worldwide. AI-SL offers intelligent tutoring, automated feedback, adaptive assessment, personalized learning pathways, and advanced analytics that support both learners and instructors [
11,
12,
13]. Recent studies indicate that generative AI can improve learning efficiency and academic outcomes, while persistent concerns remain about academic integrity, cognitive offloading, and independent thinking [
14]. Scholarly work, hence, stresses the need for responsible AI integration guided by institutional policies that encourage ethical use and sustained student engagement [
15]. Empirical research also indicates that AI-SL effectiveness depends not only on technological sophistication but also on learners’ cognitive readiness and self-regulatory capacity [
16].
Recent work by Putthidech et al. [
17] builds on earlier studies and reports that students’ views of AI-based learning systems, e.g., adaptability, feedback accuracy, and trust, bear a strong relation to calculus achievement through cognitive and emotional pathways. The investigation, however, did not clarify the separate contributions of working memory, metacognition, and reasoning ability, and it left out sustainability-oriented competencies, including Sustainable Development Goal awareness (SDGA).
Sustainability has likewise become a strategic priority in higher education. Integrating sustainability content into STEM education promotes systems thinking and sustainability literacy [
4,
18,
19,
20], while learning about the SDGs has been shown to enhance student motivation, prosocial learning, and engagement [
21,
22]. SDGA may strengthen learner engagement (ENG) by increasing the perceived relevance of academic content to real-world and societal challenges. When students recognize the social purpose and sustainability implications of their learning, they are more likely to invest greater cognitive effort and emotional commitment. Moreover, SDG-informed instruction can foster systems thinking, encouraging learners to integrate knowledge across domains and engage more deeply with complex and ill-structured problems. Despite these promising insights, relatively few studies have examined how SDGA interacts with cognitive processes (WM, MET, and REA) and AI-SL to jointly shape ENG and learning outcomes (LO) in higher education.
Relationships among learning factors may vary across academic disciplines. STEM programmes place greater emphasis on abstract reasoning and algorithmic thinking, which aligns more closely with working memory and reasoning ability, whereas social science programmes prioritize reflective and conceptual thinking that relates more strongly to metacognition [
23,
24]. Systematic comparisons of disciplinary differences within a single analytical framework remain limited. The present study seeks to clarify how cognitive processes comprising WM, MET, and REA, AI-supported learning, and Sustainable Development Goal awareness relate to learning engagement and learning outcomes. A multi-group structural equation modeling analysis compares STEM and social science students to examine disciplinary variation. The central argument holds that linking cognitive, technological, and sustainability dimensions is essential for understanding learning engagement and outcomes in contemporary higher education.
Proposed hypotheses are, thus, grouped into three themes aligned with the study objectives. Group H1–H6 examines links between WM, MET, and REA with ENG and LO. Group H7–H12 addresses roles of AI-SL and SDGA in ENG and LO and their indirect paths through ENG. Group H13–H18 investigates disciplinary differences by testing through MG-SEM whether structural relations among WM, MET, REA, AI-SL, SDGA, ENG, and LO differ between STEM and social science students. All hypotheses (H1–H18) follow a consistent format and align directly with the proposed conceptual framework.
H1. WM is positively associated with ENG.
H2. MET is positively associated with ENG.
H3. REA is positively associated with ENG.
H4. WM is positively associated with LO.
H5. MET is positively associated with LO.
H6. REA is positively associated with LO.
H7. AI-SL is positively associated with ENG.
H8. AI-SL is positively associated with LO.
H9. SDGA is positively associated with ENG.
H10. SDGA is positively associated with LO.
H11. ENG is positively associated with LO.
H12. ENG mediates the relationship between WM and LO.
H13. ENG mediates the relationship between MET and LO.
H14. ENG mediates the relationship between REA and LO.
H15. ENG mediates the relationship between AI-SL and LO.
H16. ENG mediates the relationship between SDGA and LO.
H17. The relationships between cognitive processes (WM, MET, and REA) and ENG differ between STEM and social science students.
H18. The effects of AI-SL and SDGA on ENG and LO differ between STEM and social science students.
The research anticipates that STEM students may experience stronger changes associated with AI implementation, whereas social science students may achieve better results through SDG education.
2. Materials and Methods
The study was approved by Rajamangala University of Technology Suvarnabhumi on Ethics in Human Research (IRB-RUS-2568-104).
2.1. Research Design
This study used a survey-based research design to examine the links among mental processes, AI-supported learning (AI-SL), sustainable development goal awareness (SDGA), learner engagement (ENG), and learning outcomes (LO) in higher education. Undergraduate students from nine campuses within the Rajamangala University of Technology system in Thailand completed a straightforward questionnaire, and data were collected at a single point in time.
Following data collection, Structural Equation Modeling (SEM) served as the main analytical method for the theoretically specified relationships among latent constructs. The analysis began with a confirmatory factor analysis (CFA) to validate the measurement model, followed by estimation of the structural model to examine association patterns among constructs.
The analysis compared STEM students with Social Science students to examine disciplinary differences. Measurement invariance testing was conducted to confirm that both groups interpreted the survey items in a comparable way before any group comparisons were made. Data were collected at a single point in time, so the findings indicate associations rather than causal effects. The design supports testing theoretically guided relationships and identifying links that can be examined more rigorously through longitudinal or experimental research.
2.2. Measures
All study constructs were measured with a structured self-report questionnaire. This instrument captured students’ perceptions, cognitive processes (WM, MET, and REA), and learning behaviors. Instruments were adapted from the established literature and modified for use in higher education in Thailand. Items were reviewed for clarity, relevance, and content validity before data collection. To record responses, a five-point Likert scale was used, ranging from 1 (‘strongly disagree’) to 5 (‘strongly agree’), unless noted. Higher scores indicate greater levels of each construct.
2.2.1. Cognitive Factors
- (a)
Working Memory (WM)
WM is a core cognitive system that supports the temporary storage and manipulation of information required for complex learning, reasoning, and problem-solving. In higher education, WM enables students to process academic material and manage cognitively demanding tasks, particularly in mathematics and technology-supported learning environments. Recent research consistently demonstrates that individual differences in WM capacity are strongly associated with ENG, problem-solving efficiency, and academic achievement [
8,
25].
Building upon this, maintaining information during academic tasks (WM1) refers to students’ ability to keep task-relevant information actively available while performing learning activities. This enables learners to retain instructions, key concepts, or problem conditions long enough for meaningful cognitive processing. When students can do this effectively, they are better equipped to sustain attention and avoid cognitive disruption during demanding academic tasks [
25].
Updating information while learning (WM2) reflects students’ capacity to revise and replace information in WM as new or more accurate input becomes available. This process supports cognitive flexibility, enabling learners to adjust their understanding and strategies in response to feedback or changing task requirements. Effective updating has been identified as a central executive function that facilitates deep learning and adaptive problem solving in dynamic learning environments [
26,
27].
Managing multiple ideas simultaneously (WM3) captures the ability to coordinate and integrate multiple pieces of information. This capability is essential for higher-order reasoning, conceptual integration, and decision-making in complex academic contexts. Furthermore, research grounded in cognitive load theory suggests that students who efficiently manage multiple ideas experience reduced cognitive overload and improved learning performance [
28,
29,
30].
Tracking multistep problem-solving processes (WM4) refers to students’ capacity to monitor and retain sequential steps in tasks that demand extended reasoning. This aspect of WM is especially important in mathematics and analytical disciplines, where solutions depend on interconnected steps and intermediate results. Strong multistep tracking ability supports logical coherence and accuracy in problem solving and relates to stronger mathematical reasoning and academic performance [
6,
7]. Sample items include “I can remember important details even when I am focused on other things at the moment” and “I am able to update and adjust information in my mind while working through a task.” The scale assesses WM capacity through a scoring system in which higher scores indicate stronger capacity. A pilot study found that the WM demonstrated acceptable internal consistency, with a Cronbach’s alpha of 0.83, supporting its suitability for research in higher education contexts.
- (b)
Metacognition (MET)
MET refers to individuals’ awareness and regulation of their own cognitive processes during learning. It involves the ability to plan, monitor, and evaluate one’s understanding and strategies to achieve learning goals effectively [
31]. In higher education, MET is widely recognised as a key component of self-regulated learning and has been shown to enhance academic engagement, problem-solving efficiency, and LO, particularly in complex and technology-supported learning environments [
10,
27].
Planning learning strategies (MET1) reflects students’ ability to set goals, select appropriate strategies, and allocate cognitive resources before engaging in learning tasks. Effective planning helps learners approach academic activities with clear intentions and structured methods, which supports deeper engagement and more efficient learning processes. Research indicates that students who actively plan their learning tend to demonstrate higher levels of academic control and persistence [
10].
Complementing planning, monitoring, and understanding (MET2) refers to students’ ongoing awareness of their comprehension and performance during learning. This ability enables learners to detect confusion, identify errors, and recognise when additional effort or alternative strategies are needed. Continuous monitoring is essential for maintaining learning effectiveness, especially in complex tasks that require sustained attention and adaptive thinking [
27,
32].
Evaluating learning progress (MET3) captures students’ ability to reflect on their LO and assess whether their goals have been achieved. In turn, this evaluative process enables learners to assess the effectiveness of their strategies and identify areas for improvement. Furthermore, empirical studies show that reflective evaluation supports long-term learning and improves academic performance in higher education contexts [
10,
33].
Adjusting strategies when ineffective (MET4) refers to students’ capacity to modify learning approaches when current strategies do not lead to successful outcomes. Adaptive regulation enables flexible responses to challenges and changing task demands. Strategic adjustment characterizes effective self-regulated learners and relates to stronger engagement and resilience in demanding academic contexts [
27,
32]. Sample items include “I monitor my reading comprehension throughout my work on all course assignments” and “I change my learning approach when I understand that my current methods are not leading to successful results.” The assessment evaluates MET ability by capturing awareness and regulation of cognitive processes through a scoring system in which higher ratings reflect stronger skills. Pilot testing indicated high internal consistency for the MET items, with a Cronbach’s alpha of 0.88, supporting their use in higher education research.
- (c)
Reasoning Ability (REA)
REA refers to individuals’ capacity to think logically, identify relationships among concepts, and draw valid conclusions to solve problems. In higher education, REA is a fundamental cognitive competence that underpins problem-solving, decision-making, and knowledge transfer across disciplines. Recent studies highlight that strong reasoning skills are particularly critical in mathematics, STEM-related fields, and technology-enhanced learning environments, where learners must analyse information systematically and apply concepts to novel situations [
2,
6].
Building on this foundation, logical reasoning and inference (REA1) reflects students’ ability to evaluate arguments, identify logical relationships, and draw sound conclusions from available information. This process enables students to interpret data, assess the validity of solutions, and avoid reasoning errors in academic tasks. Empirical evidence shows strong associations between logical reasoning and both academic engagement and performance in analytically demanding subjects [
6].
In addition, analytical problem-solving (REA2) means students can break down complex problems into manageable components and use systematic strategies to reach solutions. This skill helps learners structure their thinking, compare approaches, and select solution paths. Research suggests that analytical problem-solving is key to engagement and improved outcomes in higher education, especially in problem-based and AI-supported learning contexts [
2,
27].
Applying reasoning to new situations (REA3) refers to students’ capacity to reason in unfamiliar contexts. The construct reflects flexible thinking and the ability to extend reasoning beyond routine tasks. Mathematics education research links cross-context reasoning with advanced cognitive competence and regards it as essential for meaningful learning and long-term academic success [
6,
24]. Sample items include “The AI tools I use enable me to learn complex subjects at a better level” and “AI tools, including essay grading software and automated math solvers, provide feedback that helps me enhance my learning activities.” Higher REA is indicated by higher scores on the scale. A Cronbach’s alpha of 0.86 indicated strong internal consistency for the REA scale, supporting its use in higher education research.
2.2.2. AI-Supported Learning (AI-SL)
AI-SL refers to students’ perceptions of how artificial intelligence–based tools and systems support, enhance, and personalise their learning processes in higher education. In contemporary university settings, AI-SL plays an increasingly important role by providing adaptive learning pathways, automated feedback, intelligent tutoring, and real-time learning analytics. These features are designed to support students’ cognitive processing, sustain engagement, and improve LO, particularly in mathematically and analytically demanding subjects [
11,
12].
Perceived usefulness of AI tools for learning (AI-SL1) reflects students’ beliefs that AI systems help them better understand course content. Learners who perceive AI tools as useful are more likely to integrate them into routines and stay engaged. Prior studies suggest that perceived usefulness (the belief that technology improves learning) is a key factor in students’ acceptance of AI-based learning systems. This perception also affects continued use in higher education [
11,
17].
The accuracy of AI-generated feedback (AI-SL2) captures how correct, clear, and relevant students find AI-SL2. Accurate, timely feedback helps students spot errors, refine understanding, and use adaptive learning strategies. Studies show high-quality AI feedback improves learning and supports deeper engagement, especially in problem-solving and quantitative tasks [
12,
14].
Trust in AI-supported learning systems (AI-SL3) reflects students’ confidence in the reliability (consistency and dependability) and fairness (impartiality and equal treatment) of AI tools. Students must trust AI recommendations (suggestions or solutions) and feedback to rely on them without doubt. Evidence shows that trust in AI increases ENG and mediates the link between AI use and LO [
17].
AI assistance in understanding complex content (AI-SL4) matches students’ views. They say AI tools explain difficult concepts, guide problem-solving (help break down and solve problems), and reduce cognitive load (lessen the mental effort needed for learning). This dimension shows how AI-SL supports students’ WM (ability to hold and manipulate information), MET (awareness of one’s thinking), and REA (logical analysis of information) [
16,
29]. Studies show this assistance helps sustain engagement and improve academic performance in higher education [
11,
27]. Example items include “The AI tools I use help me understand complex topics more clearly.” And “AI-generated feedback helps me improve my learning strategies.” Higher scores on the AI-SL scale (a measurement assessing students’ views of AI-SL) indicate stronger perceived support from AI-based learning systems. Pilot testing revealed high internal consistency for the AI-SL scale. Cronbach’s alpha coefficient (a statistic measuring internal consistency) exceeded 0.90. This supports its suitability for research on AI-enhanced learning in higher education.
2.2.3. SDG Awareness (SDGA)
SDGA encompasses students’ comprehension of the UN Sustainable Development Goals (SDGs) and awareness of sustainability challenges at both global and societal scales. In higher education, SDGA has emerged as an important LO that supports students’ motivation, ethical responsibility, and ENG with socially relevant academic content. Recent studies indicate that integrating SDG-related knowledge into university curricula is positively related to students’ engagement in learning and contributes to more meaningful, purpose-driven learning experiences [
1,
22].
Building on this foundation, understanding Sustainable Development Goal (SDG1) reflects students’ knowledge of the core principles, objectives, and scope of the SDGs. This dimension captures learners’ awareness of how sustainability goals relate to education, society, and future professional roles. Students who clearly understand the SDGs tend to place greater value on sustainability-oriented learning and show higher levels of academic engagement [
22].
In addition to understanding the SDGs, awareness of global sustainability challenges (SDG2) involves students recognizing the major environmental, social, and economic issues they address, such as climate change, inequality, and resource scarcity. This awareness enables learners to connect academic content with real-world problems and fosters systems thinking. Empirical evidence shows that exposure to global sustainability challenges is positively related to students’ critical thinking and motivation to engage in interdisciplinary learning [
18,
21].
Responsibility for sustainable development (SDG3) reflects students’ personal and social responsibility. The construct encourages support for sustainable development through individual actions and future career choices. The dimension captures the attitudinal and ethical aspects of SDGA. Higher education research links a strong sense of sustainability responsibility with ENG, prosocial orientations, and long-term societal commitment [
4,
22]. Sample items include “I understand the importance of the United Nations Sustainable Development Goals” and “I take it as my duty to help solve worldwide sustainability problems which affect the planet.” The SDGA assessment yields higher scores for stronger performance. Pilot testing reported acceptable internal consistency for the SDGA scale, with Cronbach’s alpha of 0.87, supporting its suitability for higher education research.
2.2.4. Learner Engagement (ENG)
ENG refers to the degree to which students are actively involved in their learning activities through behavioral participation, cognitive investment, and emotional commitment. In higher education, ENG is widely recognised as a central mechanism linking instructional practices, cognitive processes (WM, MET, and REA), and LO. Recent research consistently demonstrates that higher levels of engagement are associated with improved academic performance, persistence, and deeper learning, particularly in technology-enhanced and student-centred learning environments [
34,
35].
Active participation in learning activities (ENG1) means students are visibly involved in academic tasks, such as joining discussions, completing assignments, and putting effort into coursework. Here, “active participation” refers to the direct, purposeful actions students take to engage with materials, peers, or instructors. This readiness to engage has been linked to higher achievement and persistence in higher education [
34].
In addition to active participation, sustained attention and effort (ENG2) means students stay focused, concentrated, and persistent with academic tasks, even when they are challenging. Sustained engagement supports deeper thinking and lowers the chance of disengagement in complex situations. Studies suggest that students who sustain effort achieve better outcomes and remain motivated [
35].
Cognitive and emotional involvement in learning (ENG3) reflects students’ psychological investment, including interest, enjoyment, and perceived relevance of the content. This dimension combines cognitive engagement and emotional commitment, enhancing meaningful learning and long-term success. Recent research shows that emotional and cognitive engagement are critical for deep understanding and adaptive learning in higher education [
34,
36].
2.2.5. Learning Outcomes (LO)
LO refers to the extent to which students achieve intended academic goals as a result of their learning experiences. In higher education, learning outcomes encompass both perceived learning gains and demonstrated academic performance, reflecting students’ cognitive development, problem-solving ability, and mastery of course content. Recent research emphasizes that LO are shaped not only by instructional quality but also by students’ cognitive processes (WM, MET, and REA), ENG, and self-regulation, particularly in technology-supported learning environments [
35,
36].
Achievement of learning objectives (LO1) reflects students’ perceived attainment of course goals and competencies. This dimension captures whether learners believe they have met academic standards. Studies indicate that perceived achievement of learning objectives closely relates to motivation, engagement, and persistence and serves as an important indicator of instructional effectiveness [
35].
Improvement in subject understanding (LO2) refers to students’ perceived growth in conceptual knowledge and comprehension of course material. This dimension reflects deeper learning beyond rote memorization, emphasizing meaningful understanding and the ability to explain or apply concepts. Empirical evidence suggests that improvements in subject understanding are strongly associated with active engagement and cognitively demanding learning activities, particularly in mathematics and problem-based learning contexts [
27,
36].
Confidence in problem-solving ability (LO3) refers to students’ judgments about their capacity to solve academic problems. The construct signals readiness to transfer learning to new challenges. Research links stronger confidence with higher achievement, sustained engagement, and positive academic progress in higher education [
6,
17]. Sample items include “I have developed better skills to tackle problems that the course presents” and “I have reached all the learning targets that the subject established.” Higher values on perceived and objective measures indicate stronger LO. The Perceived Learning Scale demonstrated strong internal consistency, with a Cronbach’s alpha of 0.91, supporting its use in higher education research.
The constructs and measurement items are shown in
Table 1, along with item codes, brief descriptions, number of items, and source references.
2.2.6. Content Validity and Expert Review
All measurement items underwent a systematic expert review before data collection. The questionnaire was developed by adapting validated instruments from previous studies by Schraw and Moshman [
37]. These instruments were refined to fit higher education, AI-SL, and sustainability education. Clarity and construct relevance were emphasized. Alignment between theoretical definitions and observed indicators was also ensured. Building upon this foundation, the following steps enhanced the instrument’s rigor.
A panel of three experts in educational psychology, mathematics education, and instructional technology evaluated each item independently. Item–Objective Congruence (IOC) values were calculated to assess content validity. Items with IOC values below 0.50 were revised or removed, based on expert feedback. Minor wording adjustments enhanced conceptual clarity and contextual appropriateness. After incorporating these revisions, additional validation steps were taken.
The revised instrument was pilot-tested with undergraduate students who were not included in the final sample to assess item clarity and response consistency. Reliability analysis from the pilot data indicated acceptable internal consistency across all constructs. These combined phases of expert review and pilot testing confirmed that the measurement instrument was valid, reliable, and suitable for investigating cognitive processes (WM, MET, and REA), ENG, and LO in higher education settings.
2.2.7. Pilot Testing
A pilot study was conducted with 50 undergraduate students from Rajamangala University of Technology Suvarnabhumi. These students were excluded from the final study sample. The pilot testing was intended to assess the appropriateness of the measurement tools prior to extensive data collection. It specifically evaluated item clarity, suitability for higher education and AI-SL environments, and response variability across all constructs.
The results showed that all questionnaire items were clearly understood by participants. The items were appropriately worded for the target population and showed sufficient response variance. No items were reported as confusing or ambiguous. Only minor wording refinements were made to enhance clarity. Reliability analysis indicated strong internal consistency, with Cronbach’s alpha coefficients exceeding 0.82 for all constructs.
Due to the sample size limitations, factor analysis was not performed during pilot testing. Stable factor structure assessment requires larger samples. Confirmatory factor analysis (CFA) was therefore conducted on the full dataset (N = 762) to examine construct validity and the factorial structure, following established guidelines for structural equation modeling (SEM). The combination of pilot reliability findings and expert content validation demonstrated strong measurement quality across the study instruments.
2.3. Participants
Undergraduate students from mathematics, general education, and interdisciplinary courses at the nine campuses of the Rajamangala University of Technology system in Thailand participated in the study, offering academic and institutional diversity. Including these campuses ensured a range of educational environments and approaches, including AI-based learning systems, thereby enhancing the study’s findings.
The research employed stratified sampling by dividing undergraduates into two strata: STEM fields (engineering, industrial technology) and applied/social sciences (business administration, humanities, social sciences). A total of 800 participants were randomly selected, including 412 from STEM and 388 from the social sciences. After screening for missing values, identifying multivariate outliers using Mahalanobis distance, and checking for response inconsistencies, 762 valid responses remained for analysis.
The research sample size exceeded the minimum requirements for SEM. With 35 model parameters, Kline [
38] suggests at least 350–400 participants. For multi-group SEM (MG-SEM), Hair et al. [
39] recommend a minimum of 200 participants per group to achieve analytical stability and power. The final dataset thus met these criteria, supporting reliable SEM and Multi-Group analyses of configural, metric, and scalar invariance.
Participants ranged in age from 18 to 24 years old (M = 20.6, SD = 1.4); 56% were female, 43% were male, and 1% identified as other. All participants provided electronic consent, and the study complied with the university’s research ethics guidelines.
2.4. Data Collection Procedure
The Institutional Review Board of Rajamangala University of Technology Suvarnabhumi approved the study. The target population included undergraduate students from nine campuses in the university system. Before collecting data, we obtained permission from academic units. We informed participants about the study’s purpose, voluntary participation, and data confidentiality.
The questionnaire was administered using an online survey platform, which facilitated access across campuses and ensured standardized data collection procedures. To achieve adequate representation of students from both STEM and Social Science disciplines, a stratified sampling approach was employed. The survey link was then distributed through official university channels and course instructors during the academic semester.
Participants provided electronic informed consent before responding to the questionnaire. Only students who agreed to participate continued to the survey. The questionnaire took about 15–20 min to complete. Data collection lasted for four weeks to allow sufficient participation time. After data collection, responses were screened for missing values, inconsistencies, and outliers. Consequently, incomplete or invalid responses were excluded. As a result, the final dataset was suitable for statistical analyses, including CFA and SEM.
2.5. Data Analysis
Data analysis used a two-stage SEM approach. Prior to hypothesis testing, researchers checked the dataset for missing values, outliers, and distributional issues. The expectation–maximization method handled minimal missing data, while skewness and kurtosis were used to assess univariate normality. Researchers identified multivariate outliers using Mahalanobis distance. Variance inflation factors (VIFs) were used to assess multicollinearity; all values were below the acceptable threshold.
The first stage consisted of CFA to assess the measurement model. This process examined factor loadings, which indicate the strength of the relationship between each item and its underlying factor, alongside construct reliability and validity, which measure the consistency and accuracy with which each construct is assessed. To evaluate convergent validity (ensuring related items are genuinely related), analysis included standardized factor loadings, composite reliability (CR), and average variance extracted (AVE). In contrast, discriminant validity (the distinction between separate constructs) was assessed using the Fornell–Larcker criterion and the heterotrait–monotrait (HTMT) ratio. Model fit, reflecting how well the model aligned with the data, was assessed using indices such as the comparative fit index (CFI), the Tucker–Lewis index (TLI), the root mean square error of approximation (RMSEA), and the standardized root mean square residual (SRMR). The second stage, SEM, was used to examine the proposed connections among cognitive processes (WM, MET, and REA), AI-SL, SDGA, ENG, and LO. SEM is a statistical technique used to test complex relationships. Standardized path coefficients, which express the strength of relationships on a common scale, estimated both direct and indirect effects. Bootstrapping, a resampling technique, was used with 5000 samples and bias-corrected confidence intervals to examine mediation effects. Mediation occurs when one variable is associated with the relationship between two others.
MG-SEM analysis compared STEM (Science, Technology, Engineering, and Mathematics) and Social Science students to examine differences between disciplines. Researchers sequentially tested measurement invariance at three levels: configural (the same basic factor structure), metric (equal factor loadings), and scalar (equal item intercepts). After confirming these levels, they made structural comparisons. Analyses adhered to established SEM guidelines for educational research.
2.5.1. Preliminary Screening and Assumptions
The research team conducted data screening to verify that the data met SEM requirements, following established procedures. The team allowed for less than 5% missing data points. They determined these values could be imputed using the expectation–maximization method [
40]. The researchers used skewness values between −2 and 2 and kurtosis values between −2 and 2 to check for univariate normality [
38]. VIF values below 5 indicated the absence of significant multicollinearity [
39]. The Mahalanobis distance test (
p < 0.001) identified multivariate outliers using the chi-square distribution method by Kline [
38]. The analysis used robust maximum likelihood estimation (MLR) when researchers detected non-normality, as this method performs best for such data [
41]. This process established that the data fulfilled all necessary conditions for CFA and MG-SEM analysis.
2.5.2. Confirmatory Factor Analysis (CFA)
The CFA evaluated the measurement model of all latent variables using the approach of Hair et al. [
39]. The model fit assessment included several indicators. CFI and TLI values were above 0.90, RMSEA values were below 0.08, and SRMR values were below 0.08. These values indicated acceptable model fit [
42]. Convergent validity was established through factor loadings, all of which exceeded 0.50. CR values surpassed 0.70, and AVE values exceeded 0.50 [
43]. The research team applied two verification methods to establish discriminant validity. The Forell–Larcker criterion required the square root of each AVE to exceed the inter-construct correlations. The HTMT ratio also had to remain below 0.85 [
44]. All constructs met the required thresholds, confirming their reliability and validity for future SEM analyses.
2.5.3. Measurement Invariance Testing
The research employed MG-SEM to assess STEM students relative to Social Science students. Measurement invariance was examined in three hierarchical steps following Vandenberg and Lance [
45]:
Configural invariance was first examined to determine whether the STEM and Social. Science groups shared the same underlying factor structure. The evaluation process in this step determines whether the combination of fixed and free parameters yields equivalent patterns across groups, without requiring parameter equality. The configural invariance testing results demonstrate that both groups possess identical construct structures, which allows us to develop an initial model for upcoming research. The model showed acceptable fit in both groups, indicating that the factor structure remained identical across the two groups, as reported by Byrne [
41].
- (b)
Metric Invariance
Following the establishment of configural invariance, metric invariance was examined by constraining factor loadings to be equal across the two groups. The evaluation process in this stage determines whether the measured indicators show equivalent relationships with the hidden constructs, enabling researchers to study the structural connections between variables. The assessment of metric invariance involved examining changes in global fit indices rather than relying solely on chi-square difference tests, as these tests are sensitive to sample size. Consistent with contemporary methodological recommendations, invariance was supported when changes in comparative fit index (ΔCFI) did not exceed 0.01, and changes in root mean square error of approximation (ΔRMSEA) did not exceed 0.015 [
46,
47]. The research findings showed that restricting factor loadings did not cause major deterioration in model accuracy which supported metric invariance and demonstrated that the latent constructs had similar measurement characteristics between academic fields.
- (c)
Scalar Invariance
The research team evaluated scalar invariance using a set-based process. The item intercepts are set to be equal across groups. The demonstration of scalar invariance shows that observed score differences between groups stem from true differences in latent variables rather than from flaws in the testing instrument. Similarly to the metric level, invariance was evaluated based on ΔCFI ≤ 0.01 and ΔRMSEA ≤ 0.015. The analysis results confirmed partial scalar invariance, enabling researchers to compare latent means and structural parameters across groups. The observed group differences in constructs become meaningful for interpretation because of this level of invariance.
2.5.4. Structural Model Evaluation and Multi-Group Analysis
After confirming adequate measurement invariance, the structural model was estimated to examine the hypothesized relationships among cognitive processes (WM, MET, and REA), AI-SL, SDGA, ENG, and LO. The research used SEM with robust maximum likelihood estimation to address minor violations of normality in the data. Model fit was evaluated using multiple indices, including CFI, TLI, RMSEA, and SRMR, to determine whether the model met acceptable standards of adequacy.
The team examined path coefficients to determine the strength and direction of relationships among variables. This analysis allowed them to evaluate how WM, MET, REA, AI-SL, and SDGA affected student ENG and LO. They used 5000 bootstrapped resamples to assess indirect effects and created bias-corrected confidence intervals for mediation analysis. Moreover, the study compared effects across academic subjects through MG-SEM analyses of STEM and Social Science students. Sequential constraints were applied to structural paths, and changes in fit indices (ΔCFI ≤ 0.01) were tested for significant group differences. This approach enables researchers to detect effects specific to each group when examining responses to cognitive, technological, and sustainability factors.
2.5.5. Additional Diagnostics
SEM was conducted using the lavaan engine in Jamovi (version 2.6.26). Maximum likelihood (ML) estimation was used. This estimator was appropriate given the adequate sample size and distributional properties. The skewness and kurtosis values of the observed variables were within recommended thresholds. Bias-corrected bootstrapping with 5000 resamples strengthened the assessment of mediation effects by providing robust standard errors and confidence intervals for indirect effects. Additional diagnostics included multicollinearity checks using VIF. Multiple goodness-of-fit indices were also evaluated.
Table 2 summarizes the complete data analysis procedure, detailing analytical steps, statistical techniques, software, and rationales. These procedures enhanced the robustness, reliability, and replicability of the analytical results.
4. Discussion
The research data reveals essential knowledge about students’ thinking abilities and their understanding of AI-based education and sustainability, which determine their university classroom participation and academic achievement. The pooled data showed that metacognition was the leading cognitive factor in determining learner engagement (ENG), according to current research, demonstrating that students need planning and monitoring abilities to maintain deep academic task involvement [
10,
27]. Research shows that working memory (WM), together with reasoning ability (REA), strongly determines student engagement [
6,
8]. Studies indicate that students need strong cognitive abilities and analytical skills to handle demanding academic work in math-based learning settings. The research data show that students build cognitive readiness, enabling them to understand difficult academic content when solving problems and using technology-based learning tools [
52].
The research data exhibited that student participation in educational activities had the strongest direct impact on academic results, supporting modern educational concepts that stress student engagement as the primary factor connecting classroom work to academic success [
34,
35]. The mediation analysis validated this role by demonstrating that learning outcomes (LO) stem from students’ cognitive activities, AI-based educational support, and SDG knowledge, as these elements influence how students become engaged in their learning [
53]. Students who develop their thinking abilities through continuous learning, dedicated work, and persistent effort will achieve higher academic success, according to the study. Research on self-regulated learning with technology support now investigates student engagement as a surface indicator, reflecting their mental control and active participation in educational environments through AI tool use [
11,
36,
52,
54].
The implementation of AI-supported learning (AI-SL) systems brought about major changes, affecting how students engaged with their work and their unanticipated learning achievements. Research indicates that students maintain their motivation through intelligent tutoring systems, adaptive feedback, and automated explanations, which support their problem-solving and learning activities [
12,
14]. Student engagement is an indirect effect of AI-based learning because AI provides educational value through digital learning environments that actively support students during their learning activities [
54].
The program’s success depended on participants’ SDGA, as it determined their level of involvement and LO from the experience. The research results from this study expand existing knowledge about sustainability education, as students who learn about global sustainability issues show improved academic participation and a stronger natural drive and life direction [
4,
22,
53]. The research shows that sustainability competencies serve as motivational factors that enable students to achieve their academic goals while producing positive effects for society and ethics.
The research used multi-group analyses to show that academic fields produced substantial variation that affected the research results. STEM students demonstrated higher engagement because they used their reasoning skills and AI learning tools, which STEM education demands that students solve problems through technological analytical work [
24]. The Social Science students achieved better results in metacognition and SDGA because their academic program required them to evaluate their work and identify its social effects [
18,
23]. The research results demonstrate that educational design needs to be tailored to specific subjects, aligning learning requirements with AI and sustainable content that reflects how students learn in each subject area.
The research establishes theoretical value through its combination of cognitive and technological approaches and sustainability, developing a single structural model that shows institutions how to create learning spaces with AI technology, SDG principles, and inclusive design. Practically, the results highlight the importance of instructional design that considers discipline-specific needs. For STEM programmes, learning environments may benefit from strengthening reasoning-focused tasks within AI-enhanced systems, such as problem-based activities supported by adaptive feedback and intelligent tutoring. In contrast, Social Science programmes may place greater emphasis on metacognitive scaffolding and SDG-focused reflective activities, encouraging students to connect course content with social relevance and sustainability challenges. Such targeted approaches may help maximize learner engagement and learning outcomes across disciplinary contexts.
The study used a cross-sectional design, so the observed relationships represent associations rather than causal effects. The structural model located significant links among cognitive processes (WM, MET, REA), AI-SL, SDGA, ENG, and LO, yet causal claims remain unwarranted. Data collection relied on a single online, self-reported survey. Procedural steps and statistical tests reduced common method bias, but complete elimination remains unlikely. Unmeasured factors, including prior academic achievement, instructor practices, course difficulty, and institutional context, may also affect ENG and LO. Future work should adopt longitudinal or experimental designs, gather data from multiple sources, or use objective performance indicators to clarify causal pathways and lower potential method bias.
5. Conclusions
The research introduces new insights into students’ mental processes when using AI tools for learning, enabling them to acquire sustainability knowledge that leads to better academic results. The research data showed that working memory (WM), along with metacognition (MET) and reasoning ability (REA), served as fundamental factors in how students engaged (ENG) with challenging academic tasks. ENG proved to be the main factor in determining academic success, according to research data. ENG served as a link that enabled students to use their abilities, motivations, and technological experiences to achieve better academic results. The research results confirm theoretical models that identify engagement as the primary driver of learning success in self-regulated, technology-based educational systems.
The research findings show that AI-based learning systems boost student participation, leading to better academic achievement. The research indicates that AI tools create value through their dual functionality: providing individualized feedback and boosting student engagement with their coursework. SDGA positively affected student participation, leading to better LO because sustainability competencies enhance academic achievement by fostering student motivation, purposeful learning, and subject relevance.
The multi-group analysis produced key results, revealing that academic disciplines yielded unique outcomes. STEM students depended on reasoning skills and AI tools for success, while Social Science students excelled through MET and knowledge of SDGs. These results indicate that instructional strategies should account for students’ learning approaches in each discipline and tailor methods to fit specific cognitive, technological, and sustainability needs. The research develops a theoretical framework integrating cognitive, technological, and sustainability perspectives to construct a unified model that guides the design of AI-based educational environments, enhancing student engagement and academic achievement across disciplines.