You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

23 May 2025

Can Generative Artificial Intelligence Outperform Self-Instructional Learning in Computer Programming?: Impact on Motivation and Knowledge Acquisition

and
1
Escuela de Comercio, Pontificia Universidad Católica de Valparaíso, Valparaíso 2340025, Chile
2
Escuela de Ingeniería Informática, Pontificia Universidad Católica de Valparaíso, Valparaíso 2340025, Chile
*
Author to whom correspondence should be addressed.

Abstract

Generative artificial intelligence tools, such as Microsoft Copilot, are transforming the teaching of programming by providing real-time feedback and personalized assistance; however, their impact on learning, motivation, and cognitive absorption remains underexplored, particularly in university settings. This study evaluates the effectiveness of Microsoft Copilot compared to instructional videos in teaching web programming in PHP, implementing a quasi-experimental design with 71 industrial engineering students in Chile, divided into two groups: one using Microsoft Copilot and the other following instructional videos, with pre- and post-tests applied to measure knowledge acquisition while surveys based on the Hedonic-Motivation System Adoption Model (HMSAM) assessed cognitive absorption (enjoyment, control, immersion, curiosity) and technology acceptance (perceived usefulness, ease of use, and intention to adopt). The results show that, while both methods improved learning, students who used instructional videos achieved greater knowledge gains, higher levels of curiosity, and a stronger intention to continue using the technique, suggesting that instructional videos, by providing structured explanations and reducing cognitive load, may be more effective in the early stages of programming learning. In contrast, AI tools could be more beneficial in advanced stages where students require adaptive feedback, providing empirical evidence on the comparative effectiveness of AI-based and video-based instruction in teaching programming and highlighting the importance of balancing structured learning with AI-driven interactivity, with the recommendation that educators integrate both approaches to optimize the learning experience, using videos for initial instruction and AI tools for personalized support.

1. Introduction

1.1. Background

In recent years, generative artificial intelligence (AI) tools, such as Microsoft Copilot, Google Gemini, and ChatGPT, have begun to transform the educational landscape significantly [1]. These technologies assist in solving complex problems and promote a more personalized and adaptive learning experience [2]. Tools like Microsoft Copilot have proven particularly useful in programming by offering real-time code suggestions, facilitating understanding concepts, and resolving errors [3]. However, their impact on student learning and motivation has yet to be sufficiently explored, especially within university educational contexts.
One of the main challenges in adopting technological systems in education is measuring the motivation and enjoyment that users experience when interacting with these tools. The theory of cognitive absorption suggests that immersion and enjoyment during technology use can significantly influence adoption and effective learning [4]. However, measuring these emotional and cognitive dimensions remains challenging, requiring validated instruments and rigorous methodological approaches [5]. In the case of generative AI tools, it is crucial to understand how they impact the learning experience and whether they generate positive emotional effects that enhance knowledge retention.
Learning to program web systems, particularly in languages like PHP, represents a significant challenge for university students [6,7,8]. The complexity of the concepts, the need for logical skills, and the steep learning curve contribute to the difficulty [9]. Universities face the challenge of designing effective pedagogical strategies that enable students to overcome these barriers. Traditionally, instructional videos have been a common tool for teaching programming, but their effectiveness compared to more interactive methods, such as generative AI, has not yet been sufficiently studied [10].
In Latin America, research on the impact of AI tools in education is still in its early stages [11,12]. Students in this region exhibit levels of digital literacy that differ from those in Europe, Asia, and the United States, which may influence the adoption and effectiveness of these technologies [13]. Additionally, there is a significant gap in the scientific literature addressing these differences and exploring how AI tools can be adapted to specific educational contexts. This study aims to contribute to closing this gap by evaluating the impact of Microsoft Copilot on the learning and motivation of university students in the Latin American context.

1.2. Theoretical Framework

Learning programming, especially in languages like PHP for web development, presents significant challenges for university students. Among the most common difficulties are the comprehension of abstract concepts, the resolution of syntactic and logical errors, and the lack of motivation to persist in the face of complex problems [9]. These barriers impact academic performance, generating frustration and disinterest among students and hindering their learning process [14]. In response to this scenario, educational institutions have sought to implement innovative pedagogical strategies that allow these limitations to be overcome and promote more effective and motivating learning.
In recent years, active learning techniques have gained relevance in education, particularly in technical fields such as programming [15]. Instructional videos have been widely used to facilitate the understanding of complex concepts, as they allow students to learn at their own pace and review the content as many times as necessary [10]. However, these resources are often unidirectional and do not always promote active interaction between the student and the learning material.
On the other hand, generative artificial intelligence tools, such as Microsoft Copilot, represent an evolution in active learning strategies. These tools provide immediate feedback and adapt their assistance based on the individual needs of the student [3]. Generative AI fosters a more interactive and personalized approach by offering real-time code suggestions and contextualized explanations, enhancing comprehension and student motivation.
The Hedonic-Motivation System Adoption Model (HMSAM) offers a valuable theoretical framework for understanding how generative AI tools, such as Microsoft Copilot, can influence technology adoption and effective use in educational contexts. This model focuses on hedonic motivation, the pleasure or enjoyment users experience when interacting with a technology, as a key factor in its adoption [16]. According to HMSAM, cognitive absorption, which includes dimensions such as enjoyment, immersion, and curiosity, plays a crucial role in perceiving a technology’s usefulness and ease of use.
The HMSAM is based on three main constructs: perceived enjoyment, perceived control, and immersion. Perceived enjoyment refers to the extent to which the use of a system is perceived as pleasurable and fun, serving as a critical predictor in hedonic contexts [17]. Perceived control refers to the sense of autonomy and mastery that users experience when interacting with a system, which increases their confidence and willingness to adopt the technology [18]. Finally, immersion describes a psychological state in which the user feels fully absorbed by the technological experience, a key factor in systems such as virtual reality and video games [19].
In the context of programming education, each construct of the HMSAM plays a distinct role [20,21]. For example, “perceived usefulness” and “ease of use” help assess how well students perceive these tools as effective in supporting coding tasks, especially important in introductory programming, where clear scaffolding is essential. Enjoyment and curiosity are crucial for maintaining engagement during error-prone debugging or logic design activities. “Focused immersion” and “temporal dissociation” are frequently reported by students when deeply engaged in solving complex programming challenges. “Control” reflects the learner’s perceived autonomy when navigating through AI-suggested solutions, and “behavioral intention to use” is critical for understanding whether students would continue using AI tools beyond the classroom setting. These constructs were, therefore, selected not only for their theoretical grounding in HMSAM but also for their practical alignment with the motivational, emotional, and cognitive demands inherent to programming education [20].
In line with this, Chakraborty [22] noted that GenAI fosters human-machine collaboration, enabling personalized and adaptive learning while supporting experiential approaches within the HMSAM framework. The study highlights the integration of GenAI in curriculum design, teaching, and assessment, and its potential to equip students with key competencies for the future workforce. These constructs influence the intention to use and the effective adoption of hedonic systems. For example, in video games, perceived enjoyment and immersion are significant predictors of the intention for continued use [23]. In social media applications, perceived control and immersion explain the adoption of new functionalities [24].
In the context of programming education, HMSAM suggests that tools that generate a pleasurable and motivating user experience are more likely to be adopted and used effectively. For instance, if students find using Microsoft Copilot enjoyable and immersive, allowing them to become absorbed in the coding process, they are more likely to perceive the tool as useful and easy to use, which in turn may increase their intention for continued use [16]. This theoretical approach is particularly relevant to the present study, as it allows for the evaluation of the cognitive impact of generative AI tools on learning and their ability to foster states of immersion, enjoyment, and motivation. By considering both the utilitarian and hedonic aspects of the experience, HMSAM offers a comprehensive perspective for understanding how the adoption and continued use of technologies like Microsoft Copilot can positively influence academic performance and students’ willingness to engage actively and sustainably with programming.
Furthermore, the theory of cognitive absorption can be enriched by considering knowledge dynamics and knowledge fields—rational, emotional, and spiritual—as proposed by [25,26,27] in the context of learning. These approaches suggest that constructs such as enjoyment, curiosity, and immersion depend on the transformation of knowledge across domains: for instance, the rational understanding of PHP programming structures may evolve into an emotional state of satisfaction or curiosity when solving practical problems or even a spiritual connection when perceiving a broader purpose in technological learning. Within the scope of this study, such transformations may occur during active interaction with Microsoft Copilot, which encourages autonomous exploration, as well as through instructional videos, which provide structured guidance, thereby enhancing the adoption and effectiveness of both tools under the HMSAM framework.

1.3. Objectives and Research Questions

The present study aims to evaluate the capacity of a generative AI (Microsoft Copilot) compared to an instructional video to generate learning and positive emotional effects when university students practice web programming topics in PHP.
The research questions to be considered in this study are:
  • RQ1: What differences in learning exist between students who practice web programming using Microsoft Copilot versus an instructional video?
  • RQ2: What differences exist between students who practice web programming using Microsoft Copilot and those using an instructional video regarding cognitive absorption effects: enjoyment, control, focused immersion, temporal dissociation, and curiosity?
  • RQ3: What differences in effects exist between students who practice using Microsoft Copilot and those using an instructional video in the dimensions of technology acceptance: ease of use, perceived usefulness, and intention to use?

3. Experimental Design

For the present study, a quasi-experimental design was chosen, following a quantitative research methodology based on the approach used by Mellado et al. [56]. A pre-test–intervention–post-test scheme was implemented, where learning outcomes and affective variables were measured during the pre- and post-test stages, per the HMSAM model. Microsoft Copilot (Microsoft Copilot) was selected as the generative AI tool due to its free availability to students through their university-provided Office 365 accounts, ensuring accessibility, and its robust capabilities for real-time code generation and feedback, which align with the objectives of teaching PHP programming. While other AI tools, such as ChatGPT or Google Gemini (also accessible via university accounts), could offer user-friendly interfaces or education-specific features, Microsoft Copilot was preferred due to students’ prior familiarity with it from previous course activities, making it a practical choice for this context.

3.1. Participants

This study involved third-year students from an industrial engineering program at a university in Chile as participants. A total of 71 students (53 men and 18 women), aged between 19 and 21, participated in the intervention. The activity was conducted during the first semester of 2024. Participants were randomly assigned to two groups: 35 students (23 men and 12 women) used Microsoft Copilot, and 36 (27 men and 9 women) utilized instructional videos as their study medium. The randomization was performed using the random group assignment feature in the Moodle platform, which automatically allocates participants into groups randomly.
To ensure the integrity of the random assignment, the Moodle group assignment function was configured to conceal group composition from the participants, thereby guaranteeing that students could not ascertain the assignment of their peers. Each student had access solely to the information about their group (Microsoft Copilot or instructional videos) via the platform, with no possibility of viewing the resources allocated to others. The assignment was executed automatically through the Moodle randomization feature before the intervention commenced, without manual intervention, thereby minimizing the risk of bias in the allocation process. However, owing to the study design and logistical constraints, stratification strategies were not implemented, nor was the balance in additional baseline variables, such as gender, prior programming experience beyond the course modules, or digital literacy, verified. The pre-test focused exclusively on assessing knowledge of PHP, as this constituted the focus of the learning module, and all students possessed a similar exposure to the preceding Java and database modules, thereby minimizing initial variability within the context of this study. This process was completed before the intervention, ensuring each student had an equal probability of being assigned to either group. It was managed through Moodle, which facilitates access to the respective resources. Participation was voluntary; thus, this description excludes students who began but did not complete the activity.
We relied on the pre-test to assess initial PHP knowledge to control for potential pre-existing differences, as detailed in Section 4.1. An ANOVA applied to the pre-test scores confirmed no significant differences between the groups (F (1,69) = 0.451, p = 0.50), indicating that the randomization effectively balanced prior knowledge. Additional variables such as age, gender, or previous academic performance were not analyzed in this study due to its scope and resource limitations. However, the random assignment via Moodle, combined with the pre-test equivalence, supports the comparability of the groups at baseline for this quasi-experimental design.

3.2. Curriculum

The activity was conducted as part of a software systems course, which includes a module on data structures in Java (8 weeks), another on databases (4 weeks), and a final module on web design with HTML and PHP (4 weeks). While the prior modules provided a foundation in programming logic and database management, the PHP module introduced a new language and web-specific concepts, representing a shift that required students to adapt their existing knowledge. The course consists of 3 weekly lecture sessions and 2 weekly laboratory workshop sessions. The learning objectives (LOs) considered in the activity correspond to the web module and were as follows:
  • LO1: The student analyzes basic PHP elements such as blocks, variables, loops, and decision-making in algorithmic solutions to simple problems.
  • LO2: The student correctly handles arrays, superglobal variables, and their concatenation with strings in algorithmic solutions.
  • LO3: The student correctly uses the “mysqli_” functions to interact with databases.

3.3. Process

Figure 1 illustrates the overall process used, which consists of four stages: (1) explanation of objectives and modality, (2) diagnostic assessment, (3) intervention with the exercise, and (4) final evaluation. In the first stage, during the lecture session, the objectives, modality, and deadlines for the exercise activity were explained. Later in the same week, during the two workshop sessions, Stage 2 was carried out, which involved the application of diagnostic instruments to establish a baseline for comparison, both for initial knowledge and for affective perceptions about the upcoming activity (under the HMSAM model). During the same workshop sessions, Stage 3 began, which consisted of the actual exercise, dividing participants into two groups: Microsoft Copilot and an instructional video.
Figure 1. Experimental process used.
The group division and activity sequence were managed through the Moodle platform assigned to the course. Participants were given 7 days to complete the exercise guide and the final questionnaires (Stage 4), which included a final knowledge test and a perception questionnaire (HMSAM).
Although no explicit or formal training was provided immediately before the intervention, students assigned to the Microsoft Copilot group were assumed to be familiar with generative AI tools, including Microsoft Copilot specifically. This familiarity stemmed from their prior experiences within the course, where they had previously engaged informally with generative AI tools such as Microsoft Copilot, ChatGPT, and Google Gemini via their institutional Office 365 accounts. Consequently, only brief general instructions were given, allowing participants to interact freely with Microsoft Copilot during the exercises without additional detailed guidance or standardization protocols.
Specific diagnostic instruments (pre- and post-tests) were used to evaluate learning, such as the example presented in Figure 2, which was drawn from a repository of questions and selected randomly. The example in Figure 2, extracted from the repository of questions used in the pre-test and post-test, focuses on identifying and correcting errors in a PHP code that interacts with a MySQL database. The objective of the code is to process name and surname variables sent through a form, insert them into a table called “personas”, and then display the results in an HTML table. However, the code contains several errors that must be corrected to function properly. These errors include incorrect use of SQL syntax in the INSERT statement, lack of data validation, and potential security issues such as SQL injection. The PHP instructions must also be reviewed to ensure the data is handled and displayed correctly in the browser. This exercise allowed for evaluating both the students’ technical knowledge and their ability to apply programming concepts in practical contexts.
Figure 2. Example of the type of question used in pre and post-test.
Both groups used the same PHP practice guide, including a series of exercises with code snippets. Each exercise presents 3 to 4 non-exclusive alternatives containing statements about possible outcomes or consequences of the given code, specific lines of code, or potential replacements for certain lines of code. It is important to emphasize that this exercise guide is a compilation of various exercises used in previous semesters and captures the most frequent errors made by students who, despite having learned the material, begin coding in PHP.
Figure 3 presents an exercise from the PHP practice guide used in Stage 3 by the Microsoft Copilot group, where a PHP block surrounds HTML code. Specifically, Line 5 with $objetos = 2; is outside the PHP block, meaning it will not be preprocessed and will be displayed as plain text. The figure has been enlarged to enhance the legibility of the code and accompanying text.
Figure 3. Exercise 1 of the guide used.
Likewise, Figure 4 presents a PHP code snippet that utilizes the $POST superglobal variable to construct a database query in the $sql variable by concatenating strings with variables. Additionally, it uses Mysqli_ functions.
Figure 4. Exercise 5 of the guide used.
The control group used a video (see Figure 5) specifically created to review each exercise in the guide. The video explained why the different alternatives presented were either correct or incorrect.
Figure 5. Explanatory video of the exercises is in the guide.
The experimental group used the same exercises but with the instruction to query Microsoft Copilot. Additionally, the exercises for this group included follow-up questions to guide students in evaluating the correctness of the AI’s responses. For instance, in Exercise 1 (see Figure 3), as a follow-up, Question 1.1 asks whether the AI considered the variable $objetos outside a PHP block. If not, the students were instructed to correct the AI.
To scaffold students’ critical engagement with Microsoft Copilot, the exercise guide used in this group included follow-up questions accompanying specific tasks (see Figure 3). Given the context of each exercise, these questions prompted students to evaluate whether the AI’s suggestions were syntactically and logically appropriate. For instance, one prompt asked: “Did the assistant consider that the variable $objetos is outside the PHP block?”. Students were encouraged to reformulate their queries or adjust the proposed code when discrepancies were detected. These prompts aimed to foster metacognitive awareness, reduce overreliance on AI suggestions, and support the iterative refinement of solutions.

3.4. Instruments

For the initial and final knowledge tests, exercises similar to those in the guide were used, following the format of code snippets with four non-exclusive alternatives. The pre-test and post-test included six exercises (two associated with each Learning Objective, LO), randomly selected from a pool of 18 exercises (distinct from the guide), with scores ranging from 0.0 to 10.0. These exercises were developed from materials used in prior semesters of the software systems course and aligned with the PHP module’s learning objectives. Content validity was ensured through review by two instructors with over five years of PHP teaching experience. The pre- and post-test exercises were extracted from a broader item bank applied in the software systems course for over four academic years (seven semesters), primarily as practice and assessment tools aligned with the course’s PHP module learning objectives. These items were developed by instructors with extensive experience in programming instruction and have undergone iterative refinement based on student performance and instructional feedback. While detailed item analysis was not conducted, the sustained application of these items across cohorts supports their empirical reliability and content validity. Post-hoc analysis of the post-test scores showed a Cronbach’s alpha of 0.82, indicating good reliability. Figure 6 shows an example of a question used in the pre-test.
Figure 6. Sample question in the knowledge pre-test.
It is important to note that both experimental conditions—Microsoft Copilot and instructional video—worked with the same practice guide and were assessed using pre- and post-tests drawn from a shared item bank. This uniformity ensures that familiarity with item formats or content is applied equally across both groups, minimizing the risk of biased learning gains due to prior exposure.
On the other hand, for the initial and final perception tests, the HMSAM model was used with its eight dimensions: usefulness, enjoyment, ease of use, intention to use, control, focused immersion, temporal dissociation, and curiosity [16]. The total number of statements for both tests was 24 items, measured on a Likert scale from 1 to 7, where 1 corresponded to “Strongly Disagree”, 4 to “Neutral”, and 7 to “Strongly Agree”. The list of statements by construct is shown in Table 1.
Table 1. List with questions (statements) by construct of the HMSAM model (for pre-test).

4. Results

The statistical software SPSS 29 was used to analyze the results. Analysis of variance (ANOVA) and analysis of covariance (ANCOVA) tests were considered to measure the differences in initial and final learning outcomes. To calculate the size of the effects found, partial eta squared (η2) was used, with values of 0.01 indicating a small effect, 0.06 a medium effect, and 0.14 or greater a large effect [84].

4.1. Learning Effects

Table 2 presents the descriptive statistics for the knowledge tests administered at the beginning and end of the intervention, separated by experimental condition (video vs. Microsoft Copilot).
Table 2. Descriptive statistics of pre-test and post-test per condition.
For assessing normality Python pre-test scores a Shapiro–Wilk test was performed for the Copilot (W = 0.97, p = 0.36) and the Video (W = 0.98, p = 0.54) groups, showing non-significant differences from normality. On post-test scores, Shapiro–Wilk test provided (W = 0.95, p = 0.13) for the Copilot and (W = 0.95, p = 0.11) for the video conditions, indicative for normality. A Levene’s test showed homogeneity of variances between conditions for pre-test (F = 0.38, p = 0.54) and post-test (F = 3.18, p = 0.08) scores.
When an ANCOVA test was applied to the post-test scores, using the pre-test as a covariate, to measure possible differences between groups, significant differences were found between the two groups. The group that used videos showed a higher post-test score than the group that used Microsoft Copilot, with F (1,68) = 32.621, p < 0.001, η2 = 0.32 (see Figure 7).
Figure 7. Estimated marginal means for post-test among conditions (video vs. Microsoft Copilot).
Additionally, an analysis of variance (ANOVA) test was applied to the pre-test scores by the group to verify whether there were differences between the groups, yielding F (1,69) = 0.451, p = 0.50, and no significant differences were detected in the prior knowledge levels of the subjects in both groups. Similarly, an ANOVA was applied to the post-test scores by group, resulting in F (1,69) = 31.920, p < 0.001, η2 = 0.32, indicating that the video group had a significantly higher post-test score than the Microsoft Copilot group.

4.2. HMSAM Effects

Table 3 details the descriptive statistics for the pre- and post-intervention measures of the eight dimensions of the HMSAM model, broken down by condition. The constructs considered were usefulness, enjoyment, ease of use, intention to use, control, focused immersion, temporal dissociation, and curiosity. A Cronbach’s alpha of 0.94 for the pre-IMI and 0.93 for the post-IMI perception tests was obtained.
Table 3. Descriptive statistics of HMSAM pre-test and post-test perceptions per condition.
On post-perception values, homogeneity of variances between groups was assessed by the Levene’s test for each of the eight constructs: Utility (F = 2.69, p = 0.11), Enjoyment (F = 2.84, p = 0.70), Ease of Use (F = 0.71, p = 0.40), BIU (F = 1.60, p = 0.21), Control (F = 0.96, p = 0.33), FI (F = 0.82, p = 0.37), TD (F = 2.25, p = 0.14), and Curiosity (F = 0.02, p = 0.90).
On pre-post-perception values, the Levene’s test showed equality of variances on Utility (F = 1.84, p = 0.18), Enjoyment (F = 0.24, p = 0.63), Ease of Use (F = 1.34, p = 0.25), BIU (F = 0.10, p = 0.75), Control (F = 0.74, p = 0.39), FI (F = 1.54, p = 0.22), TD (F = 3.57, p = 0.06), and Curiosity (F = 0.14, p = 0.71).
A two-way ANOVA was conducted considering the post-intervention measures, with the condition (video/Microsoft Copilot) and the HMSAM dimensions as factors on the scores obtained. The result was F (1,69) = 3.820, p < 0.055, η2 = 0.05, indicating a marginally significant interaction. After performing pairwise comparisons with Bonferroni adjustment, it was found that only the construct of curiosity showed a significant difference in favor of the video condition, with F (1,69) = 4.630, p < 0.035, η2 = 0.06.
This was followed by the dimensions of usefulness (F (1,69) = 3.301, p < 0.074, η2 = 0.05), enjoyment (F (1,69) = 3.716, p < 0.058, η2 = 0.05), and intention to use (F (1,69) = 3.536, p < 0.064, η2 = 0.05), all showing differences in favor of the video condition, though not reaching conventional levels of significance. The remaining dimensions also showed differences favoring the video condition but were not statistically significant, as shown in Figure 8.
Figure 8. Post-perception scores of HMSAM constructs per condition.
An ANOVA was also conducted with the condition and HMSAM dimensions as factors, but this time on the score differences (post-test—pre-test), yielding F (1,69) = 2.912, p < 0.092, η2 = 0.04. Pairwise comparisons with Bonferroni adjustment revealed significant differences for the dimensions of usefulness (F (1,69) = 5.135, p < 0.027, η2 = 0.07), intention to use (F (1,69) = 5.867, p < 0.018, η2 = 0.08), and curiosity (F (1,69) = 4.437, p < 0.039, η2 = 0.06), all favoring the video condition over the Microsoft Copilot condition (see Figure 9). It is important to note that, although eight separate ANOVA tests were conducted across the HMSAM constructs, no correction for family-wise error rate (FWER), such as Bonferroni or False Discovery Rate (FDR), was applied. This constitutes a limitation of the current analysis. However, the observed effects were consistent in direction and supported by medium effect sizes in key dimensions (e.g., η2 = 0.06 for curiosity, η2 = 0.08 for intention to use), suggesting that the findings are not solely due to random variation. Future studies should implement appropriate p-value correction techniques to ensure more conservative statistical inference and control for inflated Type I error risk when testing multiple constructs.
Figure 9. Pre-post-perception differences of HMSAM constructs per condition.

5. Discussion

5.1. Learning Effects

The findings confirm that, although both methods promote improvements in student performance, those who followed the video-based instruction route achieved significantly greater progress. Several factors can explain this phenomenon.
From the cognitive load theory perspective [46,85], learning complex content (such as web programming logic) benefits from resources that optimize the distribution of mental load. In video-based instruction, students can process information visually and audibly simultaneously, facilitating the formation of integrated mental models [44,47]. This advantage is amplified by the fact that the instructional videos used in this study were specifically designed to teach PHP programming, with structured explanations and examples tailored to address common challenges in understanding complex concepts and algorithms. In contrast, while Microsoft Copilot provides immediate and adaptive feedback [86,87], it is a general-purpose GenAI tool not explicitly optimized for pedagogical purposes, requiring students to formulate questions and validate responses independently. Additionally, the ability to pause and rewind allows learners to control the pace of learning and focus on the most complex elements [28,48], further enhancing the effectiveness of these purposefully crafted audiovisual resources. This adaptive control over the learning pace may help reduce extraneous cognitive load, allowing students to understand better the code’s logic and structures [46].
Although Microsoft Copilot provides immediate and adaptive feedback [86,87], students must formulate questions and validate the relevance of the generated responses. This process requires a higher level of metacognition and digital competencies to evaluate the quality of the feedback [38,42]. In contrast, videos present carefully sequenced examples and explanations, reducing uncertainty about appropriate practices. In this way, students perceive constant reinforcement that increases their perceived competence [39] and motivates them to continue exploring without fear of initial failure [33].
The inclusion of follow-up questions in the Copilot group (Figure 3) sought to support this evaluative process. These prompts encouraged learners to critically analyze the AI’s output and identify potential inaccuracies. However, the impact of these questions likely varied depending on the student’s level of engagement. Some participants may have utilized them effectively to guide iterative improvement, while others may have skipped them or provided superficial answers, resulting in inconsistent benefits across the group.
In addition to these cognitive factors, it is important to consider that the quality and structure of the instructional content may have differed significantly between groups. The video provided a consistent, pedagogically sequenced explanation for each exercise, ensuring uniform content delivery across all participants. In contrast, the AI group relied on individually formulated prompts, resulting in variability in the relevance, depth, and accuracy of the responses compared to Microsoft Copilot. This inconsistent content delivery may have contributed to the observed differences in learning outcomes. Future studies should control instructional structure across conditions to ensure more comparable and reliable evaluations of effectiveness.
While Levene’s test indicated homogeneity of variances between groups, the observed differences in post-test standard deviations (2.32 vs. 1.23) may suggest potential variance heterogeneity. However, given the substantial difference in post-test means, the coefficient of variation (CV) offers a more appropriate metric of relative dispersion, revealing that variability was not disproportionate when adjusted for group means. Moreover, ANCOVA is generally robust to moderate violations of the homogeneity of variances assumption, particularly in balanced designs [88,89]. Nonetheless, we acknowledge this as a potential limitation. As a future line of research, simulation studies [90] could be implemented to evaluate how deviations from normality or homoscedasticity might influence the robustness of ANCOVA results in educational contexts involving generative AI.
The Hedonic-Motivation System Adoption Model (HMSAM) establishes that the sense of enjoyment and perceived usefulness are key determinants in the adoption and effectiveness of learning systems [16]. On the other hand, audiovisual resources (with narrative, practical examples, and demonstrations) tend to generate greater affective and cognitive engagement by stimulating attention, curiosity, and interest [44,91]. In line with this, previous studies have shown that video-based learning can foster deeper emotional involvement, facilitating concept retention [29,31]. In contrast, interaction with generative artificial intelligence may be less engaging and require self-regulation strategies that not all students have developed, especially at the early stages of programming [30].
From an instructional perspective, Microsoft Copilot promotes a more active learning approach regarding exploration and constantly testing hypotheses within the code [82,92,93]. However, this constructivist approach may generate uncertainty when students lack a robust foundation in programming syntax and logic. On the other hand, instructional videos adopt an expository approach, where the teacher’s guidance is explicit, and students assimilate problem-solving strategies more directly [28]. This difference becomes particularly noticeable in the early stages of learning when familiarity with basic concepts is essential to avoid cognitive overload [47,85].
Thus, despite the clear inclination of the results toward the effectiveness of videos, the findings do not diminish the value of generative artificial intelligence as a reinforcement tool or for advanced tutoring [80,81,94]. In particular, a mixed instructional strategy that combines the systematic and exemplified presentation of a video with the personalized feedback of Microsoft Copilot could maximize learning by providing a solid initial conceptual framework, followed by guided and immediate experimentation [29,33]. Future research could explore the optimal integration of both methods based on student profiles, such as their level of experience, learning styles, and intrinsic motivations.
The greater effectiveness of video in improving web programming learning can be explained by the reduced extraneous cognitive load, which facilitates sequential information reception through multiple modes; sustained attention and motivation derived from audiovisual resources; and the immediate and accessible support that reduces uncertainty for novice students. With this, we can answer the research question: Are there differences in learning between students who practice web programming using Microsoft Copilot versus instructional videos? Affirmatively, learning differences favor video-based instruction over generative artificial intelligence-based practice.
The profile of the participants (third-year industrial engineering students rather than computer science students) may further explain the observed learning effects. Unlike computer science students, who typically have extensive prior exposure to programming, industrial engineering students in this study had limited experience, primarily from prior course modules in Java and databases. This relative novelty of programming, particularly in PHP, likely heightened the importance of affective factors such as curiosity, enjoyment, and immersion, as these students relied heavily on structured guidance (e.g., videos) to build confidence and competence. Consequently, the greater effectiveness of instructional videos may reflect their ability to meet the needs of learners with less programming expertise, suggesting that the comparative advantage of videos over Microsoft Copilot could vary with students who have stronger technical foundations.
An additional consideration that may limit the generalizability of our findings is the exclusive focus on PHP. While PHP was chosen for its curricular relevance and role in web development, different programming languages pose distinct syntactic, semantic, and conceptual challenges. For example, languages such as Python or JavaScript offer various levels of abstraction and readability, which could influence how learners interact with generative AI tools or benefit from structured instructional materials. Thus, the effectiveness observed in this study may not directly translate to other programming contexts. Future research should replicate this design using diverse programming languages to assess whether the comparative impact of instructional videos and AI tools remains consistent.
These conclusions open the door to the strategic use of both modalities, combining their strengths while mitigating their limitations. Thus, this integrated approach optimizes teaching and learning processes in web programming, databases, and Java. This integrated approach holds significant potential for enhancing educational outcomes and addressing the diverse needs of learners at different stages of their programming journey.
Nonetheless, it is important to underscore that the present findings are based solely on immediate post-intervention assessments. As such, they reflect short-term knowledge acquisition rather than long-term learning or retention. Without longitudinal follow-up data, we cannot determine whether the observed advantages of instructional videos persist over time. Nevertheless, it is important to note that the scope and design of this study were deliberately aligned with short-term, well-defined learning objectives. As established in instructional design theory and supported by Bloom’s taxonomy, limited interventions can be appropriate and effective when the targeted learning outcomes are specific, foundational, and skill-oriented [95,96,97]. Therefore, while long-term retention merits future study, the present design remains methodologically sound for its intended scope. Future studies should include delayed post-tests or follow-up assessments to examine the durability and transferability of these learning outcomes.
Although the pre-test confirmed equivalence in baseline PHP knowledge between groups, it is important to note that this instrument did not directly assess broader aspects of digital literacy or prior experience with generative AI tools. While not measured, these factors could have contributed to the variability in how students interacted with Microsoft Copilot. For example, two students with similar programming knowledge may differ significantly in their ability to formulate prompts, interpret responses, or detect inaccuracies in AI outputs due to differences in their digital fluency. This represents a potential source of uncontrolled variance in the AI group. We recommend that future research incorporate explicit instruments to assess digital literacy and prior AI exposure, potentially including them as covariates or segmentation variables in experimental designs.
While the present study included a structured exercise guide with follow-up prompts designed to scaffold critical engagement with Microsoft Copilot, it did not incorporate a system for recording the number, content, or quality of student interactions with the tool. This limits the ability to assess the consistency or depth of engagement across participants, as actual usage behaviors (e.g., number of prompts submitted, adherence to correction instructions, or time on task) were not tracked. Although all participants in the AI condition received the same guided activities, designed to foster metacognitive reflection and verification of Copilot’s responses, there remains a gap between the intended instructional design and the unobserved execution of that design. Future studies should integrate interaction logging or usage analytics to evaluate the relationship between engagement patterns and learning outcomes more precisely.
Additionally, we did not record the number of queries or interactions each student had with Microsoft Copilot, which limits our ability to evaluate engagement consistency across participants. The absence of data on the quantity, quality, or nature of prompts (e.g., specificity, complexity, or frequency) precludes a detailed analysis of how student engagement with Microsoft Copilot influenced learning outcomes. Variability in prompt formulation and interaction patterns likely contributed to differences in the tool’s effectiveness, as its performance heavily depends on the user’s ability to craft effective prompts and critically assess AI-generated responses. Future studies should incorporate usage tracking or interaction logs to quantify engagement levels and explore their correlation with learning gains, thereby offering deeper insights into the role of AI tool proficiency in programming education.
Another critical factor potentially influencing the learning outcomes was the absence of explicit and standardized instructions for students interacting with Microsoft Copilot. While previous informal exposure to Microsoft Copilot and other generative artificial intelligence tools provided students with practical knowledge and competence in querying and interpreting AI outputs, individual variability in proficiency and approaches likely emerged. This variability might have influenced the consistency and effectiveness of interactions during the experimental activity. Future studies should explicitly standardize training sessions and develop structured interaction protocols for generative AI tools, thereby ensuring greater methodological rigor, consistency in usage, and improved comparability of results across experimental conditions.

5.2. HMSAM Effects

5.2.1. Cognitive Absorption

The results of the present study focused on evaluating differences in the effects of cognitive absorption (enjoyment, control, focused immersion, temporal dissociation, and curiosity) between students who used Microsoft Copilot (Microsoft Copilot) and instructional videos to practice web programming, revealed that curiosity showed a significant difference in favor of the group that used videos over those who used generative artificial intelligence under the Microsoft Copilot model. This finding aligns with prior research that highlights the role of structured multimedia resources in stimulating intrinsic interest [35,36,47].
Most likely, curiosity, as the foundation of exploratory learning, was favored by instructional videos due to their ability to reduce extraneous cognitive load [85]. By integrating visual and auditory elements sequentially, videos facilitate the formation of coherent mental schemas, allowing students to focus on the logic of the code without informational overload [44]. This structured instructional design focuses attention and generates a sense of progressive competence, which is key to sparking curiosity [32]. For example, pausing and reviewing complex segments empowers students to explore concepts at their own pace, fostering self-directed curiosity [28].
In contrast, interaction with a generative artificial intelligence like Microsoft Copilot, while offering immediate feedback, may introduce beginner uncertainty by requiring the formulation of precise questions and the critical validation of responses generated by artificial models—processes that demand metacognitive skills still under development [38]. This dynamic could inhibit curiosity by being perceived as an obstacle to autonomous exploration.
Regarding the other constructs, such as enjoyment, control, focused immersion, and temporal dissociation, no significant differences were found between the groups. However, relevant trends were observed. In terms of enjoyment, videos scored slightly higher, suggesting that audiovisual storytelling might generate greater satisfaction by engaging emotional stimuli [91]. In terms of control, both methods showed similar levels, indicating that the flexibility of videos (pausing, rewinding) and the interactivity of Microsoft Copilot address different needs for autonomy. Although not statistically significant, focused immersion and temporal dissociation reflected that both approaches require comparable sustained attention, a critical aspect in programming environments [31].
These findings have significant pedagogical implications. Instructional videos emerge as effective tools for fostering curiosity in the early stages of learning, where clarity and structure are prioritized. However, Microsoft Copilot could be integrated into advanced stages, where students, already familiar with basic concepts, require personalized feedback for complex problems [42].
In response to the research question: Are there differences between students who practice web programming using Microsoft Copilot and those using instructional videos regarding cognitive absorption effects: enjoyment, control, focused immersion, temporal dissociation, and curiosity? We can state that only curiosity significantly differed in favor of the group that used videos. This result suggests that videos better foster intrinsic interest by structuring content clearly and through multisensory means. In contrast, no significant differences were found in the other constructs. This indicates that while structured multimedia resources enhance curiosity, different dimensions of cognitive absorption may depend on contextual factors or individual preferences.
Future research should explore hybrid models that combine both methodologies, adapting to student profiles and experience levels. It should also assess the long-term impact of curiosity on the retention of technical skills. Such investigations could provide deeper insights into optimizing learning strategies in web programming and related fields, ensuring that instructional approaches align with cognitive and motivational needs.

5.2.2. Dimensions of Technological Acceptance

The analysis of score differences (post-test—pre-test) revealed significant differences in perceived usefulness and intention to use, all favoring the group that used instructional videos compared to those that employed generative artificial intelligence. These findings suggest that videos are more effective in fostering curiosity, as previously discussed, and generate a higher perception of usefulness and a stronger intention to use the tool in the future.
First, perceived usefulness, defined as the extent to which students consider a tool to enhance their learning, was significantly higher in the video group. This aligns with the Technology Acceptance Model (TAM) [39], which posits that perceived usefulness is a key predictor of adopting educational tools. Videos provide structured and sequential explanations of programming concepts, allowing students to visualize how the content relates to their learning objectives and reinforcing their perception of usefulness [47].
On the other hand, while offering immediate feedback, Microsoft Copilot may generate uncertainty among students by requiring them to formulate precise questions and critically evaluate the generated responses. This process, which demands advanced metacognitive skills, could dilute their perception of usefulness, especially in the early stages of learning [38].
In the second place, intention to use, which reflects students’ willingness to continue using a tool in the future, also showed significant differences in favor of videos. This result aligns with technology acceptance and usage studies that highlight the importance of ease of use and perceived usefulness in adopting technologies [28,41]. Being more intuitive and less demanding regarding metacognitive skills, videos may generate a more satisfying learning experience, increasing the intention to use them [28]. In contrast, interaction with Microsoft Copilot, while innovative and novel, may be perceived as more complex and less accessible for students at early programming levels, potentially reducing their intention to use it [42].
In response to research question RQ3: Are there differences in effects between students who practice using Microsoft Copilot and those using instructional videos regarding technology acceptance dimensions: ease of use, perceived usefulness, and intention to use? There are significant differences in perceived usefulness and intention to use, both favoring the group that used videos. These results suggest that videos, by offering a more structured and accessible learning experience, generate a higher perception of usefulness and a stronger intention to use them in the future. In contrast, despite providing immediate feedback, Microsoft Copilot may not be perceived as equally useful or easy to use in early learning stages, limiting its adoption.
These findings underscore the importance of designing educational tools that balance innovation and accessibility, adapting to students’ needs and experience levels. By addressing these factors, educators and developers can create solutions that maximize engagement and learning outcomes, ensuring that technological advancements are enablers rather than barriers to effective education.

6. Conclusions

This study evaluated the impact of using a generative artificial intelligence (Microsoft Copilot) compared to video-based instruction on university students’ learning of web programming in PHP. Through a pre-test–intervention–post-test experimental design, the effects on learning and affective perceptions of 71 participants were analyzed. Participants were randomly divided into two groups: one that used Microsoft Copilot and another that followed instructional videos. The Hedonic-Motivation System Adoption Model (HMSAM) was employed to assess perceptions.
The results revealed that students who received instruction through videos made greater progress in the post-test knowledge assessment than those who used Microsoft Copilot. This finding suggests that the sequential and multimodal structuring of information in videos, specifically designed to teach complex programming concepts and algorithms, facilitates the assimilation of concepts by reducing extraneous cognitive load. In contrast, Microsoft Copilot, as a general-purpose GenAI tool, lacks the tailored pedagogical focus of the videos, which may explain its relatively lower effectiveness for novice learners in this context.
These conclusions highlight the potential of instructional videos as a powerful tool for introductory programming education, particularly in contexts where clarity, structure, and reduced cognitive load are critical for student success, such as with industrial engineering students who may lack extensive programming experience. At the same time, they suggest that while generative AI tools like Microsoft Copilot offer innovative possibilities, their adoption may require careful scaffolding and integration, especially for novice learners. Notably, although the participants (third-year industrial engineering students) had prior programming experience from courses in Java and databases, their foundations might not have been sufficient to leverage Microsoft Copilot in this context fully. Learning PHP, a new language for them, alongside the autonomous use of Microsoft Copilot without specific training, likely demanded advanced metacognitive skills (e.g., crafting effective prompts and critically evaluating AI responses) that were not yet fully developed. This suggests that Microsoft Copilot could be more beneficial for students in advanced stages or with a computer science background, where a stronger, language-specific conceptual foundation and familiarity with AI interaction enable them to maximize their potential for real-time problem-solving. Future research should investigate hybrid approaches that integrate the strengths of both methods to enhance learning outcomes across various stages of programming education and diverse learner profiles.
This study’s main contribution provides empirical evidence on the differential impact of generative AI tools and traditional instructional methods on programming learning. These findings underscore the importance of considering cognitive demands and students’ affective perceptions when designing technology-based pedagogical strategies. Additionally, this study contributes relevant knowledge to the Latin American context, where research on the adoption of educational technologies is still scarce, and factors such as digital literacy and resource availability may influence the effectiveness of technological tools. While our results are rooted in the context of industrial engineering students learning PHP, they may hold broader implications. Due to their structured guidance and reduced cognitive load, the preference for instructional videos could generalize to other introductory programming courses (e.g., Python, JavaScript) or even non-programming domains (e.g., mathematics or engineering design) where novices benefit from clear, sequential instruction. Similarly, the potential of generative AI tools like Microsoft Copilot for advanced learners might extend to contexts requiring adaptive feedback. However, this would depend on learners’ prior knowledge, the complexity of the subject matter, and the specific instructional design. However, such generalizations require caution, as differences in course objectives, disciplinary conventions, and student backgrounds could alter the observed effects.
Nevertheless, this research has some limitations. First, the sample was limited to students from a single university in Chile, which restricts the generalization of the findings to other educational and cultural contexts. Second, the intervention was a single-session activity conducted over seven days, with no follow-up assessments beyond the immediate post-test. This short duration means that the study primarily captures immediate learning outcomes rather than long-term retention or the ability to apply PHP skills in diverse contexts, potentially limiting the external validity of the results. The lack of follow-up assessments prevents us from determining whether the observed advantages of instructional videos or the potential of Microsoft Copilot persist over time. This is particularly relevant for programming education, where sustained practice is key to mastery. However, the study focused on comparing the immediate effectiveness of the two methods in a controlled setting, and the pre- and post-test design provides a valid measure of short-term learning gains within this scope. Finally, the study focused exclusively on learning PHP, so it would be relevant to examine whether the results are replicated in other programming languages and levels of complexity or entirely different study domains.
Another significant limitation is the lack of control over potential confounding variables, such as digital literacy, prior programming experience beyond the course modules, or gender, due to the absence of stratification or balance verification in the random assignment. Although the pre-test confirmed the equivalence in PHP knowledge between the groups, these unmeasured variables could have influenced the interaction with Microsoft Copilot, particularly considering that digital literacy may affect the ability to formulate prompts and validate AI responses. This limitation reflects the scope of the study, which prioritized evaluating the effects on learning and affective perceptions rather than technology adoption factors, which are often the primary focus in technology acceptance studies. Future research should include specific instruments to measure these variables and consider them as covariates or stratification criteria to enhance the robustness of the experimental design.
Future studies should address these limitations by expanding the sample to include diverse educational settings, extending the duration of interventions to evaluate long-term effects, and exploring the applicability of the findings across various programming languages (e.g., Python, C++) and non-programming disciplines (e.g., physics, statistics). Such efforts will help determine the extent to which the comparative advantages of instructional videos and generative AI tools can be generalized, providing a more comprehensive understanding of their role in technology-enhanced education across diverse contexts.
Future studies should address these limitations by expanding the sample to include diverse educational settings, extending the duration of interventions to evaluate long-term effects, and exploring the applicability of the findings across different programming languages and learner profiles. Such efforts will contribute to a more comprehensive understanding of how emerging technologies can be effectively integrated into programming education while addressing the unique needs of students in various contexts.
From a practical perspective, these findings provide valuable insights for curriculum designers and computer science educators. The results suggest that video-based instruction is an effective strategy for introductory programming teaching. At the same time, generative AI tools like Microsoft Copilot may be more useful in advanced stages when students already have solid conceptual foundations and require immediate feedback for problem-solving. At the technological level, these findings can guide the development of hybrid platforms that combine the pedagogical structure of videos with the adaptability of generative AI to optimize the learning experience.
As future lines of research, longitudinal studies are recommended to evaluate the long-term effects of these tools on learning. Such studies could include follow-up assessments at multiple intervals (e.g., one month, three months) to examine retention and skill application, addressing the limitation of the current short-term focus. Additionally, it would be valuable to explore combined strategies that integrate video-based instruction with generative AI, analyzing their effectiveness across different experience levels and learning profiles. Finally, expanding the sample to include diverse institutions and countries will help validate the generalization of these findings in varied educational contexts, strengthening the understanding of the impact of technology on education in Latin America. However, it is important to note that these conclusions are based on short-term post-intervention assessments. Future research should include delayed post-tests or follow-up studies to evaluate long-term retention and the sustainability of these instructional effects.

Author Contributions

Conceptualization, R.M. and C.C.; methodology, R.M. and C.C.; software, R.M.; validation, R.M. and C.C.; formal analysis, C.C.; investigation, R.M. and C.C.; resources, R.M. and C.C.; data curation, C.C.; writing—original draft preparation, R.M.; writing—review and editing, C.C.; visualization, R.M. and C.C.; supervision, C.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is funded by the Regular Fondecyt Project of the National Research and Development Agency of the State of Chile: Project No. 1252211, whose responsible researcher is Dr. Claudio Cubillos.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and the protocol was approved by the Bioethics and Biosecurity Committee of Pontificia Universidad Católica de Valparaíso (protocol code BIOEPUCV-H 659-2023 on 29 June 2023).

Data Availability Statement

Data are available in the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Bahroun, Z.; Anane, C.; Ahmed, V.; Zacca, A. Transforming Education: A Comprehensive Review of Generative Artificial Intelligence in Educational Settings through Bibliometric and Content Analysis. Sustainability 2023, 15, 12983. [Google Scholar] [CrossRef]
  2. Luckin, R.; Holmes, W. Intelligence Unleashed: An Argument for AI in Education; UCL Knowledge Lab: London, UK, 2016; pp. 849–851. [Google Scholar]
  3. Jackson, V.; Vasilescu, B.; Russo, D.; Ralph, P.; Izadi, M.; Prikladnicki, R.; D’angelo, S.; Inman, S.; Andrade, A.; van der Hoek, A. The Impact of Generative AI on Creativity in Software Development: A Research Agenda. ACM Trans. Softw. Eng. Methodol. 2024; just accepted. [Google Scholar] [CrossRef]
  4. Agarwal, R.; Karahanna, E. Time Flies When You’re Having Fun: Cognitive Absorption and Beliefs about Information Technology Usage. MIS Q. 2000, 24, 665–694. [Google Scholar] [CrossRef]
  5. Saadé, R.; Bahli, B. The Impact of Cognitive Absorption on Perceived Usefulness and Perceived Ease of Use in On-Line Learning: An Extension of the Technology Acceptance Model. Inf. Manag. 2005, 42, 317–327. [Google Scholar] [CrossRef]
  6. Park, T.H.; Wiedenbeck, S. Learning Web Development: Challenges at an Earlier Stage of Computing Education. In Proceedings of the ICER’11: ACM SIGCSE 2011 International Computing Education Research Workshop, Providence, RI, USA, 8–9 August 2011; pp. 125–132. [Google Scholar] [CrossRef]
  7. Ling, H.C.; Chiang, H. Sen Learning Performance in Adaptive Learning Systems: A Case Study of Web Programming Learning Recommendations. Front. Psychol. 2022, 13, 770637. [Google Scholar] [CrossRef] [PubMed]
  8. Liu, I.F.; Hung, H.C.; Liang, C.T. A Study of Programming Learning Perceptions and Effectiveness under a Blended Learning Model with Live Streaming: Comparisons between Full-Time and Working Students. Interact. Learn. Environ. 2024, 32, 4396–4410. [Google Scholar] [CrossRef]
  9. Robins, A.; Rountree, J.; Rountree, N. Learning and Teaching Programming: A Review and Discussion. Comput. Sci. Educ. 2003, 21, 137–172. [Google Scholar] [CrossRef]
  10. Guo, P.J.; Kim, J.; Rubin, R. How Video Production Affects Student Engagement: An Empirical Study of MOOC Videos. In Proceedings of the L@S 2014—1st ACM Conference on Learning at Scale, Atlanta, GA, USA, 4–5 March 2014; pp. 41–50. [Google Scholar] [CrossRef]
  11. Guerrero-Quiñonez, A.J.; Bedoya-Flores, M.C.; Mosquera-Quiñonez, E.F.; Mesías-Simisterra, Á.E.; Bautista-Sánchez, J.V. Artificial Intelligence and Its Scope in Latin American Higher Education. Ibero-Am. J. Educ. Soc. Res. 2023, 3, 264–271. [Google Scholar] [CrossRef]
  12. De La Torre, A.; Baldeon-Calisto, M. Generative Artificial Intelligence in Latin American Higher Education: A Systematic Literature Review. In Proceedings of the 12th International Symposium on Digital Forensics and Security, ISDFS, San Antonio, TX, USA, 29–30 April 2024. [Google Scholar] [CrossRef]
  13. Hinostroza, J.E.; Isaacs, S.; Bougroum, M. Information and Communications Technologies for Improving Learning Opportunities and Outcomes in Developing Countries. In Learning and Education in Developing Countries; Palgrave Pivot: New York, NY, USA, 2014; pp. 42–57. [Google Scholar] [CrossRef]
  14. Lahtinen, E.; Ala-Mutka, K.; Järvinen, H.M. A Study of the Difficulties of Novice Programmers. ACM SIGCSE Bull. 2005, 37, 14–18. [Google Scholar] [CrossRef]
  15. Gosavi, C.S.; Arora, S. Active Learning Strategies for Engaging Students in Higher Education. J. Eng. Educ. Transform. 2022, 36, 1–7. [Google Scholar] [CrossRef]
  16. Lowry, P.B.; Gaskin, J.; Twyman, N.; Hammer, B.; Roberts, T. Taking “Fun and Games” Seriously: Proposing the Hedonic-Motivation System Adoption Model (HMSAM). J. Assoc. Inf. Syst. 2013, 14, 2. [Google Scholar] [CrossRef]
  17. Davis, F.D.; Bagozzi, R.P.; Warshaw, P.R. Extrinsic and Intrinsic Motivation to Use Computers in the Workplace1. J. Appl. Soc. Psychol. 1992, 22, 1111–1132. [Google Scholar] [CrossRef]
  18. Venkatesh, V.; Thong, J.Y.L.; Xu, X. Consumer Acceptance and Use of Information Technology: Extending the Unified Theory of Acceptance and Use of Technology. MIS Q. 2012, 36, 157–178. [Google Scholar] [CrossRef]
  19. Jennett, C.; Cox, A.L.; Cairns, P.; Dhoparee, S.; Epps, A.; Tijs, T.; Walton, A. Measuring and Defining the Experience of Immersion in Games. Int. J. Hum. Comput. Stud. 2008, 66, 641–661. [Google Scholar] [CrossRef]
  20. Ma, J.Y.; Xie, J.F.; Chen, C.C. Exploring the Structural Relationships of Microinteractions in Perception and Behavior by the Hedonic Motivation System Adoption Model. Int. J. Hum. Comput. Interact. 2025, 41, 592–607. [Google Scholar] [CrossRef]
  21. Qomarul Huda, M.; Aeni Hidayah, N.; Nur Hafizah Hersyaf, T.; Sujoko, I. Asmawi Analysis of Continuance Use of Video on Demand Applications by Using the Hedonic Motivation System Adoption Model. In Proceedings of the 2020 8th International Conference on Cyber and IT Service Management, CITSM, Pangkal, Indonesia, 23–24 October 2020. [Google Scholar] [CrossRef]
  22. Chakraborty, S. Generative AI in Modern Education Society. arXiv 2024, arXiv:2412.08666. [Google Scholar]
  23. Hamari, J.; Koivisto, J. Why Do People Use Gamification Services? Int. J. Inf. Manag. 2015, 35, 419–431. [Google Scholar] [CrossRef]
  24. Zhang, X.; Guo, X.; Lai, K.H.; Guo, F.; Li, C. Understanding Gender Differences in M-Health Adoption: A Modified Theory of Reasoned Action Model. Telemed. J. E Health 2014, 20, 39–46. [Google Scholar] [CrossRef]
  25. Qadhi, S.; Qadhi, S. Knowledge Dynamics: Educational Pathways from Theories to Tangible Outcomes. In From Theory of Knowledge Management to Practice; IntechOpen: London, UK, 2023. [Google Scholar] [CrossRef]
  26. Bratianu, C.; Garcia-Perez, A. Knowledge Dynamics and Expert Knowledge Translation: A Case Study. In Proceedings of the 24th European Conference on Knowledge Management 2023, Lisbon, Portugal, 7–8 September 2023; Volume 24, pp. 140–147. [Google Scholar] [CrossRef]
  27. Bratianu, C.; Bejinaru, R. Knowledge Dynamics: A Thermodynamics Approach. Kybernetes 2020, 49, 6–21. [Google Scholar] [CrossRef]
  28. Bishop, J.L.; Verleger, M.A. The Flipped Classroom: A Survey of the Research. In Proceedings of the ASEE Annual Conference and Exposition, Atlanta, GA, USA, 23–26 June 2013. [Google Scholar] [CrossRef]
  29. Sung, Y.T.; Chang, K.E.; Liu, T.C. The Effects of Integrating Mobile Devices with Teaching and Learning on Students’ Learning Performance: A Meta-Analysis and Research Synthesis. Comput. Educ. 2016, 94, 252–275. [Google Scholar] [CrossRef]
  30. Bernard, R.M.; Abrami, P.C.; Borokhovski, E.; Wade, C.A.; Tamim, R.M.; Surkes, M.A.; Bethel, E.C. A Meta-Analysis of Three Types of Interaction Treatments in Distance Education. Rev. Educ. Res. 2009, 79, 1243–1289. [Google Scholar] [CrossRef]
  31. Gordillo, A.; Lopez-Fernandez, D.; Tovar, E. Comparing the Effectiveness of Video-Based Learning and Game-Based Learning Using Teacher-Authored Video Games for Online Software Engineering Education. IEEE Trans. Educ. 2022, 65, 524–532. [Google Scholar] [CrossRef]
  32. Deci, E.L.; Ryan, R.M. Intrinsic Motivation and Self-Determination in Human Behavior. In Intrinsic Motivation and Self-Determination in Human Behavior; Springer: New York, NY, USA, 1985. [Google Scholar] [CrossRef]
  33. Jena, A.K.; Bhattacharjee, S.; Gupta, S.; Das, J.; Debnath, R. Exploring the Effects of Web 2.0 Technology on Individual and Collaborative Learning Performance in Relation to Self-Regulation of Learners. J. Sch. Educ. Technol. 2018, 13, 20–35. [Google Scholar]
  34. Huang, J.; Mizumoto, A. Examining the Effect of Generative AI on Students’ Motivation and Writing Self-Efficacy. Digit. Appl. Linguist. 2024, 1, 102324. [Google Scholar] [CrossRef]
  35. Litman, J.A. Curiosity and the Pleasures of Learning: Wanting and Liking New Information. Cogn. Emot. 2005, 19, 793–814. [Google Scholar] [CrossRef]
  36. Dousay, T.A. Multimedia Design and Situational Interest: A Look at Juxtaposition and Measurement. In Educational Media and Technology Yearbook; Springer: Cham, Switzerland, 2014; pp. 69–82. [Google Scholar] [CrossRef]
  37. Krouska, A.; Mylonas, P.; Kabassi, K.; Caro, J.; Sgouropoulou, C.; Hmoud, M.; Swaity, H.; Hamad, N.; Karram, O.; Daher, W. Higher Education Students’ Task Motivation in the Generative Artificial Intelligence Context: The Case of ChatGPT. Information 2024, 15, 33. [Google Scholar] [CrossRef]
  38. Abdelshiheed, M.; Maniktala, M.; Barnes, T.; Chi, M. Assessing Competency Using Metacognition and Motivation: The Role of Time-Awareness in Preparation for Future Learning. arXiv 2023, arXiv:2303.14609. [Google Scholar]
  39. Davis, F.D. Perceived Usefulness, Perceived Ease of Use, and User Acceptance of Information Technology. MIS Q. 1989, 13, 319–339. [Google Scholar] [CrossRef]
  40. Ghimire, A.; Edwards, J. Generative AI Adoption in Classroom in Context of Technology Acceptance Model (TAM) and the Innovation Diffusion Theory (IDT). arXiv 2024, arXiv:2406.15360. [Google Scholar]
  41. Venkatesh, V.; Morris, M.G.; Davis, G.B.; Davis, F.D. User Acceptance of Information Technology: Toward a Unified View. MIS Q. 2003, 27, 425–478. [Google Scholar] [CrossRef]
  42. Steinert, S.; Avila, K.E.; Ruzika, S.; Kuhn, J.; Küchemann, S. Harnessing Large Language Models to Enhance Self-Regulated Learning via Formative Feedback. arXiv 2023, arXiv:2311.13984. [Google Scholar]
  43. Lin, Z.; Ng, Y.L. Unraveling Gratifications, Concerns, and Acceptance of Generative Artificial Intelligence. Int. J. Hum. Comput. Interact. 2024, 1–18. [Google Scholar] [CrossRef]
  44. Clark, R.C.; Mayer, R.E. E-Learning and the Science of Instruction: Proven Guidelines for Consumers and Designers of Multimedia Learning; John Wiley & Sons: Hoboken, NJ, USA, 2016. [Google Scholar]
  45. Al-Abdullatif, A.M. Modeling Teachers’ Acceptance of Generative Artificial Intelligence Use in Higher Education: The Role of AI Literacy, Intelligent TPACK, and Perceived Trust. Educ. Sci. 2024, 14, 1209. [Google Scholar] [CrossRef]
  46. Paas, F.G.W.C.; Van Merriënboer, J.J.G. Instructional Control of Cognitive Load in the Training of Complex Cognitive Tasks. Educ. Psychol. Rev. 1994, 6, 351–371. [Google Scholar] [CrossRef]
  47. Mayer, R.E. Multimedia Learning. In Multimedia Learning, 2nd ed.; Cambridge University Press: Cambridge, UK, 2009; pp. 1–304. [Google Scholar] [CrossRef]
  48. Martins, D.S.; Cunha, B.C.R.; Yaguinuma, C.A.; Zaine, I.; Pimentel, M. da G.C. Effects of Interactive Video Annotations on Students’ Browsing Behavior and Perceived Workload. ACM SIGAPP Appl. Comput. Rev. 2019, 19, 44–57. [Google Scholar] [CrossRef]
  49. Bennedsen, J.; Caspersen, M.E. Failure Rates in Introductory Programming. ACM SIGCSE Bull. 2007, 39, 32–36. [Google Scholar] [CrossRef]
  50. Suzuki, R.; Kato, J.; Yatani, K. ClassCode: An Interactive Teaching and Learning Environment for Programming Education in Classrooms. arXiv 2020, arXiv:2001.08194. [Google Scholar]
  51. Yang, C.; Thung, F.; Lo, D. Efficient Search of Live-Coding Screencasts from Online Videos. In Proceedings of the 2022 IEEE International Conference on Software Analysis, Evolution and Reengineering, SANER, Honolulu, HI, USA, 15–18 March 2022; pp. 73–77. [Google Scholar] [CrossRef]
  52. Khandwala, K.; Guo, P.J. Codemotion: Expanding the Design Space of Learner Interactions with Computer Programming Tutorial Videos. In Proceedings of the 5th Annual ACM Conference on Learning at Scale, New York, NY, USA, 26–28 June 2018. [Google Scholar] [CrossRef]
  53. Li, W.; Pea, R.; Haber, N.; Subramonyam, H. Tutorly: Turning Programming Videos Into Apprenticeship Learning Environments with LLMs. arXiv 2024, arXiv:2405.12946. [Google Scholar]
  54. Buffardi, K.; Wang, R. Integrating Videos with Programming Practice. In Proceedings of the Annual Conference on Innovation and Technology in Computer Science Education, ITiCSE, Dublin, Ireland, 8–13 July 2022; Volume 1, pp. 241–247. [Google Scholar] [CrossRef]
  55. McGowan, A.; Anderson, N.; Trombino, G.; Sage, P.; Adhikari, J.; Galway, L. Learning to Code—Investigating the Adoption of Video-Based Lab Solutions for University-Level Novice Programmers. In Proceedings of the ICERI2023 Proceedings, Seville, Spain, 13–15 November 2023; IATED: Valencia, Spain, 2023; p. 1136. [Google Scholar]
  56. Mellado, R.; Cubillos, C. Gamification Improves Learning: Experience in a Training Activity of Computer Programming in Higher Education. J. Comput. Assist. Learn. 2024, 40, 1959–1973. [Google Scholar] [CrossRef]
  57. Ferreira, D.J.; da Silva, H.C.; Melo, T.F.N.; Ambrósio, A.P. Investigation of Continuous Assessment of Correctness in Introductory Programming. Educ. Technol. Soc. 2017, 20, 182–194. [Google Scholar]
  58. Kadar, R.; Mahlan, S.B.; Shamsuddin, M.; Othman, J.; Wahab, N.A. Analysis of Factors Contributing to the Difficulties in Learning Computer Programming among Non-Computer Science Students. In Proceedings of the 2022 12th IEEE Symposium on Computer Applications and Industrial Electronics, ISCAIE, Penang, Malaysia, 21–22 May 2022; pp. 89–94. [Google Scholar] [CrossRef]
  59. Esche, S.; Weihe, K. Choosing a Didactic Basis for an Instructional Video: What Are the Implications for Novice Programmers? In Proceedings of the Annual Conference on Innovation and Technology in Computer Science Education, ITiCSE, Turku, Finland, 7–12 July 2023; Volume 1, pp. 450–456. [Google Scholar] [CrossRef]
  60. Nguyen-Duc, A.; Cabrero-Daniel, B.; Przybylek, A.; Arora, C.; Khanna, D.; Herda, T.; Rafiq, U.; Melegati, J.; Guerra, E.; Kemell, K.-K.; et al. Generative Artificial Intelligence for Software Engineering—A Research Agenda. arXiv 2023, arXiv:2310.18648. [Google Scholar]
  61. Ebert, C.; Louridas, P. Generative AI for Software Practitioners. IEEE Softw. 2023, 40, 30–38. [Google Scholar] [CrossRef]
  62. Bilgram, V.; Laarmann, F. Accelerating Innovation With Generative AI: AI-Augmented Digital Prototyping and Innovation Methods. IEEE Eng. Manag. Rev. 2023, 51, 18–25. [Google Scholar] [CrossRef]
  63. Ulfsnes, R.; Moe, N.B.; Stray, V.; Skarpen, M. Transforming Software Development with Generative AI: Empirical Insights on Collaboration and Workflow. In Generative AI for Effective Software Development; Springer: Cham, Switzerland, 2024; pp. 219–234. [Google Scholar] [CrossRef]
  64. Calegario, F.; Burégio, V.; Erivaldo, F.; Moraes, D.; Andrade, C.; Felix, K.; Barbosa, N.; Lucas Da, P.; Lucena, S.; França, C. Exploring the Intersection of Generative AI and Software Development. arXiv 2023, arXiv:2312.14262. [Google Scholar]
  65. Nettur, S.B.; Karpurapu, S.; Nettur, U.; Gajja, L.S. Cypress Copilot: Development of an AI Assistant for Boosting Productivity and Transforming Web Application Testing. IEEE Access 2024, 13, 3215–3229. [Google Scholar] [CrossRef]
  66. Mahadevappa, P.; Muzammal, S.M.; Tayyab, M.; Mahadevappa, P.; Muzammal, S.M.; Tayyab, M. Introduction to Generative AI in Web Engineering: Concepts and Applications; IGI Global: Hershey, PA, USA, 2025; pp. 297–330. [Google Scholar] [CrossRef]
  67. Ho, C.L.; Liu, X.Y.; Qiu, Y.W.; Yang, S.Y. Research on Innovative Applications and Impacts of Using Generative AI for User Interface Design in Programming Courses. In Proceedings of the International Conference on Information Technology, Data Science, and Optimization, Taipei, Taiwan, 22–24 May 2024; pp. 68–72. [Google Scholar] [CrossRef]
  68. Jayachandran, D.; Maldikar, P.; Love, T.S.; Blum, J.J. Leveraging Generative Artificial Intelligence to Broaden Participation in Computer Science. In Proceedings of the 2024 AAAI Spring Symposium Series, Stanford, CA, USA, 25–27 March 2024; Volume 3, pp. 486–492. [Google Scholar] [CrossRef]
  69. Keuning, H.; Alpizar-Chacon, I.; Lykourentzou, I.; Beehler, L.; Köppe, C.; de Jong, I.; Sosnovsky, S. Students’ Perceptions and Use of Generative AI Tools for Programming Across Different Computing Courses. In Proceedings of the 24th Koli Calling International Conference on Computing Education Research, Koli, Finland, 12–17 November 2024. [Google Scholar]
  70. Yilmaz, R.; Karaoglan Yilmaz, F.G. The Effect of Generative Artificial Intelligence (AI)-Based Tool Use on Students’ Computational Thinking Skills, Programming Self-Efficacy and Motivation. Comput. Educ. Artif. Intell. 2023, 4, 100147. [Google Scholar] [CrossRef]
  71. Wilson, S.E.; Nishimoto, M. Assessing Learning of Computer Programing Skills in the Age of Generative Artificial Intelligence. J. Biomech. Eng. 2024, 146, 051003. [Google Scholar] [CrossRef]
  72. Shanshan, S.; Sen, G. Empowering Learners with AI-Generated Content for Programming Learning and Computational Thinking: The Lens of Extended Effective Use Theory. J. Comput. Assist. Learn. 2024, 40, 1941–1958. [Google Scholar] [CrossRef]
  73. Sajja, A.; Thakur, D.; Mehra, A.; Sajja, A.; Thakur, D.; Mehra, A. Integrating Generative AI into the Software Development Lifecycle: Impacts on Code Quality and Maintenance. Int. J. Sci. Res. Arch. 2024, 13, 1952–1960. [Google Scholar] [CrossRef]
  74. Yehia, E. Developments on Generative AI. In AI and Emerging Technologies: Automated Decision-Making, Digital Forensics, and Ethical Considerations; CRC Press: Boca Raton, FL, USA, 2024; pp. 139–160. [Google Scholar] [CrossRef]
  75. Liu, J.; Li, S. Toward Artificial Intelligence-Human Paired Programming: A Review of the Educational Applications and Research on Artificial Intelligence Code-Generation Tools. J. Educ. Comput. Res. 2024, 65, 1165–1195. [Google Scholar] [CrossRef]
  76. Boguslawski, S.; Deer, R.; Dawson, M.G. Programming Education and Learner Motivation in the Age of Generative AI: Student and Educator Perspectives. Inf. Learn. Sci. 2024; ahead-of-print. [Google Scholar] [CrossRef]
  77. Cubillos, C.; Mellado, R.; Cabrera-Paniagua, D.; Urra, E. Generative Artificial Intelligence in Computer Programming: Does It Enhance Learning, Motivation, and the Learning Environment? IEEE Access 2025, 13, 40438–40455. [Google Scholar] [CrossRef]
  78. Groothuijsen, S.; van den Beemt, A.; Remmers, J.C.; van Meeuwen, L.W. AI Chatbots in Programming Education: Students’ Use in a Scientific Computing Course and Consequences for Learning. Comput. Educ. Artif. Intell. 2024, 7, 100290. [Google Scholar] [CrossRef]
  79. Pesovski, I.; Santos, R.; Henriques, R.; Trajkovik, V. Generative AI for Customizable Learning Experiences. Sustainability 2024, 16, 3034. [Google Scholar] [CrossRef]
  80. Frankford, E.; Sauerwein, C.; Bassner, P.; Krusche, S.; Breu, R. AI-Tutoring in Software Engineering Education Experiences with Large Language Models in Programming Assessments. In Proceedings of the 46th International Conference on Software Engineering, Lisbon, Portugal, 14–20 April 2024; pp. 309–319. [Google Scholar] [CrossRef]
  81. Li, H.; Xu, T.; Zhang, C.; Chen, E.; Liang, J.; Fan, X.; Li, H.; Tang, J.; Wen, Q. Bringing Generative AI to Adaptive Learning in Education. arXiv 2024, arXiv:2402.14601. [Google Scholar]
  82. Maphoto, K.B.; Sevnarayan, K.; Mohale, N.E.; Suliman, Z.; Ntsopi, T.J.; Mokoena, D. Advancing Students’ Academic Excellence in Distance Education: Exploring the Potential of Generative AI Integration to Improve Academic Writing Skills. Open Prax. 2024, 16, 142–159. [Google Scholar] [CrossRef]
  83. Ko, S.; Chan, S.C.H.; Ko, S.; Chan, S.C.H. A Framework for the Responsible Integration of Generative AI Tools in Learning; IGI Global: Hershey, PA, USA, 2024; pp. 163–194. [Google Scholar] [CrossRef]
  84. Lakens, D. Calculating and Reporting Effect Sizes to Facilitate Cumulative Science: A Practical Primer for t-Tests and ANOVAs. Front. Psychol. 2013, 4, 62627. [Google Scholar] [CrossRef] [PubMed]
  85. Sweller, J. Cognitive Load during Problem Solving: Effects on Learning. Cogn. Sci. 1988, 12, 257–285. [Google Scholar] [CrossRef]
  86. Du, X.; Liu, M.; Wang, K.; Wang, H.; Liu, J.; Chen, Y.; Feng, J.; Sha, C.; Peng, X.; Lou, Y. Evaluating Large Language Models in Class-Level Code Generation. In Proceedings of the 46th International Conference on Software Engineering, Lisbon, Portugal, 14–20 April 2024; pp. 982–994. [Google Scholar] [CrossRef]
  87. Vaithilingam, P.; Zhang, T.; Glassman, E.L. Expectation vs. Experience: Evaluating the Usability of Code Generation Tools Powered by Large Language Models. In Proceedings of the CHI Conference on Human Factors in Computing Systems, New Orleans, LA, USA, 29 April 2022. [Google Scholar] [CrossRef]
  88. Rheinheimer, D.C.; Penfield, D.A. The Effects of Type I Error Rate and Power of the ANCOVA F-Test and Selected Alternatives under Non-Normality and Variance Heterogeneity. J. Exp. Educ. 1998, 69, 373–391. [Google Scholar] [CrossRef]
  89. Ch, K.P.; Poremba, K.D.; Rowell, R.K. Testing for Homogeneity of Slopes in Analysis of Covariance: A Tutorial; ERIC Institute of Education Science: Washington, DC, USA, 1997. [Google Scholar]
  90. Ateş, C.; Kaymaz, Ö.; Tekindal, M.A.; Erdoğan, B.D. Robustness of Analysis of Covariance (Ancova) under the Distributions Assumptions and Variance Homogeneity. Eurasian J. Vet. Sci. 2020, 36, 58–65. [Google Scholar] [CrossRef]
  91. Huang, Y.M.; Chiu, P.S. The Effectiveness of a Meaningful Learning-Based Evaluation Model for Context-Aware Mobile Learning. Br. J. Educ. Technol. 2015, 46, 437–447. [Google Scholar] [CrossRef]
  92. Kharrufa, A.; Johnson, I. The Potential and Implications of Generative AI on HCI Education. In Proceedings of the 6th Annual Symposium on HCI Education, New York, NY, USA, 5–7 June 2024. [Google Scholar] [CrossRef]
  93. Hou, I.; Mettille, S.; Li, Z.; Man, O.; Zastudil, C.; MacNeil, S. The Effects of Generative AI on Computing Students’ Help-Seeking Preferences. In Proceedings of the 26th Australian Computing Education Conference, Sydney, Australia, 29 January 2024; pp. 39–48. [Google Scholar] [CrossRef]
  94. Chen, E.; Lee, J.E.; Lin, J.; Koedinger, K. GPTutor: Great Personalized Tutor with Large Language Models for Personalized Learning Content Generation. In Proceedings of the L@S 2024—11th ACM Conference on Learning @ Scale, Atlanta, GA, USA, 18–20 July 2024; pp. 539–541. [Google Scholar] [CrossRef]
  95. Guskey, T.R. Closing Achievement Gaps: Revisiting Benjamin S. Bloom’s “Learning for Mastery”. J. Adv. Acad. 2007, 19, 8–31. [Google Scholar] [CrossRef]
  96. Conklin, J. Review of A Taxonomy for Learning, Teaching, and Assessing: A Revision of Bloom’s Taxonomy of Educational Objectives Complete Edition. Educ. Horiz. 2005, 83, 154–159. [Google Scholar]
  97. Bloom, B.S. Learning for Mastery. Instruction and Curriculum. Regional Education Laboratory for the Carolinas and Virginia, Topical Papers and Reprints, Number 1. Eval. Comment 1968, 1, n2. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.