Applying Self-Optimised Feedback to a Learning Management System for Facilitating Personalised Learning Activities on Massive Open Online Courses

: Web-based educational systems collect tremendous amounts of electronic data, ranging from simple histories of students’ interactions with the system to detailed traces of their reasoning. However, less attention has been given to the pedagogical interaction data of customised learning in a gamiﬁcation environment. This study aims to research user experience, communication methods, and feedback and survey functionalities among the existing learning management system; understand the data ﬂow of the online study feedback loop; propose the design of a personalised feedback system; and conclude with a discussion of the ﬁndings from the collected experiment data. We will test the importance of gamiﬁcation learning and analytics learning, as well as conduct a literature review and laboratory experiment to examine how they can inﬂuence the effectiveness of monetary incentive schemes. This research will explain the signiﬁcance of gamiﬁcation learning and personalised feedback on motivation and performance during the learning journey.


Introduction
Amidst the relentless advancement of information and technology and the evergrowing prominence of digitalisation, the relevance of e-learning systems with personalised feedback in students' educational journeys has intensified. The concept of "Personalised learning refers to instruction in which the pace of learning and the instructional approach are optimised for the needs of each learner [1]. Learning objectives, instructional approaches, and instructional content may all vary based on learner needs". Leveraging the potential of data analysis and statistics, combined with powerful computing resources, facilitates real-time multimodal data-processing capabilities, benefiting various fields, including education. Within the domain of academia, there is a heated debate surrounding the topic of feedback, which plays a pivotal role in the learning process. Dissatisfaction among students with the feedback they receive on their work is prevalent, with timeliness and relevance being key issues [2]. In this regard, an adeptly designed system grounded in the principles of human-computer interaction (HCI) proves superior to conventional methodologies. Such a system can identify user behaviours effectively, enabling immediate and accurate responses, thus supporting learners with diverse backgrounds to achieve incremental milestones based on their strengths and motivation. The system also empowers learners to exercise greater control over the learning process and adaptability, thereby enabling them to construct personalised learning paths and attain their individual educational objectives. This paper introduces a personalised feedback strategy focused on bolstering student learning within the Learning Management System (LMS). The strategy incorporates automated feedback and surveys to provide learners with valuable insights into their weaknesses and expected learning experiences in the MOOC. The primary emphasis is on delivering adaptive learning resources to learners in manageable phases, while also modelling the learners and learning context within a well-structured environment.
Aim1: To comprehensively understand the generic feedback process through extensive research, including the design, pipeline, structure, and content of feedback messages, among other aspects. Aim2: To extract pertinent information from website resources based on learners' demands and individual backgrounds, facilitating the design of modules and a planner system to optimise the learning process accordingly. Aim3: To design a system capable of recording platform communications, including interaction logs and quiz attempts, thereby enhancing the system's capacity for automated feedback in future iterations.

Background
The backdrop against which these objectives are pursued consists of the ever-evolving landscape of information and technology, coupled with the pervasive influence of digitalisation.
As personalised feedback is empowered by data analysis and statistical techniques, it harnesses the potential of real-time multimodal data processing to yield valuable insights for stakeholders, including students, lecturers, and moderators. Enhanced synchronisation between stakeholders allows for more frequent interactions, leading to a profound impact on students' learning progress, as influenced by their unique backgrounds and feedback.
Within the domain of e-learning, the process of attaining personalisation in feedback describes how learners assimilate performance-relevant information to optimise their educational progress. The significance of feedback has been duly acknowledged by academic institutions, wherein, for instance, the students' satisfaction rate functions as a real-time measure of their learning experience, exhibiting considerable potential in identifying patterns and challenges inherent in the educational process.
To foster an enhanced educational journey, a well-structured system aligned with HCI is proposed to identify user behaviours effectively. This approach facilitates prompt and precise responses, catering to students with diverse backgrounds, and enriching their learning experience by offering controlled learning processes. The system empowers users to construct personalised study plans, thereby attaining their individual learning objectives through the e-learning platform.

Learning Analytics
The paper refers to a study on the affective responses of students concerning personalised feedback [3]. The findings indicate the successful application of learning analytics in expanding personalised feedback. In addition, the influence of student perceptions and emotions in relation to their engagement with feedback is considered.

Effect of Gamification
During their initial foray into university studies, freshmen undergo a period of adjustment to meet professional demands. Research demonstrates that gamified approaches, replete with practical tips and tricks, considerably ease their academic transition, thereby fostering a seamless beginning to their educational journey [4]. Nevertheless, extant literature highlights a prevalent issue of low retention rates and limited learner engagement, especially in Massive Online Open Courses (MOOCs) [5]. To address this, gamification has been advocated as an efficacious strategy to augment user engagement, particularly in the domain of online education [6]. The implementation of gaming activities on interactive displays fosters cognitive stimulation and imaginative exploration among students, en-hancing engagement through rewards and collaborative endeavours [7]. However, recent studies posit that completion rates alone may not serve as a definitive measure of MOOCs' success [8], considering the varied goals and motivations of participants [9]. Therefore, integrating gamification into MOOCs empowers users to develop personalised plans, thereby facilitating the accomplishment of their individual objectives throughout the course [10]. This integration further intensifies user engagement within the gamified environment. Customised learner modes, rooted in learners' motivations and engagement profiles, are proposed for incorporation into the system [11]. Notably, a systematic literature review conducted by Alessandra, Roland, and Marcus highlights the nascent nature of gamification and its application in online learning, particularly MOOCs. The paucity of empirical experiments and evidence is noted, with gamification largely employed as a mechanism for external rewards [10]. Anna Faust from Technology University Berlin explores the impact of gamification as a non-monetary incentive on motivation and performance [12]. Three studies, including a literature review and laboratory experiment, evaluate the influence of gamification on motivation and performance in the context of task complexity and monetary incentive schemes.

Learning Process
First, the system must identify the learner's status; if learners are using the system for the first time, the original course material list (non-personalisation list) is provided to them based on their query terms. After learners visit some course materials and respond to the assigned questionnaires, the proposed system re-evaluates their abilities, adjusts the difficulty parameters of the selected course material, and recommends appropriate course materials.

Learning Analytics Visualisation
Dispositional learning analytics is implemented to evaluate the characteristics and provide constructive suggestions to students and teachers. The data are extracted from student self-report, instructor perception, and conduct follows.
Riiid Labs, an AI arrangements supplier delivering innovative disruption to the training market, enables global schooling players to reconsider conventional methods of learning utilizing AI [13]. In 2017, with a strong belief in equal opportunities in education, Riiid launched the AI mentor based on deep learning analytics for over 100 million students in South Korea [13]. The innovative algorithm will assist in addressing global challenges in education. If successful, it is conceivable that any student with an Internet connection can participate in the benefits of a personalised learning experience.
Drawing on Clark and Brennan's influential Common Ground theory, a Debate Dashboard augmented with a CCSAV tool and a set of widgets that deliver meta-information about participants and the interaction process has been designed [14]. An empirical study simulating a moderately sized collective deliberation scenario provides evidence that this experimental version outperformed the control version on a range of indicators, including usability, mutual understanding, quality of perceived collaboration, and accuracy of individual decisions [14]. Using representations that highlight conceptual relationships between contributions will promote more critical thinking and evidence-based reasoning.
"The goal of KnowledgeTree is to bridge the gap between the currently popular approach to Web-based education, which is centred on learning management systems vs. the powerful but underused technologies in intelligent tutoring and adaptive hypermedia [15]." From the literature sources such as web pages and ontology libraries to the middleware layer of web crawlers (literature information) and bidirectional LSTM-CRF training model (meteorological simulation) for knowledge acquisition, the final knowledge graph construction will be built up through knowledge fusion.
Meanwhile, Tailored Recruitment Analytics and Curriculum Knowledge (TRACK) is another vital information-/examination-fueled undertaking that is building student-and staff-facing tools to assist in illuminating these and many other questions) [16]. Uniting individuals from the Connected Intelligence Center (CIC), Student Services Unit (SSU), Information Technology Division (ITD), and the Institute for Interactive Media and Learning (IML), handling live data for these tools is another curriculum. The Application Programming Interface (API) service will also make it workable for individuals to extract curriculum data for use in different apparatuses.

Learning Feedback
One of the challenges associated with creating a personalised e-learning system is to achieve viable personalisation functionalities, such as customised content administration, a student model, a student plan, and versatile moment collaboration. The architecture for the Personalised Virtual Learning Environment was proposed by Dongming Xu from City University of Hong Kong [17], utilizing intelligent decision-making agents with autonomous capabilities, along with pre-active and proactive behaviours, to enhance elearning effectiveness. Additionally, in the study of "Stimulated Planning Game Element Embedded(SPGEE) in a MOOC Platform" [18], Alessandra et al. explored the usability of Stimulated Planning through usability tests, eye tracking, and retrospective think-aloud methods. The user feedback was then applied to assess the alignment between the design choices and the conceptual planning process of the users. The progress tracking step, which links learning activities and progress to the plan, provides immediate feedback to the students, informing them about their intentions and plan-related progress. This enables learners to initiate re-planning at the earliest opportunity.
On the other hand, the OnTask project developed a software tool that collects and assesses data on students' activities, allowing instructors to design personalised feedback. By providing frequent suggestions on specific tasks in the course, students can quickly adapt and improve their learning progressively [19]. At the University of South Australia, using UniSA LMS and course dashboard analytics, the OnTask platform increased student engagement with the course and their studies. Similarly, at UNSW, with Moodle LMS, TMGrouper, MHCampus, and custom learning analytics reporting/dashboard, OnTask enhanced communication through personalised feedback and fostered students' sense of connectedness to the course. At the University of Adelaide, OnTask was utilised by the course coordinator to encourage early student engagement with the course [20]. In addition, OnTask has been applied for personalised, learning analytics-based feedback. The diagram of assessed minor methods for fruition rate shows that for the benchmark group, there was a decline in Connect action in the semester, followed by an increase in activity near the end of the semester. The pattern for the treatment group was reversed, with lower rate fruition rates at the beginning, and then a sharp increase during the semester (corresponding to Feedback Point 1), which was sustained over the remainder of the semester ( Table 1).
The outcomes from direct relapse analysis demonstrated that both the program subsection score and the OnTask group were significant indicators of the final course performance. The findings highlight the recurring nature of SRL: based on the external standards provided by LA-based feedback, students assess and adjust their actions to enhance their performance, leading to new targets for evaluation in subsequent learning iterations. Students responded effectively to personalised, learning analysis-based feedback. Using a combination of deductive coding and ENA, they found that students' perceptions of their feedback were generally positive in terms of quality, and although there was a range of emotional responses, most of them were associated with improved motivation for learning. The model below depicts the various messages students receive based on test scores. They are written in the academics' own voice but are integrated and automatically sent to the relevant students (Table 2).

Learning Result Analysis
Student learning results can be derived from summative and formative assessments. A summative assessment includes the evaluation of a midterm examination, academic paper, and final project. A formative assessment entails the analysis of gamification elements, social interaction, and attitudinal surveys. The research will primarily focus on assessing each student's actual learning level based on evidence of their knowledge and abilities. Hence, ongoing feedback will be collected through a five-point Likert scale questionnaire and social network interaction.
To investigate the impact of personalised gamification, the University of Tehran categorised gamification elements according to related motivation types, leading to increased average participation with a personalised approach [23]. Another study compared the effects of gamification and social networking on e-learning. Learning modules were designed with corresponding social network components and participation on the social networking site was based on interactions and contributions to the system [24]. The results demonstrated that participants in social networking achieved higher academic success in terms of knowledge acquisition. However, further study is required to thoroughly discuss the synthesis of social network feedback and its impact. Furthermore, the personalised system reduced the number of errors made by learners. Those in the adaptive situation exhibited a greater reduction in errors during the second session compared to learners using the non-adaptive situation [25].

Condition
Feedback Message

(no attempt)
The Accounting A Team have noticed that you are yet to attempt the learning quizzes from last week on the topic of 'Recording Accounting Transactions'.
• Why Complete? To understand the topic next week, you must have a strong understanding of this topic (it forms the foundation of your studies in accounting so PLEASE PRACTISE!) • Resources to Help:the tutorial guide contains tips and tricks for the homework questions as well as self study questions with fully worked solutions. Also watch the screencast made by one our academics on UTS Online-it will walk you through the steps necessary to record a transaction Assistance Outside Class: • Attend U:PASS Sessions: Student mentors are happy to answer your questions and provide you with additional learning material • Post all subject related questions on the discussion board

(low performance: less than 50%)
The Accounting A Team and Iwould like to commended on your efforts to complete last weeks' quizzes on 'Recording Accounting Transactions', but would like to encourage you to continue practising this topic.
• Step 1: retake the quiz-it is important you understand this topic!! • Step 2: view the screencast made by one our academics on UTS Online-it will walk you through the steps necessary to record a transaction • Step 3: post all your questions on the discussion board and refer to the tutorial guide for assistance with homework questions and additional self-study questions

(average performance: 50-74%)
Congratulations on making the effort to complete the learning quiz last week on the topic of 'Recording Accounting Transactions'. To improve your understanding: • Retake:the quiz-they can be done as many times as you like!! o Why? next week's lecture builds upon this knowledge so it's important you can get 100% on this topic • View: the screencast made by one our academics on UTS Online-it will walk you through the steps necessary to record a transaction • Practise!!: using the self-study questions contained in the tutorial guide

Gaps of Existing Research
In the body of literature, there exists an array of untracked, offline "embedded tasks" [20] that have yet to be accounted for. Previous studies have not thoroughly examined the intricate details available from the LMS logs to measure the extent to which any observed changes endure, spanning across multiple weeks or more extended durations. Consideration of a partner investigation, rather than a randomised controlled trial, would be advantageous in examining the impact of the intercession.
Moreover, communication methods [19] currently applied in text format, such as Email and messages in LMS, could be expanded to include video call data, voice call data, and other multimedia information during the trend of remote learning. This would enhance learning analytics, measure student engagement, and provide accurate personalised feedback.
OnTask [22] can play a crucial role in delivering personalised feedback on a student's work quality and quantity, along with recommending specific curriculum content. However, integration sources, such as video watching, quiz completion, practical attendance, and assignment viewing, remain limited in individual university online learning portals, thereby potentially constraining the comprehensive evaluation of individual student progress. Considering other Massive Open Online Courses (MOOCs) and activities outside of the LMS could provide diverse options for action items.
Furthermore, in consideration of the absence of a student end portal facilitating the organisation of various online channels and ownership of learning data, an ordinary student confronts the challenge of efficiently managing the required resources and sharing personal learning data with the corresponding university or employer, solely under authorised and specific circumstances.

Motivation
The objective underlying this research seeks to develop a novel generation system capable of documenting multi-source communications within a platform that maintains students' records and enables the generation of statistics and reports. During the ongoing transition from the current system to the personalised feedback-enabled e-learning, the goal is also to devise an approach that bridges the gap in essential information required for evaluating each individual student's portfolio matrix and learning progress. Moreover, the system is intended to enhance communication between instructors and feedback design, thus enabling the retrieval and storage of desired information while eliminating data redundancy issues. Additionally, through surveys, we seek to enhance the user experience by improving system interactions and providing diverse support based on the actual needs of the students.

Impact of Learning Analytics-Based(LAB) Process Feedback in a Large Course
Notwithstanding the developing number of LA-based frameworks intended to give feedback to students, Lisa Lim's investigation [3] observed how handling criticism and processing feedback, in view of students' learning data, impacts their Self Regulated Learning (SRL) and academic performance at a coarse level and academic performance within the context of a biological science course. The study utilised multimodal data, including log data generated from the Learning Management System (LMS) and the Connect digital book, just as students' performance in the assigned course assessment.
Two sets of conditional probabilities were computed: the probability of SCORM accessed with feedback (p = 0.63 ) and the probability of Sharable Content Object Reference Model (SCORM) accessed without feedback (p = 0.59). These findings indicate that students engaged with Connect differently across the platforms, and the pattern of engagement varied between groups. The treatment group achieved higher final course grades despite recording significantly lower Connect activity.
The diagram of assessed minor methods for fruition rate shows that, for the benchmark group, there was a decline in Connect activity during the semester, followed by an increase in activity towards the end of the semester. Conversely, the pattern for the treatment group was reversed, with lower completion rates at the beginning, followed by a sharp increase during the semester, which was sustained for the remainder of the semester.
Meanwhile, linear regression was conducted to predict the final course mark based on the OnTask grouping (1 = with OnTask feedback; 0 = no OnTask feedback), prior academic achievement as measured by program entry score, and the interaction between prior academic achievement and program entry score. The results indicated that the program entry score (b = 0.26, p < 0.001), OnTask group (b = 0.47, p < 0.001), and the program entry score by OnTask group interaction (b = 0.10, p = 0.31) significantly influenced the prediction. The regression equation demonstrated a significant relationship, with F(3353) = 24.05, p < 0.001, and an R2 of 0.16.

Metacognitive Control
Metacognitive insight proves to be valuable only to the extent that it can be utilised to guide and regulate behaviour. Therefore, extensive research has been conducted to explore the mechanisms by which individuals modify their responses based on the selfunderstanding acquired through metacognitive observing [26].
Metacognition models posit that when individuals assess tasks and recognise shortcomings in their current psychological or social processes for achieving their goals, they engage in control activities to ensure more effective or efficient goal attainment.
Although limited research has directly explored the link between monitoring and control actions in children, it has been hypothesised that a potential reason for children exhibiting poor metacognitive control skills is their struggle in translating monitoring assessments into fitting changes in conduct.
Ineffective metacognitive monitoring hampers the flow of goal progress information available for processing and simultaneously limits the ability to identify cues that could guide action. In certain contexts, such as alcohol consumption, metacognitive monitoring is disrupted. For instance, individuals may hold positive metacognitive beliefs about the benefits of alcohol, helping them manage worrisome thoughts. However, alcohol impairs their ability to introspect and monitor progress towards their goal of reducing stressing considerations.

Multimodal Learning Analytics
Multimodal Learning Analytics (MLA) includes traditional log-file data captured by online systems, along with learning artefacts and more natural human signals, including gestures, gaze, speech, or writing [27].
MLA constitutes a sub-field that aims to integrate diverse sources of learning traces into Learning Analytics research and practice, focusing on understanding and enhancing learning in both digital and real-world environments, where interactions may not necessarily be mediated through a computer or digital device. In MLA, traces are derived not only from log records or digital documents but also from recorded video and audio, pen strokes, positional GPS trackers, biosensors, and any other techniques that facilitate understanding or measurement of the learning process. Moreover, in MLA, traces extracted from multiple modalities are combined to provide a more comprehensive view of students' actions and internal states.
The application of diverse modalities for learning contemplation is a customary practice in conventional trial-based educational research, albeit relatively new in the context of LA. The incorporation of a human observer, inherently a multimodal sensor, into an authentic learning environment represents the established approach to studying learning in natural settings. Technological advancements, such as video and audio recording and labelling tools, have made this process less intrusive yet more measurable. However, the traditional approach to educational research encounters issues due to the manual nature of data collection and analysis, which proves to be costly and non-scalable. The limited scale and time constraints of data collection, along with the delayed availability of data analysis results, impede their utility for the learners under study. Nevertheless, if multiple modalities could be recorded, and learning traces could be automatically extracted from them, LA tools could offer a seamless real-time feedback loop to enhance the learning process [26].
In the context of communication theory, multimodality refers to the use of multiple communication methods to exchange information and meaning among individuals. The media-motion pictures, books, site pages-serve as the physical or digital substrate through which a mode of communication can be encoded. Each mode can be expressed through one or several media. For instance, discourse can be encoded as varieties of weight noticeable all around (in a vis-à-vis exchange), as varieties of attractive direction on a tape (in a tape recording), or as varieties of computerised numbers (in an MP3 file). Additionally, a similar medium can be utilised to convey multiple modes. For instance, a video recording can contain information regarding body language, emotions (e.g., facial expressions), and apparatuses or utilities used.

Hypothesis
The proposed new system aims to capture multi-source communications within a platform that maintains records for all students and allows the generation of statistics and reports; although transitioning from the current system to the Personalised Feedback-Enabled (PFE) e-learning, the design will also address the need for missing information to assess each individual student's portfolio matrix and learning progress. Moreover, it will facilitate improved communication between instructors and feedback design, enabling selective retrieval and storage of desired information while resolving data redundancy issues. Additionally, the system will enhance the user experience by improving system interactions and providing diverse support tailored to the specific needs of students.
The learner model will contain learner-related information to adapt to individual demands, comprising three components: personal information, prior knowledge, and learning style. Once the learner's profile and chosen area are identified, the associated information will be saved and processed by the knowledge extraction model. This model is divided into two phases: the relevance phase and the ranking phase.

Personalised Feedback
The research aims to explore the perception, interpretation, and actions associated with personalised feedback in the context of an experiment's design. The effective implementation of personalised e-learning depends on well-designed task-focused and improvementoriented feedback, structured content, and timely feedback loops. Without these elements, Massive Open Online Courses (MOOCs) might struggle to guide students adequately based on their individual learning intentions. In the proposed learning process, the learner's status is identified through pre-course exams and profile analysis. For first-time users, the system presents an original course material list (non-personalisation list) based on the learner's query terms. After the learners interact with course materials and complete assignments and quizzes, the system reevaluates their abilities and progress, adjusts difficulty parameters, and recommends appropriate materials accordingly.
To achieve this, the system should align human-computer interaction (HCI) with suitable methodologies, perceiving user behaviours to provide immediate and accurate responses. This approach facilitates attendees with varying backgrounds to achieve their objectives based on individual strengths and motivations. The framework or tool should utilise data collected from students' interactions with web-based educational systems to derive pedagogically personalised feedback. Defining the type of information beneficial to stakeholders is crucial, and the framework's objectives are as follows: Instructors: • Track students' course completion patterns to evaluate students' engagement with the online course. • Discern the concepts/topics that are most confusing for students by analysing forums. • Evaluate the quality and difficulty of quizzes/assignments by examining results. • Evaluate the study patterns by checking clicks per student for pre-reading materials. • Analyse surveys for queries regarding course instructions.
Learning researchers: • Observe the relation between students' course completion patterns and their results. • Analyse keystroke rhythm recognition for each student to maintain the integrity of online submissions.

Design of Feedback
Teachers can improve the instructional quality of their courses by making a number of small changes such as providing detailed task-focused and improvement-oriented feedback. As stated in "Implementing summative assessment with a formative flavour: a case study in a large class," several considerations should be made for the design of feedback. These include having a set of exemplars available online, providing detailed explanations of rubric/marking criteria through the video, breaking down larger assessments into linked summative assessments, utilizing audio feedback as a replacement for written feedback, and offering formative audio feedback to markers to enhance consistency in marking and feedback [28]. Teachers can enhance the instructional quality of their courses by implementing small changes, such as providing detailed task-focused and improvementoriented feedback.

The Pipeline of Feedback Data
Multiple innovative data-processing methodologies may be adopted during the entire journey of engagement in the e-learning system for every platform user. The data captured from video calls or Learning Management System (LMS) communication can be retrieved from diverse platforms through proper authorisation and subsequently classified according to our proposed data model structure. Amongst the data propagation, the data flow stemming from streamlined Event Input can undergo Natural Language Processing (NLP) for expressing meta-cognitive recognition. Additionally, deep learning (DL) methodologies, such as Bayesian Active Learning (BAL), can further aid in applying uncertainly estimated sequence labelling with promising performance. Based on the output stream design, the feedback data can be delivered to end users through various appropriate channels.

Structure and Content of Feedback Message
To present the feedback information effectively to the audience, it is needed to identify the necessity to connect data with actions that impact the learning experience and recognise the potential effect of the message. For instance, the feedback can be categorised as the context of criteria scope, achieved items, insufficient items, and actions. By adhering to the defined structure of feedback, all the necessary datasets can be linked and retrieved from the specified sources. Simultaneously, in-depth research should be conducted to determine the best practices for delivering the feedback message, including the syntax, wording, length, and emotional expression of the content.

Learning Data
Information Finder [29] utilises heuristics to extract significant phrases from documents for learning, whereas the agent learns standard decision trees for each user category. The system's objective is to monitor and log the selected actions of the learners during the educational process. All recordings are conducted with the user's knowledge and consent, imperceptibly at the user level, without any intervention in the educational process. The data collected during this process can be correlated with the learner's prior knowledge (theoretical background), skills in using new technologies, attitudes, and patterns during their use of the learning environment, as well as the instances and points where they sought assistance. The subsystem aims to log sufficient data, both in terms of quantity and quality, which can be utilised to build a personalised profile for the learner and provide personalised materials and assistance (Figure 1). In this study, the learning data can be divided into summative assessments, such as quizzes, assignments, papers, and online exams, and human factors feedback, such as selfefficacy, enjoyment, and usefulness, showed a strong correlation; thus, they are expected to have a significant positive impact on satisfaction, which, in turn, positively influences effectiveness. Control, attitude, teaching style, learning style, promptness, feedback, availability, and interaction exhibited varying degrees of moderate correlations [30]. The human factor learning data can be collected through questionnaires, interviews, and student surveys. Accordingly, survey questions will be attached after each of our summative assessments to understand the emotional feedback. If formative evaluation protocols are provided and the workload can be met, there is a question regarding how to motivate students to participate in the assignments. The self-regulator learns to balance all types of feedback instead of relying solely on the strongest or cheapest option [31].
In our experimental design, users will undertake the course with the learning progress data monitoring, including:

Classification, Regression, and Clustering
The author of "Deep Bayesian Active Learning for Natural Language Processing: Results of a Large-Scale Empirical Study" [32] has discovered that Bayesian active learning by disagreement, utilizing uncertainty estimates provided by either Dropout or Bayesby-Backprop, significantly enhances performance compared to i.i.d. baselines. Moreover, it often outperforms classic uncertainty sampling. In this context, prediction can aid in identifying "at-risk" students. This research treats the problem as a regression task and compares the F1 scores of two models, Linear regression and Support Vector Regression (SVR), in the realm of supervised learning. Specifically, Support Vector Regression is employed, wherein slack variables are introduced to handle errors, similar to the soft margin loss function in SVMs for classification. By treating it as an optimisation problem, the objective is to fit a maximum number of points within the decision boundary lines. Representatively, a generic hyperplane equation Y = wx + b, −n < Y < +n should be satisfied by hyperplane(s).

Natural Language Processing
Natural Language Processing (NLP) constitutes a prominent component of the ongoing AI revolution, including various tasks, from text tagging to understanding semantic content. In the context of online learning analysis, NLP techniques can be employed to validate the contents against learning outcomes. Linguistic knowledge and word features are utilised to extract significant key phrases and keywords representing each content, facilitating the assessment of whether a website fulfils the learning outcomes.
Evidence suggests that formative assessment positively impacts student outcomes, including enhanced academic performance, self-regulated learning, and self-efficacy. In this regard, there exists a research tool, such as MetaTutor, which can measure the deployment of self-regulatory processes through the collection of rich, multi-stream data [33].
Our system aims to leverage NLP to enable learners to express their meta-cognitive monitoring and control processes. For instance, learners can interact with the SRL palette and rate their level of understanding on a six-item Likert scale before taking a quiz. Moreover, the interface allows students to summarise a static illustration related to the circulatory system. The difficulty level of quizzes or exams can be dynamically adjusted based on students' interactions with historical data. Additionally, our research can utilise near state-of-the-art results for a variety of unstructured and structured language-processing tasks [34], while remaining more memory and computationally efficient than deep recurrent models. Small feed-forward networks are sufficient to achieve accuracy on various tasks under the deep recurrent model. In the context of NLP, AL serves as a straightforward technique for data labelling. It involves selecting instances from an unlabelled dataset, which are then annotated by a human oracle, and this process is repeated until a termination criterion is met. Meanwhile, Stream-Based Learning can incorporate unlabelled data into real-time streams for agent decision-making and state classification [35].

Experiment Hypothesis
Hypothesis 1. The same student cohort will exhibit similar levels of responsiveness to feedback, even when presented with different questions.
This hypothesis is premised on, that irrespective of the question type, students who excel in one AI/NLP quiz item are expected to demonstrate similar performance in other AI/NLP quiz items. The personalised feedback (PF) group will be offered customised feedback tailored to their quiz responses and understanding level, whereas the natural feedback (NF) group will receive standardised, non-personalised feedback.
Should this hypothesis hold true, it is anticipated that students within each group will exhibit similar levels of satisfaction and engagement, despite encountering different quiz questions.

Hypothesis 2.
The variation in survey result between the PF group and NF group is expected due to the provision of superior user experience and more detailed feedback in the PF group.
The underlying assumption is that personalised feedback (PF) will result in an enhanced user experience and more detailed feedback compared to natural feedback (NF). The survey questions will assess aspects such as feedback clarity, usefulness, and fairness, enabling the evaluation of satisfaction levels within the PF and NF groups. Should this hypothesis prove valid, it is expected that students in the PF group will express higher levels of satisfaction than those in the NF group, as the former will have received more custom-tailored and informative feedback.

Experiment Setup
Under normal circumstances, the curriculum is followed in a sequential manner. Sometimes, there is no direct link between the knowledge points covered in two consecutive chapters (i.e., no prerequisite of learning "a" before "b"). Thus, an automated approach can be adopted to devise the most appropriate learning sequence for students at a particular level, which is known as learning path planning ( Figure 2).
The objective of the system revolves around extracting information from publicly available educational resources on the Internet. It takes into account the learners' backgrounds and requirements to develop modules and a planner system that facilitates the learning process. This process is further supported by the creation of an ontology, which optimises the information extraction procedure. The module's aim is to introduce these students to machine-learning concepts and improve their problem-solving abilities. Throughout the course, students' learning progress will be monitored, and the system will collect their participation data and incorporate it into a matrix. Upon completing quizzes, participants will receive personalised feedback. The experiment incorporates two distinct levels of personalised feedback and two varying task-related features (Table 3).

RCT Group
The process of randomisation is crucial to ensuring that any differences in outcomes between the two groups are solely attributed to the intervention, rather than any other factors, such as demographic differences or pre-existing conditions.
Participants will be enrolled in the study and randomly assigned to either the RCT group or the control group. Both groups will have access to identical online learning activities, including quizzes and interactive exercises, within the Learning Management System (LMS). These activities are designed to teach the same concepts and skills to all participants. However, the RCT group will also have access to additional adaptive learning features within the LMS, such as personalised feedback, targeted practice exercises, and customised review materials. For instance, different blurbs can be written for learners who either passed or failed a set of questions in the same quiz, or separate suggestions can be provided for those who are minimally, partially, or completely engaged with the course activities. These rules are applied to each individual learner to generate personalised text (or resources). The features are designed to offer enhanced support and reinforcement to participants in the RCT group. To compare the outcomes between the RCT group and the control group, the data collected from assessments and LMS usage will be analysed.
In our Moodle LMS, two courses will be established, both containing the same tutor slides and quizzes. However, the RCT course setup will include personalised feedback and targeted exercises in the configuration, and the control group will receive low-level feedback.
Gamified Elements A/B Testing In addition, two student cohorts will be set up to observe the efficacy of gamified learning. One of the groups shall be exposed to badges and leaderboards, including points, metres, and resource allocation dashboards, along with instant event feedback loops and real-time responses. Pertinent feedback will be provided, stating, "Remember that decision you made or did not make in the previous level-well, you need that tool or information now, so go back and earn it". Conversely, the other group of students will only have access to an unadorned dashboard, offering delayed and simplistic instructional feedback such as "Go back and learn it".

Evaluation
Subsequently, by analysing students' submitted answers, completion rates, time taken, and satisfaction surveys, the students' Learning Performance (LP) shall be calculated. Here, k represents the number of items, and the correctness of the students' responses on each item i is considered, with zi ∈ 0,1, in conjunction with the item's difficulty level. Each item has been previously weighted based on its difficulty level and contributes differently to the overall self-assessment score, ranging from 0.5 points (easy) to 1 point (medium) to 1.5 points (hard). The final score will be presented on a [0-10] scale.
During the evaluation phase, participants will be shown feedback with the Details Relevant level, and the results from the learning stage will be emphasised. All forthcoming enhancements provided by participants will be recorded through an additional round of course slides and quizzes from the system in the Learning Stage (Table 4).

Non Task Related
The way you crafted these algorithms to fit together to take the machine learning tasks through the Nomral Language Process is very good.
good job

Task Related
Your goal was to write about all of the things you did (the events) in order. Yes, you have written the first thing first, but after that it becomes muddled. Your next step is to go back to your plan and check your numbering of the order of the events as they happened. Then you can use your plan to rewrite them in that order.
Provides information on which steps were done or not

Data Collection Process
Step 1: Participant Selection and Random Assignment For the experimental phase, a total of eight participants will be chosen and randomly allocated to either the personalised feedback (PF) group or the natural feedback (NF) group. These participants will be profiled into two categories: four of them possess a background in information technology and software engineering, whereas the remaining four have a background in marketing and company strategy. All participants share a common interest in learning artificial intelligence or machine learning, along with a certain level of discovery in the domain.
According to the research settings, two participants from the IT background and two from the marketing background will be placed in the personalised feedback (PF) group, whereas the rest (maintaining the same configuration) will be assigned to the natural feedback (NF) group. The NF group will receive limited feedback subsequent to quiz submission. Both groups will undergo five AI/NLP quiz questions, specially designed to assess their comprehension and knowledge of the subject matter.
Considering that the above-mentioned experiment serves as our pilot test, we intend to carry out further experiments, refining the questions and feedback mechanisms, and collecting more data to validate our hypothesis.
Step 2: AI/NLP Quiz Questions Both groups will undertake five AI/NLP quiz questions, which will be designed to test their knowledge and understanding of the subject matter.
Step 3: Feedback and Scoring Upon submitting their answers, participants will receive varying degrees of feedback and scoring in accordance with their respective group assignments. The PF group will receive personalised feedback tailored to their quiz responses and level of understanding, whereas the NF group will receive standard, non-personalised feedback. Participants will receive their scores immediately after submitting their answers.
Step 4: Survey Questions The researchers will administer three survey questions to both groups, aiming to assess their satisfaction with the feedback and scoring process.
Survey 1: Rate your confidence level for completing the knowledge or skill presented. Survey 2: Was the quiz feedback timely and relevant? Survey 3: Did the answer feedback help you gain a clearer understanding of the subject?
Step 5: Data Collection and Analysis The data collected will include participants' quiz scores, feedback, and survey responses. Statistical methods will be employed to analyse this data and determine if there is a significant difference in quiz performance, feedback satisfaction, and overall engagement between the PF and NF groups.
For quantitative data, descriptive statistics will be utilised to summarise and describe the quantitative data associated with quiz questions and feedback scoring. Correlation analysis will be conducted to appraise the relationship between variables concerning quiz questions and feedback scoring. The variables included in the correlation analysis will include quiz scores and feedback ratings. Moreover, regression analysis will be employed to assess the relationship between variables related to quiz questions and feedback scoring. The regression analysis will involve utilizing the dependent variable (e.g., feedback ratings) and independent variables (e.g., quiz scores, time spent on the quiz).
Regarding qualitative survey questions, the researchers will adopt theoretical sampling, selectively choosing new participants or data sources to gather additional information that can further refine the emerging theory. Additionally, data that challenges or expands the existing concepts and relationships will be sought. This iterative process of data collection, coding, and theory development will continue until theoretical saturation is achieved.
Overall, the experiment is designed to test whether personalised feedback significantly impacts quiz performance and feedback satisfaction compared to standard feedback. By randomly assigning participants to different groups and carefully collecting and analysing data, the effectiveness of personalised feedback can be evaluated, thus providing insights for future educational practices and technology development.

Data Import
The Import is as follows.
import pandas as pd import seaborn as sns import numpy as np data = pd . read_csv ( '~/Downloads/ P e r s o n a l i s e d L e a r n i n g . csv ' ) data . head ( )

Data Prepossessing
Preprocess the collected data for the preparation of analysis, fill any missing data with zeros. data = pd . get_dummies ( data ) data = data . f i l l n a ( 0 ) Scale the data to derive zero mean and unit variance. from s k l e a r n . p r e p r o c e s s i n g import S t a n d a r d S c a l e r data = S t a n d a r d S c a l e r ( ) . f i t _ t r a n s f o r m ( data )

Data Clustering and Shaping
Unsupervised Learning: Used unsupervised learning techniques to identify interesting patterns in the data. Perform k-means clustering on the data. kmeans = KMeans ( n _ c l u s t e r s =3) c l u s t e r s = kmeans . f i t _ p r e d i c t ( data ) from s k l e a r n . c l u s t e r import KMeans from s k l e a r n . c l u s t e r import KMeans kmeans = KMeans ( n _ c l u s t e r s =3) c l u s t e r s = kmeans . f i t _ p r e d i c t ( data )

Provide Personalised Feedback to Each Student Based on Cluster
Design the interactive feedback system that provides personalised feedback to the students based on the identified patterns as well as feedback to students based on their cluster membership.
Define feedback messages for each cluster. The analysis will be conducted through the next step when we have collected the responses concerning the feedback from students. Hypothesis 1: The same student should receive similar levels of feedback, even for different subject questions (Q1: NLP related and Q3: GPT model related). To test this hypothesis, a t-test was performed using the feedback ratings for Q1 [3,5,5,4,4,3,5,5] and Q3 [4,5,5,4,4,4,5,5].
import pandas as pd import seaborn as sns import numpy as np import s c i p y . s t a t s as s t a t s Hypothesis 2: Use ANOVA to establish a significant gap in the survey results between various groups of students: Students A and B of the PF group, and Students A and B of the natural group. The PF group, which provided detailed feedback, was expected to yield a better user experience as inferred from the collected survey ratings. Tukey's HSD (Tukey's Honestly Significant Difference) post hoc test was planned for further examination of the treatment levels.
Pick first student of natural group and first student from pf group.

Conclusions
The t-test represents a well-established statistical analysis method with extensive applications in scientific research. Its purpose is to determine the presence of a significant effect resulting from an intervention or treatment: through the comparison of means between two groups, researchers can identify whether the intervention or treatment has yielded a significant effect, while also considering whether the observed results are attributed to the intervention or treatment being tested or if they arise from other influencing factors.
The t-test results indicated that there was no significant difference between the feedback on different questions. The p-Value of 0.5056732339622882 exceeded the conventional threshold of 0.05, providing evidence to support the hypothesis that there is no significant difference in student test scores concerning the different feedback for the questions.
The experimental results indicate that the evaluated feedback on different types of questions did not lead to a significant impact on student test scores. Accordingly, it suggests that employing variations in feedback approaches or techniques may not result in noticeable differences in student performance within this particular context. Nevertheless, it is important to acknowledge that different types of feedback could still hold value in terms of enhancing student understanding, motivation, or metacognitive abilities, even if they do not directly influence test scores. To gain further insights, future research efforts could explore the potential benefits of various feedback strategies beyond their impact on test performance. Replicating this study or conducting similar experiments with larger sample sizes or diverse populations could serve to corroborate or challenge these findings.
Analysis of variance (ANOVA) testing is a statistical analysis method employed to ascertain whether a significant difference exists among the means of three or more groups. ANOVA testing proves to be a potent instrument for researchers appraising extensive datasets with multiple groups and is commonly utilised in experiments where efforts are made to control variables that could potentially influence the results.
ANOVA testing is a well-established statistical analysis approach with a considerable history of application in scientific research. Therefore, there exists an extensive body of knowledge and research guiding the effective use of ANOVA testing, which researchers can draw upon to design and analyse their experiments.
From ANOVA testing, when comparing the natural and personalised feedback groups with data collected from two students, the F-value of 6.249999999999997 indicates a significant difference between the groups. The p-value of 0.01716202358048234 falls below the conventional threshold of 0.05, providing robust evidence against the null hypothesis.
The finding suggests that personalised feedback exhibits a significant impact on individual students compared to natural feedback. This indicates that the tailored and individualised nature of personalised feedback likely contributed to differences in student experiences, perceptions, or outcomes. The significant difference between the two feedback groups implies that providing personalised feedback may be more effective in engaging students, enhancing their understanding, or addressing their specific learning needs compared to generic or natural feedback. This finding highlights the potential benefits of tailoring feedback to individual students.
Educators and instructional designers should contemplate integrating personalised feedback strategies into their teaching practices. Our finding suggests that dedicating time and effort to provide personalised and targeted feedback, considering each student's strengths, weaknesses, and learning styles, can yield more positive outcomes. Meanwhile, educators and researchers may need to explore other variables or instructional strategies that could exert a more substantial influence on student learning and test outcomes.

Feedback of Feedback
This study has focused on understanding how students respond to receiving feedback on their performance within an educational setting. Specifically, the impact of personalised feedback on student satisfaction, engagement, and learning outcomes will be examined. Surveys will be conducted, and qualitative data will be collected from students who have received feedback on their work, as well as those who have not. By analysing the feedback of feedback, insight into the effectiveness of different types of feedback will be gained, and areas for improvement in the feedback process will be identified. Additionally, the extent of engagement and learning in offline settings, such as peer-to-peer study and mentoring, will be considered.

Tukey's HSD Post-Hoc Test
Tukey's HSD post hoc test is a statistical analysis method employed to compare the means of multiple groups to identify differences between them. Within the scope of this study, Tukey's HSD test shall be utilised to analyse the results of the survey data and determine potential significant distinctions in feedback satisfaction and engagement between the groups receiving personalised feedback (PF) and natural feedback (NF). This test will enable the identification of specific groups exhibiting significant differences and offer insights into the potential drivers of those differences.

Bridge the Required but Missing Information to Evaluate Each Individual's Portfolio Matrix and Learning Progress
To appraise each student's portfolio matrix and learning progress, access to exhaustive and accurate information concerning their performance, strengths, and areas for improvement holds vital importance. However, it is frequently encountered that certain fragments of this information are missing or incomplete, impeding a comprehensive understanding of each student's educational journey. In order to surmount this challenge, an approach will be devised that integrates data from various sources, including assessment results, learning logs, peer feedback, and teacher observations. Employing machine-learning algorithms, this approach will uncover patterns and trends within the data, subsequently rendering personalised recommendations for each student, thereby aiding them in attaining their learning objectives. By bridging the information gap, we aspire to provide a more precise and all-encompassing assessment of each student's portfolio matrix and learning progress.

Investigate the Various Forms of Participant Support Offered in These Digital Learning Environments
The debate over the most suitable pedagogical approach in MOOCs has profound implications for the future of online education; although the pedagogy of abundance promises scalability and accessibility, concerns arise regarding the quality of learning experiences and the potential detachment of learners from human interactions. On the other hand, the human-centric pedagogy champions the creation of meaningful learning environments but might struggle to accommodate a vast number of participants.

The Design and Development of an Innovative Online Virtual Simulation Course Platform
Creative thinking has become a highly sought-after skill in the contemporary job market, with employers valuing individuals who can think critically, generate novel ideas, and solve complex problems. To address this demand and empower college students with essential skills, the CreativeThinker platform was conceived. By leveraging the power of virtual simulations, we are looking forward to inspiring and guiding learners in cultivating their creative thinking capabilities with such an add-on.