Skip Content
You are currently on the new version of our website. Access the old version .
ElectronicsElectronics
  • Article
  • Open Access

2 February 2026

Cross-Course Knowledge Tracing for Student Performance Prediction in MOOCs

,
and
1
School of Information Technology in Education, South China Normal University, Guangzhou 510631, China
2
School of Artificial Intelligence, South China Normal University, Foshan 528225, China
3
School of Computer Science, South China Normal University, Guangzhou 510631, China
*
Author to whom correspondence should be addressed.

Abstract

Knowledge tracing (KT), which uses machine learning models to predict students’ future performance, has received a lot of attention in intelligence education. However, in Massive Open Online Courses (MOOCs), most of the existing KT methods can only track students’ performance in one course. In addition, when there is scant learning record data on new courses, training a new KT model becomes challenging. Furthermore, existing KT methods tend to excel in specific courses, but their generalization ability is inferior when faced with similar or distinct courses. To address these challenges, this paper proposes a MOOC-oriented cross-course knowledge tracing model (MCKT). In MCKT, we first construct two attribute relationship graphs to obtain the student and KC representations from the source course and the target course, respectively. Then, the element-wise attention mechanism is used to fuse the student representations from both courses. Next, MLPs are used to reconstruct the interaction between students and KCs in each course to enhance students’ cross-course representation. Finally, recurrent neural networks (RNNs) are used to predict the students’ performance in each course. Experiments show that our proposed approach outperforms existing KT methods in predicting students’ performance across diverse MOOCs, proving its effectiveness and superiority.

1. Introduction

The fourth industrial revolution, driven by the integration of intelligent technologies such as the Internet of Things (IoT), big data analytics, and Artificial Intelligence (AI), has transformed the way data is generated and decisions are made in both industrial and societal domains. Machine learning architectures like Transformers have become essential for automating complex analytical tasks. In industrial settings, these models learn from sensor data and operational logs to carry out tasks such as predictive maintenance and quality control. On social media and digital platforms, they process vast amounts of sequential and graph-structured interaction data to drive content recommendations and user behavior modeling. This also provides important analogies and a technical foundation for the field of education, enabling the emergence of knowledge tracing (KT) models. KT models can help intelligent tutoring systems dynamically understand the students’ mastery of knowledge, make timely adjustments to the teaching plans, and recommend relevant teaching resources for students. To study this issue in depth, many efforts have been made to devise various methods of KT. One of the most widely used is deep knowledge tracing (DKT) [1] based on the deep learning method. DKT typically employs long short-term memory networks (LSTMs) to model students’ learning interaction state. DKT converts students’ exercise-answering performance history into a time series to predict students’ answers in the next step and infer their mastery of KCs. On this basis, many deep learning-based KT models, such as CNN-KT [2], Transformer-KT [3], and GNN-KT [4], have also been proposed to improve KT models’ performance.
Although these KT models have proven to be effective, it is still difficult to play to their strengths in Massive Open Online Course (MOOC) scenarios. (i) Single-Course Limitation: The traditional KT models only track the knowledge mastery status of students in a certain course and cannot predict their performance in another course. However, MOOCs include a large number of courses, and it is common in practice for a student to study more than one course at the same time. Figure 1 show four students in course a and four students in course b, where K a t e and T o m are enrolled in both courses (referred to as common students in this study). Students learn different KCs in different courses, but most existing KT models cannot capture the performance of a student in different courses.
Figure 1. Schematic diagram of MOOC user learning.
(ii) Data Scarcity Issue: There is a common problem of scarcity of learning record data in MOOCs [5], which will lead to difficulty in training the KT model of a single course. KT methods rely on large amounts of historical data from students to build models and make accurate predictions. Yet, in MOOCs, especially in newly opened courses or in the educational setting of some schools, learning record data may be very limited or non-existent, making it difficult to apply training KT. (iii) Generalization Difficulty: A course in the MOOCs usually has a long learning cycle and requires students to continuously invest time. As a result, students often want to predict how they will perform on a new course before they start that course in order to make better decisions about whether to start learning or how to plan for it. Nevertheless, the existing KT models only focus on one course, lack generalization ability, and cannot be used to predict the performance of students in the new course. (iv) Lack of Video Interaction Mapping: In addition, most of the courses in MOOCs contain many instructional videos, and the task of answering questions after watching the videos is not mandatory. However, most of the existing KT models map the level of mastery of KCs based on the results of students answering exercises. Due to the lack of a method to map the mastery level of KCs according to students’ video browsing behavior, the existing KT models cannot be widely used in MOOCs. These limitations hinder the realization of robust, automated learner modeling in MOOCs—a context characterized by massive scale and data-rich interactions where such technology should offer significant value.
To overcome these hurdles and advance towards truly adaptive MOOC platforms, we propose a transfer learning-based MOOC cross-course knowledge tracing method, namely MCKT, to predict students’ learning performance in the next step of a different course. The framework of MCKT can be divided into three parts, i.e., data pre-processing, pre-training, and prediction. In the part of data pre-processing, we extract the sequence of students’ interaction with knowledge concepts (KCs) based on the interaction records between students and items (e.g., exercises and videos) in the MOOC dataset; this approach follows the classic way of dealing with knowledge tracing datasets. In the part of pre-training, two graph structures are used: (i) we use the skip-gram algorithm to obtain the initial student representation in the structure of the attribute relationship graph (ARG); (ii) we use element-wise attention and a multi-layer perception (MLP) to reconstruct the interaction relationship between the students and KCs in the Interaction Relationship Graph (IRG) structure after the cross-course method. By using the above two graph structures in the pre-training, the students’ representations not only combine the similarity of students’ attributes in the source course and the target course, but also combine their relationship with KCs in the source course. In the prediction phase, the RNNs are used to model the results of the interaction between students and KCs and predict students’ mastery of the next KC in both the source and target courses.
In summary, the contributions of this study are threefold:
  • This paper proposes a MOOC-oriented cross-course KT method based on transfer learning, which predicts the status of the next stage of students’ learning in the different MOOCs with unbalanced learning records.
  • This paper proposes a method to evaluate the mastery of KCs by using the information of video viewing in MOOCs so that the KT model can be applied to video viewing behavior, not only to answer exercises.
  • This paper reports on the extensive experiments that were conducted in six scenarios based on two real MOOC datasets. The experiments verify that the proposed model is more effective than the existing KT methods in cross-course KT tasks.
The rest of this paper is organized as follows. Section 2 briefly reviews the related work. Section 3 introduce the problem statement. Section 4 describes the framework and the implementation of the proposed approach in detail. Extensive experimental studies to validate the effectiveness of our model and several research questions are addressed in Section 5. Section 6 concludes this work, analyzes this study’s limitations, and proposes future research directions.

3. Problem Statement

In this section, we first review the definition of knowledge tracing. Then, we introduce the graph structure used to represent the relationships between the students and items in the MOOC. Finally, we explain the cross-course knowledge tracing in MOOC scenarios. Some important notations are listed in Table 1, and we present a more detailed explanation of their role in the context.
Table 1. Some important notations.
Knowledge Tracing. Given the sequence of students’ learning interactions in online learning systems, knowledge tracing aims to monitor students’ changing knowledge states during the learning process and accurately predict their performance on the next items. Formally, in an intelligent education system, suppose the number of students, items, and KCs are denoted as | S | , | V | , and | K | , respectively. For a student s S , the sequence of items (e.g., video or exercise) with which the student interacted is recorded, denoted as I = { ( v t 0 , α t 0 ) , , ( v t i , α t i ) , } , where v t i V is the item that appeared at time step t i and α t i indicates the corresponding score. For instance, if the answer of the exercise is correct, α t i is 1 (or tick); otherwise, α t i is 0 (or cross). Knowledge tracing can be formulated as follows:
P r s ( α t n + 1 = 1 ) = Ω ( v t n + 1 | I t n s , V t n ) ,
where, P r s denotes the probability of correctly answering. Interaction Relationship Graph. In the process of viewing videos or answering exercises during the course learning, there are heterogeneous relationships among students, items, and KCs [17,18]. An Interaction Relationship Graph (IRG) is denoted as G = { V ; E } , consisting of an object set V and a link set E . V is a set of object types (e.g., student, item, or KC). E is a set of relation types, and E = { r A , r C } , where r A represents the relation interact and r C represents the relation contains.
As shown in Figure 1, for the IRG of course a, M a r y views the video v 1 a and v 2 a . v 2 a requires KCs k 2 a and k 3 a . v 1 a requires KC k 1 a . From the figure, we can observe that there is a relationship between k 2 a and k 3 a through v 2 a , and there is also a relationship between v 1 a and v 2 a through M a r y . The reason is that M a r y views both videos during her learning process. In the KT task, the prediction of KC mastery is more general, so in this work, the IRG is used to model the relationship between students and KCs in a MOOC.
Cross-Course Knowledge Tracing. Students often enroll in more than one course in MOOCs. As shown in Figure 1, suppose there are two courses a and b, denoted by C a = ( S c o m a , S d i s a , V a , I a ) and C b = ( S c o m b , S d i s b , V b , I b ) , where S c o m a and S c o m b represent students who have taken both course a and course b (i.e., Kate and Tom, referred to as ‘common students’). S d i s a represents the distinct students in course a (i.e., John and Mary), and S d i s b represents the distinct students in course b (i.e., James and Jone). Students enrolled in different courses are referred to as ‘distinct students’ in this paper. V a and V b represent the set of items of C a and C b separately. I signifies the set of interactions. Given the respective interaction records in these two courses, the goal of the cross-course knowledge tracing is to predict students’ mastery level of knowledge in b/a courses using the interaction data observed by a student in the a/b course.
Given a sequence of learning resources items V t n a = { v t 1 , v t 2 , , v t n } of the course a, if student s has learned on V t n a , and the interaction recording is denoted by I t n s , then the probability of the knowledge mastery of this student on the course b is calculated as
P r s ( α t n + 1 b = 1 ) = Ω ( v t n + 1 b | I t n s , V t n a ) ,
where P r ( α t n + 1 b = 1 ) represents the probability of answering correctly, which usually also reflects the mastery level of student s on the corresponding KC of the course b at the next time step, and the closer its value is to 1, the higher the student’s mastery level.

4. Proposed Approach

To address the problem introduced in Section 1, this paper proposes a novel KT method for cross-course student performance prediction in MOOCs, namely MCKT. As shown in Figure 2, the framework of MCKT is divided into the three modules, i.e., the Course graph construction module, the Cross-course embedding combination module , and the Knowledge performance prediction module. At the beginning, we extract data from the MOOC dataset that can be used for cross-course KT, including students, items, the text of KCs, the KCs contained in the items, and the results of students’ interaction with the items. Then, we calculate the similarity of the textual information of students, projects, and KCs in two courses (e.g., C a and C b ) to form an attribute relationship graph (ARG). Next, an attention mechanism and the MLP are employed in the Cross-course embedding combination module to generate the cross-course fusion representations of the common students, which is combined with the features from courses a and b. Finally, all students’ representations and the interaction results are sent to the Knowledge performance prediction module to predict students’ performance.
Figure 2. The framework of MCKT taking course a and course b as examples. The A R G and I R G are constructed by pre-processing the dataset, and the representations of the students and KCs are obtained. Then the representations of the cross-course common students and distinct students are obtained by pre-training. Next, the cross-course representations of the students and KCs from time t 0 to t n are input into the prediction layer to predict the knowledge state at the t n + 1 step.

4.1. Data Pre-Processing

In reality, some MOOC platforms provide exercises or tasks containing KCs for the courses, such as MOOPer, which has many online exercises for each course. However, to increase utilization [5], some MOOC platforms provide learning videos that contain KCs, such as MOOC-Cube, and do not necessarily provide online exercises for all courses, nor do they force students to answer online exercises. Video-watching sequences (e.g., play, pause, fast-forward, and replay) exhibit temporal patterns that have been shown in prior research [19,20] to correlate with learning attention and self-regulation (e.g., studies on clickstream analysis for engagement inference). Therefore, converting these sequences into simulated knowledge states provides a plausible, data-driven approximation. Based on this assumption, for courses that provide only video-watching logs but no exercise response records, we follow the method described in Reference [9], which uses the processed video-watching log data as the input data.
The first step involves consolidating information from the students watching the video. The initial representation of a student is constructed from five kinds of data, i.e., the learned courses C; the learned videos V; the number of times a student watched a learned video m; the duration of each learned video l; and the average duration that the s-th student spent on each learned video τ . We consider that the student’s mastery level of each KC is determined by the ratio of the average duration of the corresponding video learned by them to the total duration of the video. The process of obtaining the result of the interaction is formalized as
α v = t a n h ( τ v l v × m v ) ,
where α v represents the result of the student’s interaction with video v.
In the second step, the student’s ID, the learned video’s ID, and the mastery probability of the KC in the learned video (each video contains one KC) are combined into a triplet I to form a sequence of the student’s interaction–response to the video:
I t = ( s , v t , α t v ) ,
where I t represents an interaction between the student and video v at time t.
The third step is to generate two specific representations of each student. One reflects the student’s association with the learning videos (or other items, such as exercises), denoted by E ( g ) ,
E ( g ) t n = { ( s 1 , v 1 , k 1 ) t 1 , ( s 2 , v 2 , k 2 ) t 2 , , ( s i , v j , k m ) t n } ,
where E ( g ) is the initial data for the IRG, which is used to further generate cross-course representations in the Cross-course embedding combination module. And the other reflects the KCs and state time series of the learning interaction, which is denoted by E ( x ) ,
E ( x ) t n = { I t 1 , I t 2 , , I t n } ,
where E ( x ) is employed to further model the student’s mastery of KCs in the Knowledge performance prediction module.

4.2. Course Graph Construction

In this module, a weighted attribute relationship graph (ARG) for each course is formed. E ( g ) is input to this module. The similarity of the text representation vectors is calculated to generate edge weights between corresponding nodes. The text representations of students and KCs are derived using Doc2Vec [21] from their respective textual attributes: student ID, name, and enrolled courses for students and KC ID and textual description for knowledge components. The pairwise similarities among these resulting embedding vectors are then used to construct the weighted edges of the ARG connecting students to students, students to KCs, and KCs to KCs. The text representations’ similarities are formalized as the following Equations (7)–(9).
ω ( s , k ) = W s k × θ ( s , k ) ,
ω ( s , s ) = W s s × θ ( s i , s j ) ,
ω ( k , k ) = W k k × θ ( k i , k j ) ,
where s denotes the student node, k represents the KC node, W s k , W s s , W k k R d × V are the constant coefficient matrices, and ω ( · ) represents the relationship weight between nodes. θ ( · ) represents the cosine similarity. Through the above steps, a weighted ARG is created for each course.
Then, the Node2vec [22] module is employed to generate the node embeddings of these two ARGs. The embedding of a node is denoted by x ^ . The Node2vec module adopts the principle of similarity of adjacent nodes, uses the central node to predict the context node, and obtains the embedding of the target node after continuous iterative optimization. Node2Vec uses the skip-gram model to learn the embedding representation of the node. The goal of the skip-gram model is to predict the node itself based on its context. The optimization objective of Node2vec is
L n o d e 2 v e c = m a x f u V [ l o g Z u + n i N S ( u ) f ( n i ) · f ( u ) ] ,
where f is a function that maps each node u to a d-dimensional space vector, e.g., f = V R d . u is the current node, n i is the neighbor node of the u node obtained through the paths of the ARGs, and Z u = v V e x p ( f ( u ) · f ( v ) ) . Stochastic gradient descent is used to optimize L n o d e 2 v e c .

4.3. Cross-Course Embedding Combination

Considering that students are the subjects of cross-course learning behaviors and the importance of users in information systems [23], we propose to leverage the representation of students who have taken both the source and target courses as a transferable ’bridge’. Specifically, these overlapping students’ behavioral patterns in both courses are modeled and fused to establish a shared latent space, enabling knowledge transfer for new students in the target course based on their similarity to these bridge students in the source course. The representation of students in two different courses (take course a and course b as examples) are the embeddings of the student nodes in two ARGs. The task of this module is to generate the cross-course representations of students shared by course a and course b. In this module, the comment students’ representations are generated through element-wise attention [24]. Consequently, it attains high expressivity with neuron-level degrees of freedom [25]. We denote the representations of student s i in course a by E ( C a | s i ) ; we denote the representations of student s j in course b by E ( C b | s j ) . The process of cross-course feature combination is shown in the middle part of Figure 2.
The cross-course representations of students must be further processed before sequence modeling can be performed. The element-wise attention mechanism combines the common students’ embedding representations in courses a and b, and then a matrix multiplication calculation is performed on these cross-course representations. Specifically, the representation of common students in courses a and b can be expressed as follows:
E ~ ( C a | s i ) = W a E ( C a | s i ) + ( 1 W a ) E ( C b | s i ) ,
E ~ ( C b | s j ) = W b E ( C b | s j ) + ( 1 W b ) E ( C a | s j ) ,
where ⊙ indicates element multiplication and W a and W b are, respectively, the weight matrices of courses a and b in the element-wise attention mechanism. For distinct students and items in different courses, we only keep their embeddings and do not use the attention mechanism because they do not have double embeddings in courses a and b.
We then used two MLPs to reconstruct the non-linear relationship between students and items. We denote the representations of two kinds of students in course a i.e., [ E ~ ( C a | s c o m ) , E ( C a | s d i s ) ] , by S a and denote items in course a by Q a . Then, S a and Q a are input to the neural node; the result is h i a s and h j a q following Equations (13)–(16),
h i a s = g ( S a W i a s ) ,
h j a q = g ( Q a W 1 a q ) ,
S a ^ = g ( h k a s ) = g ( g ( g ( h i a s W 1 a s ) , W 2 a s ) k ) ,
Q a ^ = g ( h l a q ) = g ( g ( g ( h j a q W 1 a q ) , W 2 a q ) l ) ,
where g ( * ) is the activation function ReLU and W 1 a s , W 2 a s , , W 1 a q , W 2 a q , , R d × | V | are the learnable weights of each layer of the MLP of the student and items, respectively. The label ⊙ denotes element-wise multiplication. We use a cross-entropy loss to reconstruct the relationship between students and items, following Equations (17) and (18):
y a ^ = σ ( S a ^ · Q a ^ ) ,
L ( y a , y a ^ ) = y a l o g ( y a ^ ) + ( 1 y a ^ ) l o g ( 1 y a ) ,
where y a ^ is the predicted interaction in course a, σ is the activation function sigmoid, and ( · ) is the inner product. y a is an observed interaction in course a. By this way, we can also obtain the same kind of loss in course b. The algorithm of the student cross-course representation generation is shown in Algorithm 1.
Algorithm 1 Algorithm of the student cross-course representation generation
   Input: The A R G s of two courses A R G a and A R G b ; The weight matrix of the course a and b in the element-wise attention mechanism W a and W b . θ T is a termination condition.
   Output: Two sets of representations of students used for prediction layer.
1:
  for s a in C a do
2:
    { E ( C a | s c o m a ) , E ( C a | s d i   f a ) } skip - gram ( A R G a ) .
3:
  end for
4:
  for s b in C b do
5:
    { E ( C b | s c o m b ) , E ( C b | s d i   f b ) } skip - gram ( A R G b ) .
6:
  end for
7:
  while θ T not converged do
8:
    E ( C b | s c o m a ) Attention ( E ( C b | s c o m b ) , E ( C a | s c o m a ) , W a ) {Equation (11)}
9:
    { E ^ ( C b | s c o m a ) , E ^ ( C a | s d i   f a ) } MLP ( E ( C b | s c o m a ) , E ( C a | s d i   f a ) ) {Equation (17)}
10:
E ( C a | s c o m b ) Attention ( E ( C a | s c o m a ) , E ( C b | s c o m b ) , W b ) {Equation (12)}
11:
{ E ^ ( C a | s c o m b ) , E ^ ( C b | s d i   f b ) } MLP ( E ( C a | s c o m b ) , E ( C b | s d i   f b ) ) {Refer to Equation (17) but apply to course b}
12:
 end while
13:
 return

4.4. Knowledge Performance Prediction

In this module, an LSTM model is utilized to model students’ behavior and response sequences when learning items. Specifically, the cross-domain representations of students and KCs generated by using Algorithm 1 are integrated into E ( x ) t n to form the sequence of the learning interaction (e.g., performing exercises or watching videos) and responses containing cross-course information. Here, each student’s learning interaction records and student representations are individually spliced to form the input data. These input data are fed into the LSTM model, which predicts the student’s knowledge state at the next step. The variables are related using a simple network defined by the equations
h t n + 1 = t a n h ( W h x E ( x ) t n + W h h h t n + b h ) ,
p t n + 1 = σ ( W p h h t n + 1 + b p ) ,
where σ is the activation function s i g m o i d . W h x is the input weight matrix, W h h is the loop weight matrix, h t 0 is the initial state, and W p h is the output weight matrix. b h and b p are latent bias and output bias, respectively.
We denote the actual performance of the i-th student on the j-th KC by r i j , while the learning interaction results are obtained by Equation (3). We train the model with two strategies based on two types of label values. The first strategy is to treat r i j as a continuous value; i.e.,
r t i j = Φ ( α i j v ) , w . r . t Φ ( α i j v ) ( 0 , 1 ) , n u l l , o t h e r w i s e .
where Φ is the normalization operation.
The other is to set r i j to a binary value of 0 or 1; that is,
r t i j = 1 , w . r . t Φ ( α i j v ) θ , 0 , w . r . t Φ ( α i j v ) < θ , n u l l , o t h e r w i s e .
where θ denotes the threshold used to divide the mastery of KCs. The loss functions of these two strategies are both defined as the cross-entropy of the prediction performance p and the real one r; i.e.,
L ( r t i j , p t ) = r t i j l o g ( p t ) + ( 1 r t i j ) l o g ( 1 p t ) ,
L = t = 0 n L ( r t i j , p t ) ,
Remark: The task of transfer learning is to transfer the model from the source domain to the target domain. However, in our research scenario, a course serves as both the target and source course. In each round of iterative training, we use the two courses in the experiment as source courses for cross-training. In this way, the model we train will consist of two sub-models, each capable of performing two-way cross-domain functions.

5. Experiments

In this section, we first provide detailed information about the dataset and experiment setup. This section then provides details of the extensive experiments that were conducted across six cross-course scenarios and the in-depth analyses. Additionally, to provide a clearer understanding of the model, we address the following key research questions in the subsequent sections:
RQ1: Is MCKT more effective than existing KT models in terms of cross-course knowledge performance prediction or not?
RQ2: What impact do the constituent modules of MCKT have on its performance?
RQ3: How effective is MCKT in predicting similar or non-similar courses?
RQ4: Can MCKT predict a student’s performance in another course?
RQ5: Is there is a difference in the representation of common students and distinct students?

5.1. Overview of Datasets

Our experiments are performed on two classic MOOC platforms that contain multiple courses, i.e., MOOPer [26] and MOOC-Cube [27]. MOOPer extracts the interactive data on users’ participation in practical exercises on the EduCoder platform from 2018 to 2019 and models the attribute information of courses, practices, levels, and KCs and the interrelationships between them in a knowledge graph to construct a large-scale practice-oriented online learning dataset. MOOC-Cube is an open large-scale data repository for natural language processing, knowledge graphs, data mining, and other research areas concerned with massive open online courses. It comprises 706 MOOCs, 38,181 videos, 114,563 KCs, and 199,199 real MOOC users. This data source also contains a large-scale concept graph and related academic papers as additional resources for further utilization. We selected student video interaction data, students’ personal characteristics text, videos, and their corresponding courses’ KCs and KCs’ text. The student video interaction data includes video ID, student ID, the number of times students watch the video, the video duration, and the average length of time students watch the video, and the student’s personal characteristics text encompasses the name of the student, the title of the course, etc.; the KC’s text relates to its name and information summary.
To analyze the effect of the proposed method on courses in the same discipline and interdisciplinary courses, we select part of the course data from MOOPer and MOOC-Cube based on the discipline division to form three experiment datasets.
  • Dataset I—All belong to the same discipline
    Course A comes from MOOPer’s course “C/C++ programming”. We randomly selected 20,500 interactive data items for it, including 1224 students and 26 KCs.
    Course B comes from MOOPer’s course “Data Structures & Algorithms (C/C++)”. We randomly selected 55,578 interactive data items for it, including 2925 students and 134 KCs.
    Course C comes from MOOPer’s course “Introduction to Artificial Intelligence”. We randomly selected 12,048 interactive data items for it, including 679 students and 91 KCs.
  • Dataset II—From basic sciences to social sciences
    Course a comes from the MOOC-Cube course “College Physics—Mechanics and Thermal”. We randomly selected 68,131 interactive data items for it, including 5459 students and 36 KCs;
    Course b comes from the MOOC-Cube course “College Physics—Vibration, Waves and Optics”. We randomly selected 7350 interactive data items for it, including 516 students and 36 KCs.
    Course c comes from the MOOC-Cube course “Exhibition Planning and Management”. We randomly selected 103,973 interactive data items for it, including 16,256 students, and 119 KCs.
  • Dataset III—From social sciences to basic sciences
    Course x comes from the MOOC-Cube course “History as a Mirror” (Chineses name is “Zi Zhi Tong Jian”). We randomly selected 94,807 interactive data items for it, including 4138 students and 56 KCs.
    Course y comes from the MOOC-Cube course “Rise and Fall of the Tang Dynasty”. We randomly selected 23,110 interactive data items for it, including 1230 students and 47 KCs.
    Course z comes from the MOOC-Cube course “College Physics-Electromagnetism”. We randomly selected 18,422 interactive data items for it, including 969 students and 85 KCs.
Table 2 provides a summary of the experimental datasets.
Table 2. Summary of the experiment datasets.
The symbol course A → course B indicates that course A is the source course and course B is the target course. In a cross-course scenario, three courses (course A, B, and C) are divided into two scenarios for the experiment. Each scenario consists of two courses, which alternate as the source and target courses, denoted by ↔. In this way, the three datasets are pre-processed into six cross-course scenarios, .i.e.,
  • Scenario1: Course A ↔ Course B;
  • Scenario2: Course A ↔ Course C;
  • Scenario3: Course a ↔ Course b;
  • Scenario4: Course a ↔ Course c;
  • Scenario5: Course x ↔ Course y;
  • Scenario6: Course x ↔ Course z.
Table 3 provides a detailed overview of these six cross-course KT scenarios along with the corresponding dataset information.
Table 3. Statistics of six cross-course KT scenarios.

5.2. Experiment Setup

We obtain E ( g ) t n using Equation (5), which is used to further generate cross-course representations, and obtain E ( x ) t n from Equation (6), which is used to input the LSTM for modeling the students’ mastery of the KCs. Furthermore, we obtain the data of the common students in the two courses. In the Course graph construction module, the Stanford CoreNLP [28] is used to compress the text information; when constructing the HGC of the courses, we set the sampling probability parameter to 0.05, and the values of the students’ mastery of KCs are all mapped to [0,1].
In the Cross-course embedding combination module, the length of the students’ representation vector is set to 8, 16, 32, 64, and 128, as a comparative experiment. The learning rate l r of the cross-domain training is set to 0.001, and the training batch size is set to 32. When splicing the training data, we utilize 70% of the distinct students in the source course and the common students in both the source course and target course as the training set, and the remaining 30% of the distinct students in the source course are used as the test set.
In the Knowledge performance prediction module, the max size of the parameter is set to 200. The learning rate l r is set to 0.001, and the training batch size is set to 8. The dimension of the hidden layer of the model can be either 100 or 50, and the number of hidden layers can be either 1 or 2. Suppose the number of KCs contained in the course corresponding to the model is n and the length of the student representation vector is m; then the dimension of the input data is 3 n + m , and the output one is n. The training adopts the Adam optimizer.

5.3. Baseline Methods

As far as we know, most existing KT models lack cross-domain capabilities. To better demonstrate the effectiveness of MCKT, we selected four mainstream KT models from recent years as baselines and made them have cross-domain capabilities through some methods. The specific introductions of the baselines are as follows:
  • Original: After training on the source course data, the model is directly applied to the target course for prediction without cross-domain training. We select the best result for all baseline models to ensure a fair comparison.
  • DKT [1]: This is a knowledge tracing model based on an RNN that uses students’ historical answer sequences as input to predict students’ answer results at each time step in the future.
  • DKVMN [10]: This is a knowledge tracing model based on memory networks that employs a memory matrix to store students’ knowledge status and uses an attention mechanism to dynamically update and retrieve information in the memory matrix.
  • FGKT [29]: This is a knowledge tracing model based on matrix factorization, which decomposes the student’s answer result matrix into a student ability matrix and a KC difficulty matrix to infer the student’s knowledge status.
  • CKT [2]: This is an improvement on the DKT model, which uses a KC correlation matrix to model the relationship between students’ answer results and KCs, thereby more accurately predicting students’ knowledge status.
  • SGKT [18]: This is a knowledge tracing model based on the session graph and relationship graph, where the session graph is used to model the students’ answering process and the relationship graph is used to model the relationship between the exercises and KCs.
  • DDTCD-KT [30]: DDTCDR is a cross-domain recommendation model that uses latent orthogonal mapping to extract users’ preferences in multiple domains and retain the relationship between users in different latent spaces and combines the autoencoder method to extract the latent essence of feature information. We adapt it to the knowledge tracing task using a target-value mapping transformation (i.e., matching the user’s rating of the items to the right or wrong values of the student’s answers to the exercises) and name it DDTCD-KT.
We use the experiment settings in AdaptKT, which adds fine-tuning operations to each baseline model, endowing them with certain cross-domain capabilities. Therefore, in our experiment, the learning record data of the common students in the two courses is added to the training data of each baseline model to achieve the cross-course function of knowledge tracing. Furthermore, our model also incorporates student representations obtained from cross-course training, while other baseline models do not.

5.4. Evaluation Metrics

In the knowledge tracing dataset, 1 or 0 is usually used to indicate whether a student answered an exercise correctly or not. However, watching videos is the most common learning behavior for course learning in MOOCs. In view of this, we calculate the students’ performance on KCs based on the interactive indicator records of the students watching the videos in scenarios 3–6; and in scenarios 1–2, the students’ performance on KCs is still represented by their correct or incorrect answers to the exercises. In the experiments comparing the baseline models, to maintain the consistency of the evaluation indicators and data format, we still use 0 or 1 to indicate whether students have mastered KCs or not.
In subsequent ablation experiments and self-control experiments, we used probability to indicate the students’ performance on KCs and adjusted the evaluation indicators accordingly. For the case of using 0 and 1, we use AUC, F1, recall, and precision as the evaluation indicators. AUC is used to measure the performance of the classifier, where the closer the value is to 1, the better the performance. The F1 score comprehensively considers the precision rate and recall rate, where the closer the value is to 1, the better the overall performance is. Recall measures the number of positive examples correctly predicted by the model as a proportion of all true positive examples. Precision measures the proportion of the number of positive examples that were correctly predicted by the model out of all the samples predicted as positive examples.
When using probability values, we adopt MAE, RMSE, and R-squared as the evaluation indicators. MAE represents the average absolute difference between the predicted value and the actual value, with a smaller value indicating a lower prediction error; RMSE represents the square root of the mean square error between the predicted value and the true value, where the smaller the value, the smaller the prediction error; and R-squared measures the degree to which the model explains the dependent variable, with a value ranging from 0 to 1, where the closer the value is to 1, the stronger the explanatory power.

5.5. Performance Comparison Experiment (RQ1)

The results of the performance comparison experiments on the six scenarios are shown in Table 4, Table 5 and Table 6. The original model performs worst because it does not perform any cross-course transfer operations. Moreover, the model trained from the source course did not play a positive role in the target course, with an AUC of about 0.5, which is close to random estimation. After adding the alternating training operation using some of the source course students’ data and the common students’ data, the AUC of +co models exceeds 0.65, with the best results of the +co models underlined. However, the AUC of DDTCD-KT and the proposed MCKT are near or over 0.75, and the other performance indicators have improved too. These results also prove that the method of alternating training with rich training set content can enhance the effect of cross-course knowledge tracing.
Table 4. Comparison of the results for scenarios 1 and 2.
Table 5. Comparison of the results for scenarios 3 and 4.
Table 6. Comparison of the results for scenarios 5 and 6.

5.6. Ablation Study (RQ2)

The Course graph construction and the Cross-course embedding combination are two core modules of the proposed approach, where the use of students’ cross-course representations (ccrs) and the use of common students’ data to train the prediction layer (cst) are two important parts. The ablation study evaluates the utility of these two parts. In particular, ablation studies were conducted on the cross-course representation of students and the effectiveness of training using data from the common students. They consist of the following three variants:
  • MCKT-ori Removing students’ cross-course representations, and the prediction module is not trained by using the learning record data of common students.
  • MCKT-cst Removing learning record data from common students’ data when training the prediction module.
  • MCKT-ccr Removing students’ cross-course representations.
Removing student representations after cross-domain training. In the results in Table 7, Table 8 and Table 9, we remove the original spliced student representations and only use the data of students watching videos for training, and the AUC is about 68%. Although this is not as good as the training effect of the complete model, it is still slightly higher than DKT, CKT, FGKT, and DKVMN after adding alternate training operations. This is attributed to MCKT-cst’s continued utilization of the number of times students watch videos as auxiliary training data, which proves, from the perspective of the knowledge tracing effect, that the splicing of the number of times students watch videos has a positive effect on the performance of this knowledge tracing model.
Table 7. Ablation results for scenarios 1–2.
Table 8. Ablation results for scenarios 3–4.
Table 9. Ablation results for scenarios 5–6.
Remove learning data training of common students in the source course. Judging by the results in Table 7, Table 8 and Table 9, if the step of alternating training using common student data is removed, the accuracy reduces to only about 61%. The performance of this model in cross-course transfer is even worse than that of DKT, CKT, FGKT, and DKT, which have additional alternate training steps. DKVMN is only slightly better than the original model in cross-course transfer. This is because the baseline model that adds alternate training does not have a complete student representation vector, but it also learns certain characteristics of the target course through co-students, while the model that removes the alternate training step, although it has a complete student representation, still has the training data limited to a single course, which does not allow it to learn any features of the target course. Therefore, when the model is transferred to the target course, the features learned in the source course are not well utilized.

5.7. Impact of Course Similarity on Model Performance (RQ3)

From the results of the ablation study, we know that if the transfer training step is removed, the model will not perform well for cross-course transfers due to the lack of data from the common students. To test the effect of the similarity of course content on the transfer performance of the model, we carry out related experiments on dataset I, which includes course A, course B, and course C from MOOPer. Course A and course B are both programming courses, and the content is very similar. Course C is an introductory course and its content is significantly different from course A and course B. From the results shown in Figure 3, we found that models that use data from the source courses as the primary training data achieve better prediction performance in the target course when the content of the target course is similar to that of the source course compared to two courses with significantly different content. Furthermore, when the transfer training steps are removed, the transfer effect of course A and course B will be slightly better than that of course A and course C.
Figure 3. Comparison of the prediction performance between two similar courses (i.e., courses A and B) and two significantly different courses (i.e., courses A and C) in scenario 1 for all models. (a) Transfer from course A to courses B and C; (b) transfer from courses B and C to course A. The results show that all the models have better prediction performance for two similar courses than two significantly different courses.
To further verify this conclusion, we conducted an additional experiment on dataset II and dataset III (see Section 5.1 for details). The goal of this experiment is to quantitatively analyze the mechanism through which course similarity influences transfer effectiveness, rather than to compare classification performance directly. Therefore, we employ MAE, RMSE, and R-squared as evaluation metrics. These metrics are suited to measure the continuous error in the model’s predicted probabilities in the target course and to elucidate the relationship between this error and course similarity. As shown in Figure 4, course a and course b (and similarly, courses x and y) represent similar course pairs, whereas course a and course c (and courses x and z) represent dissimilar pairs. Notably, when the transfer training step is omitted, the transfer effect between the similar courses (a and b) remains marginally better than that between the dissimilar ones (a and c). This result substantiates that the content and nature of the courses have a significant impact on the model’s transfer effect.
Figure 4. Comparison of cross-course prediction performance on the group of similar courses and the group of significantly different courses on datasets II and III. (a) MAE; (b) RMSE; and (c) R-squared. The results also show that models have better prediction performance between two similar courses than between two significantly different courses.

5.8. Cross-Course Performance Evolution Description (RQ4)

The evolution description of learning performance is an important application of the knowledge tracing model, which is a visual way to dynamically trace students’ performance. This task usually employs a heat-map to show the student’s level of performance. The cross-course performance evolution description also uses a heat-map to show the strength and weakness of each student’s mastery level of each KC on the source course and the target course.
Figure 5 shows the heat-map of the students in scenario 1 whose learning performance is constantly changing on certain KCs. Figure 5 comprises three sub-figures, which show the performance evolution of the same student in source course A and two target courses, i.e., course B and course C, respectively.
Figure 5. Visualization of a common student’s knowledge tracing. (a) MOOPer course A; (b) MOOPer course B; and (c) MOOPer course C.
For each sub-figure in Figure 5, the label in the vertical dimension denotes KCs. The horizontal dimension shows a sequence of KCs and corresponding answers (0 indicates incorrectly answered and 1 indicates correctly answered) from the source course, where the labels refer to the KCs input into the model at each time step. The heat-map color indicates the predicted probability that the student will demonstrate the corresponding KCs correctly in the next step. We observed the evolution of the predicted values for the five KCs in source course A. As Figure 5a shows, after answering the 20 exercises, the student improves these five KCs. Furthermore, Figure 5b shows that the prediction on KC 58 in target course B also improves as the learning progresses. However, as Figure 5c shows, the student masters KCs 28 and 20 but fails to master KCs 7 and 4, and there is no noticeable improvement on KC 46.

5.9. Visualization of Student Embedding (RQ5)

We employ T-SNE to visualize the representations of common students and distinct students in six courses after dimensionality reduction, as shown in Figure 6, where the area of the two color scatters in each figure reflects the number of students in each course.
Figure 6. Visualization of common and distinct students’ embeddings in the different courses. (a) MOOPer course A; (b) MOOPer course B; (c) MOOC-Cube course a; (d) MOOC-Cube course b; (e) MOOC-Cube course x; and (f) MOOC-Cube course y.
For instance, Figure 6a,b show the difference between the MOOPer courses A and B. The relative positions of the representations of the common students and the distinct students are similar in the low-dimensional space; that is, the representations of the common students are mostly distributed on the left side of the figure, and the representations of the distinct students are mostly distributed on the right side of the figure. Figure 6c,e show that in course a and course x of MOOC-Cube, the representation distributions of the common and distinct students are relatively clustered. In addition, Figure 6d,f show that in course b and course y of MOOC-Cube, the number of common students is much higher than that of distinct students, but the representation of the distinct students is also relatively concentrated. However, Figure 6d,f also show that the representation distribution of the distinct students overlaps more with that of the common students. A reason may be that the distinct student representations of the target course (i.e., course b and course y) are more affected by the common students when training in the cross-domain stage.

6. Conclusions and Future Work

In this paper, we propose a cross-course knowledge tracing model. Through comparative experiments with different KT models, the superiority of MCKT in performance has been demonstrated. Of course, our method still has some limitations. After enhancing the model’s cross-course learning capability, its predictive performance can be somewhat affected by information from other courses, making it difficult to surpass single-target knowledge tracing. Our next objective is to trace these effects and develop methods to filter out “noise” introduced by other courses. Here, “noise” refers not to data errors, but to irrelevant information originating from source courses, which can hinder or mislead the prediction of knowledge states in the target course during cross-domain transfer—a phenomenon commonly termed “negative transfer”. Currently, our graph embedding layer utilizes the Doc2Vec model to obtain textual representations. However, our future work will prioritize architectural simplification and efficiency. Instead of pre-training methods, we will shift our focus towards exploring end-to-end trainable and lightweight alternatives. This includes investigating unified architectures, such as fine-tuning pre-trained language models or employing graph neural networks capable of jointly processing textual and structural information in a differentiable manner. Similarly, for graph representation learning, rather than simply substituting Node2Vec with other similar sampling-based methods, we plan to explore integrating trainable graph encoding layers that can be optimized jointly with the sequential model, thereby moving towards a more streamlined, efficient pipeline. The proposed MCKT model still carries potential risks in practical applications, primarily manifested in external threats from the following two aspects: First, the multi-stage pipeline—relying on separate Doc2Vec and Node2Vec modules—creates a dependency on pre-trained, fixed representations. The performance is sensitive to the parameters of these independent stages , and any noise or bias introduced during graph construction or text processing may propagate through the model. Second, the specific architecture of the fusion MLP and the LSTM is one of many possible configurations; while ablation studies validate our design, the causal link between the fusion mechanism itself and the performance gain could be further isolated from the specific hyperparameters selected. The main external threat comes from the applicability of the courses. Our evaluation was conducted on specific MOOC datasets. Although we selected courses from different disciplines such as IT, physics, history, and management, it is ultimately impossible to cover all disciplines. In addition, the computational complexity of the current process poses a significant threat to its feasibility in real-time educational environments, where low-latency predictions are usually required.

Author Contributions

The study was conceived and designed by D.L. and Z.W., and Y.L. was responsible for the experiments, data collection, and analysis. Y.L. and Z.W. jointly drafted the manuscript. Z.W. provided critical feedback and supervised the research work. D.L. revised the manuscript to ensure the accuracy of important content. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Project of Philosophy and Social Science Planning of Guangdong Province (No. GD25CJY40), and an industry-sponsored horizontal project (No. 2541503).

Data Availability Statement

Suggested Data Availability Statements are available at https://github.com/THU-KEG/MOOCCubeX?tab=readme-ov-file (accessed on 15 October 2025), and http://data.openkg.cn/dataset/mooper (accessed on 18 October 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Piech, C.; Bassen, J.; Huang, J.; Ganguli, S.; Sahami, M.; Guibas, L.J.; Sohl-Dickstein, J. Deep knowledge tracing. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 505–513. [Google Scholar]
  2. Shen, S.; Liu, Q.; Chen, E.; Wu, H.; Huang, Z.; Zhao, W.; Su, Y.; Ma, H.; Wang, S. Convolutional Knowledge Tracing: Modeling Individualization in Student Learning Process. In Proceedings of the the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 25–30 July 2020; pp. 1857–1860. [Google Scholar]
  3. Pu, S.; Yudelson, M.; Ou, L.; Huang, Y. Deep Knowledge Tracing with Transformers. In Proceedings of the Artificial Intelligence in Education, Ifrane, Morocco, 6–10 July 2020; pp. 252–256. [Google Scholar]
  4. Yang, Y.; Shen, J.; Qu, Y.; Liu, Y.; Wang, K.; Zhu, Y.; Zhang, W.; Yu, Y. GIKT: A Graph-Based Interaction Model for Knowledge Tracing. In Proceedings of the Machine Learning and Knowledge Discovery in Databases—European Conference, ECML PKDD 2020, Ghent, Belgium, 14–18 September 2020; Volume 12457, pp. 299–315. [Google Scholar]
  5. Feng, W.; Tang, J.; Liu, T.X. Understanding Dropouts in MOOCs. In Proceedings of the The Thirty-Third AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 517–524. [Google Scholar]
  6. Corbett, A.T.; Anderson, J.R. Knowledge tracing: Modeling the acquisition of procedural knowledge. User Model.-User-Adapt. Interact. 1994, 4, 253–278. [Google Scholar] [CrossRef]
  7. Lin, C.; Chi, M. Intervention-bkt: Incorporating instructional interventions into bayesian knowledge tracing. In Proceedings of the International Conference on Intelligent Tutoring Systems, Zagreb, Croatia, 7–10 June 2016; pp. 208–218. [Google Scholar]
  8. Wang, L.; Sy, A.; Liu, L.; Piech, C. Deep knowledge tracing on programming exercises. In Proceedings of the Fourth ACM Conference on Learning@ Scale, Cambridge, MA, USA, 20–21 April 2017; pp. 201–204. [Google Scholar]
  9. Mongkhonvanit, K.; Kanopka, K.; Lang, D. Deep Knowledge Tracing and Engagement with MOOCs. In Proceedings of the 9th International Conference on Learning Analytics and Knowledge, Tempe, AZ, USA, 4–8 March 2019; pp. 340–342. [Google Scholar]
  10. Zhang, J.; Shi, X.; King, I.; Yeung, D. Dynamic Key-Value Memory Networks for Knowledge Tracing. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 765–774. [Google Scholar]
  11. Jiang, B.; Wei, Y.; Zhang, T.; Zhang, W. Improving the performance and explainability of knowledge tracing via Markov blanket. Inf. Process. Manag. 2024, 61, 103620. [Google Scholar] [CrossRef]
  12. Shukla, P.K.; Veerasamy, B.D.; Alduaiji, N.; Addula, S.R.; Sharma, S.; Shukla, P.K. Encoder only attention-guided transformer framework for accurate and explainable social media fake profile detection. Peer Peer Netw. Appl. 2025, 18, 232. [Google Scholar] [CrossRef]
  13. Wu, Z.; Liu, Y.; Cen, J.; Zheng, Z.; Xu, G. A cross-domain knowledge tracing model based on graph optimal transport. World Wide Web 2025, 28, 10. [Google Scholar] [CrossRef]
  14. Cheng, S.; Liu, Q.; Chen, E.; Zhang, K.; Huang, Z.; Yin, Y.; Huang, X.; Su, Y. AdaptKT: A Domain Adaptable Method for Knowledge Tracing. In Proceedings of the WSDM ’22: The Fifteenth ACM International Conference on Web Search and Data Mining, Virtual, 21–25 February 2022; pp. 123–131. [Google Scholar]
  15. Wilson, G.; Cook, D.J. A Survey of Unsupervised Deep Domain Adaptation. ACM Trans. Intell. Syst. Technol. 2020, 11, 51. [Google Scholar] [CrossRef] [PubMed]
  16. Huang, Y.; Huang, W.; Tong, S.; Huang, Z.; Liu, Q.; Chen, E.; Ma, J.; Wan, L.; Wang, S. STAN: Adversarial Network for Cross-domain Question Difficulty Prediction. In Proceedings of the IEEE International Conference on Data Mining, ICDM 2021, Auckland, New Zealand, 7–10 December 2021; pp. 220–229. [Google Scholar]
  17. Wu, Z.; Liang, Q.; Zhan, Z. Course Recommendation Based on Enhancement of Meta-Path Embedding in Heterogeneous Graph. Appl. Sci. 2023, 13, 2404. [Google Scholar] [CrossRef]
  18. Wu, Z.; Huang, L.; Huang, Q.; Huang, C.; Tang, Y. SGKT: Session graph-based knowledge tracing for student performance prediction. Expert Syst. Appl. 2022, 206, 117681. [Google Scholar] [CrossRef]
  19. Liu, A.; Wei, Y.; Xiu, Q.; Yao, H.; Liu, J. How Learning Time Allocation Make Sense on Secondary School Students’ Academic Performance: A Chinese Evidence Based on PISA 2018. Behav. Sci. 2023, 13, 237. [Google Scholar] [CrossRef] [PubMed]
  20. Ozyildirim, G. Time Spent on Homework and Academic Achievement: A Meta-analysis Study Related to Results of TIMSS. Psicol. Educ. 2022, 28, 13–21. [Google Scholar] [CrossRef]
  21. Le, Q.; Mikolov, T. Distributed Representations of Sentences and Documents. In Proceedings of the the 31st International Conference on International Conference on Machine Learning, ICML’14, Beijing, China, 21–26 June 2014; Volume 32, pp. 1188–1196. [Google Scholar]
  22. Grover, A.; Leskovec, J. node2vec: Scalable Feature Learning for Networks. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 855–864. [Google Scholar]
  23. Jansen, B.J.; Jung, S.; Salminen, J. Finetuning Analytics Information Systems for a Better Understanding of Users: Evidence of Personification Bias on Multiple Digital Channels. Inf. Syst. Front. 2024, 26, 775–798. [Google Scholar] [CrossRef]
  24. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar] [CrossRef]
  25. Dengsheng, C.; Li, J.; Xu, K. AReLU: Attention-based Rectified Linear Unit. arXiv 2020, arXiv:2006.13858. [Google Scholar]
  26. Liu, K.; Zhao, X.; Tang, J.; Zeng, W.; Liao, J.; Tian, F.; Zheng, Q.; Huang, J.; Dai, A. MOOPer: A Large-Scale Dataset of Practice-Oriented Online Learning. In Proceedings of the Knowledge Graph and Semantic Computing: Knowledge Graph Empowers New Infrastructure Construction—6th China Conference, CCKS 2021, Guangzhou, China, 4–7 November 2021; Volume 1466, pp. 281–287. [Google Scholar]
  27. Yu, J.; Luo, G.; Xiao, T.; Zhong, Q.; Wang, Y.; Feng, W.; Luo, J.; Wang, C.; Hou, L.; Li, J.; et al. MOOCCube: A Large-scale Data Repository for NLP Applications in MOOCs. In Proceedings of the The 58th Annual Meeting of the Association for Computational Linguistics, ACL 2020, Online, 5–10 July 2020; pp. 3135–3142. [Google Scholar]
  28. Manning, C.D.; Surdeanu, M.; Bauer, J.; Finkel, J.R.; Bethard, S.; McClosky, D. The Stanford CoreNLP Natural Language Processing Toolkit. In Proceedings of the the 52nd Annual Meeting of the Association for Computational Linguistics, ACL 2014, Baltimore, MD, USA, 22–27 June 2014; pp. 55–60. [Google Scholar]
  29. Chen, Y.; Liu, Q.; Huang, Z.; Wu, L.; Chen, E.; Wu, R.; Su, Y.; Hu, G. Tracking Knowledge Proficiency of Students with Educational Priors. In Proceedings of the the 2017 ACM on Conference on Information and Knowledge Management, Singapore, 6–10 November 2017; pp. 989–998. [Google Scholar]
  30. Li, P.; Tuzhilin, A. DDTCDR: Deep Dual Transfer Cross Domain Recommendation. In Proceedings of the WSDM ’20: The Thirteenth ACM International Conference on Web Search and Data Mining, Houston, TX, USA, 3–7 February 2020; pp. 331–339. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Article metric data becomes available approximately 24 hours after publication online.