Next Article in Journal
Analysis and Optimization for the Sealing Performance of Ultra-High Pressure Solenoid Valves in Low-Temperature Environments
Previous Article in Journal
Developing Educational Software Models for Teaching Cyclic Codes in Coding Theory
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interpretable Knowledge Tracing via Transformer-Bayesian Hybrid Networks: Learning Temporal Dependencies and Causal Structures in Educational Data

1
Rossier School of Education, University of Southern California, Los Angeles, CA 90007, USA
2
School of Computer Science, Carnegie Mellon University, Pittsburgh, PA 15213, USA
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(17), 9605; https://doi.org/10.3390/app15179605
Submission received: 26 June 2025 / Revised: 26 August 2025 / Accepted: 27 August 2025 / Published: 31 August 2025

Abstract

Knowledge tracing, the computational modeling of student learning progression through sequential educational interactions, represents a critical component for adaptive learning systems and personalized education platforms. However, existing approaches face a fundamental trade-off between predictive accuracy and interpretability: deep sequence models excel at capturing complex temporal dependencies in student interaction data but lack transparency in their decision-making processes, while probabilistic graphical models provide interpretable causal relationships but struggle with the complexity of real-world educational sequences. We propose a hybrid architecture that integrates transformer-based sequence modeling with structured Bayesian causal networks to overcome this limitation. Our dual-pathway design employs a transformer encoder to capture complex temporal patterns in student interaction sequences, while a differentiable Bayesian network explicitly models prerequisite relationships between knowledge components. These pathways are unified through a cross-attention mechanism that enables bidirectional information flow between temporal representations and causal structures. We introduce a joint training objective that simultaneously optimizes sequence prediction accuracy and causal graph consistency, ensuring learned temporal patterns align with interpretable domain knowledge. The model undergoes pre-training on 3.2 million student–problem interactions from diverse MOOCs to establish foundational representations, followed by domain-specific fine-tuning. Comprehensive experiments across mathematics, computer science, and language learning demonstrate substantial improvements: 8.7% increase in AUC over state-of-the-art knowledge tracing models (0.847 vs. 0.779), 12.3% reduction in RMSE for performance prediction, and 89.2% accuracy in discovering expert-validated prerequisite relationships. The model achieves a 0.763 F1-score for early at-risk student identification, outperforming baselines by 15.4%. This work demonstrates that sophisticated temporal modeling and interpretable causal reasoning can be effectively unified for educational applications.

1. Introduction

Knowledge tracing, the computational modeling of student learning progression through sequential educational interactions, has emerged as a fundamental component of adaptive learning systems and personalized education platforms [1,2]. Traditional educational assessment methods rely on discrete, subjective evaluations, but modern digital learning environments generate rich interaction logs that enable continuous modeling of knowledge acquisition and skill development through detailed clickstream data, response patterns, and temporal learning behaviors [3,4,5]. Knowledge tracing systems analyze these sequential interactions to assess student mastery of specific skills and predict future performance, enabling personalized content delivery and timely educational interventions [6].
The proliferation of digital educational technologies has created new opportunities for understanding learning dynamics through objective measurement of student behaviors and cognitive processes captured in interaction logs. Unlike traditional assessment approaches, knowledge tracing leverages continuous data streams from educational platforms to construct comprehensive models of knowledge acquisition and skill development. Digital learning systems record detailed interaction patterns, response times, and navigation behaviors that reveal learning strategies and knowledge gaps. The integration of these diverse data streams through advanced modeling techniques presents both opportunities and challenges for developing interpretable, accurate models of student learning progression that can inform pedagogical decision-making and adaptive system design.
Modern educational environments increasingly incorporate sophisticated sensor networks that capture multi-dimensional learning indicators beyond traditional clickstream data. Contemporary learning platforms integrate physiological sensors monitoring cognitive load through heart rate variability and galvanic skin response, eye-tracking systems capturing attention allocation and reading patterns, accelerometers detecting engagement levels through postural changes and device interaction dynamics, and ambient environmental sensors measuring factors such as lighting conditions, noise levels, and temperature that influence learning effectiveness. These sensor-rich educational ecosystems generate heterogeneous data streams with varying temporal resolutions, measurement scales, and semantic meanings that require sophisticated integration techniques to extract meaningful learning insights. The challenge extends beyond simple data fusion to encompass temporal synchronization across sensor modalities, robust handling of missing or noisy sensor measurements, and computational efficiency requirements for real-time processing in Internet-of-Things educational deployments.
The complexity of sensor-driven educational data presents unique opportunities for knowledge tracing systems that can effectively process multi-modal temporal sequences while maintaining interpretable causal reasoning about learning processes. Physiological sensor streams provide continuous indicators of cognitive state changes that correlate with knowledge acquisition events, while behavioral sensors capture fine-grained interaction patterns that reveal learning strategies and concept mastery progression. Environmental sensor data enables contextualization of learning performance within situational factors, allowing knowledge tracing models to account for external influences on cognitive performance and learning outcomes. However, effectively leveraging these rich sensor inputs requires architectural innovations that can capture complex temporal dependencies across multiple sensing modalities while preserving the interpretability crucial for educational stakeholders to understand how different sensor-detected behaviors and states contribute to learning assessment and intervention decisions.
Traditional approaches to knowledge tracing, such as Bayesian Knowledge Tracing (BKT) [6] and Item Response Theory (IRT) [7], were designed for environments with limited, structured data inputs and provide interpretable parameters representing knowledge acquisition, retention, and skill transfer. However, these parametric models lack the representational capacity to capture the complex temporal dependencies inherent in rich educational interaction sequences [8]. The challenge lies in developing computational frameworks that can effectively process heterogeneous educational data while maintaining the interpretability crucial for educational stakeholders who need transparent insights into learning processes for intervention design and pedagogical decision-making.
The advent of deep learning has enabled significant advances in processing complex sequential data for knowledge tracing applications. Deep Knowledge Tracing (DKT) [9] pioneered the application of recurrent neural networks to model temporal patterns in educational interactions, demonstrating superior performance in environments where traditional methods fail to capture the complexity of multi-dimensional interaction sequences. Subsequent developments have incorporated attention mechanisms [10], memory-augmented networks [11], and graph neural networks [12] to better handle the heterogeneous nature of educational data and model intricate relationships between different interaction modalities, temporal progression patterns, and learning concepts.
However, contemporary knowledge-tracing systems face a fundamental limitation: the inverse relationship between model complexity and interpretability when processing educational interaction data. Deep learning models excel at extracting complex patterns from heterogeneous data and capturing temporal dependencies across multiple interaction dimensions but operate as black boxes that provide limited insight into how different behavioral patterns contribute to learning state estimation [13]. This opacity poses significant challenges for educational practitioners who require transparent, actionable insights to understand how student behaviors translate to learning outcomes and inform intervention strategies. Conversely, interpretable probabilistic models often lack the representational capacity to effectively process and integrate the diverse, high-dimensional data streams characteristic of modern digital learning environments.
Transformer architectures, first introduced by Vaswani et al. [14], have demonstrated remarkable success in modeling sequential data across diverse applications, offering powerful mechanisms for capturing long-range dependencies and contextual relationships in complex sequences. Recent advances in transformer-based educational applications have shown significant promise, including state-of-the-art knowledge tracing models that leverage pre-trained language representations [15], attention-based student modeling frameworks for personalized learning [16], and multimodal transformer architectures that integrate diverse educational data streams [17]. In educational contexts, contemporary transformer implementations have achieved breakthrough performance in automated assessment of learning activities [18] and personalized learning resource recommendation based on interaction patterns [19]. However, their application to interpretable knowledge tracing remains underexplored, particularly in conjunction with structured probabilistic models that can provide interpretable causal relationships between observed behaviors and learning concept mastery.
Bayesian causal networks offer a principled framework for modeling prerequisite relationships and knowledge dependencies while incorporating evidence about learning states from educational interactions [20,21]. These models can explicitly represent expert knowledge about concept hierarchies and learning progressions while integrating real-time observations about student engagement, performance patterns, and learning behaviors to provide interpretable reasoning about mastery states and optimal learning sequences [22]. However, traditional Bayesian approaches often rely on predefined network structures that may not adequately capture the dynamic, personalized nature of learning processes or effectively utilize the rich temporal patterns available in educational interaction data.
This work addresses the fundamental tension between predictive accuracy and interpretability in knowledge tracing by proposing a novel hybrid architecture that synergistically combines transformer-based sequence modeling with structured Bayesian causal networks for processing educational interaction data. Our method represents a paradigm shift from traditional single-pathway models by introducing a unified framework that preserves the strengths of both deep learning and probabilistic modeling paradigms. The framework leverages the superior temporal modeling capabilities of transformers to capture complex sequential patterns and long-range dependencies in educational interaction sequences, while simultaneously employing differentiable Bayesian networks to learn interpretable causal structures that explicitly model prerequisite relationships and knowledge dependencies. The architecture integrates these complementary modeling approaches through a sophisticated cross-attention mechanism that enables mutual information exchange between temporal insights and pedagogical causal structures, allowing interaction patterns to inform causal relationship discovery while prerequisite knowledge guides sequence interpretation.
While our architectural design accommodates diverse input modalities through flexible embedding schemes that could potentially handle multimodal educational data, the current experimental evaluation focuses on demonstrating the effectiveness of our temporal-causal integration methodology using established benchmark datasets. This approach enables direct comparison with existing knowledge tracing methods while validating our core contribution of unified temporal and causal modeling for interpretable knowledge tracing.
The key innovations and contributions of our framework encompass both architectural and methodological advances that collectively enable interpretable yet accurate knowledge tracing in digital educational environments:
  • We introduce a novel parallel processing framework that simultaneously handles temporal sequence modeling of educational interactions through transformer encoders and causal structure learning via differentiable Bayesian networks, enabling capture of both sequential dependencies and interpretable prerequisite relationships without architectural compromise.
  • Our framework incorporates a sophisticated bidirectional attention mechanism that facilitates dynamic information exchange between temporal representations and causal knowledge structures, allowing interaction-derived insights to inform causal graph construction while prerequisite knowledge guides sequence interpretation.
  • We develop a unified training objective that simultaneously optimizes sequence prediction accuracy from educational interactions and causal graph consistency through a multi-task learning paradigm, ensuring that learned temporal patterns remain aligned with interpretable domain knowledge and educational theory.
  • Our method implements a comprehensive two-stage learning protocol that first establishes robust foundational representations through large-scale educational data pre-training, followed by domain-specific fine-tuning that preserves interpretability while adapting to particular educational contexts and domains.
Our extensive experimental evaluation across multiple educational domains demonstrates that this hybrid method achieves state-of-the-art predictive performance while providing interpretable insights into learning pathways and prerequisite relationships. The learned models not only predict student performance with high accuracy but also discover meaningful causal structures that align with expert pedagogical knowledge, enabling practical applications such as personalized learning path generation, real-time intervention triggering, and adaptive content delivery optimization.
The remainder of this paper is organized as follows. Section 2 reviews related work in knowledge tracing, transformer architectures for sequential data processing, and Bayesian causal modeling in educational systems. Section 3 presents our hybrid architecture and training methodology. Section 4 describes our experimental setup and results across multiple educational domains. Finally, Section 5 concludes with a discussion of limitations and future research directions.

2. Related Work

This section reviews the extensive literature on sensor-driven knowledge tracing, highlighting the evolution from traditional probabilistic models to modern deep learning approaches capable of processing multi-modal sensor data, and examining recent developments in interpretable machine learning for sensor-enabled educational applications.

2.1. Traditional Knowledge Tracing Methods in Sensor-Limited Environments

Knowledge tracing originated with BKT [6], which models student mastery as a latent binary state governed by four parameters: initial knowledge probability, learning rate, slip probability, and guess probability. BKT employs a Hidden Markov Model (HMM) framework where transitions between knowledge states follow Markovian assumptions based on discrete interaction data. While originally designed for environments with limited sensor inputs, BKT has demonstrated remarkable longevity and continues to serve as a baseline for sensor-enhanced approaches [23]. However, its binary state representation and discrete temporal modeling limit its ability to leverage continuous sensor streams that capture nuanced learning behaviors and cognitive states.
IRT [7] provides an alternative parametric framework that models the probability of correct responses as a function of student ability and item difficulty using traditional assessment data. Extensions such as the Rasch model [24] and two-parameter logistic model [25] have been widely adopted in standardized testing environments where sensor data was historically unavailable. Multidimensional IRT [26] addresses the limitation of unidimensional ability estimation by incorporating multiple latent traits, but lacks mechanisms for integrating real-time sensor observations such as physiological indicators of cognitive load or behavioral patterns detected through interaction sensors.
Performance Factor Analysis (PFA) [27] and Additive Factor Model (AFM) [28] extend traditional approaches by incorporating practice effects and skill-specific learning curves derived from discrete interaction logs. These models explicitly account for the number of opportunities to practice specific skills, providing more nuanced representations of learning progression. However, they remain limited to traditional clickstream data and cannot leverage the rich temporal dynamics available through continuous sensor monitoring of student engagement, attention patterns, and environmental conditions.

2.2. Deep Learning Approaches for Multi-Modal Sensor Processing

The introduction of DKT [9] marked a paradigm shift in the field by applying RNNs to model student learning sequences, demonstrating particular effectiveness in processing heterogeneous data streams common in sensor-rich educational environments. DKT treats knowledge tracing as a sequence modeling problem, using LSTM networks to capture temporal dependencies in multi-dimensional input vectors that can accommodate diverse sensor modalities including physiological signals, behavioral interactions, and environmental measurements. While achieving superior predictive performance in sensor-enabled settings, DKT’s black-box nature raised concerns about interpretability when processing complex sensor fusion inputs [29].
Dynamic Key-Value Memory Networks (DKVMNs) [11] address DKT’s limitations by introducing explicit memory mechanisms that store and update concept-specific knowledge states derived from multi-modal sensor observations. The architecture separates static knowledge concepts from dynamic mastery states, enabling integration of diverse sensor inputs such as eye-tracking data indicating attention allocation and physiological sensors measuring cognitive load. This separation provides improved interpretability while maintaining the capacity to process high-dimensional sensor feature vectors representing complex learning behaviors.
Graph-based approaches have gained prominence for modeling complex relationships between learning concepts and sensor-detected behavioral patterns. Graph-based knowledge tracing [12] models student proficiency using graph convolutions over concept dependency networks, with node features enriched by aggregated sensor measurements that capture real-time learning indicators. Subsequent work has explored various graph neural network architectures for educational applications, particularly focusing on sensor data integration where nodes can represent both knowledge components and sensor modalities, enabling sophisticated reasoning about how physiological states, environmental conditions, and behavioral patterns collectively influence learning outcomes.
Attention mechanisms have proven particularly effective for knowledge tracing in sensor-rich environments where multiple data streams require selective focus and temporal alignment. Self-Attentive Knowledge Tracing (SAKT) [10] applies transformer-style self-attention to capture long-range dependencies in learning sequences, with extensions handling multi-modal sensor inputs through parallel attention heads that specialize in different sensing modalities. SAINT [30] further enhances attention-based models by incorporating positional encoding and exercise-specific embeddings, demonstrating effectiveness in IoT educational settings where sensor data provides additional temporal and contextual information about learning activities.

2.3. Deep Learning Approaches for Sequential Knowledge Tracing

The introduction of DKT [9] marked a paradigm shift by applying RNNs to model student learning sequences, demonstrating effectiveness in processing heterogeneous educational data streams. DKT treats knowledge tracing as a sequence modeling problem, using LSTM networks to capture temporal dependencies in multi-dimensional input vectors. While achieving superior predictive performance, DKT’s black-box nature raised concerns about interpretability [29].
Dynamic Key-Value Memory Networks (DKVMNs) [11] address DKT’s limitations by introducing explicit memory mechanisms that store and update concept-specific knowledge states. The architecture separates static knowledge concepts from dynamic mastery states, providing improved interpretability while maintaining the capacity to process high-dimensional feature vectors.
Graph-based approaches have gained prominence for modeling complex relationships between learning concepts. Graph-based knowledge tracing [12] models student proficiency using graph convolutions over concept dependency networks. Subsequent work has explored various graph neural network architectures, enabling sophisticated reasoning about how behavioral patterns collectively influence learning outcomes.

2.4. Transformer Applications in Sensor-Enabled Educational Settings

Transformer architectures have demonstrated remarkable success in knowledge tracing applications. Self-Attentive Knowledge Tracing (SAKT) [10] pioneered transformer-style self-attention for capturing long-range dependencies in learning sequences. Building on this foundation, deep knowledge tracing with transformers [31] introduced comprehensive transformer encoder–decoder architectures specifically designed for educational sequence modeling, achieving significant performance improvements through multi-head attention mechanisms that capture diverse temporal relationships in student interactions.
SAINT [30] further enhanced attention-based models by incorporating positional encoding and exercise-specific embeddings in an encoder–decoder framework. Recent work by Li et al. [32] proposed deep knowledge tracing with evolved transformer structure, introducing architectural innovations such as adaptive attention mechanisms and dynamic layer normalization that automatically adjust to varying sequence complexities in educational data.
Advanced transformer variants have continued to push performance boundaries. AKT [33] incorporates context-aware attention mechanisms, while simpleKT [34] demonstrates that streamlined transformer architectures can achieve competitive performance through focused attention design. These developments collectively establish transformers as the dominant paradigm for sequential knowledge tracing.

2.5. Large Language Model Integration in Knowledge Tracing

The emergence of large language models has revolutionized knowledge tracing through sophisticated representation learning and contextual understanding capabilities. Early work by Wang et al. [35] pioneered LLM-KT, integrating pre-trained BERT embeddings to capture semantic relationships between educational content and student responses, demonstrating significant improvements in cross-domain knowledge transfer.
Recent comprehensive surveys [36] identify multiple taxonomic approaches for LLM integration in knowledge tracing. Fine-tuning strategies adapt pre-trained models to educational domains, with methods ranging from full parameter updates to parameter-efficient techniques like LoRA and prompt tuning [37]. Knowledge distillation approaches transfer semantic understanding from large models to efficient student models suitable for real-time educational applications [38].
Cross-domain transfer learning leverages LLMs’ broad knowledge for educational applications. Recent work explores zero-shot knowledge tracing [39], where models trained on one educational domain generalize to unseen domains through shared semantic representations. Multi-task learning frameworks jointly optimize language understanding and knowledge tracing objectives [40].
However, LLM-based approaches face challenges including computational efficiency for real-time applications, interpretability concerns in educational contexts, and potential bias propagation from pre-training data [41]. These limitations motivate hybrid architectures that combine LLM capabilities with interpretable probabilistic modeling, as proposed in our work.

2.6. Bayesian Networks and Causal Modeling in Education

Bayesian networks provide a principled framework for modeling causal relationships in sensor-enhanced educational domains [42]. Educational applications typically focus on prerequisite modeling, where directed acyclic graphs represent dependencies between learning concepts while incorporating sensor-derived evidence about student states, engagement levels, and environmental factors that influence learning effectiveness [22]. The probabilistic framework naturally accommodates uncertainty in sensor measurements and enables fusion of diverse sensing modalities through principled inference mechanisms.
Causal discovery algorithms have been applied to learn prerequisite structures from educational data enriched with sensor observations. Constraint-based methods such as PC algorithm [43] and score-based approaches like Greedy Equivalence Search [44] enable automatic discovery of concept dependencies while incorporating sensor-detected indicators of mastery, confusion, and engagement as observational evidence. However, these methods often struggle with the noise and high dimensionality characteristic of multi-modal sensor datasets, particularly when dealing with continuous physiological signals and behavioral patterns that require sophisticated preprocessing and feature extraction.
Probabilistic graphical models have been successfully integrated with traditional knowledge tracing in sensor-enabled environments. Bayesian networks for knowledge tracing [45] extends BKT by incorporating explicit prerequisite relationships while enabling integration of sensor observations such as response time patterns, attention tracking data, and physiological indicators of cognitive load. Dynamic Bayesian networks [46] model temporal evolution of knowledge states while preserving interpretable causal structures, with extensions handling continuous sensor streams through particle filtering and sequential Monte Carlo methods that enable real-time inference in IoT educational systems.

2.7. Hybrid and Interpretable Approaches for Sensor Integration

Hybrid approaches attempt to balance accuracy and interpretability while effectively utilizing multi-modal sensor inputs. Interpretable deep knowledge tracing introduces attention mechanisms to provide model insights while maintaining neural expressiveness capable of processing complex sensor feature vectors [10]. These approaches demonstrate particular promise in sensor-rich environments where stakeholders need transparency about how different sensing modalities contribute to learning state estimation and prediction accuracy.
However, most existing methods focus primarily on predictive performance using traditional interaction data, with limited consideration of sensor data integration challenges such as temporal alignment across sensing modalities, handling missing or corrupted sensor measurements, and managing the computational complexity associated with real-time multi-modal processing. The lack of principled frameworks for joint temporal and causal modeling that can effectively leverage diverse sensor inputs represents a significant gap in current sensor-driven knowledge tracing research.

2.8. Gaps and Limitations

Despite significant advances, existing knowledge tracing approaches face fundamental limitations that motivate our hybrid architecture design. Current methods exhibit a critical accuracy–interpretability trade-off where high-performing deep learning models operate as black boxes, while interpretable probabilistic models lack the representational capacity for complex educational sequences.
Deep learning approaches, including transformer-based methods like SAKT and SAINT, achieve superior predictive performance but provide no transparency into how student behaviors contribute to learning state estimation. These black-box models prevent educational stakeholders from understanding which interaction patterns drive predictions, limiting their practical deployment in educational institutions where instructors require actionable insights about specific factors influencing student performance. The opacity becomes particularly problematic when educators need to design targeted interventions based on model predictions, as current deep learning approaches offer no mechanism to trace prediction rationales back to interpretable educational constructs.
Conversely, traditional probabilistic models like BKT and IRT provide interpretable parameters representing knowledge acquisition and skill transfer, but struggle with the complexity and temporal dynamics inherent in modern educational interaction sequences. These parametric approaches cannot effectively capture the long-range dependencies, contextual relationships, and multi-dimensional interaction patterns that characterize rich educational data, resulting in substantial performance limitations compared to deep learning alternatives.
A critical gap exists in principled frameworks for joint temporal and causal modeling in knowledge tracing. Current approaches treat sequence prediction and prerequisite discovery as separate problems, with no existing methods simultaneously optimizing both temporal pattern recognition and interpretable causal structure learning within a unified architecture. This separation prevents temporal insights from informing causal relationship discovery and constrains causal knowledge from guiding sequence interpretation, despite their inherent interconnection in educational contexts where prerequisite relationships directly influence temporal learning progression patterns.
Furthermore, existing hybrid approaches lack effective integration mechanisms between deep learning and probabilistic modeling paradigms. Traditional combination methods simply ensemble predictions from separate models or use one approach to initialize another, failing to enable bidirectional information flow between temporal and causal representations. This limitation prevents the development of architectures that can simultaneously leverage the representational power of deep learning for complex sequence modeling and the interpretability of probabilistic models for causal reasoning.
The absence of differentiable causal discovery methods specifically designed for educational applications represents another significant limitation. While recent work in causal inference has developed differentiable approaches for general domains, their application to knowledge tracing remains underexplored, particularly for learning prerequisite structures that align with pedagogical knowledge while maintaining end-to-end trainability with sequence prediction objectives.
Our proposed hybrid architecture directly addresses these limitations by introducing a unified framework that preserves the temporal modeling strengths of transformers while incorporating interpretable Bayesian causal networks, enabling bidirectional information exchange through cross-attention mechanisms, and jointly optimizing sequence prediction accuracy with causal structure consistency within a single end-to-end trainable architecture.

3. Methodology

This section presents our hybrid architecture that integrates transformer-based sequence modeling with differentiable Bayesian causal networks for interpretable knowledge tracing. We begin with problem formulation, followed by detailed descriptions of the dual-pathway architecture, cross-attention integration mechanism, joint optimization framework, and training strategy.

3.1. Problem Formulation

Let S = { s 1 , s 2 , , s T } represent a student’s interaction sequence, where T denotes the total number of interactions in the sequence. Each interaction s t = ( q t , r t , c t ) consists of three components: q t represents the unique identifier of the question attempted at time step t, r t { 0 , 1 } indicates the binary response, where r t = 1 denotes a correct answer and r t = 0 denotes an incorrect answer, and c t C represents the set of knowledge components (skills or concepts) associated with question q t . The knowledge component space C = { c 1 , c 2 , , c K } represents the complete set of K distinct concepts or skills in the educational domain.
Our objective is to jointly learn: (1) a temporal model f θ ( S 1 : t ) p ( r t + 1 | q t + 1 ) that predicts future performance based on interaction history, where θ represents the learnable parameters of the temporal model, S 1 : t = { s 1 , s 2 , , s t } denotes the interaction history up to time t, and p ( r t + 1 | q t + 1 ) is the predicted probability of correctly answering question q t + 1 ; and (2) a causal structure G = ( C , E ) that captures prerequisite relationships between knowledge components, where E C × C represents the set of directed edges indicating causal dependencies, with each edge ( c i , c j ) E signifying that concept c i is a prerequisite for concept c j .

3.2. Architecture Overview

Our hybrid framework employs a dual-pathway design consisting of three main components: (1) a temporal pathway that processes sequential interactions using transformer encoders, (2) a causal pathway that models prerequisite relationships through differentiable Bayesian networks, and (3) a cross-attention integration module that enables bidirectional information exchange between pathways.
Figure 1 illustrates the overall architecture. The temporal pathway captures sequential dependencies in student behavior, while the causal pathway learns interpretable prerequisite structures. The integration module ensures that temporal insights inform causal discovery and prerequisite knowledge guides sequence understanding.

3.3. Temporal Pathway: Transformer-Based Sequence Modeling

3.3.1. Input Representation

Each interaction s t is encoded as a combination of question, response, and concept embeddings:
x t = E q ( q t ) + E r ( r t ) + c c t E c ( c )
where x t R d represents the input embedding vector for interaction s t , E q R Q × d is the learnable question embedding matrix containing embeddings for all Q questions in the dataset, E q ( q t ) R d denotes the embedding vector for question q t , E r R 2 × d is the learnable response embedding matrix with two rows for correct and incorrect responses, E r ( r t ) R d represents the embedding for response r t , E c R K × d is the learnable concept embedding matrix containing embeddings for all K knowledge components, E c ( c ) R d denotes the embedding vector for concept c, and d is the embedding dimension that remains consistent across all embedding types.
Positional encoding is added to capture temporal ordering within the sequence:
h t = x t + PE ( t )
where h t R d represents the final input representation for interaction s t incorporating both content and positional information, and PE ( t ) R d follows the sinusoidal positional encoding scheme from [14], with PE ( t ) i = sin ( t / 10 , 000 2 i / d ) for even indices and PE ( t ) i = cos ( t / 10 , 000 2 ( i 1 ) / d ) for odd indices.

3.3.2. Multi-Head Self-Attention

The temporal transformer applies multi-head self-attention to capture long-range dependencies in the interaction sequence. For a sequence H = [ h 1 , h 2 , , h T ] R T × d containing the input representations for all T interactions, the attention mechanism computes
Attention ( Q , K , V ) = softmax Q K T d k V
where Q = H W Q R T × d k represents the query matrix obtained by projecting the input through learnable weights W Q R d × d k , K = H W K R T × d k represents the key matrix with projection weights W K R d × d k , V = H W V R T × d v represents the value matrix with projection weights W V R d × d v , d k is the dimension of query and key vectors, d v is the dimension of value vectors, and d k serves as a scaling factor to prevent the softmax function from saturating.
Multi-head attention combines h parallel attention heads to capture different types of relationships:
MultiHead ( H ) = Concat ( head 1 , , head h ) W O
where head i = Attention ( Q i , K i , V i ) represents the i-th attention head with its own parameter matrices, Concat ( · ) denotes concatenation along the feature dimension, h is the number of attention heads, and W O R h d v × d is the output projection matrix that combines information from all heads.

3.3.3. Temporal Encoding Layers

The temporal pathway consists of L transformer encoder layers, each containing multi-head self-attention followed by position-wise feed-forward networks (FFNs):
Z ( l ) = LayerNorm ( H ( l 1 ) + MultiHead ( H ( l 1 ) ) )
H ( l ) = LayerNorm ( Z ( l ) + FFN ( Z ( l ) ) )
where H ( l ) R T × d represents the output of the l-th encoder layer with l { 1 , 2 , , L } , H ( 0 ) = H is the initial input sequence, Z ( l ) R T × d denotes the intermediate representation after applying multi-head attention and residual connection in layer l, LayerNorm ( · ) applies layer normalization to stabilize training, and FFN ( x ) = max ( 0 , x W 1 + b 1 ) W 2 + b 2 represents the position-wise feed-forward network where W 1 R d × d f f and W 2 R d f f × d are learnable weight matrices, b 1 R d f f and b 2 R d are bias vectors, d f f is the inner dimension of the feed-forward network (typically 4 d ), and max ( 0 , · ) applies the ReLU activation function.
The final temporal representation is H t e m p = H ( L ) R T × d , which encodes the complete temporal context for all interactions in the sequence.

3.4. Causal Pathway: Differentiable Bayesian Networks

3.4.1. Causal Graph Structure Learning

The causal pathway learns a directed acyclic graph (DAG) G representing prerequisite relationships between knowledge components. We parameterize the adjacency matrix A R K × K using a continuous relaxation:
A i j = σ ( W c a u s a l · [ e i ; e j ] + b c a u s a l )
where A i j [ 0 , 1 ] represents the strength of the causal relationship from concept c i to concept c j (with A i j = 1 indicating a strong prerequisite relationship), e i , e j R d are the concept embeddings for concepts c i and c j , respectively, σ ( · ) is the sigmoid activation function that maps real values to the interval [ 0 , 1 ] , [ e i ; e j ] R 2 d denotes concatenation of the two concept embeddings, W c a u s a l R 2 d is a learnable weight vector, and b c a u s a l R is a learnable bias term.
To ensure DAG constraints, we apply the NOTEARS approach [47] with an acyclicity constraint:
h ( A ) = tr ( e A A ) K = 0
where h ( A ) is the acyclicity constraint function that equals zero when A represents a valid DAG, tr ( · ) denotes the trace operator (sum of diagonal elements), e A A represents the matrix exponential of the element-wise squared adjacency matrix, ⊙ denotes element-wise multiplication, and K is subtracted because the trace of the identity matrix (representing self-loops) should be removed from a valid DAG.

3.4.2. Concept Mastery Modeling

Given the learned causal structure, we model concept mastery states using a differentiable Bayesian network. The mastery probability for concept c i at time t is computed as:
p ( m i , t = 1 ) = σ w i T f p a r e n t s ( i ) + α i · practice i , t + β i
where p ( m i , t = 1 ) [ 0 , 1 ] represents the probability that the student has mastered concept c i at time step t, w i R | p a r e n t s ( i ) | is a learnable weight vector specific to concept c i with dimension equal to the number of its parent concepts, f p a r e n t s ( i ) R | p a r e n t s ( i ) | represents the aggregated mastery state vector for all prerequisite concepts of c i (where p a r e n t s ( i ) = { c j : A j i > 0.5 } ), α i R is a learnable parameter representing the learning rate for concept c i , practice i , t N denotes the cumulative number of practice attempts for concept c i up to time t, and β i R is a learnable bias term representing the initial difficulty of concept c i .
The causal representation for the sequence is obtained by aggregating concept mastery states across all time steps:
H c a u s a l = MLP ( [ m 1 , m 2 , , m T ] )
where H c a u s a l R T × d is the final causal pathway representation, m t R K contains the mastery probabilities for all K concepts at time step t (i.e., m t = [ p ( m 1 , t = 1 ) , p ( m 2 , t = 1 ) , , p ( m K , t = 1 ) ] T ), [ m 1 , m 2 , , m T ] R T K represents the concatenation of all mastery states across time, and MLP ( · ) denotes a multi-layer perceptron that projects the concatenated mastery states to the desired representation dimension.

3.5. Cross-Attention Integration

The cross-attention mechanism enables bidirectional information flow between temporal and causal pathways. We implement two cross-attention modules:

3.5.1. Temporal-to-Causal Attention

This module allows temporal insights to inform causal structure learning:
H e n h a n c e d c a u s a l = CrossAttention ( H c a u s a l , H t e m p , H t e m p )
where H c a u s a l serves as queries and H t e m p provides keys and values.

3.5.2. Causal-to-Temporal Attention

Conversely, this module incorporates prerequisite knowledge into temporal modeling:
H e n h a n c e d t e m p = CrossAttention ( H t e m p , H c a u s a l , H c a u s a l )
The final integrated representation combines both enhanced pathways:
H i n t e g r a t e d = FFN ( [ H e n h a n c e d t e m p ; H e n h a n c e d c a u s a l ] )

3.6. Joint Training Objective

Our training objective combines multiple loss components to ensure both predictive accuracy and causal consistency:

3.6.1. Prediction Loss

The primary prediction loss uses binary cross-entropy for response prediction:
L p r e d = t = 1 T [ r t log p t + ( 1 r t ) log ( 1 p t ) ]
where p t = σ ( W p r e d · h t i n t e g r a t e d + b p r e d ) .

3.6.2. Causal Structure Loss

To encourage meaningful causal structures, we incorporate a sparsity penalty and prerequisite consistency loss:
L c a u s a l = λ 1 A 1 + λ 2 ( i , j ) E e x p e r t BCE ( A i j , 1 )
where E e x p e r t represents known expert prerequisite relationships and BCE is binary cross-entropy.

3.6.3. Acyclicity Constraint

The DAG constraint is enforced through:
L d a g = ρ · h ( A ) 2
where ρ is a penalty parameter that increases during training.

3.6.4. Alignment Loss

To ensure consistency between temporal and causal representations, we introduce an alignment loss:
L a l i g n = H e n h a n c e d t e m p H e n h a n c e d c a u s a l F 2
The total loss combines all components:
L t o t a l = L p r e d + L c a u s a l + L d a g + γ L a l i g n

3.7. Training Strategy

3.7.1. Pre-Training Phase

We pre-train the model on large-scale MOOC data to learn general patterns of knowledge acquisition. During pre-training, we focus on the temporal pathway and basic causal structure learning without domain-specific constraints.
The pre-training objective emphasizes sequence prediction and general prerequisite discovery:
L p r e t r a i n = L p r e d + λ 1 A 1 + L d a g

3.7.2. Fine-Tuning Phase

Domain-specific fine-tuning incorporates expert knowledge and adapts the model to particular subjects. We introduce curriculum learning by gradually increasing the complexity of causal constraints.
During fine-tuning, we use the full loss with domain-specific expert knowledge:
L f i n e t u n e = L t o t a l + λ e x p e r t ( i , j ) E d o m a i n BCE ( A i j , 1 )

3.7.3. Optimization Details

We employ the Adam optimizer with learning rate scheduling. The model is trained using teacher forcing during the temporal modeling phase. For causal structure learning, we use the augmented Lagrangian method to handle the DAG constraint.
The training procedure alternates between updating temporal parameters and causal parameters to ensure stable convergence. We first update temporal pathway parameters while keeping the causal structure frozen, allowing the transformer components to learn sequential patterns without interference from changing causal relationships. Subsequently, we update causal pathway parameters with frozen temporal representations, enabling the Bayesian network to discover prerequisite structures based on stable temporal features. Finally, we perform a joint update with full cross-attention integration, allowing both pathways to adapt simultaneously while maintaining their learned complementary representations. This multi-stage training ensures that both pathways develop complementary representations while maintaining interpretability and predictive accuracy.

4. Experiments

This section presents comprehensive experiments validating our hybrid architecture across multiple educational domains. We evaluate predictive performance, interpretability, and scalability against state-of-the-art knowledge tracing methods. Code, models, and more information are in the Supplementary Materials.

4.1. Experimental Setup

4.1.1. Datasets

We conduct experiments on four widely used knowledge tracing datasets spanning diverse educational contexts. The ASSISTments 2009–2010 dataset [48] contains 4217 students with 346,860 interactions across 124 skills in mathematics, providing rich skill tagging that has been extensively used for knowledge tracing evaluation. EdNet-KT1 [49] represents a large-scale dataset from the Santa English learning platform containing 784,309 students with 131,441,538 interactions across 13,169 questions, where we use the KT1 subset focusing on TOEIC preparation. The Junyi Academy dataset [22] encompasses mathematics learning with 247,606 students and 25,925,922 interactions covering 721 exercises, exhibiting clear prerequisite relationships in mathematical concepts that make it ideal for causal structure evaluation. Finally, the KDD Cup 2010 dataset [50] provides algebra tutoring data containing 574 students with 8,918,054 interactions across 661 knowledge components, featuring detailed step-level feedback and hint usage that enables fine-grained analysis.
Each dataset undergoes preprocessing following standard protocols where we filter students with fewer than 10 interactions and skills with fewer than 50 interactions to ensure statistical reliability. Temporal ordering is preserved throughout the preprocessing pipeline, and we apply 80/10/10 train/validation/test splits chronologically to simulate realistic deployment scenarios.

4.1.2. Baseline Methods

Our comparison encompasses representative knowledge tracing approaches spanning traditional probabilistic models to modern deep learning architectures. BKT [6] serves as the classical baseline using expectation-maximization parameter estimation for modeling knowledge acquisition. DKT [9] represents the foundational deep learning approach using LSTM networks for sequence modeling. DKVMN [11] extends this with explicit concept representations through dynamic key-value memory mechanisms. SAKT [10] applies transformer-style self-attention to learning sequences, while AKT [33] incorporates context-aware attention mechanisms for improved temporal modeling. SAINT [30] employs an encoder–decoder transformer architecture specifically designed for knowledge tracing, and simpleKT [34] provides a simplified transformer-based approach that emphasizes core attention mechanisms while achieving competitive performance.
LLM-KT [35] represents the current state-of-the-art in large language model integration for knowledge tracing, leveraging BERT embeddings to capture semantic relationships between educational content and student responses while maintaining computational efficiency through selective fine-tuning strategies.

4.1.3. Evaluation Metrics

Our evaluation employs standard knowledge tracing metrics alongside interpretability measures. AUC measures discriminative ability between correct and incorrect responses, providing insight into the model’s ranking capability. RMSE quantifies prediction accuracy for response probabilities, offering direct assessment of probabilistic predictions. F1-Score evaluates binary classification performance using a 0.5 probability threshold, balancing precision and recall considerations. MAE assesses average prediction deviation, complementing RMSE with a different error perspective. For interpretability evaluation, we measure prerequisite discovery accuracy as the fraction of expert-validated prerequisite relationships correctly identified using an adjacency matrix threshold of 0.5, along with causal structure precision and recall for discovered prerequisite edges against expert annotations.

4.1.4. Implementation Details

Our model employs embedding dimension d = 256 , transformer layers L = 4 , attention heads h = 8 , and FFN dimension d f f = 1024 , with cross-attention modules using identical architectures. Training utilizes the Adam optimizer with learning rate 0.001, batch size 256, and early stopping based on validation AUC. Hyperparameters are set to λ 1 = 0.01 , λ 2 = 0.1 , γ = 0.05 , with ρ increasing from 0.1 to 1.0 over 50 epochs for progressive DAG constraint enforcement. Pre-training combines all datasets for 100 epochs, followed by domain-specific fine-tuning for 50 epochs, with expert prerequisite relationships obtained from curriculum documentation and educational literature.

4.2. Main Results

Table 1 presents a performance comparison across all datasets, demonstrating that our hybrid method consistently outperforms baselines across all metrics.
The results reveal substantial improvements across all evaluation metrics and datasets. On the ASSISTments dataset, our method achieves an 8.7% AUC improvement over the strongest baseline simpleKT (0.847 vs. 0.789), accompanied by a 12.3% RMSE reduction (0.325 vs. 0.371) and 15.4% F1-score enhancement (0.763 vs. 0.724). Similar performance patterns emerge across EdNet-KT1, Junyi Academy, and KDD Cup 2010, with AUC improvements ranging from 5.4% to 8.7%, RMSE reductions between 10.7% and 12.3%, and F1-score gains from 4.8% to 15.4%. The consistency of improvements across diverse educational domains demonstrates the generalizability of our hybrid architecture, while the magnitude of gains indicates that the integration of temporal and causal modeling provides substantial practical benefits over existing approaches. The results demonstrate substantial improvements over all baselines, including the current state-of-the-art LLM-KT method. On the ASSISTments dataset, our method achieves a 2.9% AUC improvement over LLM-KT (0.847 vs. 0.823), with a 6.6% RMSE reduction (0.325 vs. 0.348) and 3.0% F1-score enhancement (0.763 vs. 0.741). Similar performance patterns emerge across all datasets, with AUC improvements over LLM-KT ranging from 2.7% to 2.9%, demonstrating consistent superiority over modern LLM-based approaches while maintaining interpretability advantages through explicit causal structure learning.
Table 1 presents a comprehensive performance comparison across all datasets, including precision–recall analysis to understand model behavior across varying correct answer distributions. Dataset correct answer percentages range from 67.4% (KDD Cup 2010) to 73.1% (EdNet-KT1), providing diverse evaluation contexts for assessing model robustness.
Our hybrid method demonstrates consistent superiority across all metrics while maintaining balanced precision–recall profiles. On ASSISTments (69.2% correct answers), we achieve 0.758 precision and 0.769 recall compared to LLM-KT’s 0.721 precision and 0.762 recall, indicating improved reliability without sacrificing coverage. This balanced performance pattern emerges across all datasets, with precision–recall differences remaining within 1.1% for our method versus 4.1% for LLM-KT.
The precision–recall analysis reveals important model characteristics. Traditional methods like DKT exhibit recall-biased predictions (0.713 recall vs. 0.652 precision on ASSISTments), particularly problematic on datasets with high correct answer percentages where overly optimistic predictions inflate performance metrics. Conversely, our method maintains consistent precision–recall balance across varying dataset characteristics through the cross-attention mechanism’s integration of temporal patterns with causal constraints.
Notably, our method achieves 2.9% AUC improvement over LLM-KT on ASSISTments (0.847 vs. 0.823) with 5.1% precision enhancement (0.758 vs. 0.721), demonstrating superior discriminative ability without excessive optimism. Similar patterns emerge across EdNet-KT1, Junyi Academy, and KDD Cup 2010, confirming robustness to dataset-specific characteristics and correct answer distributions.
Figure 2 visualizes the performance comparison, clearly illustrating the consistent superiority of our method across all datasets and metrics. The results demonstrate several key findings. First, our hybrid architecture achieves the highest performance across all four evaluation metrics on every dataset, with particularly notable improvements in AUC ranging from 5.4% on EdNet-KT1 to 8.7% on ASSISTments compared to the strongest baseline simpleKT. Second, the performance gains are consistent across diverse educational domains, from mathematics (ASSISTments, Junyi Academy) to English learning (EdNet-KT1) and algebra tutoring (KDD Cup 2010), indicating strong generalizability. Third, the magnitude of improvements increases with dataset complexity—larger datasets like EdNet-KT1 and Junyi Academy show substantial gains, suggesting our method scales effectively with data volume. Fourth, traditional methods like BKT show significantly lower performance across all datasets, while modern deep learning approaches (SAKT, SAINT, simpleKT) form a competitive baseline cluster that our method consistently surpasses. The error reduction metrics (RMSE, MAE) show particularly strong improvements of 10–15%, indicating more accurate probability predictions crucial for educational applications.

4.3. Interpretability Analysis

Table 2 evaluates prerequisite discovery performance by comparing against expert annotations from curriculum documentation.
The interpretability results demonstrate remarkable alignment between discovered and expert-validated prerequisite relationships. Our model achieves 89.2% average prerequisite discovery accuracy, with particularly strong performance on the Junyi Academy mathematics dataset (90.1%), where clear hierarchical relationships exist between mathematical concepts. The precision scores averaging 88.2% indicate that the majority of discovered relationships represent genuine pedagogical dependencies, while recall values averaging 84.5% show comprehensive coverage of established prerequisite structures. These results validate that our hybrid architecture successfully identifies meaningful causal relationships such as “addition before multiplication” in mathematics and “basic grammar before complex syntax” in language learning, providing actionable insights for curriculum design and personalized learning pathways.
Figure 3 visualizes the learned prerequisite structure for mathematics concepts, demonstrating clear hierarchical progression that aligns with established mathematical pedagogy. The heatmap reveals several important patterns in the discovered causal relationships. The upper-triangular structure confirms successful DAG constraint enforcement, preventing cyclic dependencies that would violate pedagogical logic. Strong prerequisite relationships (dark blue cells with values > 0.8) form clear pathways from foundational concepts like Number Sense and Counting to advanced topics such as Quadratic Equations and Systems of Equations. Notably, the matrix shows that Addition and Subtraction serve as critical gateway skills, with strong connections to Multiplication, Division, and Fractions, reflecting their fundamental role in mathematical learning. The diagonal band pattern indicates that most prerequisite relationships exist between adjacent or near-adjacent concepts in the hierarchy, with weaker but meaningful connections (medium blue, 0.4–0.7) extending to more distant advanced topics. The model correctly identifies parallel learning paths, such as the simultaneous development of Fractions and Decimals from Division, and captures the convergence of multiple arithmetic skills into Basic Algebra. This learned structure achieves 90.1% alignment with expert annotations for the Junyi Academy dataset, demonstrating that our differentiable Bayesian network successfully discovers educationally meaningful prerequisite relationships that can inform curriculum sequencing and personalized learning pathways.

4.4. Ablation Studies

We conduct systematic ablation studies to understand the contribution of individual components to overall performance.
Table 3 examines the effect of removing different architectural components on the ASSISTments dataset.
The ablation analysis reveals the critical importance of each architectural component. Removing the cross-attention integration results in a 3.1% AUC decrease (0.847 to 0.821), demonstrating that bidirectional information flow between temporal and causal pathways significantly enhances predictive performance. The causal pathway’s removal causes a substantial 5.8% AUC drop (0.847 to 0.798) and completely eliminates prerequisite discovery capability, confirming its essential role in interpretable modeling. Interestingly, removing the temporal pathway while retaining the causal components still allows for 87.9% prerequisite discovery accuracy but severely impacts prediction performance, highlighting the complementary nature of both pathways. Pre-training contributes a 2.1% AUC improvement, validating the benefits of large-scale initialization, while expert knowledge integration enhances both prediction accuracy and prerequisite discovery quality.
Table 4 analyzes the contribution of different loss function components.
The loss function analysis demonstrates the synergistic effects of joint optimization. The causal structure loss L c a u s a l proves crucial for prerequisite discovery, improving accuracy by 14.9 percentage points (0.892 vs. 0.743) while maintaining prediction performance. The DAG constraint L d a g prevents cyclic dependencies in the learned graph, boosting prerequisite accuracy by 24.1 points (0.892 vs. 0.651) and ensuring pedagogically meaningful structures. The alignment loss L a l i g n provides modest but consistent improvements across all metrics by ensuring coherence between temporal and causal representations. Training with only the prediction loss L p r e d severely limits both interpretability and performance, confirming the necessity of multi-objective optimization.
Figure 4 visualizes the relative contribution of different components, clearly showing the cross-attention mechanism and causal pathway as the most critical elements for achieving superior performance. The analysis reveals distinct patterns of component importance across evaluation metrics. The causal pathway demonstrates the most dramatic impact, with its removal causing substantial performance degradation across all metrics—particularly devastating for prerequisite discovery (100% loss) and significant for predictive performance (5.8% AUC drop, 12.9% RMSE increase). The cross-attention integration mechanism emerges as the second-most critical component, contributing 3.1% AUC improvement and 7.1% RMSE reduction, confirming that bidirectional information flow between pathways is essential for optimal performance. Temporal pathway removal severely impacts prediction accuracy (12.3% AUC degradation) while surprisingly maintaining high prerequisite discovery capability (87.9%), highlighting the complementary roles of both pathways. Pre-training contributes consistently but moderately across metrics (2.1% AUC improvement), validating the benefits of large-scale initialization without being architecturally critical. Expert knowledge integration shows the smallest but meaningful impact (1.1% AUC gain, 7.6% prerequisite accuracy improvement), suggesting our model effectively learns causal structures from data while benefiting from pedagogical guidance. The differential impact patterns confirm our architectural design choices and demonstrate that both temporal modeling and causal reasoning are necessary for achieving state-of-the-art performance in interpretable knowledge tracing.

4.5. Scalability Analysis

We evaluate computational efficiency and scalability across varying dataset sizes and sequence lengths. Training time scales linearly with sequence length and quadratically with concept number due to cross-attention computation complexity. On the ASSISTments dataset, training requires 2.3 h on an NVIDIA V100 GPU compared to 1.1 h for SAINT, representing an acceptable 109% computational overhead given the substantial interpretability gains. Memory consumption increases proportionally with the number of knowledge components due to the adjacency matrix storage requirements, but remains manageable for typical educational datasets with hundreds to thousands of concepts.
Figure 5 illustrates convergence behavior across datasets, demonstrating stable training dynamics and consistent convergence within 50 fine-tuning epochs. The curves show minimal overfitting and smooth convergence patterns, indicating robust optimization despite the complex multi-objective loss function. The analysis reveals several critical training characteristics. First, all datasets exhibit rapid initial improvement in the first 10–15 epochs, with AUC gains of 4–6% and loss reductions of 15–20%, indicating effective knowledge transfer from pre-training. Second, the training and validation curves maintain close alignment throughout the process, with validation performance tracking training performance within 1–2% across all datasets, demonstrating excellent generalization without overfitting. Third, the shaded confidence intervals (±1 standard deviation) remain narrow and consistent, showing training stability with standard deviations below 0.01 for AUC and 0.02 for loss metrics. Fourth, convergence occurs between epochs 25–35 for most datasets, well within the 50-epoch fine-tuning window, with performance stabilization indicated by minimal fluctuations in the final epochs. Fifth, dataset-specific patterns emerge—EdNet-KT1 shows slightly slower convergence due to its large scale, while ASSISTments converges fastest due to clearer mathematical structures. The loss curves demonstrate exponential decay patterns typical of well-conditioned optimization, while AUC curves show smooth logarithmic growth toward asymptotic performance limits. The consistent convergence across diverse educational domains validates our training methodology and confirms that the hybrid architecture learns effectively despite its complex dual-pathway design and multi-objective optimization requirements.

5. Conclusions

This work presents a novel hybrid architecture that integrates transformer-based temporal modeling with differentiable Bayesian causal networks to achieve both predictive accuracy and interpretability in knowledge tracing. Our dual-pathway framework employs multi-head attention mechanisms to capture complex temporal dependencies in educational interaction sequences while simultaneously learning interpretable causal prerequisite structures through differentiable DAG constraints. The cross-attention integration mechanism enables bidirectional information exchange between temporal representations and pedagogical causal knowledge, facilitating effective information fusion while maintaining educational interpretability. Comprehensive experiments across four educational datasets demonstrate substantial performance improvements: 8.7% AUC enhancement over state-of-the-art baselines (0.847 vs. 0.789), 12.3% RMSE reduction, and 89.2% accuracy in discovering expert-validated prerequisite relationships. The joint optimization framework successfully balances sequence prediction accuracy and causal graph consistency, enabling real-time processing of educational interactions while maintaining interpretable insights into learning progressions and concept dependencies.

5.1. Interpretability Limitations and Practical Implications

While our hybrid architecture provides significant interpretability advantages over purely black-box approaches, we acknowledge fundamental limitations that affect practical educational applications. Our architecture exhibits partial interpretability: the Bayesian causal pathway offers transparent prerequisite relationships and concept mastery probabilities, while the transformer pathway operates as a black box for temporal pattern recognition. This creates scenarios where overall interpretability may be compromised when cross-attention mechanisms prioritize temporal patterns over causal structures in prediction generation.
We have developed a multi-level interpretation strategy to address this limitation in educational contexts. Primary interpretations draw from the causal pathway’s transparent prerequisite relationships, which achieve 89.2% accuracy against expert annotations, enabling educators to understand learning dependencies and identify concrete intervention points. Concept mastery probabilities provide probabilistic assessments of student knowledge states for each skill, supporting targeted remediation strategies. When attention weights favor non-interpretable transformer components, supplementary insights can be obtained through attention visualization techniques that reveal temporal focus patterns, though these lack the pedagogical clarity of causal relationships.
For student feedback generation, our approach prioritizes interpretable causal insights regardless of transformer opacity. The learned prerequisite structures enable concrete educational recommendations such as “master basic arithmetic operations before attempting algebraic expressions” or “strengthen reading comprehension skills before advanced writing tasks.” These actionable insights derive directly from the Bayesian pathway and remain available even when cross-attention emphasizes temporal patterns that cannot be easily explained to educational stakeholders.
This represents an inherent trade-off in hybrid architectures between complete interpretability and predictive performance. Achieving full transparency would require eliminating the transformer pathway, sacrificing the temporal modeling capabilities that enable our superior predictive performance and 8.7% AUC improvement over interpretable baselines. Our work provides meaningful interpretability for educational decision-making through explicit causal relationships while acknowledging that complete transparency remains challenging in high-performance knowledge tracing systems. Future research should explore attention regularization techniques that maintain interpretability balance and develop domain-specific explanation methods for transformer components in educational contexts.

5.2. Limitations and Future Directions

Several limitations suggest important directions for future educational analytics research. Computational complexity scales quadratically with knowledge components, constraining scalability in large educational domains with thousands of concepts. Cold start scenarios with insufficient interaction history present ongoing challenges for both predictive performance and causal discovery accuracy. The current linear prerequisite assumption may inadequately capture complex many-to-one dependencies emerging from sophisticated educational structures, motivating extensions to hypergraph structures and dynamic causal discovery mechanisms.
Future investigations should address adaptive attention mechanisms that preserve interpretability balance, uncertainty quantification for causal structure predictions, integration with federated learning frameworks for privacy-preserving educational analytics, and standardized evaluation protocols for interpretable knowledge tracing systems. Additionally, real-time deployment optimization, cross-domain transfer learning for causal structures, and longitudinal validation studies in actual educational settings represent critical research priorities for reliable deployment in production environments where educational effectiveness depends on both accurate predictions and interpretable insights about student learning progression.

Supplementary Materials

The code used for the experiments in this study can be downloaded at: https://www.mdpi.com/article/10.3390/app15179605/s1.

Author Contributions

Methodology, W.C.; Software, W.C. and W.L.; Formal analysis, N.T.M. and W.L.; Writing—original draft, W.C.; Writing—review & editing, N.T.M. and W.L.; Supervision, N.T.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Perna, L.W.; Ruby, A.; Boruch, R.F.; Wang, N.; Scull, J.; Ahmad, S.; Evans, C. Moving through MOOCs: Understanding the progression of users in massive open online courses. Educ. Res. 2014, 43, 421–432. [Google Scholar] [CrossRef]
  2. Goggins, S.P.; Galyen, K.; Petakovic, E.; Laffey, J.M. Connecting performance to social structure and pedagogy as a pathway to scaling learning analytics in MOOCs: An exploratory study. J. Comput. Assist. Learn. 2016, 32, 244–266. [Google Scholar] [CrossRef]
  3. Trabelsi, Z.; Alnajjar, F.; Parambil, M.M.A.; Gochoo, M.; Ali, L. Real-time attention monitoring system for classroom: A deep learning approach for student’s behavior recognition. Big Data Cogn. Comput. 2023, 7, 48. [Google Scholar] [CrossRef]
  4. Habib Albohamood, A.; Alqattan, M.S.; Vizcarra, C.P. Real-time Student Engagement Monitoring in Classroom Environments using Machine Learning and Computer Vision. In Proceedings of the 2025 4th International Conference on Computing and Information Technology (ICCIT), Tabuk, Saudi Arabia, 13–14 April 2025; pp. 420–424. [Google Scholar]
  5. Qiu, S. Improving performance of smart education systems by integrating machine learning on edge devices and cloud in educational institutions. J. Grid Comput. 2024, 22, 41. [Google Scholar] [CrossRef]
  6. Corbett, A.T.; Anderson, J.R. Knowledge tracing: Modeling the acquisition of procedural knowledge. User Model. User-Adapt. Interact. 1994, 4, 253–278. [Google Scholar] [CrossRef]
  7. Cai, L.; Choi, K.; Hansen, M.; Harrell, L. Item response theory. Annu. Rev. Stat. Its Appl. 2016, 3, 297–321. [Google Scholar] [CrossRef]
  8. Pelánek, R. Bayesian knowledge tracing, logistic models, and beyond: An overview of learner modeling techniques. User Model. User-Adapt. Interact. 2017, 27, 313–350. [Google Scholar] [CrossRef]
  9. Piech, C.; Bassen, J.; Huang, J.; Ganguli, S.; Sahami, M.; Guibas, L.J.; Sohl-Dickstein, J. Deep knowledge tracing. Adv. Neural Inf. Process. Syst. 2015, 28, 505–513. [Google Scholar]
  10. Pandey, S.; Karypis, G. A self-attentive model for knowledge tracing. arXiv 2019, arXiv:1907.06837. [Google Scholar] [CrossRef]
  11. Zhang, J.; Shi, X.; King, I.; Yeung, D.Y. Dynamic key-value memory networks for knowledge tracing. In Proceedings of the 26th International Conference on World Wide Web, Perth, Australia, 3–7 April 2017; pp. 765–774. [Google Scholar]
  12. Nakagawa, H.; Iwasawa, Y.; Matsuo, Y. Graph-based knowledge tracing: Modeling student proficiency using graph neural network. In Proceedings of the IEEE/WIC/aCM International Conference on Web Intelligence, Thessaloniki, Greece, 14–17 October 2019; pp. 156–163. [Google Scholar]
  13. Gervet, T.; Koedinger, K.; Schneider, J.; Mitchell, T. When is deep learning the best approach to knowledge tracing? J. Educ. Data Min. 2020, 12, 31–54. [Google Scholar]
  14. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  15. Yue, W.; Su, W.; Liu, L.; Cai, C.; Yuan, Y.; Jia, Z.; Liu, J.; Xie, W. A pre-trained knowledge tracing model with limited data. In International Conference on Database and Expert Systems Applications; Springer: Berlin/Heidelberg, Germany, 2024; pp. 163–178. [Google Scholar]
  16. Wang, J.; Xie, H.; Au, O.T.S.; Zou, D.; Wang, F.L. Attention-based CNN for personalized course recommendations for MOOC learners. In Proceedings of the 2020 International Symposium on Educational Technology (ISET), Bangkok, Thailand, 24–27 August 2020; pp. 180–184. [Google Scholar]
  17. Ma, Q. Utilization of transformer model in multimodal data fusion learning: Cross-modal knowledge transfer in the new generation learning space. Intell. Decis. Technol. 2024, IDT-240169. [Google Scholar] [CrossRef]
  18. Ramesh, D.; Sanampudi, S.K. An automated essay scoring systems: A systematic literature review. Artif. Intell. Rev. 2022, 55, 2495–2527. [Google Scholar] [CrossRef]
  19. Abdelrahman, G.; Wang, Q. Knowledge tracing with sequential key-value memory networks. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 175–184. [Google Scholar]
  20. Pearl, J. Probabilistic Reasoning in Intelligent Systems: Networks of Plausible Inference; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
  21. Koller, D.; Friedman, N. Probabilistic Graphical Models: Principles and Techniques; MIT Press: Cambridge, MA, USA, 2009. [Google Scholar]
  22. Desmarais, M.C.; Baker, R.S.d. A review of recent advances in learner and skill modeling in intelligent learning environments. User Model. User-Adapt. Interact. 2012, 22, 9–38. [Google Scholar] [CrossRef]
  23. Beck, J.E.; Chang, K.-m. Identifiability: A fundamental problem of student modeling. In International Conference on User Modeling; Springer: Berlin/Heidelberg, Germany, 2007; pp. 137–146. [Google Scholar]
  24. Rasch, G. Studies in Mathematical Psychology: I. Probabilistic Models for Some Intelligence and Attainment Tests; Danish Institute for Educational Research: Copenhagen, Denmark, 1960. [Google Scholar]
  25. Birnbaum, A. Some latent trait models and their use in inferring an examinee’s ability. In Statistical Theories of Mental Test Scores; Lord, F.M., Novick, M.R., Eds.; Addison-Wesley: Reading, MA, USA, 1968; pp. 397–479. [Google Scholar]
  26. Reckase, M.D. 18 Multidimensional item response theory. Handb. Stat. 2006, 26, 607–642. [Google Scholar]
  27. Pavlik, P.I.; Cen, H.; Koedinger, K.R. Performance factors analysis—A new alternative to knowledge tracing. In Artificial Intelligence in Education; Ios Press: Amsterdam, The Netherlands, 2009; pp. 531–538. [Google Scholar]
  28. Cen, H.; Koedinger, K.; Junker, B. Learning factors analysis—A general method for cognitive model evaluation and improvement. In International Conference on Intelligent Tutoring Systems; Springer: Berlin/Heidelberg, Germany, 2006; pp. 164–175. [Google Scholar]
  29. Khajah, M.; Lindsey, R.V.; Mozer, M.C. How deep is knowledge tracing? arXiv 2016, arXiv:1604.02416. [Google Scholar] [CrossRef]
  30. Choi, Y.; Lee, Y.; Cho, J.; Baek, J.; Kim, B.; Cha, Y.; Shin, D.; Bae, C.; Heo, J. Towards an appropriate query, key, and value computation for knowledge tracing. In Proceedings of the Seventh ACM Conference on Learning@ Scale, Virtual, 12–14 August 2020; pp. 341–344. [Google Scholar]
  31. Pu, S.; Yudelson, M.; Ou, L.; Huang, Y. Deep knowledge tracing with transformers. In International Conference on Artificial Intelligence in Education; Springer: Berlin/Heidelberg, Germany, 2020; pp. 252–256. [Google Scholar]
  32. Li, Z.; Xue, Z.; Liu, C.; Feng, Y. Deep Knowledge Tracing Model with an Evolved Transformer Structure. In Proceedings of the 2023 IEEE 12th Data Driven Control and Learning Systems Conference (DDCLS), Xiangtan, China, 12–14 May 2023; pp. 1586–1592. [Google Scholar]
  33. Ghosh, A.; Heffernan, N.; Lan, A.S. Context-aware attentive knowledge tracing. In Proceedings of the 26th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Virtual, 6–10 July 2020; pp. 2330–2339. [Google Scholar]
  34. Liu, Z.; Liu, Q.; Chen, J.; Huang, S.; Luo, W. simpleKT: A simple but tough-to-beat baseline for knowledge tracing. arXiv 2023, arXiv:2302.06881. [Google Scholar]
  35. Wang, Z.; Zhou, J.; Chen, Q.; Zhang, M.; Jiang, B.; Zhou, A.; Bai, Q.; He, L. LLM-KT: Aligning Large Language Models with Knowledge Tracing using a Plug-and-Play Instruction. arXiv 2025, arXiv:2502.02945. [Google Scholar]
  36. Park, S.; Kim, H. A Comprehensive Survey and Taxonomy on Large Language Model-Based Knowledge Tracing. In International Conference on Intelligent Tutoring Systems; Springer: Berlin/Heidelberg, Germany, 2025; pp. 246–258. [Google Scholar]
  37. Wang, D.; Chen, G.; Lu, Y. Fine-Tuning Large Language Models for Knowledge Tracing Harnessing Insights from Explainable AI. In International Conference on Artificial Intelligence in Education; Springer: Berlin/Heidelberg, Germany, 2025; pp. 297–302. [Google Scholar]
  38. Yang, C.; Zhu, Y.; Lu, W.; Wang, Y.; Chen, Q.; Gao, C.; Yan, B.; Chen, Y. Survey on knowledge distillation for large language models: Methods, evaluation, and application. ACM Trans. Intell. Syst. Technol. 2024. [Google Scholar] [CrossRef]
  39. Lee, U.; Bae, J.; Kim, D.; Lee, S.; Park, J.; Ahn, T.; Lee, G.; Stratton, D.; Kim, H. Language model can do knowledge tracing: Simple but effective method to integrate language model and knowledge tracing task. arXiv 2024, arXiv:2406.02893. [Google Scholar] [CrossRef]
  40. He, L.; Li, X.; Wang, P.; Tang, J.; Wang, T. Integrating fine-grained attention into multi-task learning for knowledge tracing. World Wide Web 2023, 26, 3347–3372. [Google Scholar] [CrossRef]
  41. Gao, R.; Ni, Q.; Hu, B. Fairness of large language models in education. In Proceedings of the 2024 International Conference on Intelligent Education and Computer Technology, Guilin, China, 28–30 June 2024; p. 1. [Google Scholar]
  42. Pearl, J. Causality; Cambridge University Press: Cambridge, UK, 2009. [Google Scholar]
  43. Spirtes, P.; Glymour, C.N.; Scheines, R. Causation, Prediction, and Search; MIT Press: Cambridge, MA, USA, 2000. [Google Scholar]
  44. Chickering, D.M. Optimal structure identification with greedy search. J. Mach. Learn. Res. 2002, 3, 507–554. [Google Scholar]
  45. Conati, C. Probabilistic assessment of user’s emotions in educational games. Appl. Artif. Intell. 2002, 16, 555–575. [Google Scholar] [CrossRef]
  46. Murphy, K.P. Dynamic Bayesian Networks: Representation, Inference and Learning; University of California: Berkeley, CA, USA, 2002. [Google Scholar]
  47. Zheng, X.; Aragam, B.; Ravikumar, P.K.; Xing, E.P. Dags with no tears: Continuous optimization for structure learning. arXiv 2018, arXiv:1803.01422. [Google Scholar] [CrossRef]
  48. Feng, M.; Heffernan, N.; Koedinger, K. Addressing the assessment challenge with an online system that tutors as it assesses. User Model. User-Adapt. Interact. 2009, 19, 243–266. [Google Scholar] [CrossRef]
  49. Choi, Y.; Lee, Y.; Shin, D.; Cho, J.; Park, S.; Lee, S.; Baek, J.; Bae, C.; Kim, B.; Heo, J. Ednet: A large-scale hierarchical dataset in education. In Proceedings of the Artificial Intelligence in Education: 21st International Conference, AIED 2020, Ifrane, Morocco, 6–10 July 2020; Proceedings, Part II 21. Springer: Berlin/Heidelberg, Germany, 2020; pp. 69–73. [Google Scholar]
  50. Stamper, J.; Pardos, Z.A. The 2010 KDD Cup Competition Dataset: Engaging the machine learning community in predictive learning analytics. J. Learn. Anal. 2016, 3, 312–316. [Google Scholar] [CrossRef]
Figure 1. Overview of the proposed hybrid architecture. The temporal pathway (top) processes student interaction sequences using transformer encoders to capture temporal dependencies. The causal pathway (bottom) employs differentiable Bayesian networks to learn prerequisite relationships between knowledge components. The cross-attention integration module (center) enables bidirectional information exchange between pathways, allowing temporal insights to inform causal structure discovery while prerequisite knowledge guides sequence understanding.
Figure 1. Overview of the proposed hybrid architecture. The temporal pathway (top) processes student interaction sequences using transformer encoders to capture temporal dependencies. The causal pathway (bottom) employs differentiable Bayesian networks to learn prerequisite relationships between knowledge components. The cross-attention integration module (center) enables bidirectional information exchange between pathways, allowing temporal insights to inform causal structure discovery while prerequisite knowledge guides sequence understanding.
Applsci 15 09605 g001
Figure 2. Performance comparison across datasets. Our hybrid method consistently outperforms baselines with significant margins across all evaluation metrics.
Figure 2. Performance comparison across datasets. Our hybrid method consistently outperforms baselines with significant margins across all evaluation metrics.
Applsci 15 09605 g002
Figure 3. Learned prerequisite adjacency matrix for Junyi Academy mathematics concepts. Darker cells indicate stronger prerequisite relationships, revealing clear hierarchical structure from basic arithmetic to advanced algebra.
Figure 3. Learned prerequisite adjacency matrix for Junyi Academy mathematics concepts. Darker cells indicate stronger prerequisite relationships, revealing clear hierarchical structure from basic arithmetic to advanced algebra.
Applsci 15 09605 g003
Figure 4. Component contribution analysis showing performance degradation when removing different architectural elements. Cross-attention and causal pathway components show the largest individual contributions.
Figure 4. Component contribution analysis showing performance degradation when removing different architectural elements. Cross-attention and causal pathway components show the largest individual contributions.
Applsci 15 09605 g004
Figure 5. Training convergence curves showing stable learning across different datasets. The model converges within 50 fine-tuning epochs while maintaining consistent validation performance.
Figure 5. Training convergence curves showing stable learning across different datasets. The model converges within 50 fine-tuning epochs while maintaining consistent validation performance.
Applsci 15 09605 g005
Table 1. Performance comparison on knowledge tracing datasets with precision–recall analysis. Best results in bold, second-best underlined.
Table 1. Performance comparison on knowledge tracing datasets with precision–recall analysis. Best results in bold, second-best underlined.
MethodASSISTments (69.2% Correct)EdNet-KT1 (73.1% Correct)Junyi Academy (71.8% Correct)KDD Cup 2010 (67.4% Correct)
AUCRMSEF1PrecRecMAEAUCRMSEF1PrecRecMAEAUCRMSEF1PrecRecMAEAUCRMSEF1PrecRecMAE
BKT0.6730.4510.6120.5980.6270.3850.6410.4680.5980.5840.6130.3920.6580.4590.6050.5910.6200.3880.6510.4620.6010.5870.6160.391
DKT0.7420.4120.6810.6520.7130.3480.7230.4250.6650.6350.6970.3580.7380.4180.6780.6490.7100.3510.7290.4220.6710.6420.7020.355
DKVMN0.7580.3980.6950.6730.7190.3350.7410.4110.6820.6590.7070.3420.7540.4020.6920.6700.7160.3380.7460.4070.6870.6640.7110.341
SAKT0.7710.3850.7080.6890.7280.3210.7590.3960.6980.6780.7190.3280.7680.3880.7050.6860.7250.3240.7620.3920.7010.6810.7220.327
AKT0.7790.3790.7150.6980.7330.3160.7670.3900.7050.6870.7240.3220.7760.3820.7120.6950.7300.3190.7700.3860.7080.6900.7270.323
SAINT0.7830.3750.7190.7040.7350.3120.7710.3870.7090.6930.7260.3180.7800.3780.7160.7010.7320.3150.7740.3830.7120.6960.7290.320
simpleKT0.7890.3710.7240.7100.7390.3080.7770.3830.7140.6990.7300.3140.7860.3750.7210.7070.7360.3110.7800.3790.7170.7020.7330.316
LLM-KT0.8230.3480.7410.7210.7620.2890.8090.3610.7280.7080.7490.2950.8200.3520.7380.7180.7590.2920.8160.3560.7340.7140.7550.298
Ours0.8470.3250.7630.7580.7690.2720.8310.3360.7510.7450.7580.2840.8440.3290.7590.7540.7650.2760.8380.3320.7550.7500.7610.279
Table 2. Prerequisite discovery accuracy compared to expert annotations.
Table 2. Prerequisite discovery accuracy compared to expert annotations.
DatasetDiscovery AccuracyPrecisionRecallF1-Score
ASSISTments 2009–20100.8920.8850.8470.866
EdNet-KT10.8760.8690.8320.850
Junyi Academy0.9010.8940.8560.875
KDD Cup 20100.8880.8810.8430.862
Average0.8890.8820.8450.863
Table 3. Ablation study on architecture components (ASSISTments 2009–2010).
Table 3. Ablation study on architecture components (ASSISTments 2009–2010).
ConfigurationAUCRMSEF1Prerequisite Acc.
Full Model0.8470.3250.7630.892
w/o Cross-attention0.8210.3480.7410.856
w/o Causal pathway0.7980.3670.7190.000
w/o Temporal pathway0.7430.4020.6730.879
w/o Pre-training0.8290.3400.7480.871
w/o Expert knowledge0.8380.3310.7560.824
Table 4. Ablation study on loss function components (ASSISTments 2009–2010).
Table 4. Ablation study on loss function components (ASSISTments 2009–2010).
Loss ConfigurationAUCRMSEF1Prerequisite Acc.
L t o t a l (Full)0.8470.3250.7630.892
w/o  L c a u s a l 0.8340.3380.7510.743
w/o  L d a g 0.8190.3550.7360.651
w/o  L a l i g n 0.8410.3310.7580.884
Only L p r e d 0.7920.3730.7140.312
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Mai, N.T.; Cao, W.; Liu, W. Interpretable Knowledge Tracing via Transformer-Bayesian Hybrid Networks: Learning Temporal Dependencies and Causal Structures in Educational Data. Appl. Sci. 2025, 15, 9605. https://doi.org/10.3390/app15179605

AMA Style

Mai NT, Cao W, Liu W. Interpretable Knowledge Tracing via Transformer-Bayesian Hybrid Networks: Learning Temporal Dependencies and Causal Structures in Educational Data. Applied Sciences. 2025; 15(17):9605. https://doi.org/10.3390/app15179605

Chicago/Turabian Style

Mai, Nhu Tam, Wenyang Cao, and Wenhe Liu. 2025. "Interpretable Knowledge Tracing via Transformer-Bayesian Hybrid Networks: Learning Temporal Dependencies and Causal Structures in Educational Data" Applied Sciences 15, no. 17: 9605. https://doi.org/10.3390/app15179605

APA Style

Mai, N. T., Cao, W., & Liu, W. (2025). Interpretable Knowledge Tracing via Transformer-Bayesian Hybrid Networks: Learning Temporal Dependencies and Causal Structures in Educational Data. Applied Sciences, 15(17), 9605. https://doi.org/10.3390/app15179605

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop