You are currently viewing a new version of our website. To view the old version click .
Sensors
  • Article
  • Open Access

11 December 2025

LightEdu-Net: Noise-Resilient Multimodal Edge Intelligence for Student-State Monitoring in Resource-Limited Environments

,
,
,
,
,
and
1
National School of Development, Peking University, Beijing 100871, China
2
China Agricultural University, Beijing 100083, China
3
Artificial Intelligence Research Institute, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Multimodal Perception Modeling Based on Advanced Computational Technologies

Abstract

Multimodal perception for student-state monitoring is difficult to deploy in rural classrooms because sensors are noisy and computing resources are highly constrained. This work targets these challenges by enabling noise-resilient, multimodal, real-time student-state recognition on low-cost edge devices. We propose LightEdu-Net, a sensor-noise-adaptive Transformer-based multimodal network that integrates visual, physiological, and environmental signals in a unified lightweight architecture. The model incorporates three key components: a sensor noise adaptive module (SNAM) to suppress degraded sensor inputs, a cross-modal attention fusion module (CMAF) to capture complementary temporal dependencies across modalities, and an edge-aware knowledge distillation module (EAKD) to transfer knowledge from high-capacity teachers to an embedded-friendly student network. We construct a multimodal behavioral dataset from several rural schools and formulate student-state recognition as a multimodal classification task with explicit evaluation of noise robustness and edge deployability. Experiments show that LightEdu-Net achieves 92.4% accuracy with an F1-score of 91.4%, outperforming representative lightweight CNN and Transformer baselines. Under a noise level of 0.3, accuracy drops by only 1.1%, indicating strong robustness to sensor degradation. Deployment experiments further show that the model operates in real time on Jetson Nano with a latency of 42.8 ms (23.4 FPS) and maintains stable high accuracy on Raspberry Pi 4B and Intel NUC platforms. Beyond technical performance, the proposed system provides a low-cost and quantifiable mechanism for capturing fine-grained learning process indicators, offering new data support for educational economics studies on instructional efficiency and resource allocation in underdeveloped regions.

1. Introduction

Against the backdrop of rapidly advancing digital education, classroom instruction is moving from experience-driven practices toward data-driven and intelligent decision-making paradigms [1]. Real-time monitoring of student attention, fatigue, and classroom engagement can help teachers track learning dynamics, adjust instructional strategies, and provide an empirical basis for personalized support and resource allocation [2]. However, in rural and under-resourced areas, persistent issues such as insufficient teaching staff, limited classroom supervision, and delayed instructional feedback make it difficult to obtain such fine-grained behavioral information [3]. Although multimodal sensing and artificial intelligence techniques now enable automated analysis of student states from visual, physiological, and environmental data [4], most existing systems assume reliable sensors, stable networks, and cloud- or GPU-based computation, which rarely hold in these settings. Low-cost devices in real classrooms must cope with severe sensor noise, heterogeneous sampling rates, and strict constraints on computation, memory, and energy. As a result, intelligent tutoring systems, while theoretically valuable for large classes and high teaching workloads [5], remain hard to deploy in the very schools that would benefit most from them. This gap between methodological advances and deployable solutions motivates the design of LightEdu-Net, a multimodal perception framework explicitly tailored to noisy sensors and resource-constrained edge hardware in real rural classrooms.
Although intelligent tutoring systems have achieved notable methodological progress, many existing approaches remain difficult to apply in real classrooms, especially in rural schools [6]. Early work on student behavior monitoring typically relied on single-modal signals (e.g., camera images or wearable physiological data) and traditional machine learning models such as SVM, random forest, and KNN operating on handcrafted features [7,8]. These shallow representations struggle to capture the complexity and variability of classroom environments and are highly sensitive to illumination changes, occlusions, and sensor noise, leading to poor robustness in deployment [9,10]. Moreover, low-dimensional features are inadequate for describing temporal dynamics such as attention fluctuation and cumulative fatigue [11], and most pipelines assume desktop or server-side processing, which prevents real-time inference on the low-cost, low-power devices typically available in rural settings [12].
With the advancement of deep learning, student behaviors have increasingly been modeled using convolutional neural networks, Transformer architectures, and graph neural networks [13]. CNNs and Transformers can extract rich spatial–temporal features from visual, speech, and physiological signals, and multimodal fusion strategies such as early fusion, late fusion, and cross-attention improve recognition accuracy by exploiting complementary information across modalities [14,15]. However, most of these models are computationally heavy, rely on high-performance GPUs or cloud infrastructure, and assume relatively clean and homogeneous multimodal data [16,17]. In contrast, sensors in rural classrooms often produce noisy, low-fidelity, and asynchronously sampled signals, and cloud-based solutions may be unreliable due to limited bandwidth and raise privacy concerns [18]. Recent edge-oriented multimodal frameworks such as TinyM2Net, Farm-LightSeek, EduVQA, and lightweight maritime recognition systems demonstrate that compact models can be deployed on resource-constrained hardware in other domains [19,20], but there is still a lack of designs that jointly address sensor noise, heterogeneous educational modalities, and strict edge-computing constraints in real classrooms. This gap motivates our development of LightEdu-Net.
In this study, we focus on how to perform reliable multimodal student-state recognition in rural classrooms where sensing is noisy and computation must run on low-power edge devices. To address this problem, we propose LightEdu-Net, a lightweight multimodal framework designed to be both noise-resilient and deployable. The framework integrates hybrid feature encoding, cross-modal attention, and edge-aware optimization, enabling high-accuracy and low-latency inference on embedded platforms. The main contributions of this work are threefold.
  • We construct a multimodal behavioral dataset from rural schools, capturing visual, physiological, and environmental signals under real classroom conditions. The dataset explicitly contains noise artifacts, illumination changes, and motion disturbances, providing realistic supervision for evaluating robustness in resource-constrained environments.
  • We design LightEdu-Net, a sensor-noise adaptive multimodal network that combines a lightweight convolution–Transformer backbone with a sensor noise adaptive module (SNAM) and a cross-modal attention fusion module (CMAF). This architecture suppresses abnormal sensor inputs, aligns heterogeneous modalities, and extracts discriminative temporal–spatial features for student-state recognition.
  • We validate the framework through targeted experiments that reflect the core challenge: robustness and deployability. Results show that LightEdu-Net maintains high accuracy under increasing noise levels and operates in real time on edge devices such as Jetson Nano, Raspberry Pi 4B, and Intel NUC, demonstrating its suitability for practical deployment in resource-constrained educational settings.
  • From an educational economics perspective, the proposed framework provides low-cost and quantifiable learning process indicators for resource-constrained classrooms, enabling data-driven analysis of teaching efficiency and educational resource utilization in underdeveloped regions.
The remainder of this manuscript is organized as follows. Section 2 reviews related work on intelligent education, multimodal sensor fusion, edge computing, and the educational economics perspective that motivates our focus on rural and resource-constrained classrooms. Section 3 describes the materials and methods, including the construction of the multimodal educational behavior dataset, data preprocessing and augmentation strategies, and the proposed LightEdu-Net framework with its sensor noise adaptive module, cross-modal attention fusion module, and edge-aware knowledge distillation module. Section 4 presents the experimental setup and results, comparing LightEdu-Net with representative baselines, reporting ablation studies and noise-robustness analyses, and evaluating deployment performance on different edge devices, followed by a discussion of practical implications and remaining limitations. Section 5 concludes the paper and outlines directions for future research.

3. Materials and Method

3.1. Data Collection

The data collection process was conducted from March 2024 to January 2025 across authentic classroom environments in three rural primary schools and one county-level middle school in western China, as shown in Table 1 and Figure 1. A total of 12 multimodal acquisition terminals were deployed, integrating high-definition cameras, wearable physiological sensors, and environmental monitoring units. This setting reflects the conditions under which LightEdu-Net is expected to operate: highly heterogeneous signals, frequent illumination changes, and unstable network connectivity. Visual data were captured using 1080p cameras mounted at the front and sides of classrooms at 30 FPS, providing continuous observation of posture, facial expressions, and gaze direction. Automatic exposure control and illumination compensation were enabled to cope with alternating natural and artificial light. Each 40-min lesson yielded approximately 72,000 sequential frames per camera, and infrared auxiliary lighting was used to improve face and keypoint visibility under dim or reflective conditions.
Table 1. Statistics of the multimodal educational behavior dataset.
Figure 1. Overview of the multimodal data collection framework used in this study.
In total, data were collected from 268 students across Grades 3–9, with class sizes ranging from 32 to 48 students. The cohort includes both boys and girls and reflects the typical demographic composition of rural schools in western China in terms of age distribution and socio-economic background. Classroom subjects covered Chinese language, mathematics, English, and science, as well as a smaller number of sessions in moral education and integrated practical activities, thereby capturing a variety of instructional formats (lecture, discussion, group work, and individual practice).
Physiological signals were recorded through multi-channel wearable devices attached to students’ wrists or arms, providing heart rate (HR), electrodermal activity (EDA), temperature (TEMP), and blood oxygen saturation (SpO2) at 10 Hz. Data were streamed via Bluetooth 5.0 to a local edge node and synchronized with visual data. A caching and resume mechanism was used to address intermittent intranet connectivity, ensuring completeness. To reduce inter-individual variability, physiological signals were normalized and smoothed before model input. Environmental conditions were monitored at 1-second intervals using sensor modules placed at the front, rear, and window areas of the classroom. Measured variables included illumination (lux), noise level (dB), temperature (°C), and humidity (RH). These signals capture rapid fluctuations associated with instructional activities, such as reading aloud, classroom transitions, or external disturbances.
All data collection procedures were approved by the institutional ethics committee of the collaborating university and the local education authorities. Written informed consent was obtained from school administrators, teachers, and the legal guardians of all participating students, and assent was obtained from the students themselves. During acquisition and subsequent processing, personal identifiers were removed or anonymized, and only aggregated indicators are used in analysis and system feedback to protect privacy.

3.2. Dataset Processing

Reliable multimodal modeling requires that input signals exhibit comparable scale, temporal consistency, and noise stability. Classroom environments inherently contain illumination fluctuation, partial occlusions, loose-wearing artifacts, and motion-induced perturbations. To ensure that the downstream lightweight architecture receives stable inputs, we apply a hybrid processing strategy that combines basic signal filtering and normalization with targeted augmentation. These operations are not intended as contribution points, but rather provide necessary signal conditioning for the subsequent SNAM and CMAF modules to operate effectively under resource constraints.

3.2.1. Time-Series Data

For physiological signals, filtering represents the first step to suppress power-line interference and high-frequency noise. Using linear time-invariant filtering expressed in convolutional form, the filtered signal is given by
y ( t ) = + x ( τ ) h ( t τ )   d τ ,
where x ( t ) is the raw signal and h ( t ) the impulse response of the filter. In practice, band-pass filters retain cognitive-related frequencies (e.g., 0.1–40 Hz) while removing out-of-band disturbances commonly observed in wearable classroom measurements.
Since inconsistent magnitudes across modalities may disturb optimization, all physiological and visual features are standardized. The standardized signal is defined as
x ^ = x μ σ ,
where μ and σ denote mean and standard deviation. To accommodate non-stationary statistics in long classroom recordings, a sliding-window standardization strategy is adopted:
x ^ t = x t μ t ( W ) σ t ( W ) ,
where μ t ( W ) and σ t ( W ) are the window-based mean and variance computed within W samples.
Sudden spikes caused by electrode displacement or motion artifacts are further smoothed by windowed averaging,
x ˜ t = 1 | W | i = t k t + k x i ,
where | W | = 2 k + 1 . This preserves global trends while attenuating high-frequency jitter, producing temporally coherent features that SNAM later refines via learnable denoising.

3.2.2. Image Data

Visual signals are augmented to simulate realistic disturbances. Brightness perturbation is modeled by a linear intensity transformation:
I = γ I + δ ,
where γ scales global brightness and δ models ambient light shift. This augmentation improves robustness to lighting variance often observed in classrooms with windows or ceiling lights.
Occlusion is simulated using a Cutout strategy. A random rectangular region Ω is selected and replaced:
I ( x , y ) = 0 , ( x , y ) Ω , I ( x , y ) , otherwise .
This encourages the backbone to rely on multiple cues rather than single facial or postural regions, improving resilience when students move, lean, or are partially blocked.

3.3. Data Enhancement

To reduce over-sensitivity to sampling jitter, a temporal perturbation strategy is introduced. For each time step, a small offset ϵ t is added:
x t = x t + ϵ t ,
where ϵ t follows a zero-mean, small-variance distribution. Such temporal jittering prevents overfitting to rigid time alignment and simulates realistic sensor drift.
Multimodal signals often differ in sampling rate. To allow cross-modal attention to operate on a unified temporal axis, all modalities are resampled via interpolation. For an original signal x ( t ) and resampled timeline t , we compute
x ( t ) = Interp ( x ( t ) , t ) ,
where Interp ( · ) denotes linear interpolation. This temporal synchronization ensures that CMAF processes inputs within consistent windows, and that modality interactions reflect true behavioral dynamics rather than sampling discrepancies.
Taken together, these preprocessing steps provide stable and comparable multimodal inputs. They do not constitute novel algorithmic contributions, but they are essential to enable SNAM to learn frequency-domain denoising and CMAF to exploit cross-modal mutual information under noisy classroom conditions and edge-computing resource limits.

3.4. Proposed Method

3.4.1. Overall

We propose LightEdu-Net, a noise-resilient multimodal student-state recognition framework designed for edge deployment in real classrooms. The model takes temporally aligned visual, physiological, and environmental features as input and passes them through three lightweight modality-specific encoders: a convolution–Transformer hybrid backbone for visual signals, a 1D convolutional–recurrent backbone for physiological signals, and multi-channel convolutions for environmental measurements. The encoded features are then refined by the sensor noise adaptive module (SNAM), which performs learnable band-pass denoising and channel reweighting to improve the signal-to-noise ratio before fusion. The denoised modality features are subsequently fed into the cross-modal attention fusion (CMAF) layer, where intra-modal attention emphasizes salient temporal segments and inter-modal attention captures complementary dependencies across modalities, yielding a compact joint representation. On top of this fused representation, a lightweight prediction head produces student-state labels suitable for classroom monitoring on embedded devices. During training, an edge-aware knowledge distillation (EAKD) scheme is applied, in which a high-capacity teacher network guides the LightEdu-Net student to approximate its decision boundaries while keeping the student architecture small and quantization-friendly. At inference time, only the student network is used: data flow sequentially through the encoders, SNAM, CMAF, and the prediction head, enabling robust multimodal perception with low latency on low-power edge hardware.

3.4.2. Sensor Noise Adaptive Module

The sensor noise adaptive module (SNAM) is inserted between the modality-specific encoders and the multimodal fusion layer, following the noise-rectification pipeline illustrated in Figure 2. For each modality, the encoder output is first projected to a compact temporal feature sequence. Concretely, the visual backbone produces a tensor V R T × 14 × 14 × 64 , which is passed through a 1 × 1 pointwise convolution (32 output channels) and global spatial average pooling to obtain V ˜ R T × 32 . The physiological branch yields P ˜ R T × 64 through one-dimensional convolutions and GRU layers, and the environmental branch produces E ˜ R T × 32 via multi-channel convolutions. SNAM processes each modality feature sequence X { V ˜ , P ˜ , E ˜ } independently with shared parameters, so that the same denoising mechanism is applied to visual, physiological, and environmental signals.
Figure 2. This figure illustrates the overall workflow of the sensor noise adaptive module (SNAM).
For a given modality, we first apply a lightweight temporal encoder to capture local dynamics and smooth high-frequency fluctuations. A depthwise–separable temporal convolution stack DS     TCN is used,
Z = DS     TCN ( X ) ,
where two depthwise–separable Conv1d layers (kernel sizes k = 5 and k = 3 , dilations d = 1 and d = 2 ) with LayerNorm and SiLU activation preserve the shape Z R T × C (with C { 32 , 64 } ). To further suppress noise in the spectral domain, we perform a learnable band-pass filtering. Let F t denote the FFT along the temporal axis and Z ^ = F t ( Z ) the corresponding spectrum. We define a parameterized mask
M ( ω ) = σ s ( ω ω low )   σ s ( ω high ω ) ,
where ω low and ω high are learnable cutoff frequencies, s > 0 is a learnable slope, and σ ( · ) denotes the sigmoid function. The filtered features in the time domain are then obtained by
U = F t 1 M Z ^ ,
which attenuates frequency components outside the band of interest while preserving the temporal resolution.
Residual noise is often concentrated in a subset of channels, especially under sensor jitter or loose wearing. To adaptively down-weight such channels, we introduce a channel reweighting unit. Global temporal average pooling GAP t followed by two bottleneck fully connected layers produces a set of channel importance scores w R C :
w = σ W 2   ϕ W 1   GAP t ( U ) ,           W 1 R C r × C ,   W 2 R C × C r ,
where ϕ ( · ) is the SiLU activation and r = 8 is the reduction ratio. The final SNAM output is
Y = U w ,
which preserves the temporal dimension T and channel dimension C of the input. Channels dominated by noise receive smaller weights, whereas informative channels are retained.
To stabilize training and encourage meaningful masks, we introduce two auxiliary objectives. A simple spectral regularization penalizes excessive suppression of useful bands,
L band = ( 1 M ) Z ^ 2 2 ,
where M is the learnable frequency mask and Z ^ is the temporal spectrum of the intermediate features. In addition, when teacher features or clean augmented views are available, we encourage the denoised features to stay close to a reference representation. Let Y R T × C denote the SNAM output for the current modality, where Y t , c is the value at time step t and channel c, and let T R T × C denote the corresponding teacher features (or clean augmented views) extracted at the same layer and temporal resolution, with T t , c aligned to the same position. The alignment loss is then defined as
L align = 1 T C t , c | Y t , c T t , c | .
In the overall LightEdu-Net framework, the modality-wise outputs Y ( v ) R T × 32 , Y ( p ) R T × 64 , and Y ( e ) R T × 32 are concatenated and projected to a unified embedding dimension before being fed into the cross-modal attention fusion module. In this rectification-before-fusion design, SNAM improves the signal-to-noise characteristics of each modality so that CMAF can allocate its attention capacity to physically meaningful temporal patterns rather than corrupted sensor artifacts, which is particularly important in noisy, resource-constrained classroom environments.

3.4.3. Cross-Modal Attention Fusion

The cross-modal attention fusion (CMAF) module is designed to model interactive representations across modalities rather than within a single modality. Conventional self-attention computes dependencies using query, key, and value matrices ( Q , K , V ) derived from the same input, and thus primarily captures intra-modal structure. In contrast, CMAF explicitly exchanges information among visual, physiological, and environmental branches, allowing each modality to attend to others in a shared embedding space.
As illustrated in Figure 3, CMAF operates on the SNAM-processed features and consists of stacked Transformer-style blocks. Each block contains intra-modal self-attention, inter-modal cross-attention, residual connections with LayerNorm, and a two-layer multilayer perceptron (MLP). The visual, physiological, and environmental branches receive inputs of shape T × C m , where T denotes the sequence length and C m { 32 , 64 , 32 } is the channel dimension for the three modalities, respectively. All modality-specific features are first projected to a unified embedding dimension d = 96 so that they can be processed jointly. In our implementation, CMAF uses three such blocks with h = 8 attention heads per block and head dimension d h = d / h = 12 .
Figure 3. Architecture of the cross-modal attention fusion (CMAF) module.
Formally, let X v , X p , X e R T × d denote the visual, physiological, and environmental features after linear projection. For notational simplicity, we describe the update of the visual branch using cross-attention to the physiological and environmental branches; the other branches are handled analogously. We define linear projections
Q v = X v W q ,       K p = X p W k ,       V p = X p W v ,       K e = X e W k ,       V e = X e W v ,
where W q , W k , W v R d × d h are learnable projection matrices shared across modalities within a block. Cross-modal attention from modality p to modality v is computed as
A v p = Softmax   Q v K p d h V p ,
and similarly for attention from modality e to modality v,
A v e = Softmax   Q v K e d h V e .
The Softmax ( · ) operator is applied row-wise over the temporal dimension to yield attention weights for each time step. The updated visual representation is given by
Y v = X v + α p A v p + α e A v e ,
where α p and α e are learnable scalar coefficients satisfying α p + α e = 1 and controlling the relative contribution of physiological and environmental information. In this way, the visual branch can adaptively integrate cues such as physiological rhythm changes and environmental disturbances at each time step. The same procedure is used to update Y p and Y e by exchanging roles of the modalities. After normalization and concatenation along the channel axis, we obtain a fused representation F R T × 3 d , which is then projected back to T × d through an MLP and forwarded to the final prediction head.
From an information-theoretic perspective, CMAF is designed to increase the mutual information among modalities. Let I ( X ; Y ) denote the mutual information between two random variables. While single-modal self-attention mainly enhances I ( X v ; X v ) , I ( X p ; X p ) , and I ( X e ; X e ) , CMAF aims to strengthen cross-modal terms such as I ( X v ; X p ) , I ( X v ; X e ) , and I ( X p ; X e ) . A common variational lower bound for I ( X i ; X j ) can be written as
I ( X i ; X j ) E ( x i , x j ) log D θ ( x i , x j ) log E x i E x j exp ( D θ ( x i , x j ) ) ,
where D θ ( · , · ) is a learnable critic function. In our setting, the attention score matrices implicitly play a similar role to D θ by assigning higher weights to modality pairs that are more informative for the task, thereby encouraging the network to align semantically consistent patterns across modalities.
When combined with the preceding SNAM module, CMAF operates in a rectification-before-alignment manner. SNAM first enhances the signal-to-noise ratio of each modality and suppresses corrupted channels, and CMAF then performs cross-modal alignment and fusion on these rectified features. Together with the edge-aware knowledge distillation mechanism, which regularizes the fused representations via a high-capacity teacher network, this hierarchical design yields a robust semantic-fusion core that can handle noisy, asynchronous, and heterogeneous multimodal signals in resource-constrained educational environments.

3.4.4. Edge-Aware Knowledge Distillation

The edge-aware knowledge distillation (EAKD) module aims to balance lightweight feature extraction and high-accuracy prediction on edge devices by introducing a dual-teacher distillation strategy. In this way, the lightweight student network deployed on devices such as Jetson Nano and Raspberry Pi can inherit the decision capability of more expressive teacher models trained on high-performance servers. As illustrated in Figure 4, EAKD involves two teachers and one student: a CNN teacher ( T CNN ), a ViT teacher, and a lightweight CNN student (denoted as S CNN ). The teacher models possess stronger local and global modeling capacity, whereas the student model contains only essential convolutional and attention layers and is designed to meet edge-computing constraints.
Figure 4. Illustration of the edge-aware knowledge distillation (EAKD) module.
To adaptively combine the two teachers, we adopt a discrepancy-aware teacher weighting (DATW) scheme. For each teacher k { CNN , ViT } , we define a confidence score
C k ( x ) = max c   p T k , c ( x ) ,
that reflects how certain the teacher is about its prediction, and a discrepancy measure
D k ( x ) = z T k ( x ) z S ( x ) 2 ,
that reflects how far the student currently deviates from that teacher in logit space. The unnormalized teacher weight is then given by
ω ˜ T k ( x ) = C k ( x ) D k ( x ) + ϵ ,
where ϵ > 0 is a small constant to avoid division by zero. The normalized weights are obtained as
ω T k ( x ) = ω ˜ T k ( x ) ω ˜ T CNN ( x ) + ω ˜ T ViT ( x ) ,
so that ω T CNN ( x ) + ω T ViT ( x ) = 1 for all x. The combined teacher target, illustrated as Z t a r g e t ( x ) in Figure 4, is a convex combination of the two teacher logits:
z target ( x ) = ω T CNN ( x )   z T CNN ( x ) + ω T ViT ( x )   z T ViT ( x ) .
The EAKD loss integrates logit matching and distribution-level distillation. The logit-level term encourages the student to approximate the combined teacher target Z t a r g e t ( x ) ,
L logit = z S ( x ) z target ( x ) 2 2 ,
and the distribution-level term uses Kullback–Leibler divergence to pull the student probabilities toward the teacher probabilities,
L KL = k { CNN , ViT } KL p T k ( x ) p S ( x ) .
The overall EAKD objective is then written as
L EAKD = η   L logit + ( 1 η )   L KL ,
where η [ 0 , 1 ] is a balancing coefficient tuned on the validation set. When the student already matches a given teacher well (small D k ( x ) ), the corresponding weight ω T k ( x ) increases, and that teacher dominates the distillation signal; when the discrepancy is large or the teacher is uncertain (low C k ( x ) ), its influence is automatically reduced.
By combining structurally different teachers (CNN and ViT), explicitly modeling teacher confidence and student–teacher discrepancy, and constraining the student to a lightweight architecture, EAKD enables LightEdu-Net to approach the decision quality of high-capacity models while maintaining low computational cost. In the intelligent tutoring task considered here, this module increases recognition accuracy on low-power platforms, reduces inference latency, and alleviates dependence on high-bandwidth transmission, thereby achieving a favorable balance between computation and intelligence for edge-deployed educational models.

4. Results and Discussion

4.1. Experimental Settings

4.1.1. Platforms and Parameters

The experimental hardware platforms employed in this study encompass three typical categories of edge and near-edge computing devices to comprehensively validate the model’s adaptability and performance under varying computational capacities. NVIDIA Jetson Nano, equipped with a 128-core Maxwell GPU, serves as a representative low-power embedded platform commonly available in rural schools. Raspberry Pi 4B contains a quad-core ARM Cortex-A72 CPU and performs inference purely on processor resources, aligning with extremely low-compute deployment scenarios. Intel NUC, as a lightweight desktop-class system, integrates a multi-core CPU with higher clock frequency, enabling evaluation of real-time performance and energy consumption under moderate computational conditions. Deployment across this hardware spectrum provides a realistic demonstration of the computational heterogeneity and operational demands likely encountered in real-world educational environments.
Regarding software configuration, the experiments are conducted using PyTorch 2.2 as the primary deep learning framework for model construction and training. GPU-accelerated inference is enabled through CUDA 12.1 to ensure efficient computation for convolutional and Transformer-based operations. Python 3.10 is adopted as the primary programming language to support data preprocessing, model training, visualization, and general experimental workflow management. All components are executed within an Ubuntu environment to ensure dependency compatibility and stable model deployment, thereby facilitating consistent migration to edge devices. For hyperparameter configuration, the constructed dataset is partitioned into training, validation, and test sets with a ratio of 7:2:1, ensuring adequate multimodal feature learning during training while preserving sufficient samples for performance assessment. A mini-batch strategy with batch size = 32 is adopted to balance memory consumption and stable gradient updates. The learning rate is set to 1 × 10 4 and optimized with the Adam optimizer to achieve adaptive gradient updates while maintaining stable convergence. A total of 100 training epochs is used to allow sufficient convergence time. To further evaluate the generalization capability under multimodal input conditions, a five-fold cross-validation strategy is incorporated. Specifically, the dataset is randomly divided into five subsets, with four subsets used for training and one for validation in each iteration. After five iterations, the averaged performance metrics are obtained, reducing bias introduced by data partitioning and yielding a more reliable estimation of model performance.

4.1.2. Baseline Models and Evaluation Metrics

A set of representative lightweight and multimodal baseline models is selected for comparison, including MobileNetV3 + SVM [45,46], ResNet [47], EfficientNet-lite [48], Tiny-Transformer [49], and CLIP-Fusion [50]. MobileNetV3 + SVM combines an efficient feature extraction backbone with a traditional classifier, offering advantages in structural simplicity and training stability. ResNet, as a classical residual architecture, provides strong robustness and consistent performance. EfficientNet-lite adopts a compound scaling strategy across depth, width, and resolution, achieving high parameter efficiency with relatively strong accuracy. Tiny-Transformer employs a lightweight self-attention structure by reducing attention heads and channel dimensions, enabling long-range dependency modeling under constrained computational budgets. CLIP-Fusion exploits pretrained vision–language alignment models as multimodal encoders, enabling feature concatenation or matching across modalities and providing strong cross-modal transferability. For all baselines, network depth, width, learning rate, and batch size are selected according to their original papers or official lightweight configurations and then fine-tuned on a held-out validation set to ensure a fair comparison under similar computational budgets.
The evaluation in this study covers classification metrics (accuracy, precision, recall, F1-score), efficiency metrics (model size, FLOPs, inference latency, power consumption), and robustness metrics (performance degradation under noise, expressed as Δ A c c u r a c y ), providing a comprehensive assessment of model suitability in resource-constrained educational environments. The mathematical definitions of the evaluation metrics are as follows:
A c c u r a c y = T P + T N T P + T N + F P + F N ,
P r e c i s i o n = T P T P + F P ,
R e c a l l = T P T P + F N ,
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l ,
Δ A c c u r a c y = A c c u r a c y c l e a n A c c u r a c y n o i s e .
The symbols in these equations are defined as follows: T P denotes the number of true positives, representing instances correctly identified as positive; T N denotes true negatives, referring to instances correctly identified as negative; F P denotes false positives, referring to negative instances incorrectly classified as positive; F N denotes false negatives, representing positive instances incorrectly classified as negative. A c c u r a c y c l e a n denotes the accuracy obtained under noise-free conditions, whereas A c c u r a c y n o i s e represents the accuracy under specific noise perturbations, thereby reflecting the model’s sensitivity to noise.

4.2. Comparison of LightEdu-Net and Baseline Models

The objective of this experiment is to verify whether the proposed LightEdu-Net can maintain high recognition accuracy while remaining lightweight and computationally efficient in multimodal student-state recognition tasks, thereby satisfying real-time inference demands on resource-constrained edge devices. By comparing LightEdu-Net with five representative lightweight and multimodal baseline models, a comprehensive evaluation of the trade-offs among classification capacity, representational ability, and computational efficiency under different architectural designs can be achieved.
As shown in Table 2 and Figure 5, conventional convolution-based models such as MobileNetV3 + SVM and ResNet18 exhibit considerably lower accuracy and recall. Their feature extraction capacity is restricted by shallow convolutional kernels with limited receptive fields, which are insufficient for capturing long-range temporal dependencies and subtle cross-modal cues, such as gradual attention changes or cumulative physiological fatigue. EfficientNet-lite and Tiny-Transformer provide stronger representational capacity but still fall short in explicitly modeling and exploiting complementary information across visual, physiological, and environmental inputs. CLIP-Fusion, as a large pretrained multimodal model, achieves relatively high accuracy; however, its large parameter scale and high FLOPs make deployment on edge devices impractical in rural classroom settings. In contrast, LightEdu-Net achieves the best performance across all metrics, with particularly notable gains in F1-score and recall, because SNAM improves the signal-to-noise ratio at the feature level, CMAF explicitly aligns and fuses modalities through cross-attention, and EAKD transfers knowledge from high-capacity teachers into a compact student. This combination yields a model that is both more discriminative for complex multimodal patterns and better suited to low-power edge deployment.
Table 2. Comparison of LightEdu-Net and baseline models in classification performance and efficiency.
Figure 5. The violin plot of accuracy comparision results across different models.
From a theoretical perspective, performance differences arise from variations in intrinsic structural design and mathematical representational capacity. MobileNetV3 + SVM and ResNet18 rely on local convolutional kernels and lack mechanisms for long-range or multimodal dependency modeling, resulting in representational bottlenecks when handling high-dimensional heterogeneous features from visual, physiological, and environmental signals. EfficientNet-lite increases network capacity through compound scaling, yet it primarily captures patterns within the visual domain, limiting cross-modal fusion ability. Tiny-Transformer can model long-range dependencies but constrains attention computation within the visual modality, preventing deeper multimodal interaction. CLIP-Fusion optimizes a cross-modal embedding space through large-scale pretraining but suffers from high-dimensional embeddings and dense attention structures, resulting in prohibitive computational cost for edge inference. LightEdu-Net achieves superior performance by incorporating a sensor noise adaptive module (SNAM) to enhance feature stability, using a cross-modal attention fusion (CMAF) module to maximize mutual information across modalities, and adopting an edge-aware knowledge distillation (EAKD) mechanism to compress the teacher model’s decision boundaries into a lightweight student network. This combination allows the lightweight model to approximate the expressiveness of a high-capacity model while maintaining efficiency, thereby achieving optimal performance under resource constraints.

4.3. Ablation Study

This experiment evaluates whether the overall performance of LightEdu-Net relies on its core modules and quantifies the contributions of each component to multimodal recognition capability and noise robustness.
As shown in Table 3, the three unimodal variants (Visual-only, Physiological-only, and Environmental-only) all perform noticeably worse than the full multimodal LightEdu-Net, indicating that each single modality provides only partial information and that robust student-state recognition benefits from integrating complementary cues. Visual-only typically captures posture and facial dynamics but lacks direct information about internal arousal and external disturbances; Physiological-only is sensitive to fatigue and stress but cannot distinguish specific classroom behaviors; Environmental-only reflects illumination and noise changes but is insufficient for fine-grained engagement recognition. The performance of these variants highlights the necessity of multimodal fusion. retaining only the backbone network leads to a significant decline across all performance metrics, demonstrating that shallow convolutional or limited long-range modeling is insufficient for capturing deep multimodal dependencies. Removing SNAM results in the most severe performance degradation, especially under noise level 0.3 , indicating that illumination variations, sensor jitter, and physiological noise—frequently present in real-world rural classrooms—substantially affect feature stability. Removing CMAF moderately reduces performance, confirming that cross-modal attention plays a critical role in exploiting complementary information among modalities. Removing EAKD leads to a mild performance drop, showing that the distillation mechanism effectively enhances the expressive capacity of the lightweight student model without increasing computational cost. From a theoretical standpoint, SNAM stabilizes temporal and spectral characteristics by reducing distributional shifts from noise perturbations. CMAF maximizes mutual information across modalities, equivalent to learning conditional dependency structures in a high-dimensional feature space. Removing CMAF degrades joint representations into weakly coupled features. EAKD provides a richer teacher-defined feature space, allowing the lightweight model to maintain high-level decision boundaries; removing it reduces classification granularity and generalization capability.
Table 3. Ablation study of LightEdu-Net showing the contribution of each module and input configuration.

4.4. Deployment Performance

This experiment evaluates the deployability and stability of LightEdu-Net under different computational resources by testing latency, throughput, power consumption, and accuracy on three representative edge and near-edge platforms. The goal is to determine whether the model satisfies real-time, energy-efficient, and reliable recognition requirements in practical educational environments.
As shown in Table 4, Jetson Nano, despite limited computational capability, achieves relatively high throughput and low latency due to the lightweight convolutional architecture and compact fused features, while maintaining server-level accuracy. Raspberry Pi 4B, relying solely on CPU inference, exhibits higher latency but retains stable accuracy under low power consumption, confirming that the model design minimizes convolutional depth and attention complexity to accommodate CPU-based sequential operations. Intel NUC, with higher clock frequencies and multi-core processing, provides the lowest latency and highest throughput, indicating strong scalability on more capable platforms.
Table 4. Deployment performance of LightEdu-Net across different edge and near-edge platforms.

4.5. Discussion

4.5.1. Effectiveness of LightEdu-Net on Core Challenges

The experimental results validate that LightEdu-Net effectively addresses the key challenges of noisy sensing, multimodal fusion, and edge deployability in rural classroom environments. As shown in Table 2 and Figure 5, LightEdu-Net achieves the highest accuracy, precision, recall, and F1-score among all compared methods while maintaining a moderate model size and FLOPs, indicating a favorable balance between representational capacity and computational efficiency. The ablation results in Table 3 further clarify the contribution of each core component: the “Backbone Only” variant exhibits pronounced performance degradation, removing SNAM causes the largest drop and severe sensitivity to noise, removing CMAF weakens the exploitation of complementary multimodal information, and removing EAKD reduces performance despite unchanged inference cost. Together with the deployment results in Table 4, which show that LightEdu-Net meets latency, throughput, and power constraints on Jetson Nano, Raspberry Pi 4B, and Intel NUC, these findings confirm that SNAM, CMAF, and EAKD work synergistically to provide robust, accurate, and deployable multimodal perception under realistic edge-computing conditions.
From an educational perspective, field feedback from teachers suggests that the system’s continuous indicators of attention, fatigue, and participation provide actionable information that is difficult to obtain in large classes, enabling timely adjustments in pacing, interaction strategies, and classroom management. This indicates that LightEdu-Net can help increase the effective use of limited teaching resources and improve instructional efficiency in rural and under-resourced schools. At the same time, the sensitivity of multimodal behavioral and physiological data highlights the need for privacy-preserving extensions, such as federated learning, in which model updates rather than raw data are shared across schools. Future work will also involve more fine-grained power-consumption measurements under varying workloads (e.g., different class sizes and sampling rates) to better assess long-term sustainability. Overall, the current experiments show that the core challenges identified in this study—sensor noise, multimodal fusion, and edge deployment—are substantially mitigated, and that LightEdu-Net provides a practical and scalable foundation for multimodal perception in real classroom environments.

4.5.2. Classroom Usage, Annotation Protocol, and Educational Impact

Beyond quantitative performance, it is important to clarify how LightEdu-Net operates at the behavioral level and how its outputs are obtained. In our implementation, the model predicts discrete engagement-related states, including multi-level attention labels (high/medium/low), fatigue levels (alert/mild/high), and participation indicators (active/passive), derived from the fused multimodal representation produced by CMAF and generated at regular time intervals for each student and the class as a whole. Ground truth labels are constructed through an offline annotation protocol: synchronized video, physiological signals, and classroom logs are reviewed by trained annotators with educational backgrounds, who assign attention, fatigue, and participation levels based on observable cues such as gaze orientation, posture, facial expressions, interaction frequency, and characteristic changes in physiological signals. Each segment is labeled independently by at least two annotators, and disagreements are resolved by consensus, ensuring that the supervisory signals correspond to pedagogically meaningful states rather than purely low-level patterns.
In real deployments, the outputs of LightEdu-Net are delivered to teachers via a lightweight classroom dashboard designed to impose minimal cognitive load. At the individual level, the interface presents smoothed timelines of attention and fatigue using simple color-coded bars without exposing raw physiological values, while at the class level it provides aggregated indicators (e.g., proportion of students with low attention, average fatigue level) as trend curves or compact heatmaps. This allows teachers to react during class to sudden drops in engagement by adjusting pacing, questioning strategies, or instructional modalities, and to reflect after class on when and where engagement declined to inform lesson redesign and targeted support. Over longer periods, aggregated statistics across lessons help administrators identify structural inefficiencies (such as overloaded time slots or problematic subject scheduling) and optimize timetables and resource allocation. From an educational economics perspective, this workflow converts previously unobservable behavioral processes into measurable process data, enabling more efficient use of limited teaching resources and supporting equity-oriented decision-making in resource-constrained schools, with modest hardware cost due to edge-based deployment.

4.6. Limitations and Future Work

Although the proposed study provides a comprehensive evaluation of multimodal student-state recognition and edge deployment, several limitations remain with respect to the core challenges of noisy sensing, multimodal fusion, and resource-constrained inference, and these should be addressed in future research. First, the dataset was collected from multiple rural primary schools and one county-level middle school and primarily reflects typical classroom behaviors in this specific regional and cultural context. While the experiments demonstrate that LightEdu-Net is robust to sensor noise and heterogeneous modalities under these conditions, its generalization ability has not yet been fully validated for other subject types (e.g., laboratory sessions, project-based learning), special learning behaviors (e.g., students with learning difficulties), or classrooms with different cultural, linguistic, or curriculum backgrounds. In addition, the current dataset, though multimodal, still focuses on a limited set of physiological and environmental variables. Future work should expand data collection across more diverse schools and regions, include richer behavioral and contextual modalities, and explore unified annotation and preprocessing standards so that the noise-resilient and fusion mechanisms of LightEdu-Net can be systematically evaluated and adapted in broader settings. Second, although the proposed architecture achieves real-time inference on mainstream edge platforms such as Jetson Nano, Raspberry Pi 4B, and Intel NUC, the current design is primarily optimized for these mid-range devices. Under ultra-low-power or highly constrained hardware (e.g., wearable devices, low-end tablets, or legacy educational terminals), the computational cost of cross-modal attention and feature alignment remains a bottleneck, and the latency may no longer satisfy strict real-time requirements. The present SNAM, CMAF, and EAKD modules effectively alleviate the original deployment challenge within the tested resource envelope, but they do not yet provide a principled solution for the extreme edge regime. Future research should therefore investigate more efficient low-rank and sparse attention structures, event-driven or adaptive-inference strategies that selectively update only salient segments, and lighter distillation frameworks tailored to on-device training or continual learning. These directions will be important for further reducing computational and energy overhead while preserving the robustness and multimodal discriminative capability demonstrated in this work.

5. Conclusions

This study addresses a core challenge in deploying intelligent tutoring systems within resource-constrained educational environments, namely, how to perform reliable multimodal student-state recognition under sensor noise, heterogeneous inputs, and limited computational resources. To tackle this problem, we propose LightEdu-Net, a lightweight multimodal framework that combines a sensor-noise adaptive module to stabilize feature quality, a cross-modal attention fusion mechanism to enhance information interaction across visual, physiological, and environmental signals, and an edge-aware knowledge distillation strategy to preserve discriminative capacity on low-power devices. Experimental evaluations demonstrate clear progress in overcoming the identified challenges: LightEdu-Net achieves 92.4% accuracy with balanced precision and recall and maintains stable performance under noise perturbation while supporting real-time inference on Jetson Nano, Raspberry Pi 4B, and Intel NUC. These results confirm that the proposed solutions improve robustness, fusion expressiveness, and deployability, providing a viable pathway for multimodal perception in rural and under-resourced classrooms.
Despite these advances, several limitations remain when addressing the deployment challenge in its full scope. The current approach has been validated primarily on a limited set of schools and typical classroom settings, and its generalization to diverse learning activities, cultural contexts, or extreme hardware constraints has not been fully established. Moreover, although latency and power consumption are acceptable for mainstream edge devices, further reductions will be required for ultra-low-power platforms such as wearable systems or legacy terminals. Future work will therefore focus on expanding multimodal datasets across broader regions, investigating more efficient attention and distillation mechanisms, and exploring privacy-preserving training strategies such as federated learning to enable scalable deployment across multiple schools. These directions will further enhance the practicality and sustainability of LightEdu-Net and accelerate its application in real-world educational environments.

Author Contributions

Conceptualization, C.H., Y.C., B.Z., X.C. and Y.Z.; Data curation, Z.Z. and J.Z.; Funding acquisition, Y.Z.; Methodology, C.H., Y.C., B.Z. and X.C.; Project administration, Y.Z.; Resources, Z.Z. and J.Z.; Software, C.H., Y.C., B.Z. and X.C.; Supervision, Y.Z.; Validation, J.Z.; Visualization, Z.Z.; Writing—original draft, C.H., Y.C., B.Z., X.C., Z.Z., J.Z. and Y.Z.; C.H., Y.C., B.Z. and X.C. contributed equally to this work. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Natural Science Foundation of China grant number 61202479.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kayumova, M.; Akbarova, S.; Azizova, M.; Hojiyeva, I.; Karimova, F.; Makhmudova, F. Data Driven Teaching and Real Time Decision Making in Education Management. In Proceedings of the 8th International Conference on Future Networks & Distributed Systems, Marrakech, Morocoo, 11–12 December 2024; pp. 756–763. [Google Scholar]
  2. Trabelsi, Z.; Alnajjar, F.; Parambil, M.M.A.; Gochoo, M.; Ali, L. Real-time attention monitoring system for classroom: A deep learning approach for student’s behavior recognition. Big Data Cogn. Comput. 2023, 7, 48. [Google Scholar] [CrossRef]
  3. Wei, N.; Yang, F.; Muthu, B.; Shanthini, A. Human machine interaction-assisted smart educational system for rural children. Comput. Electr. Eng. 2022, 99, 107812. [Google Scholar] [CrossRef]
  4. TS, A.; Guddeti, R.M.R. Automatic detection of students’ affective states in classroom environment using hybrid convolutional neural networks. Educ. Inf. Technol. 2020, 25, 1387–1415. [Google Scholar]
  5. Nye, B.D. Intelligent tutoring systems by and for the developing world: A review of trends and approaches for educational technology in a global context. Int. J. Artif. Intell. Educ. 2015, 25, 177–203. [Google Scholar] [CrossRef]
  6. Hwang, K.A.; Yang, C.H. Automated inattention and fatigue detection system in distance education for elementary school students. J. Educ. Technol. Soc. 2009, 12, 22–35. [Google Scholar]
  7. Pabba, C.; Kumar, P. An intelligent system for monitoring students’ engagement in large classroom teaching through facial expression recognition. Expert Syst. 2022, 39, e12839. [Google Scholar] [CrossRef]
  8. Alsayigh, H.K.S.; Khidhir, A.S.M. Using IoT to predict student attention levels in e-learning classes: A review. AIP Conf. Proc. 2024, 3091, 040009. [Google Scholar]
  9. Gupta, S.; Kumar, P.; Tekchandani, R. A machine learning-based decision support system for temporal human cognitive state estimation during online education using wearable physiological monitoring devices. Decis. Anal. J. 2023, 8, 100280. [Google Scholar] [CrossRef]
  10. Chen, S.; Hu, Z.; Li, S.; Hu, X.; Liu, G.; Wen, W. Autonomic nervous pattern recognition of students’ learning states in real classroom situation. IEEE Trans. Comput. Soc. Syst. 2021, 10, 220–233. [Google Scholar] [CrossRef]
  11. Miao, Y. Online and offline mixed intelligent teaching assistant mode of english based on mobile information system. Mob. Inf. Syst. 2021, 2021, 7074629. [Google Scholar] [CrossRef]
  12. Sajja, R.; Sermet, Y.; Cikmaz, M.; Cwiertny, D.; Demir, I. Artificial intelligence-enabled intelligent assistant for personalized and adaptive learning in higher education. Information 2024, 15, 596. [Google Scholar] [CrossRef]
  13. Liu, Y.; Chen, L.; Yao, Z. The application of artificial intelligence assistant to deep learning in teachers’ teaching and students’ learning processes. Front. Psychol. 2022, 13, 929175. [Google Scholar] [CrossRef]
  14. Yang, Y. Design and Implementation of Algorithm-Based Intelligent Auxiliary System for Physical Education Teaching Equipment. Int. J. High Speed Electron. Syst. 2025, 8, 2540571. [Google Scholar] [CrossRef]
  15. Wang, J. Generative Pre-Trained Transformer-4 based on English Assistant Teaching System in Higher Vocational Colleges. In Proceedings of the 2024 First International Conference on Software, Systems and Information Technology (SSITCON), Tumkur, India, 18–19 October 2024; IEEE: New York, NY, USA, 2024; pp. 1–5. [Google Scholar]
  16. Wang, J.; Lu, T.; Li, L.; Huang, D. Enhancing personalized search with AI: A hybrid approach integrating deep learning and cloud computing. J. Adv. Comput. Syst. 2024, 4, 1–13. [Google Scholar] [CrossRef]
  17. Ji, X.; Sun, L.; Huang, K. The construction and implementation direction of personalized learning model based on multimodal data fusion in the context of intelligent education. Cogn. Syst. Res. 2025, 92, 101379. [Google Scholar] [CrossRef]
  18. Zhang, Y.; He, S.; Wa, S.; Zong, Z.; Lin, J.; Fan, D.; Fu, J.; Lv, C. Symmetry GAN detection network: An automatic one-stage high-accuracy detection network for various types of lesions on CT images. Symmetry 2022, 14, 234. [Google Scholar] [CrossRef]
  19. Rashid, H.A.; Ovi, P.R.; Busart, C.; Gangopadhyay, A.; Mohsenin, T. Tinym2net: A flexible system algorithm co-designed multimodal learning framework for tiny devices. arXiv 2022, arXiv:2202.04303. [Google Scholar]
  20. Jiang, D.; Shen, Z.; Zheng, Q.; Zhang, T.; Xiang, W.; Jin, J. Farm-LightSeek: An Edge-centric Multimodal Agricultural IoT Data Analytics Framework with Lightweight LLMs. IEEE Internet Things Mag. 2025, 26, 9982–10001. [Google Scholar] [CrossRef]
  21. Xiong, Y.; Guo, X.; Xu, J. CNN-Transformer: A deep learning method for automatically identifying learning engagement. Educ. Inf. Technol. 2024, 29, 9989–10008. [Google Scholar] [CrossRef]
  22. Cruz, Y.J.; Villalonga, A.; Castaño, F.; Rivas, M.; Haber, R.E. Automated machine learning methodology for optimizing production processes in small and medium-sized enterprises. Oper. Res. Perspect. 2024, 12, 100308. [Google Scholar] [CrossRef]
  23. Pei, Y.; Lu, G. Design of an intelligent educational evaluation system using deep learning. IEEE Access 2023, 11, 29790–29799. [Google Scholar] [CrossRef]
  24. Bouafoud, C.; Zine-Dine, K.; Madani, A. The Evolution of Transformers in Education: A Literature Review. In Proceedings of the 2024 International Conference on Circuit, Systems and Communication (ICCSC), Fez, Morocco, 28–29 June 2024; IEEE: New York, NY, USA, 2024; pp. 1–7. [Google Scholar]
  25. Sun, P.; Gu, L. Fuzzy knowledge graph system for artificial intelligence-based smart education. J. Intell. Fuzzy Syst. 2021, 40, 2929–2940. [Google Scholar] [CrossRef]
  26. Dan, Y.; Lei, Z.; Gu, Y.; Li, Y.; Yin, J.; Lin, J.; Ye, L.; Tie, Z.; Zhou, Y.; Wang, Y.; et al. Educhat: A large-scale language model-based chatbot system for intelligent education. arXiv 2023, arXiv:2308.02773. [Google Scholar]
  27. Xu, Q.; Wu, Y.; Zheng, H.; Yan, H.; Wu, H.; Qian, Y.; Wu, Y.; Liu, B. Standardization in artificial general intelligence model for education. Comput. Stand. Interfaces 2025, 94, 104006. [Google Scholar] [CrossRef]
  28. Gan, W.; Dao, M.S.; Zettsu, K.; Sun, Y. IoT-based multimodal analysis for smart education: Current status, challenges and opportunities. In Proceedings of the 3rd ACM Workshop on Intelligent Cross-Data Analysis and Retrieval, Newark, NJ, USA, 27–30 June 2022; pp. 32–40. [Google Scholar]
  29. Li, C.; Liu, C.; Ju, W.; Zhong, Y.; Li, Y. Prediction of teaching quality in the context of smart education: Application of multimodal data fusion and complex network topology structure. Discov. Artif. Intell. 2025, 5, 19. [Google Scholar] [CrossRef]
  30. Chango, W.; Lara, J.A.; Cerezo, R.; Romero, C. A review on data fusion in multimodal learning analytics and educational data mining. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2022, 12, e1458. [Google Scholar] [CrossRef]
  31. Luo, Z.; Zheng, C.; Gong, J.; Chen, S.; Luo, Y.; Yi, Y. 3DLIM: Intelligent analysis of students’ learning interest by using multimodal fusion technology. Educ. Inf. Technol. 2023, 28, 7975–7995. [Google Scholar] [CrossRef]
  32. Shi, W.; Nie, Z.; Shi, Y. Application and Challenges of Multi-Modal Data Fusion in Adaptive Learning System. In Proceedings of the 2024 6th International Conference on Computer Science and Technologies in Education (CSTE), Xi’an, China, 19–21 April 2024; IEEE: New York, NY, USA, 2024; pp. 82–87. [Google Scholar]
  33. Qianyi, Z.; Zhiqiang, L. Research on multimodal based learning evaluation method in smart classroom. Learn. Motiv. 2023, 84, 101943. [Google Scholar] [CrossRef]
  34. Zhao, F.; Zhang, C.; Geng, B. Deep multimodal data fusion. ACM Comput. Surv. 2024, 56, 216. [Google Scholar] [CrossRef]
  35. Dai, Z.; Zhang, Q.; Zhao, L.; Zhu, X.; Zhou, D. Cloud-edge computing technology-based internet of things system for smart classroom environment. Int. J. Emerg. Technol. Learn. (iJET) 2023, 18, 79–96. [Google Scholar] [CrossRef]
  36. Kumari, K.A.; Sadasivam, G.S.; Dharani, D.; Niranjanamurthy, M. Edge Computing: Fundamentals, Advances and Applications; CRC Press: Boca Raton, FL, USA, 2021. [Google Scholar]
  37. Shuvo, M.M.H.; Islam, S.K.; Cheng, J.; Morshed, B.I. Efficient acceleration of deep learning inference on resource-constrained edge devices: A review. Proc. IEEE 2022, 111, 42–91. [Google Scholar] [CrossRef]
  38. Bauravindah, A.; Fudholi, D.H. Lightweight models for real-time steganalysis: A Comparison of MobileNet, ShuffleNet, and EfficientNet. J. RESTI (Rekayasa Sist. Dan Teknol. Inf.) 2024, 8, 737–747. [Google Scholar] [CrossRef]
  39. Ahmadi, M.; Mirmahboub, B.; Karimi, N.; Khadivi, P.; Samavi, S. AI-Assisted Breast Cancer Classification: A Deep Learning Model Integrating MobileNet and ShuffleNet Features. In Proceedings of the 2025 IEEE World AI IoT Congress (AIIoT), Seattle, WA, USA, 28–30 May 2025; IEEE: New York, NY, USA, 2025; pp. 280–284. [Google Scholar]
  40. Khediri, N.; Ben Ammar, M.; Kherallah, M. A real-time multimodal intelligent tutoring emotion recognition system (MITERS). Multimed. Tools Appl. 2024, 83, 57759–57783. [Google Scholar] [CrossRef]
  41. Xia, C. Research on the Application of Large Language Models in Intelligent Tutoring System. In Proceedings of the 2025 3rd International Conference on Image, Algorithms, and Artificial Intelligence (ICIAAI 2025), Singapore, 23–25 May 2025; Atlantis Press: Dordrecht, The Netherlands, 2025; pp. 993–1003. [Google Scholar]
  42. You, C.; Lu, H.; Li, P.; Zhao, X.; Yao, Y. AI-Driven Intelligent Learning Companions: A Multimodal Fusion Framework for Personalized Education. In Proceedings of the 2025 IEEE 34th Wireless and Optical Communications Conference (WOCC), Macau, China, 20–22 May 2025; IEEE: New York, NY, USA, 2025; pp. 424–428. [Google Scholar]
  43. El Hajji, M.; Ait Baha, T.; Berka, A.; Ait Nacer, H.; El Aouifi, H.; Es-Saady, Y. An Architecture for Intelligent Tutoring in Virtual Reality: Integrating LLMs and Multimodal Interaction for Immersive Learning. Information 2025, 16, 556. [Google Scholar] [CrossRef]
  44. Jia, J.; He, Y.; Le, H. A multimodal human-computer interaction system and its application in smart learning environments. In Proceedings of the International Conference on Blended Learning, Bangkok, Thailand, 21–24 July 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 3–14. [Google Scholar]
  45. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  46. Howard, A.; Sandler, M.; Chu, G.; Chen, L.C.; Chen, B.; Tan, M.; Wang, W.; Zhu, Y.; Pang, R.; Vasudevan, V.; et al. Searching for mobilenetv3. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1314–1324. [Google Scholar]
  47. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  48. Ab Wahab, M.N.; Nazir, A.; Ren, A.T.Z.; Noor, M.H.M.; Akbar, M.F.; Mohamed, A.S.A. Efficientnet-lite and hybrid CNN-KNN implementation for facial expression recognition on raspberry Pi. IEEE Access 2021, 9, 134065–134080. [Google Scholar] [CrossRef]
  49. Barbieri, L.; Brambilla, M.; Stefanutti, M.; Romano, C.; De Carlo, N.; Roveri, M. A tiny transformer-based anomaly detection framework for IoT solutions. IEEE Open J. Signal Process. 2023, 4, 462–478. [Google Scholar] [CrossRef]
  50. Çökmez, G.M.; Zhang, Y.; Schroers, C.; Aydin, T.O. CLIP-Fusion: A Spatio-Temporal Quality Metric for Frame Interpolation. In Proceedings of the 2025 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Tucson, Arizona, 28 February–4 March 2025; IEEE: New York, NY, USA, 2025; pp. 7450–7459. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.