Next Article in Journal
Biomimetic Innovations: Exploring Bubble-Trapping Organisms for Manufacturing Breakthroughs
Previous Article in Journal
Digital Twin Cognition: AI-Biomarker Integration in Biomimetic Neuropsychology
Previous Article in Special Issue
Brain–Computer Interfaces in Parkinson’s Disease Rehabilitation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MASleepNet: A Sleep Staging Model Integrating Multi-Scale Convolution and Attention Mechanisms

1
Xi’an Key Laboratory of High Precision Industrial Intelligent Vision Measurement Technology, School of Electronic Information, Xijing University, Xi’an 710123, China
2
School of Mathematics and Statistics, Zhengzhou University, Zhengzhou 450001, China
3
School of Electronic Information, Xijing University, Xi’an 710123, China
*
Authors to whom correspondence should be addressed.
Biomimetics 2025, 10(10), 642; https://doi.org/10.3390/biomimetics10100642
Submission received: 24 August 2025 / Revised: 16 September 2025 / Accepted: 19 September 2025 / Published: 23 September 2025

Abstract

With the rapid development of modern industry, people’s living pressures are gradually increasing, and an increasing number of individuals are affected by sleep disorders such as insomnia, hypersomnia, and sleep apnea syndrome. Many cardiovascular and psychiatric diseases are also closely related to sleep. Therefore, the early detection, accurate diagnosis, and treatment of sleep disorders an urgent research priority. Traditional manual sleep staging methods have many problems, such as being time-consuming and cumbersome, relying on expert experience, or being subjective. To address these issues, researchers have proposed multiple algorithmic strategies for sleep staging automation based on deep learning in recent years. This paper studies MASleepNet, a sleep staging neural network model that integrates multimodal deep features. This model takes multi-channel Polysomnography (PSG) signals (including EEG (Fpz-Cz, Pz-Oz), EOG, and EMG) as input and employs a multi-scale convolutional module to extract features at different time scales in parallel. It then adaptively weights and fuses the features from each modality using a channel-wise attention mechanism. The integrated temporal features are integrated into a Bidirectional Long Short-Term Memory (BiLSTM) sequence encoder, where an attention mechanism is introduced to identify key temporal segments. The final classification result is produced by the fully connected layer. The proposed model was experimentally evaluated on the Sleep-EDF dataset (consisting of two subsets, Sleep-EDF-78 and Sleep-EDF-20), achieving classification accuracies of 82.56% and 84.53% on the two subsets, respectively. These results demonstrate that deep models that integrate multimodal signals and an attention mechanism offer the possibility to enhance the efficiency of automatic sleep staging compared to cutting-edge methods.

Graphical Abstract

1. Introduction

Sleep is a vital physiological process, occupying over one-third of the human lifespan. Good sleep plays a vital role in maintaining both physical and mental health [1]. However, in modern society, factors such as life stress, excessive use of electronic devices, and circadian rhythm disorders have led to the increasing prevalence of insomnia, sleep apnea, and other problems, seriously endangering people’s health and quality of life [2,3]. With the increasing attention paid to sleep health issues, the study of sleep staging has gradually become a major research focus in sleep medicine [4,5,6]. Traditional sleep staging studies require professional physicians to manually segment the polysomnography (PSG) signal and divide it into five stages (W, REM, N1, N2, N3) as defined by internationally accepted sleep staging standards [7]. This method is highly time-consuming, requires substantial human effort, and often suffers from inconsistencies across different scorers [8,9,10]. Therefore, developing an automated sleep staging algorithm holds substantial significance.
Significant advancements in classification have been driven by the integration of machine learning and deep learning methods in recent years [11,12,13,14,15,16,17]. However, most early studies mainly used single-channel electroencephalogram (EEG) signals [18,19] and relied on expert prior experience for feature extraction. A single EEG channel is insufficient to comprehensively capture the diverse physiological changes during sleep. For example, changes in eye movement and muscle tension are also crucial for identifying stages such as REM. Therefore, methods that integrate multimodal signals are expected to increase the efficiency of automatic sleep staging. Deep neural network models have proven effective in sleep staging tasks [20,21,22]. For example, DeepSleepNet, proposed by Supratak et al. [23], integrates CNNs with Bidirectional Long Short-Term Memory (BiLSTM) networks to automatically extract features and model stage sequences from single-channel EEG raw data. The model extracts features of different frequency bands such as slow waves and fast waves through multi-scale convolution, and improves staging performance by combining sequence information.
A series of subsequent works further improved on this basis. For example, Eldele et al. [24] proposed the AttnSleep model, which introduced multi-resolution convolutional feature extraction mechanism to capture features of different frequencies and adjust feature channel weights, thereby improving the quality of feature representation. At the same time, by modeling long-range dependencies through the attention mechanism, the accuracy of single-channel staging and the ability to detect each stage were effectively improved. The 1D-ResNet-SE-LSTM model proposed by Li et al. [25] combines the residual network with the SE attention module, works with LSTM to capture stage transition patterns, and introduces a weighted loss function to reduce the impact of category imbalance. The above studies show that multi-scale convolution, channel attention, and temporal context modeling are crucial to improving sleep staging performance.
In terms of multi-channel and multi-modal signals, some automatic staging models that integrate EEG, EOG, and EMG have also emerged in recent years. For example, the SleepPrintNet proposed by Jia et al. [26] uses a multivariate and multimodal neural network to capture salient features from multiple physiological modalities, significantly improving the staging accuracy. Subsequently, they proposed the SalientSleepNet, which further achieved a parameter-efficient and performance-leading staging model through multimodal saliency wave detection and multi-scale time series feature extraction [27]. The above work proves the effectiveness of multimodal data fusion and attention mechanism in sleep staging.
Deep learning has made remarkable advances in automatic sleep staging, but the following major challenges still exist: (1) Single-channel EEG-based models have difficulty fully utilizing complementary information from modalities; (2) An uneven distribution of sleep data among different stages is observed, resulting in insufficient recognition performance of the model for minority classes, especially the N1 stage [28,29]; (3) The cross-scale fusion of time-frequency features and the modeling of long-range temporal dependencies are still insufficient, which makes it difficult for the model to learn complex sleep dynamics.
To overcome these limitations, this study introduces a MASleepNet sleep staging model. This model integrates multi-scale convolution, channel attention, temporal Bidirectional Long Short-Term Memory (BiLSTM), an attention mechanism, and a class weighting strategy into a unified architecture to automatically classify sleep stages from multi-channel polysomnography (PSG) signals. Compared to existing methods, this model strives to improve overall accuracy while strengthening the distinguish of the minority class N1.
The main contributions of the study are: (1) Multimodal and multiscale feature extraction: The dual-branch convolutional network is designed to process different time domain scales in parallel, extracting cross-band features for EEG/EOG/EMG signals while taking into account transient spikes and slow wave patterns. One branch uses a larger convolution kernel to capture slow wave information spanning several seconds, while the other branch uses a smaller one to capture sub-second transient features. The outputs of the two branches are spliced to obtain a rich time-frequency feature representation. (2) Channel attention fusion: The Squeeze-and-Excitation (SE) module is applied to each modal convolution feature to achieve channel-level adaptive weighting. The SE module learns channel importance weights by applying global pooling and subsequent fully connected network to highlight key channel features. This mechanism can effectively enhance discriminative channel information such as REM stage eye movement features and N3 stage slow wave features. (3) BiLSTM+attention temporal sequence modeling: The fused multimodal feature sequence is input into a BiLSTM to capture temporal correlations, and then the LSTM output is weighted and summed through the attention mechanism to screen the time segments that contribute most to the current epoch judgment. The BiLSTM’s forward and backward hidden states can simultaneously utilize contextual information, helping to learn the evolution of sleep stages. The attention layer assigns weights to features at different moments in accordance with the trained query vector, allowing the model to automatically ignore irrelevant information and highlight key features.
The organization of this paper is as follows: Section 2 introduces the MASleepNet model architecture and its key components. Section 3 presents the dataset, experimental setup, evaluation metrics, and results. Section 4 discusses ablation studies. Section 5 provides the conclusions, and Section 6 outlines future work.

2. Methods

This section describes the MASleepNet model for classification using multimodal data, including its overall framework and key components. The model uses multi-channel polysomnography (PSG) signals as input.

2.1. Overall Model Architecture

The architecture of the proposed MASleepNet sleep staging model is illustrated in Figure 1. It has three functional modules: (1) Multi-modal Scalable Feature Extraction Module (MFEM); (2) Bi-directional Time Series Modeling and Focusing Module (BiLSTM-Attention Context Module, ACM); and (3) Sleep Stage Classifier (Classifier).
MASleepNet receives polysomnography (PSG) signals comprising four channels in total: two EEG channels (Fpz–Cz, Pz–Oz), one EOG channel, and one EMG channel. It first extracts multimodal time-frequency features using a multi-scale convolutional network. A channel-wise attention mechanism (SELayer) is then introduced after each modal channel for weighted fusion. The integrated multimodal features are subsequently input into a bidirectional LSTM + Attention module (BiLSTM-Attention) to extract contextual information and focus on temporal features at key moments. Finally, the classification results are output through the fully connected layer.
The entire model architecture is end-to-end trainable. The components of the model are designed to work synergistically: the MFEM extracts modal features across time scales, the SE channel attention highlights key channels, and the temporal evolution and focuses on discriminative key segments, thereby improving overall classification performance.

2.2. Multimodal Multiscale Feature Extraction Module (MFEM)

The MFEM is designed to extract modality-specific features from EEG, EOG, and EMG signals. As shown in Figure 2. This module constructs independent convolutional feature extraction branches for the feature, EOG, and EMG signals. Each modality branch uses a parallel dual-branch one-dimensional convolutional network to capture key information at multiple time scales, as shown in Figure 2. Multi-branch structures have been widely shown in previous studies to enhance model robustness and generalization [30]. Although multi-branch structures have certain advantages in feature extraction, introducing too many branches will significantly increase computational costs and may even lead to side effects such as overfitting. Existing studies have shown that additional branches have limited impact on performance improvement [31]. Based on this consideration, this paper adopts two convolutional branches in the model design to balance classification accuracy and computational efficiency.
Specifically, the branch of each modality contains the following two convolution branches:
(1)
Short-time-scale convolution branch (local feature branch): This path uses a smaller convolution kernel (kernel size = 50) to accurately capture short-time local information in the signal, such as the wake phase (W) and high-frequency components like spindles;
(2)
Long-time convolution branch (global feature branch): This path uses a larger convolution kernel (kernel size = 400) and the receptive field covers a signal length of about 4 s to fully extract the slower trend features in the signal, such as the slow wave features in the N3 stage.
The basic structure of each convolution branch includes convolution, batch normalization (BN), LeakyReLU activation, max pooling, and dropout regularization. The specific convolution calculation process can be defined as Formula (1):
X j l + 1 = f i M j   X i l × W i j + b j
Among them, X i l represents the i channel of the l layer input feature, W i j denotes the weight and b j corresponds the bias of the convolution kernel, and f(⋅) is the LeakyReLU activation function.
After the two convolution branches extract features respectively, they are concatenated in the time dimension to form a fused multi-scale feature.
Furthermore, considering that EEG, EOG, and EMG signals have different levels of importance in identifying different stages (for example, EOG eye movement features are more critical in the REM stage), this study introduced the a SE channel attention module following the convolutional fusion features of each modality to achieve adaptive weighting of channel features and reduce the inevitable redundancy generated when extracting large-scale features and small-scale features [32]. The SE module first captures the statistical characteristics of each channel via global average pooling. The calculation process is shown in Formula (2):
z = 1 T t = 1 T   X [ : , t ]
Then, the attention weight vector s of each channel is generated using two successive fully connected layers. The specific calculation formula is (3):
s = σ W 2 δ W 1 z
Finally, the weight vector is multiplied element-by-element with the original channel feature to achieve adaptive channel weighting of the feature, so as to highlight the channel features with stronger discriminative ability and suppress the interference of redundant or noisy channels. The final channel-weighted feature is expressed as Formula (4):
X ^ c = X c s c , c = 1,2 , , C
Among them, W 1 and W 2 are the learnable weight parameter matrices, σ denotes the sigmoid activation function, and δ corresponds to the ReLU activation function.
Through the above feature extraction process, the weighted features of EEG (2 channels), EOG (1 channel), and EMG (1 channel) are concatenated along the channel dimension to construct a unified multimodal fused feature tensor. This fused feature tensor is further rearranged into a format suitable for sequence modeling (i.e., (batch size, time steps, features)) to be passed to the subsequent bidirectional LSTM and attention focusing modules for sequential modeling and classification of sleep stages.

2.3. BiLSTM-Attention Context Module (ACM)

Sleep stage sequences exhibit strong contextual relevance in the temporal dimension, with strong dependencies across different stages. To effectively model this long-range temporal dependency, this study designed a BiLSTM-Attention Context Module (ACM). This module combines a BiLSTM network with an attention mechanism to enhance context modeling capabilities in sequential data. Such architectures have been extensively applied in diverse domains.
The first employs a Bi-LSTM network to simultaneously capture contextual information of the feature sequence in both forward and backward directions. As illustrated in Figure 3, the input feature sequence X = (x_1,x_2,…,x_T), obtained after multi-modal feature fusion, is fed into two LSTM layers: one processes the sequence in the forward direction while the other in the backward direction. At each time step t, the hidden states from both directions are concatenated to form the bidirectional representation, as expressed in Formula (5):
h t = h t ; h t , t = 1,2 , , T
Subsequently, a sequence-wise attention module is employed to automatically focus on the most discriminative key segments in the temporal feature sequence. The specific steps involved in the calculation of the attention mechanism are outlined below.
First, the attention weight α t of the hidden state h t at each time step as shown in Formula (6):
α t = exp v t a n h W h h t + b h t = 1 T   e x p v t a n h W h h t + b h
Among them, the attention mechanism employs W h as the weight matrix and b h as the associated bias term, v denotes the attention vector parameter, tanh(⋅) is the hyperbolic tangent activation function, and exp(⋅) represents the exponential function.
Subsequently, according to the above weights, the hidden state sequence output by BiLSTM is weighted and summed to form the final attention context vector c, as shown in Formula (7):
c = t = 1 T   α t h t
Through these calculations, the automatically prioritizes the time segments that contribute most to classification, while suppressing less contributive or redundant features. This module demonstrates a significant performance advantage in the identification of easily confused sleep stages, including N1 and REM.

2.4. Classifier and Loss Function

The classifier of the model maps the attention context feature vector c output by the to the final classification space.
To avoid model overfitting, a Dropout regularization strategy is used on the context vector c, with a Dropout ratio of p = 0.5 set to obtain the regularized feature vector representation c . The regularized context feature vector is input to the fully connected classification layer, and the predicted class probabilities of the model distribution for each sleep stage is finally obtained y . The output calculation process of the classifier can be expressed as Formula (8):
y = s o f t m a x W f c c + b f c
Among them, the weight parameter matrix is represented by W f c , the bias vector of the fully connected layer is indicated by b f c , and the softmax(⋅) function is used to convert the linear output into a probability form to denote the model’s predicted probabilities for each sleep stage class.
The cross-entropy loss quantifies the divergence between the predicted distribution y and the actual labels y ^ . The definition of the loss function is provided in Formula (9):
L ( y ^ , y ) = k = 1 K   y ^ k l o g y k
where K = 5 represents the complete set of sleep stage categories, y ^ k is the one-hot encoding vector of the true label, and y k is the probability value predicted by the classifier.

3. Experiments and Results

3.1. Dataset and Preprocessing

This study utilized the Sleep-EDF dataset provided by PhysioNet [33,34], which comprises two subsets, Sleep-EDF-20 and Sleep-EDF-78 (Table 1). Sleep-EDF-20 includes two-night recordings from 20 subjects, Sleep-EDF-78 is an extended dataset containing single-night recordings from 78 subjects. Each polysomnography (PSG) recording comprises two EEG channels (Fpz-Cz, Pz-Oz), one EOG channel, and one submental (chin) EMG channel, all sampled at 100 Hz. The original dataset was annotated by experts following the AASM guidelines into five stages: W, N1, N2, N3, and REM [35]. During preprocessing, each recording was divided into epochs and annotated accordingly. For Sleep-EDF-20 and Sleep-EDF-78, this study the same preprocessing pipeline was applied and without extracting additional intervals.

3.2. Experimental Plan

To evaluate the generalization ability of MASleepNet, this study employed 5-fold cross-subject validation. The data from 20 subjects were randomly partitioned into five subsets, with four subsets used for training and validation in each iteration, and the remaining subset reserved as test set. To mitigate overfitting and enable early stopping, 10% of the training samples were randomly held out as a validation set. During training, the validation loss was continuously monitored. Training was terminated if the validation loss did not decrease for five successive epochs, and the model parameters corresponding to the best result were saved.
Model training was conducted in an end-to-end manner. The Adam optimizer was selected due to its widespread application in sleep staging and its stable convergence properties. The initial learning rate was set to 0.001, which is commonly adopted in related studies and was further confirmed through preliminary tuning experiments to provide a balance between convergence stability and computational efficiency. The cross-entropy loss function was employed since it is well-suited for multi-class classification tasks. A batch size of 64 was chosen to ensure both computational efficiency and gradient stability, while the maximum number of epochs was limited to 100, which was sufficient for convergence under the given settings. These hyperparameter configurations were determined based on prior research and empirical validation, ensuring consistent and reliable training performance.

3.3. Performance Evaluation Index

The evaluation metrics of the model are: overall accuracy (ACC), F1 score (F1Macro), and Kappa coefficient. Sensitivity (SEN), specificity (SPE), and precision (PRE) were also calculated.
The accuracy score is calculated using Formula (10):
A C C = T P + T N T P + F P + T N + F N
The F1 is calculated using Formula (11):
F 1 M a c r o = 1 K k = 1 K   2 · P R E i · S E N i P R E i · S E N i
The K a p p a coefficient is calculated using Formula (12):
K a p p a = P 0 P e 1 P e
Here, P 0 represents observed consistency and P e represents random expected consistency.
Sensitivity and specificity are calculated according to Formula (13):
S E N = T P T P + F N ,   S P E = T N T N + F P
The precision is calculated using Formula (14):
P R E = T P T P + F P
In the formula above, TP denotes true positives, TN denotes true negatives, FP denotes false positives, and FN denotes false negatives.

3.4. Classification Performance of the MASleepNet Model

To intuitively demonstrate the model’s prediction performance at the individual subject level, we visualized the complete sleep process of subject SC406 from the SleepEDF-78 dataset, as shown in Figure 4. As illustrated in Figure 4, panel (a) depicts the subject’s truth hypnogram, and panel (b) illustrates the hypnogram predicted by the proposed model.
Figure 4 clearly shows that the model’s output hypnograms closely match the ground-truth annotations in terms of overall trends and cyclical variations, demonstrating particularly good fit for the identification of N3, REM, and W stage. While some localized errors exist in individual stages, such as N1, the overall trend remains highly consistent, demonstrating the model’s robust performance in staging complex individual sleep structures.
Figure 5 and Figure 6 present the overall confusion matrices of MASleepNet on the Sleep-EDF-78 and Sleep-EDF-20 datasets under five-fold cross-validation. The results show that stages W, N2, N3, and REM achieve consistently high classification accuracy in both datasets, confirming the model’s strong ability to recognize the major sleep stages. In contrast, the N1 stage remains the most challenging, as reflected by the higher proportion of off-diagonal elements in its row. This indicates frequent misclassification of N1 as the adjacent W, N2, or REM stages, which can be attributed to its short duration and smooth signal transitions that overlap with neighboring stages.
Despite differences in sample size, signal composition, and subject diversity between the two datasets, MASleepNet maintains stable performance across both, demonstrating strong cross-dataset generalization. The overall distribution patterns of the confusion matrices remain consistent, while the slightly higher N1 confusion observed on Sleep-EDF-78 may be due to greater inter-subject variability. These results confirm that MASleepNet achieves competitive accuracy while preserving robustness and adaptability across heterogeneous datasets.

3.5. Baseline Networks and Comparison

As a reference, the study incorporated a classical and widely recognized sleep staging model for comparative evaluation such as AttnSleep [24], DeepSleepNet [23], ResNetLSTM [36], SeqSleepNet [15] and TinySleepNet [37] as comparison methods, namely: (1) AttnSleep adopts multi-resolution convolution and sequence attention mechanism, and is good at capturing long-range features; (2) DeepSleepNet: combines CNN and BiLSTM, and introduces end-to-end deep structure for the first time; (3) ResNetLSTM: uses ResNet structure to extract local features and combines LSTM to capture temporal information; (4) SeqSleepNet: a two-layer BiLSTM structure, and introduces attention mechanism at the frame level and stage level; (5) TinySleepNet: a lightweight CNN network that takes into account both computational efficiency and classification accuracy.
All comparison models were tested under the same data partitioning, experimental configuration, training parameters, and adopted the same experimental and evaluation settings as this study to ensure fairness and comparability of the results. Table 2 and Table 3 summarize the overall classification performance of MASleepNet and other baseline models on the Sleep-EDF-20 and Sleep-EDF-78 datasets, respectively, while Table 2 and Table 3 present the per-class precision results for each sleep stage.
An examination of Table 2 and Table 3 reveals that the proposed MASleepNet consistently achieves excellent performance, including accuracy, F1Macro, Kappa, sensitivity, specificity, and precision. In particular, the MASleepNet model significantly outperforms other methods in challenging sleep stages such as N1 and REM, demonstrating its effectiveness in capturing subtle and discriminative multimodal temporal features.
Specifically, MASleepNet achieved an average accuracy of 84% and an F1Macro of 0.78 on the Sleep-EDF-20 test set; and an average accuracy of 82% and an F1Macro of 0.76 on the Sleep-EDF-78 test set. Compared to methods such as AttnSleep, which only use single-channel EEG, this model improves both accuracy and F1Macro (AttnSleep achieves approximately 81% accuracy on Sleep-EDF-20, but only approximately 0.37 F1Macro for the minority class N1). In particular, MASleepNet maintains high accuracy for the major phases (N1, N3, and REM), with N3 performing best with an accuracy exceeding 85%. Misclassifications primarily occur between N3 and N2, and between REM and N1 (these phases share similar transition signals). The sensitivity of MASleepNet in recognizing the most challenging N1 stage is significantly greater than that of competing models. On the Sleep-EDF-20 dataset, MASleepNet’s accuracy for the N1 stage increased to 53.55%, significantly outperforming other models. This demonstrates that the incorporation of EOG and EMG information, along with the attention mechanism, effectively alleviates the difficulty in identifying the N1 stage. This model achieved classification accuracy exceeding 80% for all stages except N1 (nearly 90% for wake, N2, and N3). While approximately 40% of N1 samples were misclassified as N2 or REM, the overall false positive rate was lower than that of the baseline model. Comprehensive metrics and comparisons confirm that MASleepNet achieves superior classification capability and balanced performance in the automatic sleep staging task. This experiment validates MASleepNet as an effective and robust method for automatic classification, outperforming existing current leading approaches and providing data support for clinical applications.

4. Ablation Experiment

To gain deeper insights into the contribution of individual components, this study designed the following ablation experiments:
(1)
No-MSConv (no multi-scale convolution structure): This variant removes the large kernel convolution branch in the multi-scale feature extraction module and only retains the small kernel (local) convolution path in the original network to evaluate the impact of different receptive field combinations on feature extraction capabilities.
(2)
EEG-Only: This variant starts from the input layer and only retains EEG as input, completely removing the EOG and EMG signal paths, and correspondingly deleting the feature extraction and attention fusion modules of these two modalities. It is used to evaluate the gain effect of multimodal signal fusion on classification performance.
(3)
No-SE (remove squeeze-and-excitation module): This variant removes the SE channel attention module in each branch, such that feature maps are directly propagated without channel-wise recalibration. This design evaluates the effect of adaptive channel weighting on performance.
(4)
No-LSTM (remove BiLSTM): This variant removes the BiLSTM temporal modeling module and instead directly applies attention-based global pooling over the extracted features. It is used to evaluate the contribution of recurrent temporal modeling to sequence classification.
(5)
No-Attn (remove attention mechanism): This variant retains the BiLSTM module for time series modeling, but removes the attention mechanism; the classifier only receives the last time step in the LSTM output sequence as the global representation vector for classification, which is used to compare the attention context modeling ability.
Table 4 and Table 5 present the ablation results of the MASleepNet model and its variants on the Sleep-EDF-20 and Sleep-EDF-78 datasets, with detailed comparisons across Accuracy, F1Macro, Kappa, SEN, SPE, and PRE.
Comparing the results of EEG-Only and MASleepNet, we found that the introduction of multimodal fusion significantly improved classification performance, increasing accuracy by 1.80% and 2.62% on the Sleep-EDF-20 and Sleep-EDF-78 datasets, respectively. This strongly supports that EOG and EMG provide effective information complementary to EEG, playing a particularly positive role in the identification of easily confused phases such as N1 and REM.
Comparing the EEG-Only and MASleepNet models, multimodal fusion improved accuracy by 1.80% and 2.62% on Sleep-EDF-20 and Sleep-EDF-78, respectively, confirming the complementary role of EOG and EMG in supporting EEG, particularly in the recognition of challenging stages such as N1 and REM.
Removing the SE module (No-SE) led to a decline in overall performance (ACC decreased by 0.31% and 0.51%), demonstrating that adaptive channel weighting improves the discriminability of multimodal features by highlighting informative channels.
Eliminating the BiLSTM (No-LSTM) resulted in accuracy reductions of 0.35% and 1.04% on the two datasets, highlighting the importance of recurrent temporal modeling for capturing sequential dependencies in sleep signals.
Finally, the No-Attn variant also showed decreased accuracy (−0.40% and −0.71%), indicating that the attention mechanism further enhances feature utilization by focusing on discriminative time segments, especially improving recognition of the minority N1 stage.
Through comparative analysis, the effectiveness and necessity of the five key modules—multi-scale convolution, multimodal fusion, SE attention, BiLSTM, and attention mechanism—are validated, and their specific contributions to the overall performance of MASleepNet are systematically demonstrated.

5. Conclusions

This paper addresses the shortcomings of existing sleep staging models in modality fusion, multi-scale feature extraction, and sequential information capture by proposing a novel multimodal deep learning framework, MASleepNet. The model integrates a multi-scale convolutional neural network, a channel-level SE attention mechanism, and a BiLSTM network with a sequence-level attention mechanism. By leveraging the complementary characteristics of EEG, EOG, and EMG signals, MASleepNet effectively captures both local and global features across different time scales, as well as the long-term sequential dependencies between sleep stages.
Experimental results demonstrate that MASleepNet delivers superior performance on both the Sleep-EDF-20 and Sleep-EDF-78 datasets, significantly outperforming widely recognized baseline models such as AttnSleep, DeepSleepNet, ResNetLSTM, SeqSleepNet, and TinySleepNet. The model achieves notable improvements in overall accuracy, F1 score, and Cohen’s kappa, and shows remarkable effectiveness in recognizing the challenging N1 and REM stages. Ablation studies further confirm the contributions of each module: the multi-scale convolutional structure captures frequency-specific sleep characteristics, the SE attention module enhances discriminative feature channels, and the sequence-level attention mechanism identifies critical temporal dependencies. Collectively, these findings substantiate the effectiveness of the MASleepNet framework and highlight its potential for advancing sleep stage classification.

6. Future Work

Future research will further emphasize improving the computational efficiency and lightweight design of the model, with the aim of enabling real-time deployment in portable and embedded sleep monitoring devices. Additionally, validating the model’s generalizability on diverse and large-scale clinical datasets will be a key focus, ensuring its applicability to real-world medical environments. Extending the framework to integrate additional physiological signals and exploring advanced fusion strategies may also enhance its robustness and clinical utility.

Author Contributions

Conceptualization, data curation, investigation, methodology, validation, visualization, writing—original draft, Z.W.; data curation, project administration, resources, supervision, Y.M.; supervision, writing—review and editing, Y.M. and S.Z.; investigation, Q.D., Z.H., Z.G. and T.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the General Projects of Shaanxi Science and Technology Plan (No. 2023-JC-YB-504), the National Natural Science Foundation of China (No. 62172338), the Shaanxi province innovation capacity support program (No. 2018KJXX-095), and the Xijing Universityspecial talent research fund (No. XJ7T03).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data were obtained from [physiobank] and are available from the authors/at Sleep-EDF Database Expanded.

Acknowledgments

We would like to thank the staff at the Xi’an Key Laboratory of Intelligent Rehabilitation Human–Machine Fusion and Control Technology for their technical assistance and the administrative team for their support throughout the project.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Weber, F.; Dan, Y. Circuit-based interrogation of sleep control. Nature 2016, 538, 51–59. [Google Scholar] [CrossRef] [PubMed]
  2. Ma, H.; Li, H.; Chen, C.; Xu, N.; Chen, Q.; Ma, J. Sleep effects on memory consolidation in adolescent English vocabulary memorization: Investigating links between sleep quality, memory, and mood disturbance. Curr. Psychol. 2024, 43, 28429–28437. [Google Scholar] [CrossRef]
  3. He, Y.; Yang, T.; Guo, Q.; Wu, S.; Liu, W.; Xu, T. Innovative Analysis of the Interconnected Network Structure Between Anxiety and Sleep Quality Among College Students. Psychol. Res. Behav. Manag. 2025, 18, 607–618. [Google Scholar] [CrossRef] [PubMed]
  4. Jadhav, P.; Mukhopadhyay, S. Automated sleep stage scoring using time-frequency spectra convolution neural network. IEEE Trans. Instrum. Meas. 2022, 71, 2510309. [Google Scholar] [CrossRef]
  5. Wang, Z.; Zhang, J.; Xia, Y.; Chen, P.; Wang, B. A general and scalable vision framework for functional near-infrared spectroscopy classification. IEEE Trans. Neural Syst. Rehabil. Eng. 2022, 30, 1982–1991. [Google Scholar] [CrossRef]
  6. Choi, M.; Lee, S.-J. Oscillometry-based blood pressure estimation using convolutional neural networks. IEEE Access 2022, 10, 56813–56822. [Google Scholar] [CrossRef]
  7. Iber, C.; Ancoli-Israel, S.; Chesson, A.; Quan, S. The AASM Manual for the Scoring of Sleep and Associated Events: Rules, Terminology and Technical Specifications; American Academy of Sleep Medicine: Westchester, IL, USA, 2007. [Google Scholar]
  8. Özşen, S. Classification of sleep stages using class-dependent sequential feature selection and artificial neural network. Neural Comput. Appl. 2013, 23, 1239–1250. [Google Scholar] [CrossRef]
  9. Eldele, E.; Ragab, M.; Chen, Z.; Wu, M.; Kwoh, C.K.; Li, X. Selfsupervised learning for label-efficient sleep stage classification: A comprehensive evaluation. IEEE Trans. Neural Syst. Rehabil. Eng. 2023, 31, 1333–1342. [Google Scholar] [CrossRef]
  10. Malhotra, A.; Younes, M.; Kuna, S.T.; Benca, R.; Kushida, C.A.; Walsh, J.; Hanlon, A.; Staley, B.; Pack, A.I.; Pien, G.W. Performance of an automated polysomnography scoring system versus computer-assisted manual scoring. Sleep 2013, 36, 573–582. [Google Scholar] [CrossRef] [PubMed]
  11. Anandakumar, M.; Pradeepkumar, J.; Kappel, S.L.; Edussooriya, C.U.; de Silva, A.C. A knowledge distillation framework for enhancing Ear-EEG based sleep staging with scalp-EEG data. In Proceedings of the 2023 IEEE International Conference on Systems, Man, and Cybernetics (SMC), Honolulu, HI, USA, 1–4 October 2023; pp. 514–519. [Google Scholar]
  12. Fiorillo, L.; Favaro, P.; Faraci, F.D. DeepSleepNet-Lite: A simplified automatic sleep stage scoring model with uncertainty estimates. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 2076–2085. [Google Scholar] [CrossRef]
  13. Casal, R.; Di Persia, L.E.; Schlotthauer, G. Classifying sleep–wake stages through recurrent neural networks using pulse oximetry signals. Biomed. Signal Process. Control 2021, 63, 102195. [Google Scholar] [CrossRef]
  14. Phan, H.; Chén, O.Y.; Tran, M.C.; Koch, P.; Mertins, A.; De Vos, M. XSleepNet: Multi-view sequential model for automatic sleep staging. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 5903–5915. [Google Scholar] [CrossRef] [PubMed]
  15. Phan, H.; Andreotti, F.; Cooray, N.; Chen, O.Y.; De Vos, M. SeqSleepNet: End-to-end hierarchical recurrent neural network for sequence-to sequence automatic sleep staging. IEEE Trans. Neural Syst. Rehabil. Eng. 2019, 27, 400–410. [Google Scholar] [CrossRef]
  16. Chambon, S.; Galtier, M.N.; Arnal, P.J.; Wainrib, G.; Gramfort, A. A deep learning architecture for temporal sleep stage classification us-ing multivariate and multimodal time series. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 758–769. [Google Scholar] [CrossRef] [PubMed]
  17. Dong, H.; Supratak, A.; Pan, W.; Wu, C.; Matthews, P.M.; Guo, Y. Mixed neural network approach for temporal sleep stage classification. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 324–333. [Google Scholar] [CrossRef]
  18. Lee, C.-H.; Kim, H.; Han, H.-J.; Jung, M.-K.; Yoon, B.C.; Kim, D.-J. NeuroNet: A novel hybrid self-supervised learning framework for sleep stage classification using single-channel EEG. arXiv 2024, arXiv:2404.17585. [Google Scholar]
  19. Siddiqa, H.A.; Tang, Z.; Xu, Y.; Wang, L.; Irfan, M.; Abbasi, S.F.; Nawaz, A.; Chen, C.; Chen, W. Single-Channel EEG data analysis using a multi-branch CNN for neonatal sleep staging. IEEE Access 2024, 12, 29910–29925. [Google Scholar] [CrossRef]
  20. Ronzhina, M.; Janoušek, O.; Kolářová, J.; Nováková, M.; Honzík, P.; Provazník, I. Sleep scoring usingartificial neural networks. Sleep Med. Rev. 2012, 16, 251–263. [Google Scholar] [CrossRef]
  21. Chriskos, P.; Frantzidis, C.A.; Nday, C.M.; Gkivogkli, P.T.; Bamidis, P.D.; Kourtidou-Papadeli, C.P.; Chriskos, C.A.; Frantzidis, C.M. A review on currenttrends in automatic sleep staging through bio-signal recordings and future challenges. Sleep Med. Rev. 2021, 55, 101377. [Google Scholar] [CrossRef]
  22. Phan, H.; Mikkelsen, K. Automatic sleep staging of EEG signals: Recent development, challenges, and future directions. Physiol Meas 2022, 43, 04TR01. [Google Scholar] [CrossRef]
  23. Supratak, A.; Dong, H.; Wu, C.; Guo, Y. DeepSleepNet: A model for automatic sleep stage scoring based on raw single-channel EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2017, 25, 1998–2008. [Google Scholar] [CrossRef]
  24. Eldele, E.; Chen, Z.; Liu, C.; Wu, M.; Kwoh, C.-K.; Li, X.; Guan, C. An Attention-Based Deep Learning Approach for Sleep Stage Classification With Single-Channel EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 809–818. [Google Scholar] [CrossRef]
  25. Li, W.; Gao, J. Automatic sleep staging by a hybrid model based on deep 1D-ResNet-SE and LSTM with single-channel raw EEG signals. PeerJ Comput. Sci. 2023, 9, e1561. [Google Scholar] [CrossRef]
  26. Jia, Z.; Cai, X.; Zheng, G.; Wang, J.; Lin, Y. SleepPrintNet: A Multivariate Multimodal Neural Network Based on Physiological Time-Series for Automatic Sleep Staging. IEEE Trans. Artif. Intell. 2020, 1, 248–257. [Google Scholar] [CrossRef]
  27. Jia, Z.; Lin, Y.; Wang, J.; Wang, X.; Xie, P.; Zhang, Y. SalientSleepNet: Multimodal salient wave detection network for sleep stagingCCC. In Proceedings of the IJCAI 2021 (30th Int Joint Conf on Artificial Intelligence), Virtual, 19–26 August 2021; pp. 3617–3623. [Google Scholar]
  28. Fraiwan, L.; Lweesy, K.; Khasawneh, N.; Wenz, H.; Dickhaus, H. Automated sleep stage identification system based on time-frequency analysis of a single EEG channel and random forest classifer. Comput. Methods Programs Biomed. 2012, 108, 10–19. [Google Scholar] [CrossRef] [PubMed]
  29. Aboalayon, K.A.I.; Faezipour, M.; Almuhammadi, W.S.; Moslehpour, S. Sleep stage classifcation using EEG signal analysis: A comprehensive survey and new investigation. Entropy 2016, 18, 272. [Google Scholar] [CrossRef]
  30. Cui, Z.; Chen, W.; Chen, Y. Multi-scale convolutional neuralnetworks for time series classification. arXiv 2016, arXiv:1603.06995. [Google Scholar]
  31. Phan, H.; Andreotti, F.; Cooray, N.; Chén, O.Y.; De Vos, M. Jointclassification and prediction CNN framework for automatic sleep stageclassifcation. IEEE Trans. Biomed. Eng. 2019, 66, 1285–1296. [Google Scholar] [CrossRef]
  32. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  33. Kemp, B.; Zwinderman, A.; Tuk, B.; Kamphuisen, H.; Oberye, J. Analysis of a sleep-dependent neuronal feedback loop: The slow-wave microcontinuity of the EEG. IEEE Trans. Biomed. Eng. 2000, 47, 1185–1194. [Google Scholar] [CrossRef] [PubMed]
  34. Goldberger, A.L.; Amaral, L.A.; Glass, L.; Hausdorff, J.M.; Ivanov, P.C.; Mark, R.G.; Mietus, J.E.; Moody, G.B.; Peng, C.-K.; Stanley, H.E. PhysioBank, PhysioToolkit, and PhysioNet: Components of a new research resource for complex physiologicsignals. Cirulation 2000, 101, e215–e220. [Google Scholar] [CrossRef]
  35. Wolpert, E.A. A manual of standardized terminology, techniques and scoring system for sleep stages of human subjects. Arch. Gen. Psychiatry 1969, 20, 246–247. [Google Scholar] [CrossRef]
  36. Sun, Y.; Wang, B.; Jin, J.; Wang, X. Deep convolutional network method for automatic sleep stage classification based on neurophysiological signals. In Proceedings of the 11th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Beijing, China, 13–15 October 2018; pp. 1–5. [Google Scholar]
  37. Supratak, A.; Guo, Y. TinySleepNet: An efficient deep learning model for sleep stage scoring based on raw single-channel EEG. In Proceedings of the 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society, EMBC, Montréal, QC, Canada, 20–24 July 2020; IEEE: New York, NY, USA, 2020; pp. 641–644. [Google Scholar]
Figure 1. MASleepNet Model.
Figure 1. MASleepNet Model.
Biomimetics 10 00642 g001
Figure 2. Multi-modal Scalable Feature Extraction Module.
Figure 2. Multi-modal Scalable Feature Extraction Module.
Biomimetics 10 00642 g002
Figure 3. Bi-LSTM Network.
Figure 3. Bi-LSTM Network.
Biomimetics 10 00642 g003
Figure 4. Visualization of the estimated hypnogram for subject SC406 of the: SleepEDF-78 dataset: (a) ground-truth hypnogram and (b) output hypnogram.
Figure 4. Visualization of the estimated hypnogram for subject SC406 of the: SleepEDF-78 dataset: (a) ground-truth hypnogram and (b) output hypnogram.
Biomimetics 10 00642 g004
Figure 5. Confusion matrix on the Sleep-EDF-78 dataset.
Figure 5. Confusion matrix on the Sleep-EDF-78 dataset.
Biomimetics 10 00642 g005
Figure 6. Confusion matrix on the Sleep-EDF-20 dataset.
Figure 6. Confusion matrix on the Sleep-EDF-20 dataset.
Biomimetics 10 00642 g006
Table 1. Sleep-EDF dataset.
Table 1. Sleep-EDF dataset.
DatasetWN1N2N3REMTOTAL
Sleep-EDF-208285280417,7995703771742,308
19.6%6.6%42.1%13.5%18.2%
Sleep-EDF-7865,95121,52269,13213,03925,835195,479
14.3%3.2%43.7%18.5%20.3%
Table 2. Overall experimental results on the Sleep-EDF-20 dataset and Per-class precision results on the Sleep-EDF-20 dataset.
Table 2. Overall experimental results on the Sleep-EDF-20 dataset and Per-class precision results on the Sleep-EDF-20 dataset.
ModelOverall Results (%)
ACCF1MacroKappaSenSpePre
DeepSleepNet [23]80.6676.6074.1480.0795.2175.75
ResNetLSTM [36]74.5869.9765.9671.9793.5471.14
AttnSleep [24]81.3977.0774.9879.7195.3277.02
TinySleepNet [37]80.5071.4073.1672.0994.8172.62
SeqSleepNet [15]75.0861.6865.0063.0093.0363.10
MASleepNet84.5378.8878.5378.4395.6980.21
ModelPre-Class Precisions (%)
Pre (N1)Pre (N2)Pre (N3)Pre (REM)Pre (Wake)
DeepSleepNet [23]36.8591.6479.0880.1289.58
ResNetLSTM [36]28.2985.9178.5374.8382.38
AttnSleep [24]37.7890.5182.0483.8488.45
TinySleepNet [37]36.4587.6983.1270.6782.05
SeqSleepNet [15]24.0981.0282.5659.3278.66
MASleepNet53.5586.3088.0484.0687.29
Table 3. Overall experimental results on the Sleep-EDF-78 dataset and Per-class precision results on the Sleep-EDF-78 dataset.
Table 3. Overall experimental results on the Sleep-EDF-78 dataset and Per-class precision results on the Sleep-EDF-78 dataset.
ModelResults (%)
ACCF1MacroKappaSenSpePre
DeepSleepNet [23]76.8772.2569.0676.6494.3670.86
ResNetLSTM [36]76.7772.8968.9076.4394.3171.75
AttnSleep [24]75.7071.7267.5475.3194.0770.70
TinySleepNet [37]78.0669.3969.2969.3194.0771.01
SeqSleepNet [15]73.8462.0263.2763.5292.9264.75
MASleepNet82.5676.1274.9575.8595.1676.94
ModelPre-Class Precisions (%)
Pre (N1)Pre (N2)Pre (N3)Pre (REM)Pre (Wake)
DeepSleepNet [23]39.4885.2757.8275.0596.27
ResNetLSTM [36]37.6386.1165.5774.2994.65
AttnSleep [24]35.7985.0864.0371.0496.48
TinySleepNet [37]40.1678.9680.2165.3389.06
SeqSleepNet [15]32.4479.8870.5055.4780.66
MASleepNet47.6481.2377.9679.9792.78
Table 4. Comparative experimental results on the Sleep-EDF-20 dataset.
Table 4. Comparative experimental results on the Sleep-EDF-20 dataset.
ModelResults (%)
ACCF1MacroKappaSenSpePre
No-MSConv82.0775.6675.2375.5795.0976.71
EEG-Only82.7374.9076.1775.0195.3277.31
No-SE84.2278.3378.3478.1495.6279.06
No-LSTM84.1878.3478.1278.1895.6579.56
No-Attn84.1378.2278.0678.0895.6378.95
MASleepNet84.5378.8878.5378.4395.6980.21
Table 5. Comparative experimental results on the Sleep-EDF-78 dataset.
Table 5. Comparative experimental results on the Sleep-EDF-78 dataset.
DatasetResults (%)
ACCF1MacroKappaSenSpePre
No-MSConv79.7772.6971.7672.8694.4973.80
EEG-Only79.9472.5272.0772.4194.6573.57
No-SE82.0575.9574.6175.3695.0876.62
No-LSTM81.5273.5674.2972.3494.9076.51
No-Attn81.8575.8874.7875.6895.1476.48
MASleepNet82.5676.1274.9575.8595.1676.94
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Gong, Z.; Wang, T.; Dong, Q.; Huang, Z.; Zhang, S.; Ma, Y. MASleepNet: A Sleep Staging Model Integrating Multi-Scale Convolution and Attention Mechanisms. Biomimetics 2025, 10, 642. https://doi.org/10.3390/biomimetics10100642

AMA Style

Wang Z, Gong Z, Wang T, Dong Q, Huang Z, Zhang S, Ma Y. MASleepNet: A Sleep Staging Model Integrating Multi-Scale Convolution and Attention Mechanisms. Biomimetics. 2025; 10(10):642. https://doi.org/10.3390/biomimetics10100642

Chicago/Turabian Style

Wang, Zhiyuan, Zian Gong, Tengjie Wang, Qi Dong, Zhentao Huang, Shanwen Zhang, and Yahong Ma. 2025. "MASleepNet: A Sleep Staging Model Integrating Multi-Scale Convolution and Attention Mechanisms" Biomimetics 10, no. 10: 642. https://doi.org/10.3390/biomimetics10100642

APA Style

Wang, Z., Gong, Z., Wang, T., Dong, Q., Huang, Z., Zhang, S., & Ma, Y. (2025). MASleepNet: A Sleep Staging Model Integrating Multi-Scale Convolution and Attention Mechanisms. Biomimetics, 10(10), 642. https://doi.org/10.3390/biomimetics10100642

Article Metrics

Back to TopTop