Next Article in Journal
Technology Acceptance and Usability of a Therapy System with a Humanoid Robot Serving as Therapeutic Assistant for Post-Stroke Arm and Neurovisual Rehabilitation—An Evaluation Based on Stroke Survivors’ Experience
Previous Article in Journal
Hierarchical Structure of the Program Used by Filamentous Fungi to Navigate in Confining Microenvironments
Previous Article in Special Issue
Surface EMG Sensing and Granular Gesture Recognition for Rehabilitative Pouring Tasks: A Case Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MultiSEss: Automatic Sleep Staging Model Based on SE Attention Mechanism and State Space Model

1
Xi’an Key Laboratory of High Precision Industrial Intelligent Vision Measurement Technology, School of Electronic Information, Xijing University, Xi’an 710123, China
2
School of Physics and Electronic-Electrical Engineering, ABA Teachers College, Wenchuan 623002, China
3
School of Electronic Information, Xijing University, Xi’an 710123, China
*
Authors to whom correspondence should be addressed.
Biomimetics 2025, 10(5), 288; https://doi.org/10.3390/biomimetics10050288
Submission received: 2 April 2025 / Revised: 30 April 2025 / Accepted: 2 May 2025 / Published: 3 May 2025
(This article belongs to the Special Issue Artificial Intelligence (AI) in Biomedical Engineering)

Abstract

:
Sleep occupies about one-third of human life and is crucial for health, but traditional sleep staging relies on experts manually performing polysomnography (PSG), a process that is time-consuming, labor-intensive, and susceptible to subjective differences between evaluators. With the development of deep learning technologies, particularly the application of convolutional neural networks and recurrent neural networks, significant progress has been made in automatic sleep staging. However, existing methods still face challenges in feature extraction and cross-modal data fusion. This paper introduces an innovative deep learning architecture, MultiSEss, aimed at solving key issues in automatic sleep stage classification. The MultiSEss architecture utilizes a multi-scale convolution module to capture signal features from different frequency bands and incorporates a Squeeze-and-Excitation attention mechanism to enhance the learning of channel feature weights. Furthermore, the architecture discards complex attention mechanisms or encoder–decoder structures in favor of a state–space sequence coupling module, which more accurately captures and integrates correlations between multi-modal data. Experiments show that MultiSEss achieved accuracy results of 83.84% and 82.30% in five-fold cross-subject testing on the Sleep-EDF-20 and Sleep-EDF-78 datasets. MultiSEss demonstrates its potential in improving sleep stage accuracy, which is significant for enhancing the diagnosis and treatment of sleep disorders.

1. Introduction

Sleep accounts for about one-third of a person’s life and is a fundamental component of human health, influencing aspects such as memory consolidation, emotional regulation, physical recovery, and immune function. Chronic sleep disorders can lead to serious consequences, including cardiovascular diseases [1], diabetes [2], obesity [3], and mental health issues [4]. Sleep plays a crucial role in maintaining overall human health, making it essential to understand and monitor sleep quality and stages.
In clinical settings, the determination of sleep stages typically relies on doctors or specialists performing polysomnography (PSG), which includes electroencephalogram (EEG), electrooculogram (EOG), electromyogram (EMG), and electrocardiogram (ECG).
The frequency of EEG typically ranges from 0.5 Hz to 100 Hz. EOG measures the potential changes around the eyes caused by eye movements, and it is not strictly defined within a fixed frequency range. However, it can capture the relatively low-frequency fluctuations caused by eye movement, typically between 0.1 Hz and 10 Hz, depending on the speed and direction of the eye movement. The frequency range of EMG is generally between 5 Hz and 500 Hz, but in clinical practice, the focus is more often on the range from 5 Hz to 150 Hz. Traditionally, sleep experts or doctors first segment the nighttime PSG data into 30 s non-overlapping segments, and then manually classify these segments into different sleep stages according to sleep manuals (such as Rechtschaffen and Kales (R&K) [5] or American Academy of Sleep Medicine (AASM) standards [6]). For example, according to AASM, sleep time is divided into five stages: wakefulness (W), non-rapid eye movement stage 1 (N1), non-rapid eye movement stage 2 (N2), non-rapid eye movement stage 3 (N3), and rapid eye movement (REM). However, the classification of sleep stages is manually performed by trained technicians through the analysis of PSG data [7]. This process is time-consuming, labor-intensive, and prone to inconsistency due to inter-evaluator variability [8]. Therefore, developing automated methods for sleep stage classification has become especially urgent [9,10,11,12].
Recent studies have shown that deep learning algorithms have achieved significant success in automatic sleep staging. To accomplish this, researchers have widely adopted technologies such as Convolutional Neural Networks (CNNs) and Recurrent Neural Networks (RNNs). Among them, the single epoch model approach has attracted particular attention. This method uses a 30 s unit (i.e., epoch) as the time frame, treating the data of each epoch as input, and constructs deep learning models using CNNs to solve time series classification problems. For example, Xia et al. [13] proposed an end-to-end automatic sleep staging classification method based on raw single-channel sleep EEG signals. This method uses Convolutional Neural Networks (CNN) to extract the time-frequency domain features of the signal, incorporates compression and excitation modules (SE-Block) on the CNN to enhance its feature representation ability, and uses a bidirectional gated recurrent unit (Bi-GRU) to learn the transition rules of sleep stages. Additionally, an attention mechanism is added to the decoding part of the Bi-GRU to strengthen its long-term memory capabilities, highlighting the impact of key features. In order to effectively learn the intrinsic features of significant waves from single-channel EEG signals and capture useful information about sleep stage transition rules, Li et al. [14] proposed a new method, SleepFC. This method includes a Convolutional Feature Pyramid Network (CFPN), Cross-Scale Temporal Context Learning (CSTCL), and a class-adaptive fine-tuning loss function (CAFTLF) based classification network. CFPN learns multi-scale features from prominent waves in EEG signals. CSTCL extracts useful multi-scale transition rules between sleep stages. To address the challenge of effectively extracting temporal information from EEG, which faces issues of future information leakage and high model training time costs, Pan et al. [15] proposed a novel network architecture, CausalAttenNet. It prevents future information leakage through causal dilated convolution and extracts time-related sleep features with a wide temporal receptive field. Traditional recurrent networks are replaced by Multi-Head Attention (MHA), significantly reducing time costs and forming a faster sleep staging network. However, existing methods based on end-to-end CNN or RNN models that implicitly learn temporal correlations may limit the performance of learning time patterns, making accurate phase recognition challenging. Duan et al. [16] proposed a new multimodal and multiscale automatic sleep staging framework, MMS-SleepNet. It uses a deep learning-based multimodal feature extraction module (MMS-FE), incorporating expert knowledge to effectively capture multimodal features for each stage and fine-grained EEG features at various frequencies. The module leverages an attention mechanism to seamlessly integrate the extracted multimodal features, significantly improving classification accuracy. At the same time, a contrastive learning module and data balancing strategy were introduced to address class confusion and data imbalance issues in existing models. Ying et al. [17] proposed a deep learning-based hybrid public–private domain sleep grading network, HybridDomainSleepNet. This network explores the correlation between multimodal data and multiple domains to enhance the representation of different sleep stages. HybridDomainSleepNet can understand both the common and individual characteristics of subjects in different sleep stages and capture long- and short-term dependencies during sleep. Additionally, a multi-domain attention framework was proposed to capture brain signals across domains using a three-branch structure, exploring the correlations of multimodal data across different sleep stages.
There are three main challenges in building an effective sleep stage classification model: first, the difficulty of extracting effective features from raw EEG signals. According to AASM sleep standards, physiological signals in different sleep stages typically exhibit distinct characteristic waveforms, and EEG is time dependent. Therefore, many researchers use spatiotemporal models to extract data features. Secondly, the insufficient capture of cross-modal contextual relationships between different modalities. Different sleep stages are characterized by specific EEG waves, which are also reflected in other signals. Experts typically combine EEG with other signals (such as EOG) to determine sleep stages. For example, EEG waveforms in the REM and N1 stages are similar, while EOG waveforms show significant differences. However, many existing deep learning methods overlook the effectiveness of multimodal data, even though some studies have explored the importance of different modalities [18,19,20]. These studies usually employ simple methods at the model input or intermediate stages, such as concatenation, addition, or dot-product methods, to fuse raw signals or features learned from different modalities together. This leads to the neglect of cross-modal contextual relationships between different modalities, even when they are not in the same epoch. In the end, many model constructions incorporated basic modules such as RNN, LSTM, or Transformer. While they offer better performance, they come with higher computational complexity and costs when processing data like multi-channel multi-lead sleep graphs.
During sleep, the body’s physiological state continuously changes. These changes are reflected not only in individual signals but also in the interactions between multiple signals. For instance, EOG can detect eye movements, which is a fundamental indicator for distinguishing between Rapid Eye Movement (REM) and Slow Eye Movement (SEM). Additionally, the EEG waveforms during the REM and N1 stages are similar, while the EOG waveforms differ significantly. Therefore, when classifying REM and N1 stages, the contribution of EOG is greater than that of EEG. In contrast, the classification of N2 and N3 stages primarily relies on significant waveforms in the EEG. Thus, this paper chose EEG and EOG to construct our model [21].
To address these challenges, MultiSEss was present, which is an innovative deep learning architecture based on convolutional neural networks (CNNs), specifically engineered for automated sleep stage classification. First, this paper designs a multi-scale convolution module to capture low and high-frequency signals across different frequency bands. Next, we implemented a Squeeze-and-Excitation (SE) attention mechanism to obtain channel feature weights. At the end of the model, a state–space sequence coupling module was employed to learn the cross-modal contextual relationships between signals. Compared to traditional one-dimensional convolution or simple weighted fusion methods, the state–space sequence coupling module can more accurately capture correlations between modalities, avoiding information loss caused by modality heterogeneity.
Overall, the contributions of this paper can be summarized as follows: First, multi-scale convolution and SE attention mechanism modules were used to extract frequency features and channel weight features from signals. Second, a state space sequence coupling module was adopted to integrate global context features from multimodal samples and shared features across different samples, enhancing the model’s ability to handle multimodal data. Finally, extensive experiments were conducted on the Sleep-EDF-20 and Sleep-EDF-78 datasets to verify the effectiveness and feasibility of the method.

2. Materials and Methods

2.1. Sleep-EDF Dataset

In the experiments described in this paper, the Sleep-EDF-20 and Sleep-EDF-78 public datasets obtained from Physiobank [22] were used, as shown in Table 1. Sleep-EDF-20 contains PSG data from two nights for 20 subjects. Sleep-EDF-78 includes PSG data from 78 subjects, serving as an extended version of the Sleep-EDF-20 dataset. Secondly, they wore miniature sleep telemetry systems on the subjects. First, the researchers tested healthy subjects aged between 25 and 101 years, aiming to explore the impact of age on sleep. Secondly, they conducted sleep telemetry. They tested the effects of Temazepam on human sleep without using any other medications. In this dataset, the generated PSG includes EEG data from two channels (Fpz-Cz and Pz-Oz). Both channels have a sampling frequency of 100 Hz. The data for the entire night is recorded in two files: SC*PSG. The hypnograms were manually scored according to the Rechtschaffen and Kales protocol and then labeled as N1, N2, N3, N4, Wake, REM, and UNKNOWN classes. Since stage M and unknown stages do not belong to any sleep stage, they have been excluded from the experiment. In addition, according to AASM standards, stage N3 and stage N4 are combined into a single stage N3 in the experiment. In order to emphasize sleep stages, only the wakefulness within half an hour before and after sleep is included.

2.2. The Structure of MultiSEss Model

Figure 1 illustrates the basic structure of the MultiSEss model. As shown in Figure 1, the MultiSEss model consists of four parts: multi-scale convolutional layers, SE channel attention mechanism layer, SSM coupling layer, and output layer. First, the EEG signals and EOG signals are processed through the multi-scale convolutional layers. The multi-scale convolutional layers capture the features of the signals at various frequencies by using convolution kernels of different sizes, which helps the model understand the complexity of the input data from different dimensions. Each convolution kernel size extracts frequency information within a specific range, ensuring the model’s broad adaptability and sensitivity to the input signals. Subsequently, the features after convolution at different scales are concatenated for each signal. Next, the data passes through the SE attention mechanism, which uses a two-layer neural network to learn the importance of each channel in the feature map from the channel dimension data. This is then multiplied by the original channel data. This module can automatically identify which channel information is more crucial for the final task and adjust the weights of each channel accordingly. Next, the data are passed through the SSM coupling layer to more accurately capture and fuse the correlations between multimodal data. Finally, the classification is output through a two-layer fully connected network and the softmax function.

2.2.1. Multiscale Convolution

In order to extract a more robust temporal representation from the raw sleep signals, this paper introduces a multi-scale convolutional neural network (MSCNN) [23,24]. The MSCNN module is shown in Figure 2 with its core mechanism lying in its feature extraction approach, which utilizes two convolutional layers with parallel filters of different sizes. Specifically, this method uses small-sized kernels to capture low-frequency information in the signal, while large-sized kernels are responsible for capturing high-frequency information. This design allows the model to consider different frequency components simultaneously when processing complex signals, thereby improving recognition accuracy. After each convolutional layer, a max-pooling layer follows, helping highlight the key features within each filter’s scanning region. The max-pooling operation not only simplifies the model’s learning process but also helps the model focus on the most important information points while disregarding less relevant details. For example, in the specific implementation of MSCNN, the large-scale convolutional kernel in the first layer is set to a size of (1, 50), with a stride of 20 and 64 filters; correspondingly, the small-scale convolutional kernel is set to a size of (1, 20), with a stride of 5, and also contains 64 filters. The subsequent max-pooling layer operates with a kernel size of (4, 4). Furthermore, signals from EEG (electroencephalogram) and EOG (electrooculogram) are combined through multiple convolutions and pooling, followed by a Concatenation operation. This step aims to merge two different types of physiological signals in order to generate more comprehensive and representative feature representations for subsequent analysis tasks. This combination not only enhances the model’s understanding of the input signals but also provides strong support for accurate recognition. Overall, MSCNN offers an effective method for processing complex sleep signal data, demonstrating its potential in the medical and health field, especially in sleep research.

2.2.2. Squeeze-and-Excitation Networks

The Squeeze-and-Excitation Networks (SENets), first proposed by Hu et al. [25], can scale channel feature maps by explicitly focusing on the interdependence between channels. Since its introduction, the SE module has been widely adopted due to its excellent performance in emphasizing key features and suppressing unnecessary ones. This structure significantly enhances the neural network’s ability to represent features, thereby improving the model’s recognition accuracy. The SE attention mechanism structure is shown in Figure 3.
The working principle of the SE module can be divided into two main stages: Squeeze and Excitation. First, as shown in Equation (1), during the squeezing stage, through the global average pooling operation, the spatial information of each channel in the input feature map U is compressed into a scalar value, forming a C-dimensional vector z, where C represents the number of channels in the feature map, H is the height, and W is the width. Specifically, the global average pooling result zc for each channel C is obtained by averaging all spatial locations of that channel, calculated as
z c = 1 H × W i = 1 H j = 1 W U i , j , c   f o r   c = 1 , 2 , , C
Here, zc represents the global average pooling result of the C channel.
Subsequently, as shown in formula (2), during the incentive phase, an adaptive weight generation mechanism calculates a corresponding weight value for each channel to achieve selective enhancement or suppression of channels. This process involves two fully connected layers (FC) and nonlinear activation functions (such as ReLU and Sigmoid) to establish relationships between channels. First, the first fully connected layer reduces the dimensionality, transforming the channel descriptor vector z into a lower-dimensional space, where r serves as the dimensionality reduction ratio. Next, after processing through the ReLU and Sigmoid activation functions, the second fully connected layer maps the low-dimensional representation back to the original dimensions, generating weights for each channel.
u = σ ( W 2 δ ( W 1 z ) )
In Equation (2), δ represents the ReLU activation function, σ represents the Sigmoid activation function, and W1 and W2 are the weighted matrices of the two fully connected layers.
In the final stage of feature recalibration, as shown in Equation (3), the SE module multiplies each channel in the original feature map by its corresponding weight to achieve adaptive channel weighting. The output feature map is calculated as follows:
X i , j , c = s c U i , j , c ,   f o r   c = 1 , 2 , , C

2.2.3. State Space Model Coupling Module

Currently, in the field of deep learning, although Transformer models are widely popular for their excellent ability to handle long-range dependencies, their enormous parameter size and high computational complexity have become significant bottlenecks. In recent years, the State Space Model (SSM), as an emerging efficient architecture, has shown great potential in handling continuous long-sequence data, and SSM is the core module of Mamba [26,27]. Compared to traditional transformers, SSM offers notable advantages, especially in terms of computational efficiency. The linear time complexity of SSM makes it far more efficient in processing long sequences than the quadratic complexity of transformers, making it particularly suitable for applications that require analysis of ultra-long contexts. Additionally, the mathematical framework of SSM, based on state transition equations and observation equations, provides the model with stronger interpretability. SSM is not limited to natural language processing tasks but can also be applied to time series prediction, speech processing, reinforcement learning, and other fields, demonstrating its powerful capability in solving complex problems. However, SSM has limitations in capturing interaction information between different modalities. To address this issue, this paper proposes a novel module that combines the SSM concept with the attention mechanism—the Multimodal Coupling Module (SSM Coupling Module)—to enhance the interaction performance between modalities. Below is the data processing flow for this module:
The SSM coupling module process is as follows. First, the modal input is defined as a three-dimensional tensor XnϵRB×T×D, where B represents the batch size, T is the time step, and D is the input feature dimension. The core of the model lies in its effective processing and integration of different modal data. The model includes several key matrices to implement the transformation process from input to output. For example, the state update matrix AϵRH×H is responsible for the transformation between hidden states, where H denotes the hidden state dimension. The input projection matrix BϵRH×D is used to map the input features to the hidden state space. The modal correlation matrix SnϵRH×H aims to capture the relationships between different modalities, thereby enhancing the model’s ability to understand multimodal data. The output transformation matrix is the key in mapping the hidden state back to the original feature space. Additionally, to dynamically evaluate and fuse information from different modalities, the weighted fusion parameter Wn is introduced. This mechanism allows the model to adaptively adjust the contribution of each modality based on the importance of the current input, thereby improving overall performance and flexibility. This design is especially suitable for applications that require handling and integrating various types of data.
State update mechanism: First, based on Equation (4), layer normalization is performed on the input of the n-th modality at time step t to standardize the input features. Then, according to Equation (5), a cross-modal interaction representation is constructed by aggregating the hidden states of all modalities.
x n ( t ) = L ayer N orm ( x n ( t ) )
H s u m ( t ) = n = 1 N h n ( t 1 )
Next, as shown in Equation (6), the current input features combined with the historical hidden states of each modality are used to generate the hidden state for the next time step.
h n ( t ) = tanh ( m = 1 M h m ( t 1 ) S n + X ( t ) B T ) R B × H
Among them, M represents the number of modalities, hm(t − 1) is the hidden state of the m-th modality at time step t − 1, Sn is the correlation matrix of modality n, and BT is the transpose of the input projection matrix.
Cross-modal Interaction and Attention Mechanism:
In order to further enhance the interaction effect between modalities, by combining the historical states of all modalities with the current input, according to Formula (7), the similarity score(hn(t),hm(t)) between modality n and modality m at time step t is calculated to achieve adaptive weighting between modalities. Finally, as shown in Formula (8), the hidden state of modality n at time step t will interact with the states of other modalities through attention weighting:
α n m ( t ) = exp ( s c o r e ( h n ( t ) , h m ( t ) ) ) m = 1 M exp ( s c o r e ( h n ( t ) , h m ( t ) ) )
h n ( t ) = tanh ( m = 1 M α n m ( t ) h n ( t 1 ) S n + X ( t ) B T )
Output:
According to Formula (9), the output Yn for each modality is obtained by projecting the corresponding hidden state hn(t) into the output space through the linear transformation matrix ET. Then, as per Formula (10), the model performs a weighted integration of the outputs from all modalities to generate the final multimodal output Y. Here, Wn is the weighting parameter for each modality, which can adaptively adjust the contribution of each modality to the final result.
Y = h n E T
Y = 1 M n = 1 M W n Y n

2.3. Evaluation Indexes

The performance of the proposed model is comprehensively evaluated by several evaluation indexes. Accuracy (ACC) measures the proportion of correct predictions made by the model, but it may not be fully representative in imbalanced datasets. The Macro-F1 score (MF1) takes both precision and recall into account by averaging the F1 scores of each category, providing a holistic evaluation that is not affected by sample distribution, making it particularly suitable for multi-class problems. Cohen’s Kappa (Kappa) measures the consistency between the model’s predictions and the true labels while considering the probability of random agreement, with values closer to 1 indicating higher consistency. Sensitivity (Sen), also known as Recall rate, focuses on the proportion of actual positive samples correctly identified by the model; Specificity (Spec) focuses on the proportion of actual negative samples correctly identified as negative by the model. Precision (Pre) reflects the proportion of samples predicted as positive by the model that are actually positive, showing the accuracy of the model’s predictions. In order to analyze the performance of the model in more detail, two calculation methods are used in this paper. Overall Results provide a comprehensive view of the model’s performance across the entire test set, offering an initial understanding of the model’s overall effectiveness. While Per-Class Precisions, starting from the individual categories, calculate the model’s performance in each category separately, allowing for a detailed display of the differences between categories, which is especially important when dealing with imbalanced class distributions. By combining these two methods, we can assess the model’s actual performance more comprehensively and accurately, ensuring that not only the general trends are observed, but also potential specific issues are identified.
The formulas for the evaluation metrics Accuracy, Macro-F1 score, Cohen’s Kappa, Sensitivity, Specificity, and Precision are shown in (11)–(15). True Positive (TP) refers to the model correctly predicting samples that are actually of the positive class as positive. False Negative (FN) refers to the model incorrectly predicting samples that are actually of the positive class as negative. False Positive (FP) refers to the model incorrectly predicting samples that are actually of the negative class as positive. True Negative (TN) refers to the model correctly predicting samples that are actually of the negative class as negative. po is the observed agreement probability, and pe is the expected agreement probability.
a c c u r a c y = T P + T N T P + T N + F P + F N
M a c r o - F 1   s c o r e = 2 1 p r e c i s i o n + 1 r e c a l l
C o h e n K a p p a = p o - p e 1 - p e
S ensitivity = T P T P + F N
S p e c i f i c i t y = T N T N + F P
p r e c i s i o n = T P T P + F P

3. Experimental Results and Analysis

3.1. Experimental Setup

In the experiment, this paper used a five-fold cross-validation method across subjects to evaluate model performance on two datasets. Specifically, participants in each dataset were randomly and evenly divided into five groups, ensuring that each fold accurately reflected the overall sample’s feature distribution. This method not only enhanced the reliability of the results but also effectively reduced the impact of individual differences on model evaluation, maximizing data utilization and evaluation effectiveness. During the model training phase, the Adam optimizer with a learning rate of 0.001 was chosen. The Adam optimizer combines the advantages of AdaGrad and RMSProp, providing an adaptive learning rate that dynamically adjusts based on each parameter’s historical gradient, making it especially suitable for handling sparse gradients. To address the class imbalance issue in the dataset, a weighted cross-entropy loss function was employed, with the importance weights of different class samples being adaptively adjusted to ensure balanced information acquisition from all categories by the model during the learning process. Additionally, to ensure consistency in the data used for both the training and testing sets, all data from the pre-trained model was set with the same random seed, shuffled randomly, and then sent to the network model. To further enhance the model’s expressiveness and generalization ability, the batch size was configured as 128 throughout the training process, with 100 training epochs being systematically executed. This configuration aims to optimize through extensive iterations, enabling the model to learn effectively from diverse data, thereby improving its predictive ability and stability. These strategies work together to ensure the reliability of the experimental results and the model’s high performance.

3.2. Result of MultiSEss Model

This study focuses on evaluating the performance of a novel MultiSEss model in sleep signal recognition from EEG data. The effectiveness of the MultiSEss model was comprehensively validated through comparative experiments with several state-of-the-art deep learning algorithms. The models involved in the comparison include U-time [28], ResnetLSTM [29], Cross-Modal Transformer [30], AttnSleep [31], and MMASleepNet [32]. The experimental results, as shown in Table 2 and Table 3, reveal that among the six models evaluated, MMASleepNet performs the worst. Specifically, on the Sleep-EDF-20 dataset, MMASleepNet’s Accuracy, Macro-F1 score, Cohen’s Kappa coefficient, sensitivity, specificity, and Precision are 79.04%, 71.87%, 70.02%, 78.33%, 94.81%, and 70.55%, respectively; In the Sleep-EDF-78 dataset, these metrics are 76.62%, 69.53%, 67.92%, 78.05%, 94.26%, and 67.39%. In comparison, the proposed MultiSEss model demonstrates superior performance. On the Sleep-EDF-20 dataset, the model achieved evaluation metrics of 83.84%, 73.46%, 75.36%, 72.20%, 95.31%, and 95.15%; for the Sleep-EDF-78 dataset, these values were 82.30%, 72.23%, 74.21%, 71.10%, 95.02%, and 74.63%.
In addition, considering model complexity, U-time and ResnetLSTM have higher computational demands due to the use of recursive or convolutional neural network structures, especially when processing long time series. Cross-Modal Transformer has a higher computational cost because of its complex attention mechanism. While AttnSleep uses attention mechanisms to enhance feature extraction capabilities, its design also results in a relatively higher number of parameters and computational load. In contrast, although MMASleepNet has lower performance, it shows certain advantages in terms of computational resource consumption, which may be related to its relatively simple architecture. However, it is worth noting that the proposed MultiSEss model, despite including several modules (such as MSCNN, SE attention mechanism module, and SSM module), ensures high performance while maintaining a reasonable level of computational complexity and parameter count through careful design. This allows it to not only provide high-precision sleep staging results in practical applications but also effectively control the use of computational resources.
Furthermore, the study observed that the classification accuracy for N1, N3, and REM sleep stages was relatively high for both the Sleep-EDF-20 and Sleep-EDF-78 datasets. This indicates that the MultiSEss model proposed in this paper can more effectively perform feature extraction and fusion operations when processing multimodal electrophysiological signals, leading to more accurate sleep staging. This result further demonstrates the advantages of the MultiSEss model over other baseline methods.

3.3. Ablation Study

To further explore the contributions of the spatial module and temporal module in the MultiSEss model, ablation tests were carried out on the Sleep-EDF-20 and Sleep-EDF-78 datasets, using a five-fold cross-subject experimental setup. The results of the ablation experiments are presented in Table 4 and Table 5, as well as Figure 4. Specifically, MultiSEss is composed of the MSCNN module and the SE attention mechanism module. and the SSM module. The MSCNN and SE attention mechanism modules collaborate to extract spatial features, while the SSM module is dedicated to capturing temporal features. The individual importance of these two modules can be better understood through systematic comparison of their effects on the overall model performance. Based on the data in Table 4 and Table 5, from the Sleep-EDF-20 dataset, the spatial feature extraction model that only includes the MSCNN and SE attention mechanism (MSCNN-SE) achieved an accuracy of 82.99%, while the complete MultiSEss model, integrating all three modules, achieved an accuracy of 83.84%. Similarly, on the Sleep-EDF-78 dataset, MSCNN-SE achieved an accuracy of 81.01%, while the complete MultiSEss model reached an accuracy of 82.30%. From these results, it can be seen that, although the MSCNN-SE combined model’s performance is slightly lower than that of the entire MultiSEss model, the performance of the individual SSM module is significantly lower. This suggests that the spatial features extracted by the MSCNN-SE module dominate the overall model performance. However, it is noteworthy that, when the SSM module is added to enhance the model’s ability to capture temporal features, the overall performance of the model is further enhanced. This finding highlights the significance of modeling the interdependence between features, particularly temporal features. In addition to focusing on spatial features when dealing with complex physiological signal data, this study demonstrates that, while spatial feature extraction is a key factor in improving classification performance, combining temporal feature modeling can further optimize the model’s performance. This is crucial for achieving more accurate sleep stage recognition.

4. Conclusions

This paper presents a multi-channel sleep signal automatic classification model, MultiSEss, based on EEG and EOG signals. With its unique design, the model efficiently extracts key features from complex EEG signals and uses these features to achieve precise sleep stage classification. MultiSEss employs a multi-scale convolutional neural network to extract rich feature information from EEG data across different frequency bands. Additionally, it incorporates a channel attention mechanism module to identify and weight the data features from each channel, thereby enhancing the model’s focus on important features. Furthermore, to ensure that the global context information of each signal is effectively captured, MultiSEss introduces an SSM module to further improve classification accuracy. In experimental evaluation, MultiSEss was tested on two widely used datasets, Sleep-EDF-20 and Sleep-EDF-78, with five-fold cross-validation accuracy reaching 83.84% and 82.30%, respectively, proving the model’s superior performance. It is worth noting that MultiSEss is an end-to-end model network that can directly learn features from raw sleep signals and perform classification without relying on manually crafted feature extraction processes. This not only improves work efficiency but also demonstrates the model’s potential in future development of efficient and accurate real-time brain–computer interface frameworks. The dataset will be systematically expanded in both scale and diversity in future work to enable more comprehensive evaluation of the model’s performance. To enhance the model’s robustness in noisy data environments, a specialized masking module is being developed to effectively mitigate noise interference.
Subsequent optimization efforts will prioritize performance enhancement of the MultiSEss model, encompassing systematic exploration of its integration within sleep signal-based usability testing frameworks. These refinements are projected to not only strengthen the model’s robustness and prediction accuracy but also significantly expand its practical deployment scenarios.

Author Contributions

Conceptualization, data curation, investigation, methodology, validation, visualization, writing—original draft, Z.H.; data curation, project administration, resources, supervision Y.M.; supervision, writing—review and editing Y.M. and S.Z.; investigation, Y.Y., Z.W., Y.L. and Z.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the General Projects of Shaanxi Science and Technology Plan (No.2023-JC-YB-504), the National Natural Science Foundation of China (No. 62172338), the Shaanxi province innovation capacity support program (No.2018KJXX-095), and the Xijing University special talent research fund (No.XJ17T03).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data were obtained from [physiobank] and are available from the authors/at Sleep-EDF Database Expanded.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kasasbeh, E.; Chi, D.S.; Krishnaswamy, G. Inflammatory aspects of sleep apnea and their cardiovascular consequences. South Med. J. 2006, 99, 58–68. [Google Scholar] [CrossRef] [PubMed]
  2. Khalil, M.; Power, N.; Graham, E.; Deschênes, S.S.; Schmitz, N. The association between sleep and diabetes outcomes—A systematic review. Diabetes Res. Clin. Pract. 2020, 161, 108035. [Google Scholar] [CrossRef]
  3. Hargens, T.A.; Kaleth, A.S.; Edwards, E.S.; Butner, K.L. Association between sleep disorders, obesity, and exercise: A review. Nat. Sci. Sleep 2013, 5, 27–35. [Google Scholar] [CrossRef]
  4. Reis, C.; Dias, S.; Rodrigues, A.M.; Sousa, R.D.; Gregório, M.J.; Branco, J.; Canhão, H.; Paiva, T. Sleep duration, lifestyles and chronic diseases: A cross-sectional population-based study. Sleep Sci. 2018, 11, 217–230. [Google Scholar] [CrossRef]
  5. Wolpert, E.A. A Manual of Standardized Terminology, Techniques and Scoring System for Sleep Stages of Human Subjects. Arch. Gen. Psychiatry 1969, 20, 246–247. [Google Scholar] [CrossRef]
  6. Iber, C. The AASM Manual for the Scoring of Sleep and Associated Events: Rules, Terminology, and Technical Specification. 2007. Available online: https://cir.nii.ac.jp/crid/1370004237604151044 (accessed on 1 May 2025).
  7. Rundo, J.V.; Downey, R., III. Polysomnography. Handb. Clin. Neurol. 2019, 160, 381–392. [Google Scholar] [PubMed]
  8. Phan, H.; Mikkelsen, K. Automatic sleep staging of EEG signals: Recent development, challenges, and future directions. Physiol. Meas. 2022, 43, 04TR01. [Google Scholar] [CrossRef]
  9. Yue, H.; Chen, Z.; Guo, W.; Sun, L.; Dai, Y.; Wang, Y.; Ma, W.; Fan, X.; Wen, W.; Lei, W. Research and application of deep learning-based sleep staging: Data, modeling, validation, and clinical practice. Sleep Med. Rev. 2024, 74, 101897. [Google Scholar] [CrossRef]
  10. Loh, H.W.; Ooi, C.P.; Vicnesh, J.; Oh, S.L.; Faust, O.; Gertych, A.; Acharya, U.R. Automated detection of sleep stages using deep learning techniques: A systematic review of the last decade (2010–2020). Appl. Sci. 2020, 10, 8963. [Google Scholar] [CrossRef]
  11. Alattar, M.; Govind, A.; Mainali, S. Artificial intelligence models for the automation of standard diagnostics in sleep medicine—A systematic review. Bioengineering 2024, 11, 206. [Google Scholar] [CrossRef]
  12. Soleimani, R.; Barahona, J.; Chen, Y.; Bozkurt, A.; Daniele, M.; Pozdin, V.; Lobaton, E. Advances in Modeling and Interpretability of Deep Neural Sleep Staging: A Systematic Review. Physiologia 2023, 4, 1–42. [Google Scholar] [CrossRef]
  13. Heng, X.; Wang, M.; Wang, Z.; Zhang, J.; He, L.; Fan, L. Leveraging discriminative features for automatic sleep stage classification based on raw single-channel EEG. Biomed. Signal Process. Control 2024, 88, 105631. [Google Scholar] [CrossRef]
  14. Li, W.; Liu, T.; Xu, B.; Song, A. SleepFC: Feature Pyramid and Cross-Scale Context Learning for Sleep Staging. IEEE Trans. Neural Syst. Rehabil. Eng. 2024, 32, 2198–2208. [Google Scholar] [CrossRef] [PubMed]
  15. Pan, J.; Feng, Y.; Zhao, P.; Zou, X.; Hou, A.; Che, X. Causalattennet: A fast and long-term-temporal network for automatic sleep staging with single-channel eeg. IEEE Trans. Instrum. Meas. 2024, 73, 1–13. [Google Scholar] [CrossRef]
  16. Duan, L.; Ma, B.; Yin, Y.; Huang, Z.; Qiao, Y. MMS-SleepNet: A knowledge-based multimodal and multiscale network for sleep staging. Biomed. Signal Process. Control 2025, 103, 107370. [Google Scholar] [CrossRef]
  17. Ying, S.; Wang, L.; Zhang, L.; Xie, J.; Ren, J.; Qin, Y.; Liu, T. HybridDomainSleepNet: A hybrid common-private domain deep learning network for automatic sleep staging. Biomed. Signal Process. Control 2025, 103, 107436. [Google Scholar] [CrossRef]
  18. Chambon, S.; Galtier, M.N.; Arnal, P.J.; Wainrib, G.; Gramfort, A. A deep learning architecture for temporal sleep stage classification using multivariate and multimodal time series. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 758–769. [Google Scholar] [CrossRef]
  19. Phan, H.; Chén, O.Y.; Tran, M.C.; Koch, P.; Mertins, A.; De Vos, M. XSleepNet: Multi-view sequential model for automatic sleep staging. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 5903–5915. [Google Scholar] [CrossRef]
  20. Kwon, K.; Kwon, S.; Yeo, W.H. Automatic and accurate sleep stage classification via a convolutional deep neural network and nanomembrane electrodes. Biosensors 2022, 12, 155. [Google Scholar] [CrossRef]
  21. Wang, X.; Zhu, Y.; Lai, W. MIASS: A multi-interactive attention model for sleep staging via EEG and EOG signals. Comput. Electr. Eng. 2025, 121, 109852. [Google Scholar] [CrossRef]
  22. Kemp, B.; Zwinderman, A.; Tuk, B.; Kamphuisen, H.; Oberye, J. Analysis of a sleep-dependent neuronal feedback loop: The slow-wave microcontinuity of the EEG. IEEE Trans. Biomed. Eng. 2000, 47, 1185–1194. [Google Scholar] [CrossRef]
  23. Hu, D.; Gao, W.; Ang, K.K.; Hu, M.; Huang, R.; Chuai, G.; Li, X. CHMMConvScaleNet: A hybrid convolutional neural network and continuous hidden Markov model with multi-scale features for sleep posture detection. Sci. Rep. 2025, 15, 12206. [Google Scholar] [CrossRef] [PubMed]
  24. Zhong, J. Dynamic Multi-Scale Feature Fusion for Robust Sleep Stage Classification Using Single-Channel EEG. J. Comput. Electron. Inf. Manag. 2025, 16, 14–20. [Google Scholar] [CrossRef]
  25. JHu, L.; Shen, G. Sun, Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  26. He, X.; Cao, K.; Zhang, J.; Yan, K.; Wang, Y.; Li, R.; Xie, C.; Hong, D.; Zhou, M. Pan-mamba: Effective pan-sharpening with state space model. Inf. Fusion 2025, 115, 102779. [Google Scholar] [CrossRef]
  27. Zhang, T.; Yuan, H.; Qi, L.; Zhang, J.; Zhou, Q.; Ji, S.; Yan, S.; Li, X. Point cloud mamba: Point cloud learning via state space model. Proc. AAAI Conf. Artif. Intell. 2025, 39, 10121–10130. [Google Scholar] [CrossRef]
  28. Perslev, M.; Jensen, M.; Darkner, S.; Jennum, P.J.; Igel, C. U-time: A fully convolutional network for time series segmentation applied to sleep staging. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2019; Volume 32. [Google Scholar]
  29. Sun, Y.; Wang, B.; Jin, J.; Wang, X. Deep convolutional network method for automatic sleep stage classification based on neurophysiological signals. In Proceedings of the 2018 11th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Beijing, China, 13–15 October 2018; pp. 1–5. [Google Scholar]
  30. Pradeepkumar, J.; Anandakumar, M.; Kugathasan, V.; Suntharalingham, D.; Kappel, S.L.; De Silva, A.C.; Edussooriya, C.U.S. Towards interpretable sleep stage classification using cross-modal transformers. IEEE Trans. Neural Syst. Rehabil. Eng. 2024, 32, 2893–2904. [Google Scholar] [CrossRef]
  31. Eldele, E.; Chen, Z.; Liu, C.; Wu, M.; Kwoh, C.-K.; Li, X.; Guan, C. An attention-based deep learning approach for sleep stage classification with single-channel EEG. IEEE Trans. Neural Syst. Rehabil. Eng. 2021, 29, 809–818. [Google Scholar] [CrossRef]
  32. Yubo, Z.; Yingying, L.; Bing, Z.; Lin, Z.; Lei, L. MMASleepNet: A multimodal attention network based on electrophysiological signals for automatic sleep staging. Front. Neurosci. 2022, 16, 973761. [Google Scholar] [CrossRef]
Figure 1. MultiSEss model structure.
Figure 1. MultiSEss model structure.
Biomimetics 10 00288 g001
Figure 2. Multiscale convolution module.
Figure 2. Multiscale convolution module.
Biomimetics 10 00288 g002
Figure 3. SE attention mechanism module.
Figure 3. SE attention mechanism module.
Biomimetics 10 00288 g003
Figure 4. MultiSEss ablated experimental results on two datasets, Sleep-EDF-20 and Sleep-EDF-78.
Figure 4. MultiSEss ablated experimental results on two datasets, Sleep-EDF-20 and Sleep-EDF-78.
Biomimetics 10 00288 g004
Table 1. Sample size and sample proportion in the SLEEP-EDF dataset.
Table 1. Sample size and sample proportion in the SLEEP-EDF dataset.
DatasetWN1N2N3REMTotal
Sleep-EDF-208285
19.6%
2804
6.6%
17,799
42.1%
5703
13.5%
7717
18.2%
42,308
Sleep-EDF-7865,951
14.3%
21,522
3.2%
69,132
43.7%
13,039
18.5%
25,835
20.3%
195,479
Table 2. Performance of automatic installment on the Sleep-EDF-20 dataset.
Table 2. Performance of automatic installment on the Sleep-EDF-20 dataset.
ModelOverall Results (%)Per-Class Precisions (%)
ACCMF1KappaSenSpecPrePre(N1)Pre(N2)Pre(N3)Pre(REW)Pre(Wake)
U-time80.5272.0771.8777.8695.0370.5637.6692.9957.2475.8989.02
ResnetLSTM81.9673.6073.5075.5195.1572.7337.9290.7768.8776.1289.97
Cross-Modal Transformer81.2972.4972.7376.1295.1370.6036.7692.7261.7673.8287.94
AttnSleep81.3372.4472.6675.5995.0970.5136.0292.8164.4273.5385.70
MMASleepNet79.0471.8770.0278.3394.8170.5533.3694.2457.1677.1390.83
MultiSEss83.8473.4675.3672.2095.1575.8745.0787.6979.0178.0489.56
Table 3. Performance of automatic installment on the Sleep-EDF-78 dataset.
Table 3. Performance of automatic installment on the Sleep-EDF-78 dataset.
ModelOverall Results(%)Per-Class Precisions(%)
ACCMF1KappaSenSpecPrePre(N1)Pre(N2)Pre(N3)Pre(REW)Pre(Wake)
U-time79.7471.8470.9773.5094.7969.7838.9784.4260.0472.9492.53
ResnetLSTM79.1871.5070.8776.5794.7569.1240.1588.8652.0870.4894.18
Cross-Modal Transformer79.1671.8470.9777.8294.7969.5039.2689.5250.1574.6393.92
AttnSleep80.0673.6272.2779.4995.0170.7242.0089.1653.9173.2595.29
MMASleepNet76.6269.5367.9278.0594.2667.3938.5788.3741.0973.9095.04
MultiSEss82.3072.2374.2171.1095.0274.6347.6283.5175.3675.1191.58
Table 4. Ablation experiment results of MultiSEss on the Sleep-EDF-20 dataset.
Table 4. Ablation experiment results of MultiSEss on the Sleep-EDF-20 dataset.
ModelOverall Results (%)Per-Class Precisions (%)
ACCMF1KappaSenSpecPrePre(N1)Pre(N2)Pre(N3)Pre(REW)Pre(Wake)
MSCNN-SE82.9969.3273.9367.9994.8975.7050.1087.9982.1375.4482.84
SSM51.8323.3913.2325.4182.5930.802.5056.3030.3432.3532.55
MultiSEss83.8473.4675.3672.2095.1575.8745.0787.6979.0178.0489.56
Table 5. Ablation Experiment Results of MultiSEss on the Sleep-EDF-78 Dataset.
Table 5. Ablation Experiment Results of MultiSEss on the Sleep-EDF-78 Dataset.
ModelOverall Results (%)Per-Class Precisions (%)
ACCMF1KappaSenSpecPrePre(N1)Pre(N2)Pre(N3)Pre(REW)Pre(Wake)
MSCNN-SE81.0167.9872.1766.3894.6073.2045.2381.9677.3070.9990.52
SSM53.1528.5427.0230.7285.4538.5125.3558.1129.2129.7950.12
MultiSEss82.3072.2374.2171.1095.0274.6347.6283.5175.3675.1191.58
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Huang, Z.; Yang, Y.; Wang, Z.; Li, Y.; Chen, Z.; Ma, Y.; Zhang, S. MultiSEss: Automatic Sleep Staging Model Based on SE Attention Mechanism and State Space Model. Biomimetics 2025, 10, 288. https://doi.org/10.3390/biomimetics10050288

AMA Style

Huang Z, Yang Y, Wang Z, Li Y, Chen Z, Ma Y, Zhang S. MultiSEss: Automatic Sleep Staging Model Based on SE Attention Mechanism and State Space Model. Biomimetics. 2025; 10(5):288. https://doi.org/10.3390/biomimetics10050288

Chicago/Turabian Style

Huang, Zhentao, Yuyao Yang, Zhiyuan Wang, Yuan Li, Zuowen Chen, Yahong Ma, and Shanwen Zhang. 2025. "MultiSEss: Automatic Sleep Staging Model Based on SE Attention Mechanism and State Space Model" Biomimetics 10, no. 5: 288. https://doi.org/10.3390/biomimetics10050288

APA Style

Huang, Z., Yang, Y., Wang, Z., Li, Y., Chen, Z., Ma, Y., & Zhang, S. (2025). MultiSEss: Automatic Sleep Staging Model Based on SE Attention Mechanism and State Space Model. Biomimetics, 10(5), 288. https://doi.org/10.3390/biomimetics10050288

Article Metrics

Back to TopTop