Next Article in Journal
Efficient Scheduling of Heterogeneous Messages in the FlexRay Dynamic Segment
Previous Article in Journal
Output Feedback Adaptive Tracking Control for Uncertain Strict-Feedback Nonlinear Systems with Full-State Constraints and Unknown Output Gain
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

MFSleepNet: An Interactive Multimodal Fusion Framework for Automatic Sleep Staging

1
School of Electronic and Electrical Engineering, Zhengzhou University of Science and Technology, Zhengzhou 450064, China
2
College of Electrical Engineering, Henan University of Technology, Zhengzhou 450001, China
*
Authors to whom correspondence should be addressed.
Sensors 2026, 26(10), 3085; https://doi.org/10.3390/s26103085
Submission received: 30 March 2026 / Revised: 5 May 2026 / Accepted: 7 May 2026 / Published: 13 May 2026
(This article belongs to the Section Biomedical Sensors)

Abstract

Accurate automatic sleep staging remains challenging due to complex temporal dynamics, inter-subject variability, and the difficulty of effectively integrating heterogeneous physiological signals. Electroencephalogram (EEG) and electrooculogram (EOG) recordings provide complementary information for sleep analysis; however, most existing multimodal approaches rely on simple feature concatenation, which limits their ability to capture structured inter-modality relationships. This paper proposes MFSleepNet, a multimodal sleep staging framework that explicitly models interactions between EEG and EOG signals. The proposed system incorporates a multimodal feature fusion module to enable bidirectional information exchange between modality-specific representations, followed by a gated temporal-channel attention mechanism to adaptively emphasize informative temporal segments and signal channels, facilitating joint representation learning while preserving modality-specific characteristics. Experiments on three public datasets (Sleep-EDF, SHHS, and HSP) under an epoch-level cross-validation protocol show that MFSleepNet consistently outperforms representative single-modality and multimodal baseline methods in terms of overall accuracy, Cohen’s κ, and Macro-F1. Ablation studies further demonstrate the contribution of each functional module. Correlation analysis indicates stage-dependent variations in EEG–EOG relationships, while interaction-based experiments show that explicit feature interaction improves both joint and modality-specific representations. Grad-CAM visualizations provide interpretability of model decisions. External validation on unseen subjects reveals a noticeable performance drop, highlighting the challenges of inter-subject variability and the limited baseline generalization capability of the model. To address this, a lightweight subject-specific adaptation strategy is introduced, which improves performance using a small amount of labeled subject-specific data. Overall, the proposed framework provides an effective and interpretable solution for multimodal sleep staging while emphasizing the importance of structured inter-modality interaction and subject-adaptive modeling in practical applications.
Keywords: sleep staging; physiological signal processing; EEG and EOG fusion; multimodal learning; subject-specific adaptation sleep staging; physiological signal processing; EEG and EOG fusion; multimodal learning; subject-specific adaptation

Share and Cite

MDPI and ACS Style

Gui, R.; Wang, C.; Niu, Q.; Wang, L. MFSleepNet: An Interactive Multimodal Fusion Framework for Automatic Sleep Staging. Sensors 2026, 26, 3085. https://doi.org/10.3390/s26103085

AMA Style

Gui R, Wang C, Niu Q, Wang L. MFSleepNet: An Interactive Multimodal Fusion Framework for Automatic Sleep Staging. Sensors. 2026; 26(10):3085. https://doi.org/10.3390/s26103085

Chicago/Turabian Style

Gui, Ranran, Chen Wang, Qunfeng Niu, and Li Wang. 2026. "MFSleepNet: An Interactive Multimodal Fusion Framework for Automatic Sleep Staging" Sensors 26, no. 10: 3085. https://doi.org/10.3390/s26103085

APA Style

Gui, R., Wang, C., Niu, Q., & Wang, L. (2026). MFSleepNet: An Interactive Multimodal Fusion Framework for Automatic Sleep Staging. Sensors, 26(10), 3085. https://doi.org/10.3390/s26103085

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop