Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (392)

Search Parameters:
Keywords = STFT

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
27 pages, 3545 KB  
Article
Deep Learning-Based Alzheimer’s Detection from Multi-Channel EEG Using Fused Time–Frequency Image Grids
by Abdulnasır Yıldız and Hasan Zan
Diagnostics 2026, 16(5), 746; https://doi.org/10.3390/diagnostics16050746 - 2 Mar 2026
Abstract
Background/Objectives: Dementia is a progressive neurodegenerative disorder for which accurate and timely diagnosis remains a major clinical challenge. Electroencephalography (EEG) offers a noninvasive and cost-effective means of capturing neurophysiological alterations, motivating the development of reliable EEG-based automated diagnostic frameworks. This study aims to [...] Read more.
Background/Objectives: Dementia is a progressive neurodegenerative disorder for which accurate and timely diagnosis remains a major clinical challenge. Electroencephalography (EEG) offers a noninvasive and cost-effective means of capturing neurophysiological alterations, motivating the development of reliable EEG-based automated diagnostic frameworks. This study aims to systematically examine how different time–frequency representations (TFRs) affect dementia classification performance within a unified multi-channel EEG image fusion framework. Methods: Resting-state, eyes-closed EEG recordings from 88 subjects, including Alzheimer’s disease, frontotemporal dementia, and cognitively normal controls, were preprocessed and segmented. Channel-wise signals were converted into two-dimensional time–frequency images using Short-Time Fourier Transform (STFT), Continuous Wavelet Transform (CWT), Hilbert–Huang Transform (HHT), Wigner–Ville Distribution (WVD), or Constant-Q Transform (CQT). Images from 19 EEG channels were fused into a structured grid and classified using pretrained convolutional neural networks, including MobileNetV2, ResNet-50, and InceptionV3. Results: Results indicate that classification performance is highly dependent on the chosen TFR. The STFT-based representation combined with InceptionV3 achieved the highest accuracy, reaching 98.8% with random splitting and 84.3% with subject-wise splitting, outperforming previous studies. CQT also showed competitive performance, whereas HHT and WVD were less effective. Gradient-weighted class activation mapping provided interpretable visualization of physiologically relevant EEG channel contributions. Conclusions: The proposed framework demonstrates the importance of structured multi-channel fusion and systematic TFR evaluation for robust and interpretable EEG-based dementia classification and serves as a foundation for future cross-dataset validation. Full article
19 pages, 2282 KB  
Article
Lung Disease Diagnosis Using Radial STFT and a Lightweight Convolutional Neural Network
by Uriel Calderon-Uribe, Rocio A. Lizarraga-Morales and Igor V. Guryev
Electronics 2026, 15(5), 983; https://doi.org/10.3390/electronics15050983 (registering DOI) - 27 Feb 2026
Viewed by 121
Abstract
Lung diseases are among the leading causes of death worldwide. Nowadays, to detect lung diseases, a specialist uses auscultation to make a diagnosis. Newer auscultation devices based on stethoscopes allow these sounds to be recorded for later analysis. However, the diagnosis process is [...] Read more.
Lung diseases are among the leading causes of death worldwide. Nowadays, to detect lung diseases, a specialist uses auscultation to make a diagnosis. Newer auscultation devices based on stethoscopes allow these sounds to be recorded for later analysis. However, the diagnosis process is time-consuming and relies on medical expertise to generate an accurate diagnosis. For these reasons, automated and objective diagnostic systems are crucial for the early detection of lung diseases and preventing them from worsening. In this study, a computer-aided diagnostic system that integrates the Radial Short-Time Fourier Transform (RSTFT) with a Convolutional Neural Network (CNN) enhanced by attention mechanisms is presented. The RSTFT is employed to convert lung sound recordings from a public dataset into angular frequency representations, which are used as input to the CNN. The network automatically extracts discriminative features and classifies the recordings into five categories: Chronic Obstructive Pulmonary Disease (COPD), bronchiectasis, pneumonia, asthma, and healthy lungs. Experimental results demonstrate that the proposed method outperforms several state-of-the-art approaches in terms of accuracy, precision, recall, and F1-score. These findings indicate that the proposed RSTFT–CNN framework provides an effective and reliable solution for the automated diagnosis of lung diseases, offering valuable support for clinical decision-making and early intervention. Full article
(This article belongs to the Special Issue Artificial Intelligence and Deep Learning Techniques for Healthcare)
Show Figures

Figure 1

18 pages, 973 KB  
Article
How Far Can a U-Net Go? An Empirical Analysis of Music Source Separation Performance
by Daniel Kostrzewa, Mikolaj Kondziolka, Robert Brzeski, Jeremiah Abimbola and Pawel Benecki
Appl. Sci. 2026, 16(5), 2195; https://doi.org/10.3390/app16052195 - 25 Feb 2026
Viewed by 168
Abstract
Music source separation (MSS) focuses on decomposing a mixed audio signal into individual instrumental components and is increasingly relevant for music production, restoration, remixing, education, and music information retrieval. Deep learning methods, particularly U-Net architectures operating on time–frequency representations, have recently advanced the [...] Read more.
Music source separation (MSS) focuses on decomposing a mixed audio signal into individual instrumental components and is increasingly relevant for music production, restoration, remixing, education, and music information retrieval. Deep learning methods, particularly U-Net architectures operating on time–frequency representations, have recently advanced the state of the art beyond traditional signal-processing techniques. This work presents an optimized multi-source U-Net model for separating selected musical instruments from stereo mixtures. The system uses magnitude spectrograms generated by the short-time Fourier transform and is trained and evaluated on the MUSDB18 dataset. We systematically examine architectural and training-related factors, including normalization strategies, dropout placement, optimizer selection, loss weighting, data augmentation, and spectrogram-domain modifications. Separation quality is measured using BSS Eval metrics, assessing artifacts, interference, and distortion. Experimental results show that the proposed configuration achieves competitive performance relative to established convolutional and U-Net-based open-source systems, especially in terms of vocal track separation, offering practical insights into designing efficient models for multi-instrument separation. Full article
(This article belongs to the Special Issue Advances in Audio Signal Processing)
Show Figures

Figure 1

18 pages, 999 KB  
Article
Image-Based Fault Detection and Severity Classification of Broken Rotor Bars in Induction Motors Using EfficientNetB3
by Shahil Kumar, Meshach Kumar and Rahul Ranjeev Kumar
Energies 2026, 19(4), 1110; https://doi.org/10.3390/en19041110 - 23 Feb 2026
Viewed by 224
Abstract
Broken rotor bar faults (BRBFs) in induction motors (IMs) present significant challenges in industrial applications, particularly due to the need for large labeled datasets and fast processing. This study addresses these issues by leveraging transfer learning with classical diagnostic techniques, using experimental 3-phase [...] Read more.
Broken rotor bar faults (BRBFs) in induction motors (IMs) present significant challenges in industrial applications, particularly due to the need for large labeled datasets and fast processing. This study addresses these issues by leveraging transfer learning with classical diagnostic techniques, using experimental 3-phase current and 3-axes vibration signals. The Gramian Angular Field (GAF) technique has been utilized to transform time series data into 2D images, enabling fine-tuning of an EfficientNetB3 model, which achieved 99.83% accuracy in classifying five BRBF severity levels. The proposed strategy also outperforms the state-of-the-art methods using the same experimental data. Similarly, validation with features extracted using Continuous Wavelet Transform (CWT) and Short-Time Fourier Transform (STFT) further confirmed its reliability and superiority. This study also offers enhanced interpretability through Grad-CAM visualizations of the best model, which highlights the critical regions contributing to fault classification. These visualizations enable deeper and simpler understanding of fault mechanisms and support subsequent risk analysis, making the developed model actionable and user-friendly for industrial applications. Full article
Show Figures

Figure 1

36 pages, 14856 KB  
Article
Multi-Source Fusion CNN-RF Framework for Intelligent Fault Diagnosis of Head Sheave Devices in Mining Hoists
by Chi Ma, Jian Fei, Zhiyuan Shi, Md Abdur Rob, Md Ashraful Islam and Md Habibullah
Machines 2026, 14(2), 244; https://doi.org/10.3390/machines14020244 - 21 Feb 2026
Viewed by 202
Abstract
Accurate fault diagnosis of mining hoisting head sheave systems is critical for ensuring operational safety in harsh underground environments. This study proposes a multi-source fault diagnosis framework that fuses vibration and acoustic information using a Convolutional Neural Network and Random Forest (CNN-RF). To [...] Read more.
Accurate fault diagnosis of mining hoisting head sheave systems is critical for ensuring operational safety in harsh underground environments. This study proposes a multi-source fault diagnosis framework that fuses vibration and acoustic information using a Convolutional Neural Network and Random Forest (CNN-RF). To support mechanism understanding and validate the experimental platform, finite element and multi-body dynamics simulations (ANSYS/ADAMS) are employed for physical verification and fault signature analysis, while the CNN-RF model is trained and tested exclusively using experimentally acquired vibration and acoustic data. For feature construction, vibration signals are transformed into time–frequency representations (including STFT, CWT, and generalized S-Transform (GST)), and acoustic signals are characterized using Mel-Frequency Cepstral Coefficients (MFCCs). Experimental results demonstrate that vibration–acoustic fusion improves diagnostic performance compared with single-modality baselines; the best performance is achieved by GST+MFCC with the proposed CNN-RF classifier, reaching an accuracy of 98.96%. Future work will conduct cross-condition validation under varying speeds and loads and investigate missing-modality robustness to further assess generalization and deployment reliability. Full article
Show Figures

Figure 1

24 pages, 17028 KB  
Article
Lithology Identification via MSC-Transformer Network with Time-Frequency Feature Fusion
by Shiyi Xu, Sheng Wang, Jun Bai, Kun Lai, Jie Zhang, Qingfeng Wang and Jie Zhang
Appl. Sci. 2026, 16(4), 1949; https://doi.org/10.3390/app16041949 - 15 Feb 2026
Viewed by 266
Abstract
Real-time lithology identification during drilling faces challenges such as indistinct boundaries and difficulties in feature extraction. To address these, this study proposes the MSC-Transformer, a novel model integrating time-frequency features with a deep neural network. A series of drilling experiments were conducted using [...] Read more.
Real-time lithology identification during drilling faces challenges such as indistinct boundaries and difficulties in feature extraction. To address these, this study proposes the MSC-Transformer, a novel model integrating time-frequency features with a deep neural network. A series of drilling experiments were conducted using an intelligent drilling platform, during which triaxial vibration signals were collected from five types of rock specimens: anthracite, granite, bituminous coal, sandstone, and shale. Short-time Fourier Transform (STFT) was applied to generate multi-channel power spectral density (PSD) maps, which were then fused into a three-channel tensor to preserve directional frequency information and used as inputs to the model. The proposed MSC-Transformer combines a multi-scale convolutional (MSC) module with a lightweight Transformer encoder to jointly capture local texture patterns and global dependency features, thereby enabling accurate classification of complex lithologies. Experimental results demonstrate that the model achieves an average accuracy of 98.21 ± 0.49% on the test set, outperforming convolutional neural networks (CNNs), visual geometry group (VGG), residual network (ResNet), and bidirectional long short-term memory (Bi-LSTM) by 5.93 ± 0.90%, 2.54 ± 1.11%, 6.38 ± 2.63%, and 10.56 ± 3.11%, respectively, with statistically significant improvements (p < 0.05). Ablation studies and visualization analyses further validate the effectiveness and interpretability of the model architecture. These findings indicate that lithology recognition based on time-frequency representations of vibration signals is both stable and generalizable, offering technical support for real-time intelligent lithology identification during drilling operations. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

29 pages, 8492 KB  
Article
Dual-Stream Hybrid Attention Network for Robust Intelligent Spectrum Sensing
by Bixue Song, Yongxin Feng, Fan Zhou and Peiying Zhang
Computers 2026, 15(2), 120; https://doi.org/10.3390/computers15020120 - 11 Feb 2026
Viewed by 187
Abstract
UAV communication, leveraging high mobility and flexible deployment, is gradually becoming an important component of 6G integrated air–ground networks. With the expansion of aerial services, air–ground spectrum resources are increasingly scarce, and spectrum sharing and opportunistic access have become key technologies for improving [...] Read more.
UAV communication, leveraging high mobility and flexible deployment, is gradually becoming an important component of 6G integrated air–ground networks. With the expansion of aerial services, air–ground spectrum resources are increasingly scarce, and spectrum sharing and opportunistic access have become key technologies for improving spectrum utilization. Spectrum sensing is the prerequisite for UAVs to perform dynamic access and avoid causing interference to primary users. However, in air–ground links, the channel time variability caused by Doppler effects, carrier frequency offset, and Rician fading can weaken feature separability, making it difficult for deep learning-based spectrum sensing methods to maintain reliable detection in complex environments. In this paper, a dual-stream hybrid-attention spectrum sensing method (DSHA) is proposed, which represents the received signal simultaneously as a time-domain I/Q sequence and an STFT time-frequency map to extract complementary features and employs a hybrid attention mechanism to model key intra-branch dependencies and achieve inter-branch interaction and fusion. Furthermore, a noise-consistent paired training strategy is introduced to mitigate the bias induced by noise randomness, thereby enhancing weak-signal discrimination capability. Simulation results show that under different false-alarm constraints, the proposed method achieves higher detection probability in low-SNR scenarios as well as under fading and CFO perturbations. In addition, compared with multiple typical baselines, DSHA exhibits better robustness and generalization; under Rician channels, its detection probability is improved by about 28.6% over the best baseline. Full article
(This article belongs to the Special Issue Wireless Sensor Networks in IoT)
Show Figures

Figure 1

9 pages, 1865 KB  
Proceeding Paper
Detection of Respiratory Diseases Based on Poultry Vocalizations Using Deep Learning
by Farook Sattar
Biol. Life Sci. Forum 2025, 54(1), 18; https://doi.org/10.3390/blsf2025054018 - 9 Feb 2026
Viewed by 173
Abstract
In this study, we design a deep learning-based intelligent recognition method capable of accurately distinguishing abnormal chicken vocalizations among complex sound signals. Our proposed framework is based on the wavelet scattering transform (WST) and a Long Short-Term Memory (LSTM) network, and uses preprocessed [...] Read more.
In this study, we design a deep learning-based intelligent recognition method capable of accurately distinguishing abnormal chicken vocalizations among complex sound signals. Our proposed framework is based on the wavelet scattering transform (WST) and a Long Short-Term Memory (LSTM) network, and uses preprocessed chicken vocalizations processed through a denoising scheme, adopting an audio image generation model (AIGM) based on rectified STFT (Short-Term Fourier Transform). We have used a public chicken language dataset that consists of a total of segments for each of the two categories (Healthy or Sick), totaling 4000 five-second audio clips from actual farming environments, which are labeled by veterinary experts. The proposed method achieves promising performance, outperforming state-of-the-art methods for detecting poultry respiratory diseases and enabling poultry personnel to accurately assess the health and well-being of the chickens. Full article
(This article belongs to the Proceedings of The 3rd International Online Conference on Agriculture)
Show Figures

Figure 1

19 pages, 3671 KB  
Article
Detecting Rail Surface Contaminants Using a Combined Short-Time Fourier Transform and Convolutional Neural Network Approach
by Gerardo Hurtado-Hurtado, Tania Elizabeth Sandoval-Valencia, Luis Morales-Velázquez and Juan Carlos Jáuregui-Correa
Modelling 2026, 7(1), 35; https://doi.org/10.3390/modelling7010035 - 9 Feb 2026
Viewed by 259
Abstract
Condition monitoring of railway track surfaces is crucial for ensuring the safety, operational efficiency, and effective maintenance of railway systems. This work presents a data-driven modelling and an experimental methodology for identifying and classifying contaminants on railway tracks using vibration analysis and artificial [...] Read more.
Condition monitoring of railway track surfaces is crucial for ensuring the safety, operational efficiency, and effective maintenance of railway systems. This work presents a data-driven modelling and an experimental methodology for identifying and classifying contaminants on railway tracks using vibration analysis and artificial intelligence techniques. In this study, the railway dynamics were physically simulated using a 1:20 scaled test rig, where the rails were treated with various contaminants (oil, water, and sand), and the resulting vehicle vibrations were recorded by on-board accelerometers and gyroscopes. To construct the predictive model, a hybrid architecture was designed integrating Short-Time Fourier Transform (STFT) for time-frequency feature extraction and a multi-channel Convolutional Neural Network (CNN) for pattern recognition. Initial results indicate that accelerometer data, particularly from longitudinal and lateral vibrations, are more effective than gyroscope data for classifying certain contaminants. To enhance classification robustness, this work introduces a multi-channel CNN that simultaneously processes the most informative signals, leading to a significant improvement in detection accuracy across all tested contaminants. This study validates the effectiveness of the proposed methodology as a robust and reliable solution for contaminant detection, while also confirming the utility of the scaled testbed as a valuable platform for future research in railway dynamics. Full article
(This article belongs to the Section Modelling in Artificial Intelligence)
Show Figures

Figure 1

16 pages, 6191 KB  
Article
A Hybrid Millimeter-Wave Radar–Ultrasonic Fusion System for Robust Human Activity Recognition with Attention-Enhanced Deep Learning
by Liping Yao, Kwok L. Chung, Luxin Tang, Tao Ye, Shiquan Wang, Pingchuan Xu, Yuhao Bi and Yaowen Wu
Sensors 2026, 26(3), 1057; https://doi.org/10.3390/s26031057 - 6 Feb 2026
Viewed by 355
Abstract
To address the tradeoff between environmental robustness and fine-grained accuracy in single-sensor human behavior recognition, this paper proposes a non-contact system fusing 77 GHz SIFT (mmWave) radar and a 40 kHz ultrasonic array. The system leverages radar’s long-range penetration and low-visibility adaptability, paired [...] Read more.
To address the tradeoff between environmental robustness and fine-grained accuracy in single-sensor human behavior recognition, this paper proposes a non-contact system fusing 77 GHz SIFT (mmWave) radar and a 40 kHz ultrasonic array. The system leverages radar’s long-range penetration and low-visibility adaptability, paired with ultrasound’s centimeter-level short-range precision and electromagnetic clutter immunity. A synchronized data acquisition platform ensures multi-modal signal consistency, while wavelet transform (for radar) and STFT (for ultrasound) extract complementary time–frequency features. The proposed Attention-CNN-BiLSTM architecture integrates local spatial feature extraction, bidirectional temporal dependency modeling, and salient cue enhancement. Experimental results on 1600 synchronized sequences (four behaviors: standing, sitting, walking, falling) show a 98.6% mean class accuracy with subject-wise generalization, outperforming single-sensor baselines and traditional deep learning models. As a privacy-preserving, lighting-agnostic solution, it offers promising applications in smart homes, healthcare monitoring, and intelligent surveillance, providing a robust technical foundation for contactless behavior recognition. Full article
(This article belongs to the Special Issue Electromagnetic Sensors and Their Applications)
Show Figures

Figure 1

24 pages, 3287 KB  
Article
Neonatal Seizure Detection Based on Spatiotemporal Feature Decoupling and Domain-Adversarial Learning
by Tiannuo Xu and Wei Zheng
Sensors 2026, 26(3), 938; https://doi.org/10.3390/s26030938 - 1 Feb 2026
Viewed by 318
Abstract
Neonatal seizures are a critical early indicator of neurological injury, yet effective automated detection is challenged by significant inter-subject variability in electroencephalogram (EEG) signals. To address this generalization gap, this study introduces the Domain-Adversarial Spatiotemporal Network (DA-STNet) for robust cross-subject seizure detection. Utilizing [...] Read more.
Neonatal seizures are a critical early indicator of neurological injury, yet effective automated detection is challenged by significant inter-subject variability in electroencephalogram (EEG) signals. To address this generalization gap, this study introduces the Domain-Adversarial Spatiotemporal Network (DA-STNet) for robust cross-subject seizure detection. Utilizing Short-Time Fourier Transform (STFT) spectrograms, the architecture employs a hierarchical backbone comprising a Channel-Independent CNN (CI-CNN) for local texture extraction, a Spatial Bidirectional Long Short-Term Memory (Bi-LSTM) for modeling topological dependencies, and Attention Pooling to dynamically prioritize pathological channels while suppressing noise. Crucially, a Gradient Reversal Layer (GRL) is integrated to enforce domain-adversarial training, decoupling pathological features from subject-specific identity to ensure domain invariance. Under rigorous 5-fold cross-validation, the model achieves State-of-the-Art performance with an average Area Under the Curve (AUC) of 0.9998 and an F1-score of 0.9952. Data scaling experiments further reveal that optimal generalization is attainable using only 80% of source data, highlighting the model’s superior data efficiency. These findings demonstrate the proposed method’s capability to reduce reliance on extensive clinical annotations while maintaining high diagnostic precision in complex clinical scenarios. Full article
(This article belongs to the Special Issue AI and Big Data Analytics for Medical E-Diagnosis)
Show Figures

Figure 1

15 pages, 3959 KB  
Technical Note
Airborne SAR Imaging Algorithm for Ocean Waves Oriented to Sea Spike Suppression
by Yawei Zhao, Yongsheng Xu, Yanlei Du and Jinsong Chong
Remote Sens. 2026, 18(3), 397; https://doi.org/10.3390/rs18030397 - 24 Jan 2026
Viewed by 353
Abstract
Synthetic aperture radar (SAR) is widely used in the field of ocean remote sensing. However, SAR images are usually affected by sea spikes, which appear as strong echo and azimuth defocus characteristics. The texture features of ocean waves in SAR images are submerged [...] Read more.
Synthetic aperture radar (SAR) is widely used in the field of ocean remote sensing. However, SAR images are usually affected by sea spikes, which appear as strong echo and azimuth defocus characteristics. The texture features of ocean waves in SAR images are submerged by sea spikes, making them weak or even invisible. This seriously affects the further applications of SAR technology in ocean remote sensing. To address this issue, an airborne SAR imaging algorithm for ocean waves oriented to sea spike suppression is proposed in this paper. The non-stationary characteristics of sea spikes are taken into account in the proposed algorithm. The SAR echo data is transformed into the time–frequency domain by short-time Fourier transform (STFT). And the echo signals of sea spikes are suppressed in the time–frequency domain. Then, the ocean waves are imaged in focus by applying focus settings. In order to verify the effectiveness of the proposed algorithm, airborne SAR data was processed using the proposed algorithm, including SAR data with completely invisible waves and other data with weakly visible waves under sea spike influence. Through analyzing the ocean wave spectrum and imaging quality, it is confirmed that the proposed algorithm can significantly suppress sea spikes and improve the texture features of ocean waves in SAR images. Full article
(This article belongs to the Special Issue Microwave Remote Sensing on Ocean Observation)
Show Figures

Figure 1

27 pages, 4802 KB  
Article
Fine-Grained Radar Hand Gesture Recognition Method Based on Variable-Channel DRSN
by Penghui Chen, Siben Li, Chenchen Yuan, Yujing Bai and Jun Wang
Electronics 2026, 15(2), 437; https://doi.org/10.3390/electronics15020437 - 19 Jan 2026
Viewed by 317
Abstract
With the ongoing miniaturization of smart devices, fine-grained hand gesture recognition using millimeter-wave radar has attracted increasing attention, yet practical deployment remains challenging in continuous-gesture segmentation, robust feature extraction, and reliable classification. This paper presents an end-to-end fine-grained gesture recognition framework based on [...] Read more.
With the ongoing miniaturization of smart devices, fine-grained hand gesture recognition using millimeter-wave radar has attracted increasing attention, yet practical deployment remains challenging in continuous-gesture segmentation, robust feature extraction, and reliable classification. This paper presents an end-to-end fine-grained gesture recognition framework based on frequency modulated continuous wave(FMCW) millimeter-wave radar, including gesture design, data acquisition, feature construction, and neural network-based classification. Ten gesture types are recorded (eight valid gestures and two return-to-neutral gestures); for classification, the two return-to-neutral gesture types are merged into a single invalid class, yielding a nine-class task. A sliding-window segmentation method is developed using short-time Fourier transformation(STFT)-based Doppler-time representations, and a dataset of 4050 labeled samples is collected. Multiple signal classification(MUSIC)-based super-resolution estimation is adopted to construct range–time and angle–time representations, and instance-wise normalization is applied to Doppler and range features to mitigate inter-individual variability without test leakage. For recognition, a variable-channel deep residual shrinkage network (DRSN) is employed to improve robustness to noise, supporting single-, dual-, and triple-channel feature inputs. Results under both subject-dependent evaluation with repeated random splits and subject-independent leave one subject out(LOSO) cross-validation show that DRSN architecture consistently outperforms the RefineNet-based baseline, and the triple-channel configuration achieves the best performance (98.88% accuracy). Overall, the variable-channel design enables flexible feature selection to meet diverse application requirements. Full article
Show Figures

Figure 1

23 pages, 4752 KB  
Article
Modulation-Based Feature Extraction for Robust Sleep Stage Classification Across Apnea-Based Cohorts
by Unaza Tallal, Rupesh Agrawal and Shruti Kshirsagar
Biosensors 2026, 16(1), 56; https://doi.org/10.3390/bios16010056 - 13 Jan 2026
Viewed by 602
Abstract
Automated sleep staging remains challenging due to the transitional nature of certain sleep stages, particularly N1. In this paper, we explore modulation spectrograms for automatic sleep staging to capture the transitional nature of sleep stages and compare them with conventional benchmark features, such [...] Read more.
Automated sleep staging remains challenging due to the transitional nature of certain sleep stages, particularly N1. In this paper, we explore modulation spectrograms for automatic sleep staging to capture the transitional nature of sleep stages and compare them with conventional benchmark features, such as the Short-Time Fourier Transform (STFT) and the Continuous Wavelet Transform (CWT). We utilized a single-channel EEG (C4–M1) from the DREAMT dataset with subject-independent validation. We stratify participants by the Apnea–Hypopnea Index (AHI) into Normal, Mild, Moderate, and Severe groups to assess clinical generalizability. Our modulation-based framework significantly outperforms STFT and CWT in the Mild and Severe cohorts, while maintaining comparable high performance in the Normal and Moderate AHI groups. Notably, the proposed framework maintained robust performance in severe apnea cohorts, effectively mitigating the degradation observed in standard time–frequency baselines. These findings demonstrate the effectiveness of modulation spectrograms for sleep staging while emphasizing the importance of medical stratification for reliable outcomes in clinical populations. Full article
(This article belongs to the Section Biosensors and Healthcare)
Show Figures

Figure 1

20 pages, 2906 KB  
Article
Research on Oil and Gas Pipeline Leakage Detection Based on MSCNN-Transformer
by Yingtao Zhang, Wenhe Li, Yang Wu and Huili Wei
Appl. Sci. 2026, 16(1), 480; https://doi.org/10.3390/app16010480 - 2 Jan 2026
Viewed by 474
Abstract
The leakage detection of oil and gas is very important for the safe operation of pipelines. The existing working condition recognition methods have limitations in processing and capturing complex multi-category leakage signal characteristics. In order to improve the accuracy of oil and gas [...] Read more.
The leakage detection of oil and gas is very important for the safe operation of pipelines. The existing working condition recognition methods have limitations in processing and capturing complex multi-category leakage signal characteristics. In order to improve the accuracy of oil and gas pipeline leakage detection, a multi-scale convolutional neural network-Transformer (MSCNN-Transformer)-based oil and gas pipeline leakage condition recognition method is proposed. Firstly, in order to capture the global information and nonlinear characteristics of the time series signal, STFT is used to generate the time-frequency image. Furthermore, in order to enrich the feature information from different dimensions, the one-dimensional signal and the two-dimensional time-frequency image are sampled by multi-scale convolution, and the global relationship is established by combining the multi-head attention mechanism of the Transformer module. Finally, the leakage signal is accurately identified by fusing features and classifiers. The experimental results show that the proposed method shows high performance on the GPLA-12 data set, and the recognition accuracy is 96.02%. Compared with other leakage signal recognition methods, the proposed method has obvious advantages. Full article
(This article belongs to the Section Energy Science and Technology)
Show Figures

Figure 1

Back to TopTop