Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,536)

Search Parameters:
Keywords = convolutional-recurrent neural network

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
23 pages, 2976 KB  
Article
Transfer Learning-Based Piezoelectric Actuators Feedforward Control with GRU-CNN
by Yaqian Hu, Herong Jin, Xiangcheng Chu and Yali Yi
Appl. Sci. 2026, 16(3), 1305; https://doi.org/10.3390/app16031305 - 27 Jan 2026
Abstract
To compensate for hysteresis, low damping vibration, and their coupling effects, this paper proposes a gated recurrent unit and convolutional neural network (GRU-CNN) model as a feedforward control model that maps desired displacement trajectories to driving voltages. The GRU-CNN integrates a gated recurrent [...] Read more.
To compensate for hysteresis, low damping vibration, and their coupling effects, this paper proposes a gated recurrent unit and convolutional neural network (GRU-CNN) model as a feedforward control model that maps desired displacement trajectories to driving voltages. The GRU-CNN integrates a gated recurrent unit (GRU) layer to capture long-term temporal dependencies, a multi-layer convolutional neural network (CNN) to extract local data features, and residual connections to mitigate information distortion. The GRU-CNN is then combined with transfer learning (TL) for feedforward control of cross-batch and cross-type piezoelectric actuators (PEAs), so as to reduce reliance on training datasets. The analysis focuses on the impacts of target PEA data volume and source-target similarity on transfer learning strategies. The GRU-CNN trained on PEA #1 achieves high control accuracy, with a mean absolute error (MAE) of 0.077, a root mean square error (RMSE) of 0.129, and a coefficient of determination (R2) of 0.997. When transferred to cross-batch PEA #2 and cross-type PEA #3, the GRU-CNN feedforward controller still delivers favorable performance; R2 values all exceed 0.98, representing at least a 27% improvement compared to training from scratch. These results indicate that the proposed transfer learning-based feedforward control method can effectively reduce retraining effort, suggesting its potential applicability to batch production scenarios. Full article
30 pages, 2761 KB  
Article
HST–MB–CREH: A Hybrid Spatio-Temporal Transformer with Multi-Branch CNN/RNN for Rare-Event-Aware PV Power Forecasting
by Guldana Taganova, Jamalbek Tussupov, Assel Abdildayeva, Mira Kaldarova, Alfiya Kazi, Ronald Cowie Simpson, Alma Zakirova and Bakhyt Nurbekov
Algorithms 2026, 19(2), 94; https://doi.org/10.3390/a19020094 - 23 Jan 2026
Viewed by 71
Abstract
We propose the Hybrid Spatio-Temporal Transformer with Multi-Branch CNN/RNN and Extreme-Event Head (HST–MB–CREH), a hybrid spatio-temporal deep learning architecture for joint short-term photovoltaic (PV) power forecasting and the detection of rare extreme events, to support the reliable operation of renewable-rich power systems. The [...] Read more.
We propose the Hybrid Spatio-Temporal Transformer with Multi-Branch CNN/RNN and Extreme-Event Head (HST–MB–CREH), a hybrid spatio-temporal deep learning architecture for joint short-term photovoltaic (PV) power forecasting and the detection of rare extreme events, to support the reliable operation of renewable-rich power systems. The model combines a spatio-temporal transformer encoder with three convolutional neural network (CNN)/recurrent neural network (RNN) branches (CNN → long short-term memory (LSTM), LSTM → gated recurrent unit (GRU), CNN → GRU) and a dense pathway for tabular meteorological and calendar features. A multitask output head simultaneously performs the regression of PV power and binary classification of extremes defined above the 95th percentile. We evaluate HST–MB–CREH on the publicly available Renewable Power Generation and Weather Conditions dataset with hourly resolutions from 2017 to 2022, using a 5-fold TimeSeriesSplit protocol to avoid temporal leakage and to cover multiple seasons. Compared with tree ensembles (RandomForest, XGBoost), recurrent baselines (Stacked GRU, LSTM), and advanced hybrid/transformer models (Hybrid Multi-Branch CNN–LSTM/GRU with Dense Path and Extreme-Event Head (HMB–CLED) and Spatio-Temporal Multitask Transformer with Extreme-Event Head (STM–EEH)), the proposed architecture achieves the best overall trade-off between accuracy and rare-event sensitivity, with normalized performance of RMSE_z = 0.2159 ± 0.0167, MAE_z = 0.1100 ± 0.0085, mean absolute percentage error (MAPE) = 9.17 ± 0.45%, R2 = 0.9534 ± 0.0072, and AUC_ext = 0.9851 ± 0.0051 across folds. Knowledge extraction is supported via attention-based analysis and permutation feature importance, which highlight the dominant role of global horizontal irradiance, diurnal harmonics, and solar geometry features. The results indicate that hybrid spatio-temporal multitask architectures can substantially improve both the forecast accuracy and robustness to extremes, making HST–MB–CREH a promising building block for intelligent decision-support tools in smart grids with a high share of PV generation. Full article
(This article belongs to the Section Evolutionary Algorithms and Machine Learning)
52 pages, 3528 KB  
Review
Advanced Fault Detection and Diagnosis Exploiting Machine Learning and Artificial Intelligence for Engineering Applications
by Davide Paolini, Pierpaolo Dini, Abdussalam Elhanashi and Sergio Saponara
Electronics 2026, 15(2), 476; https://doi.org/10.3390/electronics15020476 - 22 Jan 2026
Viewed by 128
Abstract
Modern engineering systems require reliable and timely Fault Detection and Diagnosis (FDD) to ensure operational safety and resilience. Traditional model-based and rule-based approaches, although interpretable, exhibit limited scalability and adaptability in complex, data-intensive environments. This survey provides a systematic overview of recent studies [...] Read more.
Modern engineering systems require reliable and timely Fault Detection and Diagnosis (FDD) to ensure operational safety and resilience. Traditional model-based and rule-based approaches, although interpretable, exhibit limited scalability and adaptability in complex, data-intensive environments. This survey provides a systematic overview of recent studies exploring Machine Learning (ML) and Artificial Intelligence (AI) techniques for FDD across industrial, energy, Cyber-Physical Systems (CPS)/Internet of Things (IoT), and cybersecurity domains. Deep architectures such as Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), Transformers, and Graph Neural Networks (GNNs) are compared with unsupervised, hybrid, and physics-informed frameworks, emphasizing their respective strengths in adaptability, robustness, and interpretability. Quantitative synthesis and radar-based assessments suggest that AI-driven FDD approaches offer increased adaptability, scalability, and early fault detection capabilities compared to classical methods, while also introducing new challenges related to interpretability, robustness, and deployment. Emerging research directions include the development of foundation and multimodal models, federated learning (FL), and privacy-preserving learning, as well as physics-guided trustworthy AI. These trends indicate a paradigm shift toward self-adaptive, interpretable, and collaborative FDD systems capable of sustaining reliability, transparency, and autonomy across critical infrastructures. Full article
Show Figures

Figure 1

22 pages, 5103 KB  
Article
On the Development of an AI-Based Tool to Assess the Instantaneous Modal Properties of Nonlinear SDOF Systems
by Alvaro Iglesias-Pordomingo, Guillermo Fernandez, Alvaro Magdaleno and Antolin Lorenzana
Appl. Sci. 2026, 16(2), 1070; https://doi.org/10.3390/app16021070 - 20 Jan 2026
Viewed by 156
Abstract
In this article, a data-driven algorithm is developed to assess the natural frequency and damping ratio of a nonlinear oscillating single-degree-of-freedom (SDOF) system. The algorithm is based on hybrid convolutional–long short-term memory neural networks (CNN-LSTM) that process a short moving window belonging to [...] Read more.
In this article, a data-driven algorithm is developed to assess the natural frequency and damping ratio of a nonlinear oscillating single-degree-of-freedom (SDOF) system. The algorithm is based on hybrid convolutional–long short-term memory neural networks (CNN-LSTM) that process a short moving window belonging to a free-decay response and provide estimates of both parameters over time. The novelty of the study resides in the fact that the neural network is trained exclusively using synthetic data issued from linear SDOF models. Since the recurrent neural network (RNN) requires relatively small amounts of data to operate effectively, the nonlinear system locally behaves as a quasi-linear model, allowing each data segment to be processed under this assumption. The proposed RecuID tool is experimentally validated on a laboratory-scale nonlinear SDOF system. To demonstrate its effectiveness, it is compared to conventional identification algorithms. The experimental study yields a maximum mean absolute error (MAE) of 0.244 Hz for the natural frequency and 0.015 for the damping ratio. RecuID proves to be a faster and more robust methodology, capable of handling time-varying damping ratios up to 0.2 and a wide range of natural frequencies defined relative to the sampling rate. Full article
(This article belongs to the Special Issue Nonlinear Dynamics and Vibration)
Show Figures

Figure 1

34 pages, 7567 KB  
Article
Enhancing Demand Forecasting Using the Formicary Zebra Optimization with Distributed Attention Guided Deep Learning Model
by Ikhalas Fandi and Wagdi Khalifa
Appl. Sci. 2026, 16(2), 1039; https://doi.org/10.3390/app16021039 - 20 Jan 2026
Viewed by 103
Abstract
In the modern era, demand forecasting enhances the decision-making tasks of industries for controlling production planning and reducing inventory costs. However, the dynamic nature of the fashion and apparel retail industry necessitates precise demand forecasting to optimize supply chain operations and meet customer [...] Read more.
In the modern era, demand forecasting enhances the decision-making tasks of industries for controlling production planning and reducing inventory costs. However, the dynamic nature of the fashion and apparel retail industry necessitates precise demand forecasting to optimize supply chain operations and meet customer expectations. Consequently, this research proposes the Formicary Zebra Optimization-Based Distributed Attention-Guided Convolutional Recurrent Neural Network (FZ-DACR) model for improving the demand forecasting. In the proposed approach, the combination of the Formicary Zebra Optimization and Distributed Attention mechanism enabled deep learning architectures to assist in capturing the complex patterns of the retail sales data. Specifically, the neural networks, including convolutional neural networks (CNNs) and recurrent neural networks (RNNs), facilitate extracting the local features and temporal dependencies to analyze the volatile demand patterns. Furthermore, the proposed model integrates visual and textual data to enhance forecasting accuracy. By leveraging the adaptive optimization capabilities of the Formicary Zebra Algorithm, the proposed model effectively extracts features from product images and historical sales data while addressing the complexities of volatile demand patterns. Based on extensive experimental analysis of the proposed model using diverse datasets, the FZ-DACR model achieves superior performance, with minimum error values including MAE of 1.34, MSE of 4.7, RMS of 2.17, and R2 of 93.3% using the DRESS dataset. Moreover, the findings highlight the ability of the proposed model in managing the fluctuating trends and supporting inventory and pricing strategies effectively. This innovative approach has significant implications for retailers, enabling more agile supply chains and improved decision making in a highly competitive market. Full article
(This article belongs to the Special Issue Advanced Methods for Time Series Forecasting)
Show Figures

Figure 1

17 pages, 5869 KB  
Article
Research on Tool Wear Prediction Method Based on CNN-ResNet-CBAM-BiGRU
by Bo Sun, Hao Wang, Jian Zhang, Lixin Zhang and Xiangqin Wu
Sensors 2026, 26(2), 661; https://doi.org/10.3390/s26020661 - 19 Jan 2026
Viewed by 177
Abstract
Aiming to address insufficient feature extraction, vanishing gradients, and low prediction accuracy in tool wear prediction, this paper proposes a hybrid deep neural network based on a Convolutional Neural Network (CNN), Residual Network (ResNet) residual connections, the Convolutional Block Attention Module (CBAM), and [...] Read more.
Aiming to address insufficient feature extraction, vanishing gradients, and low prediction accuracy in tool wear prediction, this paper proposes a hybrid deep neural network based on a Convolutional Neural Network (CNN), Residual Network (ResNet) residual connections, the Convolutional Block Attention Module (CBAM), and a Bidirectional Gated Recurrent Unit (BiGRU). First, a 34-dimensional multi-domain feature set covering the time domain, frequency domain, and time–frequency domain is constructed, and multi-sensor signals are standardized using z-score normalization. A CNN–BiGRU backbone is then established, where ResNet-style residual connections are introduced to alleviate training degradation and mitigate vanishing-gradient issues in deep networks. Meanwhile, CBAM is integrated into the feature extraction module to adaptively reweight informative features in both channel and spatial dimensions. In addition, a BiGRU layer is embedded for temporal modeling to capture bidirectional dependencies throughout the wear evolution process. Finally, a fully connected layer is used as a regressor to map high-dimensional representations to tool wear values. Experiments on the PHM2010 dataset demonstrate that the proposed hybrid architecture is more stable and achieves better predictive performance than several mainstream deep learning baselines. Systematic ablation studies further quantify the contribution of each component: compared with the baseline CNN model, the mean absolute error (MAE) is reduced by 47.5%, the root mean square error (RMSE) is reduced by 68.5%, and the coefficient of determination (R2) increases by 14.5%, enabling accurate tool wear prediction. Full article
(This article belongs to the Section Sensor Networks)
Show Figures

Figure 1

26 pages, 544 KB  
Article
Physics-Aware Deep Learning Framework for Solar Irradiance Forecasting Using Fourier-Based Signal Decomposition
by Murad A. Yaghi and Huthaifa Al-Omari
Algorithms 2026, 19(1), 81; https://doi.org/10.3390/a19010081 - 17 Jan 2026
Viewed by 121
Abstract
Photovoltaic Systems have been a long-standing challenge to integrate with electrical Power Grids due to the randomness of solar irradiance. Deep Learning (DL) has potential to forecast solar irradiance; however, black-box DL models typically do not offer interpretation, nor can they easily distinguish [...] Read more.
Photovoltaic Systems have been a long-standing challenge to integrate with electrical Power Grids due to the randomness of solar irradiance. Deep Learning (DL) has potential to forecast solar irradiance; however, black-box DL models typically do not offer interpretation, nor can they easily distinguish between deterministic astronomical cycles, and random meteorological variability. The objective of this study was to develop and apply a new Physics-Aware Deep Learning Framework that identifies and utilizes physical attributes of solar irradiance via Fourier-based signal decomposition. The proposed method decomposes the time-series into polynomial trend, Fourier-based seasonal component and stochastic residual, each of which are processed within different neural network paths. A wide variety of architectures were tested (Recurrent Neural Network (RNN), Long Short-Term Memory (LSTM), Gated Recurrent Unit (GRU), Convolutional Neural Network (CNN)), at multiple historical window sizes and forecast horizons on a diverse dataset from a three-year span. All of the architectures tested demonstrated improved accuracy and robustness when using the physics aware decomposition as opposed to all other methods. Of the architectures tested, the GRU architecture was the most accurate and performed well in terms of overall evaluation. The GRU model had an RMSE of 78.63 W/m2 and an R2 value of 0.9281 for 15 min ahead forecasting. Additionally, the Fourier-based methodology was able to reduce the maximum absolute error by approximately 15% to 20%, depending upon the architecture used, and therefore it provided a way to reduce the impact of the larger errors in forecasting during periods of unstable weather. Overall, this framework represents a viable option for both physically interpretive and computationally efficient real-time solar forecasting that provides a bridge between Physical Modeling and Data-Driven Intelligence. Full article
(This article belongs to the Special Issue Artificial Intelligence in Sustainable Development)
Show Figures

Figure 1

22 pages, 5927 KB  
Article
Research on a Temperature and Humidity Prediction Model for Greenhouse Tomato Based on iT-LSTM-CA
by Yanan Gao, Pingzeng Liu, Yuxuan Zhang, Fengyu Li, Ke Zhu, Yan Zhang and Shiwei Xu
Sustainability 2026, 18(2), 930; https://doi.org/10.3390/su18020930 - 16 Jan 2026
Viewed by 153
Abstract
Constructing a temperature and humidity prediction model for greenhouse-grown tomatoes is of great significance for achieving resource-efficient and sustainable greenhouse environmental control and promoting healthy tomato growth. However, traditional models often struggle to simultaneously capture long-term temporal trends, short-term local dynamic variations, and [...] Read more.
Constructing a temperature and humidity prediction model for greenhouse-grown tomatoes is of great significance for achieving resource-efficient and sustainable greenhouse environmental control and promoting healthy tomato growth. However, traditional models often struggle to simultaneously capture long-term temporal trends, short-term local dynamic variations, and the coupling relationships among multiple variables. To address these issues, this study develops an iT-LSTM-CA multi-step prediction model, in which the inverted Transformer (iTransformer, iT) is employed to capture global dependencies across variables and long temporal scales, the Long Short-Term Memory (LSTM) network is utilized to extract short-term local variation patterns, and a cross-attention (CA) mechanism is introduced to dynamically fuse the two types of features. Experimental results show that, compared with models such as Gated Recurrent Unit (GRU), Temporal Convolutional Network (TCN), Recurrent Neural Network (RNN), LSTM, and Bidirectional Long Short-Term Memory (Bi-LSTM), the iT-LSTM-CA achieves the best performance in multi-step forecasting tasks at 3 h, 6 h, 12 h, and 24 h horizons. For temperature prediction, the R2 ranges from 0.96 to 0.98, with MAE between 0.42 °C and 0.79 °C and RMSE between 0.58 °C and 1.06 °C; for humidity prediction, the R2 ranges from 0.95 to 0.97, with MAE between 1.21% and 2.49% and RMSE between 1.78% and 3.42%. These results indicate that the iT-LSTM-CA model can effectively capture greenhouse environmental variations and provide a scientific basis for environmental control and management in tomato greenhouses. Full article
Show Figures

Figure 1

31 pages, 1742 KB  
Article
Federated Learning Frameworks for Intelligent Transportation Systems: A Comparative Adaptation Analysis
by Mario Steven Vela Romo, Carolina Tripp-Barba, Nathaly Orozco Garzón, Pablo Barbecho, Xavier Calderón Hinojosa and Luis Urquiza-Aguiar
Smart Cities 2026, 9(1), 12; https://doi.org/10.3390/smartcities9010012 - 16 Jan 2026
Viewed by 172
Abstract
Intelligent Transportation Systems (ITS) have progressively incorporated machine learning to optimize traffic efficiency, enhance safety, and improve real-time decision-making. However, the traditional centralized machine learning (ML) paradigm faces critical limitations regarding data privacy, scalability, and single-point vulnerabilities. This study explores FL as a [...] Read more.
Intelligent Transportation Systems (ITS) have progressively incorporated machine learning to optimize traffic efficiency, enhance safety, and improve real-time decision-making. However, the traditional centralized machine learning (ML) paradigm faces critical limitations regarding data privacy, scalability, and single-point vulnerabilities. This study explores FL as a decentralized alternative that preserves privacy by training local models without transferring raw data. Based on a systematic literature review encompassing 39 ITS-related studies, this work classifies applications according to their architectural detail—distinguishing systems from models—and identifies three families of federated learning (FL) frameworks: privacy-focused, integrable, and advanced infrastructure. Three representative frameworks—Federated Learning-based Gated Recurrent Unit (FedGRU), Digital Twin + Hierarchical Federated Learning (DT + HFL), and Transfer Learning with Convolutional Neural Networks (TFL-CNN)—were comparatively analyzed against a client–server baseline to assess their suitability for ITS adaptation. Our qualitative, architecture-level comparison suggests that DT + HFL and TFL-CNN, characterized by hierarchical aggregation and edge-level coordination, are conceptually better aligned with scalability and stability requirements in vehicular and traffic deployments than pure client–server baselines. FedGRU, while conceptually relevant as a meta-framework for coordinating multiple organizational models, is primarily intended as a complementary reference rather than as a standalone architecture for large-scale ITS deployment. Through application-level evaluations—including traffic prediction, accident detection, transport-mode identification, and driver profiling—this study demonstrates that FL can be effectively integrated into ITS with moderate architectural adjustments. This work does not introduce new experimental results; instead, it provides a qualitative, architecture-level comparison and adaptation guideline to support the migration of ITS applications toward federated learning. Overall, the results establish a solid methodological foundation for migrating centralized ITS architectures toward federated, privacy-preserving intelligence, in alignment with the evolution of edge and 6G infrastructures. Full article
(This article belongs to the Special Issue Big Data and AI Services for Sustainable Smart Cities)
Show Figures

Figure 1

19 pages, 1973 KB  
Article
Continuous Smartphone Authentication via Multimodal Biometrics and Optimized Ensemble Learning
by Chia-Sheng Cheng, Ko-Chien Chang, Hsing-Chung Chen and Chao-Lung Chou
Mathematics 2026, 14(2), 311; https://doi.org/10.3390/math14020311 - 15 Jan 2026
Viewed by 378
Abstract
The ubiquity of smartphones has transformed them into primary repositories of sensitive data; however, traditional one-time authentication mechanisms create a critical trust gap by failing to verify identity post-unlock. Our aim is to mitigate these vulnerabilities and align with the Zero Trust Architecture [...] Read more.
The ubiquity of smartphones has transformed them into primary repositories of sensitive data; however, traditional one-time authentication mechanisms create a critical trust gap by failing to verify identity post-unlock. Our aim is to mitigate these vulnerabilities and align with the Zero Trust Architecture (ZTA) framework and philosophy of “never trust, always verify,” as formally defined by the National Institute of Standards and Technology (NIST) in Special Publication 800-207. This study introduces a robust continuous authentication (CA) framework leveraging multimodal behavioral biometrics. A dedicated application was developed to synchronously capture touch, sliding, and inertial sensor telemetry. For feature modeling, a heterogeneous deep learning pipeline was employed to capture modality-specific characteristics, utilizing Convolutional Neural Networks (CNNs) for sensor data, Long Short-Term Memory (LSTM) networks for curvilinear sliding, and Gated Recurrent Units (GRUs) for discrete touch. To resolve performance degradation caused by class imbalance in Zero Trust environments, a Grid Search Optimization (GSO) strategy was applied to optimize a weighted voting ensemble, identifying the global optimum for decision thresholds and modality weights. Empirical validation on a dataset of 35,519 samples from 15 subjects demonstrates that the optimized ensemble achieves a peak accuracy of 99.23%. Sensor kinematics emerged as the primary biometric signature, followed by touch and sliding features. This framework enables high-precision, non-intrusive continuous verification, bridging the critical security gap in contemporary mobile architectures. Full article
Show Figures

Figure 1

35 pages, 1875 KB  
Review
FPGA-Accelerated ECG Analysis: Narrative Review of Signal Processing, ML/DL Models, and Design Optimizations
by Laura-Ioana Mihăilă, Claudia-Georgiana Barbura, Paul Faragó, Sorin Hintea, Botond Sandor Kirei and Albert Fazakas
Electronics 2026, 15(2), 301; https://doi.org/10.3390/electronics15020301 - 9 Jan 2026
Viewed by 260
Abstract
Recent advances in deep learning have had a significant impact on biomedical applications, driving precise actions in automated diagnostic processes. However, integrating neural networks into medical devices requires meeting strict requirements regarding computing power, energy efficiency, reconfigurability, and latency, essential conditions for real-time [...] Read more.
Recent advances in deep learning have had a significant impact on biomedical applications, driving precise actions in automated diagnostic processes. However, integrating neural networks into medical devices requires meeting strict requirements regarding computing power, energy efficiency, reconfigurability, and latency, essential conditions for real-time inference. Field-Programmable Gate Array (FPGA) architectures provide a high level of flexibility, performance, and parallel execution, thus making them a suitable option for the real-world implementation of machine learning (ML) and deep learning (DL) models in systems dedicated to the analysis of physiological signals. This paper presents a review of intelligent algorithms for electrocardiogram (ECG) signal classification, including Support Vector Machines (SVMs), Artificial Neural Networks (ANNs), Recurrent Neural Networks (RNNs), Long Short-Term Memory Networks (LSTMs), and Convolutional Neural Networks (CNNs), which have been implemented on FPGA platforms. A comparative evaluation of the performances of these hardware-accelerated solutions is provided, focusing on their classification accuracy. At the same time, the FPGA families used are analyzed, along with the reported performances in terms of operating frequency, power consumption, and latency, as well as the optimization strategies applied in the design of deep learning hardware accelerators. The conclusions emphasize the popularity and efficiency of CNN architectures in the context of ECG signal classification. The study aims to offer a current overview and to support specialists in the field of FPGA design and biomedical engineering in the development of accelerators dedicated to physiological signals analysis. Full article
(This article belongs to the Special Issue Emerging Biomedical Electronics)
Show Figures

Figure 1

16 pages, 2092 KB  
Article
Bidirectional Temporal Attention Convolutional Networks for High-Performance Network Traffic Anomaly Detection
by Feng Wang, Yufeng Huang and Yifei Shi
Information 2026, 17(1), 61; https://doi.org/10.3390/info17010061 - 9 Jan 2026
Viewed by 177
Abstract
Deep learning-based network traffic anomaly detection, particularly using Recurrent Neural Networks (RNNs), often struggles with high computational overhead and difficulties in capturing long-range temporal dependencies. To address these limitations, this paper proposes a Bidirectional Temporal Attention Convolutional Network (Bi-TACN) for robust and efficient [...] Read more.
Deep learning-based network traffic anomaly detection, particularly using Recurrent Neural Networks (RNNs), often struggles with high computational overhead and difficulties in capturing long-range temporal dependencies. To address these limitations, this paper proposes a Bidirectional Temporal Attention Convolutional Network (Bi-TACN) for robust and efficient network traffic anomaly detection. Specifically, dilated causal convolutions with expanding receptive fields and residual modules are employed to capture multi-scale temporal patterns while effectively mitigating the vanishing gradient. Furthermore, a bidirectional structure integrated with Efficient Channel Attention (ECA) is designed to adaptively weight contextual features, preventing sparse attack indicators from being overwhelmed by dominant normal traffic. A Softmax-based classifier then leverages these refined representations to execute high-performance anomaly detection. Extensive experiments on the NSL-KDD and UNSW-NB15 datasets demonstrate that Bi-TACN achieves average accuracies of 88.51% and 82.5%, respectively, significantly outperforming baseline models such as Bi-TCN and Bi-GRU in terms of both precision and convergence speed. Full article
Show Figures

Figure 1

27 pages, 7153 KB  
Article
State-Dependent CNN–GRU Reinforcement Framework for Robust EEG-Based Sleep Stage Classification
by Sahar Zakeri, Somayeh Makouei and Sebelan Danishvar
Biomimetics 2026, 11(1), 54; https://doi.org/10.3390/biomimetics11010054 - 8 Jan 2026
Viewed by 355
Abstract
Recent advances in automated learning techniques have enhanced the analysis of biomedical signals for detecting sleep stages and related health abnormalities. However, many existing models face challenges with imbalanced datasets and the dynamic nature of evolving sleep states. In this study, we present [...] Read more.
Recent advances in automated learning techniques have enhanced the analysis of biomedical signals for detecting sleep stages and related health abnormalities. However, many existing models face challenges with imbalanced datasets and the dynamic nature of evolving sleep states. In this study, we present a robust algorithm for classifying sleep states using electroencephalogram (EEG) data collected from 33 healthy participants. We extracted dynamic, brain-inspired features, such as microstates and Lempel–Ziv complexity, which replicate intrinsic neural processing patterns and reflect temporal changes in brain activity during sleep. An optimal feature set was identified based on significant spectral ranges and classification performance. The classifier was developed using a convolutional neural network (CNN) combined with gated recurrent units (GRUs) within a reinforcement learning framework, which models adaptive decision-making processes similar to those in biological neural systems. Our proposed biomimetic framework illustrates that a multivariate feature set provides strong discriminative power for sleep state classification. Benchmark comparisons with established approaches revealed a classification accuracy of 98% using the optimized feature set, with the framework utilizing fewer EEG channels and reducing processing time, underscoring its potential for real-time deployment. These findings indicate that applying biomimetic principles in feature extraction and model design can improve automated sleep monitoring and facilitate the development of novel therapeutic and diagnostic tools for sleep-related disorders. Full article
(This article belongs to the Section Bioinspired Sensorics, Information Processing and Control)
Show Figures

Graphical abstract

34 pages, 6460 KB  
Article
Explainable Gait Multi-Anchor Space-Aware Temporal Convolutional Networks for Gait Recognition in Neurological, Orthopedic, and Healthy Cohorts
by Abdullah Alharthi
Mathematics 2026, 14(2), 230; https://doi.org/10.3390/math14020230 - 8 Jan 2026
Viewed by 222
Abstract
Gait recognition using wearable sensor data is crucial for healthcare, rehabilitation, and monitoring neurological and musculoskeletal disorders. This study proposes a deep learning framework for gait classification using inertial measurements from four body-mounted IMU sensors (head, lower back, and both feet). The data [...] Read more.
Gait recognition using wearable sensor data is crucial for healthcare, rehabilitation, and monitoring neurological and musculoskeletal disorders. This study proposes a deep learning framework for gait classification using inertial measurements from four body-mounted IMU sensors (head, lower back, and both feet). The data were collected from a publicly available, clinically annotated dataset comprising 1356 gait trials from 260 individuals with diverse pathologies. The framework, G-MASA-TCN (Gait Multi-Anchor, Space-Aware Temporal Convolutional Network), integrates multi-scale temporal fusion, graph-informed spatial modeling, and residual dilated convolutions to extract discriminative gait signatures. To ensure both high performance and interpretability, Integrated Gradients is incorporated as an explainable AI (XAI) method, providing sensor-level and temporal attributes that reveal the features driving model decisions. The framework is evaluated via repeated cross-validation experiments, reporting detailed metrics with cross-run statistical analysis (mean ± standard deviation) to assess robustness. Results show that G-MASA-TCN achieves 98% classification accuracy for neurological, orthopedic, and healthy cohorts, demonstrating superior stability and resilience compared to baseline architectures, including Gated Recurrent Unit (GRU), Transformer neural networks, and standard TCNs, and 98.4% accuracy in identifying individual subjects based on gait. Furthermore, the model offers clinically meaningful insights into which sensors and gait phases contribute most to its predictions. This work presents an accurate, interpretable, and reliable tool for gait pathology recognition, with potential for translation to real-world clinical settings. Full article
(This article belongs to the Special Issue Deep Neural Network: Theory, Algorithms and Applications)
Show Figures

Graphical abstract

24 pages, 1788 KB  
Article
Uncertainty-Aware Machine Learning for NBA Forecasting in Digital Betting Markets
by Matteo Montrucchio, Enrico Barbierato and Alice Gatti
Information 2026, 17(1), 56; https://doi.org/10.3390/info17010056 - 8 Jan 2026
Viewed by 388
Abstract
This study introduces a fully uncertainty-aware forecasting framework for NBA games that integrates team-level performance metrics, rolling-form indicators, and spatial shot-chart embeddings. The predictive backbone is a recurrent neural network equipped with Monte Carlo dropout, yielding calibrated sequential probabilities. The model is evaluated [...] Read more.
This study introduces a fully uncertainty-aware forecasting framework for NBA games that integrates team-level performance metrics, rolling-form indicators, and spatial shot-chart embeddings. The predictive backbone is a recurrent neural network equipped with Monte Carlo dropout, yielding calibrated sequential probabilities. The model is evaluated against strong baselines including logistic regression, XGBoost, convolutional models, a GRU sequence model, and both market-only and non-market-only benchmarks. All experiments rely on strict chronological partitioning (train ≤ 2022, validation 2023, test 2024), ablation tests designed to eliminate any circularity with bookmaker odds, and cross-season robustness checks spanning 2012–2024. Predictive performance is assessed through accuracy, Brier score, log-loss, AUC, and calibration metrics (ECE/MCE), complemented by SHAP-based interpretability to verify that only pre-game information influences predictions. To quantify economic value, calibrated probabilities are fed into a frictionless betting simulator using fractional-Kelly staking, an expected-value threshold, and bootstrap-based uncertainty estimation. Empirically, the uncertainty-aware model delivers systematically better calibration than non-Bayesian baselines and benefits materially from the combination of shot-chart embeddings and recent-form features. Economic value emerges primarily in less-efficient segments of the market: The fused predictor outperforms both market-only and non-market-only variants on moneylines, while spreads and totals show limited exploitable edge, consistent with higher pricing efficiency. Sensitivity studies across Kelly multipliers, EV thresholds, odds caps, and sequence lengths confirm that the findings are robust to modelling and decision-layer perturbations. The paper contributes a reproducible, decision-focused framework linking uncertainty-aware prediction to economic outcomes, clarifying when predictive lift can be monetized in NBA markets, and outlining methodological pathways for improving robustness, calibration, and execution realism in sports forecasting. Full article
Show Figures

Graphical abstract

Back to TopTop