Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (859)

Search Parameters:
Keywords = BiLSTM neural network

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 1516 KB  
Article
Comparative Benchmarking of Deep Learning Architectures for Detecting Adversarial Attacks on Large Language Models
by Oleksandr Kushnerov, Ruslan Shevchuk, Serhii Yevseiev and Mikołaj Karpiński
Information 2026, 17(2), 155; https://doi.org/10.3390/info17020155 - 4 Feb 2026
Abstract
The rapid adoption of large language models (LLMs) in corporate and governmental systems has raised critical security concerns, particularly prompt injection attacks exploiting LLMs’ inability to differentiate control instructions from untrusted user inputs. This study systematically benchmarks neural network architectures for malicious prompt [...] Read more.
The rapid adoption of large language models (LLMs) in corporate and governmental systems has raised critical security concerns, particularly prompt injection attacks exploiting LLMs’ inability to differentiate control instructions from untrusted user inputs. This study systematically benchmarks neural network architectures for malicious prompt detection, emphasizing robustness against character-level adversarial perturbations—an aspect that remains comparatively underemphasized in the specific context of prompt-injection detection despite its established significance in general adversarial NLP. Using the Malicious Prompt Detection Dataset (MPDD) containing 39,234 labeled instances, eight architectures—Dense DNN, CNN, BiLSTM, BiGRU, Transformer, ResNet, and character-level variants of CNN and BiLSTM—were evaluated based on standard performance metrics (accuracy, F1-score, and AUC-ROC), adversarial robustness coefficients against spacing and homoglyph perturbations, and inference latency. Results indicate that the word-level 3_Word_BiLSTM achieved the highest performance on clean samples (accuracy = 0.9681, F1 = 0.9681), whereas the Transformer exhibited lower accuracy (0.9190) and significant vulnerability to spacing attacks (adversarial robustness ρspacing=0.61). Conversely, the Character-level BiLSTM demonstrated superior resilience (ρspacing=1.0, ρhomoglyph=0.98), maintaining high accuracy (0.9599) and generalization on external datasets with only 2–4% performance decay. These findings highlight that character-level representations provide intrinsic robustness against obfuscation attacks, suggesting Char_BiLSTM as a reliable component in defense-in-depth strategies for LLM-integrated systems. Full article
(This article belongs to the Special Issue Public Key Cryptography and Privacy Protection)
Show Figures

Graphical abstract

35 pages, 7867 KB  
Article
Inter-Comparison of Deep Learning Models for Flood Forecasting in Ethiopia’s Upper Awash Basin
by Girma Moges Mengistu, Addisu G. Semie, Gulilat T. Diro, Natei Ermias Benti, Emiola O. Gbobaniyi and Yonas Mersha
Water 2026, 18(3), 397; https://doi.org/10.3390/w18030397 - 3 Feb 2026
Abstract
Flood events driven by climate variability and change pose significant risks for socio-economic activities in the Awash Basin, necessitating advanced forecasting tools. This study benchmarks five deep learning (DL) architectures, Convolutional Neural Network (CNN), Long Short-Term Memory (LSTM), Gated Recurrent Unit (GRU), Bidirectional [...] Read more.
Flood events driven by climate variability and change pose significant risks for socio-economic activities in the Awash Basin, necessitating advanced forecasting tools. This study benchmarks five deep learning (DL) architectures, Convolutional Neural Network (CNN), Long Short-Term Memory (LSTM), Gated Recurrent Unit (GRU), Bidirectional LSTM (BiLSTM), and a Hybrid CNN–LSTM, for daily discharge forecasting for the Hombole catchment in the Upper Awash Basin (UAB) using 40 years of hydrometeorological observations (1981–2020). Rainfall, lagged discharge, and seasonal indicators were used as predictors. Model performance was evaluated against two baseline approaches, a conceptual HBV rainfall–runoff model as well as a climatology, using standard and hydrological metrics. Of the two baselines (climatology and HBV), the climatology showed limited skill with large bias and negative NSE, whereas the HBV model achieved moderate skill (NSE = 0.64 and KGE = 0.82). In contrast, all DL models substantially improved predictive performance, achieving test NSE values above 0.83 and low overall bias. Among them, the Hybrid CNN–LSTM provided the most balanced performance, combining local temporal feature extraction with long-term memory and yielding stable efficiency (NSE ≈ 0.84, KGE ≈ 0.90, and PBIAS ≈ −2%) across flow regimes. The LSTM and GRU models performed comparably, offering strong temporal learning and robust daily predictions, while BiLSTM improved flood timing through bidirectional sequence modeling. The CNN captured short-term variability effectively but showed weaker representation of extreme peaks. Analysis of peak-flow metrics revealed systematic underestimation of extreme discharge magnitudes across all models. However, a post-processing flow-regime classification based on discharge quantiles demonstrated high extreme-event detection skill, with deep learning models exceeding 89% accuracy in identifying extreme-flow occurrences on the test set. These findings indicate that, while magnitude errors remain for rare floods, DL models reliably discriminate flood regimes relevant for early warning. Overall, the results show that deep learning models provide clear improvements over climatology and conceptual baselines for daily streamflow forecasting in the UAB, while highlighting remaining challenges in peak-flow magnitude prediction. The study indicates promising results for the integration of deep learning methods into flood early-warning workflows; however, these results could be further improved by adopting a probabilistic forecasting framework that accounts for model uncertainty. Full article
(This article belongs to the Section Hydrology)
Show Figures

Figure 1

25 pages, 33109 KB  
Article
Spatio-Temporal Shoreline Changes and AI-Based Predictions for Sustainable Management of the Damietta–Port Said Coast, Nile Delta, Egypt
by Hesham M. El-Asmar, Mahmoud Sh. Felfla and Amal A. Mokhtar
Sustainability 2026, 18(3), 1557; https://doi.org/10.3390/su18031557 - 3 Feb 2026
Abstract
The Damietta–Port Said coast, Nile Delta, has experienced extreme morphological change over the past four decades due to sediment reduction due to Aswan High Dam and continued anthropogenic pressures. Using multi-temporal Landsat (1985–2025) and high-resolution RapidEye and PlanetScope imagery with 50 m-spaced transects, [...] Read more.
The Damietta–Port Said coast, Nile Delta, has experienced extreme morphological change over the past four decades due to sediment reduction due to Aswan High Dam and continued anthropogenic pressures. Using multi-temporal Landsat (1985–2025) and high-resolution RapidEye and PlanetScope imagery with 50 m-spaced transects, the study documents major shoreline shifts: the Damietta sand spit retreated by >1 km at its proximal apex while its distal tip advanced by ≈3.1 km southeastward under persistent longshore drift. Sectoral analyses reveal typical structure-induced patterns of updrift accretion (+180 to +210 m) and downdrift erosion (−50 to −330 m). To improve predictive capability beyond linear DSAS extrapolation, Nonlinear Autoregressive Exogenous (NARX) and Bidirectional Long Short-Term Memory (BiLSTM) neural networks were applied to forecast the 2050 shoreline. BiLSTM demonstrated superior stability, capturing nonlinear sediment transport patterns where NARX produced unstable over-predictions. Furthermore, coupled wave–flow modeling validates a sustainable management strategy employing successive short groins (45–50 m length, 150 m spacing). Simulations indicate that this configuration reduces longshore current velocities by 40–60% and suppresses rip-current eddies, offering a sediment-compatible alternative to conventional breakwaters and seawalls. This integrated remote sensing, hydrodynamic, and AI-based framework provides a robust scientific basis for adaptive, sediment-compatible shoreline management, supporting the long-term resilience of one of Egypt’s most vulnerable deltaic coasts under accelerating climatic and anthropogenic pressures. Full article
Show Figures

Figure 1

21 pages, 1315 KB  
Article
Ensemble Deep Learning Models for Multi-Class DNA Sequence Classification: A Comparative Study of CNN, BiLSTM, and GRU Architectures
by Elias Tabane, Ernest Mnkandla and Zenghui Wang
Appl. Sci. 2026, 16(3), 1545; https://doi.org/10.3390/app16031545 - 3 Feb 2026
Abstract
DNA sequence classification is a fundamental problem in bioinformatics, playing an indispensable role in gene annotation and disease prediction. Whereas most deep learning models, such as CNNs, BiLSTM networks, and GRUs, have been found individually optimal, each of these methods excels in modeling [...] Read more.
DNA sequence classification is a fundamental problem in bioinformatics, playing an indispensable role in gene annotation and disease prediction. Whereas most deep learning models, such as CNNs, BiLSTM networks, and GRUs, have been found individually optimal, each of these methods excels in modeling a specific aspect of sequence data: local motifs, long-range dependencies, and efficient temporal modeling of the sequences. Here, we present and evaluate an ensemble model that integrates CNN, BiLSTM, and GRU architectures via a majority voting combination scheme so that their complementary strengths can be harnessed. We trained and evaluated each standalone and the integrated model on a DNA dataset comprising 4380 sequences falling under five functional categories. The ensemble model achieved a classification accuracy of 90.6% with precision, recall, and F1 score equal to 0.91, thereby outperforming the state-of-the-art techniques by large margins. Although previous studies have tried analyzing each Deep Learning method individually for DNA classification tasks, none have attempted a systematic combination of CNN, BiLSTM, and GRU based on their ability to extract features simultaneously. The current research aims at presenting a novel method that combines these architectures based on a Majority Voting strategy and proves how their combination is better at extracting local patterns and long dependency information when compared individually. In particular, the proposed ensemble model smoothed the high recall of BiLSTM with the high precision of CNN, leading to more robust and reliable classification. The experiments involved a publicly available DNA sequence data set of 4380 sequences distributed over 5 classes. Our results emphasized the prospect of hybrid ensemble deep learning as a strong approach for complex genomic data analysis, opening ways toward more accurate and interpretable bioinformatics research. Full article
(This article belongs to the Special Issue Advances in Deep Learning and Intelligent Computing)
Show Figures

Figure 1

18 pages, 1921 KB  
Article
Prediction of Sleep Apnea Occurrence from a Single-Lead Electrocardiogram Using Stacking Hybrid Architecture with Gated Recurrent Neural Network Architectures and Logistic Regression
by Tan-Hsu Tan, Guan-Hua Chen, Shing-Hong Liu and Wenxi Chen
Technologies 2026, 14(2), 92; https://doi.org/10.3390/technologies14020092 - 1 Feb 2026
Viewed by 49
Abstract
Obstructive sleep apnea (OSA) is a common sleep disorder that impacts patient health and imposes a burden on families and healthcare systems. The diagnosis of OSA is usually performed through overnight polysomnography (PSG) in a hospital setting. In recent years, OSA detection using [...] Read more.
Obstructive sleep apnea (OSA) is a common sleep disorder that impacts patient health and imposes a burden on families and healthcare systems. The diagnosis of OSA is usually performed through overnight polysomnography (PSG) in a hospital setting. In recent years, OSA detection using a single-lead electrocardiogram (ECG) has been explored. The advantage of this method is that patients can be measured in home environments. Thus, the aim of this study was to predict occurrences of sleep apnea with parameters extracted from previous single-lead ECG measurements. The parameters were the R-R interval (RRI) and R-wave amplitude (RwA). The dataset was the single-lead ECG Apnea-ECG Database, and a stacking hybrid architecture (SHA) including three gated recurrent neural network architectures (GRNNAs) and logistic regression was proposed to improve the accuracy of OSA detection. Three GRNNAs used three different recurrent neural networks: Bidirectional Long Short-Term Memory (BiLSTM), Gated Recurrent Unit (GRU), and Bidirectional GRU (BiGRU). The challenge of this method was in exploring how many minutes of previous RRI and RwA measurements (n minutes) have the best performance in predicting occurrences of sleep apnea in the future (h minutes). The results showed that the SHA under an n of 20 min had the best performance in predicting occurrences of sleep apnea in the following 10 min: the SHA achieved a precision of 95.79%, sensitivity of 94.74%, specificity of 97.48%, F1-score of 95.26%, and accuracy of 96.45%. The proposed SHA was successful in predicting future sleep apnea occurrence with a single-lead ECG. Thus, this approach could be used in the development of wearable sleep monitors for the management of sleep apnea. Full article
(This article belongs to the Special Issue AI-Enabled Smart Healthcare Systems)
Show Figures

Figure 1

19 pages, 1638 KB  
Article
An Intrusion Detection Method for the Internet of Things Based on Spatiotemporal Fusion
by Junzhong He and Xiaorui An
Mathematics 2026, 14(3), 504; https://doi.org/10.3390/math14030504 - 30 Jan 2026
Viewed by 155
Abstract
In the information age, Internet of Things (IoT) devices are more susceptible to intrusion due to today’s complex network attack methods. Therefore, accurately detecting evolving network attacks from complex and ever-changing IoT environments has become a key research goal in the current intrusion [...] Read more.
In the information age, Internet of Things (IoT) devices are more susceptible to intrusion due to today’s complex network attack methods. Therefore, accurately detecting evolving network attacks from complex and ever-changing IoT environments has become a key research goal in the current intrusion detection field. Due to the spatial and temporal characteristics of IoT data, this paper proposes a Spatiotemporal Feature Weighted Fusion Approach Combining Gating Attention Transformation (STWGA). STWGA consists of three parts, namely spatial feature learning, the gated attention transformer, and the temporal feature learning module. It integrates improved convolutional neural networks (CNN), batch normalization, and Bidirectional Long Short-Term Memory Network (Bi-LSTM) to fully learn the deep spatial and temporal features of the data, achieving the goal of global deep spatiotemporal feature extraction. The gated attention transformer introduces an attention mechanism. In addition, an additional control mechanism is introduced in the self-attention module to more effectively improve detection accuracy. Finally, the experimental results show that STWGA has better spatiotemporal feature extraction ability and can effectively improve the intrusion detection effect of anomalies. Full article
Show Figures

Figure 1

19 pages, 2692 KB  
Article
A Hybrid Deep Learning Model Based on Spatio-Temporal Feature Mining for Traffic Analysis in Industrial Internet Gateway
by Danpei Li, Pinglai He, Jiayi Li, Panfeng Xu, Yan Song and Xiaoping Bai
Symmetry 2026, 18(2), 245; https://doi.org/10.3390/sym18020245 - 30 Jan 2026
Viewed by 104
Abstract
As the scale of the Industrial Internet continues to expand, the number of network connections and data traffic are experiencing explosive growth. Security threats and attack types targeting the Industrial Internet are becoming increasingly complex, rendering traditional firewalls and encryption/decryption technologies inadequate for [...] Read more.
As the scale of the Industrial Internet continues to expand, the number of network connections and data traffic are experiencing explosive growth. Security threats and attack types targeting the Industrial Internet are becoming increasingly complex, rendering traditional firewalls and encryption/decryption technologies inadequate for addressing diverse and sophisticated attack scenarios. Furthermore, traffic characteristics within the Industrial Internet environment exhibit significant asymmetry, such as a highly imbalanced distribution between benign and malicious traffic. To address this challenge, this paper proposes CBiNet—a hybrid deep learning model that integrates a one-dimensional convolutional neural network (1D-CNN) with a bidirectional long short-term memory network (BiLSTM). Designed to effectively learn and leverage such asymmetric spatio-temporal patterns, experimental validation demonstrates that the CBiNet model can efficiently tackle complex traffic identification tasks in industrial internet environments. It provides a highly accurate, scalable intrusion detection method for securing industrial internet gateways. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

26 pages, 6698 KB  
Article
A Novel Decomposition-Prediction Framework for Predicting InSAR-Derived Ground Displacement: A Case Study of the XMLC Landslide in China
by Mimi Peng, Jing Xue, Zhuge Xia, Jiantao Du and Yinghui Quan
Remote Sens. 2026, 18(3), 425; https://doi.org/10.3390/rs18030425 - 28 Jan 2026
Viewed by 178
Abstract
Interferometric Synthetic Aperture Radar (InSAR) is an advanced imaging geodesy technique for detecting and characterizing surface deformation with high spatial resolution and broad spatial coverage. However, as an inherently post-event observation method, InSAR suffers from limited capability for near-real-time and short-term updates of [...] Read more.
Interferometric Synthetic Aperture Radar (InSAR) is an advanced imaging geodesy technique for detecting and characterizing surface deformation with high spatial resolution and broad spatial coverage. However, as an inherently post-event observation method, InSAR suffers from limited capability for near-real-time and short-term updates of deformation time series. In this paper, we proposed a data-driven adaptive framework for deformation prediction based on a hybrid deep learning method to accurately predict the InSAR-derived deformation time series and take the Xi’erguazi−Mawo landslide complex (XMLC) as a case study. The InSAR-derived time series was initially decomposed into trend and periodic components with a two-step decomposition process, which were thereafter modeled separately to enhance the characterization of motion kinematics and prediction accuracy. After retrieving the observations from the multi-temporal InSAR method, two-step signal decomposition was then performed using the Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) and Variational Mode Decomposition (VMD). The decomposed trend and periodic components were further evaluated using statistical hypothesis testing to verify their significance and reliability. Compared with the single-decomposition model, the further decomposition leads to an overall improvement in prediction accuracy, i.e., the Mean Absolute Errors (MAEs) and the Root Mean Square Errors (RMSEs) are reduced by 40–49% and 36–42%, respectively. Subsequently, the Radial Basis Function (RBF) neural network and the proposed CNN-BiLSTM-SelfAttention (CBS) models were constructed to predict the trend and periodic variations, respectively. The CNN and self-attention help to extract local features in time series and strengthen the ability to capture global dependencies and key fluctuation patterns. Compared with the single network model in prediction, the MAEs and RMSEs are reduced by 22–57% and 4–33%, respectively. Finally, the two predicted components were integrated to generate the fused deformation prediction results. Ablation experiments and comparative experiments show that the proposed method has superior ability. Through rapid and accurate prediction of InSAR-derived deformation time series, this research could contribute to the early-warning systems of slope instabilities. Full article
Show Figures

Figure 1

20 pages, 1124 KB  
Article
Scalable Neural Cryptanalysis of Block Ciphers in Federated Attack Environments
by Ongee Jeong, Seonghwan Park and Inkyu Moon
Mathematics 2026, 14(2), 373; https://doi.org/10.3390/math14020373 - 22 Jan 2026
Viewed by 94
Abstract
This paper presents an extended investigation into deep learning-based cryptanalysis of block ciphers by introducing and evaluating a multi-server attack environment. Building upon our prior work in centralized settings, we explore the practicality and scalability of deploying such attacks across multiple distributed edge [...] Read more.
This paper presents an extended investigation into deep learning-based cryptanalysis of block ciphers by introducing and evaluating a multi-server attack environment. Building upon our prior work in centralized settings, we explore the practicality and scalability of deploying such attacks across multiple distributed edge servers. We assess the vulnerability of five representative block ciphers—DES, SDES, AES-128, SAES, and SPECK32/64—under two neural attack models: Encryption Emulation (EE) and Plaintext Recovery (PR), using both fully connected neural networks and Recurrent Neural Networks (RNNs) based on bidirectional Long Short-Term Memory (BiLSTM). Our experimental results show that the proposed federated learning-based cryptanalysis framework achieves performance nearly identical to that of centralized attacks, particularly for ciphers with low round complexity. Even as the number of edge servers increases to 32, the attack models maintain high accuracy in reduced-round settings. We validate our security assessments through formal statistical significance testing using two-tailed binomial tests with 99% confidence intervals. Additionally, our scalability analysis demonstrates that aggregation times remain negligible (<0.01% of total training time), confirming the computational efficiency of the federated framework. Overall, this work provides both a scalable cryptanalysis framework and valuable insights into the design of cryptographic algorithms that are resilient to distributed, deep learning-based threats. Full article
(This article belongs to the Section E: Applied Mathematics)
Show Figures

Figure 1

17 pages, 2935 KB  
Article
A Hybrid Deep Learning Framework for Non-Intrusive Load Monitoring
by Xiangbin Kong, Zhihang Gui, Minghu Wu, Chuyu Miao and Zhe Luo
Electronics 2026, 15(2), 453; https://doi.org/10.3390/electronics15020453 - 21 Jan 2026
Viewed by 211
Abstract
In recent years, load disaggregation and non-intrusive load-monitoring (NILM) methods have garnered widespread attention for optimizing energy management systems, becoming crucial tools for achieving energy efficiency and analyzing power consumption. However, existing NILM methods face challenges in accurately handling appliances with multiple operational [...] Read more.
In recent years, load disaggregation and non-intrusive load-monitoring (NILM) methods have garnered widespread attention for optimizing energy management systems, becoming crucial tools for achieving energy efficiency and analyzing power consumption. However, existing NILM methods face challenges in accurately handling appliances with multiple operational states and suffer from low accuracy and poor computational efficiency, particularly in modeling long-term dependencies and complex appliance load patterns. This article proposes an improved NILM model optimized based on transformers. The model first utilizes a convolutional neural network (CNN) to extract features from the input sequence and employs a bidirectional long short-term memory (BiLSTM) network to model long-term dependencies. Subsequently, multiple transformer blocks are used to capture dependencies within the sequence. To validate the effectiveness of the proposed model, we applied it to real-world household energy datasets: UK-DALE and REDD. Compared with suboptimal models, our model significantly improves the F1 score by 24.5% and 22.8%. Full article
(This article belongs to the Section Artificial Intelligence)
Show Figures

Figure 1

14 pages, 11925 KB  
Technical Note
Detecting Mowed Tidal Wetlands Using Time-Series NDVI and LSTM-Based Machine Learning
by Mayeesha Humaira, Stephen Aboagye-Ntow, Chuyuan Wang, Alexi Sanchez de Boado, Mark Burchick, Leslie Wood Mummert and Xin Huang
Land 2026, 15(1), 193; https://doi.org/10.3390/land15010193 - 21 Jan 2026
Viewed by 217
Abstract
This study presents the first application of machine learning (ML) to detect and map mowed tidal wetlands in the Chesapeake Bay region of Maryland and Virginia, focusing on emergent estuarine intertidal (E2EM) wetlands. Monitoring human disturbances like mowing is essential because repeated mowing [...] Read more.
This study presents the first application of machine learning (ML) to detect and map mowed tidal wetlands in the Chesapeake Bay region of Maryland and Virginia, focusing on emergent estuarine intertidal (E2EM) wetlands. Monitoring human disturbances like mowing is essential because repeated mowing stresses wetland vegetation, reducing habitat quality and diminishing other ecological services wetlands provide, including shoreline stabilization and water filtration. Traditional field-based monitoring is labor-intensive and impractical for large-scale assessments. To address these challenges, this study utilized 2021 and 2022 Sentinel-2 satellite imagery and a time-series analysis of the Normalized Difference Vegetation Index (NDVI) to distinguish between mowed and unmowed (control) wetlands. A bidirectional Long Short-Term Memory (BiLSTM) neural network was created to predict NDVI patterns associated with mowing events, such as rapid decreases followed by slow vegetation regeneration. The training dataset comprised 204 field-verified and desktop-identified samples, accounting for under 0.002% of the research area’s herbaceous E2EM wetlands. The model obtained 97.5% accuracy on an internal test set and was verified at eight separate Chesapeake Bay locations, indicating its promising generality. This work demonstrates the potential of remote sensing and machine learning for scalable, automated monitoring of tidal wetland disturbances to aid in conservation, restoration, and resource management. Full article
(This article belongs to the Section Land – Observation and Monitoring)
Show Figures

Figure 1

23 pages, 2529 KB  
Article
Loss Prediction and Global Sensitivity Analysis for Distribution Transformers Based on NRBO-Transformer-BiLSTM
by Qionglin Li, Yi Wang and Tao Mao
Electronics 2026, 15(2), 420; https://doi.org/10.3390/electronics15020420 - 18 Jan 2026
Viewed by 188
Abstract
As distributed energy resources and nonlinear loads are integrated into power grids on a large scale, power quality issues have grown increasingly prominent, triggering a substantial rise in distribution transformer losses. Traditional approaches struggle to accurately forecast transformer losses under complex power quality [...] Read more.
As distributed energy resources and nonlinear loads are integrated into power grids on a large scale, power quality issues have grown increasingly prominent, triggering a substantial rise in distribution transformer losses. Traditional approaches struggle to accurately forecast transformer losses under complex power quality conditions and lack quantitative analysis of the influence of various power quality indicators on losses. This study presents a data-driven methodology for transformer loss prediction and sensitivity analysis in such environments. First, an experimental platform is designed and built to measure transformer losses under composite power quality conditions, enabling the collection of actual measurement data when multi-source disturbances exist. Second, a high-precision loss prediction model—dubbed Newton-Raphson-Based Optimizer-Transformer-Bidirectional Long Short-Term Memory (NRBO-Transformer-BiLSTM)—is developed on the basis of an enhanced deep neural network. Finally, global sensitivity analysis methods are utilized to quantitatively evaluate the impact of different power quality indicators on transformer losses. Experimental results reveal that the proposed prediction model achieves an average error rate of less than 0.18% and a similarity coefficient of over 0.9989. Among all power quality indicators, voltage deviation has the most significant impact on transformer losses (with a sensitivity of 0.3268), followed by three-phase unbalance (sensitivity: 0.0109) and third harmonics (sensitivity: 0.0075). This research offers a theoretical foundation and technical support for enhancing the energy efficiency of distribution transformers and implementing effective power quality management. Full article
Show Figures

Figure 1

47 pages, 17315 KB  
Article
RNN Architecture-Based Short-Term Forecasting Framework for Rooftop PV Surplus to Enable Smart Energy Scheduling in Micro-Residential Communities
by Abdo Abdullah Ahmed Gassar, Mohammad Nazififard and Erwin Franquet
Buildings 2026, 16(2), 390; https://doi.org/10.3390/buildings16020390 - 17 Jan 2026
Viewed by 141
Abstract
With growing community awareness of greenhouse gas emissions and their environmental consequences, distributed rooftop photovoltaic (PV) systems have emerged as a sustainable energy alternative in residential settings. However, the high penetration of these systems without effective operational strategies poses significant challenges for local [...] Read more.
With growing community awareness of greenhouse gas emissions and their environmental consequences, distributed rooftop photovoltaic (PV) systems have emerged as a sustainable energy alternative in residential settings. However, the high penetration of these systems without effective operational strategies poses significant challenges for local distribution grids. Specifically, the estimation of surplus energy production from these systems, closely linked to complex outdoor weather conditions and seasonal fluctuations, often lacks an accurate forecasting approach to effectively capture the temporal dynamics of system output during peak periods. In response, this study proposes a recurrent neural network (RNN)- based forecasting framework to predict rooftop PV surplus in the context of micro-residential communities over time horizons not exceeding 48 h. The framework includes standard RNN, long short-term memory (LSTM), bidirectional LSTM (BiLSTM), and gated recurrent unit (GRU) networks. In this context, the study employed estimated surplus energy datasets from six single-family detached houses, along with weather-related variables and seasonal patterns, to evaluate the framework’s effectiveness. Results demonstrated the significant effectiveness of all framework models in forecasting surplus energy across seasonal scenarios, with low MAPE values of up to 3.02% and 3.59% over 24-h and 48-h horizons, respectively. Simultaneously, BiLSTM models consistently demonstrated a higher capacity to capture surplus energy fluctuations during peak periods than their counterparts. Overall, the developed data-driven framework demonstrates potential to enable short-term smart energy scheduling in micro-residential communities, supporting electric vehicle charging from single-family detached houses through efficient rooftop PV systems. It also provides decision-making insights for evaluating renewable energy contributions in the residential sector. Full article
(This article belongs to the Section Building Energy, Physics, Environment, and Systems)
Show Figures

Figure 1

22 pages, 5927 KB  
Article
Research on a Temperature and Humidity Prediction Model for Greenhouse Tomato Based on iT-LSTM-CA
by Yanan Gao, Pingzeng Liu, Yuxuan Zhang, Fengyu Li, Ke Zhu, Yan Zhang and Shiwei Xu
Sustainability 2026, 18(2), 930; https://doi.org/10.3390/su18020930 - 16 Jan 2026
Viewed by 171
Abstract
Constructing a temperature and humidity prediction model for greenhouse-grown tomatoes is of great significance for achieving resource-efficient and sustainable greenhouse environmental control and promoting healthy tomato growth. However, traditional models often struggle to simultaneously capture long-term temporal trends, short-term local dynamic variations, and [...] Read more.
Constructing a temperature and humidity prediction model for greenhouse-grown tomatoes is of great significance for achieving resource-efficient and sustainable greenhouse environmental control and promoting healthy tomato growth. However, traditional models often struggle to simultaneously capture long-term temporal trends, short-term local dynamic variations, and the coupling relationships among multiple variables. To address these issues, this study develops an iT-LSTM-CA multi-step prediction model, in which the inverted Transformer (iTransformer, iT) is employed to capture global dependencies across variables and long temporal scales, the Long Short-Term Memory (LSTM) network is utilized to extract short-term local variation patterns, and a cross-attention (CA) mechanism is introduced to dynamically fuse the two types of features. Experimental results show that, compared with models such as Gated Recurrent Unit (GRU), Temporal Convolutional Network (TCN), Recurrent Neural Network (RNN), LSTM, and Bidirectional Long Short-Term Memory (Bi-LSTM), the iT-LSTM-CA achieves the best performance in multi-step forecasting tasks at 3 h, 6 h, 12 h, and 24 h horizons. For temperature prediction, the R2 ranges from 0.96 to 0.98, with MAE between 0.42 °C and 0.79 °C and RMSE between 0.58 °C and 1.06 °C; for humidity prediction, the R2 ranges from 0.95 to 0.97, with MAE between 1.21% and 2.49% and RMSE between 1.78% and 3.42%. These results indicate that the iT-LSTM-CA model can effectively capture greenhouse environmental variations and provide a scientific basis for environmental control and management in tomato greenhouses. Full article
Show Figures

Figure 1

23 pages, 1151 KB  
Article
CNN–BiLSTM–Attention-Based Hybrid-Driven Modeling for Diameter Prediction of Czochralski Silicon Single Crystals
by Pengju Zhang, Hao Pan, Chen Chen, Yiming Jing and Ding Liu
Crystals 2026, 16(1), 57; https://doi.org/10.3390/cryst16010057 - 13 Jan 2026
Viewed by 211
Abstract
High-precision prediction of the crystal diameter during the growth of electronic-grade silicon single crystals is a critical step for the fabrication of high-quality single crystals. However, the process features high-temperature operation, strong nonlinearities, significant time-delay dynamics, and external disturbances, which limit the accuracy [...] Read more.
High-precision prediction of the crystal diameter during the growth of electronic-grade silicon single crystals is a critical step for the fabrication of high-quality single crystals. However, the process features high-temperature operation, strong nonlinearities, significant time-delay dynamics, and external disturbances, which limit the accuracy of conventional mechanism-based models. In this study, mechanism-based models denote physics-informed heat-transfer and geometric models that relate heater power and pulling rate to diameter evolution. To address this challenge, this paper proposes a hybrid deep learning model combining a convolutional neural network (CNN), a bidirectional long short-term memory network (BiLSTM), and self-attention to improve diameter prediction during the shoulder-formation and constant-diameter stages. The proposed model leverages the CNN to extract localized spatial features from multi-source sensor data, employs the BiLSTM to capture temporal dependencies inherent to the crystal growth process, and utilizes the self-attention mechanism to dynamically highlight critical feature information, thereby substantially enhancing the model’s capacity to represent complex industrial operating conditions. Experiments on operational production data collected from an industrial Czochralski (Cz) furnace, model TDR-180, demonstrate improved prediction accuracy and robustness over mechanism-based and single data-driven baselines, supporting practical process control and production optimization. Full article
(This article belongs to the Section Inorganic Crystalline Materials)
Show Figures

Figure 1

Back to TopTop