Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (2,345)

Search Parameters:
Keywords = convolutional long short-term memory

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 7086 KiB  
Article
Gas Leak Detection and Leakage Rate Identification in Underground Utility Tunnels Using a Convolutional Recurrent Neural Network
by Ziyang Jiang, Canghai Zhang, Zhao Xu and Wenbin Song
Appl. Sci. 2025, 15(14), 8022; https://doi.org/10.3390/app15148022 - 18 Jul 2025
Abstract
An underground utility tunnel (UUT) is essential for the efficient use of urban underground space. However, current maintenance systems rely on patrol personnel and professional equipment. This study explores industrial detection methods for identifying and monitoring natural gas leaks in UUTs. Via infrared [...] Read more.
An underground utility tunnel (UUT) is essential for the efficient use of urban underground space. However, current maintenance systems rely on patrol personnel and professional equipment. This study explores industrial detection methods for identifying and monitoring natural gas leaks in UUTs. Via infrared thermal imaging gas experiments, data were acquired and a dataset established. To address the low-resolution problem of existing imaging devices, video super-resolution (VSR) was used to improve the data quality. Based on a convolutional recurrent neural network (CRNN), the image features at each moment were extracted, and the time series data were modeled to realize the risk-level classification mechanism based on the automatic classification of the leakage rate. The experimental results show that when the sliding window size was set to 10 frames, the classification accuracy of the CRNN was the highest, which could reach 0.98. This method improves early warning precision and response efficiency, offering practical technical support for UUT maintenance management. Full article
(This article belongs to the Special Issue Applications of Artificial Intelligence in Industrial Engineering)
11 pages, 1540 KiB  
Article
Extraction of Clinically Relevant Temporal Gait Parameters from IMU Sensors Mimicking the Use of Smartphones
by Aske G. Larsen, Line Ø. Sadolin, Trine R. Thomsen and Anderson S. Oliveira
Sensors 2025, 25(14), 4470; https://doi.org/10.3390/s25144470 - 18 Jul 2025
Abstract
As populations age and workforces decline, the need for accessible health assessment methods grows. The merging of accessible and affordable sensors such as inertial measurement units (IMUs) and advanced machine learning techniques now enables gait assessment beyond traditional laboratory settings. A total of [...] Read more.
As populations age and workforces decline, the need for accessible health assessment methods grows. The merging of accessible and affordable sensors such as inertial measurement units (IMUs) and advanced machine learning techniques now enables gait assessment beyond traditional laboratory settings. A total of 52 participants walked at three speeds while carrying a smartphone-sized IMU in natural positions (hand, trouser pocket, or jacket pocket). A previously trained Convolutional Neural Network and Long Short-Term Memory (CNN-LSTM)-based machine learning model predicted gait events, which were then used to calculate stride time, stance time, swing time, and double support time. Stride time predictions were highly accurate (<5% error), while stance and swing times exhibited moderate variability and double support time showed the highest errors (>20%). Despite these variations, moderate-to-strong correlations between the predicted and experimental spatiotemporal gait parameters suggest the feasibility of IMU-based gait tracking in real-world settings. These associations preserved inter-subject patterns that are relevant for detecting gait disorders. Our study demonstrated the feasibility of extracting clinically relevant gait parameters using IMU data mimicking smartphone use, especially parameters with longer durations such as stride time. Robustness across sensor locations and walking speeds supports deep learning on single-IMU data as a viable tool for remote gait monitoring. Full article
(This article belongs to the Special Issue Sensor Systems for Gesture Recognition (3rd Edition))
Show Figures

Figure 1

23 pages, 14080 KiB  
Article
Regional Ecological Environment Quality Prediction Based on Multi-Model Fusion
by Yiquan Song, Zhengwei Li and Baoquan Wei
Land 2025, 14(7), 1486; https://doi.org/10.3390/land14071486 - 17 Jul 2025
Abstract
Regional ecological environmental quality (EEQ) is a vital indicator for environmental management and supporting sustainable development. However, the absence of robust and accurate EEQ prediction models has hindered effective environmental strategies. This study proposes a novel approach to address this gap by integrating [...] Read more.
Regional ecological environmental quality (EEQ) is a vital indicator for environmental management and supporting sustainable development. However, the absence of robust and accurate EEQ prediction models has hindered effective environmental strategies. This study proposes a novel approach to address this gap by integrating the ecological index (EI) model with several predictive models, including autoregressive integrated moving average (ARIMA), convolutional neural network (CNN), long short-term memory (LSTM), and cellular automata (CA), to forecast regional EEQ. Initially, the spatiotemporal evolution of the input data used to calculate the EI score was analyzed. Subsequently, tailored prediction models were developed for each dataset. These models were sequentially trained and validated, and their outputs were integrated into the EI model to enhance the accuracy and coherence of the final EEQ predictions. The novelty of this methodology lies not only in integrating existing predictive models but also in employing an innovative fusion technique that significantly improves prediction accuracy. Despite data quality issues in the case study dataset led to higher prediction errors in certain regions, the overall results exhibit a high degree of accuracy. A comparison of long-term EI predictions with EI assessment results reveals that the R2 value for the EI score exceeds 0.96, and the kappa value surpasses 0.76 for the EI level, underscoring the robust performance of the integrated model in forecasting regional EEQ. This approach offers valuable insights into exploring regional EEQ trends and future challenges. Full article
Show Figures

Figure 1

25 pages, 4363 KiB  
Article
Method for Predicting Transformer Top Oil Temperature Based on Multi-Model Combination
by Lin Yang, Minghe Wang, Liang Chen, Fan Zhang, Shen Ma, Yang Zhang and Sixu Yang
Electronics 2025, 14(14), 2855; https://doi.org/10.3390/electronics14142855 - 17 Jul 2025
Abstract
The top oil temperature of a transformer is a vital sign reflecting its operational condition. The accurate prediction of this parameter is essential for evaluating insulation performance and extending equipment lifespan. At present, the prediction of oil temperature is mainly based on single-feature [...] Read more.
The top oil temperature of a transformer is a vital sign reflecting its operational condition. The accurate prediction of this parameter is essential for evaluating insulation performance and extending equipment lifespan. At present, the prediction of oil temperature is mainly based on single-feature prediction. However, it overlooks the influence of other features. This has a negative effect on the prediction accuracy. Furthermore, the training dataset is often made up of data from a single transformer. This leads to the poor generalization of the prediction. To tackle these challenges, this paper leverages large-scale data analysis and processing techniques, and presents a transformer top oil temperature prediction model that combines multiple models. The Convolutional Neural Network was applied in this method to extract spatial features from multiple input variables. Subsequently, a Long Short-Term Memory network was employed to capture dynamic patterns in the time series. Meanwhile, a Transformer encoder enhanced feature interaction and global perception. The spatial characteristics extracted by the CNN and the temporal characteristics extracted by LSTM were further integrated to create a more comprehensive representation. The established model was optimized using the Whale Optimization Algorithm to improve prediction accuracy. The results of the experiment indicate that the maximum RMSE and MAPE of this method on the summer and winter datasets were 0.5884 and 0.79%, respectively, demonstrating superior prediction accuracy. Compared with other models, the proposed model improved prediction performance by 13.74%, 36.66%, and 43.36%, respectively, indicating high generalization capability and accuracy. This provides theoretical support for condition monitoring and fault warning of power equipment. Full article
Show Figures

Figure 1

18 pages, 533 KiB  
Article
Comparative Analysis of Deep Learning Models for Intrusion Detection in IoT Networks
by Abdullah Waqas, Sultan Daud Khan, Zaib Ullah, Mohib Ullah and Habib Ullah
Computers 2025, 14(7), 283; https://doi.org/10.3390/computers14070283 - 17 Jul 2025
Abstract
The Internet of Things (IoT) holds transformative potential in fields such as power grid optimization, defense networks, and healthcare. However, the constrained processing capacities and resource limitations of IoT networks make them especially susceptible to cyber threats. This study addresses the problem of [...] Read more.
The Internet of Things (IoT) holds transformative potential in fields such as power grid optimization, defense networks, and healthcare. However, the constrained processing capacities and resource limitations of IoT networks make them especially susceptible to cyber threats. This study addresses the problem of detecting intrusions in IoT environments by evaluating the performance of deep learning (DL) models under different data and algorithmic conditions. We conducted a comparative analysis of three widely used DL models—Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM), and Bidirectional LSTM (biLSTM)—across four benchmark IoT intrusion detection datasets: BoTIoT, CiCIoT, ToNIoT, and WUSTL-IIoT-2021. Each model was assessed under balanced and imbalanced dataset configurations and evaluated using three loss functions (cross-entropy, focal loss, and dual focal loss). By analyzing model efficacy across these datasets, we highlight the importance of generalizability and adaptability to varied data characteristics that are essential for real-world applications. The results demonstrate that the CNN trained using the cross-entropy loss function consistently outperforms the other models, particularly on balanced datasets. On the other hand, LSTM and biLSTM show strong potential in temporal modeling, but their performance is highly dependent on the characteristics of the dataset. By analyzing the performance of multiple DL models under diverse datasets, this research provides actionable insights for developing secure, interpretable IoT systems that can meet the challenges of designing a secure IoT system. Full article
(This article belongs to the Special Issue Application of Deep Learning to Internet of Things Systems)
Show Figures

Figure 1

21 pages, 4238 KiB  
Article
Fault Prediction of Hydropower Station Based on CNN-LSTM-GAN with Biased Data
by Bei Liu, Xiao Wang, Zhaoxin Zhang, Zhenjie Zhao, Xiaoming Wang and Ting Liu
Energies 2025, 18(14), 3772; https://doi.org/10.3390/en18143772 - 16 Jul 2025
Viewed by 65
Abstract
Fault prediction of hydropower station is crucial for the stable operation of generator set equipment, but the traditional method struggles to deal with data with an imbalanced distribution and untrustworthiness. This paper proposes a fault detection method based on a convolutional neural network [...] Read more.
Fault prediction of hydropower station is crucial for the stable operation of generator set equipment, but the traditional method struggles to deal with data with an imbalanced distribution and untrustworthiness. This paper proposes a fault detection method based on a convolutional neural network (CNNs) and long short-term memory network (LSTM) with a generative adversarial network (GAN). Firstly, a reliability mechanism based on principal component analysis (PCA) is designed to solve the problem of data bias caused by multiple monitoring devices. Then, the CNN-LSTM network is used to predict time series data, and the GAN is used to expand fault data samples to solve the problem of an unbalanced data distribution. Meanwhile, a multi-scale feature extraction network with time–frequency information is designed to improve the accuracy of fault detection. Finally, a dynamic multi-task training algorithm is proposed to ensure the convergence and training efficiency of the deep models. Experimental results show that compared with RNN, GRU, SVM, and threshold detection algorithms, the proposed fault prediction method improves the accuracy performance by 5.5%, 4.8%, 7.8%, and 9.3%, with at least a 160% improvement in the fault recall rate. Full article
(This article belongs to the Special Issue Optimal Schedule of Hydropower and New Energy Power Systems)
Show Figures

Figure 1

20 pages, 2632 KiB  
Article
Data-Driven Attack Detection Mechanism Against False Data Injection Attacks in DC Microgrids Using CNN-LSTM-Attention
by Chunxiu Li, Xinyu Wang, Xiaotao Chen, Aiming Han and Xingye Zhang
Symmetry 2025, 17(7), 1140; https://doi.org/10.3390/sym17071140 - 16 Jul 2025
Viewed by 50
Abstract
This study presents a novel spatio-temporal detection framework for identifying False Data Injection (FDI) attacks in DC microgrid systems from the perspective of cyber–physical symmetry. While modern DC microgrids benefit from increasingly sophisticated cyber–physical symmetry network integration, this interconnected architecture simultaneously introduces significant [...] Read more.
This study presents a novel spatio-temporal detection framework for identifying False Data Injection (FDI) attacks in DC microgrid systems from the perspective of cyber–physical symmetry. While modern DC microgrids benefit from increasingly sophisticated cyber–physical symmetry network integration, this interconnected architecture simultaneously introduces significant cybersecurity vulnerabilities. Notably, FDI attacks can effectively bypass conventional Chi-square detector-based protection mechanisms through malicious manipulation of communication layer data. To address this critical security challenge, we propose a hybrid deep learning framework that synergistically combines: Convolutional Neural Networks (CNN) for robust spatial feature extraction from power system measurements; Long Short-Term Memory (LSTM) networks for capturing complex temporal dependencies; and an attention mechanism that dynamically weights the most discriminative features. The framework operates through a hierarchical feature extraction process: First-level spatial analysis identifies local measurement patterns; second-level temporal analysis detects sequential anomalies; attention-based feature refinement focuses on the most attack-relevant signatures. Comprehensive simulation studies demonstrate the superior performance of our CNN-LSTM-Attention framework compared to conventional detection approaches (CNN-SVM and MLP), with significant improvements across all key metrics. Namely, the accuracy, precision, F1-score, and recall could be improved by at least 7.17%, 6.59%, 2.72% and 6.55%. Full article
Show Figures

Figure 1

24 pages, 6089 KiB  
Article
An Optimized 1-D CNN-LSTM Approach for Fault Diagnosis of Rolling Bearings Considering Epistemic Uncertainty
by Onur Can Kalay
Machines 2025, 13(7), 612; https://doi.org/10.3390/machines13070612 - 16 Jul 2025
Viewed by 40
Abstract
Rolling bearings are indispensable but also the most fault-prone components of rotating machinery, typically used in fields such as industrial aircraft, production workshops, and manufacturing. They encounter diverse mechanical stresses, such as vibration and friction during operation, which may lead to wear and [...] Read more.
Rolling bearings are indispensable but also the most fault-prone components of rotating machinery, typically used in fields such as industrial aircraft, production workshops, and manufacturing. They encounter diverse mechanical stresses, such as vibration and friction during operation, which may lead to wear and fatigue cracks. From this standpoint, the present study combined a 1-D convolutional neural network (1-D CNN) with a long short-term memory (LSTM) algorithm for classifying different ball-bearing health conditions. A physics-guided method that adopts fault characteristics frequencies was used to calculate an optimal input size (sample length). Moreover, grid search was utilized to optimize (1) the number of epochs, (2) batch size, and (3) dropout ratio and further enhance the efficacy of the proposed 1-D CNN-LSTM network. Therefore, an attempt was made to reduce epistemic uncertainty that arises due to not knowing the best possible hyper-parameter configuration. Ultimately, the effectiveness of the physics-guided optimized 1-D CNN-LSTM was tested by comparing its performance with other state-of-the-art models. The findings revealed that the average accuracies could be enhanced by up to 20.717% with the help of the proposed approach after testing it on two benchmark datasets. Full article
(This article belongs to the Section Machines Testing and Maintenance)
Show Figures

Figure 1

22 pages, 11458 KiB  
Article
Convolutional Neural Networks—Long Short-Term Memory—Attention: A Novel Model for Wear State Prediction Based on Oil Monitoring Data
by Ying Du, Hui Wei, Tao Shao, Shishuai Chen, Jianlei Wang, Chunguo Zhou and Yanchao Zhang
Lubricants 2025, 13(7), 306; https://doi.org/10.3390/lubricants13070306 - 15 Jul 2025
Viewed by 171
Abstract
Wear state prediction based on oil monitoring technology enables the early identification of potential wear and failure risks of friction pairs, facilitating optimized equipment maintenance and extended service life. However, the complexity of lubricating oil monitoring data often poses challenges in extracting discriminative [...] Read more.
Wear state prediction based on oil monitoring technology enables the early identification of potential wear and failure risks of friction pairs, facilitating optimized equipment maintenance and extended service life. However, the complexity of lubricating oil monitoring data often poses challenges in extracting discriminative features, limiting the accuracy of wear state prediction. To address this, a CNN–LSTM–Attention network is specially constructed for predicting wear state, which hierarchically integrates convolutional neural networks (CNNs) for spatial feature extraction, long short-term memory (LSTM) networks for temporal dynamics modeling, and self-attention mechanisms for adaptive feature refinement. The proposed architecture implements a three-stage computational pipeline. Initially, the CNN performs hierarchical extraction of localized patterns from multi-sensor tribological signals. Subsequently, the self-attention mechanism conducts adaptive recalibration of feature saliency, prioritizing diagnostically critical feature channels. Ultimately, bidirectional LSTM establishes cross-cyclic temporal dependencies, enabling cascaded fully connected layers with Gaussian activation to generate probabilistic wear state estimations. Experimental results demonstrate that the proposed model not only achieves superior predictive accuracy but also exhibits robust stability, offering a reliable solution for condition monitoring and predictive maintenance in industrial applications. Full article
Show Figures

Figure 1

24 pages, 5534 KiB  
Article
Enhancing Healthcare Assistance with a Self-Learning Robotics System: A Deep Imitation Learning-Based Solution
by Yagna Jadeja, Mahmoud Shafik, Paul Wood and Aaisha Makkar
Electronics 2025, 14(14), 2823; https://doi.org/10.3390/electronics14142823 - 14 Jul 2025
Viewed by 214
Abstract
This paper presents a Self-Learning Robotic System (SLRS) for healthcare assistance using Deep Imitation Learning (DIL). The proposed SLRS solution can observe and replicate human demonstrations, thereby acquiring complex skills without the need for explicit task-specific programming. It incorporates modular components for perception [...] Read more.
This paper presents a Self-Learning Robotic System (SLRS) for healthcare assistance using Deep Imitation Learning (DIL). The proposed SLRS solution can observe and replicate human demonstrations, thereby acquiring complex skills without the need for explicit task-specific programming. It incorporates modular components for perception (i.e., advanced computer vision methodologies), actuation (i.e., dynamic interaction with patients and healthcare professionals in real time), and learning. The innovative approach of implementing a hybrid model approach (i.e., deep imitation learning and pose estimation algorithms) facilitates autonomous learning and adaptive task execution. The environmental awareness and responsiveness were also enhanced using both a Convolutional Neural Network (CNN)-based object detection mechanism using YOLOv8 (i.e., with 94.3% accuracy and 18.7 ms latency) and pose estimation algorithms, alongside a MediaPipe and Long Short-Term Memory (LSTM) framework for human action recognition. The developed solution was tested and validated in healthcare, with the aim to overcome some of the current challenges, such as workforce shortages, ageing populations, and the rising prevalence of chronic diseases. The CAD simulation, validation, and verification tested functions (i.e., assistive functions, interactive scenarios, and object manipulation) of the system demonstrated the robot’s adaptability and operational efficiency, achieving an 87.3% task completion success rate and over 85% grasp success rate. This approach highlights the potential use of an SLRS for healthcare assistance. Further work will be undertaken in hospitals, care homes, and rehabilitation centre environments to generate complete holistic datasets to confirm the system’s reliability and efficiency. Full article
Show Figures

Figure 1

27 pages, 3705 KiB  
Article
A Method for Selecting the Appropriate Source Domain Buildings for Building Energy Prediction in Transfer Learning: Using the Euclidean Distance and Pearson Coefficient
by Chuyi Luo, Liang Xia and Sung-Hugh Hong
Energies 2025, 18(14), 3706; https://doi.org/10.3390/en18143706 - 14 Jul 2025
Viewed by 88
Abstract
Building energy prediction faces challenges such as data scarcity while Transfer Learning (TL) demonstrates significant potential by leveraging source building energy data to enhance target building energy prediction. However, the accuracy of TL heavily relies on selecting appropriate source buildings as the source [...] Read more.
Building energy prediction faces challenges such as data scarcity while Transfer Learning (TL) demonstrates significant potential by leveraging source building energy data to enhance target building energy prediction. However, the accuracy of TL heavily relies on selecting appropriate source buildings as the source data. This study proposes a novel, easy-to-understand, statistics-based visualization method that combines the Euclidean distance and Pearson correlation coefficient for selecting source buildings in TL for target building electricity prediction. Long Short-Term Memory, the Gated Recurrent Unit, and the Convolutional Neural Network were applied to verify the appropriateness of the source domain buildings. The results showed the source building, selected via the method proposed by this research, could reduce 65% of computational costs, while the RMSE was approximately 6.5 kWh, and the R2 was around 0.92. The method proposed in this study is well suited for scenes requiring rapid response times and exhibiting low tolerance for prediction errors. Full article
(This article belongs to the Special Issue Innovations in Low-Carbon Building Energy Systems)
Show Figures

Figure 1

18 pages, 6924 KiB  
Article
A Method Based on CNN–BiLSTM–Attention for Wind Farm Line Fault Distance Prediction
by Ming Zhang, Qingzhong Gao, Baoliang Liu, Chen Zhang and Guangkai Zhou
Energies 2025, 18(14), 3703; https://doi.org/10.3390/en18143703 - 14 Jul 2025
Viewed by 159
Abstract
In view of the complex operating environments of wind farms and the characteristics of multi-branch mixed collector lines, in order to improve the accuracy of single-phase grounding fault location, the convolutional neural network (CNN), bidirectional long short-term memory network (BiLSTM), and attention mechanism [...] Read more.
In view of the complex operating environments of wind farms and the characteristics of multi-branch mixed collector lines, in order to improve the accuracy of single-phase grounding fault location, the convolutional neural network (CNN), bidirectional long short-term memory network (BiLSTM), and attention mechanism (attention) were combined to construct a single-phase grounding fault location strategy for the CNN–BiLSTM–attention hybrid model. Using a zero-sequence current as the fault information identification method, through the deep fusion of the CNN–BiLSTM–attention hybrid model, the single-phase grounding faults in the collector lines of the wind farm can be located. The simulation modeling was carried out using the MATLAB R2022b software, and the effectiveness of the hybrid model in the single-phase grounding fault location of multi-branch mixed collector lines was studied and verified. The research results show that, compared with the random forest algorithm, decision tree algorithm, CNN, and LSTM neural network, the proposed method significantly improved the location accuracy and is more suitable for the fault distance measurement requirements of collector lines in the complex environments of wind farms. The research conclusions provide technical support and a reference for the actual operation and maintenance of wind farms. Full article
(This article belongs to the Section A3: Wind, Wave and Tidal Energy)
Show Figures

Figure 1

33 pages, 7266 KiB  
Article
Temperature Prediction and Fault Warning of High-Speed Shaft of Wind Turbine Gearbox Based on Hybrid Deep Learning Model
by Min Zhang, Jijie Wei, Zhenli Sui, Kun Xu and Wenyong Yuan
J. Mar. Sci. Eng. 2025, 13(7), 1337; https://doi.org/10.3390/jmse13071337 - 13 Jul 2025
Viewed by 215
Abstract
Gearbox failure represents one of the most time-consuming maintenance challenges in wind turbine operations. Abnormal temperature variations in the gearbox high-speed shaft (GHSS) serve as reliable indicators of potential faults. This study proposes a Spatio-Temporal Attentive (STA) synergistic architecture for GHSS fault detection [...] Read more.
Gearbox failure represents one of the most time-consuming maintenance challenges in wind turbine operations. Abnormal temperature variations in the gearbox high-speed shaft (GHSS) serve as reliable indicators of potential faults. This study proposes a Spatio-Temporal Attentive (STA) synergistic architecture for GHSS fault detection and early warning by utilizing the in situ monitoring data from a wind farm. This comprehensive architecture involves five modules: data preprocessing, multi-dimensional spatial feature extraction, temporal dependency modeling, global relationship learning, and hyperparameter optimization. It was achieved by using real-time monitoring data to predict the GHSS temperature in 10 min, with an accuracy of 1 °C. Compared to the long short-term memory (LSTM) and convolutional neural network and LSTM hybrid models, the STA architecture reduces the root mean square error of the prediction by approximately 37% and 13%, respectively. Furthermore, the architecture establishes a normal operating condition model and provides benchmark eigenvalues for subsequent fault warnings. The model was validated to issue early warnings up to seven hours before the fault alert is triggered by the supervisory control and data acquisition system of the wind turbine. By offering reliable, cost-effective prognostics without additional hardware, this approach significantly improves wind turbine health management and fault prevention. Full article
(This article belongs to the Section Ocean Engineering)
Show Figures

Figure 1

15 pages, 1296 KiB  
Article
Predicting Photovoltaic Energy Production Using Neural Networks: Renewable Integration in Romania
by Grigore Cican, Adrian-Nicolae Buturache and Valentin Silivestru
Processes 2025, 13(7), 2219; https://doi.org/10.3390/pr13072219 - 11 Jul 2025
Viewed by 268
Abstract
Photovoltaic panels are pivotal in transforming solar irradiance into electricity, making them a key technology in renewable energy. Despite their potential, the distribution of photovoltaic systems in Romania remains sparse, requiring advanced data analytics for effective management, particularly in addressing the intermittent nature [...] Read more.
Photovoltaic panels are pivotal in transforming solar irradiance into electricity, making them a key technology in renewable energy. Despite their potential, the distribution of photovoltaic systems in Romania remains sparse, requiring advanced data analytics for effective management, particularly in addressing the intermittent nature of photovoltaic energy. This study investigates the predictive capabilities of Long Short-Term Memory (LSTM) and Convolutional Neural Network (CNN) architectures for forecasting hourly photovoltaic energy production in Romania. The results indicate that CNN models significantly outperform LSTM models, with 77% of CNNs achieving an R2 of 0.9 or higher compared to only 13% for LSTMs. The best-performing CNN model reached an R2 of 0.9913 with a mean absolute error (MAE) of 9.74, while the top LSTM model achieved an R2 of 0.9880 and an MAE of 12.57. The rapid convergence of the CNN model to stable error rates illustrates its superior generalization capabilities. Moreover, the model’s ability to accurately predict photovoltaic production over a two-day timeframe, which is not included in the testing dataset, confirms its robustness. This research highlights the critical role of accurate energy forecasting in optimizing the integration of photovoltaic energy into Romania’s power grid, thereby supporting sustainable energy management strategies in line with the European Union’s climate goals. Through this methodology, we aim to enhance the operational safety and efficiency of photovoltaic systems, facilitating their large-scale adoption and ultimately contributing to the fight against climate change. Full article
(This article belongs to the Special Issue Design, Modeling and Optimization of Solar Energy Systems)
Show Figures

Figure 1

16 pages, 1730 KiB  
Article
Retail Demand Forecasting: A Comparative Analysis of Deep Neural Networks and the Proposal of LSTMixer, a Linear Model Extension
by Georgios Theodoridis and Athanasios Tsadiras
Information 2025, 16(7), 596; https://doi.org/10.3390/info16070596 - 11 Jul 2025
Viewed by 314
Abstract
Accurate retail demand forecasting is integral to the operational efficiency of any retail business. As demand is described over time, the prediction of demand is a time-series forecasting problem which may be addressed in a univariate manner, via statistical methods and simplistic machine [...] Read more.
Accurate retail demand forecasting is integral to the operational efficiency of any retail business. As demand is described over time, the prediction of demand is a time-series forecasting problem which may be addressed in a univariate manner, via statistical methods and simplistic machine learning approaches, or in a multivariate fashion using generic deep learning forecasters that are well-established in other fields. This study analyzes, optimizes, trains and tests such forecasters, namely the Temporal Fusion Transformer and the Temporal Convolutional Network, alongside the recently proposed Time-Series Mixer, to accurately forecast retail demand given a dataset of historical sales in 45 stores with their accompanied features. Moreover, the present work proposes a novel extension of the Time-Series Mixer architecture, the LSTMixer, which utilizes an additional Long Short-Term Memory block to achieve better forecasts. The results indicate that the proposed LSTMixer model is the better predictor, whilst all the other aforementioned models outperform the common statistical and machine learning methods. An ablation test is also performed to ensure that the extension within the LSTMixer design is responsible for the improved results. The findings promote the use of deep learning models for retail demand forecasting problems and establish LSTMixer as a viable and efficient option. Full article
(This article belongs to the Special Issue Artificial Intelligence (AI) for Economics and Business Management)
Show Figures

Figure 1

Back to TopTop