Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,327)

Search Parameters:
Keywords = recurrent neural network (RNN)

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
30 pages, 1142 KiB  
Review
Beyond the Backbone: A Quantitative Review of Deep-Learning Architectures for Tropical Cyclone Track Forecasting
by He Huang, Difei Deng, Liang Hu, Yawen Chen and Nan Sun
Remote Sens. 2025, 17(15), 2675; https://doi.org/10.3390/rs17152675 - 2 Aug 2025
Viewed by 119
Abstract
Accurate forecasting of tropical cyclone (TC) tracks is critical for disaster preparedness and risk mitigation. While traditional numerical weather prediction (NWP) systems have long served as the backbone of operational forecasting, they face limitations in computational cost and sensitivity to initial conditions. In [...] Read more.
Accurate forecasting of tropical cyclone (TC) tracks is critical for disaster preparedness and risk mitigation. While traditional numerical weather prediction (NWP) systems have long served as the backbone of operational forecasting, they face limitations in computational cost and sensitivity to initial conditions. In recent years, deep learning (DL) has emerged as a promising alternative, offering data-driven modeling capabilities for capturing nonlinear spatiotemporal patterns. This paper presents a comprehensive review of DL-based approaches for TC track forecasting. We categorize all DL-based TC tracking models according to the architecture, including recurrent neural networks (RNNs), convolutional neural networks (CNNs), Transformers, graph neural networks (GNNs), generative models, and Fourier-based operators. To enable rigorous performance comparison, we introduce a Unified Geodesic Distance Error (UGDE) metric that standardizes evaluation across diverse studies and lead times. Based on this metric, we conduct a critical comparison of state-of-the-art models and identify key insights into their relative strengths, limitations, and suitable application scenarios. Building on this framework, we conduct a critical cross-model analysis that reveals key trends, performance disparities, and architectural tradeoffs. Our analysis also highlights several persistent challenges, such as long-term forecast degradation, limited physical integration, and generalization to extreme events, pointing toward future directions for developing more robust and operationally viable DL models for TC track forecasting. To support reproducibility and facilitate standardized evaluation, we release an open-source UGDE conversion tool on GitHub. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

25 pages, 2515 KiB  
Article
Solar Agro Savior: Smart Agricultural Monitoring Using Drones and Deep Learning Techniques
by Manu Mundappat Ramachandran, Bisni Fahad Mon, Mohammad Hayajneh, Najah Abu Ali and Elarbi Badidi
Agriculture 2025, 15(15), 1656; https://doi.org/10.3390/agriculture15151656 - 1 Aug 2025
Viewed by 233
Abstract
The Solar Agro Savior (SAS) is an innovative solution that is assisted by drones for the sustainable utilization of water and plant disease observation in the agriculture sector. This system integrates an alerting mechanism for humidity, moisture, and temperature variations, which affect the [...] Read more.
The Solar Agro Savior (SAS) is an innovative solution that is assisted by drones for the sustainable utilization of water and plant disease observation in the agriculture sector. This system integrates an alerting mechanism for humidity, moisture, and temperature variations, which affect the plants’ health and optimization in water utilization, which enhances plant yield productivity. A significant feature of the system is the efficient monitoring system in a larger region through drones’ high-resolution cameras, which enables real-time, efficient response and alerting for environmental fluctuations to the authorities. The machine learning algorithm, particularly recurrent neural networks, which is a pioneer with agriculture and pest control, is incorporated for intelligent monitoring systems. The proposed system incorporates a specialized form of a recurrent neural network, Long Short-Term Memory (LSTM), which effectively addresses the vanishing gradient problem. It also utilizes an attention-based mechanism that enables the model to assign meaningful weights to the most important parts of the data sequence. This algorithm not only enhances water utilization efficiency but also boosts plant yield and strengthens pest control mechanisms. This system also provides sustainability through the re-utilization of water and the elimination of electric energy through solar panel systems for powering the inbuilt irrigation system. A comparative analysis of variant algorithms in the agriculture sector with a machine learning approach was also illustrated, and the proposed system yielded 99% yield accuracy, a 97.8% precision value, 98.4% recall, and a 98.4% F1 score value. By encompassing solar irrigation and artificial intelligence-driven analysis, the proposed algorithm, Solar Argo Savior, established a sustainable framework in the latest agricultural sectors and promoted sustainability to protect our environment and community. Full article
(This article belongs to the Section Agricultural Technology)
Show Figures

Figure 1

21 pages, 4147 KiB  
Article
OLTEM: Lumped Thermal and Deep Neural Model for PMSM Temperature
by Yuzhong Sheng, Xin Liu, Qi Chen, Zhenghao Zhu, Chuangxin Huang and Qiuliang Wang
AI 2025, 6(8), 173; https://doi.org/10.3390/ai6080173 - 31 Jul 2025
Viewed by 235
Abstract
Background and Objective: Temperature management is key for reliable operation of permanent magnet synchronous motors (PMSMs). The lumped-parameter thermal network (LPTN) is fast and interpretable but struggles with nonlinear behavior under high power density. We propose OLTEM, a physics-informed deep model that combines [...] Read more.
Background and Objective: Temperature management is key for reliable operation of permanent magnet synchronous motors (PMSMs). The lumped-parameter thermal network (LPTN) is fast and interpretable but struggles with nonlinear behavior under high power density. We propose OLTEM, a physics-informed deep model that combines LPTN with a thermal neural network (TNN) to improve prediction accuracy while keeping physical meaning. Methods: OLTEM embeds LPTN into a recurrent state-space formulation and learns three parameter sets: thermal conductance, inverse thermal capacitance, and power loss. Two additions are introduced: (i) a state-conditioned squeeze-and-excitation (SC-SE) attention that adapts feature weights using the current temperature state, and (ii) an enhanced power-loss sub-network that uses a deep MLP with SC-SE and non-negativity constraints. The model is trained and evaluated on the public Electric Motor Temperature dataset (Paderborn University/Kaggle). Performance is measured by mean squared error (MSE) and maximum absolute error across permanent-magnet, stator-yoke, stator-tooth, and stator-winding temperatures. Results: OLTEM tracks fast thermal transients and yields lower MSE than both the baseline TNN and a CNN–RNN model for all four components. On a held-out generalization set, MSE remains below 4.0 °C2 and the maximum absolute error is about 4.3–8.2 °C. Ablation shows that removing either SC-SE or the enhanced power-loss module degrades accuracy, confirming their complementary roles. Conclusions: By combining physics with learned attention and loss modeling, OLTEM improves PMSM temperature prediction while preserving interpretability. This approach can support motor thermal design and control; future work will study transfer to other machines and further reduce short-term errors during abrupt operating changes. Full article
Show Figures

Figure 1

20 pages, 1346 KiB  
Article
Integrated Smart Farm System Using RNN-Based Supply Scheduling and UAV Path Planning
by Dongwoo You, Yukai Chen and Donkyu Baek
Drones 2025, 9(8), 531; https://doi.org/10.3390/drones9080531 - 28 Jul 2025
Viewed by 337
Abstract
Smart farming has emerged as a promising solution to address challenges such as climate change, population growth, and limited agricultural infrastructure. To enhance the operational efficiency of smart farms, this paper proposes an integrated system that combines Recurrent Neural Networks (RNNs) and Unmanned [...] Read more.
Smart farming has emerged as a promising solution to address challenges such as climate change, population growth, and limited agricultural infrastructure. To enhance the operational efficiency of smart farms, this paper proposes an integrated system that combines Recurrent Neural Networks (RNNs) and Unmanned Aerial Vehicles (UAVs). The proposed framework forecasts future resource shortages using an RNN model and recent environmental data collected from the field. Based on these forecasts, the system schedules a resource supply plan and determines the UAV path by considering both dynamic energy consumption and priority levels, aiming to maximize the efficiency of the resource supply. Experimental results show that the proposed integrated smart farm framework achieves an average reduction of 81.08% in the supply miss rate. This paper demonstrates the potential of an integrated AI- and UAV-based smart farm management system in achieving both environmental responsiveness and operational optimization. Full article
(This article belongs to the Section Drones in Agriculture and Forestry)
Show Figures

Figure 1

24 pages, 2815 KiB  
Article
Blockchain-Powered LSTM-Attention Hybrid Model for Device Situation Awareness and On-Chain Anomaly Detection
by Qiang Zhang, Caiqing Yue, Xingzhe Dong, Guoyu Du and Dongyu Wang
Sensors 2025, 25(15), 4663; https://doi.org/10.3390/s25154663 - 28 Jul 2025
Viewed by 263
Abstract
With the increasing scale of industrial devices and the growing complexity of multi-source heterogeneous sensor data, traditional methods struggle to address challenges in fault detection, data security, and trustworthiness. Ensuring tamper-proof data storage and improving prediction accuracy for imbalanced anomaly detection for potential [...] Read more.
With the increasing scale of industrial devices and the growing complexity of multi-source heterogeneous sensor data, traditional methods struggle to address challenges in fault detection, data security, and trustworthiness. Ensuring tamper-proof data storage and improving prediction accuracy for imbalanced anomaly detection for potential deployment in the Industrial Internet of Things (IIoT) remain critical issues. This study proposes a blockchain-powered Long Short-Term Memory Network (LSTM)–Attention hybrid model: an LSTM-based Encoder–Attention–Decoder (LEAD) for industrial device anomaly detection. The model utilizes an encoder–attention–decoder architecture for processing multivariate time series data generated by industrial sensors and smart contracts for automated on-chain data verification and tampering alerts. Experiments on real-world datasets demonstrate that the LEAD achieves an F0.1 score of 0.96, outperforming baseline models (Recurrent Neural Network (RNN): 0.90; LSTM: 0.94; and Bi-directional LSTM (Bi-LSTM, 0.94)). We simulate the system using a private FISCO-BCOS network with a multi-node setup to demonstrate contract execution, anomaly data upload, and tamper alert triggering. The blockchain system successfully detects unauthorized access and data tampering, offering a scalable solution for device monitoring. Full article
(This article belongs to the Section Internet of Things)
Show Figures

Figure 1

26 pages, 5325 KiB  
Article
Spatiotemporal Dengue Forecasting for Sustainable Public Health in Bandung, Indonesia: A Comparative Study of Classical, Machine Learning, and Bayesian Models
by I Gede Nyoman Mindra Jaya, Yudhie Andriyana, Bertho Tantular, Sinta Septi Pangastuti and Farah Kristiani
Sustainability 2025, 17(15), 6777; https://doi.org/10.3390/su17156777 - 25 Jul 2025
Viewed by 368
Abstract
Accurate dengue forecasting is essential for sustainable public health planning, especially in tropical regions where the disease remains a persistent threat. This study evaluates the predictive performance of seven modeling approaches—Seasonal Autoregressive Integrated Moving Average (SARIMA), Extreme Gradient Boosting (XGBoost), Recurrent Neural Network [...] Read more.
Accurate dengue forecasting is essential for sustainable public health planning, especially in tropical regions where the disease remains a persistent threat. This study evaluates the predictive performance of seven modeling approaches—Seasonal Autoregressive Integrated Moving Average (SARIMA), Extreme Gradient Boosting (XGBoost), Recurrent Neural Network (RNN), Long Short-Term Memory (LSTM), Bidirectional LSTM (BiLSTM), Convolutional LSTM (CNN–LSTM), and a Bayesian spatiotemporal model—using monthly dengue incidence data from 2009 to 2023 in Bandung City, Indonesia. Model performance was assessed using MAE, sMAPE, RMSE, and Pearson’s correlation (R). Among all models, the Bayesian spatiotemporal model achieved the best performance, with the lowest MAE (5.543), sMAPE (62.137), and RMSE (7.482), and the highest R (0.723). While SARIMA and XGBoost showed signs of overfitting, the Bayesian model not only delivered more accurate forecasts but also produced spatial risk estimates and identified high-risk hotspots via exceedance probabilities. These features make it particularly valuable for developing early warning systems and guiding targeted public health interventions, supporting the broader goals of sustainable disease management. Full article
(This article belongs to the Section Health, Well-Being and Sustainability)
Show Figures

Figure 1

30 pages, 9222 KiB  
Article
Using Deep Learning in Forecasting the Production of Electricity from Photovoltaic and Wind Farms
by Michał Pikus, Jarosław Wąs and Agata Kozina
Energies 2025, 18(15), 3913; https://doi.org/10.3390/en18153913 - 23 Jul 2025
Viewed by 306
Abstract
Accurate forecasting of electricity production is crucial for the stability of the entire energy sector. However, predicting future renewable energy production and its value is difficult due to the complex processes that affect production using renewable energy sources. In this article, we examine [...] Read more.
Accurate forecasting of electricity production is crucial for the stability of the entire energy sector. However, predicting future renewable energy production and its value is difficult due to the complex processes that affect production using renewable energy sources. In this article, we examine the performance of basic deep learning models for electricity forecasting. We designed deep learning models, including recursive neural networks (RNNs), which are mainly based on long short-term memory (LSTM) networks; gated recurrent units (GRUs), convolutional neural networks (CNNs), temporal fusion transforms (TFTs), and combined architectures. In order to achieve this goal, we have created our benchmarks and used tools that automatically select network architectures and parameters. Data were obtained as part of the NCBR grant (the National Center for Research and Development, Poland). These data contain daily records of all the recorded parameters from individual solar and wind farms over the past three years. The experimental results indicate that the LSTM models significantly outperformed the other models in terms of forecasting. In this paper, multilayer deep neural network (DNN) architectures are described, and the results are provided for all the methods. This publication is based on the results obtained within the framework of the research and development project “POIR.01.01.01-00-0506/21”, realized in the years 2022–2023. The project was co-financed by the European Union under the Smart Growth Operational Programme 2014–2020. Full article
Show Figures

Figure 1

24 pages, 3714 KiB  
Article
DTCMMA: Efficient Wind-Power Forecasting Based on Dimensional Transformation Combined with Multidimensional and Multiscale Convolutional Attention Mechanism
by Wenhan Song, Enguang Zuo, Junyu Zhu, Chen Chen, Cheng Chen, Ziwei Yan and Xiaoyi Lv
Sensors 2025, 25(15), 4530; https://doi.org/10.3390/s25154530 - 22 Jul 2025
Viewed by 266
Abstract
With the growing global demand for clean energy, the accuracy of wind-power forecasting plays a vital role in ensuring the stable operation of power systems. However, wind-power generation is significantly influenced by meteorological conditions and is characterized by high uncertainty and multiscale fluctuations. [...] Read more.
With the growing global demand for clean energy, the accuracy of wind-power forecasting plays a vital role in ensuring the stable operation of power systems. However, wind-power generation is significantly influenced by meteorological conditions and is characterized by high uncertainty and multiscale fluctuations. Traditional recurrent neural network (RNN) and long short-term memory (LSTM) models, although capable of handling sequential data, struggle with modeling long-term temporal dependencies due to the vanishing gradient problem; thus, they are now rarely used. Recently, Transformer models have made notable progress in sequence modeling compared to RNNs and LSTM models. Nevertheless, when dealing with long wind-power sequences, their quadratic computational complexity (O(L2)) leads to low efficiency, and their global attention mechanism often fails to capture local periodic features accurately, tending to overemphasize redundant information while overlooking key temporal patterns. To address these challenges, this paper proposes a wind-power forecasting method based on dimension-transformed collaborative multidimensional multiscale attention (DTCMMA). This method first employs fast Fourier transform (FFT) to automatically identify the main periodic components in wind-power data, reconstructing the one-dimensional time series as a two-dimensional spatiotemporal representation, thereby explicitly encoding periodic features. Based on this, a collaborative multidimensional multiscale attention (CMMA) mechanism is designed, which hierarchically integrates channel, spatial, and pixel attention to adaptively capture complex spatiotemporal dependencies. Considering the geometric characteristics of the reconstructed data, asymmetric convolution kernels are adopted to enhance feature extraction efficiency. Experiments on multiple wind-farm datasets and energy-related datasets demonstrate that DTCMMA outperforms mainstream methods such as Transformer, iTransformer, and TimeMixer in long-sequence forecasting tasks, achieving improvements in MSE performance by 34.22%, 2.57%, and 0.51%, respectively. The model’s training speed also surpasses that of the fastest baseline by 300%, significantly improving both prediction accuracy and computational efficiency. This provides an efficient and accurate solution for wind-power forecasting and contributes to the further development and application of wind energy in the global energy mix. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Figure 1

31 pages, 7723 KiB  
Article
A Hybrid CNN–GRU–LSTM Algorithm with SHAP-Based Interpretability for EEG-Based ADHD Diagnosis
by Makbal Baibulova, Murat Aitimov, Roza Burganova, Lazzat Abdykerimova, Umida Sabirova, Zhanat Seitakhmetova, Gulsiya Uvaliyeva, Maksym Orynbassar, Aislu Kassekeyeva and Murizah Kassim
Algorithms 2025, 18(8), 453; https://doi.org/10.3390/a18080453 - 22 Jul 2025
Viewed by 462
Abstract
This study proposes an interpretable hybrid deep learning framework for classifying attention deficit hyperactivity disorder (ADHD) using EEG signals recorded during cognitively demanding tasks. The core architecture integrates convolutional neural networks (CNNs), gated recurrent units (GRUs), and long short-term memory (LSTM) layers to [...] Read more.
This study proposes an interpretable hybrid deep learning framework for classifying attention deficit hyperactivity disorder (ADHD) using EEG signals recorded during cognitively demanding tasks. The core architecture integrates convolutional neural networks (CNNs), gated recurrent units (GRUs), and long short-term memory (LSTM) layers to jointly capture spatial and temporal dynamics. In addition to the final hybrid architecture, the CNN–GRU–LSTM model alone demonstrates excellent accuracy (99.63%) with minimal variance, making it a strong baseline for clinical applications. To evaluate the role of global attention mechanisms, transformer encoder models with two and three attention blocks, along with a spatiotemporal transformer employing 2D positional encoding, are benchmarked. A hybrid CNN–RNN–transformer model is introduced, combining convolutional, recurrent, and transformer-based modules into a unified architecture. To enhance interpretability, SHapley Additive exPlanations (SHAP) are employed to identify key EEG channels contributing to classification outcomes. Experimental evaluation using stratified five-fold cross-validation demonstrates that the proposed hybrid model achieves superior performance, with average accuracy exceeding 99.98%, F1-scores above 0.9999, and near-perfect AUC and Matthews correlation coefficients. In contrast, transformer-only models, despite high training accuracy, exhibit reduced generalization. SHAP-based analysis confirms the hybrid model’s clinical relevance. This work advances the development of transparent and reliable EEG-based tools for pediatric ADHD screening. Full article
Show Figures

Graphical abstract

27 pages, 3019 KiB  
Article
New Deep Learning-Based Approach for Source Code Generation: Application to Computer Vision Systems
by Wafa Alshehri, Salma Kammoun Jarraya and Arwa Allinjawi
AI 2025, 6(7), 162; https://doi.org/10.3390/ai6070162 - 21 Jul 2025
Viewed by 499
Abstract
Deep learning has enabled significant progress in source code generation, aiming to reduce the manual, error-prone, and time-consuming aspects of software development. While many existing models rely on recurrent neural networks (RNNs) with sequence-to-sequence architectures, these approaches struggle with the long and complex [...] Read more.
Deep learning has enabled significant progress in source code generation, aiming to reduce the manual, error-prone, and time-consuming aspects of software development. While many existing models rely on recurrent neural networks (RNNs) with sequence-to-sequence architectures, these approaches struggle with the long and complex token sequences typical in source code. To address this, we propose a grammar-based convolutional neural network (CNN) combined with a tree-based representation to enhance accuracy and efficiency. Our model achieves state-of-the-art results on the benchmark HEARTHSTONE dataset, with a BLEU score of 81.4 and an Acc+ of 62.1%. We further evaluate the model on our proposed dataset, AST2CVCode, designed for computer vision applications, achieving 86.2 BLEU and 51.9% EM. Additionally, we introduce BLEU+, an enhanced evaluation metric tailored for functional correctness in code generation, which achieves a BLEU+ score of 92.0% on the AST2CVCode dataset. These results demonstrate the effectiveness of our approach in both model architecture and evaluation methodology. Full article
(This article belongs to the Section AI Systems: Theory and Applications)
Show Figures

Figure 1

16 pages, 1251 KiB  
Article
Enhanced Detection of Intrusion Detection System in Cloud Networks Using Time-Aware and Deep Learning Techniques
by Nima Terawi, Huthaifa I. Ashqar, Omar Darwish, Anas Alsobeh, Plamen Zahariev and Yahya Tashtoush
Computers 2025, 14(7), 282; https://doi.org/10.3390/computers14070282 - 17 Jul 2025
Viewed by 341
Abstract
This study introduces an enhanced Intrusion Detection System (IDS) framework for Denial-of-Service (DoS) attacks, utilizing network traffic inter-arrival time (IAT) analysis. By examining the timing between packets and other statistical features, we detected patterns of malicious activity, allowing early and effective DoS threat [...] Read more.
This study introduces an enhanced Intrusion Detection System (IDS) framework for Denial-of-Service (DoS) attacks, utilizing network traffic inter-arrival time (IAT) analysis. By examining the timing between packets and other statistical features, we detected patterns of malicious activity, allowing early and effective DoS threat mitigation. We generate real DoS traffic, including normal, Internet Control Message Protocol (ICMP), Smurf attack, and Transmission Control Protocol (TCP) classes, and develop nine predictive algorithms, combining traditional machine learning and advanced deep learning techniques with optimization methods, including the synthetic minority sampling technique (SMOTE) and grid search (GS). Our findings reveal that while traditional machine learning achieved moderate accuracy, it struggled with imbalanced datasets. In contrast, Deep Neural Network (DNN) models showed significant improvements with optimization, with DNN combined with GS (DNN-GS) reaching 89% accuracy. However, we also used Recurrent Neural Networks (RNNs) combined with SMOTE and GS (RNN-SMOTE-GS), which emerged as the best-performing with a precision of 97%, demonstrating the effectiveness of combining SMOTE and GS and highlighting the critical role of advanced optimization techniques in enhancing the detection capabilities of IDS models for the accurate classification of various types of network traffic and attacks. Full article
Show Figures

Figure 1

17 pages, 2117 KiB  
Article
On-Orbit Life Prediction and Analysis of Triple-Junction Gallium Arsenide Solar Arrays for MEO Satellites
by Huan Liu, Chenjie Kong, Yuan Shen, Baojun Lin, Xueliang Wang and Qiang Zhang
Aerospace 2025, 12(7), 633; https://doi.org/10.3390/aerospace12070633 - 16 Jul 2025
Viewed by 260
Abstract
This paper focuses on the triple-junction gallium arsenide solar array of a MEO (Medium Earth Orbit) satellite that has been in orbit for 7 years. Through a combination of theoretical and data-driven methods, it conducts research on anti-radiation design verification and life prediction. [...] Read more.
This paper focuses on the triple-junction gallium arsenide solar array of a MEO (Medium Earth Orbit) satellite that has been in orbit for 7 years. Through a combination of theoretical and data-driven methods, it conducts research on anti-radiation design verification and life prediction. This study integrates the Long Short-Term Memory (LSTM) algorithm into the full life cycle management of MEO satellite solar arrays, providing a solution that combines theory and engineering for the design of high-reliability energy systems. Based on semiconductor physics theory, this paper establishes an output current calculation model. By combining radiation attenuation factors obtained from ground experiments, it derives the theoretical current values for the initial orbit insertion and the end of life. Aiming at the limitations of traditional physical models in addressing solar performance degradation under complex radiation environments, this paper introduces an LSTM algorithm to deeply mine the high-density current telemetry data (approximately 30 min per point) accumulated over 7 years in orbit. By comparing the prediction accuracy of LSTM with traditional models such as Recurrent Neural Network (RNN) and Feedforward Neural Network (FNN), the significant advantage of LSTM in capturing the long-term attenuation trend of solar arrays is verified. This study integrates deep learning technology into the full life cycle management of solar arrays, constructs a closed-loop verification system of “theoretical modeling–data-driven intelligent prediction”, and provides a solution for the long-life and high-reliability operation of the energy system of MEO orbit satellites. Full article
Show Figures

Figure 1

20 pages, 5700 KiB  
Article
Multimodal Personality Recognition Using Self-Attention-Based Fusion of Audio, Visual, and Text Features
by Hyeonuk Bhin and Jongsuk Choi
Electronics 2025, 14(14), 2837; https://doi.org/10.3390/electronics14142837 - 15 Jul 2025
Viewed by 458
Abstract
Personality is a fundamental psychological trait that exerts a long-term influence on human behavior patterns and social interactions. Automatic personality recognition (APR) has exhibited increasing importance across various domains, including Human–Robot Interaction (HRI), personalized services, and psychological assessments. In this study, we propose [...] Read more.
Personality is a fundamental psychological trait that exerts a long-term influence on human behavior patterns and social interactions. Automatic personality recognition (APR) has exhibited increasing importance across various domains, including Human–Robot Interaction (HRI), personalized services, and psychological assessments. In this study, we propose a multimodal personality recognition model that classifies the Big Five personality traits by extracting features from three heterogeneous sources: audio processed using Wav2Vec2, video represented as Skeleton Landmark time series, and text encoded through Bidirectional Encoder Representations from Transformers (BERT) and Doc2Vec embeddings. Each modality is handled through an independent Self-Attention block that highlights salient temporal information, and these representations are then summarized and integrated using a late fusion approach to effectively reflect both the inter-modal complementarity and cross-modal interactions. Compared to traditional recurrent neural network (RNN)-based multimodal models and unimodal classifiers, the proposed model achieves an improvement of up to 12 percent in the F1-score. It also maintains a high prediction accuracy and robustness under limited input conditions. Furthermore, a visualization based on t-distributed Stochastic Neighbor Embedding (t-SNE) demonstrates clear distributional separation across the personality classes, enhancing the interpretability of the model and providing insights into the structural characteristics of its latent representations. To support real-time deployment, a lightweight thread-based processing architecture is implemented, ensuring computational efficiency. By leveraging deep learning-based feature extraction and the Self-Attention mechanism, we present a novel personality recognition framework that balances performance with interpretability. The proposed approach establishes a strong foundation for practical applications in HRI, counseling, education, and other interactive systems that require personalized adaptation. Full article
(This article belongs to the Special Issue Explainable Machine Learning and Data Mining)
Show Figures

Figure 1

24 pages, 26672 KiB  
Article
Short-Term Electric Load Forecasting Using Deep Learning: A Case Study in Greece with RNN, LSTM, and GRU Networks
by Vasileios Zelios, Paris Mastorocostas, George Kandilogiannakis, Anastasios Kesidis, Panagiota Tselenti and Athanasios Voulodimos
Electronics 2025, 14(14), 2820; https://doi.org/10.3390/electronics14142820 - 14 Jul 2025
Viewed by 585
Abstract
The increasing volatility in energy markets, particularly in Greece where electricity costs reached a peak of 236 EUR/MWh in 2022, underscores the urgent need for accurate short-term load forecasting models. In this study, the application of deep learning techniques, specifically Recurrent Neural Network [...] Read more.
The increasing volatility in energy markets, particularly in Greece where electricity costs reached a peak of 236 EUR/MWh in 2022, underscores the urgent need for accurate short-term load forecasting models. In this study, the application of deep learning techniques, specifically Recurrent Neural Network (RNN), Long Short-Term Memory (LSTM), and Gated Recurrent Unit (GRU), to forecast hourly electricity demand is investigated. The proposed models were trained on historical load data from the Greek power system spanning the years 2013 to 2016. Various deep learning architectures were implemented and their forecasting performances using statistical metrics such as Root Mean Squared Error (RMSE) and Mean Absolute Percentage Error (MAPE) were evaluated. The experiments utilized multiple time horizons (1 h, 2 h, 24 h) and input sequence lengths (6 h to 168 h) to assess model accuracy and robustness. The best performing GRU model achieved an RMSE of 83.2 MWh and a MAPE of 1.17% for 1 h ahead forecasting, outperforming both LSTM and RNN in terms of both accuracy and computational efficiency. The predicted values were integrated into a dynamic Power BI dashboard, to enable real-time visualization and decision support. These findings demonstrate the potential of deep learning architectures, particularly GRUs, for operational load forecasting and their applicability to intelligent energy systems in a market-strained environment. Full article
Show Figures

Figure 1

16 pages, 1351 KiB  
Article
A Comparative Study on Machine Learning Methods for EEG-Based Human Emotion Recognition
by Shokoufeh Davarzani, Simin Masihi, Masoud Panahi, Abdulrahman Olalekan Yusuf and Massood Atashbar
Electronics 2025, 14(14), 2744; https://doi.org/10.3390/electronics14142744 - 8 Jul 2025
Viewed by 460
Abstract
Electroencephalogram (EEG) signals provide a direct and non-invasive means of interpreting brain activity and are increasingly becoming valuable in embedded emotion-aware systems, particularly for applications in healthcare, wearable electronics, and human–machine interactions. Among various EEG-based emotion recognition techniques, deep learning methods have demonstrated [...] Read more.
Electroencephalogram (EEG) signals provide a direct and non-invasive means of interpreting brain activity and are increasingly becoming valuable in embedded emotion-aware systems, particularly for applications in healthcare, wearable electronics, and human–machine interactions. Among various EEG-based emotion recognition techniques, deep learning methods have demonstrated superior performance compared to traditional approaches. This advantage stems from their ability to extract complex features—such as spectral–spatial connectivity, temporal dynamics, and non-linear patterns—from raw EEG data, leading to a more accurate and robust representation of emotional states and better adaptation to diverse data characteristics. This study explores and compares deep and shallow neural networks for human emotion recognition from raw EEG data, with the goal of enabling real-time processing in embedded and edge-deployable systems. Deep learning models—specifically convolutional neural networks (CNNs) and recurrent neural networks (RNNs)—have been benchmarked against traditional approaches such as the multi-layer perceptron (MLP), support vector machine (SVM), and k-nearest neighbors (kNN) algorithms. This comparative study investigates the effectiveness of deep learning techniques in EEG-based emotion recognition by classifying emotions into four categories based on the valence–arousal plane: high arousal, positive valence (HAPV); low arousal, positive valence (LAPV); high arousal, negative valence (HANV); and low arousal, negative valence (LANV). Evaluations were conducted using the DEAP dataset. The results indicate that both the CNN and RNN-STM models have a high classification performance in EEG-based emotion recognition, with an average accuracy of 90.13% and 93.36%, respectively, significantly outperforming shallow algorithms (MLP, SVM, kNN). Full article
(This article belongs to the Special Issue New Advances in Embedded Software and Applications)
Show Figures

Figure 1

Back to TopTop