Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (66)

Search Parameters:
Keywords = bidirectional long short-term memory (BLSTM)

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 4452 KiB  
Article
Upper Limb Joint Angle Estimation Using a Reduced Number of IMU Sensors and Recurrent Neural Networks
by Kevin Niño-Tejada, Laura Saldaña-Aristizábal, Jhonathan L. Rivas-Caicedo and Juan F. Patarroyo-Montenegro
Electronics 2025, 14(15), 3039; https://doi.org/10.3390/electronics14153039 - 30 Jul 2025
Abstract
Accurate estimation of upper-limb joint angles is essential in biomechanics, rehabilitation, and wearable robotics. While inertial measurement units (IMUs) offer portability and flexibility, systems requiring multiple inertial sensors can be intrusive and complex to deploy. In contrast, optical motion capture (MoCap) systems provide [...] Read more.
Accurate estimation of upper-limb joint angles is essential in biomechanics, rehabilitation, and wearable robotics. While inertial measurement units (IMUs) offer portability and flexibility, systems requiring multiple inertial sensors can be intrusive and complex to deploy. In contrast, optical motion capture (MoCap) systems provide precise tracking but are constrained to controlled laboratory environments. This study presents a deep learning-based approach for estimating shoulder and elbow joint angles using only three IMU sensors positioned on the chest and both wrists, validated against reference angles obtained from a MoCap system. The input data includes Euler angles, accelerometer, and gyroscope data, synchronized and segmented into sliding windows. Two recurrent neural network architectures, Convolutional Neural Network with Long-short Term Memory (CNN-LSTM) and Bidirectional LSTM (BLSTM), were trained and evaluated using identical conditions. The CNN component enabled the LSTM to extract spatial features that enhance sequential pattern learning, improving angle reconstruction. Both models achieved accurate estimation performance: CNN-LSTM yielded lower Mean Absolute Error (MAE) in smooth trajectories, while BLSTM provided smoother predictions but underestimated some peak movements, especially in the primary axes of rotation. These findings support the development of scalable, deep learning-based wearable systems and contribute to future applications in clinical assessment, sports performance analysis, and human motion research. Full article
(This article belongs to the Special Issue Wearable Sensors for Human Position, Attitude and Motion Tracking)
Show Figures

Figure 1

12 pages, 351 KiB  
Article
A Combined Method for Localizing Two Overlapping Acoustic Sources Based on Deep Learning
by Alexander Lyapin, Ghiath Shahoud and Evgeny Agafonov
Appl. Sci. 2025, 15(12), 6768; https://doi.org/10.3390/app15126768 - 16 Jun 2025
Viewed by 437
Abstract
Deep learning approaches for multi-source sound localization face significant challenges, particularly the need for extensive training datasets encompassing diverse spatial configurations to achieve robust generalization. This requirement leads to substantial computational demands, which are further exacerbated when localizing overlapping sources in complex acoustic [...] Read more.
Deep learning approaches for multi-source sound localization face significant challenges, particularly the need for extensive training datasets encompassing diverse spatial configurations to achieve robust generalization. This requirement leads to substantial computational demands, which are further exacerbated when localizing overlapping sources in complex acoustic environments with reverberation and noise. In this paper, a new methodology is proposed for simultaneous localization of two overlapping sound sources in the time–frequency domain in a closed, reverberant environment with a spatial resolution of 10° using a small-sized microphone array. The proposed methodology is based on the integration of the sound source separation method with a single-source sound localization model. A hybrid model was proposed to separate the sound source signals received by each microphone in the array. The model was built using a bidirectional long short-term memory (BLSTM) network and trained on a dataset using the ideal binary mask (IBM) as the training target. The modeling results show that the proposed localization methodology is efficient in determining the directions for two overlapping sources simultaneously, with an average localization accuracy of 86.1% for the test dataset containing short-term signals of 500 ms duration with different signal-to-signal ratio values. Full article
(This article belongs to the Section Acoustics and Vibrations)
Show Figures

Figure 1

23 pages, 3006 KiB  
Article
Enhancing Upper Limb Exoskeletons Using Sensor-Based Deep Learning Torque Prediction and PID Control
by Farshad Shakeriaski and Masoud Mohammadian
Sensors 2025, 25(11), 3528; https://doi.org/10.3390/s25113528 - 3 Jun 2025
Viewed by 648
Abstract
Upper limb assistive exoskeletons help stroke patients by assisting arm movement in impaired individuals. However, effective control of these systems to help stroke survivors is a complex task. In this paper, a novel approach is proposed to enhance the control of upper limb [...] Read more.
Upper limb assistive exoskeletons help stroke patients by assisting arm movement in impaired individuals. However, effective control of these systems to help stroke survivors is a complex task. In this paper, a novel approach is proposed to enhance the control of upper limb assistive exoskeletons by using torque estimation and prediction in a proportional–integral–derivative (PID) controller loop to more optimally integrate the torque of the exoskeleton robot, which aims to eliminate system uncertainties. First, a model for torque estimation from Electromyography (EMG) signals and a predictive torque model for the upper limb exoskeleton robot for the elbow are trained. The trained data consisted of two-dimensional high-density surface EMG (HD-sEMG) signals to record myoelectric activity from five upper limb muscles (biceps brachii, triceps brachii, anconeus, brachioradialis, and pronator teres) during voluntary isometric contractions for twelve healthy subjects performing four different isometric tasks (supination/pronation and elbow flexion/extension) for one minute each, which were trained on long short-term memory (LSTM), bidirectional LSTM (BLSTM), and gated recurrent units (GRU) deep neural network models. These models estimate and predict torque requirements. Finally, the estimated and predicted torque from the trained network is used online as input to a PID control loop and robot dynamic, which aims to control the robot optimally. The results showed that using the proposed method creates a strong and innovative approach to greater independence and rehabilitation improvement. Full article
Show Figures

Figure 1

16 pages, 1174 KiB  
Article
Natural Language Processing for Aviation Safety: Predicting Injury Levels from Incident Reports in Australia
by Aziida Nanyonga, Keith Joiner, Ugur Turhan and Graham Wild
Modelling 2025, 6(2), 40; https://doi.org/10.3390/modelling6020040 - 28 May 2025
Viewed by 1166
Abstract
This study investigates the application of advanced deep learning models for the classification of aviation safety incidents, focusing on four models: Simple Recurrent Neural Network (sRNN), Gated Recurrent Unit (GRU), Bidirectional Long Short-Term Memory (BLSTM), and DistilBERT. The models were evaluated based on [...] Read more.
This study investigates the application of advanced deep learning models for the classification of aviation safety incidents, focusing on four models: Simple Recurrent Neural Network (sRNN), Gated Recurrent Unit (GRU), Bidirectional Long Short-Term Memory (BLSTM), and DistilBERT. The models were evaluated based on key performance metrics, including accuracy, precision, recall, and F1-score. DistilBERT achieved perfect performance with an accuracy of 1.00 across all metrics, while BLSTM demonstrated the highest performance among the deep learning models, with an accuracy of 0.9896, followed by GRU (0.9893) and sRNN (0.9887). Class-wise evaluations revealed that DistilBERT excelled across all injury categories, with BLSTM outperforming the other deep learning models, particularly in detecting fatal injuries, achieving a precision of 0.8684 and an F1-score of 0.7952. The study also addressed the challenges of class imbalance by applying class weighting, although the use of more sophisticated techniques, such as focal loss, is recommended for future work. This research highlights the potential of transformer-based models for aviation safety classification and provides a foundation for future research to improve model interpretability and generalizability across diverse datasets. These findings contribute to the growing body of research on applying deep learning techniques to aviation safety and underscore opportunities for further exploration. Full article
Show Figures

Figure 1

20 pages, 3519 KiB  
Article
Attention-Based Hybrid Deep Learning Models for Classifying COVID-19 Genome Sequences
by A. M. Mutawa
AI 2025, 6(1), 4; https://doi.org/10.3390/ai6010004 - 2 Jan 2025
Cited by 3 | Viewed by 1633
Abstract
Background: COVID-19 genetic sequence research is crucial despite immunizations and pandemic control. COVID-19-causing SARS-CoV-2 must be understood genomically for several reasons. New viral strains may resist vaccines. Categorizing genetic sequences helps researchers track changes and assess immunization efficacy. Classifying COVID-19 genome sequences with [...] Read more.
Background: COVID-19 genetic sequence research is crucial despite immunizations and pandemic control. COVID-19-causing SARS-CoV-2 must be understood genomically for several reasons. New viral strains may resist vaccines. Categorizing genetic sequences helps researchers track changes and assess immunization efficacy. Classifying COVID-19 genome sequences with other viruses helps to understand its evolution and interactions with other illnesses. Methods: The proposed study introduces a deep learning-based COVID-19 genomic sequence categorization approach. Attention-based hybrid deep learning (DL) models categorize 1423 COVID-19 and 11,388 other viral genome sequences. An unknown dataset is also used to assess the models. The five models’ accuracy, f1-score, area under the curve (AUC), precision, Matthews correlation coefficient (MCC), and recall are evaluated. Results: The results indicate that the Convolutional neural network (CNN) with Bidirectional long short-term memory (BLSTM) with attention layer (CNN-BLSTM-Att) achieved an accuracy of 99.99%, which outperformed the other models. For external validation, the model shows an accuracy of 99.88%. It reveals that DL-based approaches with an attention layer can accurately classify COVID-19 genomic sequences with a high degree of accuracy. This method might assist in identifying and classifying COVID-19 virus strains in clinical situations. Immunizations have lowered COVID-19 danger, but categorizing its genetic sequences is crucial to global health activities to plan for recurrence or future viral threats. Full article
Show Figures

Graphical abstract

24 pages, 4002 KiB  
Article
An Intelligent Approach for Early and Accurate Predication of Cardiac Disease Using Hybrid Artificial Intelligence Techniques
by Hazrat Bilal, Yibin Tian, Ahmad Ali, Yar Muhammad, Abid Yahya, Basem Abu Izneid and Inam Ullah
Bioengineering 2024, 11(12), 1290; https://doi.org/10.3390/bioengineering11121290 - 19 Dec 2024
Cited by 15 | Viewed by 1821
Abstract
This study proposes a new hybrid machine learning (ML) model for the early and accurate diagnosis of heart disease. The proposed model is a combination of two powerful ensemble ML models, namely ExtraTreeClassifier (ETC) and XGBoost (XGB), resulting in a hybrid model named [...] Read more.
This study proposes a new hybrid machine learning (ML) model for the early and accurate diagnosis of heart disease. The proposed model is a combination of two powerful ensemble ML models, namely ExtraTreeClassifier (ETC) and XGBoost (XGB), resulting in a hybrid model named ETCXGB. At first, all the features of the utilized heart disease dataset were given as input to the ETC model, which processed it by extracting the predicted probabilities and produced an output. The output of the ETC model was then added to the original feature space by producing an enriched feature matrix, which is then used as input for the XGB model. The new feature matrix is used for training the XGB model, which produces the final result that whether a person has cardiac disease or not, resulting in a high diagnosis accuracy for cardiac disease. In addition to the proposed model, three other hybrid DL models, such as convolutional neural network + recurrent neural network (CNN-RNN), convolutional neural network + long short-term memory (CNN-LSTM), and convolutional neural network + bidirectional long short-term memory (CNN-BLSTM), were also investigated. The proposed ETCXGB model improved the prediction accuracy by 3.91%, while CNN-RNN, CNN-LSTM, and CNN-BLSTM enhanced the prediction accuracy by 1.95%, 2.44%, and 2.45%, respectively, for the diagnosis of cardiac disease. The simulation outcomes illustrate that the proposed ETCXGB hybrid ML outperformed the classical ML and DL models in terms of all performance measures. Therefore, using the proposed hybrid ML model for the diagnosis of cardiac disease will help the medical practitioner make an accurate diagnosis of the disease and will help the healthcare society decrease the mortality rate caused by cardiac disease. Full article
Show Figures

Figure 1

14 pages, 1739 KiB  
Article
Older Adult Fall Risk Prediction with Deep Learning and Timed Up and Go (TUG) Test Data
by Josu Maiora, Chloe Rezola-Pardo, Guillermo García, Begoña Sanz and Manuel Graña
Bioengineering 2024, 11(10), 1000; https://doi.org/10.3390/bioengineering11101000 - 5 Oct 2024
Cited by 1 | Viewed by 3070
Abstract
Falls are a major health hazard for older adults; therefore, in the context of an aging population, predicting the risk of a patient suffering falls in the near future is of great impact for health care systems. Currently, the standard prospective fall risk [...] Read more.
Falls are a major health hazard for older adults; therefore, in the context of an aging population, predicting the risk of a patient suffering falls in the near future is of great impact for health care systems. Currently, the standard prospective fall risk assessment instrument relies on a set of clinical and functional mobility assessment tools, one of them being the Timed Up and Go (TUG) test. Recently, wearable inertial measurement units (IMUs) have been proposed to capture motion data that would allow for the building of estimates of fall risk. The hypothesis of this study is that the data gathered from IMU readings while the patient is performing the TUG test can be used to build a predictive model that would provide an estimate of the probability of suffering a fall in the near future, i.e., assessing prospective fall risk. This study applies deep learning convolutional neural networks (CNN) and recurrent neural networks (RNN) to build such predictive models based on features extracted from IMU data acquired during TUG test realizations. Data were obtained from a cohort of 106 older adults wearing wireless IMU sensors with sampling frequencies of 100 Hz while performing the TUG test. The dependent variable is a binary variable that is true if the patient suffered a fall in the six-month follow-up period. This variable was used as the output variable for the supervised training and validations of the deep learning architectures and competing machine learning approaches. A hold-out validation process using 75 subjects for training and 31 subjects for testing was repeated one hundred times to obtain robust estimations of model performances At each repetition, 5-fold cross-validation was carried out to select the best model over the training subset. Best results were achieved by a bidirectional long short-term memory (BLSTM), obtaining an accuracy of 0.83 and AUC of 0.73 with good sensitivity and specificity values. Full article
Show Figures

Figure 1

23 pages, 9001 KiB  
Article
Remaining Useful Life Prediction of Lithium-Ion Batteries Based on Neural Network and Adaptive Unscented Kalman Filter
by Lingtao Wu, Wenhao Guo, Yuben Tang, Youming Sun and Tuanfa Qin
Electronics 2024, 13(13), 2619; https://doi.org/10.3390/electronics13132619 - 4 Jul 2024
Cited by 9 | Viewed by 2222
Abstract
Accurate prediction of remaining useful life (RUL) plays an important role in maintaining the safe and stable operation of Lithium-ion battery management systems. Aiming at the problem of poor prediction stability of a single model, this paper combines the advantages of data-driven and [...] Read more.
Accurate prediction of remaining useful life (RUL) plays an important role in maintaining the safe and stable operation of Lithium-ion battery management systems. Aiming at the problem of poor prediction stability of a single model, this paper combines the advantages of data-driven and model-based methods and proposes a RUL prediction method combining convolutional neural network (CNN), bi-directional long and short-term memory neural network (Bi-LSTM), SE attention mechanism (AM) and adaptive unscented Kalman filter (AUKF). First, three types of indirect features that are highly correlated with RUL decay are selected as inputs to the model to improve the accuracy of RUL prediction. Second, a CNN-BLSTM-AM network is used to further extract, select and fuse the indirect features to form predictive measurements of the identified degradation metrics. In addition, we introduce the AUKF model to increase the uncertainty representation of the RUL prediction. Finally, the method is validated on the NASA dataset and the CALCE dataset and compared with other methods. The experimental results show that the method is able to achieve an accurate estimation of RUL, a minimum RMSE of up to 0.0030, and a minimum MAE of up to 0.0024, which has high estimation accuracy and robustness. Full article
(This article belongs to the Special Issue Energy Storage, Analysis and Battery Usage)
Show Figures

Figure 1

19 pages, 10556 KiB  
Article
Self-Adaptive Server Anomaly Detection Using Ensemble Meta-Reinforcement Learning
by Bao Rong Chang, Hsiu-Fen Tsai and Guan-Ru Chen
Electronics 2024, 13(12), 2348; https://doi.org/10.3390/electronics13122348 - 15 Jun 2024
Cited by 5 | Viewed by 1682
Abstract
As the user’s behavior changes at any time with cloud computing and network services, abnormal server resource utilization traffic will lead to severe service crashes and system downtime. The traditional single anomaly detection model cannot handle the rapid failure prediction ahead. Therefore, this [...] Read more.
As the user’s behavior changes at any time with cloud computing and network services, abnormal server resource utilization traffic will lead to severe service crashes and system downtime. The traditional single anomaly detection model cannot handle the rapid failure prediction ahead. Therefore, this study proposed ensemble learning combined with model-agnostic meta-reinforcement learning called ensemble meta-reinforcement learning (EMRL) to implement self-adaptive server anomaly detection rapidly and precisely, according to the time series of server resource utilization. The proposed ensemble approach combines hidden Markov model (HMM), variational autoencoder (VAE), temporal convolutional autoencoder (TCN-AE), and bidirectional long short-term memory (BLSTM). The EMRL algorithm trains this combination with several tasks to learn the implicit representation of various anomalous traffic, where each task executes trust region policy optimization (TRPO) to quickly adapt the time-varying data distribution and make rapid decisions precisely for an agent response. As a result, our proposed approach can improve the precision of anomaly prediction by 2.4 times and reduce the model deployment speed by 5.8 times on average because a meta-learner can immediately be applied to new tasks. Full article
Show Figures

Figure 1

15 pages, 1505 KiB  
Article
A Deep Learning-Based Framework for Strengthening Cybersecurity in Internet of Health Things (IoHT) Environments
by Sarah A. Algethami and Sultan S. Alshamrani
Appl. Sci. 2024, 14(11), 4729; https://doi.org/10.3390/app14114729 - 30 May 2024
Cited by 11 | Viewed by 2768
Abstract
The increasing use of IoHT devices in healthcare has brought about revolutionary advancements, but it has also exposed some critical vulnerabilities, particularly in cybersecurity. IoHT is characterized by interconnected medical devices sharing sensitive patient data, which amplifies the risk of cyber threats. Therefore, [...] Read more.
The increasing use of IoHT devices in healthcare has brought about revolutionary advancements, but it has also exposed some critical vulnerabilities, particularly in cybersecurity. IoHT is characterized by interconnected medical devices sharing sensitive patient data, which amplifies the risk of cyber threats. Therefore, ensuring healthcare data’s integrity, confidentiality, and availability is essential. This study proposes a hybrid deep learning-based intrusion detection system that uses an Artificial Neural Network (ANN) with Bidirectional Long Short-Term Memory (BLSTM) and Gated Recurrent Unit (GRU) architectures to address critical cybersecurity threats in IoHT. The model was tailored to meet the complex security demands of IoHT and was rigorously tested using the Electronic Control Unit ECU-IoHT dataset. The results are impressive, with the system achieving 100% accuracy, precision, recall, and F1-Score in binary classifications and maintaining exceptional performance in multiclass scenarios. These findings demonstrate the potential of advanced AI methodologies in safeguarding IoHT environments, providing high-fidelity detection while minimizing false positives. Full article
Show Figures

Figure 1

27 pages, 4204 KiB  
Article
Evaluation of EEG Signals by Spectral Peak Methods and Statistical Correlation for Mental State Discrimination Induced by Arithmetic Tasks
by Daniela Andreea Coman, Silviu Ionita and Ioan Lita
Sensors 2024, 24(11), 3316; https://doi.org/10.3390/s24113316 - 22 May 2024
Cited by 2 | Viewed by 2302
Abstract
Bringing out brain activity through the interpretation of EEG signals is a challenging problem that involves combined methods of signal analysis. The issue of classifying mental states induced by arithmetic tasks can be solved through various classification methods, using diverse characteristic parameters of [...] Read more.
Bringing out brain activity through the interpretation of EEG signals is a challenging problem that involves combined methods of signal analysis. The issue of classifying mental states induced by arithmetic tasks can be solved through various classification methods, using diverse characteristic parameters of EEG signals in the time, frequency, and statistical domains. This paper explores the results of an experiment that aimed to highlight arithmetic mental tasks contained in the PhysioNet database, performed on a group of 36 subjects. The majority of publications on this topic deal with machine learning (ML)-based classification methods with supervised learning support vector machine (SVM) algorithms, K-Nearest Neighbor (KNN), Linear Discriminant Analysis (LDA), and Decision Trees (DTs). Also, there are frequent approaches based on the analysis of EEG data as time series and their classification with Recurrent Neural Networks (RNNs), as well as with improved algorithms such as Long Short-Term Memory (LSTM), Bidirectional Long Short-Term Memory (BLSTM), and Gated Recurrent Units (GRUs). In the present work, we evaluate the classification method based on the comparison of domain limits for two specific characteristics of EEG signals: the statistical correlation of pairs of signals and the size of the spectral peak detected in theta, alpha, and beta bands. This study provides some interpretations regarding the electrical activity of the brain, consolidating and complementing the results of similar research. The classification method used is simple and easy to apply and interpret. The analysis of EEG data showed that the theta and beta frequency bands were the only discriminators between the relaxation and arithmetic calculation states. Notably, the F7 signal, which used the spectral peak criterion, achieved the best classification accuracy (100%) in both theta and beta bands for the subjects with the best results in performing calculations. Also, our study found the Fz signal to be a good sensor in the theta band for mental task discrimination for all subjects in the group with 90% accuracy. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

22 pages, 7742 KiB  
Article
Ensemble Meta-Learning-Based Robust Chipping Prediction for Wafer Dicing
by Bao Rong Chang, Hsiu-Fen Tsai and Hsiang-Yu Mo
Electronics 2024, 13(10), 1802; https://doi.org/10.3390/electronics13101802 - 7 May 2024
Cited by 3 | Viewed by 1446
Abstract
Our previous study utilized importance analysis, random forest, and Barnes–Hut t-SNE dimensionality reduction to analyze critical dicing parameters and used bidirectional long short-term memory (BLSTM) to predict wafer chipping occurrence successfully in a single dicing machine. However, each dicing machine of the same [...] Read more.
Our previous study utilized importance analysis, random forest, and Barnes–Hut t-SNE dimensionality reduction to analyze critical dicing parameters and used bidirectional long short-term memory (BLSTM) to predict wafer chipping occurrence successfully in a single dicing machine. However, each dicing machine of the same type may produce unevenly distributed non-IID dicing signals, which may lead to the undesirable result that a pre-trained model trained by dicing machine #1 could not effectively predict chipping occurrence in dicing machine #2. Therefore, regarding the model robustness, this study introduces an ensemble meta-learning-based model that can evaluate many dicing machines for chipping prediction with high stability and accuracy. This approach constructs several base learners, such as the hidden Markov model (HMM), the variational autoencoder (VAE), and BLSTM, to form an ensemble learning. We use model-agnostic meta-learning (MAML) to train and test the ensemble learning model by several prediction tasks from machine #1. After MAML learning, we call the trained model a meta learner. Then, we successfully apply a retrieved data set from machine #2 to the meta learner for training and testing wafer chipping occurrence in this machine. As a result, our contribution to the robust chipping prediction on cross-machines can improve the yield of wafer dicing with a prediction accuracy of 93.21%, preserve the practical wearing of dicing kerfs, and significantly cut wafer manufacturing costs. Full article
(This article belongs to the Special Issue Novel Methods for Object Detection and Segmentation)
Show Figures

Figure 1

20 pages, 3626 KiB  
Article
Determining Optimal Assembly Condition for Lens Module Production by Combining Genetic Algorithm and C-BLSTM
by Hyegeun Min, Yeonbin Son and Yerim Choi
Processes 2024, 12(3), 452; https://doi.org/10.3390/pr12030452 - 23 Feb 2024
Viewed by 1564
Abstract
Mobile camera modules are manufactured by aligning and assembling multiple differently shaped part lenses. Therefore, selecting the part lenses to assemble from candidates (called cavities) and determining the directional angle of each part lens for assembly have been important issues to maximize production [...] Read more.
Mobile camera modules are manufactured by aligning and assembling multiple differently shaped part lenses. Therefore, selecting the part lenses to assemble from candidates (called cavities) and determining the directional angle of each part lens for assembly have been important issues to maximize production yield. Currently, this process is manually conducted by experts at the manufacturing site, and the manual assembly condition optimization carries the risk of reduced production yield and increased failure cost as it largely depends on one’s expertise. Herein, we propose an AI framework that determines the optimal assembly condition including the combination of part lens cavities and the directional angles of part lenses. To achieve this, we combine the genetic algorithm with convolutional bidirectional long-term short-term memory (C-BLSTM). To the best of our knowledge, this is the first study on lens module production finding the optimal combination of part lens cavities and directional angles at the same time using machine learning methods. Based on experimental results using real-world datasets collected by lens module manufacturers, the proposed framework outperformed existing algorithms with an F1 score of 0.89. Moreover, the proposed method (S2S-AE) for predicting the directional angles exhibited the best performance compared to existing algorithms with an accuracy of 78.19%. Full article
(This article belongs to the Special Issue Advances in Intelligent Manufacturing Systems and Process Control)
Show Figures

Figure 1

17 pages, 3341 KiB  
Article
Informer-Based Temperature Prediction Using Observed and Numerical Weather Prediction Data
by Jimin Jun and Hong Kook Kim
Sensors 2023, 23(16), 7047; https://doi.org/10.3390/s23167047 - 9 Aug 2023
Cited by 16 | Viewed by 2911
Abstract
This paper proposes an Informer-based temperature prediction model to leverage data from an automatic weather station (AWS) and a local data assimilation and prediction system (LDAPS), where the Informer as a variant of a Transformer was developed to better deal with time series [...] Read more.
This paper proposes an Informer-based temperature prediction model to leverage data from an automatic weather station (AWS) and a local data assimilation and prediction system (LDAPS), where the Informer as a variant of a Transformer was developed to better deal with time series data. Recently, deep-learning-based temperature prediction models have been proposed, demonstrating successful performances, such as conventional neural network (CNN)-based models, bi-directional long short-term memory (BLSTM)-based models, and a combination of both neural networks, CNN–BLSTM. However, these models have encountered issues due to the lack of time data integration during the training phase, which also lead to the persistence of a long-term dependency problem in the LSTM models. These limitations have culminated in a performance deterioration when the prediction time length was extended. To overcome these issues, the proposed model first incorporates time-periodic information into the learning process by generating time-periodic information and inputting it into the model. Second, the proposed model replaces the LSTM with an Informer as an alternative to mitigating the long-term dependency problem. Third, a series of fusion operations between AWS and LDAPS data are executed to examine the effect of each dataset on the temperature prediction performance. The performance of the proposed temperature prediction model is evaluated via objective measures, including the root-mean-square error (RMSE) and mean absolute error (MAE) over different timeframes, ranging from 6 to 336 h. The experiments showed that the proposed model relatively reduced the average RMSE and MAE by 0.25 °C and 0.203 °C, respectively, compared with the results of the CNN–BLSTM-based model. Full article
(This article belongs to the Section Intelligent Sensors)
Show Figures

Graphical abstract

30 pages, 11851 KiB  
Article
Load Forecasting Based on LVMD-DBFCM Load Curve Clustering and the CNN-IVIA-BLSTM Model
by Linjing Hu, Jiachen Wang, Zhaoze Guo and Tengda Zheng
Appl. Sci. 2023, 13(12), 7332; https://doi.org/10.3390/app13127332 - 20 Jun 2023
Cited by 5 | Viewed by 2003
Abstract
Power load forecasting plays an important role in power systems, and the accuracy of load forecasting is of vital importance to power system planning as well as economic efficiency. Power load data are nonsmooth, nonlinear time-series and “noisy” data. Traditional load forecasting has [...] Read more.
Power load forecasting plays an important role in power systems, and the accuracy of load forecasting is of vital importance to power system planning as well as economic efficiency. Power load data are nonsmooth, nonlinear time-series and “noisy” data. Traditional load forecasting has low accuracy and curves not fitting the load variation. It is not well predicted by a single forecasting model. In this paper, we propose a novel model based on the combination of data mining and deep learning to improve the prediction accuracy. First, data preprocessing is performed. Second, identification and correction of anomalous data, normalization of continuous sequences, and one-hot encoding of discrete sequences are performed. The load data are decomposed and denoised using the double decomposition modal (LVMD) strategy, the load curves are clustered using the double weighted fuzzy C-means (DBFCM) algorithm, and the typical curves obtained are used as load patterns. In addition, data feature analysis is performed. A convolutional neural network (CNN) is used to extract data features. A bidirectional long short-term memory (BLSTM) network is used for prediction, in which the number of hidden layer neurons, the number of training epochs, the learning rate, the regularization coefficient, and other relevant parameters in the BLSTM network are optimized using the influenza virus immunity optimization algorithm (IVIA). Finally, the historical data of City H from 1 January 2016 to 31 December 2018, are used for load forecasting. The experimental results show that the novel model based on LVMD-DBFCM load c1urve clustering combined with CNN-IVIA-BLSTM proposed in this paper has an error of only 2% for electric load forecasting. Full article
(This article belongs to the Topic Soft Computing)
Show Figures

Figure 1

Back to TopTop