Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,698)

Search Parameters:
Keywords = hybrid convolutional neural networks

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
32 pages, 4909 KB  
Article
A Lightweight Hybrid Deep Learning Model for Tuberculosis Detection from Chest X-Rays
by Majdi Owda, Ahmad Abumihsan, Amani Yousef Owda and Mobarak Abumohsen
Diagnostics 2025, 15(24), 3216; https://doi.org/10.3390/diagnostics15243216 - 16 Dec 2025
Abstract
Background/Objectives: Tuberculosis remains a significant global health problem, particularly in resource-limited environments. Its mortality and spread can be considerably decreased by early and precise detection via chest X-ray imaging. This study introduces a novel approach based on hybrid deep learning for Tuberculosis [...] Read more.
Background/Objectives: Tuberculosis remains a significant global health problem, particularly in resource-limited environments. Its mortality and spread can be considerably decreased by early and precise detection via chest X-ray imaging. This study introduces a novel approach based on hybrid deep learning for Tuberculosis detection from chest X-ray images. Methods: The introduced approach combines GhostNet, a lightweight convolutional neural network tuned for computational efficiency, and MobileViT, a transformer-based model that can capture both local spatial patterns and global contextual dependencies. Through such integration, the model attains a balanced trade-off between classification accuracy and computational efficiency. The architecture employs feature fusion, where spatial features from GhostNet and contextual representations from MobileViT are globally pooled and concatenated, which allows the model to learn discriminative and robust feature representations. Results: The suggested model was assessed on two publicly available chest X-ray datasets and contrasted against several cutting-edge convolutional neural network architectures. Findings showed that the introduced hybrid model surpasses individual baselines, attaining 99.52% accuracy on dataset 1 and 99.17% on dataset 2, while keeping low computational cost (7.73M parameters, 282.11M Floating Point Operations). Conclusions: These outcomes verify the efficacy of feature-level fusion between a convolutional neural network and transformer branches, allowing robust tuberculosis detection with low inference overhead. The model is ideal for clinical deployment and resource-constrained contexts due to its high accuracy and lightweight design. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

22 pages, 1626 KB  
Article
Genetic Algorithm Optimization for Hybrid Deep Learning Prognosis of Reverse Total Shoulder Arthroplasty Rehabilitation
by Sotiria Vrouva, Christos Raptis, George A. Koumantakis, George Anastasopoulos, Efstratios Karavasilis and Adam Adamopoulos
AI Med. 2026, 1(1), 3; https://doi.org/10.3390/aimed1010003 - 16 Dec 2025
Abstract
Within the framework of the ongoing development of application of Machine Learning models in Medicine and Physical Therapy, the development of accurate prognosis algorithms for postoperative patients during the rehabilitation phase remains an area requiring further refinement. This paper examines hybrid Deep Learning [...] Read more.
Within the framework of the ongoing development of application of Machine Learning models in Medicine and Physical Therapy, the development of accurate prognosis algorithms for postoperative patients during the rehabilitation phase remains an area requiring further refinement. This paper examines hybrid Deep Learning models that integrate Convolution Neural Networks, Long Short-Term Memory and Gated Recurrent Unit networks, as well as genetic algorithm optimization for feature selection for predicting the time needed for a patient to rehabilitate. Patient data included features like age, passive range of available movements (preoperative and postoperative) and total rehabilitation time. Genetic Algorithm optimization for feature selection indicated that 4 out of the 16 available features are adequate for predicting rehabilitation time. Hybrid Deep Learning models achieved a Root Mean Squared Error (RMSE) of 12 days (less than 0.4 months) in rehabilitation time prediction, demonstrating good performance on a relatively small dataset of 120 patients. Full article
Show Figures

Figure 1

21 pages, 1994 KB  
Article
On a Hybrid CNN-Driven Pipeline for 3D Defect Localisation in the Inspection of EV Battery Modules
by Paolo Catti, Luca Fabbro and Nikolaos Nikolakis
Sensors 2025, 25(24), 7613; https://doi.org/10.3390/s25247613 - 15 Dec 2025
Abstract
The reliability of electric vehicle (EV) batteries requires detecting surface defects but also precisely locating them on the physical module for automated inspection, repair, and process optimisation. Conventional 2D computer vision methods, though accurate in image-space, do not provide traceable, real-world defect coordinates [...] Read more.
The reliability of electric vehicle (EV) batteries requires detecting surface defects but also precisely locating them on the physical module for automated inspection, repair, and process optimisation. Conventional 2D computer vision methods, though accurate in image-space, do not provide traceable, real-world defect coordinates on complex or curved battery surfaces, limiting utility for digital twins, root cause analysis, and automated quality control. This work proposes a hybrid inspection pipeline that produces millimetre-level three-dimensional (3D) defect maps for EV battery modules. The approach integrates (i) calibrated dual-view multi-view geometry to project defect points onto the CAD geometry and triangulate them where dual-view coverage is available, (ii) single-image neural 3D shape inference calibrated to the module geometry to complement regions with limited multi-view coverage, and (iii) generative, physically informed augmentation of rare or complex defect types. Defects are first detected in 2D images using a convolutional neural network (CNN), then projected onto a dense 3D CAD model of each module, complemented by a single-image depth prediction in regions with limited dual-view coverage, yielding true as-built localisation on the battery’s surface. GenAI methods are employed to expand the dataset with synthetic defect variations. Synthetic, physically informed defect examples are incorporated during training to mitigate the scarcity of rare defect types. Evaluation on a pilot industrial dataset, with a physically measured reference subset, demonstrates that the hybrid 3D approach achieves millimetre-scale localisation accuracy and outperforms a per-view CNN baseline in both segmentation and 3D continuity. Full article
(This article belongs to the Special Issue Convolutional Neural Network Technology for 3D Imaging and Sensing)
28 pages, 2880 KB  
Article
A Novel Hybrid GWO-RFO Metaheuristic Algorithm for Optimizing 1D-CNN Hyperparameters in IoT Intrusion Detection Systems
by Eslam Bokhory Elsayed, Abdalla Sayed Yassin and Hanan Fahmy
Information 2025, 16(12), 1103; https://doi.org/10.3390/info16121103 - 15 Dec 2025
Abstract
Because Internet of Things (IoT) networks are widely deployed, they have become attractive targets for botnet and distributed denial of service (DDoS) attacks, which require effective intrusion detection. Convolutional neural networks (CNNs) can achieve strong detection performance, but their many hyperparameters are usually [...] Read more.
Because Internet of Things (IoT) networks are widely deployed, they have become attractive targets for botnet and distributed denial of service (DDoS) attacks, which require effective intrusion detection. Convolutional neural networks (CNNs) can achieve strong detection performance, but their many hyperparameters are usually tuned manually, which is costly and time-consuming. This paper proposes a new hybrid metaheuristic optimizer, FW-CNN, that combines Grey Wolf Optimization and Red Fox Optimization to automatically tune the key hyperparameters of a one-dimensional CNN for IoT intrusion detection. The Red Fox component enhances exploration and helps the search escape local optima, while the Grey Wolf component strengthens exploitation and guides convergence toward high-quality solutions. The proposed model is evaluated using the N-BaIoT dataset and compared with a feedforward neural network as well as a metaheuristic-optimized model based on the Adaptive Particle Swarm Optimization–Whale Optimization Algorithm-CNN. It achieves a final accuracy of 95.56%, improving on the feedforward network by 12.56 percentage points and outperforming the Adaptive Particle Swarm Optimization–Whale Optimization Algorithm-based CNN model by 1.02 percentage points. It also yields higher average precision, Kappa coefficient, and Jaccard similarity, and significantly reduces Hamming loss. These results indicate that the proposed hybrid optimizer is stable and effective for multi-class IoT intrusion detection in real environments. Full article
(This article belongs to the Special Issue Security and Privacy of Resource-Constrained IoT Devices)
Show Figures

Figure 1

22 pages, 3276 KB  
Article
Deep Neural Network-Based Inverse Identification of the Mechanical Behavior of Anisotropic Tubes
by Zied Ktari, Pedro Prates and Ali Khalfallah
J. Manuf. Mater. Process. 2025, 9(12), 410; https://doi.org/10.3390/jmmp9120410 - 14 Dec 2025
Viewed by 68
Abstract
Tube hydroforming is a versatile forming process widely used in lightweight structural applications, where accurate characterization of the hoop mechanical behavior is crucial for reliable design and simulation. The ring hoop tensile test (RHTT) provides valuable experimental data for evaluating the elastoplastic response [...] Read more.
Tube hydroforming is a versatile forming process widely used in lightweight structural applications, where accurate characterization of the hoop mechanical behavior is crucial for reliable design and simulation. The ring hoop tensile test (RHTT) provides valuable experimental data for evaluating the elastoplastic response of anisotropic tubes in the hoop direction, but frictional effects often distort the measured force–displacement response. This study proposes a deep learning-based inverse identification framework to accurately recover the true hoop stress–strain behavior from RHTT data. Convolutional and recurrent neural network architectures, including CNN, long short term memory (LSTM), gated recurrent unit (GRU), bidirectional GRU (BiGRU), bidirectional LSTM (BiLSTM) and ConvLSTM, were trained using numerically generated datasets from finite element simulations. Data augmentation and hyperparameter tuning were applied to generalization. The hybrid ConvLSTM model achieved superior performance, with a minimum mean absolute error (MAE) of 0.08 and a coefficient of determination (R2) value of approximately 0.97, providing a close match to the Hill48 yield criterion. The proposed approach demonstrates the potential of deep neural networks as an efficient and accurate alternative to traditional inverse methods for characterizing anisotropic tubular materials. Full article
(This article belongs to the Special Issue Innovative Approaches in Metal Forming and Joining Technologies)
Show Figures

Figure 1

27 pages, 709 KB  
Article
A Tabular Data Imputation Technique Using Transformer and Convolutional Neural Networks
by Charlène Béatrice Bridge-Nduwimana, Salah Eddine El Harrauss, Aziza El Ouaazizi and Majid Benyakhlef
Big Data Cogn. Comput. 2025, 9(12), 321; https://doi.org/10.3390/bdcc9120321 - 13 Dec 2025
Viewed by 90
Abstract
Upstream processes strongly influence downstream analysis in sequential data-processing workflows, particularly in machine learning, where data quality directly affects model performance. Conventional statistical imputations often fail to capture nonlinear dependencies, while deep learning approaches typically lack uncertainty quantification. We introduce a hybrid imputation [...] Read more.
Upstream processes strongly influence downstream analysis in sequential data-processing workflows, particularly in machine learning, where data quality directly affects model performance. Conventional statistical imputations often fail to capture nonlinear dependencies, while deep learning approaches typically lack uncertainty quantification. We introduce a hybrid imputation model that integrates a deep learning autoencoder with Convolutional Neural Network (CNN) layers and a Transformer-based contextual modeling architecture to address systematic variation across heterogeneous data sources. Performing multiple imputations in the autoencoder–transformer latent space and averaging representations provides implicit batch correction that suppresses context-specific remains without explicit batch identifiers. We performed experiments on datasets in which 10% of missing data was artificially introduced by completely random missing data (MCAR) and non-random missing data (MNAR) mechanisms. They demonstrated practical performance, jointly ranking first among the imputation methods evaluated. This imputation technique reduced the root mean square error (RMSE) by 50% compared to denoising autoencoders (DAE) and by 46% compared to iterative imputation (MICE). Performance was comparable for adversarial models (GAIN) and attention-based models (MIDA), and both provided interpretable uncertainty estimates (CV = 0.08–0.15). Validation on datasets from multiple sources confirmed the robustness of the technique: notably, on a forensic dataset from multiple laboratories, our imputation technique achieved a practical improvement over GAIN (0.146 vs. 0.189 RMSE), highlighting its effectiveness in mitigating batch effects. Full article
Show Figures

Graphical abstract

23 pages, 3125 KB  
Article
Hybrid AI Intrusion Detection: Balancing Accuracy and Efficiency
by Vandit R Joshi, Kwame Assa-Agyei, Tawfik Al-Hadhrami and Sultan Noman Qasem
Sensors 2025, 25(24), 7564; https://doi.org/10.3390/s25247564 - 12 Dec 2025
Viewed by 224
Abstract
The Internet of Things (IoT) has transformed industries, healthcare, and smart environments, but introduces severe security threats due to resource constraints, weak protocols, and heterogeneous infrastructures. Traditional Intrusion Detection Systems (IDS) fail to address critical challenges including scalability across billions of devices, interoperability [...] Read more.
The Internet of Things (IoT) has transformed industries, healthcare, and smart environments, but introduces severe security threats due to resource constraints, weak protocols, and heterogeneous infrastructures. Traditional Intrusion Detection Systems (IDS) fail to address critical challenges including scalability across billions of devices, interoperability among diverse protocols, real-time responsiveness under strict latency, data privacy in distributed edge networks, and high false positives in imbalanced traffic. This study provides a systematic comparative evaluation of three representative AI models, CNN-BiLSTM, Random Forest, and XGBoost for IoT intrusion detection on the NSL-KDD and UNSW-NB15 datasets. The analysis quantifies the achievable detection performance and inference latency of each approach, revealing a clear accuracy–latency trade-off that can guide practical model selection: CNN-BiLSTM offers the highest detection capability (F1 up to 0.986) at the cost of higher computational overhead, whereas XGBoost and Random Forest deliver competitive accuracy with significantly lower inference latency (sub-millisecond on conventional hardware). These empirical insights support informed deployment decisions in heterogeneous IoT environments where accuracy-critical gateways and latency-critical sensors coexist. Full article
(This article belongs to the Special Issue AI-Empowered Internet of Things)
Show Figures

Figure 1

29 pages, 11999 KB  
Article
Pixel-Wise Sky-Obstacle Segmentation in Fisheye Imagery Using Deep Learning and Gradient Boosting
by Némo Bouillon and Vincent Boitier
J. Imaging 2025, 11(12), 446; https://doi.org/10.3390/jimaging11120446 - 12 Dec 2025
Viewed by 179
Abstract
Accurate sky–obstacle segmentation in hemispherical fisheye imagery is essential for solar irradiance forecasting, photovoltaic system design, and environmental monitoring. However, existing methods often rely on expensive all-sky imagers and region-specific training data, produce coarse sky–obstacle boundaries, and ignore the optical properties of fisheye [...] Read more.
Accurate sky–obstacle segmentation in hemispherical fisheye imagery is essential for solar irradiance forecasting, photovoltaic system design, and environmental monitoring. However, existing methods often rely on expensive all-sky imagers and region-specific training data, produce coarse sky–obstacle boundaries, and ignore the optical properties of fisheye lenses. We propose a low-cost segmentation framework designed for fisheye imagery that combines synthetic data generation, lens-aware augmentation, and a hybrid deep-learning pipeline. Synthetic fisheye training images are created from publicly available street-view panoramas to cover diverse environments without dedicated hardware, and lens-aware augmentations model fisheye projection and photometric effects to improve robustness across devices. On this dataset, we train a convolutional neural network (CNN) and refine its output with gradient-boosted decision trees (GBDT) to sharpen sky–obstacle boundaries. The method is evaluated on real fisheye images captured with smartphones and low-cost clip-on lenses across multiple sites, achieving an Intersection over Union (IoU) of 96.63% and an F1 score of 98.29%, along with high boundary accuracy. An additional evaluation on an external panoramic baseline dataset confirms strong cross-dataset generalization. Together, these results show that the proposed framework enables accurate, low-cost, and widely deployable hemispherical sky segmentation for practical solar and environmental imaging applications. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

28 pages, 3895 KB  
Article
Advancing Machine Learning Strategies for Power Consumption-Based IoT Botnet Detection
by Almustapha A. Wakili, Saugat Guni, Sabbir Ahmed Khan, Wei Yu and Woosub Jung
Sensors 2025, 25(24), 7553; https://doi.org/10.3390/s25247553 - 12 Dec 2025
Viewed by 235
Abstract
The proliferation of Internet of Things (IoT) devices has amplified botnet risks, while traditional network-based intrusion detection systems (IDSs) struggle under encrypted and/or sparse traffic. Power consumption offers an effective side channel for device-level detection. Yet, prior studies typically focus on a single [...] Read more.
The proliferation of Internet of Things (IoT) devices has amplified botnet risks, while traditional network-based intrusion detection systems (IDSs) struggle under encrypted and/or sparse traffic. Power consumption offers an effective side channel for device-level detection. Yet, prior studies typically focus on a single model family (often a convolutional neural network (CNN)) and rarely assess generalization across devices or compare broader model classes. In this paper, we conduct unified benchmarking and comparison of classical (SVM and RF), deep (CNN, LSTM, and 1D Transformer), and hybrid (CNN + LSTM, CNN + Transformer, and CNN + RF) models on the CHASE’19 dataset and a newly curated three-class botnet dataset, using consistent preprocessing and evaluation across single- and cross-device settings, reporting both accuracy and efficiency (latency and throughput). Experimental results demonstrate that Random Forest achieves the highest single-device accuracy (99.43% on the Voice Assistant with Seed 42), while CNN + Transformer shows a strong accuracy–efficiency trade-off in cross-device scenarios (94.02% accuracy on the combined dataset at ∼60,000 samples/s when using the best-performing Seed 42). These results offer practical guidance for selecting models under accuracy, latency, and throughput constraints and establish a reproducible baseline for power-side-channel IDSs. Full article
(This article belongs to the Special Issue IoT Cybersecurity: 2nd Edition)
Show Figures

Figure 1

12 pages, 931 KB  
Article
Efficient Pulsar Candidate Recognition Algorithm Under a Multi-Scale DenseNet Framework
by Junlin Tang, Xiaoyao Xie and Xiangguang Xiong
Appl. Sci. 2025, 15(24), 13097; https://doi.org/10.3390/app152413097 - 12 Dec 2025
Viewed by 82
Abstract
The exponential growth of candidate data from large-scale radio pulsar surveys has created a pressing need for efficient and accurate classification methods. This paper presents a novel hybrid pulsar candidate recognition algorithm that integrates diagnostic plot images and structured numerical features using a [...] Read more.
The exponential growth of candidate data from large-scale radio pulsar surveys has created a pressing need for efficient and accurate classification methods. This paper presents a novel hybrid pulsar candidate recognition algorithm that integrates diagnostic plot images and structured numerical features using a multi-scale DenseNet framework. The proposed model combines convolutional neural networks (CNNs) for extracting spatial patterns from pulsar diagnostic plots and feedforward neural networks (FNNs) for processing scalar features such as SNR, DM, and pulse width. By fusing these multimodal representations, the model achieves superior classification performance, particularly in class-imbalanced settings standard to pulsar survey data. Evaluated on a synthesized dataset constructed from FAST and HTRU survey characteristics, the model demonstrates robust performance, achieving an F1-score of 0.904 and AUC-ROC of 0.978. Extensive ablation and cross-validation analyses confirm the contribution of each data modality and the model’s generalizability. Furthermore, the system maintains low inference latency (4.2 ms per candidate) and a compact architecture (~2.3 million parameters), indicating potential for real-time deployment once validated on real observational datasets. The proposed approach offers a scalable and interpretable multimodal framework for automated pulsar classification and provides a foundation for future validation and potential integration into observatories such as FAST and the Square Kilometre Array (SKA). Full article
Show Figures

Figure 1

23 pages, 6143 KB  
Article
Hybrid Cascade and Dual-Path Adaptive Aggregation Network for Medical Image Segmentation
by Junhong Ren, Sen Chen, Yange Sun, Huaping Guo, Yongqiang Tang and Wensheng Zhang
Electronics 2025, 14(24), 4879; https://doi.org/10.3390/electronics14244879 - 11 Dec 2025
Viewed by 88
Abstract
Deep learning methods based on convolutional neural networks (CNNs) and Mamba have advanced medical image segmentation, yet two challenges remain: (1) trade-off in feature extraction, where CNNs capture local details but miss global context, and Mamba captures global dependencies but overlooks fine structures, [...] Read more.
Deep learning methods based on convolutional neural networks (CNNs) and Mamba have advanced medical image segmentation, yet two challenges remain: (1) trade-off in feature extraction, where CNNs capture local details but miss global context, and Mamba captures global dependencies but overlooks fine structures, and (2) limited feature aggregation, as existing methods insufficiently integrate inter-layer common information and delta details, hindering robustness to subtle structures. To address these issues, we propose a hybrid cascade and dual-path adaptive aggregation network (HCDAA-Net). For feature extraction, we design a hybrid cascade structure (HCS) that alternately applies ResNet and Mamba modules, achieving a spatial balance between local detail preservation and global semantic modeling. We further employ a general channel-crossing attention mechanism to enhance feature expression, complementing this spatial modeling and accelerating convergence. For feature aggregation, we first propose correlation-aware aggregation (CAA) to model correlations among features of the same lesions or anatomical structures. Second, we develop a dual-path adaptive feature aggregation (DAFA) module: the common path captures stable cross-layer semantics and suppresses redundancy, while the delta path emphasizes subtle differences to strengthen the model’s sensitivity to fine details. Finally, we introduce a residual-gated visual state space module (RG-VSS), which dynamically modulates information flow via a convolution-enhanced residual gating mechanism to refine fused representations. Experiments on diverse datasets demonstrate that our HCDAA-Net outperforms some state-of-the-art (SOTA) approaches. Full article
Show Figures

Figure 1

20 pages, 6385 KB  
Article
Molecular Remodeling of Milk Fat Globules Induced by Centrifugation: Insights from Deep Learning-Based Detection of Milk Adulteration
by Grzegorz Gwardys, Grzegorz Grodkowski, Piotr Kostusiak, Wojciech Mendelowski, Jan Slósarz, Michał Satława, Bartłomiej Śmietanka, Krzysztof Gwardys, Marcin Gołębiewski and Kamila Puppel
Int. J. Mol. Sci. 2025, 26(24), 11919; https://doi.org/10.3390/ijms262411919 - 10 Dec 2025
Viewed by 138
Abstract
Milk adulteration through centrifugation, which artificially reduces the somatic cell count (SCC), represents a significant challenge to food authenticity and public health. This fraudulent practice alters the native molecular architecture of milk, masking inflammatory conditions such as subclinical mastitis and distorting product quality. [...] Read more.
Milk adulteration through centrifugation, which artificially reduces the somatic cell count (SCC), represents a significant challenge to food authenticity and public health. This fraudulent practice alters the native molecular architecture of milk, masking inflammatory conditions such as subclinical mastitis and distorting product quality. Conventional analytical and microscopic techniques remain insufficiently sensitive to detect the subtle physicochemical changes associated with centrifugation, highlighting the need for molecular-level, data-driven diagnostics. The dataset included 128 paired raw milk samples and approximately 25,000 bright-field micrographs acquired across multiple microscopes, of which 95% were confirmed to be of high quality. In this study, advanced machine learning (ML) and deep learning (DL) approaches were applied to identify centrifugation-induced alterations in raw milk microstructure. Bright-field micrographs (pixel size 0.27 µm) of paired unprocessed and centrifuged samples were obtained under standardized optical conditions and analyzed using convolutional neural networks (ResNet-18/50, Inception-v3, Xception, NasNet-Mobile) and hybrid attention architectures (MaxViT, CoAtNet). Model performance was evaluated using the harmonic average of recalls across five micrographs per sample (HAR5). Human microscopy experts (n = 4) achieved only 18% classification accuracy—below the random baseline (25%)—confirming that centrifugation-induced modifications are not visually discernible. In contrast, DL architectures reached up to 97% accuracy (HAR5, Xception), successfully identifying subtle molecular cues. Class activation and sensitivity analyses indicated that models focused not on milk fat globule (MFG) boundaries but on high-frequency nanoscale variations related to the reorganization of casein micelles and solid non-fat fractions. The findings strongly suggest that centrifugation adulteration constitutes a molecular reorganization event rather than a morphological alteration. The integration of optical microscopy with AI-driven molecular analytics establishes deep learning as a precise and objective tool for detecting fraudulent milk processing and improving food integrity diagnostics. Full article
(This article belongs to the Special Issue Application of Artificial Intelligence in Molecular Sciences)
Show Figures

Figure 1

23 pages, 3326 KB  
Article
Hybrid Multi-Scale Neural Network with Attention-Based Fusion for Fruit Crop Disease Identification
by Shakhmaran Seilov, Akniyet Nurzhaubayev, Marat Baideldinov, Bibinur Zhursinbek, Medet Ashimgaliyev and Ainur Zhumadillayeva
J. Imaging 2025, 11(12), 440; https://doi.org/10.3390/jimaging11120440 - 10 Dec 2025
Viewed by 170
Abstract
Unobserved fruit crop illnesses are a major threat to agricultural productivity worldwide and frequently cause farmers to suffer large financial losses. Manual field inspection-based disease detection techniques are time-consuming, unreliable, and unsuitable for extensive monitoring. Deep learning approaches, in particular convolutional neural networks, [...] Read more.
Unobserved fruit crop illnesses are a major threat to agricultural productivity worldwide and frequently cause farmers to suffer large financial losses. Manual field inspection-based disease detection techniques are time-consuming, unreliable, and unsuitable for extensive monitoring. Deep learning approaches, in particular convolutional neural networks, have shown promise for automated plant disease identification, although they still face significant obstacles. These include poor generalization across complicated visual backdrops, limited resilience to different illness sizes, and high processing needs that make deployment on resource-constrained edge devices difficult. We suggest a Hybrid Multi-Scale Neural Network (HMCT-AF with GSAF) architecture for precise and effective fruit crop disease identification in order to overcome these drawbacks. In order to extract long-range dependencies, HMCT-AF with GSAF combines a Vision Transformer-based structural branch with multi-scale convolutional branches to capture both high-level contextual patterns and fine-grained local information. These disparate features are adaptively combined using a novel HMCT-AF with a GSAF module, which enhances model interpretability and classification performance. We conduct evaluations on both PlantVillage (controlled environment) and CLD (real-world in-field conditions), observing consistent performance gains that indicate strong resilience to natural lighting variations and background complexity. With an accuracy of up to 93.79%, HMCT-AF with GSAF outperforms vanilla Transformer models, EfficientNet, and traditional CNNs. These findings demonstrate how well the model captures scale-variant disease symptoms and how it may be used in real-time agricultural applications using hardware that is compatible with the edge. According to our research, HMCT-AF with GSAF presents a viable basis for intelligent, scalable plant disease monitoring systems in contemporary precision farming. Full article
Show Figures

Figure 1

25 pages, 3403 KB  
Article
Hybrid Deep Learning Approach for Fractional-Order Model Parameter Identification of Lithium-Ion Batteries
by Maharani Putri, Dat Nguyen Khanh, Kun-Che Ho, Shun-Chung Wang and Yi-Hua Liu
Batteries 2025, 11(12), 452; https://doi.org/10.3390/batteries11120452 - 9 Dec 2025
Viewed by 239
Abstract
Fractional-order models (FOMs) have been recognized as superior tools for capturing the complex electrochemical dynamics of lithium-ion batteries, outperforming integer-order models in accuracy, robustness, and adaptability. Parameter identification (PI) is essential for FOMs, as its accuracy directly affects the model’s ability to predict [...] Read more.
Fractional-order models (FOMs) have been recognized as superior tools for capturing the complex electrochemical dynamics of lithium-ion batteries, outperforming integer-order models in accuracy, robustness, and adaptability. Parameter identification (PI) is essential for FOMs, as its accuracy directly affects the model’s ability to predict battery behavior and estimate critical states such as state of charge (SOC) and state of health (SOH). In this study, a hybrid deep learning approach has been introduced for FOM PI, representing the first application of deep learning in this domain. A simulation platform was developed to generate datasets using Sobol and Monte Carlo sampling methods. Five deep learning models were constructed: long short-term memory (LSTM), gated recurrent unit (GRU), one-dimensional convolutional neural network (1DCNN), and hybrid models combining 1DCNN with LSTM and GRU. Hyperparameters were optimized using Optuna, and enhancements such as Huber loss for robustness to outliers, stochastic weight averaging, and exponential moving average for training stability were incorporated. The primary contribution lies in the hybrid architectures, which integrate convolutional feature extraction with recurrent temporal modeling, outperforming standalone models. On a test set of 1000 samples, the improved 1DCNN + GRU model achieved an overall root mean square error (RMSE) of 0.2223 and a mean absolute percentage error (MAPE) of 0.27%, representing nearly a 50% reduction in RMSE compared to its baseline. This performance surpasses that of the improved LSTM model, which yielded a MAPE of 9.50%, as evidenced by tighter scatter plot alignments along the diagonal and reduced error dispersion in box plots. Terminal voltage prediction was validated with an average RMSE of 0.002059 and mean absolute error (MAE) of 0.001387, demonstrating high-fidelity dynamic reconstruction. By advancing data-driven PI, this framework is well-positioned to enable real-time integration into battery management systems. Full article
(This article belongs to the Special Issue Towards a Smarter Battery Management System: 3rd Edition)
22 pages, 3280 KB  
Article
A Novel Scenario-Based Comparative Framework for Short- and Medium-Term Solar PV Power Forecasting Using Deep Learning Models
by Elif Yönt Aydın, Kevser Önal, Cem Haydaroğlu, Heybet Kılıç, Özal Yıldırım, Oğuzhan Katar and Hüseyin Erdoğan
Appl. Sci. 2025, 15(24), 12965; https://doi.org/10.3390/app152412965 - 9 Dec 2025
Viewed by 234
Abstract
Accurate short- and medium-term forecasting of photovoltaic (PV) power generation is vital for grid stability and renewable energy integration. This study presents a comparative scenario-based approach using Long Short-Term Memory (LSTM), Convolutional Neural Network (CNN), and Gated Recurrent Unit (GRU) models trained with [...] Read more.
Accurate short- and medium-term forecasting of photovoltaic (PV) power generation is vital for grid stability and renewable energy integration. This study presents a comparative scenario-based approach using Long Short-Term Memory (LSTM), Convolutional Neural Network (CNN), and Gated Recurrent Unit (GRU) models trained with one year of real-time meteorological and production data from a 250 kWp grid-connected PV system located at Dicle University in Diyarbakır, Southeastern Anatolia, Turkey. The dataset includes hourly measurements of solar irradiance (average annual GHI 5.4 kWh/m2/day), ambient temperature, humidity, and wind speed, with missing data below 2% after preprocessing. Six forecasting scenarios were designed for different horizons (6 h to 1 month). Results indicate that the LSTM model achieved the best performance in short-term scenarios, reaching R2 values above 0.90 and lower MAE and RMSE compared to CNN and GRU. The GRU model showed similar accuracy with faster training time, while CNN produced higher errors due to the dominant temporal nature of PV output. These results align with recent studies that emphasize selecting suitable deep learning architectures for time-series energy forecasting. This work highlights the benefit of integrating real local meteorological data with deep learning models in a scenario-based design and provides practical insights for regional grid operators and energy planners to reduce production uncertainty. Future studies can improve forecast reliability by testing hybrid models and implementing real-time adaptive training strategies to better handle extreme weather fluctuations. Full article
Show Figures

Figure 1

Back to TopTop