Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (768)

Search Parameters:
Keywords = full convolutional networks

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
12 pages, 1163 KB  
Article
Sensor Input Type and Location Influence Outdoor Running Terrain Classification via Deep Learning Approaches
by Gabrielle Thibault, Philippe C. Dixon and David J. Pearsall
Sensors 2025, 25(19), 6203; https://doi.org/10.3390/s25196203 (registering DOI) - 7 Oct 2025
Abstract
Background/Objective: Understanding the training effect in high-level running is important for performance optimization and injury prevention. This includes awareness of how different running surface types (e.g., hard versus soft) may modify biomechanics. Recent studies have demonstrated that deep learning algorithms, such as convolutional [...] Read more.
Background/Objective: Understanding the training effect in high-level running is important for performance optimization and injury prevention. This includes awareness of how different running surface types (e.g., hard versus soft) may modify biomechanics. Recent studies have demonstrated that deep learning algorithms, such as convolutional neural networks (CNNs), can accurately classify human activity collected via body-worn sensors. To date, no study has assessed optimal signal type, sensor location, and model architecture to classify running surfaces. This study aimed to determine which combination of signal type, sensor location, and CNN architecture would yield the highest accuracy in classifying grass and asphalt surfaces using inertial measurement unit (IMU) sensors. Methods: Running data were collected from forty participants (27.4 years + 7.8 SD, 10.5 ± 7.3 SD years of running) with a full-body IMU system (head, sternum, pelvis, upper legs, lower legs, feet, and arms) on grass and asphalt outdoor surfaces. Performance (accuracy) for signal type (acceleration and angular velocity), sensor configuration (full body, lower body, pelvis, and feet), and CNN model architecture was tested for this specific task. Moreover, the effect of preprocessing steps (separating into running cycles and amplitude normalization) and two different data splitting protocols (leave-n-subject-out and subject-dependent split) was evaluated. Results: In general, acceleration signals improved classification results compared to angular velocity (3.8%). Moreover, the foot sensor configuration had the best performance-to-number of sensor ratio (95.5% accuracy). Finally, separating trials into gait cycles and not normalizing the raw signals improved accuracy by approximately 28%. Conclusion: This analysis sheds light on the important parameters to consider when developing machine learning classifiers in the human activity recognition field. A surface classification tool could provide useful quantitative feedback to athletes and coaches in terms of running technique effort on varied terrain surfaces, improve training personalization, prevent injuries, and improve performance. Full article
Show Figures

Figure 1

19 pages, 1948 KB  
Article
Graph-MambaRoadDet: A Symmetry-Aware Dynamic Graph Framework for Road Damage Detection
by Zichun Tian, Xiaokang Shao and Yuqi Bai
Symmetry 2025, 17(10), 1654; https://doi.org/10.3390/sym17101654 - 5 Oct 2025
Abstract
Road-surface distress poses a serious threat to traffic safety and imposes a growing burden on urban maintenance budgets. While modern detectors based on convolutional networks and Vision Transformers achieve strong frame-level performance, they often overlook an essential property of road environments—structural symmetry [...] Read more.
Road-surface distress poses a serious threat to traffic safety and imposes a growing burden on urban maintenance budgets. While modern detectors based on convolutional networks and Vision Transformers achieve strong frame-level performance, they often overlook an essential property of road environments—structural symmetry within road networks and damage patterns. We present Graph-MambaRoadDet (GMRD), a symmetry-aware and lightweight framework that integrates dynamic graph reasoning with state–space modeling for accurate, topology-informed, and real-time road damage detection. Specifically, GMRD employs an EfficientViM-T1 backbone and two DefMamba blocks, whose deformable scanning paths capture sub-pixel crack patterns while preserving geometric symmetry. A superpixel-based graph is constructed by projecting image regions onto OpenStreetMap road segments, encoding both spatial structure and symmetric topological layout. We introduce a Graph-Generating State–Space Model (GG-SSM) that synthesizes sparse sample-specific adjacency in O(M) time, further refined by a fusion module that combines detector self-attention with prior symmetry constraints. A consistency loss promotes smooth predictions across symmetric or adjacent segments. The full INT8 model contains only 1.8 M parameters and 1.5 GFLOPs, sustaining 45 FPS at 7 W on a Jetson Orin Nano—eight times lighter and 1.7× faster than YOLOv8-s. On RDD2022, TD-RD, and RoadBench-100K, GMRD surpasses strong baselines by up to +6.1 mAP50:95 and, on the new RoadGraph-RDD benchmark, achieves +5.3 G-mAP and +0.05 consistency gain. Qualitative results demonstrate robustness under shadows, reflections, back-lighting, and occlusion. By explicitly modeling spatial and topological symmetry, GMRD offers a principled solution for city-scale road infrastructure monitoring under real-time and edge-computing constraints. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

20 pages, 5116 KB  
Article
Design of Portable Water Quality Spectral Detector and Study on Nitrogen Estimation Model in Water
by Hongfei Lu, Hao Zhou, Renyong Cao, Delin Shi, Chao Xu, Fangfang Bai, Yang Han, Song Liu, Minye Wang and Bo Zhen
Processes 2025, 13(10), 3161; https://doi.org/10.3390/pr13103161 - 3 Oct 2025
Abstract
A portable spectral detector for water quality assessment was developed, utilizing potassium nitrate and ammonium chloride standard solutions as the subjects of investigation. By preparing solutions with differing concentrations, spectral data ranging from 254 to 1275 nm was collected and subsequently preprocessed using [...] Read more.
A portable spectral detector for water quality assessment was developed, utilizing potassium nitrate and ammonium chloride standard solutions as the subjects of investigation. By preparing solutions with differing concentrations, spectral data ranging from 254 to 1275 nm was collected and subsequently preprocessed using methods such as multiple scattering correction (MSC), Savitzky–Golay filtering (SG), and standardization (SS). Estimation models were constructed employing modeling algorithms including Support Vector Machine-Multilayer Perceptron (SVM-MLP), Support Vector Regression (SVR), random forest (RF), RF-Lasso, and partial least squares regression (PLSR). The research revealed that the primary variation bands for NH4+ and NO3 are concentrated within the 254–550 nm and 950–1275 nm ranges, respectively. For predicting ammonium chloride, the optimal model was found to be the SVM-MLP model, which utilized spectral data reduced to 400 feature bands after SS processing, achieving R2 and RMSE of 0.8876 and 0.0883, respectively. For predicting potassium nitrate, the optimal model was the 1D Convolutional Neural Network (1DCNN) model applied to the full band of spectral data after SS processing, with R2 and RMSE of 0.7758 and 0.1469, respectively. This study offers both theoretical and technical support for the practical implementation of spectral technology in rapid water quality monitoring. Full article
Show Figures

Figure 1

22 pages, 32792 KB  
Article
MRV-YOLO: A Multi-Channel Remote Sensing Object Detection Method for Identifying Reclaimed Vegetation in Hilly and Mountainous Mining Areas
by Xingmei Li, Hengkai Li, Jingjing Dai, Kunming Liu, Guanshi Wang, Shengdong Nie and Zhiyu Zhang
Forests 2025, 16(10), 1536; https://doi.org/10.3390/f16101536 - 2 Oct 2025
Abstract
Leaching mining of ion-adsorption rare earths degrades soil organic matter and hampers vegetation recovery. High-resolution UAV remote sensing enables large-scale monitoring of reclamation, yet vegetation detection accuracy is constrained by key challenges. Conventional three-channel detection struggles with terrain complexity, illumination variation, and shadow [...] Read more.
Leaching mining of ion-adsorption rare earths degrades soil organic matter and hampers vegetation recovery. High-resolution UAV remote sensing enables large-scale monitoring of reclamation, yet vegetation detection accuracy is constrained by key challenges. Conventional three-channel detection struggles with terrain complexity, illumination variation, and shadow effects. Fixed UAV altitude and missing topographic data further cause resolution inconsistencies, posing major challenges for accurate vegetation detection in reclaimed land. To enhance multi-spectral vegetation detection, the model input is expanded from the traditional three channels to six channels, enabling full utilization of multi-spectral information. Furthermore, the Channel Attention and Global Pooling SPPF (CAGP-SPPF) module is introduced for multi-scale feature extraction, integrating global pooling and channel attention to capture multi-channel semantic information. In addition, the C2f_DynamicConv module replaces conventional convolutions in the neck network to strengthen high-dimensional feature transmission and reduce information loss, thereby improving detection accuracy. On the self-constructed reclaimed vegetation dataset, MRV-YOLO outperformed YOLOv8, with mAP@0.5 and mAP@0.5:0.95 increasing by 4.6% and 10.8%, respectively. Compared with RT-DETR, YOLOv3, YOLOv5, YOLOv6, YOLOv7, yolov7-tiny, YOLOv8-AS, YOLOv10, and YOLOv11, mAP@0.5 improved by 6.8%, 9.7%, 5.3%, 6.5%, 6.4%, 8.9%, 4.6%, 2.1%, and 5.4%, respectively. The results demonstrate that multichannel inputs incorporating near-infrared and dual red-edge bands significantly enhance detection accuracy for reclaimed vegetation in rare earth mining areas, providing technical support for ecological restoration monitoring. Full article
(This article belongs to the Section Forest Inventory, Modeling and Remote Sensing)
Show Figures

Figure 1

14 pages, 1081 KB  
Article
Hybrid Deep Learning Approach for Secure Electric Vehicle Communications in Smart Urban Mobility
by Abdullah Alsaleh
Vehicles 2025, 7(4), 112; https://doi.org/10.3390/vehicles7040112 - 2 Oct 2025
Abstract
The increasing adoption of electric vehicles (EVs) within intelligent transportation systems (ITSs) has elevated the importance of cybersecurity, especially with the rise in Vehicle-to-Everything (V2X) communications. Traditional intrusion detection systems (IDSs) struggle to address the evolving and complex nature of cyberattacks in such [...] Read more.
The increasing adoption of electric vehicles (EVs) within intelligent transportation systems (ITSs) has elevated the importance of cybersecurity, especially with the rise in Vehicle-to-Everything (V2X) communications. Traditional intrusion detection systems (IDSs) struggle to address the evolving and complex nature of cyberattacks in such dynamic environments. To address these challenges, this study introduces a novel deep learning-based IDS designed specifically for EV communication networks. We present a hybrid model that integrates convolutional neural networks (CNNs), long short-term memory (LSTM) layers, and adaptive learning strategies. The model was trained and validated using the VeReMi dataset, which simulates a wide range of attack scenarios in V2X networks. Additionally, an ablation study was conducted to isolate the contribution of each of its modules. The model demonstrated strong performance with 98.73% accuracy, 97.88% precision, 98.91% sensitivity, and 98.55% specificity, as well as an F1-score of 98.39%, an MCC of 0.964, a false-positive rate of 1.45%, and a false-negative rate of 1.09%, with a detection latency of 28 ms and an AUC-ROC of 0.994. Specifically, this work fills a clear gap in the existing V2X intrusion detection literature—namely, the lack of scalable, adaptive, and low-latency IDS solutions for hardware-constrained EV platforms—by proposing a hybrid CNN–LSTM architecture coupled with an elastic weight consolidation (EWC)-based adaptive learning module that enables online updates without full retraining. The proposed model provides a real-time, adaptive, and high-precision IDS for EV networks, supporting safer and more resilient ITS infrastructures. Full article
Show Figures

Figure 1

25 pages, 8881 KB  
Article
Evaluating Machine Learning Techniques for Brain Tumor Detection with Emphasis on Few-Shot Learning Using MAML
by Soham Sanjay Vaidya, Raja Hashim Ali, Shan Faiz, Iftikhar Ahmed and Talha Ali Khan
Algorithms 2025, 18(10), 624; https://doi.org/10.3390/a18100624 - 2 Oct 2025
Abstract
Accurate brain tumor classification from MRI is often constrained by limited labeled data. We systematically compare conventional machine learning, deep learning, and few-shot learning (FSL) for four classes (glioma, meningioma, pituitary, no tumor) using a standardized pipeline. Models are trained on the Kaggle [...] Read more.
Accurate brain tumor classification from MRI is often constrained by limited labeled data. We systematically compare conventional machine learning, deep learning, and few-shot learning (FSL) for four classes (glioma, meningioma, pituitary, no tumor) using a standardized pipeline. Models are trained on the Kaggle Brain Tumor MRI Dataset and evaluated across dataset regimes (100%→10%). We further test generalization on BraTS and quantify robustness to resolution changes, acquisition noise, and modality shift (T1→FLAIR). To support clinical trust, we add visual explanations (Grad-CAM/saliency) and report per-class results (confusion matrices). A fairness-aligned protocol (shared splits, optimizer, early stopping) and a complexity analysis (parameters/FLOPs) enable balanced comparison. With full data, Convolutional Neural Networks (CNNs)/Residual Networks (ResNets) perform strongly but degrade with 10% data; Model-Agnostic Meta-Learning (MAML) retains competitive performance (AUC-ROC ≥ 0.9595 at 10%). Under cross-dataset validation (BraTS), FSL—particularly MAML—shows smaller performance drops than CNN/ResNet. Variability tests reveal FSL’s relative robustness to down-resolution and noise, although modality shift remains challenging for all models. Interpretability maps confirm correct activations on tumor regions in true positives and explain systematic errors (e.g., “no tumor”→pituitary). Conclusion: FSL provides accurate, data-efficient, and comparatively robust tumor classification under distribution shift. The added per-class analysis, interpretability, and complexity metrics strengthen clinical relevance and transparency. Full article
(This article belongs to the Special Issue Machine Learning Models and Algorithms for Image Processing)
Show Figures

Figure 1

32 pages, 2020 KB  
Article
From Trends to Insights: A Text Mining Analysis of Solar Energy Forecasting (2017–2023)
by Mohammed Asloune, Gilles Notton and Cyril Voyant
Energies 2025, 18(19), 5231; https://doi.org/10.3390/en18195231 - 1 Oct 2025
Abstract
This study aims to highlight key figures and organizations in solar energy forecasting research, including the most prominent authors, journals, and countries. It also clarifies commonly used abbreviations in the field, with a focus on forecasting methods and techniques, the form and type [...] Read more.
This study aims to highlight key figures and organizations in solar energy forecasting research, including the most prominent authors, journals, and countries. It also clarifies commonly used abbreviations in the field, with a focus on forecasting methods and techniques, the form and type of solar energy forecasting outputs, and the associated error metrics. Building on previous research that analyzed data up to 2017, the study updates findings to include information through 2023, incorporating metadata from 500 articles to identify key figures and organizations, along with 276 full-text articles analyzed for abbreviations. The application of text mining offers a concise yet comprehensive overview of the latest trends and insights in solar energy forecasting. The key findings of this study are threefold: First, China, followed by the United States of America and India, is the leading country in solar energy forecasting research, with shifts observed compared to the pre-2017 period. Second, numerous new abbreviations related to machine learning, particularly deep learning, have emerged in solar energy forecasting since before 2017, with Long Short-Term Memory, Convolutional Neural Networks, and Recurrent Neural Networks the most prominent. Finally, deterministic error metrics are mentioned nearly 11 times more frequently than probabilistic ones. Furthermore, perspectives on the practices and approaches of solar energy forecasting companies are also examined. Full article
(This article belongs to the Special Issue Solar Energy Utilization Toward Sustainable Urban Futures)
20 pages, 162180 KB  
Article
Annotation-Efficient and Domain-General Segmentation from Weak Labels: A Bounding Box-Guided Approach
by Ammar M. Okran, Hatem A. Rashwan, Sylvie Chambon and Domenec Puig
Electronics 2025, 14(19), 3917; https://doi.org/10.3390/electronics14193917 - 1 Oct 2025
Abstract
Manual pixel-level annotation remains a major bottleneck in deploying deep learning models for dense prediction and semantic segmentation tasks across domains. This challenge is especially pronounced in applications involving fine-scale structures, such as cracks in infrastructure or lesions in medical imaging, where annotations [...] Read more.
Manual pixel-level annotation remains a major bottleneck in deploying deep learning models for dense prediction and semantic segmentation tasks across domains. This challenge is especially pronounced in applications involving fine-scale structures, such as cracks in infrastructure or lesions in medical imaging, where annotations are time-consuming, expensive, and subject to inter-observer variability. To address these challenges, this work proposes a weakly supervised and annotation-efficient segmentation framework that integrates sparse bounding-box annotations with a limited subset of strong (pixel-level) labels to train robust segmentation models. The fundamental element of the framework is a lightweight Bounding Box Encoder that converts weak annotations into multi-scale attention maps. These maps guide a ConvNeXt-Base encoder, and a lightweight U-Net–style convolutional neural network (CNN) decoder—using nearest-neighbor upsampling and skip connections—reconstructs the final segmentation mask. This design enables the model to focus on semantically relevant regions without relying on full supervision, drastically reducing annotation cost while maintaining high accuracy. We validate our framework on two distinct domains, road crack detection and skin cancer segmentation, demonstrating that it achieves performance comparable to fully supervised segmentation models using only 10–20% of strong annotations. Given the ability of the proposed framework to generalize across varied visual contexts, it has strong potential as a general annotation-efficient segmentation tool for domains where strong labeling is costly or infeasible. Full article
Show Figures

Figure 1

22 pages, 1797 KB  
Article
A Novel Hybrid Deep Learning–Probabilistic Framework for Real-Time Crash Detection from Monocular Traffic Video
by Reşat Buğra Erkartal and Atınç Yılmaz
Appl. Sci. 2025, 15(19), 10523; https://doi.org/10.3390/app151910523 - 29 Sep 2025
Abstract
The rapid evolution of autonomous vehicle technologies has amplified the need for crash detection that operates robustly under complex traffic conditions with minimal latency. We propose a hybrid temporal hierarchy that augments a Region-based Convolutional Neural Network (R-CNN) with an adaptive time-variant Kalman [...] Read more.
The rapid evolution of autonomous vehicle technologies has amplified the need for crash detection that operates robustly under complex traffic conditions with minimal latency. We propose a hybrid temporal hierarchy that augments a Region-based Convolutional Neural Network (R-CNN) with an adaptive time-variant Kalman filter (with total-variation prior), a Hidden Markov Model (HMM) for state stabilization, and a lightweight Artificial Neural Network (ANN) for learned temporal refinement, enabling real-time crash detection from monocular video. Evaluated on simulated traffic in CARLA and real-world driving in Istanbul, the full temporal stack achieves the best precision–recall balance, yielding 83.47% F1 offline and 82.57% in real time (corresponding to 94.5% and 91.2% detection accuracy, respectively). Ablations are consistent and interpretable: removing the HMM reduces F1 by 1.85–2.16 percentage points (pp), whereas removing the ANN has a larger impact of 2.94–4.58 pp, indicating that the ANN provides the largest marginal gains—especially under real-time constraints. The transition from offline to real time incurs a modest overall loss (−0.90 pp F1), driven more by recall than precision. Compared to strong single-frame baselines, YOLOv10 attains 82.16% F1 and a real-time Transformer detector reaches 82.41% F1, while our full temporal stack remains slightly ahead in real time and offers a more favorable precision–recall trade-off. Notably, integrating the ANN into the HMM-based pipeline improves accuracy by 2.2%, while the time-variant Kalman configuration reduces detection lag by approximately 0.5 s—an improvement that directly addresses the human reaction time gap. Under identical conditions, the best RCNN-based configuration yields AP@0.50 ≈ 0.79 with an end-to-end latency of 119 ± 21 ms per frame (~8–9 FPS). Overall, coupling deep learning with probabilistic reasoning yields additive temporal benefits and advances deployable, camera-only crash detection that is cost-efficient and scalable for intelligent transportation systems. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

29 pages, 7711 KB  
Article
Fundamentals of Controlled Demolition in Structures: Real-Life Applications, Discrete Element Methods, Monitoring, and Artificial Intelligence-Based Research Directions
by Julide Yuzbasi
Buildings 2025, 15(19), 3501; https://doi.org/10.3390/buildings15193501 - 28 Sep 2025
Abstract
Controlled demolition is a critical engineering practice that enables the safe and efficient dismantling of structures while minimizing risks to the surrounding environment. This study presents, for the first time, a detailed, structured framework for understanding the fundamental principles of controlled demolition by [...] Read more.
Controlled demolition is a critical engineering practice that enables the safe and efficient dismantling of structures while minimizing risks to the surrounding environment. This study presents, for the first time, a detailed, structured framework for understanding the fundamental principles of controlled demolition by outlining key procedures, methodologies, and directions for future research. Through original, carefully designed charts and full-scale numerical simulations, including two 23-story building scenarios with different delay and blasting sequences, this paper provides real-life insights into the effects of floor-to-floor versus axis-by-axis delays on structural collapse behavior, debris spread, and toppling control. Beyond traditional techniques, this study explores how emerging technologies, such as real-time structural monitoring via object tracking, LiDAR scanning, and Unmanned Aerial Vehicle (UAV)-based inspections, can be further advanced through the integration of artificial intelligence (AI). The potential Deep learning (DL) and Machine learning (ML)-based applications of tools like Convolutional Neural Network (CNN)-based digital twins, YOLO object detection, and XGBoost classifiers are highlighted as promising avenues for future research. These technologies could support real-time decision-making, automation, and risk assessment in demolition scenarios. Furthermore, vision-language models such as SAM and Grounding DINO are discussed as enabling technologies for real-time risk assessment, anomaly detection, and adaptive control. By sharing insights from full-scale observations and proposing a forward-looking analytical framework, this work lays a foundation for intelligent and resilient demolition practices. Full article
(This article belongs to the Section Building Structures)
Show Figures

Figure 1

14 pages, 1932 KB  
Article
Skin Cancer Detection and Classification Through Medical Image Analysis Using EfficientNet
by Sima Das and Rishabh Kumar Addya
NDT 2025, 3(4), 23; https://doi.org/10.3390/ndt3040023 - 26 Sep 2025
Abstract
Skin cancer is one of the most prevalent and potentially lethal cancers worldwide, highlighting the need for accurate and timely diagnosis. Convolutional neural networks (CNNs) have demonstrated strong potential in automating skin lesion classification. In this study, we propose a multi-class classification model [...] Read more.
Skin cancer is one of the most prevalent and potentially lethal cancers worldwide, highlighting the need for accurate and timely diagnosis. Convolutional neural networks (CNNs) have demonstrated strong potential in automating skin lesion classification. In this study, we propose a multi-class classification model using EfficientNet-B0, a lightweight yet powerful CNN architecture, trained on the HAM10000 dermoscopic image dataset. All images were resized to 224 × 224 pixels and normalized using ImageNet statistics to ensure compatibility with the pre-trained network. Data augmentation and preprocessing addressed class imbalance, resulting in a balanced dataset of 7512 images across seven diagnostic categories. The baseline model achieved 77.39% accuracy, which improved to 89.36% with transfer learning by freezing the convolutional base and training only the classification layer. Full network fine-tuning with test-time augmentation increased the accuracy to 96%, and the final model reached 97.15% when combined with Monte Carlo dropout. These results demonstrate EfficientNet-B0’s effectiveness for automated skin lesion classification and its potential as a clinical decision support tool. Full article
Show Figures

Figure 1

23 pages, 3115 KB  
Article
Deep Learning-Based Prediction of Multi-Species Leaf Pigment Content Using Hyperspectral Reflectance
by Ziyu Wang and Duanyang Xu
Remote Sens. 2025, 17(19), 3293; https://doi.org/10.3390/rs17193293 - 25 Sep 2025
Abstract
Leaf pigment composition and concentration are crucial indicators of plant physiological status, photosynthetic capacity, and overall ecosystem health. While spectroscopy techniques show promise for monitoring vegetation growth, phenology, and stress, accurately estimating leaf pigments remains challenging due to the complex reflectance properties across [...] Read more.
Leaf pigment composition and concentration are crucial indicators of plant physiological status, photosynthetic capacity, and overall ecosystem health. While spectroscopy techniques show promise for monitoring vegetation growth, phenology, and stress, accurately estimating leaf pigments remains challenging due to the complex reflectance properties across diverse tree species. This study introduces a novel approach using a two-dimensional convolutional neural network (2D-CNN) coupled with a genetic algorithm (GA) to predict leaf pigment content, including chlorophyll a and b content (Cab), carotenoid content (Car), and anthocyanin content (Canth). Leaf reflectance and biochemical content measurements taken from 28 tree species were used in this study. The reflectance spectra ranging from 400 nm to 800 nm were encoded as 2D matrices with different sizes to train the 2D-CNN and compared with the one-dimensional convolutional neural network (1D-CNN). The results show that the 2D-CNN model (nRMSE = 11.71–31.58%) achieved higher accuracy than the 1D-CNN model (nRMSE = 12.79–55.34%) in predicting leaf pigment contents. For the 2D-CNN models, Cab achieved the best estimation accuracy with an nRMSE value of 11.71% (R2 = 0.92, RMSE = 6.10 µg/cm2), followed by Car (R2 = 0.84, RMSE = 1.03 µg/cm2, nRMSE = 12.29%) and Canth (R2 = 0.89, RMSE = 0.35 µg/cm2, nRMSE = 31.58%). Both 1D-CNN and 2D-CNN models coupled with GA using a subset of the spectrum produced higher prediction accuracy in all pigments than those using the full spectrum. Additionally, the generalization of 2D-CNN is higher than that of 1D-CNN. This study highlights the potential of 2D-CNN approaches for accurate prediction of leaf pigment content from spectral reflectance data, offering a promising tool for advanced vegetation monitoring. Full article
(This article belongs to the Section Forest Remote Sensing)
Show Figures

Figure 1

20 pages, 1844 KB  
Article
Detection of Rice Prolamin and Glutelin Content Using Hyperspectral Imaging Combined with Feature Selection Algorithms and Multivariate Regression Models
by Chu Zhang, Zhongjie Tang, Xiaojing Tan, Hengnian Qi, Xincheng Zhang and Shanlin Ma
Foods 2025, 14(19), 3304; https://doi.org/10.3390/foods14193304 - 24 Sep 2025
Viewed by 126
Abstract
Prolamin and glutelin are the major constituents of rice protein. The rapid and non-destructive detection of prolamin and glutelin content is conducive to the accurate assessment of rice quality. In this study, hyperspectral imaging combined with regression models and feature wavelength selection was [...] Read more.
Prolamin and glutelin are the major constituents of rice protein. The rapid and non-destructive detection of prolamin and glutelin content is conducive to the accurate assessment of rice quality. In this study, hyperspectral imaging combined with regression models and feature wavelength selection was employed to detect the rice prolamin and glutelin content. Feature wavelength selection was achieved using the successive projections algorithm (SPA), competitive adaptive reweighted sampling (CARS), and convolutional neural network (CNN)-based Gradient-weighted Class Activation Mapping++ (GradCAM++). Partial least squares regression (PLSR), support vector regression (SVR), back-propagation neural network (BPNN), and CNN models were established using the full spectra and the feature wavelengths. The BPNN models showed the best prediction performance for prolamin and glutelin. The optimal BPNN models achieved a correlation coefficient (r) greater than 0.8 for both proteins. Performance differences were observed between models using feature wavelengths and those using the full spectra. The GradCAM++ method was used to select feature wavelengths with different threshold values, and the performance of different threshold values were compared. The results demonstrated that hyperspectral imaging with multivariate data analysis was feasible for predicting the rice prolamin and glutelin content. This study provided a methodological reference for detecting prolamin and glutelin in rice, as well as the other protein types. Full article
(This article belongs to the Section Food Analytical Methods)
Show Figures

Figure 1

29 pages, 7933 KB  
Article
Hybrid Ship Design Optimization Framework Integrating a Dual-Mode CFD–Surrogate Mechanism
by Yicun Dong, Lin Du and Guangnian Li
Appl. Sci. 2025, 15(19), 10318; https://doi.org/10.3390/app151910318 - 23 Sep 2025
Viewed by 198
Abstract
Reducing hydrodynamic resistance remains a central concern in modern ship design. The Simulation-Based Design technique offers high-fidelity optimization through computational fluid dynamics, but this comes at the cost of computational efficiency. In contrast, surrogate models trained offline can accelerate the process but often [...] Read more.
Reducing hydrodynamic resistance remains a central concern in modern ship design. The Simulation-Based Design technique offers high-fidelity optimization through computational fluid dynamics, but this comes at the cost of computational efficiency. In contrast, surrogate models trained offline can accelerate the process but often compromise on accuracy. To address this issue, this study proposes a hybrid optimization framework connecting a computational fluid dynamics solver and a convolutional neural network surrogate model within a dual-mode mechanism. By comparing selected computational fluid dynamics evaluations with surrogate predictions during each iteration, the system is able to balance the precision and efficiency adaptively. The framework integrates a particle swarm optimizer, a free-form deformation modeler, and a dual-mode solver. Case studies on three benchmark hulls including KCS, KVLCC1, and JBC have shown 3.40%, 3.95%, and 2.74% resistance reduction, respectively, with computation efficiency gains exceeding 44% compared to the traditional Simulation-Based Design process using full computational fluid dynamics. This study provides a practical attempt to enhance the efficiency of hull form optimization while maintaining accuracy. Full article
(This article belongs to the Section Marine Science and Engineering)
Show Figures

Figure 1

21 pages, 1602 KB  
Article
A Forensic Odontology Application: Impact of Image Quality on CNNs for Dental Status Analysis from Orthopantomograms
by Ajla Zymber Çeshko, Ivana Savić Pavičin, Denis Milošević, Luka Banjšak, Marko Subašić and Marin Vodanović
Appl. Sci. 2025, 15(18), 10265; https://doi.org/10.3390/app151810265 - 21 Sep 2025
Viewed by 218
Abstract
Artificial Intelligence, especially Convolutional Neural Networks (CNN), is gaining importance in health sciences, including forensic odontology. This study aimed to systematically analyze elements for automated dental status registration on OPGs using CNNs, on different image segments and resolutions. A dataset of 1400 manually [...] Read more.
Artificial Intelligence, especially Convolutional Neural Networks (CNN), is gaining importance in health sciences, including forensic odontology. This study aimed to systematically analyze elements for automated dental status registration on OPGs using CNNs, on different image segments and resolutions. A dataset of 1400 manually annotated digital OPGs was divided into train, test, and validation sets (75%–12.5%–12.5%). Pre-trained and from-scratch models were developed and evaluated on images from full OPGs to individual and segmented teeth and sizes from 256 px to 1820 px. Performance was measured by Sørensen–Dice coefficient for segmentation and mean average precision (mAP) for detection. For segmentation, the UNet Big model was the most successful, using segmented or individual images, achieving 89.14% for crown and 84.90% for fillings, and UNet with 79.09% for root canal fillings. Caries presented a significant challenge, with the UNet model achieving the highest score of 64.68%. In detection, YOLOv5x6, trained from scratch, achieved the highest mAP of 98.02% on 1820 px images. Larger image resolutions and individual tooth inputs significantly improved performance. This study confirms the success of CNN models in specific tasks on OPGs. Image quality and input (individual tooth, resolutions above 640 px) critically influenced model competence. Further research with from-scratch models, higher resolutions, and smaller image segments is recommended. Full article
(This article belongs to the Special Issue Deep Learning Applied in Dentistry: Challenges and Prospects)
Show Figures

Figure 1

Back to TopTop