Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (167)

Search Parameters:
Keywords = pre-trained AlexNet

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
26 pages, 2445 KB  
Article
Image-Based Deep Learning Approach for Drilling Kick Risk Prediction
by Wei Liu, Yuansen Wei, Jiasheng Fu, Qihao Li, Yi Zou, Tao Pan and Zhaopeng Zhu
Processes 2025, 13(10), 3251; https://doi.org/10.3390/pr13103251 - 13 Oct 2025
Viewed by 318
Abstract
As oil and gas exploration and development advance into deep and ultra-deep areas, kick accidents are becoming more frequent during drilling operations, posing a serious threat to construction safety. Traditional kick monitoring methods are limited in their multivariate coupling modeling. These models rely [...] Read more.
As oil and gas exploration and development advance into deep and ultra-deep areas, kick accidents are becoming more frequent during drilling operations, posing a serious threat to construction safety. Traditional kick monitoring methods are limited in their multivariate coupling modeling. These models rely too heavily on single-feature weights, making them prone to misjudgment. Therefore, this paper proposes a drilling kick risk prediction method based on image modality. First, a sliding window mechanism is used to slice key drilling parameters in time series to extract multivariate data for continuous time periods. Second, data processing is performed to construct joint logging curve image samples. Then, classical CNN models such as VGG16 and ResNet are used to train and classify image samples; finally, the performance of the model on a number of indicators is evaluated and compared with different CNN and temporal neural network models. Finally, the model’s performance is evaluated across multiple metrics and compared with CNN and time series neural network models of different structures. Experimental results show that the image-based VGG16 model outperforms typical convolutional neural network models such as AlexNet, ResNet, and EfficientNet in overall performance, and significantly outperforms LSTM and GRU time series models in classification accuracy and comprehensive discriminative power. Compared to LSTM, the recall rate increased by 23.8% and the precision increased by 5.8%, demonstrating that its convolutional structure possesses stronger perception and discriminative capabilities in extracting local spatiotemporal features and recognizing patterns, enabling more accurate identification of kick risks. Furthermore, the pre-trained VGG16 model achieved an 8.69% improvement in accuracy compared to the custom VGG16 model, fully demonstrating the effectiveness and generalization advantages of transfer learning in small-sample engineering problems and providing feasibility support for model deployment and engineering applications. Full article
(This article belongs to the Section Energy Systems)
Show Figures

Figure 1

17 pages, 1602 KB  
Article
Deep Transfer Learning for Automatic Analysis of Ignitable Liquid Residues in Fire Debris Samples
by Ting-Yu Huang and Jorn Chi Chung Yu
Chemosensors 2025, 13(9), 320; https://doi.org/10.3390/chemosensors13090320 - 26 Aug 2025
Viewed by 781
Abstract
Interpreting chemical analysis results to identify ignitable liquid (IL) residues in fire debris samples is challenging, owing to the complex chemical composition of ILs and the diverse sample matrices. This work investigated a transfer learning approach with convolutional neural networks (CNNs), pre-trained for [...] Read more.
Interpreting chemical analysis results to identify ignitable liquid (IL) residues in fire debris samples is challenging, owing to the complex chemical composition of ILs and the diverse sample matrices. This work investigated a transfer learning approach with convolutional neural networks (CNNs), pre-trained for image recognition, to classify gas chromatography and mass spectrometry (GC/MS) data transformed into scalogram images. A small data set containing neat gasoline samples with diluted concentrations and burned Nylon carpets with varying weights was prepared to retrain six CNNs: GoogLeNet, AlexNet, SqueezeNet, VGG-16, ResNet-50, and Inception-v3. The classification tasks involved two classes: “positive of gasoline” and “negative of gasoline.” The results demonstrated that the CNNs performed very well in predicting the trained class data. When predicting untrained intra-laboratory class data, GoogLeNet had the highest accuracy (0.98 ± 0.01), precision (1.00 ± 0.01), sensitivity (0.97 ± 0.01), and specificity (1.00 ± 0.00). When predicting untrained inter-laboratory class data, GoogLeNet exhibited a sensitivity of 1.00 ± 0.00, while ResNet-50 achieved 0.94 ± 0.01 for neat gasoline. For simulated fire debris samples, both models attained sensitivities of 0.86 ± 0.02 and 0.89 ± 0.02, respectively. The new deep transfer learning approach enables automated pattern recognition in GC/MS data, facilitates high-throughput forensic analysis, and improves consistency in interpretation across various laboratories, making it a valuable tool for fire debris analysis. Full article
Show Figures

Figure 1

19 pages, 15592 KB  
Technical Note
Integration of Convolutional Neural Networks and UAV-Derived DEM for the Automatic Classification of Benthic Habitats in Shallow Water Environments
by Hassan Mohamed and Kazuo Nadaoka
Remote Sens. 2025, 17(17), 2928; https://doi.org/10.3390/rs17172928 - 23 Aug 2025
Viewed by 861
Abstract
Benthic habitats are highly complex and diverse ecosystems that are increasingly threatened by human-induced stressors and the impacts of climate change. Therefore, accurate classification and mapping of these marine habitats are essential for effective monitoring and management. In recent years, Unmanned Aerial Vehicles [...] Read more.
Benthic habitats are highly complex and diverse ecosystems that are increasingly threatened by human-induced stressors and the impacts of climate change. Therefore, accurate classification and mapping of these marine habitats are essential for effective monitoring and management. In recent years, Unmanned Aerial Vehicles (UAVs) have been increasingly used to expand the spatial coverage of surveys and to produce high-resolution imagery. These images can be processed using photogrammetry-based techniques to generate high-resolution digital elevation models (DEMs) and orthomosaics. In this study, we demonstrate that integrating descriptors extracted from pre-trained Convolutional Neural Networks (CNNs) with geomorphometric attributes derived from DEMs significantly enhances the accuracy of automatic benthic habitat classification. To assess this integration, we analyzed orthomosaics and DEMs generated from UAV imagery across three shallow reef zones along the Red Sea coast of Saudi Arabia. Furthermore, we tested various combinations of feature layers from pre-trained CNNs—including ResNet-50, VGG16, and AlexNet—together with several geomorphometric variables to evaluate classification accuracy. The results showed that features extracted from the ResNet-50 FC1000 layer, when combined with twelve geomorphometric attributes based on curvature, slope, the Topographic Ruggedness Index (TRI), and DEM-derived heights, achieved the highest overall accuracies. Moreover, training a Support Vector Machine (SVM) classifier using both pre-trained ResNet-50 features and geomorphometric variables led to an improvement in overall accuracy of up to 5%, compared to using ResNet-50 features alone. The proposed integration effectively improves the automation and accuracy of benthic habitat mapping processes. Full article
(This article belongs to the Section Ocean Remote Sensing)
Show Figures

Graphical abstract

23 pages, 1938 KB  
Article
Algorithmic Silver Trading via Fine-Tuned CNN-Based Image Classification and Relative Strength Index-Guided Price Direction Prediction
by Yahya Altuntaş, Fatih Okumuş and Adnan Fatih Kocamaz
Symmetry 2025, 17(8), 1338; https://doi.org/10.3390/sym17081338 - 16 Aug 2025
Viewed by 1241
Abstract
Predicting short-term buy and sell signals in financial markets remains a significant challenge for algorithmic trading. This difficulty stems from the data’s inherent volatility and noise, which often leads to spurious signals and poor trading performance. This paper presents a novel algorithmic trading [...] Read more.
Predicting short-term buy and sell signals in financial markets remains a significant challenge for algorithmic trading. This difficulty stems from the data’s inherent volatility and noise, which often leads to spurious signals and poor trading performance. This paper presents a novel algorithmic trading model for silver that combines fine-tuned Convolutional Neural Networks (CNNs) with a decision filter based on the Relative Strength Index (RSI). The technique allows for the prediction of buy and sell points by turning time series data into chart images. Daily silver price per ounce data were turned into chart images using technical analysis indicators. Four pre-trained CNNs, namely AlexNet, VGG16, GoogLeNet, and ResNet-50, were fine-tuned using the generated image dataset to find the best architecture based on classification and financial performance. The models were evaluated using walk-forward validation with an expanding window. This validation method made the tests more realistic and the performance evaluation more robust under different market conditions. Fine-tuned VGG16 with the RSI filter had the best cost-adjusted profitability, with a cumulative return of 115.03% over five years. This was nearly double the 61.62% return of a buy-and-hold strategy. This outperformance is especially impressive because the evaluation period was mostly upward, which makes it harder to beat passive benchmarks. Adding the RSI filter also helped models make more disciplined decisions. This reduced transactions with low confidence. In general, the results show that pre-trained CNNs fine-tuned on visual representations, when supplemented with domain-specific heuristics, can provide strong and cost-effective solutions for algorithmic trading, even when realistic cost assumptions are used. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

23 pages, 5155 KB  
Article
Enhancing Early Detection of Diabetic Foot Ulcers Using Deep Neural Networks
by A. Sharaf Eldin, Asmaa S. Ahmoud, Hanaa M. Hamza and Hanin Ardah
Diagnostics 2025, 15(16), 1996; https://doi.org/10.3390/diagnostics15161996 - 9 Aug 2025
Viewed by 1059
Abstract
Background/Objectives: Diabetic foot ulcers (DFUs) remain a critical complication of diabetes, with high rates of amputation when not diagnosed early. Despite advancements in medical imaging, current DFU detection methods are often limited by their computational complexity, poor generalizability, and delayed diagnostic performance. [...] Read more.
Background/Objectives: Diabetic foot ulcers (DFUs) remain a critical complication of diabetes, with high rates of amputation when not diagnosed early. Despite advancements in medical imaging, current DFU detection methods are often limited by their computational complexity, poor generalizability, and delayed diagnostic performance. This study presents a novel hybrid diagnostic framework that integrates traditional feature extraction methods with deep learning (DL) to improve the early real-time computer-aided detection (CAD) of DFUs. Methods: The proposed model leverages plantar thermograms to detect early thermal asymmetries associated with DFUs. It uniquely combines the oriented FAST and rotated BRIEF (ORB) algorithm with the Bag of Features (BOF) method to extract robust handcrafted features while also incorporating deep features from pretrained convolutional neural networks (ResNet50, AlexNet, and EfficientNet). These features were fused and input into a lightweight deep neural network (DNN) classifier designed for binary classification. Results: Our model demonstrated an accuracy of 98.51%, precision of 100%, sensitivity of 98.98%, and AUC of 1.00 in a publicly available plantar thermogram dataset (n = 1670 images). An ablation study confirmed the superiority of ORB + DL fusion over standalone approaches. Unlike previous DFU detection models that rely solely on either handcrafted or deep features, our study presents the first lightweight hybrid framework that integrates ORB-based descriptors with deep CNN representations (e.g., ResNet50 and EfficientNet). Compared with recent state-of-the-art models, such as DFU_VIRNet and DFU_QUTNet, our approach achieved a higher diagnostic performance (accuracy = 98.51%, AUC = 1.00) while maintaining real-time capability and a lower computational overhead, making it highly suitable for clinical deployment. Conclusions: This study proposes the first integration of ORB-based handcrafted features with deep neural representations for DFU detection from thermal images. The model delivers high accuracy, robustness to noise, and real-time capabilities, outperforming existing state-of-the-art approaches and demonstrating strong potential for clinical deployment. Full article
(This article belongs to the Topic Machine Learning and Deep Learning in Medical Imaging)
Show Figures

Figure 1

29 pages, 4203 KB  
Article
A Lightweight Deep Learning and Sorting-Based Smart Parking System for Real-Time Edge Deployment
by Muhammad Omair Khan, Muhammad Asif Raza, Md Ariful Islam Mozumder, Ibad Ullah Azam, Rashadul Islam Sumon and Hee Cheol Kim
AppliedMath 2025, 5(3), 79; https://doi.org/10.3390/appliedmath5030079 - 28 Jun 2025
Viewed by 1687
Abstract
As cities grow denser, the demand for efficient parking systems becomes more critical to reduce traffic congestion, fuel consumption, and environmental impact. This paper proposes a smart parking solution that combines deep learning and algorithmic sorting to identify the nearest available parking slot [...] Read more.
As cities grow denser, the demand for efficient parking systems becomes more critical to reduce traffic congestion, fuel consumption, and environmental impact. This paper proposes a smart parking solution that combines deep learning and algorithmic sorting to identify the nearest available parking slot in real time. The system uses several pre-trained convolutional neural network (CNN) models—VGG16, ResNet50, Xception, LeNet, AlexNet, and MobileNet—along with a lightweight custom CNN architecture, all trained on a custom parking dataset. These models are integrated into a mobile application that allows users to view and request nearby parking spaces. A merge sort algorithm ranks available slots based on proximity to the user. The system is validated using benchmark datasets (CNR-EXT and PKLot), demonstrating high accuracy across diverse weather conditions. The proposed system shows how applied mathematical models and deep learning can improve urban mobility through intelligent infrastructure. Full article
Show Figures

Figure 1

33 pages, 5602 KB  
Article
CELM: An Ensemble Deep Learning Model for Early Cardiomegaly Diagnosis in Chest Radiography
by Erdem Yanar, Fırat Hardalaç and Kubilay Ayturan
Diagnostics 2025, 15(13), 1602; https://doi.org/10.3390/diagnostics15131602 - 25 Jun 2025
Cited by 3 | Viewed by 976
Abstract
Background/Objectives: Cardiomegaly—defined as the abnormal enlargement of the heart—is a key radiological indicator of various cardiovascular conditions. Early detection is vital for initiating timely clinical intervention and improving patient outcomes. This study investigates the application of deep learning techniques for the automated diagnosis [...] Read more.
Background/Objectives: Cardiomegaly—defined as the abnormal enlargement of the heart—is a key radiological indicator of various cardiovascular conditions. Early detection is vital for initiating timely clinical intervention and improving patient outcomes. This study investigates the application of deep learning techniques for the automated diagnosis of cardiomegaly from chest X-ray (CXR) images, utilizing both convolutional neural networks (CNNs) and Vision Transformers (ViTs). Methods: We assembled one of the largest and most diverse CXR datasets to date, combining posteroanterior (PA) images from PadChest, NIH CXR, VinDr-CXR, and CheXpert. Multiple pre-trained CNN architectures (VGG16, ResNet50, InceptionV3, DenseNet121, DenseNet201, and AlexNet), as well as Vision Transformer models, were trained and compared. In addition, we introduced a novel stacking-based ensemble model—Combined Ensemble Learning Model (CELM)—that integrates complementary CNN features via a meta-classifier. Results: The CELM achieved the highest diagnostic performance, with a test accuracy of 92%, precision of 99%, recall of 89%, F1-score of 0.94, specificity of 92.0%, and AUC of 0.90. These results highlight the model’s high agreement with expert annotations and its potential for reliable clinical use. Notably, Vision Transformers offered competitive performance, suggesting their value as complementary tools alongside CNNs. Conclusions: With further validation, the proposed CELM framework may serve as an efficient and scalable decision-support tool for cardiomegaly screening, particularly in resource-limited settings such as intensive care units (ICUs) and emergency departments (EDs), where rapid and accurate diagnosis is imperative. Full article
(This article belongs to the Special Issue Machine-Learning-Based Disease Diagnosis and Prediction)
Show Figures

Figure 1

12 pages, 2782 KB  
Article
Platelets Image Classification Through Data Augmentation: A Comparative Study of Traditional Imaging Augmentation and GAN-Based Synthetic Data Generation Techniques Using CNNs
by Itunuoluwa Abidoye, Frances Ikeji, Charlie A. Coupland, Simon D. J. Calaminus, Nick Sander and Eva Sousa
J. Imaging 2025, 11(6), 183; https://doi.org/10.3390/jimaging11060183 - 4 Jun 2025
Cited by 1 | Viewed by 1616
Abstract
Platelets play a crucial role in diagnosing and detecting various diseases, influencing the progression of conditions and guiding treatment options. Accurate identification and classification of platelets are essential for these purposes. The present study aims to create a synthetic database of platelet images [...] Read more.
Platelets play a crucial role in diagnosing and detecting various diseases, influencing the progression of conditions and guiding treatment options. Accurate identification and classification of platelets are essential for these purposes. The present study aims to create a synthetic database of platelet images using Generative Adversarial Networks (GANs) and validate its effectiveness by comparing it with datasets of increasing sizes generated through traditional augmentation techniques. Starting from an initial dataset of 71 platelet images, the dataset was expanded to 141 images (Level 1) using random oversampling and basic transformations and further to 1463 images (Level 2) through extensive augmentation (rotation, shear, zoom). Additionally, a synthetic dataset of 300 images was generated using a Wasserstein GAN with Gradient Penalty (WGAN-GP). Eight pre-trained deep learning models (DenseNet121, DenseNet169, DenseNet201, VGG16, VGG19, InceptionV3, InceptionResNetV2, and AlexNet) and two custom CNNs were evaluated across these datasets. Performance was measured using accuracy, precision, recall, and F1-score. On the extensively augmented dataset (Level 2), InceptionV3 and InceptionResNetV2 reached 99% accuracy and 99% precision/recall/F1-score, while DenseNet201 closely followed, with 98% accuracy, precision, recall and F1-score. GAN-augmented data further improved DenseNet’s performance, demonstrating the potential of GAN-generated images in enhancing platelet classification, especially where data are limited. These findings highlight the benefits of combining traditional and GAN-based augmentation techniques to improve classification performance in medical imaging tasks. Full article
(This article belongs to the Topic Machine Learning and Deep Learning in Medical Imaging)
Show Figures

Figure 1

28 pages, 3438 KB  
Article
Optimizing Remote Sensing Image Retrieval Through a Hybrid Methodology
by Sujata Alegavi and Raghvendra Sedamkar
J. Imaging 2025, 11(6), 179; https://doi.org/10.3390/jimaging11060179 - 28 May 2025
Cited by 1 | Viewed by 819
Abstract
The contemporary challenge in remote sensing lies in the precise retrieval of increasingly abundant and high-resolution remotely sensed images (RS image) stored in expansive data warehouses. The heightened spatial and spectral resolutions, coupled with accelerated image acquisition rates, necessitate advanced tools for effective [...] Read more.
The contemporary challenge in remote sensing lies in the precise retrieval of increasingly abundant and high-resolution remotely sensed images (RS image) stored in expansive data warehouses. The heightened spatial and spectral resolutions, coupled with accelerated image acquisition rates, necessitate advanced tools for effective data management, retrieval, and exploitation. The classification of large-sized images at the pixel level generates substantial data, escalating the workload and search space for similarity measurement. Semantic-based image retrieval remains an open problem due to limitations in current artificial intelligence techniques. Furthermore, on-board storage constraints compel the application of numerous compression algorithms to reduce storage space, intensifying the difficulty of retrieving substantial, sensitive, and target-specific data. This research proposes an innovative hybrid approach to enhance the retrieval of remotely sensed images. The approach leverages multilevel classification and multiscale feature extraction strategies to enhance performance. The retrieval system comprises two primary phases: database building and retrieval. Initially, the proposed Multiscale Multiangle Mean-shift with Breaking Ties (MSMA-MSBT) algorithm selects informative unlabeled samples for hyperspectral and synthetic aperture radar images through an active learning strategy. Addressing the scaling and rotation variations in image capture, a flexible and dynamic algorithm, modified Deep Image Registration using Dynamic Inlier (IRDI), is introduced for image registration. Given the complexity of remote sensing images, feature extraction occurs at two levels. Low-level features are extracted using the modified Multiscale Multiangle Completed Local Binary Pattern (MSMA-CLBP) algorithm to capture local contexture features, while high-level features are obtained through a hybrid CNN structure combining pretrained networks (Alexnet, Caffenet, VGG-S, VGG-M, VGG-F, VGG-VDD-16, VGG-VDD-19) and a fully connected dense network. Fusion of low- and high-level features facilitates final class distinction, with soft thresholding mitigating misclassification issues. A region-based similarity measurement enhances matching percentages. Results, evaluated on high-resolution remote sensing datasets, demonstrate the effectiveness of the proposed method, outperforming traditional algorithms with an average accuracy of 86.66%. The hybrid retrieval system exhibits substantial improvements in classification accuracy, similarity measurement, and computational efficiency compared to state-of-the-art scene classification and retrieval methods. Full article
(This article belongs to the Topic Computational Intelligence in Remote Sensing: 2nd Edition)
Show Figures

Figure 1

21 pages, 26641 KB  
Article
A CNN-Based Method for Quantitative Assessment of Steel Microstructures in Welded Zones
by Cássio Danelon de Almeida, Thales Tozatto Filgueiras, Moisés Luiz Lagares, Bruno da Silva Macêdo, Camila Martins Saporetti, Matteo Bodini and Leonardo Goliatt
Fibers 2025, 13(5), 66; https://doi.org/10.3390/fib13050066 - 15 May 2025
Viewed by 2076
Abstract
The mechanical performance of metallic components is intrinsically linked to their microstructural features. However, the manual quantification of microconstituents in metallographic images remains a time-consuming and subjective task, often requiring over 15 min per image by a trained expert. To address this limitation, [...] Read more.
The mechanical performance of metallic components is intrinsically linked to their microstructural features. However, the manual quantification of microconstituents in metallographic images remains a time-consuming and subjective task, often requiring over 15 min per image by a trained expert. To address this limitation, this study proposes an automated approach for quantifying the microstructural constituents from low-carbon steel welded zone images using convolutional neural networks (CNNs). A dataset of 210 micrographs was expanded to 720 samples through data augmentation to improve model generalization. Two architectures (AlexNet and VGG16) were trained from scratch, while three pre-trained models (VGG19, InceptionV3, and Xception) were fine-tuned. Among these, VGG19 optimized with stochastic gradient descent (SGD) achieved the best predictive performance, with an R2 of 0.838, MAE of 5.01%, and RMSE of 6.88%. The results confirm the effectiveness of CNNs for reliable and efficient microstructure quantification, offering a significant contribution to computational metallography. Full article
Show Figures

Figure 1

20 pages, 6149 KB  
Article
Computer Vision and Transfer Learning for Grading of Egyptian Cotton Fibres
by Ahmed Rady, Oliver Fisher, Aly A. A. El-Banna, Haitham H. Emasih and Nicholas J. Watson
AgriEngineering 2025, 7(5), 127; https://doi.org/10.3390/agriengineering7050127 - 22 Apr 2025
Cited by 2 | Viewed by 1479
Abstract
Egyptian cotton fibres have worldwide recognition due to their distinct quality and luxurious textile products known by the “Egyptian Cotton“ label. However, cotton fibre trading in Egypt still depends on human grading of cotton quality, which is resource-intensive and faces challenges in terms [...] Read more.
Egyptian cotton fibres have worldwide recognition due to their distinct quality and luxurious textile products known by the “Egyptian Cotton“ label. However, cotton fibre trading in Egypt still depends on human grading of cotton quality, which is resource-intensive and faces challenges in terms of subjectivity and expertise requirements. This study investigates colour vision and transfer learning to classify the grade of five long (Giza 86, Giza 90, and Giza 94) and extra-long (Giza 87 and Giza 96) staple cotton cultivars. Five Convolutional Neural networks (CNNs)—AlexNet, GoogleNet, SqueezeNet, VGG16, and VGG19—were fine-tuned, optimised, and tested on independent datasets. The highest classifications were 75.7%, 85.0%, 80.0%, 77.1%, and 90.0% for Giza 86, Giza 87, Giza 90, Giza 94, and Giza 96, respectively, with F1-Scores ranging from 51.9–100%, 66.7–100%, 42.9–100%, 40.0–100%, and 80.0–100%. Among the CNNs, AlexNet, GoogleNet, and VGG19 outperformed the others. Fused CNN models further improved classification accuracy by up to 7.2% for all cultivars except Giza 87. These results demonstrate the feasibility of developing a fast, low-cost, and low-skilled vision system that overcomes the inconsistencies and limitations of manual grading in the early stages of cotton fibre trading in Egypt. Full article
Show Figures

Figure 1

15 pages, 4254 KB  
Proceeding Paper
A Custom Convolutional Neural Network Model-Based Bioimaging Technique for Enhanced Accuracy of Alzheimer’s Disease Detection
by Gogulamudi Pradeep Reddy, Duppala Rohan, Shaik Mohammed Abdul Kareem, Yellapragada Venkata Pavan Kumar, Kasaraneni Purna Prakash and Malathi Janapati
Eng. Proc. 2025, 87(1), 47; https://doi.org/10.3390/engproc2025087047 - 14 Apr 2025
Cited by 1 | Viewed by 805
Abstract
Alzheimer’s disease (AD), an intense neurological illness, severely impacts memory, behavior, and personality, posing a growing concern worldwide due to the aging population. Early and accurate detection is crucial as it enables preventive measures. However, current diagnostic methods are often inaccurate in identifying [...] Read more.
Alzheimer’s disease (AD), an intense neurological illness, severely impacts memory, behavior, and personality, posing a growing concern worldwide due to the aging population. Early and accurate detection is crucial as it enables preventive measures. However, current diagnostic methods are often inaccurate in identifying the disease in its early stages. Although deep learning-based bioimaging has shown promising results in medical image classification, challenges remain in achieving the highest accuracy for detecting AD. Existing approaches, such as ResNet50, VGG19, InceptionV3, and AlexNet have shown potential, but they often lack reliability and accuracy due to several issues. To address these gaps, this paper suggests a novel bioimaging technique by developing a custom Convolutional Neural Network (CNN) model for detecting AD. This model is designed with optimized layers to enhance feature extraction from medical images. The experiment’s first phase involved the construction of the custom CNN structure with three max-pooling layers, three convolutional layers, two dense layers, and one flattened layer. The Adam optimizer and categorical cross-entropy were adopted to compile the model. The model’s training was carried out on 100 epochs with the patience set to 10 epochs. The second phase involved augmentation of the dataset images and adding a dropout layer to the custom CNN model. Moreover, fine-tuned hyperparameters and advanced regularization methods were integrated to prevent overfitting. A comparative analysis of the proposed model with conventional models was performed on the dataset both before and after the data augmentation. The results validate that the proposed custom CNN model significantly overtakes pre-existing models, achieving the highest validation accuracy of 99.53% after data augmentation while maintaining the lowest validation loss of 0.0238. Its precision, recall, and F1 score remained consistently high across all classes, with perfect scores for the Moderate Demented and Non-Demented categories after augmentation, indicating superior classification capability. Full article
(This article belongs to the Proceedings of The 5th International Electronic Conference on Applied Sciences)
Show Figures

Figure 1

29 pages, 24098 KB  
Article
A Deep Transfer Learning-Based Visual Inspection System for Assembly Defects in Similar Types of Manual Tool Products
by Hong-Dar Lin, Hsiang-Ling Wu and Chou-Hsien Lin
Sensors 2025, 25(6), 1645; https://doi.org/10.3390/s25061645 - 7 Mar 2025
Cited by 1 | Viewed by 2428
Abstract
This study introduces an advanced inspection system for manual tool assembly, focusing on defect detection and classification in flex-head ratchet wrenches as a modern alternative to traditional inspection methods. Using a deep learning R-CNN approach with transfer learning, specifically utilizing the AlexNet architecture, [...] Read more.
This study introduces an advanced inspection system for manual tool assembly, focusing on defect detection and classification in flex-head ratchet wrenches as a modern alternative to traditional inspection methods. Using a deep learning R-CNN approach with transfer learning, specifically utilizing the AlexNet architecture, the system accurately identifies and classifies assembly defects across similar tools. This study demonstrates how a pre-trained defect detection model for older manual tool models can be efficiently adapted to new models with only moderate amounts of new samples and fine-tuning. Experimental evaluations at three assembly stations show that the AlexNet model achieves a classification accuracy of 98.67% at the station with the highest defect variety, outperforming the R-CNN model with randomly initialized weights. Even with a 40% reduction in sample size for new products, the AlexNet model maintains a classification accuracy of 98.66%. Additionally, compared to R-CNN, it improves average effectiveness by 9% and efficiency by 26% across all stations. A sensitivity analysis further reveals that the proposed method reduces training samples by 50% at 50% similarity while enhancing effectiveness by 13.06% and efficiency by 5.31%. Full article
Show Figures

Figure 1

15 pages, 1511 KB  
Article
Transfer Learning-Based Integration of Dual Imaging Modalities for Enhanced Classification Accuracy in Confocal Laser Endomicroscopy of Lung Cancer
by Mircea-Sebastian Șerbănescu, Liliana Streba, Alin Dragoș Demetrian, Andreea-Georgiana Gheorghe, Mădălin Mămuleanu, Daniel-Nicolae Pirici and Costin-Teodor Streba
Cancers 2025, 17(4), 611; https://doi.org/10.3390/cancers17040611 - 11 Feb 2025
Cited by 2 | Viewed by 1797
Abstract
Background/Objectives: Lung cancer remains the leading cause of cancer-related mortality, underscoring the need for improved diagnostic methods. This study seeks to enhance the classification accuracy of confocal laser endomicroscopy (pCLE) images for lung cancer by applying a dual transfer learning (TL) approach that [...] Read more.
Background/Objectives: Lung cancer remains the leading cause of cancer-related mortality, underscoring the need for improved diagnostic methods. This study seeks to enhance the classification accuracy of confocal laser endomicroscopy (pCLE) images for lung cancer by applying a dual transfer learning (TL) approach that incorporates histological imaging data. Methods: Histological samples and pCLE images, collected from 40 patients undergoing curative lung cancer surgeries, were selected to create 2 balanced datasets (800 benign and 800 malignant images each). Three CNN architectures—AlexNet, GoogLeNet, and ResNet—were pre-trained on ImageNet and re-trained on pCLE images (confocal TL) or using dual TL (first re-trained on histological images, then pCLE). Model performance was evaluated using accuracy and AUC across 50 independent runs with 10-fold cross-validation. Results: The dual TL approach statistically significant outperformed confocal TL, with AlexNet achieving a mean accuracy of 94.97% and an AUC of 0.98, surpassing GoogLeNet (91.43% accuracy, 0.97 AUC) and ResNet (89.87% accuracy, 0.96 AUC). All networks demonstrated statistically significant (p < 0.001) improvements in performance with dual TL. Additionally, dual TL models showed reductions in both false positives and false negatives, with class activation mappings highlighting enhanced focus on diagnostically relevant regions. Conclusions: Dual TL, integrating histological and pCLE imaging, results in a statistically significant improvement in lung cancer classification. This approach offers a promising framework for enhanced tissue classification. and with future development and testing, iy has the potential to improve patient outcomes. Full article
(This article belongs to the Special Issue Digital Health Technologies in Oncology)
Show Figures

Figure 1

32 pages, 3875 KB  
Article
Enhanced Multi-Model Deep Learning for Rapid and Precise Diagnosis of Pulmonary Diseases Using Chest X-Ray Imaging
by Rahul Kumar, Cheng-Tang Pan, Yi-Min Lin, Shiue Yow-Ling, Ting-Sheng Chung and Uyanahewa Gamage Shashini Janesha
Diagnostics 2025, 15(3), 248; https://doi.org/10.3390/diagnostics15030248 - 22 Jan 2025
Cited by 4 | Viewed by 3366
Abstract
Background: The global burden of respiratory diseases such as influenza, tuberculosis, and viral pneumonia necessitates rapid, accurate diagnostic tools to improve healthcare responses. Current methods, including RT-PCR and chest radiography, face limitations in accuracy, speed, accessibility, and cost-effectiveness, especially in resource-constrained settings, often [...] Read more.
Background: The global burden of respiratory diseases such as influenza, tuberculosis, and viral pneumonia necessitates rapid, accurate diagnostic tools to improve healthcare responses. Current methods, including RT-PCR and chest radiography, face limitations in accuracy, speed, accessibility, and cost-effectiveness, especially in resource-constrained settings, often delaying treatment and increasing transmission. Methods: This study introduces an Enhanced Multi-Model Deep Learning (EMDL) approach to address these challenges. EMDL integrates an ensemble of five pre-trained deep learning models (VGG-16, VGG-19, ResNet, AlexNet, and GoogleNet) with advanced image preprocessing (histogram equalization and contrast enhancement) and a novel multi-stage feature selection and optimization pipeline (PCA, SelectKBest, Binary Particle Swarm Optimization (BPSO), and Binary Grey Wolf Optimization (BGWO)). Results: Evaluated on two independent chest X-ray datasets, EMDL achieved high accuracy in the multiclass classification of influenza, pneumonia, and tuberculosis. The combined image enhancement and feature optimization strategies significantly improved diagnostic precision and model robustness. Conclusions: The EMDL framework provides a scalable and efficient solution for accurate and accessible pulmonary disease diagnosis, potentially improving treatment efficacy and patient outcomes, particularly in resource-limited settings. Full article
Show Figures

Figure 1

Back to TopTop