Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (871)

Search Parameters:
Keywords = Inception_v3

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
19 pages, 2136 KB  
Article
Transformer-Based Multi-Class Classification of Bangladeshi Rice Varieties Using Image Data
by Israt Tabassum and Vimala Nunavath
Appl. Sci. 2026, 16(3), 1279; https://doi.org/10.3390/app16031279 - 27 Jan 2026
Abstract
Rice (Oryza sativa L.) is a staple food for over half of the global population, with significant economic, agricultural, and cultural importance, particularly in Asia. Thousands of rice varieties exist worldwide, differing in size, shape, color, and texture, making accurate classification essential [...] Read more.
Rice (Oryza sativa L.) is a staple food for over half of the global population, with significant economic, agricultural, and cultural importance, particularly in Asia. Thousands of rice varieties exist worldwide, differing in size, shape, color, and texture, making accurate classification essential for quality control, breeding programs, and authenticity verification in trade and research. Traditional manual identification of rice varieties is time-consuming, error-prone, and heavily reliant on expert knowledge. Deep learning provides an efficient alternative by automatically extracting discriminative features from rice grain images for precise classification. While prior studies have primarily employed deep learning models such as CNN, VGG, InceptionV3, MobileNet, and DenseNet201, transformer-based models remain underexplored for rice variety classification. This study addresses this gap by applying two deep learning models such as Swin Transformer and Vision Transformer for multi-class classification of rice varieties using the publicly available PRBD dataset from Bangladesh. Experimental results demonstrate that the ViT model achieved an accuracy of 99.86% with precision, recall, and F1-score all at 0.9986, while the Swin Transformer model obtained an accuracy of 99.44% with a precision of 0.9944, recall of 0.9944, and F1-score of 0.9943. These results highlight the effectiveness of transformer-based models for high-accuracy rice variety classification. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

27 pages, 1579 KB  
Article
Quadra Sense: A Fusion of Deep Learning Classifiers for Mitosis Detection in Breast Cancer Histopathology
by Afnan M. Alhassan and Nouf I. Altmami
Diagnostics 2026, 16(3), 393; https://doi.org/10.3390/diagnostics16030393 - 26 Jan 2026
Abstract
Background/Objectives: The difficulties caused by breast cancer have been addressed in a number of ways. Since it is said to be the second most common cause of death from cancer among women, early intervention is crucial. Early detection is difficult because of [...] Read more.
Background/Objectives: The difficulties caused by breast cancer have been addressed in a number of ways. Since it is said to be the second most common cause of death from cancer among women, early intervention is crucial. Early detection is difficult because of the existing detection tools’ shortcomings in objectivity and accuracy. Quadra Sense, a fusion of deep learning (DL) classifiers for mitosis detection in breast cancer histopathology, is proposed to address the shortcomings of current approaches. It demonstrates a greater capacity to produce more accurate results. Methods: Initially, the raw dataset is preprocessed by using a normalization by means of color channel normalization (zero-mean normalization) and stain normalization (Macenko Stain Normalization), and the artifact can be removed via median filtering and contrast enhancement using histogram equalization; ROI identification is performed using modified Fully Convolutional Networks (FCNs) followed by the feature extraction (FE) with Modified InceptionV4 (M-IV4), by which the deep features are retrieved and the feature are selected by means of a Self-Improved Seagull Optimization Algorithm (SA-SOA), and finally, classification is performed using Mito-Quartet. Results: Ultimately, using a performance evaluation, the suggested approach achieved a higher accuracy of 99.2% in comparison with the current methods. Conclusions: From the outcomes, the recommended technique performs well. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

35 pages, 5337 KB  
Article
Enhancing Glioma Classification in Magnetic Resonance Imaging Using Vision Transformers and Convolutional Neural Networks
by Marco Antonio Gómez-Guzmán, José Jaime Esqueda-Elizondo, Laura Jiménez-Beristain, Gilberto Manuel Galindo-Aldana, Oscar Adrian Aguirre-Castro, Edgar Rene Ramos-Acosta, Cynthia Torres-Gonzalez, Enrique Efren García-Guerrero and Everardo Inzunza-Gonzalez
Electronics 2026, 15(2), 434; https://doi.org/10.3390/electronics15020434 - 19 Jan 2026
Viewed by 109
Abstract
Brain tumors, encompassing subtypes with distinct progression and risk profiles, are a serious public health concern. Magnetic resonance imaging (MRI) is the primary imaging modality for non-invasive assessment, providing the contrast and detail necessary for diagnosis, subtype classification, and individualized care planning. In [...] Read more.
Brain tumors, encompassing subtypes with distinct progression and risk profiles, are a serious public health concern. Magnetic resonance imaging (MRI) is the primary imaging modality for non-invasive assessment, providing the contrast and detail necessary for diagnosis, subtype classification, and individualized care planning. In this paper, we evaluate the capability of modern deep learning models to classify gliomas as high-grade (HGG) or low-grade (LGG) using reduced training data from MRI scans. Utilizing the BraTS 2019 best-slice dataset (2185 images in two classes, HGG and LGG) divided in two folders, training and testing, with different images obtained from different patients, we created subsets including 10%, 25%, 50%, 75%, and 100% of the dataset. Six deep learning architectures, DeiT3_base_patch16_224, Inception_v4, Xception41, ConvNextV2_tiny, swin_tiny_patch4_window7_224, and EfficientNet_B0, were evaluated utilizing three-fold cross-validation (k = 3) and increasingly large training datasets. Explainability was assessed using Grad-CAM. With 25% of the training data, DeiT3_base_patch16_224 achieved an accuracy of 99.401% and an F1-Score of 99.403%. Under the same conditions, Inception_v4 achieved an accuracy of 99.212% and a F1-Score of 99.222%. Considering how the models performed across both data subsets and their compute demands, Inception_v4 struck the best balance for MRI-based glioma classification. Both convolutional networks and vision transformers achieved superior discrimination between HGGs and LGGs, even under data-limited conditions. Architectural disparities became increasingly apparent as training data diminished, highlighting unique inductive biases and efficiency characteristics. Even with a relatively limited amount of training data, current deep learning (DL) methods can achieve reliable performance in classifying gliomas from MRI scans. Among the architectures evaluated, Inception_v4 offered the most consistent balance between accuracy, F1-Score, and computational cost, making it a strong candidate for integration into MRI-based clinical workflows. Full article
Show Figures

Figure 1

28 pages, 30101 KB  
Article
Machine Learning-Driven Soil Fungi Identification Using Automated Imaging Techniques
by Karol Struniawski, Ryszard Kozera, Aleksandra Konopka, Lidia Sas-Paszt and Agnieszka Marasek-Ciolakowska
Appl. Sci. 2026, 16(2), 855; https://doi.org/10.3390/app16020855 - 14 Jan 2026
Viewed by 125
Abstract
Soilborne fungi (Fusarium, Trichoderma, Verticillium, Purpureocillium) critically impact agricultural productivity, disease dynamics, and soil health, requiring rapid identification for precision agriculture. Current diagnostics require labor-intensive microscopy or expensive molecular assays (up to 10 days), while existing ML studies [...] Read more.
Soilborne fungi (Fusarium, Trichoderma, Verticillium, Purpureocillium) critically impact agricultural productivity, disease dynamics, and soil health, requiring rapid identification for precision agriculture. Current diagnostics require labor-intensive microscopy or expensive molecular assays (up to 10 days), while existing ML studies suffer from small datasets (<500 images), expert selection bias, and lack of public availability. A fully automated identification system integrating robotic microscopy (Keyence VHX-700) with deep learning was developed. The Soil Fungi Microscopic Images Dataset (SFMID) comprises 20,151 images (11,511 no-water, 8640 water-based)—the largest publicly available soil fungi dataset. Four CNN architectures (InceptionResNetV2, ResNet152V2, DenseNet121, DenseNet201) were evaluated with transfer learning and three-shot majority voting. Grad-CAM analysis validated biological relevance. ResNet152V2 conv2 achieved optimal SFMID-NW performance (precision: 0.6711; AUC: 0.8031), with real-time inference (20 ms, 48–49 images/second). Statistical validation (McNemar’s test: χ2=27.34,p<0.001) confirmed that three-shot classification significantly outperforms single-image prediction. Confusion analysis identified Fusarium–Trichoderma (no-water) and Fusarium–Verticillium (water-based) challenges, indicating morphological ambiguities. The publicly available SFMID provides a scalable foundation for AI-enhanced agricultural diagnostics. Full article
(This article belongs to the Special Issue Latest Research on Computer Vision and Image Processing)
Show Figures

Figure 1

28 pages, 13960 KB  
Article
Deep Learning Approaches for Brain Tumor Classification in MRI Scans: An Analysis of Model Interpretability
by Emanuela F. Gomes and Ramiro S. Barbosa
Appl. Sci. 2026, 16(2), 831; https://doi.org/10.3390/app16020831 - 14 Jan 2026
Viewed by 388
Abstract
This work presents the development and evaluation of Artificial Intelligence (AI) models for the automatic classification of brain tumors in Magnetic Resonance Imaging (MRI) scans. Several deep learning architectures were implemented and compared, including VGG-19, ResNet50, EfficientNetB3, Xception, MobileNetV2, DenseNet201, InceptionV3, Vision Transformer [...] Read more.
This work presents the development and evaluation of Artificial Intelligence (AI) models for the automatic classification of brain tumors in Magnetic Resonance Imaging (MRI) scans. Several deep learning architectures were implemented and compared, including VGG-19, ResNet50, EfficientNetB3, Xception, MobileNetV2, DenseNet201, InceptionV3, Vision Transformer (ViT), and an Ensemble model. The models were developed in Python (version 3.12.4) using the Keras and TensorFlow frameworks and trained on a public Brain Tumor MRI dataset containing 7023 images. Data augmentation and hyperparameter optimization techniques were applied to improve model generalization. The results showed high classification performance, with accuracies ranging from 89.47% to 98.17%. The Vision Transformer achieved the best performance, reaching 98.17% accuracy, outperforming traditional Convolutional Neural Network (CNN) architectures. Explainable AI (XAI) methods Grad-CAM, LIME, and Occlusion Sensitivity were employed to assess model interpretability, showing that the models predominantly focused on tumor regions. The proposed approach demonstrated the effectiveness of AI-based systems in supporting early diagnosis of brain tumors, reducing analysis time and assisting healthcare professionals. Full article
(This article belongs to the Special Issue Advanced Intelligent Technologies in Bioinformatics and Biomedicine)
Show Figures

Figure 1

31 pages, 3424 KB  
Article
Intrusion Detection in Smart Power Networks Using Inception-V4 Neural Networks Optimized by Modified Polar Fox Optimization Algorithm for Cyber-Physical Threat Mitigation
by Chao Tang, Linghao Zhang and Hongli Liu
Electronics 2026, 15(2), 360; https://doi.org/10.3390/electronics15020360 - 13 Jan 2026
Viewed by 225
Abstract
Threats that are caused by cyber-attacks on intelligent power networks promote the implementation of sophisticated intrusion detection devices, which can effectively detect advanced attacks. In this paper, a new model is introduced that combines the Modified Polar Fox Optimization Algorithm (MPFA) with an [...] Read more.
Threats that are caused by cyber-attacks on intelligent power networks promote the implementation of sophisticated intrusion detection devices, which can effectively detect advanced attacks. In this paper, a new model is introduced that combines the Modified Polar Fox Optimization Algorithm (MPFA) with an Inception-V4 deep neural network to enhance the effectiveness of the threat detection task. The MPFA optimizes inception-V4 hyperparameters and architecture to balance the exploration and exploitation processes of the courtship learning process and fitness-based scaling. The optimized model on the smart grid monitoring power is shown to perform well; it achieves over 99.5% accuracy, precision, recall, and F1-score on the detection of various attacks, including False Data Injection, Denial-of-Service, and Load Redistribution, and has a favorable computational overhead, thus it can be considered a formidable solution to protect critical smart grid infrastructure. The optimized model, evaluated on the Smart Grid Monitoring Power dataset, achieves state-of-the-art performance with an accuracy of 99.63%, a precision of 99.61%, a recall of 99.65%, and an F1-score of 99.63% for the detection of various cyber-physical attacks, including False Data Injection, Denial-of-Service, and Load Redistribution. It also maintains a favorable computational overhead, thus presenting a formidable solution for protecting critical smart grid infrastructure. Full article
Show Figures

Figure 1

14 pages, 1825 KB  
Article
CycleGAN-Based Translation of Digital Camera Images into Confocal-like Representations for Paper Fiber Imaging: Quantitative and Grad-CAM Analysis
by Naoki Kamiya, Kosuke Ashino, Yuto Hosokawa and Koji Shibazaki
Appl. Sci. 2026, 16(2), 814; https://doi.org/10.3390/app16020814 - 13 Jan 2026
Viewed by 193
Abstract
The structural analysis of paper fibers is vital for the noninvasive classification and conservation of traditional handmade paper in cultural heritage. Although digital still cameras (DSCs) offer a low-cost and noninvasive imaging solution, their inferior image quality compared to white-light confocal microscopy (WCM) [...] Read more.
The structural analysis of paper fibers is vital for the noninvasive classification and conservation of traditional handmade paper in cultural heritage. Although digital still cameras (DSCs) offer a low-cost and noninvasive imaging solution, their inferior image quality compared to white-light confocal microscopy (WCM) limits their effectiveness in fiber classification. To address this modality gap, we propose an unpaired image-to-image translation approach using cycle-consistent adversarial networks (CycleGANs). Our study targets a multifiber setting involving kozo, mitsumata, and gampi, using publicly available domain-specific datasets. Generated WCM-style images were quantitatively evaluated using peak signal-to-noise ratio, structural similarity index measure, mean absolute error, and Fréchet inception distance, achieving 8.24 dB, 0.28, 172.50, and 197.39, respectively. Classification performance was tested using EfficientNet-B0 and Inception-ResNet-v2, with F1-scores reaching 94.66% and 98.61%, respectively, approaching the performance of real WCM images (99.50% and 98.86%) and surpassing previous results obtained directly from DSC inputs (80.76% and 84.19%). Furthermore, Grad-CAM visualization confirmed that the translated images retained class-discriminative features aligned with those of the actual WCM inputs. Thus, the proposed CycleGAN-based image conversion effectively bridges the modality gap, enabling DSC images to approximate WCM characteristics and support high-accuracy paper fiber classification, which is a practical alternative for noninvasive material analysis. Full article
Show Figures

Graphical abstract

15 pages, 665 KB  
Article
Comparative Evaluation of Deep Learning Models for the Classification of Impacted Maxillary Canines on Panoramic Radiographs
by Nazlı Tokatlı, Buket Erdem, Mustafa Özcan, Begüm Turan Maviş, Çağla Şar and Fulya Özdemir
Diagnostics 2026, 16(2), 219; https://doi.org/10.3390/diagnostics16020219 - 9 Jan 2026
Viewed by 254
Abstract
Background/Objectives: The early and accurate identification of impacted teeth in the maxilla is critical for effective dental treatment planning. Traditional diagnostic methods relying on manual interpretation of radiographic images are often time-consuming and subject to variability. Methods: This study presents a deep learning-based [...] Read more.
Background/Objectives: The early and accurate identification of impacted teeth in the maxilla is critical for effective dental treatment planning. Traditional diagnostic methods relying on manual interpretation of radiographic images are often time-consuming and subject to variability. Methods: This study presents a deep learning-based approach for automated classification of impacted maxillary canines using panoramic radiographs. A comparative evaluation of four pre-trained convolutional neural network (CNN) architectures—ResNet50, Xception, InceptionV3, and VGG16—was conducted through transfer learning techniques. In this retrospective single-center study, the dataset comprised 694 annotated panoramic radiographs sourced from the archives of a university dental hospital, with a mildly imbalanced representation of impacted and non-impacted cases. Models were assessed using accuracy, precision, recall, specificity, and F1-score. Results: Among the tested architectures, VGG16 demonstrated superior performance, achieving an accuracy of 99.28% and an F1-score of 99.43%. Additionally, a prototype diagnostic interface was developed to demonstrate the potential for clinical application. Conclusions: The findings underscore the potential of deep learning models, particularly VGG16, in enhancing diagnostic workflows; however, further validation on diverse, multi-center datasets is required to confirm clinical generalizability. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

33 pages, 24811 KB  
Article
Demystifying Deep Learning Decisions in Leukemia Diagnostics Using Explainable AI
by Shahd H. Altalhi and Salha M. Alzahrani
Diagnostics 2026, 16(2), 212; https://doi.org/10.3390/diagnostics16020212 - 9 Jan 2026
Viewed by 328
Abstract
Background/Objectives: Conventional workflows, peripheral blood smears, and bone marrow assessment supplemented by LDI-PCR, molecular cytogenetics, and array-CGH, are expert-driven in the face of biological and imaging variability. Methods: We propose an AI pipeline that integrates convolutional neural networks (CNNs) and transfer [...] Read more.
Background/Objectives: Conventional workflows, peripheral blood smears, and bone marrow assessment supplemented by LDI-PCR, molecular cytogenetics, and array-CGH, are expert-driven in the face of biological and imaging variability. Methods: We propose an AI pipeline that integrates convolutional neural networks (CNNs) and transfer learning-based models with two explainable AI (XAI) approaches, LIME and Grad-Cam, to deliver both high diagnostic accuracy and transparent rationale. Seven public sources were curated into a unified benchmark (66,550 images) covering ALL, AML, CLL, CML, and healthy controls; images were standardized, ROI-cropped, and split with stratification (80/10/10). We fine-tuned multiple backbones (DenseNet-121, MobileNetV2, VGG16, InceptionV3, ResNet50, Xception, and a custom CNN) and evaluated the accuracy and F1-score, benchmarking against the recent literature. Results: On the five-class task (ALL/AML/CLL/CML/Healthy), MobileNetV2 achieved 97.9% accuracy/F1, with DenseNet-121 reaching 97.66% F1. On ALL subtypes (Benign, Early, Pre, Pro) and across tasks, DenseNet121 and MobileNetV2 were the most reliable, achieving state-of-the-art accuracy with the strongest, nucleus-centric explanations. Conclusions: XAI analyses (LIME, Grad-CAM) consistently localized leukemic nuclei and other cell-intrinsic morphology, aligning saliency with clinical cues and model performance. Compared with baselines, our approach matched or exceeded accuracy while providing stronger, corroborated interpretability on a substantially larger and more diverse dataset. Full article
Show Figures

Figure 1

23 pages, 30920 KB  
Article
A Surface Defect Detection System for Industrial Conveyor Belt Inspection Using Apple’s TrueDepth Camera Technology
by Mohammad Siami, Przemysław Dąbek, Hamid Shiri, Tomasz Barszcz and Radosław Zimroz
Appl. Sci. 2026, 16(2), 609; https://doi.org/10.3390/app16020609 - 7 Jan 2026
Viewed by 232
Abstract
Maintaining the structural integrity of conveyor belts is essential for safe and reliable mining operations. However, these belts are susceptible to longitudinal tearing and surface degradation from material impact, fatigue, and deformation. Many computer vision-based inspection methods are inefficient and unreliable in harsh [...] Read more.
Maintaining the structural integrity of conveyor belts is essential for safe and reliable mining operations. However, these belts are susceptible to longitudinal tearing and surface degradation from material impact, fatigue, and deformation. Many computer vision-based inspection methods are inefficient and unreliable in harsh mining environments characterized by dust and variable lighting. This study introduces a smartphone-driven defect detection system for the cost-effective, geometric inspection of conveyor belt surfaces. Using Apple’s iPhone 12 Pro Max (Apple Inc., Cupertino, CA, USA), the system captures 3D point cloud data from a moving belt with induced damage via the integrated TrueDepth camera. A key innovation is a 3D-to-2D projection pipeline that converts point cloud data into structured representations compatible with standard 2D Convolutional Neural Networks (CNNs). We then propose a hybrid deep learning and machine learning model, where features extracted by pre-trained CNNs (VGG16, ResNet50, InceptionV3, Xception) are classified by ensemble methods (Random Forest, XGBoost, LightGBM). The proposed system achieves high detection accuracy exceeding 0.97 F1 score in the case of all proposed model implementations with TrueDepth F1 score over 0.05 higher than RGB approach. Applied cost-effective smartphone-based sensing platform proved to support near-real-time maintenance decisions. Laboratory results demonstrate the method’s reliability, with measurement errors for defect dimensions within 3 mm. This approach shows significant potential to improve conveyor belt management, reduce maintenance costs, and enhance operational safety. Full article
(This article belongs to the Special Issue Mining Engineering: Present and Future Prospectives)
Show Figures

Figure 1

27 pages, 18339 KB  
Article
SBMEV: A Stacking-Based Meta-Ensemble Vehicle Classification Framework for Real-World Traffic Surveillance
by Preeti Pateriya, Ashutosh Trivedi and Ruchika Malhotra
Appl. Sci. 2026, 16(1), 520; https://doi.org/10.3390/app16010520 - 4 Jan 2026
Viewed by 220
Abstract
Developing vehicle classification remains a fundamental challenge for intelligent traffic management in the Indian urban environment, where traffic exhibits high heterogeneity, density and unpredictability. In the Indian subcontinent, vehicle movement is erratic, congestion is high, and vehicle types vary significantly. Conventional global benchmarks [...] Read more.
Developing vehicle classification remains a fundamental challenge for intelligent traffic management in the Indian urban environment, where traffic exhibits high heterogeneity, density and unpredictability. In the Indian subcontinent, vehicle movement is erratic, congestion is high, and vehicle types vary significantly. Conventional global benchmarks often fail to capture these complexities, highlighting the need for a region-specific dataset. To address this gap, the present study introduced the EAHVSD dataset, a novel real-world image collection comprising 10,864 vehicle images from four distinct classes, acquired from roadside surveillance cameras at multiple viewpoints and under varying conditions. This dataset is designed to support the development of an automatic traffic counter and classifier (ATCC) system. A comprehensive evaluation of eleven state-of-the-art deep learning models, namely VGG16, VGG19, MobileNetV2, Xception, AlexNet, ResNet50, ResNet152, DenseNet121, DenseNet201, InceptionV3, and NASNetMobile, was carried out. Among these, the highest accuracy result has been achieved by VGG-16, MobileNetV2, InceptionV3, DenseNet-121, and DenseNet-201. We developed a stacking-based meta-ensemble framework to leverage the complementary strengths of its components and overcome their individual limitations. In this approach, a meta-learner classifier integrates the predictions of the best-performing models, thereby improving robustness, scalability, and real-world adaptability. The proposed ensemble model achieved an overall classification accuracy of 96.04%, a Cohen’s Kappa of 0.93, and an AUC of 0.99, consistently outperforming the individual models and existing baselines. A comparative analysis with prior studies further validates the efficacy and reliability of the stacking-based meta-ensemble method. These findings position the proposed frameworks as a robust and scalable solution for efficient vehicle classification under practical surveillance constraints, with potential applications in intelligent transportation systems and traffic management. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

15 pages, 921 KB  
Article
Rethinking DeepVariant: Efficient Neural Architectures for Intelligent Variant Calling
by Anastasiia Gurianova, Anastasiia Pestruilova, Aleksandra Beliaeva, Artem Kasianov, Liudmila Mikhailova, Egor Guguchkin and Evgeny Karpulevich
Int. J. Mol. Sci. 2026, 27(1), 513; https://doi.org/10.3390/ijms27010513 - 4 Jan 2026
Viewed by 434
Abstract
DeepVariant has revolutionized the field of genetic variant identification by reframing variant detection as an image classification problem. However, despite its wide adoption in bioinformatics workflows, the tool continues to evolve mainly through the expansion of training datasets, while its core neural network [...] Read more.
DeepVariant has revolutionized the field of genetic variant identification by reframing variant detection as an image classification problem. However, despite its wide adoption in bioinformatics workflows, the tool continues to evolve mainly through the expansion of training datasets, while its core neural network architecture—Inception V3—has remained unchanged. In this study, we revisited the DeepVariant design and presented a prototype of a modernized version that supports alternative neural network backbones. As a proof of concept, we replaced the legacy Inception V3 model with a mid-sized EfficientNet model and evaluated its performance using the benchmark dataset from the Genome in a Bottle (GIAB) project. Alternative architecture demonstrated faster convergence, a twofold reduction in the number of parameters, and improved accuracy in variant identification. On the test dataset, updated workflow achieved consistent improvements of +0.1% in SNP F1-score, enabling the detection of up to several hundred additional true variants per genome. These results show that optimizing the neural architecture alone can enhance the accuracy, robustness, and efficiency of variant calling, thereby improving the overall quality of sequencing data analysis. Full article
Show Figures

Figure 1

17 pages, 2703 KB  
Article
A Method for Automatic Emotion Detection Through Machine Learning
by Jessica Lucarelli, Mario Cesarelli, Antonella Santone, Fabio Martinelli and Francesco Mercaldo
Appl. Sci. 2026, 16(1), 397; https://doi.org/10.3390/app16010397 - 30 Dec 2025
Viewed by 417
Abstract
Facial expression recognition (FER) is a fundamental component of Affective Computing and is gaining increasing relevance in mental health applications. This study presents an approach for facial expression recognition using feature extraction and machine learning techniques. Starting from a publicly available dataset, a [...] Read more.
Facial expression recognition (FER) is a fundamental component of Affective Computing and is gaining increasing relevance in mental health applications. This study presents an approach for facial expression recognition using feature extraction and machine learning techniques. Starting from a publicly available dataset, a manual cleaning and relabeling process led to the creation of a refined dataset of 35,625 facial images grouped into four emotional macroclasses. Features were extracted using the SqueezeNet and Inception v3 embedders and classified using various algorithms. The experimental results show that Inception v3 consistently outperforms SqueezeNet and that feature normalization improves classification stability and robustness. The results highlight the importance of data quality and preprocessing in applied FER systems. Full article
(This article belongs to the Section Electrical, Electronics and Communications Engineering)
Show Figures

Figure 1

25 pages, 6127 KB  
Article
Deep Learning-Based Prediction of Fish Freshness and Purchasability Using Multi-Angle Image Data
by Sakhi Mohammad Hamidy, Yusuf Kuvvetli, Yetkin Sakarya, Serya Tülin Özkütük and Yesim Özoğul
Foods 2026, 15(1), 68; https://doi.org/10.3390/foods15010068 - 25 Dec 2025
Viewed by 624
Abstract
This study aims to predict the freshness of sea bass (Dicentrarchus labrax) using deep learning models based on image data. For this purpose, 10 fish were monitored daily from the day of purchase until three days after spoilage, with multi-angle imaging [...] Read more.
This study aims to predict the freshness of sea bass (Dicentrarchus labrax) using deep learning models based on image data. For this purpose, 10 fish were monitored daily from the day of purchase until three days after spoilage, with multi-angle imaging (eight distinct perspectives per fish, both with and without background) and corresponding quality analyses. A total of 22 quality parameters—10 categorical (sensory-based) and 12 numerical (color-based)—were evaluated, with the purchasability parameter defined as the most critical indicator of freshness. Using seven popular transfer learning algorithms (EfficientNetB0, ResNet50, DenseNet121, VGG16, InceptionV3, MobileNet, and VGG19), 2464 predictive models (1120 classification and 1344 regression) were trained. Classification models were evaluated using accuracy, precision, recall, F1-score, and response time, while regression models were assessed using mean absolute error and tolerance-based error metrics. The results showed that the MobileNet algorithm achieved the best overall performance, successfully predicting 15 of the 22 parameters with the lowest error or highest accuracy. Importantly, in the prediction of the most critical parameter—purchasability—the DenseNet121 architecture yielded the best classification performance with an accuracy of 0.9894. The findings indicate that deep learning-based image analysis is a viable method for evaluating the freshness of fish. Full article
(This article belongs to the Section Food Quality and Safety)
Show Figures

Figure 1

18 pages, 428 KB  
Article
Enhancing Education Through Generative AI: A Multimodal Approach to Semantic Search and Authentic Learning
by Ahmad Raza, Amina Jameel and Freeha Azmat
Educ. Sci. 2026, 16(1), 22; https://doi.org/10.3390/educsci16010022 - 24 Dec 2025
Viewed by 267
Abstract
In contemporary education, learners face the challenge of navigating an overwhelming abundance of information. Traditional search methods, often limited to keyword matching, fail to capture the nuanced meaning and relationships within educational materials. Our multimodal approach combines Sentence Transformer for text and Inception [...] Read more.
In contemporary education, learners face the challenge of navigating an overwhelming abundance of information. Traditional search methods, often limited to keyword matching, fail to capture the nuanced meaning and relationships within educational materials. Our multimodal approach combines Sentence Transformer for text and Inception V3 for images to generate vector embeddings for textbooks which are stored in an Elasticsearch database. Learners’ queries again are converted to vector embeddings which are matched through cosine similarity with stored embeddings, resulting in retrieval of relevant material which is ranked and then synthesized using large language model (LLM) APIs. The approach retrieves answers based on semantic search rather than keywords. The system also integrates GenAI capabilities separately, specifically leveraging LLM APIs, to generate context-aware answers to user-posed questions at varying levels of complexity, e.g., beginner, intermediate, and advanced. Through comprehensive evaluation, we demonstrate the system’s ability to retrieve coherent answers across multiple sources, offering significant advancements in cross-text and cross-modal retrieval tasks. This work also contributes to the international discourse on ethical GenAI integration in curricula and fosters a collaborative human–AI learning ecosystem. Full article
(This article belongs to the Special Issue Generative-AI-Enhanced Learning Environments and Applications)
Show Figures

Figure 1

Back to TopTop