Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (738)

Search Parameters:
Keywords = inception net

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
18 pages, 7062 KiB  
Article
Multimodal Feature Inputs Enable Improved Automated Textile Identification
by Magken George Enow Gnoupa, Andy T. Augousti, Olga Duran, Olena Lanets and Solomiia Liaskovska
Textiles 2025, 5(3), 31; https://doi.org/10.3390/textiles5030031 - 2 Aug 2025
Viewed by 71
Abstract
This study presents an advanced framework for fabric texture classification by leveraging macro- and micro-texture extraction techniques integrated with deep learning architectures. Co-occurrence histograms, local binary patterns (LBPs), and albedo-dependent feature maps were employed to comprehensively capture the surface properties of fabrics. A [...] Read more.
This study presents an advanced framework for fabric texture classification by leveraging macro- and micro-texture extraction techniques integrated with deep learning architectures. Co-occurrence histograms, local binary patterns (LBPs), and albedo-dependent feature maps were employed to comprehensively capture the surface properties of fabrics. A late fusion approach was applied using four state-of-the-art convolutional neural networks (CNNs): InceptionV3, ResNet50_V2, DenseNet, and VGG-19. Excellent results were obtained, with the ResNet50_V2 achieving a precision of 0.929, recall of 0.914, and F1 score of 0.913. Notably, the integration of multimodal inputs allowed the models to effectively distinguish challenging fabric types, such as cotton–polyester and satin–silk pairs, which exhibit overlapping texture characteristics. This research not only enhances the accuracy of textile classification but also provides a robust methodology for material analysis, with significant implications for industrial applications in fashion, quality control, and robotics. Full article
26 pages, 1790 KiB  
Article
A Hybrid Deep Learning Model for Aromatic and Medicinal Plant Species Classification Using a Curated Leaf Image Dataset
by Shareena E. M., D. Abraham Chandy, Shemi P. M. and Alwin Poulose
AgriEngineering 2025, 7(8), 243; https://doi.org/10.3390/agriengineering7080243 - 1 Aug 2025
Viewed by 176
Abstract
In the era of smart agriculture, accurate identification of plant species is critical for effective crop management, biodiversity monitoring, and the sustainable use of medicinal resources. However, existing deep learning approaches often underperform when applied to fine-grained plant classification tasks due to the [...] Read more.
In the era of smart agriculture, accurate identification of plant species is critical for effective crop management, biodiversity monitoring, and the sustainable use of medicinal resources. However, existing deep learning approaches often underperform when applied to fine-grained plant classification tasks due to the lack of domain-specific, high-quality datasets and the limited representational capacity of traditional architectures. This study addresses these challenges by introducing a novel, well-curated leaf image dataset consisting of 39 classes of medicinal and aromatic plants collected from the Aromatic and Medicinal Plant Research Station in Odakkali, Kerala, India. To overcome performance bottlenecks observed with a baseline Convolutional Neural Network (CNN) that achieved only 44.94% accuracy, we progressively enhanced model performance through a series of architectural innovations. These included the use of a pre-trained VGG16 network, data augmentation techniques, and fine-tuning of deeper convolutional layers, followed by the integration of Squeeze-and-Excitation (SE) attention blocks. Ultimately, we propose a hybrid deep learning architecture that combines VGG16 with Batch Normalization, Gated Recurrent Units (GRUs), Transformer modules, and Dilated Convolutions. This final model achieved a peak validation accuracy of 95.24%, significantly outperforming several baseline models, such as custom CNN (44.94%), VGG-19 (59.49%), VGG-16 before augmentation (71.52%), Xception (85.44%), Inception v3 (87.97%), VGG-16 after data augumentation (89.24%), VGG-16 after fine-tuning (90.51%), MobileNetV2 (93.67), and VGG16 with SE block (94.94%). These results demonstrate superior capability in capturing both local textures and global morphological features. The proposed solution not only advances the state of the art in plant classification but also contributes a valuable dataset to the research community. Its real-world applicability spans field-based plant identification, biodiversity conservation, and precision agriculture, offering a scalable tool for automated plant recognition in complex ecological and agricultural environments. Full article
(This article belongs to the Special Issue Implementation of Artificial Intelligence in Agriculture)
Show Figures

Figure 1

26 pages, 4572 KiB  
Article
Transfer Learning-Based Ensemble of CNNs and Vision Transformers for Accurate Melanoma Diagnosis and Image Retrieval
by Murat Sarıateş and Erdal Özbay
Diagnostics 2025, 15(15), 1928; https://doi.org/10.3390/diagnostics15151928 - 31 Jul 2025
Viewed by 260
Abstract
Background/Objectives: Melanoma is an aggressive type of skin cancer that poses serious health risks if not detected in its early stages. Although early diagnosis enables effective treatment, delays can result in life-threatening consequences. Traditional diagnostic processes predominantly rely on the subjective expertise [...] Read more.
Background/Objectives: Melanoma is an aggressive type of skin cancer that poses serious health risks if not detected in its early stages. Although early diagnosis enables effective treatment, delays can result in life-threatening consequences. Traditional diagnostic processes predominantly rely on the subjective expertise of dermatologists, which can lead to variability and time inefficiencies. Consequently, there is an increasing demand for automated systems that can accurately classify melanoma lesions and retrieve visually similar cases to support clinical decision-making. Methods: This study proposes a transfer learning (TL)-based deep learning (DL) framework for the classification of melanoma images and the enhancement of content-based image retrieval (CBIR) systems. Pre-trained models including DenseNet121, InceptionV3, Vision Transformer (ViT), and Xception were employed to extract deep feature representations. These features were integrated using a weighted fusion strategy and classified through an Ensemble learning approach designed to capitalize on the complementary strengths of the individual models. The performance of the proposed system was evaluated using classification accuracy and mean Average Precision (mAP) metrics. Results: Experimental evaluations demonstrated that the proposed Ensemble model significantly outperformed each standalone model in both classification and retrieval tasks. The Ensemble approach achieved a classification accuracy of 95.25%. In the CBIR task, the system attained a mean Average Precision (mAP) score of 0.9538, indicating high retrieval effectiveness. The performance gains were attributed to the synergistic integration of features from diverse model architectures through the ensemble and fusion strategies. Conclusions: The findings underscore the effectiveness of TL-based DL models in automating melanoma image classification and enhancing CBIR systems. The integration of deep features from multiple pre-trained models using an Ensemble approach not only improved accuracy but also demonstrated robustness in feature generalization. This approach holds promise for integration into clinical workflows, offering improved diagnostic accuracy and efficiency in the early detection of melanoma. Full article
Show Figures

Figure 1

21 pages, 15647 KiB  
Article
Research on Oriented Object Detection in Aerial Images Based on Architecture Search with Decoupled Detection Heads
by Yuzhe Kang, Bohao Zheng and Wei Shen
Appl. Sci. 2025, 15(15), 8370; https://doi.org/10.3390/app15158370 - 28 Jul 2025
Viewed by 253
Abstract
Object detection in aerial images can provide great support in traffic planning, national defense reconnaissance, hydrographic surveys, infrastructure construction, and other fields. Objects in aerial images are characterized by small pixel–area ratios, dense arrangements between objects, and arbitrary inclination angles. In response to [...] Read more.
Object detection in aerial images can provide great support in traffic planning, national defense reconnaissance, hydrographic surveys, infrastructure construction, and other fields. Objects in aerial images are characterized by small pixel–area ratios, dense arrangements between objects, and arbitrary inclination angles. In response to these characteristics and problems, we improved the feature extraction network Inception-ResNet using the Fast Architecture Search (FAS) module and proposed a one-stage anchor-free rotation object detector. The structure of the object detector is simple and only consists of convolution layers, which reduces the number of model parameters. At the same time, the label sampling strategy in the training process is optimized to resolve the problem of insufficient sampling. Finally, a decoupled object detection head is used to separate the bounding box regression task from the object classification task. The experimental results show that the proposed method achieves mean average precision (mAP) of 82.6%, 79.5%, and 89.1% on the DOTA1.0, DOTA1.5, and HRSC2016 datasets, respectively, and the detection speed reaches 24.4 FPS, which can meet the needs of real-time detection. Full article
(This article belongs to the Special Issue Innovative Applications of Artificial Intelligence in Engineering)
Show Figures

Figure 1

27 pages, 4682 KiB  
Article
DERIENet: A Deep Ensemble Learning Approach for High-Performance Detection of Jute Leaf Diseases
by Mst. Tanbin Yasmin Tanny, Tangina Sultana, Md. Emran Biswas, Chanchol Kumar Modok, Arjina Akter, Mohammad Shorif Uddin and Md. Delowar Hossain
Information 2025, 16(8), 638; https://doi.org/10.3390/info16080638 - 27 Jul 2025
Viewed by 197
Abstract
Jute, a vital lignocellulosic fiber crop with substantial industrial and ecological relevance, continues to suffer considerable yield and quality degradation due to pervasive foliar pathologies. Traditional diagnostic modalities reliant on manual field inspections are inherently constrained by subjectivity, diagnostic latency, and inadequate scalability [...] Read more.
Jute, a vital lignocellulosic fiber crop with substantial industrial and ecological relevance, continues to suffer considerable yield and quality degradation due to pervasive foliar pathologies. Traditional diagnostic modalities reliant on manual field inspections are inherently constrained by subjectivity, diagnostic latency, and inadequate scalability across geographically distributed agrarian systems. To transcend these limitations, we propose DERIENet, a robust and scalable classification approach within a deep ensemble learning framework. It is meticulously engineered by integrating three high-performing convolutional neural networks—ResNet50, InceptionV3, and EfficientNetB0—along with regularization, batch normalization, and dropout strategies, to accurately classify jute leaf diseases such as Cercospora Leaf Spot, Golden Mosaic Virus, and healthy leaves. A key methodological contribution is the design of a novel augmentation pipeline, termed Geometric Localized Occlusion and Adaptive Rescaling (GLOAR), which dynamically modulates photometric and geometric distortions based on image entropy and luminance to synthetically upscale a limited dataset (920 images) into a significantly enriched and diverse dataset of 7800 samples, thereby mitigating overfitting and enhancing domain generalizability. Empirical evaluation, utilizing a comprehensive set of performance metrics—accuracy, precision, recall, F1-score, confusion matrices, and ROC curves—demonstrates that DERIENet achieves a state-of-the-art classification accuracy of 99.89%, with macro-averaged and weighted average precision, recall, and F1-score uniformly at 99.89%, and an AUC of 1.0 across all disease categories. The reliability of the model is validated by the confusion matrix, which shows that 899 out of 900 test images were correctly identified and that there was only one misclassification. Comparative evaluations of the various ensemble baselines, such as DenseNet201, MobileNetV2, and VGG16, and individual base learners demonstrate that DERIENet performs noticeably superior to all baseline models. It provides a highly interpretable, deployment-ready, and computationally efficient architecture that is ideal for integrating into edge or mobile platforms to facilitate in situ, real-time disease diagnostics in precision agriculture. Full article
Show Figures

Figure 1

13 pages, 2660 KiB  
Review
Pituitary Apoplexy in a Non-Functioning PitNET After Cabergoline Use: Case Report and Review of the Literature
by Federica De Luca, Margherita Paccagnella, Anna Pizzo, Giulia Zuolo, Veronica Calabrò and Stella Bernardi
J. Clin. Med. 2025, 14(14), 5089; https://doi.org/10.3390/jcm14145089 - 17 Jul 2025
Viewed by 262
Abstract
Background/Objectives: Pituitary apoplexy (PA) is a rare medical emergency characterized by the sudden onset of symptoms resulting from hemorrhage and/or infarction within the pituitary gland. Precipitating factors include the use of dopamine agonists (DAs), whose main indication is the treatment of prolactin [...] Read more.
Background/Objectives: Pituitary apoplexy (PA) is a rare medical emergency characterized by the sudden onset of symptoms resulting from hemorrhage and/or infarction within the pituitary gland. Precipitating factors include the use of dopamine agonists (DAs), whose main indication is the treatment of prolactin (PRL)-secreting pituitary neuroendocrine tumors (PitNETs), but which can also be considered in non-functioning PitNETs. Here we report a case of PA in a patient taking cabergoline for a non-functioning PitNET, followed by a review of the literature focusing on the cases of PA associated with the use of DAs. Methods: A review of the literature was performed, searching Pubmed for other clinical cases of PA associated with the use of DAs, from inception to March 2025. Results: We found 43 cases of PA associated with the use of DAs. All the patients had secreting tumors: 86% were classified as PRL-secreting PitNETs, 7% were classified as GH-secreting PitNETs, and 4.6% included a mixed PRL/GH-secreting PitNET and a TSH-secreting PitNET. By contrast, here we present a case of PA in a non-functioning PitNET during cabergoline therapy. Our patient was managed conservatively and endocrine function recovered spontaneously. In our case, cabergoline might have promoted PA, which is consistent with the reported efficacy of cabergoline in inducing tumor shrinkage of non-functioning PitNETs that express dopamine 2 receptors, including silent PIT1 and SF1 or NULL tumors. Conclusions: Our case confirms cabergoline efficacy in non-functioning PitNETs and sheds light on a possible complication of its use. Patients, particularly those with large tumors, should be closely monitored for this occurrence. Full article
Show Figures

Figure 1

29 pages, 9069 KiB  
Article
Prediction of Temperature Distribution with Deep Learning Approaches for SM1 Flame Configuration
by Gökhan Deveci, Özgün Yücel and Ali Bahadır Olcay
Energies 2025, 18(14), 3783; https://doi.org/10.3390/en18143783 - 17 Jul 2025
Viewed by 308
Abstract
This study investigates the application of deep learning (DL) techniques for predicting temperature fields in the SM1 swirl-stabilized turbulent non-premixed flame. Two distinct DL approaches were developed using a comprehensive CFD database generated via the steady laminar flamelet model coupled with the SST [...] Read more.
This study investigates the application of deep learning (DL) techniques for predicting temperature fields in the SM1 swirl-stabilized turbulent non-premixed flame. Two distinct DL approaches were developed using a comprehensive CFD database generated via the steady laminar flamelet model coupled with the SST k-ω turbulence model. The first approach employs a fully connected dense neural network to directly map scalar input parameters—fuel velocity, swirl ratio, and equivalence ratio—to high-resolution temperature contour images. In addition, a comparison was made with different deep learning networks, namely Res-Net, EfficientNetB0, and Inception Net V3, to better understand the performance of the model. In the first approach, the results of the Inception V3 model and the developed Dense Model were found to be better than Res-Net and Efficient Net. At the same time, file sizes and usability were examined. The second framework employs a U-Net-based convolutional neural network enhanced by an RGB Fusion preprocessing technique, which integrates multiple scalar fields from non-reacting (cold flow) conditions into composite images, significantly improving spatial feature extraction. The training and validation processes for both models were conducted using 80% of the CFD data for training and 20% for testing, which helped assess their ability to generalize new input conditions. In the secondary approach, similar to the first approach, studies were conducted with different deep learning models, namely Res-Net, Efficient Net, and Inception Net, to evaluate model performance. The U-Net model, which is well developed, stands out with its low error and small file size. The dense network is appropriate for direct parametric analyses, while the image-based U-Net model provides a rapid and scalable option to utilize the cold flow CFD images. This framework can be further refined in future research to estimate more flow factors and tested against experimental measurements for enhanced applicability. Full article
Show Figures

Figure 1

33 pages, 4016 KiB  
Article
Integrated Deep Learning Framework for Cardiac Risk Stratification and Complication Analysis in Leigh’s Disease
by Md Aminul Islam, Jayasree Varadarajan, Md Abu Sufian, Bhupesh Kumar Mishra and Md Ruhul Amin Rasel
Cardiogenetics 2025, 15(3), 19; https://doi.org/10.3390/cardiogenetics15030019 - 15 Jul 2025
Viewed by 274
Abstract
Background: Leigh’s Disease is a rare mitochondrial disorder primarily affecting the central nervous system, with frequent secondary cardiac manifestations such as hypertrophic and dilated cardiomyopathies. Early detection of cardiac complications is crucial for patient management, but manual interpretation of cardiac MRI is labour-intensive [...] Read more.
Background: Leigh’s Disease is a rare mitochondrial disorder primarily affecting the central nervous system, with frequent secondary cardiac manifestations such as hypertrophic and dilated cardiomyopathies. Early detection of cardiac complications is crucial for patient management, but manual interpretation of cardiac MRI is labour-intensive and subject to inter-observer variability. Methodology: We propose an integrated deep learning framework using cardiac MRI to automate the detection of cardiac abnormalities associated with Leigh’s Disease. Four CNN architectures—Inceptionv3, a custom 3-layer CNN, DenseNet169, and EfficientNetB2—were trained on preprocessed MRI data (224 × 224 pixels), including left ventricular segmentation, contrast enhancement, and gamma correction. Morphological features (area, aspect ratio, and extent) were also extracted to aid interpretability. Results: EfficientNetB2 achieved the highest test accuracy (99.2%) and generalization performance, followed by DenseNet169 (98.4%), 3-layer CNN (95.6%), and InceptionV3 (94.2%). Statistical morphological analysis revealed significant differences in cardiac structure between Leigh’s and non-Leigh’s cases, particularly in area (212,097 vs. 2247 pixels) and extent (0.995 vs. 0.183). The framework was validated using ROC (AUC = 1.00), Brier Score (0.000), and cross-validation (mean sensitivity = 1.000, std = 0.000). Feature embedding visualisation using PCA, t-SNE, and UMAP confirmed class separability. Grad-CAM heatmaps localised relevant myocardial regions, supporting model interpretability. Conclusions: Our deep learning-based framework demonstrated high diagnostic accuracy and interpretability in detecting Leigh’s disease-related cardiac complications. Integrating morphological analysis and explainable AI provides a robust and scalable tool for early-stage detection and clinical decision support in rare diseases. Full article
Show Figures

Figure 1

21 pages, 4147 KiB  
Article
AgriFusionNet: A Lightweight Deep Learning Model for Multisource Plant Disease Diagnosis
by Saleh Albahli
Agriculture 2025, 15(14), 1523; https://doi.org/10.3390/agriculture15141523 - 15 Jul 2025
Viewed by 477
Abstract
Timely and accurate identification of plant diseases is critical to mitigating crop losses and enhancing yield in precision agriculture. This paper proposes AgriFusionNet, a lightweight and efficient deep learning model designed to diagnose plant diseases using multimodal data sources. The framework integrates RGB [...] Read more.
Timely and accurate identification of plant diseases is critical to mitigating crop losses and enhancing yield in precision agriculture. This paper proposes AgriFusionNet, a lightweight and efficient deep learning model designed to diagnose plant diseases using multimodal data sources. The framework integrates RGB and multispectral drone imagery with IoT-based environmental sensor data (e.g., temperature, humidity, soil moisture), recorded over six months across multiple agricultural zones. Built on the EfficientNetV2-B4 backbone, AgriFusionNet incorporates Fused-MBConv blocks and Swish activation to improve gradient flow, capture fine-grained disease patterns, and reduce inference latency. The model was evaluated using a comprehensive dataset composed of real-world and benchmarked samples, showing superior performance with 94.3% classification accuracy, 28.5 ms inference time, and a 30% reduction in model parameters compared to state-of-the-art models such as Vision Transformers and InceptionV4. Extensive comparisons with both traditional machine learning and advanced deep learning methods underscore its robustness, generalization, and suitability for deployment on edge devices. Ablation studies and confusion matrix analyses further confirm its diagnostic precision, even in visually ambiguous cases. The proposed framework offers a scalable, practical solution for real-time crop health monitoring, contributing toward smart and sustainable agricultural ecosystems. Full article
(This article belongs to the Special Issue Computational, AI and IT Solutions Helping Agriculture)
Show Figures

Figure 1

19 pages, 3165 KiB  
Article
Majority Voting Ensemble of Deep CNNs for Robust MRI-Based Brain Tumor Classification
by Kuo-Ying Liu, Nan-Han Lu, Yung-Hui Huang, Akari Matsushima, Koharu Kimura, Takahide Okamoto and Tai-Been Chen
Diagnostics 2025, 15(14), 1782; https://doi.org/10.3390/diagnostics15141782 - 15 Jul 2025
Viewed by 430
Abstract
Background/Objectives: Accurate classification of brain tumors is critical for treatment planning and prognosis. While deep convolutional neural networks (CNNs) have shown promise in medical imaging, few studies have systematically compared multiple architectures or integrated ensemble strategies to improve diagnostic performance. This study [...] Read more.
Background/Objectives: Accurate classification of brain tumors is critical for treatment planning and prognosis. While deep convolutional neural networks (CNNs) have shown promise in medical imaging, few studies have systematically compared multiple architectures or integrated ensemble strategies to improve diagnostic performance. This study aimed to evaluate various CNN models and optimize classification performance using a majority voting ensemble approach on T1-weighted MRI brain images. Methods: Seven pretrained CNN architectures were fine-tuned to classify four categories: glioblastoma, meningioma, pituitary adenoma, and no tumor. Each model was trained using two optimizers (SGDM and ADAM) and evaluated on a public dataset split into training (70%), validation (10%), and testing (20%) subsets, and further validated on an independent external dataset to assess generalizability. A majority voting ensemble was constructed by aggregating predictions from all 14 trained models. Performance was assessed using accuracy, Kappa coefficient, true positive rate, precision, confusion matrix, and ROC curves. Results: Among individual models, GoogLeNet and Inception-v3 with ADAM achieved the highest classification accuracy (0.987). However, the ensemble approach outperformed all standalone models, achieving an accuracy of 0.998, a Kappa coefficient of 0.997, and AUC values above 0.997 for all tumor classes. The ensemble demonstrated improved sensitivity, precision, and overall robustness. Conclusions: The majority voting ensemble of diverse CNN architectures significantly enhanced the performance of MRI-based brain tumor classification, surpassing that of any single model. These findings underscore the value of model diversity and ensemble learning in building reliable AI-driven diagnostic tools for neuro-oncology. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

24 pages, 9593 KiB  
Article
Deep Learning Approaches for Skin Lesion Detection
by Jonathan Vieira, Fábio Mendonça and Fernando Morgado-Dias
Electronics 2025, 14(14), 2785; https://doi.org/10.3390/electronics14142785 - 10 Jul 2025
Viewed by 350
Abstract
Recently, there has been a rise in skin cancer cases, for which early detection is highly relevant, as it increases the likelihood of a cure. In this context, this work presents a benchmarking study of standard Convolutional Neural Network (CNN) architectures for automated [...] Read more.
Recently, there has been a rise in skin cancer cases, for which early detection is highly relevant, as it increases the likelihood of a cure. In this context, this work presents a benchmarking study of standard Convolutional Neural Network (CNN) architectures for automated skin lesion classification. A total of 38 CNN architectures from ten families (ConvNeXt, DenseNet, EfficientNet, Inception, InceptionResNet, MobileNet, NASNet, ResNet, VGG, and Xception) were evaluated using transfer learning on the HAM10000 dataset for seven-class skin lesion classification, namely, actinic keratoses, basal cell carcinoma, benign keratosis-like lesions, dermatofibroma, melanoma, melanocytic nevi, and vascular lesions. The comparative analysis used standardized training conditions, with all models utilizing frozen pre-trained weights. Cross-database validation was then conducted using the ISIC 2019 dataset to assess generalizability across different data distributions. The ConvNeXtXLarge architecture achieved the best performance, despite having one of the lowest performance-to-number-of-parameters ratios, with 87.62% overall accuracy and 76.15% F1 score on the test set, demonstrating competitive results within the established performance range of existing HAM10000-based studies. A proof-of-concept multiplatform mobile application was also implemented using a client–server architecture with encrypted image transmission, demonstrating the viability of integrating high-performing models into healthcare screening tools. Full article
Show Figures

Figure 1

30 pages, 5474 KiB  
Article
WHU-RS19 ABZSL: An Attribute-Based Dataset for Remote Sensing Image Understanding
by Mattia Balestra, Marina Paolanti and Roberto Pierdicca
Remote Sens. 2025, 17(14), 2384; https://doi.org/10.3390/rs17142384 - 10 Jul 2025
Viewed by 318
Abstract
The advancement of artificial intelligence (AI) in remote sensing (RS) increasingly depends on datasets that offer rich and structured supervision beyond traditional scene-level labels. Although existing benchmarks for aerial scene classification have facilitated progress in this area, their reliance on single-class annotations restricts [...] Read more.
The advancement of artificial intelligence (AI) in remote sensing (RS) increasingly depends on datasets that offer rich and structured supervision beyond traditional scene-level labels. Although existing benchmarks for aerial scene classification have facilitated progress in this area, their reliance on single-class annotations restricts their application to more flexible, interpretable and generalisable learning frameworks. In this study, we introduce WHU-RS19 ABZSL: an attribute-based extension of the widely adopted WHU-RS19 dataset. This new version comprises 1005 high-resolution aerial images across 19 scene categories, each annotated with a vector of 38 features. These cover objects (e.g., roads and trees), geometric patterns (e.g., lines and curves) and dominant colours (e.g., green and blue), and are defined through expert-guided annotation protocols. To demonstrate the value of the dataset, we conduct baseline experiments using deep learning models that had been adapted for multi-label classification—ResNet18, VGG16, InceptionV3, EfficientNet and ViT-B/16—designed to capture the semantic complexity characteristic of real-world aerial scenes. The results, which are measured in terms of macro F1-score, range from 0.7385 for ResNet18 to 0.7608 for EfficientNet-B0. In particular, EfficientNet-B0 and ViT-B/16 are the top performers in terms of the overall macro F1-score and consistency across attributes, while all models show a consistent decline in performance for infrequent or visually ambiguous categories. This confirms that it is feasible to accurately predict semantic attributes in complex scenes. By enriching a standard benchmark with detailed, image-level semantic supervision, WHU-RS19 ABZSL supports a variety of downstream applications, including multi-label classification, explainable AI, semantic retrieval, and attribute-based ZSL. It thus provides a reusable, compact resource for advancing the semantic understanding of remote sensing and multimodal AI. Full article
(This article belongs to the Special Issue Remote Sensing Datasets and 3D Visualization of Geospatial Big Data)
Show Figures

Figure 1

23 pages, 3645 KiB  
Article
Color-Guided Mixture-of-Experts Conditional GAN for Realistic Biomedical Image Synthesis in Data-Scarce Diagnostics
by Patrycja Kwiek, Filip Ciepiela and Małgorzata Jakubowska
Electronics 2025, 14(14), 2773; https://doi.org/10.3390/electronics14142773 - 10 Jul 2025
Viewed by 261
Abstract
Background: Limited availability of high-quality labeled biomedical image datasets presents a significant challenge for training deep learning models in medical diagnostics. This study proposes a novel image generation framework combining conditional generative adversarial networks (cGANs) with a Mixture-of-Experts (MoE) architecture and color histogram-aware [...] Read more.
Background: Limited availability of high-quality labeled biomedical image datasets presents a significant challenge for training deep learning models in medical diagnostics. This study proposes a novel image generation framework combining conditional generative adversarial networks (cGANs) with a Mixture-of-Experts (MoE) architecture and color histogram-aware loss functions to enhance synthetic blood cell image quality. Methods: RGB microscopic images from the BloodMNIST dataset (eight blood cell types, resolution 3 × 128 × 128) underwent preprocessing with k-means clustering to extract the dominant colors and UMAP for visualizing class similarity. Spearman correlation-based distance matrices were used to evaluate the discriminative power of each RGB channel. A MoE–cGAN architecture was developed with residual blocks and LeakyReLU activations. Expert generators were conditioned on cell type, and the generator’s loss was augmented with a Wasserstein distance-based term comparing red and green channel histograms, which were found most relevant for class separation. Results: The red and green channels contributed most to class discrimination; the blue channel had minimal impact. The proposed model achieved 0.97 classification accuracy on generated images (ResNet50), with 0.96 precision, 0.97 recall, and a 0.96 F1-score. The best Fréchet Inception Distance (FID) was 52.1. Misclassifications occurred mainly among visually similar cell types. Conclusions: Integrating histogram alignment into the MoE–cGAN training significantly improves the realism and class-specific variability of synthetic images, supporting robust model development under data scarcity in hematological imaging. Full article
Show Figures

Figure 1

16 pages, 1347 KiB  
Article
Detection of Helicobacter pylori Infection in Histopathological Gastric Biopsies Using Deep Learning Models
by Rafael Parra-Medina, Carlos Zambrano-Betancourt, Sergio Peña-Rojas, Lina Quintero-Ortiz, Maria Victoria Caro, Ivan Romero, Javier Hernan Gil-Gómez, John Jaime Sprockel, Sandra Cancino and Andres Mosquera-Zamudio
J. Imaging 2025, 11(7), 226; https://doi.org/10.3390/jimaging11070226 - 7 Jul 2025
Viewed by 758
Abstract
Traditionally, Helicobacter pylori (HP) gastritis has been diagnosed by pathologists through the examination of gastric biopsies using optical microscopy with standard hematoxylin and eosin (H&E) staining. However, with the adoption of digital pathology, the identification of HP faces certain limitations, particularly due to [...] Read more.
Traditionally, Helicobacter pylori (HP) gastritis has been diagnosed by pathologists through the examination of gastric biopsies using optical microscopy with standard hematoxylin and eosin (H&E) staining. However, with the adoption of digital pathology, the identification of HP faces certain limitations, particularly due to insufficient resolution in some scanned images. Moreover, interobserver variability has been well documented in the traditional diagnostic approach, which may further complicate consistent interpretation. In this context, deep convolutional neural network (DCNN) models are showing promising results in the automated detection of this infection in whole-slide images (WSIs). The aim of the present article is to detect the presence of HP infection from our own institutional dataset of histopathological gastric biopsy samples using different pretrained and recognized DCNN and AutoML approaches. The dataset comprises 100 H&E-stained WSIs of gastric biopsies. HP infection was confirmed previously using immunohistochemical confirmation. A total of 45,795 patches were selected for model development. InceptionV3, Resnet50, and VGG16 achieved AUC (area under the curve) values of 1. However, InceptionV3 showed superior metrics such as accuracy (97%), recall (100%), F1 score (97%), and MCC (93%). BoostedNet and AutoKeras achieved accuracy, precision, recall, specificity, and F1 scores less than 85%. The InceptionV3 model was used for external validation, and the predictions across all patches yielded a global accuracy of 78%. In conclusion, DCNN models showed stronger potential for diagnosing HP in gastric biopsies compared with the auto ML approach. However, due to variability across pathology applications, no single model is universally optimal. A problem-specific approach is essential. With growing WSI adoption, DL can improve diagnostic accuracy, reduce variability, and streamline pathology workflows using automation. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

20 pages, 2968 KiB  
Article
Real-Time Lightweight Morphological Detection for Chinese Mitten Crab Origin Tracing
by Xiaofei Ma, Nannan Shen, Yanhui He, Zhuo Fang, Hongyan Zhang, Yun Wang and Jinrong Duan
Appl. Sci. 2025, 15(13), 7468; https://doi.org/10.3390/app15137468 - 3 Jul 2025
Viewed by 258
Abstract
During the cultivation and circulation of Chinese mitten crab (Eriocheir sinensis), the difficulty in tracing geographic origin leads to quality uncertainty and market disorder. To address this challenge, this study proposes a two-stage origin traceability framework that integrates a lightweight object detector and [...] Read more.
During the cultivation and circulation of Chinese mitten crab (Eriocheir sinensis), the difficulty in tracing geographic origin leads to quality uncertainty and market disorder. To address this challenge, this study proposes a two-stage origin traceability framework that integrates a lightweight object detector and a high-precision classifier. In the first stage, an improved YOLOv10n-based model is designed by incorporating omni-dimensional dynamic convolution, a SlimNeck structure, and a Lightweight Shared Convolutional Detection head, which effectively enhances the detection accuracy of crab targets under complex multi-scale environments while reducing computational cost. In the second stage, an Improved GoogleNet’s Inception Net for Crab is developed based on the Inception module, with further integration of Asymmetric Convolution Blocks and Squeeze and Excitation modules to improve the feature extraction and classification ability for regional origin. A comprehensive crab dataset is constructed, containing images from diverse farming sites, including variations in species, color, size, angle, and background conditions. Experimental results show that the proposed detector achieves an mAP50 of 99.5% and an mAP50-95 of 88.5%, while maintaining 309 FPS and reducing GFLOPs by 35.3%. Meanwhile, the classification model achieves high accuracy with only 17.4% and 40% of the parameters of VGG16 and AlexNet, respectively. In conclusion, the proposed method achieves an optimal accuracy-speed-complexity trade-off, enabling robust real-time traceability for aquaculture systems. Full article
(This article belongs to the Section Computing and Artificial Intelligence)
Show Figures

Figure 1

Back to TopTop