3rd Edition: AI/ML-Based Medical Image Processing and Analysis

A special issue of Diagnostics (ISSN 2075-4418). This special issue belongs to the section "Machine Learning and Artificial Intelligence in Diagnostics".

Deadline for manuscript submissions: 30 June 2026 | Viewed by 20336

Special Issue Editors


E-Mail Website
Guest Editor
Department of Computing Science, Thompson Rivers University, Kamloops, BC V2C 0C8, Canada
Interests: artificial intelligence; large language models; medical imaging; robotics; internet of things; machine learning; embedded systems; microcontrollers
Special Issues, Collections and Topics in MDPI journals

E-Mail Website
Guest Editor
Software Engineering and Information Technology Management, University of Minnesota Crookston, Crookston, MN 56716, USA
Interests: machine learning; artificial intelligence; image processing; Internet of things (IoT)
Special Issues, Collections and Topics in MDPI journals

Special Issue Information

Dear Colleagues,

The use of artificial intelligence (AI) and machine learning (ML) in medical image processing and analysis is becoming increasingly important due to advances in image processing and analysis technology for the automated recommendation of medical diagnoses. Medical professionals and institutions could benefit from machine learning (ML)- and artificial intelligence (AI)-enabled medical devices, as they could ease the workload of professional medical personnel, increase the accuracy of diagnoses, and enable early diagnosis and intervention. The U.S. Food and Drug Administration has already approved many AI/ML-enabled medical devices, which are listed at https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-aiml-enabled-medical-devices. Researchers are encouraged to continue conducting studies in this field and establishing patentable methods and devices for use in medical institutions.

The International Conference on Advancement in Healthcare Technology and Biomedical Engineering (AHTBE 25) will be held in Vancouver, BC, Canada, on August 28–30, 2025. The conference aims to connect leading experts to discuss the latest innovations, challenges, and future directions in healthcare technology and biomedical engineering. The conference will provide a platform for the dissemination of cutting-edge research, fostering collaboration among researchers, practitioners, and policymakers from around the globe.

More information can be found on the conference website: https://ahtbe.ca/.

Dr. Ghazanfar Latif
Dr. Jaafar Alghazo
Guest Editors

Manuscript Submission Information

Manuscripts should be submitted online at www.mdpi.com by registering and logging in to this website. Once you are registered, click here to go to the submission form. Manuscripts can be submitted until the deadline. All submissions that pass pre-check are peer-reviewed. Accepted papers will be published continuously in the journal (as soon as accepted) and will be listed together on the special issue website. Research articles, review articles as well as short communications are invited. For planned papers, a title and short abstract (about 250 words) can be sent to the Editorial Office for assessment.

Submitted manuscripts should not have been published previously, nor be under consideration for publication elsewhere (except conference proceedings papers). All manuscripts are thoroughly refereed through a single-blind peer-review process. A guide for authors and other relevant information for submission of manuscripts is available on the Instructions for Authors page. Diagnostics is an international peer-reviewed open access semimonthly journal published by MDPI.

Please visit the Instructions for Authors page before submitting a manuscript. The Article Processing Charge (APC) for publication in this open access journal is 2600 CHF (Swiss Francs). Submitted papers should be well formatted and use good English. Authors may use MDPI's English editing service prior to publication or during author revisions.

Keywords

  • machine learning
  • artificial intelligence
  • medical imaging
  • medical diagnosis
  • medical
  • magnetic resonance imaging (MRI)
  • CT scan
  • X-ray
  • computer tomography
  • imaging techniques
  • medical conditions
  • convolutional neural networks
  • deep learning
  • transfer learning

Benefits of Publishing in a Special Issue

  • Ease of navigation: Grouping papers by topic helps scholars navigate broad scope journals more efficiently.
  • Greater discoverability: Special Issues support the reach and impact of scientific research. Articles in Special Issues are more discoverable and cited more frequently.
  • Expansion of research network: Special Issues facilitate connections among authors, fostering scientific collaborations.
  • External promotion: Articles in Special Issues are often promoted through the journal's social media, increasing their visibility.
  • Reprint: MDPI Books provides the opportunity to republish successful Special Issues in book format, both online and in print.

Further information on MDPI's Special Issue policies can be found here.

Published Papers (22 papers)

Order results
Result details
Select all
Export citation of selected articles as:

Research

Jump to: Review, Other

32 pages, 5551 KB  
Article
BanglaOCT2025: A Population-Specific Fovea-Centric OCT Dataset with Self-Supervised Volumetric Restoration Using Flip-Flop Swin Transformers
by Chinmay Bepery, G. M. Atiqur Rahaman, Rameswar Debnath, Sajib Saha, Md. Shafiqul Islam, Md. Emranul Islam Abir and Sanjay Kumar Sarker
Diagnostics 2026, 16(3), 420; https://doi.org/10.3390/diagnostics16030420 - 1 Feb 2026
Viewed by 134
Abstract
Background: Age-related macular degeneration (AMD) is a major cause of vision loss, yet publicly available Optical Coherence Tomography (OCT) datasets lack demographic diversity, particularly from South Asian populations. Existing datasets largely represent Western cohorts, limiting AI generalizability. Moreover, raw OCT volumes contain redundant [...] Read more.
Background: Age-related macular degeneration (AMD) is a major cause of vision loss, yet publicly available Optical Coherence Tomography (OCT) datasets lack demographic diversity, particularly from South Asian populations. Existing datasets largely represent Western cohorts, limiting AI generalizability. Moreover, raw OCT volumes contain redundant spatial information and speckle noise, hindering efficient analysis. Methods: We introduce BanglaOCT2025, a retrospective dataset collected from the National Institute of Ophthalmology and Hospital (NIOH), Bangladesh, using Nidek RS-330 Duo 2 and RS-3000 Advance systems. We propose a novel preprocessing pipeline comprising two stages: (1) A constraint-based centroid minimization algorithm automatically localizes the foveal center and extracts a fixed 33-slice macular sub-volume, robust to retinal tilt and acquisition variability; and (2) A self-supervised volumetric denoising module based on a Flip-Flop Swin Transformer (FFSwin) backbone suppresses speckle noise without requiring paired clean reference data. Results: The dataset comprises 1585 OCT volumes (202,880 B-scans), including 857 expert-annotated cases (54 DryAMD, 61 WetAMD, and 742 NonAMD). Denoising quality was evaluated using reference-free volumetric metrics, paired statistical analysis, and blinded clinical review by a retinal specialist, confirming preservation of pathological biomarkers and absence of hallucination. Under a controlled paired evaluation using the same classifier with frozen weights, downstream AMD classification accuracy improved from 69.08% to 99.88%, interpreted as an upper-bound estimate of diagnostic signal recoverability rather than independent generalization. Conclusions: BanglaOCT2025 is the first clinically validated OCT dataset representing the Bengali population and establishes a reproducible fovea-centric volumetric preprocessing and restoration framework for AMD analysis, with future validation across independent and multi-centre test cohorts. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Graphical abstract

15 pages, 2430 KB  
Article
Improved Detection of Small (<2 cm) Hepatocellular Carcinoma via Deep Learning-Based Synthetic CT Hepatic Arteriography: A Multi-Center External Validation Study
by Jung Won Kwak, Sung Bum Cho, Ki Choon Sim, Jeong Woo Kim, In Young Choi and Yongwon Cho
Diagnostics 2026, 16(2), 343; https://doi.org/10.3390/diagnostics16020343 - 21 Jan 2026
Viewed by 170
Abstract
Background/Objectives: Early detection of hepatocellular carcinoma (HCC), particularly small lesions (<2 cm), which is crucial for curative treatment, remains challenging with conventional liver dynamic computed tomography (LDCT). We aimed to develop a deep learning algorithm to generate synthetic CT during hepatic arteriography (CTHA) [...] Read more.
Background/Objectives: Early detection of hepatocellular carcinoma (HCC), particularly small lesions (<2 cm), which is crucial for curative treatment, remains challenging with conventional liver dynamic computed tomography (LDCT). We aimed to develop a deep learning algorithm to generate synthetic CT during hepatic arteriography (CTHA) from non-invasive LDCT and evaluate its lesion detection performance. Methods: A cycle-consistent generative adversarial network with an attention module [Unsupervised Generative Attentional Networks with Adaptive Layer-Instance Normalization (U-GAT-IT)] was trained using paired LDCT and CTHA images from 277 patients. The model was validated using internal (68 patients, 139 lesions) and external sets from two independent centers (87 patients, 117 lesions). Two radiologists assessed detection performance using a 5-point scale and the detection rate. Results: Synthetic CTHA significantly improved the detection of sub-centimeter (<1 cm) HCCs compared with LDCT in the internal set (69.6% vs. 47.8%, p < 0.05). This improvement was robust in the external set; synthetic CTHA detected a greater number of small lesions than LDCT. Quantitative metrics (structural similarity index measure and peak signal-to-noise ratio) indicated high structural fidelity. Conclusions: Deep-learning–based synthetic CTHA significantly enhanced the detection of small HCCs compared with standard LDCT, offering a non-invasive alternative with high detection sensitivity, which was validated across multicentric data. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

21 pages, 2615 KB  
Article
Evaluating the Impact of Demographic Factors on Subject-Independent EEG-Based Emotion Recognition Approaches
by Nathan Douglas, Maximilien Oosterhuis and Camilo E. Valderrama
Diagnostics 2026, 16(1), 144; https://doi.org/10.3390/diagnostics16010144 - 1 Jan 2026
Viewed by 429
Abstract
Background: Emotion recognition using electroencephalography (EEG) offers a non-invasive means of measuring brain responses to affective stimuli. However, since EEG signals can vary significantly between subjects, developing a deep learning model capable of accurately predicting emotions is challenging. Methods: To address [...] Read more.
Background: Emotion recognition using electroencephalography (EEG) offers a non-invasive means of measuring brain responses to affective stimuli. However, since EEG signals can vary significantly between subjects, developing a deep learning model capable of accurately predicting emotions is challenging. Methods: To address that challenge, this study proposes a deep learning approach that fuses EEG features with demographic information, specifically age, sex, and nationality, using an attention-based mechanism that learns to weigh each modality during classification. The method was evaluated using three benchmark datasets: SEED, SEED-FRA, and SEED-GER, which include EEG recordings of 31 subjects of different demographic backgrounds. Results: We compared a baseline model trained solely on the EEG-derived features against an extended model that fused the subjects’ EEG and demographic information. Including demographic information improved the performance, achieving 80.2%, 80.5%, and 88.8% for negative, neutral, and positive classes. The attention weights also revealed different contributions of EEG and demographic inputs, suggesting that the model learns to adapt based on subjects’ demographic information. Conclusions: These findings support integrating demographic data to enhance the performance and fairness of subject-independent EEG-based emotion recognition models. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

16 pages, 2601 KB  
Article
Diagnostic Accuracy of an Offline CNN Framework Utilizing Multi-View Chest X-Rays for Screening 14 Co-Occurring Communicable and Non-Communicable Diseases
by Latika Giri, Pradeep Raj Regmi, Ghanshyam Gurung, Grusha Gurung, Shova Aryal, Sagar Mandal, Samyam Giri, Sahadev Chaulagain, Sandip Acharya and Muhammad Umair
Diagnostics 2026, 16(1), 66; https://doi.org/10.3390/diagnostics16010066 - 24 Dec 2025
Viewed by 632
Abstract
Background: Chest radiography is the most widely used diagnostic imaging modality globally, yet its interpretation is hindered by a critical shortage of radiologists, especially in low- and middle-income countries (LMICs). The interpretation is both time-consuming and error-prone in high-volume settings. Artificial Intelligence (AI) [...] Read more.
Background: Chest radiography is the most widely used diagnostic imaging modality globally, yet its interpretation is hindered by a critical shortage of radiologists, especially in low- and middle-income countries (LMICs). The interpretation is both time-consuming and error-prone in high-volume settings. Artificial Intelligence (AI) systems trained on public data may lack generalizability to multi-view, real-world, local images. Deep learning tools have the potential to augment radiologists by providing real-time decision support by overcoming these. Objective: We evaluated the diagnostic accuracy of a deep learning-based convolutional neural network (CNN) trained on multi-view, hybrid (public and local datasets) for detecting thoracic abnormalities in chest radiographs of adults presenting to a tertiary hospital, operating in offline mode. Methodology: A CNN was pretrained on public datasets (Vin Big, NIH) and fine-tuned on a local dataset from a Nepalese tertiary hospital, comprising frontal (PA/AP) and lateral views from emergency, ICU, and outpatient settings. The dataset was annotated by three radiologists for 14 pathologies. Data augmentation simulated poor-quality images and artifacts. Performance was evaluated on a held-out test set (N = 522) against radiologists’ consensus, measuring AUC, sensitivity, specificity, mean average precision (mAP), and reporting time. Deployment feasibility was tested via PACS integration and standalone offline mode. Results: The CNN achieved an overall AUC of 0.86 across 14 abnormalities, with 68% sensitivity, 99% specificity, and 0.93 mAP. Colored bounding boxes improved clarity when multiple pathologies co-occurred (e.g., cardiomegaly with effusion). The system performed effectively on PA, AP, and lateral views, including poor-quality ER/ICU images. Deployment testing confirmed seamless PACS integration and offline functionality. Conclusions: The CNN trained on adult CXRs performed reliably in detecting key thoracic findings across varied clinical settings. Its robustness to image quality, integration of multiple views and visualization capabilities suggest it could serve as a useful aid for triage and diagnosis. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

16 pages, 3554 KB  
Article
Early Detection of Cystoid Macular Edema in Retinitis Pigmentosa Using Longitudinal Deep Learning Analysis of OCT Scans
by Farhang Hosseini, Farkhondeh Asadi, Reza Rabiei, Arash Roshanpoor, Hamideh Sabbaghi, Mehrnoosh Eslami and Rayan Ebnali Harari
Diagnostics 2026, 16(1), 46; https://doi.org/10.3390/diagnostics16010046 - 23 Dec 2025
Viewed by 443
Abstract
Background/Objectives: Retinitis pigmentosa (RP) is a progressive hereditary retinal disorder that frequently leads to vision loss, with cystoid macular edema (CME) occurring in approximately 10–50% of affected patients. Early detection of CME is crucial for timely intervention, yet most existing studies lack [...] Read more.
Background/Objectives: Retinitis pigmentosa (RP) is a progressive hereditary retinal disorder that frequently leads to vision loss, with cystoid macular edema (CME) occurring in approximately 10–50% of affected patients. Early detection of CME is crucial for timely intervention, yet most existing studies lack longitudinal data capable of capturing subtle disease progression. Methods: We propose a deep learning–based framework utilizing longitudinal optical coherence tomography (OCT) imaging for early detection of CME in patients with RP. A total of 2280 longitudinal OCT images were preprocessed using denoising and data augmentation techniques. Multiple pre-trained deep learning architectures were evaluated using a patient-wise data split to ensure robust performance assessment. Results: Among the evaluated models, ResNet-34 achieved the best performance, with an accuracy of 98.68%, specificity of 99.45%, and an F1-score of 98.36%. Conclusions: These results demonstrate the potential of longitudinal OCT–based artificial intelligence as a reliable, non-invasive screening tool for early CME detection in RP. To the best of our knowledge, this study is the first to leverage longitudinal OCT data for AI-driven CME prediction in this patient population. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

23 pages, 4282 KB  
Article
Scaphoid Fracture Detection and Localization Using Denoising Diffusion Models
by Zhih-Cheng Huang, Tai-Hua Yang, Zhen-Li Yang and Ming-Huwi Horng
Diagnostics 2026, 16(1), 26; https://doi.org/10.3390/diagnostics16010026 - 21 Dec 2025
Viewed by 341
Abstract
Background/Objectives: Scaphoid fractures are a common wrist injury, typically diagnosed and treated through X-ray imaging, a process that is often time-consuming. Among the various types of scaphoid fractures, occult and nondisplaced fractures pose significant diagnostic challenges due to their subtle appearance and variable [...] Read more.
Background/Objectives: Scaphoid fractures are a common wrist injury, typically diagnosed and treated through X-ray imaging, a process that is often time-consuming. Among the various types of scaphoid fractures, occult and nondisplaced fractures pose significant diagnostic challenges due to their subtle appearance and variable bone density, complicating accurate identification via X-ray images. Therefore, creating a reliable assist diagnostic system based on deep learning for the scaphoid fracture detection and localization is critical. Methods: This study proposes a scaphoid fracture detection and localization framework based on diffusion models. In Stage I, we augment the training data set by embedding fracture anomalies. Pseudofracture regions are generated on healthy scaphoid images, producing healthy and fractured data sets, forming a self-supervised learning strategy that avoids the complex and time-consuming manual annotation of medical images. In Stage II, a diffusion-based reconstruction model learns the features of healthy scaphoid images to perform high-quality reconstruction of pseudofractured scaphoid images, generating healthy scaphoid images. In Stage III, a U-Net-like network identifies differences between the target and reconstructed images, using these differences to determine whether the target scaphoid image contains a fracture. Results: After model training, we evaluated its diagnostic performance on real scaphoid images by comparing the model’s results with precise fracture locations further annotated by physicians. The proposed method achieved an image area under the receiver operating characteristic curve (AUROC) of 0.993 for scaphoid fracture detection, corresponding to an accuracy of 0.983, recall of 1.00, and precision of 0.975. For fracture localization, the model achieved a pixel AUROC of 0.978 and a pixel region overlap of 0.921. Conclusions: This study shows promise as a reliable, powerful, and scalable solution for the scaphoid fracture detection and localization. Experimental results demonstrate the strong performance of the denoising diffusion models; these models can significantly reduce diagnostic time while precisely localizing potential fracture regions, identifying conditions overlooked by the human eye. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

16 pages, 1560 KB  
Article
Performance Comparison of U-Net and Its Variants for Carotid Intima–Media Segmentation in Ultrasound Images
by Seungju Jeong, Minjeong Park, Sumin Jeong and Dong Chan Park
Diagnostics 2026, 16(1), 2; https://doi.org/10.3390/diagnostics16010002 - 19 Dec 2025
Viewed by 501
Abstract
Background/Objectives: This study systematically compared the performance of U-Net and variants for automatic analysis of carotid intima-media thickness (CIMT) in ultrasound images, focusing on segmentation accuracy and real-time efficiency. Methods: Ten models were trained and evaluated using a publicly available Carotid [...] Read more.
Background/Objectives: This study systematically compared the performance of U-Net and variants for automatic analysis of carotid intima-media thickness (CIMT) in ultrasound images, focusing on segmentation accuracy and real-time efficiency. Methods: Ten models were trained and evaluated using a publicly available Carotid Ultrasound Boundary Study (CUBS) dataset (2176 images from 1088 subjects). Images were preprocessed using histogram-based smoothing and resized to a resolution of 256 × 256 pixels. Model training was conducted using identical hyperparameters (50 epochs, batch size 8, Adam optimizer with a learning rate of 1 × 10−4, and binary cross-entropy loss). Segmentation accuracy was assessed using Dice, Intersection over Union (IoU), Precision, Recall, and Accuracy metrics, while real-time performance was evaluated based on training/inference times and the model parameter counts. Results: All models achieved high accuracy, with Dice/IoU scores above 0.80/0.67. Attention U-Net achieved the highest segmentation accuracy, while UNeXt demonstrated the fastest training/inference speeds (approximately 420,000 parameters). Qualitatively, UNet++ produced smooth and natural boundaries, highlighting its strength in boundary reconstruction. Additionally, the relationship between the model parameter count and Dice performance was visualized to illustrate the tradeoff between accuracy and efficiency. Conclusions: This study provides a quantitative/qualitative evaluation of the accuracy, efficiency, and boundary reconstruction characteristics of U-Net-based models for CIMT segmentation, offering guidance for model selection according to clinical requirements (accuracy vs. real-time performance). Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

13 pages, 457 KB  
Article
LivSCP: Improving Liver Fibrosis Classification Through Supervised Contrastive Pretraining
by Yogita Dubey, Aditya Bhongade and Punit Fulzele
Diagnostics 2025, 15(24), 3226; https://doi.org/10.3390/diagnostics15243226 - 17 Dec 2025
Viewed by 552
Abstract
Background: Deep learning models have been used in the past for non-invasive liver fibrosis classification based on liver ultrasound scans. After numerous improvements in the network architectures, optimizers, and development of hybrid methods, the performance of these models has barely improved. This [...] Read more.
Background: Deep learning models have been used in the past for non-invasive liver fibrosis classification based on liver ultrasound scans. After numerous improvements in the network architectures, optimizers, and development of hybrid methods, the performance of these models has barely improved. This creates a need for a sophisticated method that helps improve this slow-improving performance. Methods: We propose LivSCP, a method to train liver fibrosis classification models for better accuracy than the traditional supervised learning (SL). Our method needs no changes in the network architecture, optimizer, etc. Results: The proposed method achieves state-of-the-art performance, with an accuracy, precision, recall, and F1-score of 98.10% each, and an AUROC of 0.9972. A major advantage of LivSCP is that it does not require any modification to the network architecture. Our method is particularly well-suited for scenarios with limited labeled data and computational resources. Conclusions: In this work, we successfully propose a training method for liver fibrosis classification models in low-data and computation settings. By comparing the proposed method with our baseline (Vision Transformer with SL) and multiple models, we demonstrate the state-of-the-art performance of our method. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

15 pages, 3077 KB  
Article
Integrating Numerical Data with AI-Based Image Processing Techniques to Improve the Diagnostic Accuracy of Detecting Dental Caries in Panoramic Radiographs
by Bengü Başarı, Nuran Ulusoy and Kamil Dimililer
Diagnostics 2025, 15(24), 3167; https://doi.org/10.3390/diagnostics15243167 - 12 Dec 2025
Viewed by 567
Abstract
Background/Objectives: Dental caries is among the most common oral health problems, resulting from demineralization of dental hard tissues in acidic environments. Early diagnosis is essential to prevent severe tissue destruction, systemic complications and costly treatments. Conventional visual interpretation of panoramic radiographs, though [...] Read more.
Background/Objectives: Dental caries is among the most common oral health problems, resulting from demineralization of dental hard tissues in acidic environments. Early diagnosis is essential to prevent severe tissue destruction, systemic complications and costly treatments. Conventional visual interpretation of panoramic radiographs, though widely used, remains subjective and variable. This study evaluated the effectiveness of image processing techniques and artificial intelligence (AI)-assisted models for automated detection and classification of dental caries on panoramic radiographs, emphasizing numerical image data analysis. Methods: From 1084 panoramic radiographs, 405 were selected and classified into interproximal, occlusal and secondary caries groups. Each was segmented and one representative region was analyzed using the image data representation method. Numerical descriptors—brightness, contrast, entropy and histogram parameters—were extracted and evaluated with several machine learning algorithms. Results: Among tested models, the Decision Tree algorithm achieved the highest classification accuracy (0.988 at the 0.2 train-test ratio), showing superior and consistent results across caries types. Random Forest also demonstrated strong performance with limited training data, while Gaussian Naïve Bayes, KNN and RBFNN were less effective. Conclusions: The integration of numerical image features with AI-based models demonstrated high diagnostic accuracy and clinical interpretability, particularly with Decision Tree algorithm. These results highlight the potential of AI-assisted analysis of panoramic radiographs to enhance diagnostic reliability, reduce subjectivity and support more effective treatment planning. Further multicentre studies with larger and more diverse datasets are recommended to validate generalizability. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

23 pages, 2303 KB  
Article
Explainable Deep Learning for Breast Lesion Classification in Digital and Contrast-Enhanced Mammography
by Samara Acosta-Jiménez, Miguel M. Mendoza-Mendoza, Carlos E. Galván-Tejada, José M. Celaya-Padilla, Jorge I. Galván-Tejada and Manuel A. Soto-Murillo
Diagnostics 2025, 15(24), 3143; https://doi.org/10.3390/diagnostics15243143 - 10 Dec 2025
Viewed by 537
Abstract
Background: Artificial intelligence (AI) emerges as a powerful tool to assist breast cancer screening; however, its integration into different mammographic modalities remains insufficiently explored. Digital Mammography (DM) is widely accessible but presents limitations in dense breast tissue, whereas Contrast-Enhanced Spectral Mammography (CESM) [...] Read more.
Background: Artificial intelligence (AI) emerges as a powerful tool to assist breast cancer screening; however, its integration into different mammographic modalities remains insufficiently explored. Digital Mammography (DM) is widely accessible but presents limitations in dense breast tissue, whereas Contrast-Enhanced Spectral Mammography (CESM) provides functional information that enhances lesion visualization. Understanding how deep learning models behave across these modalities, and determining whether their decision-making patterns remain consistent, is essential for equitable clinical adoption. Methods: This study evaluates three convolutional neural network (CNN) architectures, ResNet-18, DenseNet-121, and EfficientNet-B0, for binary classification of breast lesions using DM and CESM images from the public CDD-CESM dataset (2006 images, three diagnostic classes). The models are trained separately on DM and CESM using three classification tasks: Normal vs. Benign, Benign vs. Malignant, and Normal vs. Malignant. A 3-fold cross-validation scheme and an independent test set are employed. Training uses transfer learning with ImageNet weights, weighted binary cross-entropy (BCE) loss, and SHapley Additive exPlanations (SHAP) analysis to visualize pixel-level relevance of model decisions. Results: CESM yields higher performance in the Normal vs. Benign and Benign vs. Malignant tasks, whereas DM achieves the highest discriminative ability in the Normal vs. Malignant comparison (EfficientNet-B0: AUC = 97%, Accuracy = 93.15%), surpassing the corresponding CESM results (AUC = 93%, Accuracy = 85.66%). SHAP attribution maps reveal anatomically coherent decision patterns in both modalities, with CESM producing sharper and more localized relevance regions due to contrast uptake, while DM exhibits broader yet spatially aligned attention. Across architectures, EfficientNet-B0 demonstrates the most stable performance and interpretability. Conclusions: CESM enhances subtle lesion discrimination through functional contrast, whereas DM, despite its simpler acquisition and wider availability, provides highly accurate and explainable outcomes when combined with modern CNNs. The consistent SHAP-based relevance observed across modalities indicates that both preserve clinically meaningful information. To the best of our knowledge, this study is the first to directly compare DM and CESM under identical preprocessing, training, and evaluation conditions using explainable deep learning models. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

16 pages, 2076 KB  
Article
Mortality Prediction from Patient’s First Day PAAC Radiograph in Internal Medicine Intensive Care Unit Using Artificial Intelligence Methods
by Orhan Gok, Türker Fedai Cavus, Ahmed Cihad Genc, Selcuk Yaylaci and Lacin Tatli Ayhan
Diagnostics 2025, 15(24), 3138; https://doi.org/10.3390/diagnostics15243138 - 10 Dec 2025
Viewed by 394
Abstract
Introduction: This study aims to predict mortality using chest radiographs obtained on the first day of intensive care admission, thereby contributing to better planning of doctors’ treatment strategies and more efficient use of limited resources through early and accurate predictions. Methods: We retrospectively [...] Read more.
Introduction: This study aims to predict mortality using chest radiographs obtained on the first day of intensive care admission, thereby contributing to better planning of doctors’ treatment strategies and more efficient use of limited resources through early and accurate predictions. Methods: We retrospectively analyzed 510 ICU patients. After data augmentation, a total of 3019 chest radiographs were used for model training and validation, while an independent, non-augmented test set of 100 patients (100 images) was reserved for final evaluation. Seventy-four (74) radiomic features were extracted from the images and analyzed using machine learning algorithms. Model performances were evaluated using the area under the ROC curve (AUC), sensitivity, and specificity metrics. Results: A total of 3019 data samples were included in the study. Through feature selection methods, the initial 74 features were gradually reduced to 10. The Subspace KNN algorithm demonstrated the highest prediction accuracy, achieving AUC 0.88, sensitivity 0.80, and specificity 0.87. Conclusions: Machine learning algorithms such as Subspace KNN and features obtained from PAAC radiographs, such as GLCM Contrast, Kurtosis, Cobb angle, Haralick, Bilateral Infiltrates, Cardiomegaly, Skewness, Unilateral Effusion, Median Intensity, and Intensity Range, are promising tools for mortality prediction in patients hospitalized in the internal medicine intensive care unit. These tools can be integrated into clinical decision support systems to provide benefits in patient management. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

15 pages, 1875 KB  
Article
MS-Detector: A Hierarchical Deep Learning Method to Detect Muscle Strain Using Bilateral Symmetric Ultrasound Images of the Body
by Le Zhu, Yifu Xiong, Huachao Wu, Li Zhu, Zihan Tang, Wenbin Pei, Jing Zhou and Zhidong Xue
Diagnostics 2025, 15(23), 3087; https://doi.org/10.3390/diagnostics15233087 - 4 Dec 2025
Viewed by 486
Abstract
Background/Objectives: Muscle strain impairs mobility and quality of life, yet ultrasound diagnosis remains dependent on subjective expert interpretation, which can lead to variability in lesion detection. This study aimed to develop and evaluate MS-detector, a symmetry-aware, two-stage deep learning model that leverages bilateral [...] Read more.
Background/Objectives: Muscle strain impairs mobility and quality of life, yet ultrasound diagnosis remains dependent on subjective expert interpretation, which can lead to variability in lesion detection. This study aimed to develop and evaluate MS-detector, a symmetry-aware, two-stage deep learning model that leverages bilateral B-mode ultrasound images to automatically detect muscle strain and provide clinicians with a consistent second-reader decision-support tool in routine practice. Methods: A YOLOv5-based detector proposes candidate regions, and a Siamese convolutional neural network (CNN) compares contralateral regions to filter false positives. The dataset comprised 559 bilateral pairs from 86 patients with consensus labels. All splits were enforced at the patient level. A fixed, independent hold-out test set of 32 pairs was never used for training, tuning, or threshold selection. Five-fold cross-validation (CV) on the remaining development set was used for model selection. The operating point was pre-specified at T1 = 0.01 and T2 = 0.20. Results: The detector achieved mAP = 0.4006 (five-fold CV mean). On the hold-out set at the pre-specified operating point, MS-detector attained recall = 0.826 and precision = 0.486, improving F1/F2 over the YOLOv5 baseline by increasing precision with an acceptable recall trade-off. A representative figure illustrates the reduction in low-confidence false positives after filtering; this example is illustrative rather than aggregate. Conclusions: Leveraging contralateral symmetry in a hierarchical scheme improves detection precision while maintaining clinically acceptable recall, supporting MS-detector as a decision-support tool. Future work will evaluate generalizability across scanners and centers and assess calibrated probabilistic fusion and lesion grading. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

13 pages, 16473 KB  
Article
Logarithmic Scaling of Loss Functions for Enhanced Self-Supervised Accelerated MRI Reconstruction
by Jaejin Cho
Diagnostics 2025, 15(23), 2993; https://doi.org/10.3390/diagnostics15232993 - 25 Nov 2025
Viewed by 395
Abstract
Background/Objectives: Magnetic resonance imaging (MRI) is a widely used non-invasive imaging modality that provides high-fidelity soft-tissue contrast without ionizing radiation. However, acquiring high-resolution MRI scans is time-consuming, necessitating accelerated acquisition and reconstruction methods. Recently, self-supervised learning approaches have been introduced for reconstructing undersampled [...] Read more.
Background/Objectives: Magnetic resonance imaging (MRI) is a widely used non-invasive imaging modality that provides high-fidelity soft-tissue contrast without ionizing radiation. However, acquiring high-resolution MRI scans is time-consuming, necessitating accelerated acquisition and reconstruction methods. Recently, self-supervised learning approaches have been introduced for reconstructing undersampled MRI data without external fully sampled ground truth. Methods: In this work, we propose a logarithmic scaled scheme for conventional loss functions (e.g., 1, 2) to enhance self-supervised MRI reconstruction. Standard self-supervised methods typically compute loss in the k-space domain, which tends to overemphasize low spatial frequencies while under-representing high-frequency information. Our method introduces a logarithmic scaling to adaptively rescale residuals, emphasizing high-frequency contributions and improving perceptual quality. Results: Experiments on public datasets demonstrate consistent quantitative improvements when the proposed log-scaled loss is applied within a self-supervised MRI reconstruction framework. Conclusions: The proposed approach improves reconstruction fidelity and perceptual quality while remaining lightweight, architecture-agnostic, and readily integrable into existing self-supervised MRI reconstruction pipelines. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

23 pages, 8309 KB  
Article
Hybrid Faster R-CNN for Tooth Numbering in Periapical Radiographs Based on Fédération Dentaire Internationale System
by Yong-Shao Su, I Elizabeth Cha, Yi-Cheng Mao, Li-Hsin Chang, Zi-Chun Kao, Shun-Yuan Tien, Yuan-Jin Lin, Shih-Lun Chen, Kuo-Chen Li and Patricia Angela R. Abu
Diagnostics 2025, 15(22), 2900; https://doi.org/10.3390/diagnostics15222900 - 15 Nov 2025
Viewed by 871
Abstract
Background/Objectives: Tooth numbering is essential because it allows dental clinicians to identify lesion locations during diagnosis, typically using the Fédération Dentaire Internationale system. However, accurate tooth numbering is challenging due to variations in periapical radiograph (PA) angles. In this study, we aimed to [...] Read more.
Background/Objectives: Tooth numbering is essential because it allows dental clinicians to identify lesion locations during diagnosis, typically using the Fédération Dentaire Internationale system. However, accurate tooth numbering is challenging due to variations in periapical radiograph (PA) angles. In this study, we aimed to develop a deep learning-based tool to assist dentists in accurately identifying teeth via tooth numbering and improve diagnostic efficiency and accuracy. Methods: We developed a Hybrid Faster Region-based Convolutional Neural Network (R-CNN) technique and a custom loss function tailored for PA tooth numbering to accelerate training. Additionally, we developed a tooth-numbering position auxiliary localization algorithm to address challenges associated with missing teeth and extensive crown loss in existing datasets. Results: We achieved a maximum precision of 95.16% utilizing the transformer-based NextViT-Faster R-CNN hybrid model, along with an accuracy increase of at least 8.5% and a 19.8% reduction in training time compared to models without the proposed tooth-numbering position auxiliary localization algorithm and conventional methods. Conclusions: The results demonstrate the effectiveness of the proposed method in overcoming challenges in PA tooth numbering within AI-assisted dental diagnostics, enhancing clinical efficiency, and reducing the risk of misdiagnosis in dental practices. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

20 pages, 6268 KB  
Article
Automated Implant Placement Pathway from Dental Panoramic Radiographs Using Deep Learning for Preliminary Clinical Assistance
by Pei-Yi Wu, Shih-Lun Chen, Yi-Cheng Mao, Yuan-Jin Lin, Pin-Yu Lu, Kai-Hsun Yu, Kuo-Chen Li, Tsun-Kuang Chi, Tsung-Yi Chen and Patricia Angela R. Abu
Diagnostics 2025, 15(20), 2598; https://doi.org/10.3390/diagnostics15202598 - 15 Oct 2025
Cited by 1 | Viewed by 1669
Abstract
Background/Objective: Dental implant therapy requires clinicians to identify edentulous regions and adjacent teeth accurately to ensure precise and efficient implant placement. However, this process is time-consuming and subject to operator bias. To address this challenge, this study proposes an AI-assisted detection framework that [...] Read more.
Background/Objective: Dental implant therapy requires clinicians to identify edentulous regions and adjacent teeth accurately to ensure precise and efficient implant placement. However, this process is time-consuming and subject to operator bias. To address this challenge, this study proposes an AI-assisted detection framework that integrates deep learning and image processing techniques to predict implant placement pathways on dental panoramic radiographs, supporting clinical decision-making. Methods: The proposed framework is first applied to YOLO models to detect edentulous regions and employs image enhancement techniques to improve image quality. Subsequently, YOLO-OBB is utilized to extract pixel-level positional information about neighboring healthy teeth. An implant pathway orientation visualization algorithm is applied to derive clinically relevant implant placement recommendations. Results: Experimental evaluation using YOLOv9m and YOLOv8n-OBB demonstrated stable performance in both recognition and accuracy. The models achieved Precision values of 88.86% and 89.82%, respectively, with an average angular error of only 1.537° compared to clinical implant pathways annotated by dentists. Conclusions: This study presents the first AI-assisted diagnostic framework for DPR-based implant pathway prediction. The results indicate strong consistency with clinical planning, confirming its potential to enhance diagnostic accuracy and provide reliable decision support in implant dentistry. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

23 pages, 1668 KB  
Article
Brain Stroke Classification Using CT Scans with Transformer-Based Models and Explainable AI
by Shomukh Qari and Maha A. Thafar
Diagnostics 2025, 15(19), 2486; https://doi.org/10.3390/diagnostics15192486 - 29 Sep 2025
Viewed by 3342
Abstract
Background & Objective: Stroke remains a leading cause of mortality and long-term disability worldwide, demanding rapid and accurate diagnosis to improve patient outcomes. Computed tomography (CT) scans are widely used in emergency settings due to their speed, availability, and cost-effectiveness. This study proposes [...] Read more.
Background & Objective: Stroke remains a leading cause of mortality and long-term disability worldwide, demanding rapid and accurate diagnosis to improve patient outcomes. Computed tomography (CT) scans are widely used in emergency settings due to their speed, availability, and cost-effectiveness. This study proposes an artificial intelligence (AI)-based framework for multiclass stroke classification (ischemic, hemorrhagic, and no stroke) using CT scan images from the Ministry of Health of the Republic of Turkey. Methods: We adopted MaxViT, a state-of-the-art Vision Transformer (ViT)-based architecture, as the primary deep learning model for stroke classification. Additional transformer variants, including Vision Transformer (ViT), Transformer-in-Transformer (TNT), and ConvNeXt, were evaluated for comparison. To improve model generalization and handle class imbalance, classical data augmentation techniques were applied. Furthermore, explainable AI (XAI) was integrated using Grad-CAM++ to provide visual insights into model decisions. Results: The MaxViT model with augmentation achieved the highest performance, reaching an accuracy and F1-score of 98.00%, outperforming the baseline Vision Transformer and other evaluated models. Grad-CAM++ visualizations confirmed that the proposed framework effectively identified stroke-related regions, enhancing transparency and clinical trust. Conclusions: This research contributes to the development of a trustworthy AI-assisted diagnostic tool for stroke, facilitating its integration into clinical practice and improving access to timely and optimal stroke diagnosis in emergency departments. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

21 pages, 2716 KB  
Article
An Explainable Deep Learning Framework for Multimodal Autism Diagnosis Using XAI GAMI-Net and Hypernetworks
by Wajeeha Malik, Muhammad Abuzar Fahiem, Tayyaba Farhat, Runna Alghazo, Awais Mahmood and Mousa Alhajlah
Diagnostics 2025, 15(17), 2232; https://doi.org/10.3390/diagnostics15172232 - 3 Sep 2025
Cited by 2 | Viewed by 2116
Abstract
Background: Autism Spectrum Disorder (ASD) is a neurodevelopmental condition characterized by heterogeneous behavioral and neurological patterns, complicating timely and accurate diagnosis. Behavioral datasets are commonly used to diagnose ASD. In clinical practice, it is difficult to identify ASD because of the complexity of [...] Read more.
Background: Autism Spectrum Disorder (ASD) is a neurodevelopmental condition characterized by heterogeneous behavioral and neurological patterns, complicating timely and accurate diagnosis. Behavioral datasets are commonly used to diagnose ASD. In clinical practice, it is difficult to identify ASD because of the complexity of the behavioral symptoms, overlap of neurological disorders, and individual heterogeneity. Correct and timely identification is dependent on the presence of skilled professionals to perform thorough neurological examinations. Nevertheless, with developments in deep learning techniques, the diagnostic process can be significantly improved by automatically identifying and automatically classifying patterns of ASD-related behaviors and neuroimaging features. Method: This study introduces a novel multimodal diagnostic paradigm that combines structured behavioral phenotypes and structural magnetic resonance imaging (sMRI) into an interpretable and personalized framework. A Generalized Additive Model with Interactions (GAMI-Net) is used to process behavioral data for transparent embedding of clinical phenotypes. Structural brain characteristics are extracted via a hybrid CNN–GNN model, which retains voxel-level patterns and region-based connectivity through the Harvard–Oxford atlas. The embeddings are then fused using an Autoencoder, compressing cross-modal data into a common latent space. A Hyper Network-based MLP classifier produces subject-specific weights to make the final classification. Results: On the held-out test set drawn from the ABIDE-I dataset, a 20% split with about 247 subjects, the constructed system achieved an accuracy of 99.40%, precision of 100%, recall of 98.84%, an F1-score of 99.42%, and an ROC-AUC of 99.99%. For another test of generalizability, five-fold stratified cross-validation on the entire dataset yielded a mean accuracy of 98.56%, an F1-score of 98.61%, precision of 98.13%, recall of 99.12%, and an ROC-AUC of 99.62%. Conclusions: These results suggest that interpretable and personalized multimodal fusion can be useful in aiding practitioners in performing effective and accurate ASD diagnosis. Nevertheless, as the test was performed on stratified cross-validation and a single held-out split, future research should seek to validate the framework on larger, multi-site datasets and different partitioning schemes to guarantee robustness over heterogeneous populations. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

24 pages, 2282 KB  
Article
Top-k Bottom All but σ Loss Strategy for Medical Image Segmentation
by Corneliu Florea, Laura Florea and Constantin Vertan
Diagnostics 2025, 15(17), 2189; https://doi.org/10.3390/diagnostics15172189 - 29 Aug 2025
Viewed by 1280
Abstract
Background/Objectives In this study we approach the problem of medical image segmentation by introducing a new loss function envelope that is derived from the Top-k loss strategy. We exploit the fact that, for semantic segmentation, the training loss is computed at two levels, [...] Read more.
Background/Objectives In this study we approach the problem of medical image segmentation by introducing a new loss function envelope that is derived from the Top-k loss strategy. We exploit the fact that, for semantic segmentation, the training loss is computed at two levels, more specifically at pixel level and at image level. Quite often, the envisaged problem has particularities that include noisy annotation at pixel level and limited data, but with accurate annotations at image level. Methods To address the mentioned issues, the Top-k strategy at image level and respectively the “Bottom all but σ” strategy at pixel level are assumed. To deal with the discontinuities of the differentials faced in the automatic learning, a derivative smoothing procedure is introduced. Results The method is thoroughly and successfully tested (in conjunction with a variety of backbone models) for several medical image segmentation tasks performed onto a variety of image acquisition types and human body regions. We present the burned skin area segmentation in standard color images, the segmentation of fetal abdominal structures in ultrasound images and ventricles and myocardium segmentation in cardiac MRI images, in all cases yielding performance improvements. Conclusions The proposed novel mechanism enhances model training by selectively emphasizing certain loss values by the use of two complementary strategies. The major benefits of the approach are clear in challenging scenarios, where the segmentation problem is inherently difficult or where the quality of pixel-level annotations is degraded by noise or inconsistencies. The proposed approach performs equally well in both convolutional neural networks (CNNs) and vision transformer (ViT) architectures. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

22 pages, 1882 KB  
Article
Optimizing CNN-Based Diagnosis of Knee Osteoarthritis: Enhancing Model Accuracy with CleanLab Relabeling
by Thomures Momenpour and Arafat Abu Mallouh
Diagnostics 2025, 15(11), 1332; https://doi.org/10.3390/diagnostics15111332 - 26 May 2025
Cited by 2 | Viewed by 3060
Abstract
Background: Knee Osteoarthritis (KOA) is a prevalent and debilitating joint disorder that significantly impacts quality of life, particularly in aging populations. Accurate and consistent classification of KOA severity, typically using the Kellgren-Lawrence (KL) grading system, is crucial for effective diagnosis, treatment planning, and [...] Read more.
Background: Knee Osteoarthritis (KOA) is a prevalent and debilitating joint disorder that significantly impacts quality of life, particularly in aging populations. Accurate and consistent classification of KOA severity, typically using the Kellgren-Lawrence (KL) grading system, is crucial for effective diagnosis, treatment planning, and monitoring disease progression. However, traditional KL grading is known for its inherent subjectivity and inter-rater variability, which underscores the pressing need for objective, automated, and reliable classification methods. Methods: This study investigates the performance of an EfficientNetB5 deep learning model, enhanced with transfer learning from the ImageNet dataset, for the task of classifying KOA severity into five distinct KL grades (0–4). We utilized a publicly available Kaggle dataset comprising 9786 knee X-ray images. A key aspect of our methodology was a comprehensive data-centric preprocessing pipeline, which involved an initial phase of outlier removal to reduce noise, followed by systematic label correction using the Cleanlab framework to identify and rectify potential inconsistencies within the original dataset labels. Results: The final EfficientNetB5 model, trained on the preprocessed and Cleanlab-remediated data, achieved an overall accuracy of 82.07% on the test set. This performance represents a significant improvement over previously reported benchmarks for five-class KOA classification on this dataset, such as ResNet-101 which achieved 69% accuracy. The substantial enhancement in model performance is primarily attributed to Cleanlab’s robust ability to detect and correct mislabeled instances, thereby improving the overall quality and reliability of the training data and enabling the model to better learn and capture complex radiographic patterns associated with KOA. Class-wise performance analysis indicated strong differentiation between healthy (KL Grade 0) and severe (KL Grade 4) cases. However, the “Doubtful” (KL Grade 1) class presented ongoing challenges, exhibiting lower recall and precision compared to other grades. When evaluated against other architectures like MobileNetV3 and Xception for multi-class tasks, our EfficientNetB5 demonstrated highly competitive results. Conclusions: The integration of an EfficientNetB5 model with a rigorous data-centric preprocessing approach, particularly Cleanlab-based label correction and outlier removal, provides a robust and significantly more accurate method for five-class KOA severity classification. While limitations in handling inherently ambiguous cases (such as KL Grade 1) and the small sample size for severe KOA warrant further investigation, this study demonstrates a promising pathway to enhance diagnostic precision. The developed pipeline shows considerable potential for future clinical applications, aiding in more objective and reliable KOA assessment. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

20 pages, 1989 KB  
Article
Deep Reinforcement Learning for CT-Based Non-Invasive Prediction of SOX9 Expression in Hepatocellular Carcinoma
by Minghui Liu, Yi Wei, Tianshu Xie, Meiyi Yang, Xuan Cheng, Lifeng Xu, Qian Li, Feng Che, Qing Xu, Bin Song and Ming Liu
Diagnostics 2025, 15(10), 1255; https://doi.org/10.3390/diagnostics15101255 - 15 May 2025
Viewed by 918
Abstract
Background: The transcription factor SOX9 plays a critical role in various diseases, including hepatocellular carcinoma (HCC), and has been implicated in resistance to sorafenib treatment. Accurate assessment of SOX9 expression is important for guiding personalized therapy in HCC patients; however, a reliable non-invasive [...] Read more.
Background: The transcription factor SOX9 plays a critical role in various diseases, including hepatocellular carcinoma (HCC), and has been implicated in resistance to sorafenib treatment. Accurate assessment of SOX9 expression is important for guiding personalized therapy in HCC patients; however, a reliable non-invasive method for evaluating SOX9 status remains lacking. This study aims to develop a deep learning (DL) model capable of preoperatively and non-invasively predicting SOX9 expression from CT images in HCC patients. Methods: We retrospectively analyzed a dataset comprising 4011 CT images from 101 HCC patients who underwent surgical resection followed by sorafenib therapy at West China Hospital, Sichuan University. A deep reinforcement learning (DRL) approach was proposed to enhance prediction accuracy by identifying and focusing on image regions highly correlated with SOX9 expression, thereby reducing the impact of background noise. Results: Our DRL-based model achieved an area under the curve (AUC) of 91.00% (95% confidence interval: 88.64–93.15%), outperforming conventional DL methods by over 10%. Furthermore, survival analysis revealed that patients with SOX9-positive tumors had significantly shorter recurrence-free survival (RFS) and overall survival (OS) compared to SOX9-negative patients, highlighting the prognostic value of SOX9 status. Conclusions: This study demonstrates that a DRL-enhanced DL model can accurately and non-invasively predict SOX9 expression in HCC patients using preoperative CT images. These findings support the clinical utility of imaging-based SOX9 assessment in informing treatment strategies and prognostic evaluation for patients with advanced HCC. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

Review

Jump to: Research, Other

36 pages, 1320 KB  
Review
A Review of U-Net Based Deep Learning Frameworks for MRI-Based Brain Tumor Segmentation
by Ayse Bastug Koc and Devrim Akgun
Diagnostics 2026, 16(4), 506; https://doi.org/10.3390/diagnostics16040506 (registering DOI) - 7 Feb 2026
Abstract
Automated segmentation of brain tumors from Magnetic Resonance Imaging (MRI) images is helpful for clinical diagnosis, surgical planning, and post-treatment monitoring. In recent years, the U-Net architecture has been observed as one of the most popular solutions among deep learning models. This article [...] Read more.
Automated segmentation of brain tumors from Magnetic Resonance Imaging (MRI) images is helpful for clinical diagnosis, surgical planning, and post-treatment monitoring. In recent years, the U-Net architecture has been observed as one of the most popular solutions among deep learning models. This article presents a review of 35 studies published between 2019 and 2025 focusing on U-Net-based brain tumor segmentation. The primary focus of this review is an in-depth analysis of commonly used U-Net architectures. The transformation of original 2D and 3D models into more advanced variants is examined in detail. Results from a wide range of studies are synthesized, and standard evaluation criteria are summarized along with benchmark datasets such as the BRATS competition to validate the effectiveness of these models. Additionally, the paper overviews the recent developments in the field, determines fundamental challenges, and provides insight into future directions, including improving model efficiency and generalization, combining multimodal data, and advancing clinical applications. This review serves as a guide for researchers to examine the impact of the U-Net architecture on brain tumor segmentation. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)

Other

Jump to: Research, Review

9 pages, 1528 KB  
Brief Report
Impact of Deep Learning-Based Reconstruction on the Accuracy and Precision of Cardiac Tissue Characterization
by Margarita Gorodezky, Linda Reichardt, Tom Geisler, Marc-André Weber, Felix G. Meinel and Ann-Christin Klemenz
Diagnostics 2026, 16(2), 348; https://doi.org/10.3390/diagnostics16020348 - 21 Jan 2026
Viewed by 155
Abstract
Background/Objectives: Interest in myocardial mapping for cardiac MRI has increased, enabling differentiation of various cardiac diseases through T1, T2, and T2* mapping. This study evaluates the impact of deep learning (DL)-based image reconstruction on the mean and standard deviation (SD) of these techniques. [...] Read more.
Background/Objectives: Interest in myocardial mapping for cardiac MRI has increased, enabling differentiation of various cardiac diseases through T1, T2, and T2* mapping. This study evaluates the impact of deep learning (DL)-based image reconstruction on the mean and standard deviation (SD) of these techniques. Methods: Fifty healthy adults underwent cardiac MRI, with images reconstructed using the AIR Recon DL prototype. This DL approach, which reduces noise and enhances image quality, was applied at three levels and compared to non-DL reconstructions. Results: Analysis focused on the septum to minimize artifacts. For T1 mapping, mean values were 988 ± 50, 981 ± 45, 982 ± 43, and 980 ± 24 ms; for T2 mapping, mean values were 53 ± 5, 54 ± 5, 54 ± 5, and 54 ± 5 ms and for T2* mapping, mean values were 37 ± 5, 37 ± 5, 37 ± 5, and 38 ± 5 ms for no DL and increasing DL levels, respectively. Results showed no significant differences in mean values for any mappings between non-DL and DL reconstructions. However, DL significantly reduced the SD within regions of interest for T1 mapping, enhancing image sharpness and registration accuracy. No significant SD reduction was observed for T2 and T2* mappings. Conclusions: These findings suggest that DL-based reconstructions improve the precision of T1 mapping without affecting mean values, supporting their clinical integration for more accurate cardiac disease diagnosis. Future studies should include patient cohorts and optimized protocols to further validate these findings. Full article
(This article belongs to the Special Issue 3rd Edition: AI/ML-Based Medical Image Processing and Analysis)
Show Figures

Figure 1

Back to TopTop