Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (76)

Search Parameters:
Keywords = nnUNet

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 27829 KB  
Article
Deep Learning Strategies for Semantic Segmentation in Robot-Assisted Radical Prostatectomy
by Elena Sibilano, Claudia Delprete, Pietro Maria Marvulli, Antonio Brunetti, Francescomaria Marino, Giuseppe Lucarelli, Michele Battaglia and Vitoantonio Bevilacqua
Appl. Sci. 2025, 15(19), 10665; https://doi.org/10.3390/app151910665 - 2 Oct 2025
Viewed by 277
Abstract
Robot-assisted radical prostatectomy (RARP) has become the most prevalent treatment for patients with organ-confined prostate cancer. Despite superior outcomes, suboptimal vesicourethral anastomosis (VUA) may lead to serious complications, including urinary leakage, prolonged catheterization, and extended hospitalization. A precise localization of both the surgical [...] Read more.
Robot-assisted radical prostatectomy (RARP) has become the most prevalent treatment for patients with organ-confined prostate cancer. Despite superior outcomes, suboptimal vesicourethral anastomosis (VUA) may lead to serious complications, including urinary leakage, prolonged catheterization, and extended hospitalization. A precise localization of both the surgical needle and the surrounding vesical and urethral tissues to coadapt is needed for fine-grained assessment of this task. Nonetheless, the identification of anatomical structures from endoscopic videos is difficult due to tissue distortions, changes in brightness, and instrument interferences. In this paper, we propose and compare two Deep Learning (DL) pipelines for the automatic segmentation of the mucosal layers and the suturing needle in real RARP videos by exploiting different architectures and training strategies. To train the models, we introduce a novel, annotated dataset collected from four VUA procedures. Experimental results show that the nnU-Net 2D model achieved the highest class-specific metrics, with a Dice Score of 0.663 for the mucosa class and 0.866 for the needle class, outperforming both transformer-based and baseline convolutional approaches on external validation video sequences. This work paves the way for computer-assisted tools that can objectively evaluate surgical performance during the critical phase of suturing tasks. Full article
Show Figures

Figure 1

32 pages, 9105 KB  
Article
Development of Semi-Automatic Dental Image Segmentation Workflows with Root Canal Recognition for Faster Ground Tooth Acquisition
by Yousef Abo El Ela and Mohamed Badran
J. Imaging 2025, 11(10), 340; https://doi.org/10.3390/jimaging11100340 - 1 Oct 2025
Viewed by 248
Abstract
This paper investigates the application of image segmentation techniques in endodontics, focusing on improving diagnostic accuracy and achieving faster segmentation by delineating specific dental regions such as teeth and root canals. Deep learning architectures, notably 3D U-Net and GANs, have advanced the image [...] Read more.
This paper investigates the application of image segmentation techniques in endodontics, focusing on improving diagnostic accuracy and achieving faster segmentation by delineating specific dental regions such as teeth and root canals. Deep learning architectures, notably 3D U-Net and GANs, have advanced the image segmentation process for dental structures, supporting more precise dental procedures. However, challenges like the demand for extensive labeled datasets and ensuring model generalizability remain. Two semi-automatic segmentation workflows, Grow From Seeds (GFS) and Watershed (WS), were developed to provide quicker acquisition of ground truth training data for deep learning models using 3D Slicer software version 5.8.1. These workflows were evaluated against a manual segmentation benchmark and a recent dental segmentation automated tool on three separate datasets. The evaluations were performed by the overall shapes of a maxillary central incisor and a maxillary second molar and by the region of the root canal of both teeth. Results from Kruskal–Wallis and Nemenyi tests indicated that the semi-automated workflows, more often than not, were not statistically different from the manual benchmark based on dice coefficient similarity, while the automated method consistently provided significantly different 3D models from their manual counterparts. The study also explores the benefits of labor reduction and time savings achieved by the semi-automated methods. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

18 pages, 2124 KB  
Article
Automated Subregional Hippocampus Segmentation Using 3D CNNs: A Computational Framework for Brain Aging Biomarker Analysis
by Eshaa Gogia, Arash Dehzangi and Iman Dehzangi
Algorithms 2025, 18(8), 509; https://doi.org/10.3390/a18080509 - 13 Aug 2025
Viewed by 713
Abstract
The hippocampus is a critical brain structure involved in episodic memory, spatial orientation, and stress regulation. Its volumetric shrinkage is among the earliest and most reliable indicators of both physiological brain aging and pathological neurodegeneration. Accurate segmentation and measurement of the hippocampal subregions [...] Read more.
The hippocampus is a critical brain structure involved in episodic memory, spatial orientation, and stress regulation. Its volumetric shrinkage is among the earliest and most reliable indicators of both physiological brain aging and pathological neurodegeneration. Accurate segmentation and measurement of the hippocampal subregions from magnetic resonance imaging (MRI) is therefore essential for neurobiological age estimation and the early identification of at-risk individuals. In this study, we present a fully automated pipeline that leverages nnU-Net, a self-configuring deep learning framework, to segment the hippocampus from high-resolution 3D T1-weighted brain MRI scans. The primary objective of this work is to enable accurate estimation of brain age through quantitative analysis of hippocampal volume. By fusing domain knowledge in neuroanatomy with data-driven learning through a highly expressive and self-optimizing model, this work advances the methodological frontier for neuroimaging-based brain-age estimation. The proposed approach demonstrates that deep learning can serve as a reliable segmentation tool as well as a foundational layer in predictive neuroscience, supporting early detection of accelerated aging and subclinical neurodegenerative processes. Full article
Show Figures

Figure 1

24 pages, 3617 KB  
Article
A Comparison Between Unimodal and Multimodal Segmentation Models for Deep Brain Structures from T1- and T2-Weighted MRI
by Nicola Altini, Erica Lasaracina, Francesca Galeone, Michela Prunella, Vladimiro Suglia, Leonarda Carnimeo, Vito Triggiani, Daniele Ranieri, Gioacchino Brunetti and Vitoantonio Bevilacqua
Mach. Learn. Knowl. Extr. 2025, 7(3), 84; https://doi.org/10.3390/make7030084 - 13 Aug 2025
Cited by 1 | Viewed by 1086
Abstract
Accurate segmentation of deep brain structures is critical for preoperative planning in such neurosurgical procedures as Deep Brain Stimulation (DBS). Previous research has showcased successful pipelines for segmentation from T1-weighted (T1w) Magnetic Resonance Imaging (MRI) data. Nevertheless, the role of T2-weighted (T2w) MRI [...] Read more.
Accurate segmentation of deep brain structures is critical for preoperative planning in such neurosurgical procedures as Deep Brain Stimulation (DBS). Previous research has showcased successful pipelines for segmentation from T1-weighted (T1w) Magnetic Resonance Imaging (MRI) data. Nevertheless, the role of T2-weighted (T2w) MRI data has been underexploited so far. This study proposes and evaluates a fully automated deep learning pipeline based on nnU-Net for the segmentation of eight clinically relevant deep brain structures. A heterogeneous dataset has been prepared by gathering 325 paired T1w and T2w MRI scans from eight publicly available sources, which have been annotated by means of an atlas-based registration approach. Three 3D nnU-Net models—unimodal T1w, unimodal T2w, and multimodal (encompassing both T1w and T2w)—have been trained and compared by using 5-fold cross-validation and a separate test set. The outcomes prove that the multimodal model consistently outperforms the T2w unimodal model and achieves comparable performance with the T1w unimodal model. On our dataset, all proposed models significantly exceed the performance of the state-of-the-art DBSegment tool. These findings underscore the value of multimodal MRI in enhancing deep brain segmentation and offer a robust framework for accurate delineation of subcortical targets in both research and clinical settings. Full article
(This article belongs to the Special Issue Deep Learning in Image Analysis and Pattern Recognition, 2nd Edition)
Show Figures

Figure 1

15 pages, 1193 KB  
Article
Enhanced Brain Stroke Lesion Segmentation in MRI Using a 2.5D Transformer Backbone U-Net Model
by Mahsa Karimzadeh, Hadi Seyedarabi, Ata Jodeiri and Reza Afrouzian
Brain Sci. 2025, 15(8), 778; https://doi.org/10.3390/brainsci15080778 - 22 Jul 2025
Viewed by 1125
Abstract
Background/Objectives: Accurate segmentation of brain stroke lesions from MRI images is a critical task in medical image analysis that is essential for timely diagnosis and treatment planning. Methods: This paper presents a novel approach for segmenting brain stroke lesions using a deep learning [...] Read more.
Background/Objectives: Accurate segmentation of brain stroke lesions from MRI images is a critical task in medical image analysis that is essential for timely diagnosis and treatment planning. Methods: This paper presents a novel approach for segmenting brain stroke lesions using a deep learning model based on the U-Net neural network architecture. We enhanced the traditional U-Net by integrating a transformer-based backbone, specifically the Mix Vision Transformer (MiT), and compared its performance against other commonly used backbones such as ResNet and EfficientNet. Additionally, we implemented a 2.5D method, which leverages 2D networks to process three-dimensional data slices, effectively balancing the rich spatial context of 3D methods and the simplicity of 2D methods. The 2.5D approach captures inter-slice dependencies, leading to improved lesion delineation without the computational complexity of full 3D models. Utilizing the 2015 ISLES dataset, which includes MRI images and corresponding lesion masks for 20 patients, we conducted our experiments with 4-fold cross-validation to ensure robustness and reliability. To evaluate the effectiveness of our method, we conducted comparative experiments with several state-of-the-art (SOTA) segmentation models, including CNN-based UNet, nnU-Net, TransUNet, and SwinUNet. Results: Our proposed model outperformed all competing methods in terms of Dice Coefficient and Intersection over Union (IoU), demonstrating its robustness and superiority. Our extensive experiments demonstrate that the proposed U-Net with the MiT Backbone, combined with 2.5D data preparation, achieves superior performance metrics, specifically achieving DICE and IoU scores of 0.8153 ± 0.0101 and 0.7835 ± 0.0079, respectively, outperforming other backbone configurations. Conclusions: These results indicate that the integration of transformer-based backbones and 2.5D techniques offers a significant advancement in the accurate segmentation of brain stroke lesions, paving the way for more reliable and efficient diagnostic tools in clinical settings. Full article
(This article belongs to the Section Neural Engineering, Neuroergonomics and Neurorobotics)
Show Figures

Figure 1

12 pages, 1493 KB  
Article
Automatic Segmentation of the Infraorbital Canal in CBCT Images: Anatomical Structure Recognition Using Artificial Intelligence
by Ismail Gumussoy, Emre Haylaz, Suayip Burak Duman, Fahrettin Kalabalık, Muhammet Can Eren, Seyda Say, Ozer Celik and Ibrahim Sevki Bayrakdar
Diagnostics 2025, 15(13), 1713; https://doi.org/10.3390/diagnostics15131713 - 4 Jul 2025
Cited by 2 | Viewed by 641
Abstract
Background/Objectives: The infraorbital canal (IOC) is a critical anatomical structure that passes through the anterior surface of the maxilla and opens at the infraorbital foramen, containing the infraorbital nerve, artery, and vein. Accurate localization of this canal in maxillofacial, dental implant, and orbital [...] Read more.
Background/Objectives: The infraorbital canal (IOC) is a critical anatomical structure that passes through the anterior surface of the maxilla and opens at the infraorbital foramen, containing the infraorbital nerve, artery, and vein. Accurate localization of this canal in maxillofacial, dental implant, and orbital surgeries is of great importance to preventing nerve damage, reducing complications, and enabling successful surgical planning. The aim of this study is to perform automatic segmentation of the infraorbital canal in cone-beam computed tomography (CBCT) images using an artificial intelligence (AI)-based model. Methods: A total of 220 CBCT images of the IOC from 110 patients were labeled using the 3D Slicer software (version 4.10.2; MIT, Cambridge, MA, USA). The dataset was split into training, validation, and test sets at a ratio of 8:1:1. The nnU-Net v2 architecture was applied to the training and test datasets to predict and generate appropriate algorithm weight factors. The confusion matrix was used to check the accuracy and performance of the model. As a result of the test, the Dice Coefficient (DC), Intersection over the Union (IoU), F1-score, and 95% Hausdorff distance (95% HD) metrics were calculated. Results: By testing the model, the DC, IoU, F1-score, and 95% HD metric values were found to be 0.7792, 0.6402, 0.787, and 0.7661, respectively. According to the data obtained, the receiver operating characteristic (ROC) curve was drawn, and the AUC value under the curve was determined to be 0.91. Conclusions: Accurate identification and preservation of the IOC during surgical procedures are of critical importance to maintaining a patient’s functional and sensory integrity. The findings of this study demonstrated that the IOC can be detected with high precision and accuracy using an AI-based automatic segmentation method in CBCT images. This approach has significant potential to reduce surgical risks and to enhance the safety of critical anatomical structures. Full article
Show Figures

Figure 1

17 pages, 1863 KB  
Article
MedSAM/MedSAM2 Feature Fusion: Enhancing nnUNet for 2D TOF-MRA Brain Vessel Segmentation
by Han Zhong, Jiatian Zhang and Lingxiao Zhao
J. Imaging 2025, 11(6), 202; https://doi.org/10.3390/jimaging11060202 - 18 Jun 2025
Viewed by 1852
Abstract
Accurate segmentation of brain vessels is critical for diagnosing cerebral stroke, yet existing AI-based methods struggle with challenges such as small vessel segmentation and class imbalance. To address this, our study proposes a novel 2D segmentation method based on the nnUNet framework, enhanced [...] Read more.
Accurate segmentation of brain vessels is critical for diagnosing cerebral stroke, yet existing AI-based methods struggle with challenges such as small vessel segmentation and class imbalance. To address this, our study proposes a novel 2D segmentation method based on the nnUNet framework, enhanced with MedSAM/MedSAM2 features, for arterial vessel segmentation in time-of-flight magnetic resonance angiography (TOF-MRA) brain slices. The approach first constructs a baseline segmentation network using nnUNet, then incorporates MedSAM/MedSAM2’s feature extraction module to enhance feature representation. Additionally, focal loss is introduced to address class imbalance. Experimental results on the CAS2023 dataset demonstrate that the MedSAM2-enhanced model achieves a 0.72% relative improvement in Dice coefficient and reduces HD95 (mm) and ASD (mm) from 48.20 mm to 46.30 mm and from 5.33 mm to 4.97 mm, respectively, compared to the baseline nnUNet, showing significant enhancements in boundary localization and segmentation accuracy. This approach addresses the critical challenge of small vessel segmentation in TOF-MRA, with the potential to improve cerebrovascular disease diagnosis in clinical practice. Full article
(This article belongs to the Section AI in Imaging)
Show Figures

Figure 1

20 pages, 1669 KB  
Article
Automated Pneumothorax Segmentation with a Spatial Prior Contrast Adapter
by Yiming Jia and Essam A. Rashed
Appl. Sci. 2025, 15(12), 6598; https://doi.org/10.3390/app15126598 - 12 Jun 2025
Viewed by 962
Abstract
Pneumothorax is a critical condition that requires rapid and accurate diagnosis from standard chest radiographs. Identifying and segmenting the location of the pneumothorax are essential for developing an effective treatment plan. nnUNet is a self-configuring, deep learning-based framework for medical image segmentation. Despite [...] Read more.
Pneumothorax is a critical condition that requires rapid and accurate diagnosis from standard chest radiographs. Identifying and segmenting the location of the pneumothorax are essential for developing an effective treatment plan. nnUNet is a self-configuring, deep learning-based framework for medical image segmentation. Despite adjusting its parameters automatically through data-driven optimization strategies and offering robust feature extraction and segmentation capabilities across diverse datasets, our initial experiments revealed that nnUNet alone struggled to achieve consistently accurate segmentation for pneumothorax, particularly in challenging scenarios where subtle intensity variations and anatomical noise obscure the target regions. This study aims to enhance the accuracy and robustness of pneumothorax segmentation in low-contrast chest radiographs by integrating spatial prior information and attention mechanism into the nnUNet framework. In this study, we introduce the spatial prior contrast adapter (SPCA)-enhanced nnUNet by implementing two modules. First, we integrate an SPCA utilizing the MedSAM foundation model to incorporate spatial prior information of the lung region, effectively guiding the segmentation network to focus on anatomically relevant areas. In the meantime, a probabilistic atlas, which shows the probability of an area prone to pneumothorax, is generated based on the ground truth masks. Both the lung segmentation results and the probabilistic atlas are used as attention maps in nnUNet. Second, we combine the two attention maps as additional input into nnUNet and integrate an attention mechanism into standard nnUNet by using a convolutional block attention module (CBAM). We validate our method by experimenting on the dataset CANDID-PTX, a benchmark dataset representing 19,237 chest radiographs. By introducing spatial awareness and intensity adjustments, the model reduces false positives and improves the precision of boundary delineations, ultimately overcoming many of the limitations associated with low-contrast radiographs. Compared with standard nnUNet, SPCA-enhanced nnUNet achieves an average Dice coefficient of 0.81, which indicates an improvement of standard nnUNet by 15%. This study provides a novel approach toward enhancing the segmentation performance of pneumothorax with low contrast in chest X-ray radiographs. Full article
(This article belongs to the Special Issue Applications of Computer Vision and Image Processing in Medicine)
Show Figures

Figure 1

14 pages, 2063 KB  
Article
Deep Learning Based Automatic Ankle Tenosynovitis Quantification from MRI in Patients with Psoriatic Arthritis: A Feasibility Study
by Saeed Arbabi, Vahid Arbabi, Lorenzo Costa, Iris ten Katen, Simon C. Mastbergen, Peter R. Seevinck, Pim A. de Jong, Harrie Weinans, Mylène P. Jansen and Wouter Foppen
Diagnostics 2025, 15(12), 1469; https://doi.org/10.3390/diagnostics15121469 - 9 Jun 2025
Cited by 1 | Viewed by 794
Abstract
Background/Objectives: Tenosynovitis is a common feature of psoriatic arthritis (PsA) and is typically assessed using semi-quantitative magnetic resonance imaging (MRI) scoring. However, visual scoring s variability. This study evaluates a fully automated, deep-learning approach for ankle tenosynovitis segmentation and volume-based quantification from MRI [...] Read more.
Background/Objectives: Tenosynovitis is a common feature of psoriatic arthritis (PsA) and is typically assessed using semi-quantitative magnetic resonance imaging (MRI) scoring. However, visual scoring s variability. This study evaluates a fully automated, deep-learning approach for ankle tenosynovitis segmentation and volume-based quantification from MRI in psoriatic arthritis (PsA) patients. Methods: We analyzed 364 ankle 3T MRI scans from 71 PsA patients. Four tenosynovitis pathologies were manually scored and used to create ground truth segmentations through a human–machine workflow. For each pathology, 30 annotated scans were used to train a deep-learning segmentation model based on the nnUNet framework, and 20 scans were used for testing, ensuring patient-level disjoint sets. Model performance was evaluated using Dice scores. Volumetric pathology measurements from test scans were compared to radiologist scores using Spearman correlation. Additionally, 218 serial MRI pairs were assessed to analyze the relationship between changes in pathology volume and changes in visual scores. Results: The segmentation model achieved promising performance on the test set, with mean Dice scores ranging from 0.84 to 0.92. Pathology volumes correlated with visual scores across all test MRIs (Spearman ρ = 0.52–0.62). Volume-based quantification captured changes in inflammation over time and identified subtle progression not reflected in semi-quantitative scores. Conclusions: Our automated segmentation tool enables fast and accurate quantification of ankle tenosynovitis in PsA patients. It may enhance sensitivity to disease progression and complement visual scoring through continuous, volume-based metrics. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

17 pages, 3455 KB  
Article
Segment Anything Model (SAM) and Medical SAM (MedSAM) for Lumbar Spine MRI
by Christian Chang, Hudson Law, Connor Poon, Sydney Yen, Kaustubh Lall, Armin Jamshidi, Vadim Malis, Dosik Hwang and Won C. Bae
Sensors 2025, 25(12), 3596; https://doi.org/10.3390/s25123596 - 7 Jun 2025
Cited by 1 | Viewed by 2521
Abstract
Lumbar spine Magnetic Resonance Imaging (MRI) is commonly used for intervertebral disc (IVD) and vertebral body (VB) evaluation during low back pain. Segmentation of these tissues can provide useful quantitative information such as shape and volume. The objective of the study was to [...] Read more.
Lumbar spine Magnetic Resonance Imaging (MRI) is commonly used for intervertebral disc (IVD) and vertebral body (VB) evaluation during low back pain. Segmentation of these tissues can provide useful quantitative information such as shape and volume. The objective of the study was to determine the performances of Segment Anything Model (SAM) and medical SAM (MedSAM), two “zero-shot” deep learning models, in segmenting lumbar IVD and VB from MRI images and compare against the nnU-Net model. This cadaveric study used 82 donor spines. Manual segmentation was performed to serve as ground truth. Two readers processed the spine MRI using SAM and MedSAM by placing points or drawing bounding boxes around regions of interest (ROI). The outputs were compared against ground truths to determine Dice score, sensitivity, and specificity. Qualitatively, results varied but overall, MedSAM produced more consistent results than SAM, but neither matched the performance of nnU-Net. Mean Dice scores for MedSAM were 0.79 for IVDs and 0.88 for VBs, and significantly higher (each p < 0.001) than those for SAM (0.64 for IVDs, 0.83 for VBs). Both were lower compared to nnU-Net (0.99 for IVD and VB). Sensitivity values also favored MedSAM. These results demonstrated the feasibility of “zero-shot” DL models to segment lumbar spine MRI. While performance falls short of recent models, these zero-shot models offer key advantages in not needing training data and faster adaptation to other anatomies and tasks. Validation of a generalizable segmentation model for lumbar spine MRI can lead to more precise diagnostics, follow-up, and enhanced back pain research, with potential cost savings from automated analyses while supporting the broader use of AI and machine learning in healthcare. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

38 pages, 13454 KB  
Article
Deep Learning Innovations: ResNet Applied to SAR and Sentinel-2 Imagery
by Giuliana Bilotta, Luigi Bibbò, Giuseppe M. Meduri, Emanuela Genovese and Vincenzo Barrile
Remote Sens. 2025, 17(12), 1961; https://doi.org/10.3390/rs17121961 - 6 Jun 2025
Viewed by 1841
Abstract
The elevated precision of data regarding the Earth’s surface, facilitated by the enhanced interoperability among various GNSSs (Global Navigation Satellite Systems), enables the classification of land use and land cover (LULC) via satellites equipped with optical sensors, such as Sentinel-2 of the Copernicus [...] Read more.
The elevated precision of data regarding the Earth’s surface, facilitated by the enhanced interoperability among various GNSSs (Global Navigation Satellite Systems), enables the classification of land use and land cover (LULC) via satellites equipped with optical sensors, such as Sentinel-2 of the Copernicus program, which is crucial for land use management and environmental planning. Likewise, data from SAR satellites, such Copernicus’ Sentinel-1 and Jaxa’s ALOS PALSAR, provide diverse environmental investigations, allowing different types of spatial information to be analysed thanks to the particular features of analysis based on radar. Nonetheless, in optical satellites, the relatively low resolution of Sentinel-2 satellites may impede the precision of supervised AI classifiers, crucial for ongoing land use monitoring, especially during the training phase, which can be expensive due to the requirement for advanced technology and extensive training datasets. This project aims to develop an AI classifier utilising high-resolution training data and the resilient architecture of ResNet, in conjunction with the Remote Sensing Image Classification Benchmark (RSI-CB128). ResNet, noted for its deep residual learning capabilities, significantly enhances the classifier’s proficiency in identifying intricate patterns and features from high-resolution images. A test dataset derived from Sentinel-2 raster images is utilised to evaluate the effectiveness of the neural network (NN). Our goals are to thoroughly assess and confirm the efficacy of an AI classifier utilised on high-resolution Sentinel-2 photos. The findings indicate substantial enhancements compared to current classification methods, such as U-Net, Vision Transformer (ViT), and OBIA, underscoring ResNet’s transformative capacity to elevate the precision of land use classification. Full article
Show Figures

Figure 1

20 pages, 5630 KB  
Article
Deep Learning for Automated Ventricle and Periventricular Space Segmentation on CT and T1CE MRI in Neuro-Oncology Patients
by Mart Wubbels, Marvin Ribeiro, Jelmer M. Wolterink, Wouter van Elmpt, Inge Compter, David Hofstede, Nikolina E. Birimac, Femke Vaassen, Kati Palmgren, Hendrik H. G. Hansen, Hiska L. van der Weide, Charlotte L. Brouwer, Miranda C. A. Kramer, Daniëlle B. P. Eekers and Catharina M. L. Zegers
Cancers 2025, 17(10), 1598; https://doi.org/10.3390/cancers17101598 - 8 May 2025
Viewed by 1071
Abstract
Purpose: This study aims to create a deep learning (DL) model capable of accurately delineating the ventricles, and by extension, the periventricular space (PVS), following the 2021 EPTN Neuro-Oncology Atlas guidelines on T1-weighted contrast-enhanced MRI scans (T1CE). The performance of this DL model [...] Read more.
Purpose: This study aims to create a deep learning (DL) model capable of accurately delineating the ventricles, and by extension, the periventricular space (PVS), following the 2021 EPTN Neuro-Oncology Atlas guidelines on T1-weighted contrast-enhanced MRI scans (T1CE). The performance of this DL model was quantitatively and qualitatively compared with an off-the-shelf model. Materials and Methods: An nnU-Net was trained for ventricle segmentation using both CT and T1CE MRI images from 78 patients. Its performance was compared to that of a publicly available pretrained segmentation model, SynthSeg. The evaluation was conducted on both internal (N = 18) and external (n = 18) test sets, with each consisting of paired CT and T1CE MRI images and expert-delineated ground truths (GTs). Segmentation accuracy was assessed using the volumetric Dice Similarity Coefficient (DSC), 95th percentile Hausdorff distance (HD95), surface DSC, and added path length (APL). Additionally, a local evaluation of ventricle segmentations quantified differences between manual and automatic segmentations across both test sets. All segmentations were scored by radiotherapy technicians for clinical acceptability using a 4-point Likert scale. Results: The nnU-Net significantly outperformed the SynthSeg model on the internal test dataset in terms of median [range] DSC, 0.93 [0.86–0.95] vs. 0.85 [0.67–0.91], HD95, 0.9 [0.7–2.5] mm vs. 2.2 [1.7–4.8] mm, surface DSC, 0.97 [0.90–0.98] vs. 0.84 [0.70–0.89], and APL, 876 [407–1298] mm vs. 2809 [2311–3622] mm, all with p < 0.001. No significant differences in these metrics were found in the external test set. However clinical ratings favored nnU-Net segmentations on the internal and external test sets. In addition, the nnU-Net had higher clinical ratings than the GT delineation on the internal and external test set. Conclusions: The nnU-Net model outperformed the SynthSeg model on the internal dataset in both segmentation metrics and clinician ratings. While segmentation metrics showed no significant differences between the models on the external set, clinician ratings favored nnU-Net, suggesting enhanced clinical acceptability. This suggests that nnU-Net could contribute to more time-efficient and streamlined radiotherapy planning workflows. Full article
Show Figures

Figure 1

26 pages, 8000 KB  
Article
Patient-Specific Hyperparameter Optimization of a Deep Learning-Based Tumor Autocontouring Algorithm on 2D Liver, Prostate, and Lung Cine MR Images: A Pilot Study
by Gawon Han, Keith Wachowicz, Nawaid Usmani, Don Yee, Jordan Wong, Arun Elangovan, Jihyun Yun and B. Gino Fallone
Algorithms 2025, 18(4), 233; https://doi.org/10.3390/a18040233 - 18 Apr 2025
Cited by 2 | Viewed by 763
Abstract
Linear accelerator–magnetic resonance (linac-MR) hybrid systems allow for real-time magnetic resonance imaging (MRI)-guided radiotherapy for more accurate dose delivery to the tumor and improved sparing of the adjacent healthy tissues. However, for real-time tumor detection, it is unfeasible for a human expert to [...] Read more.
Linear accelerator–magnetic resonance (linac-MR) hybrid systems allow for real-time magnetic resonance imaging (MRI)-guided radiotherapy for more accurate dose delivery to the tumor and improved sparing of the adjacent healthy tissues. However, for real-time tumor detection, it is unfeasible for a human expert to manually contour (gold standard) the tumor at the fast imaging rate of a linac-MR. This study aims to develop a neural network-based tumor autocontouring algorithm with patient-specific hyperparameter optimization (HPO) and to validate its contouring accuracy using in vivo MR images of cancer patients. Two-dimensional (2D) intrafractional MR images were acquired at 4 frames/s using 3 tesla (T) MRI from 11 liver, 24 prostate, and 12 lung cancer patients. A U-Net architecture was applied for tumor autocontouring and was further enhanced by implementing HPO using the Covariance Matrix Adaptation Evolution Strategy. Six hyperparameters were optimized for each patient, for which intrafractional images and experts’ manual contours were input into the algorithm to find the optimal set of hyperparameters. For evaluation, Dice’s coefficient (DC), centroid displacement (CD), and Hausdorff distance (HD) were computed between the manual contours and autocontours. The performance of the algorithm was benchmarked against two standardized autosegmentation methods: non-optimized U-Net and nnU-Net. For the proposed algorithm, the mean (standard deviation) DC, CD, and HD of the 47 patients were 0.92 (0.04), 1.35 (1.03), and 3.63 (2.17) mm, respectively. Compared to the two benchmarking autosegmentation methods, the proposed algorithm achieved the best overall performance in terms of contouring accuracy and speed. This work presents the first tumor autocontouring algorithm applicable to the intrafractional MR images of liver and prostate cancer patients for real-time tumor-tracked radiotherapy. The proposed algorithm performs patient-specific HPO, enabling accurate tumor delineation comparable to that of experts. Full article
(This article belongs to the Special Issue Machine Learning in Medical Signal and Image Processing (3rd Edition))
Show Figures

Figure 1

15 pages, 3821 KB  
Article
MATLAB Application for User-Friendly Design of Fully Convolutional Data Description Models for Defect Detection of Industrial Products and Its Concurrent Visualization
by Fusaomi Nagata, Shingo Sakata, Keigo Watanabe, Maki K. Habib and Ahmad Shahrizan Abdul Ghani
Machines 2025, 13(4), 328; https://doi.org/10.3390/machines13040328 - 17 Apr 2025
Viewed by 701
Abstract
In this paper, a fully convolutional data description (FCDD) model is applied to defect detection and its concurrent visualization for industrial products and materials. The authors’ propose a MATLAB application that enables users to efficiently and in a user-friendly way design, train, and [...] Read more.
In this paper, a fully convolutional data description (FCDD) model is applied to defect detection and its concurrent visualization for industrial products and materials. The authors’ propose a MATLAB application that enables users to efficiently and in a user-friendly way design, train, and test various kinds of neural network (NN) models for defect detection. Models supported by the application include the following original designs: convolutional neural network (CNN), transfer learning-based CNN, NN-based support vector machine (SVM), convolutional autoencoder (CAE), variational autoencoder (VAE), fully convolution network (FCN) (such as U-Net), and YOLO. However, FCDD is not yet supported. This paper includes the software development of the MATLAB R2024b application, which is extended to be able to build FCDD models. In particular, a systematic threshold determination method is proposed to obtain the best performance for defect detection from FCDD models. Also, through three different kinds of defect detection experiments, the usefulness and effectiveness of FCDD models in terms of defect detection and its concurrent visualization are quantitatively and qualitatively evaluated by comparing conventional transfer learning-based CNN models. Full article
(This article belongs to the Section Machines Testing and Maintenance)
Show Figures

Figure 1

19 pages, 691 KB  
Review
Artificial Intelligence in Decoding Ocular Enigmas: A Literature Review of Choroidal Nevus and Choroidal Melanoma Assessment
by Konstantina-Eleni Karamanli, Eirini Maliagkani, Petros Petrou, Elpiniki Papageorgiou and Ilias Georgalas
Appl. Sci. 2025, 15(7), 3565; https://doi.org/10.3390/app15073565 - 25 Mar 2025
Cited by 1 | Viewed by 1334
Abstract
This review examines the role of artificial intelligence (AI) in differentiating choroidal nevus (CN) from choroidal melanoma (CM), focusing on diagnosis, classification, segmentation, and prediction of malignant transformation. A literature search was performed in PubMed, Google Scholar, and Science Direct up to December [...] Read more.
This review examines the role of artificial intelligence (AI) in differentiating choroidal nevus (CN) from choroidal melanoma (CM), focusing on diagnosis, classification, segmentation, and prediction of malignant transformation. A literature search was performed in PubMed, Google Scholar, and Science Direct up to December 2024. Eight studies met all the inclusion criteria, evaluating machine learning (ML) and deep learning (DL) applications for CN and CM assessment using various ophthalmic imaging modalities. Performance varied across AI models. U-Net achieved 100% sensitivity and an AUC of 0.88, while DenseNet121 reached an AUC of 0.9781, LASSO logistic regression an AUC of 0.88, RETFound (a self-supervised learning model) had an AUCROC of 0.92, and ResNet50 an accuracy of 92.65% in classification tasks. DeepLabv3 achieved Dice scores of 0.87 (melanoma) and 0.81 (nevi) in lesion-based segmentation, while nnU-Net yielded a Dice score of 0.78 and a recall of 0.77 for pigmented lesion segmentation. SAINT (XGBoost-based) demonstrated a strong predictive performance (AUC: 0.910), confirming its effectiveness in ophthalmic imaging. These results highlight the potential of AI models as effective diagnostic tools in ocular oncology. However, further research is needed to improve model generalizability, enhance dataset diversity, and facilitate clinical integration. Full article
(This article belongs to the Special Issue Robotics, IoT and AI Technologies in Bioengineering)
Show Figures

Figure 1

Back to TopTop