Next Article in Journal
The Role of Insulin-like Growth Factor 2 (IGF-2) in Periodontal Regeneration: A Systematic Review
Previous Article in Journal
Comparison of Deep SMAS Lift and MACS in Facelift: A Meta-Analysis
error_outline You can access the new MDPI.com website here. Explore and share your feedback with us.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Segmentation of the Ulnar Nerve in Ultrasound Images

1
Chung-Ang University Industry-Academic Cooperation Foundation, Seoul 06974, Republic of Korea
2
Mellowing Factory Co., Ltd., Seoul 03726, Republic of Korea
3
Department of Dermatology, Korea University Guro Hospital, Seoul 08308, Republic of Korea
4
Department of Physical and Rehabilitation Medicine, Chung-Ang University Gwangmyeong Hospital, Gwangmyeong-si 14241, Republic of Korea
5
Department of Physical and Rehabilitation Medicine, Chung-Ang University College of Medicine, Seoul 06974, Republic of Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work as first authors.
These authors contributed equally to this work as corresponding authors.
Medicina 2026, 62(1), 113; https://doi.org/10.3390/medicina62010113
Submission received: 27 October 2025 / Revised: 7 December 2025 / Accepted: 9 December 2025 / Published: 5 January 2026

Abstract

Background and Objectives: We evaluate deep learning-based segmentation methods for detecting the ulnar nerve in ultrasound (US) images, leveraging the first-ever large US dataset of the ulnar nerve. We compare several widely used segmentation models, analyze their performance, and evaluate several common data augmentation techniques for the US. Materials and Methods: Our analysis is conducted on a large dataset of 4789 US images from 545 patients, with expert-annotated ground-truth segmentations of the ulnar nerve, and uses six segmentation models with several backbone architectures. Further, we analyze the statistical significance of five common data augmentation techniques on segmentation performance: flipping, rotation, shearing, contrast and brightness adjustments, and resizing. Results: In this study, the shear, rotate, and resize augmentations consistently improved segmentation performance across multiple runs, with p-values < 0.05 in a paired t-test relative to the no-augmentation baseline. Furthermore, we showed that newer architectures do not provide any metric improvements over traditional U-Net models, which achieved a Dice score of 0.88 and an IoU of 0.81. Conclusions: Through our systematic analysis of segmentation models and data augmentation strategies, we provide key insights into optimizing deep learning approaches for ulnar nerve segmentation and other US-based nerve segmentation tasks.

1. Introduction

High-resolution ultrasound (US) is vital in neuromuscular medicine for diagnosis, treatment, and prognostication. It identifies peripheral nerve structures, measures nerve parameters, and guides real-time interventions. The US is beneficial for diagnosing entrapment neuropathies, with guidelines indicating that the median nerve’s cross-sectional area (CSA) measurement at the wrist provides level A evidence for carpal tunnel syndrome (CTS) [1,2,3,4]. This has led to rapid expansion in its clinical use. Furthermore, the ulnar nerve, particularly at the elbow, is the second most common site of upper extremity entrapment neuropathy, following the median nerve at the wrist [2]. However, the efficacy of deep learning approaches for ulnar nerve segmentation in US remains understudied within the literature.
Peripheral nerves exhibit characteristic US features: appearing as linear hypoechoic areas separated by hyperechoic bands in longitudinal scans and as a honeycomb-like structure in axial scans [5,6,7]. However, accurate identification can be difficult due to anatomical variability, individual echotexture differences, and similarity to tendons, potentially leading to diagnostic errors and malpractice. Clinicians differentiate nerves using strategies such as assessing echotexture stability despite probe tilting (to account for anisotropy) and by tracing nerve pathways [6,8]. Still, these methods are time-consuming, labor-intensive, and require extensive training. Manual measurements further stress the need for automated solutions to boost efficiency and accuracy.
Deep learning-based segmentation models have become a crucial part of many image processing workflows and enhance computer-aided diagnostics (CAD) [9,10,11,12,13]. These models have proven useful in several US imaging applications ranging from breast cancer [14,15] and vessel segmentation [16,17] to peripheral nerve segmentation [18,19]. However, US segmentation remains challenging due to the often small size of target structures and the presence of noise and distortion artifacts. Obtaining high-quality images with expert-annotated segmentation labels is labor-intensive and time-consuming, usually limiting prior studies to small datasets. These dataset constraints introduce distribution bias, reducing the real-world generalizability.
Despite advances in deep learning for medical imaging, ultrasound-based peripheral nerve segmentation remains challenging. Small datasets have limited model generalizability and require restricted parameter sizes to prevent overfitting. Data augmentation for US imaging has also received insufficient attention. Techniques like shearing and resizing may improve model robustness, but their effects are not well understood. Ulnar nerve segmentation, in particular, is underrepresented in the literature, with few comprehensive evaluations of segmentation models or augmentation strategies. To address these gaps, our study compares and benchmarks deep learning segmentation models for ulnar nerve detection, using the largest expert-annotated dataset to date, comprising 4789 ultrasound images from 545 patients. We also assess the impact of various augmentation methods, including rotation, shearing, resizing, cropping, and adjustments to contrast and brightness. This work provides practical guidance for applying deep learning to peripheral nerve segmentation, with a focus on the understudied ulnar nerve.

2. Materials and Methods

We compare six common segmentation architectures, each using three different backbone architectures as feature encoders. Further, we explore the impact of common data augmentation methods applied to US images on the segmentation of the ulnar nerve.

2.1. Deep Learning Architectures

The six segmentation network architectures studied for ulnar nerve segmentation are U-Net [11], LinkNet [20], U-Net++ [21], DeepLabV3+ [22], Pyramid Attention Network [13], and Segformer [12]. U-Net++ improves upon U-Net by embedding nested, dense skip pathways, thereby reducing the semantic gap between the encoder and decoder layers. While this enhances multi-scale context capture, it significantly increases complexity and parameter count. LinkNet, in contrast, directly connects encoder inputs to corresponding decoder outputs, preserving high-resolution spatial details during upsampling while minimizing computational overhead. The Pyramid Attention Network (PAN) introduces a multi-scale attention module that aggregates features across various scales to emphasize informative regions and suppress irrelevant ones. Computational efficiency is crucial in ultrasound segmentation, as real-time assistance is essential for clinical applications [23,24]. DeepLabV3+ employs an encoder–decoder structure with atrous separable convolutions in both the Atrous Spatial Pyramid Pooling (ASPP) module and decoder, refining segmentation boundaries while maintaining computational efficiency. Segformer further optimizes the balance between performance and efficiency by leveraging transformer-based architectures, decoupling encoding and decoding to achieve robust global context modeling with minimal computational cost.
In addition, we employ various encoder architectures, each with distinct computational trade-offs, including ResNet [25], EfficientNet [26], and Mix Transformer (MiT) encoders from the Segformer framework. Table 1 summarizes the computational cost in MACs (multiply–accumulate) for a 512 × 512 image and the parameter counts across the different combinations of backbones and architectures considered in this study. Although our dataset is relatively large compared to prior studies, it remains small compared to natural image datasets. To mitigate overfitting and reduce compute costs, we use the more minor backbone variants.

2.2. Data Augmentations

Various data augmentation schemes have been proposed for use with US images including random resizing, cropping, rotation, shearing and more. We selected five data augmentation techniques for analysis: random rotation, resizing, contrast and brightness adjustments, flipping, and shearing. Random cropping was always applied during training to standardize image sizes for batching and encourage learning from different regions, enhancing focus on key anatomical features. Random resizing simulates variability in object size and probe distance, improving generalization across imaging conditions. Contrast and brightness adjustments mimic natural fluctuations in ultrasound quality due to machine settings and patient factors. Flipping, rotation, and shearing introduce orientation and distortion variations, reflecting probe movement and operator variability. Together, these augmentations enhance data diversity, reduce overfitting, and improve model generalizability for ultrasound image analysis.

2.3. Dataset Description

A computerized search of electronic medical records and picture archiving and communication systems was performed to identify patients who underwent US examinations of the upper extremities from March 2019 to December 2024. Among the 983 initially identified subjects, 545 subjects over 20 years of age with normal findings in ulnar nerve conduction studies were included. From the 545 patients, a total of 4789 US images are included as a part of this study. Eighty percent of the patients are randomly selected for training and validation (a total of 436 patients) and the remaining patients are withheld for the test set (a total of 109 patients).
An expert clinician with over 15 years of clinical experience in neuromuscular ultrasound conducted all examinations and annotations using the miniSONO (Alpinion Medical Systems, Seoul, Republic of Korea). Subjects were placed in the supine position with elbows flexed to 90 degrees, forearms fully supinated, and fingers maintained in slight flexion. To accurately identify the ulnar nerve, the clinician performed dynamic ultrasound scanning by slowly moving the transducer proximally and distally along its course. Ultrasound images of the ulnar nerve were acquired at five standardized locations per extremity: specifically, five images were captured at 1 cm intervals, spanning from 2 cm proximal to 2 cm distal to the medial epicondyle at the elbow. The examiner endeavored to capture the clearest and most representative images of the nerve by orienting the ultrasound beam perpendicular to the course of the nerve. Nerve boundaries were defined using the inner epineurial contour, excluding the outer hyperechoic rim. Ambiguous regions—where the perineurium blended with adjacent hyperechoic tissue—were resolved by reviewing adjacent frames to maintain consistent labeling.
Images with significant motion artifacts, acoustic shadowing, signal dropout, or excessive deformation caused by probe pressure were excluded from the dataset. All included images underwent expert quality review to ensure sufficient visualization of the ulnar nerve before annotation.

2.4. Outcome Variables

In this study, the performance of the deep learning architectures was verified using the following indicators: recall, precision, F1 score, Dice score, and intersection over union (IoU).
O ( D i c e ) = 2 X Y X + Y = 2 T P 2 T P + F P + F N
( I o U ) = X Y X Y = T P T P + F P + F N
where X and Y represent the predicted and ground-truth segmentation masks, respectively. Further, TP is the true positive rate, FP is the false positive rate, and FN is the false negative rate.

2.5. Implementation Details

For the experiments in this study, we used the PyTorch (2.5.1) machine learning framework, along with the segmentation-models-pytorch library, to ensure reproducible implementations of each segmentation model [27]. We adopt the Dice loss function, denoted by L, to find the boundaries of peripheral nerves in the US images. The Dice loss function L is defined as follows:
L = 1 − (Dice)
(Dice) is defined in Equation (1). The Dice loss can alleviate blurred boundary problems often observed with cross-entropy loss and is also effective at handling class imbalance, i.e., the nerves occupy only a small portion of the US image.
All models were trained using the Adam optimizer with a learning rate of 1 × 10−4 and a weight decay of 1 × 10−6. Training was conducted for 150 epochs, with the learning rate reduced to 1 × 10−5 after 100 epochs. During training, images were either padded or randomly cropped to 512 × 512 before applying data augmentations, with resizing applied first. A batch size of 32 was used for all experiments, and all models were trained on an NVIDIA RTX A6000 GPU.
To avoid data leakage, all dataset splits were performed on a patient-wise basis, such that images from the same subject were not shared across the training, validation, and test sets. Data augmentations were applied exclusively during training, while the validation set was used to monitor model convergence and guide learning-rate scheduling.
Data augmentations were applied in the following order: (1) random resize and cropping, (2) random contrast and brightness adjustments, (3) random rotations, (4) random shearing, and (5) random horizontal flipping. Resizes were chosen uniformly between a 25% decrease and a 25% increase in the image’s height, preserving the aspect ratio. Brightness and contrast adjustments were selected uniformly within ±50% of their original values. Rotations and shearing transformations were applied uniformly between −15° and 15°, while horizontal flips were performed with a 50% probability.

3. Results

3.1. Quantitative

To evaluate the impact of different data augmentation strategies, we trained a U-Net model with a ResNet-34 backbone using every possible combination of augmentations. Figure 1 presents the test-set Dice scores for all single augmentation strategies, along with no augmentation and all augmentations. Our results indicate that all augmentations, combined, provided a statistically significant improvement in the Dice score on our test set. Further, in Figure 2, we see that as the number of data augmentations increases, the Dice score also increases. For our augmentation analysis, we employ Bonferroni correction to our paired t-tests. Bonferroni correction is a statistical adjustment that accounts for the increased risk of false positives when performing multiple comparisons simultaneously. It works by dividing the significance threshold by the number of tests being conducted. For both Figure 1 and Figure 2 the overall pattern of statistical significance did not change after applying this correction. The results in Figure 1 show the impact of individual augmentations, with the most significant being the rotation augmentation, while Figure 2 shows that the number of combinations employed is also statistically significant in improving overall segmentation performance.
Table 2 shows that despite extensive testing, the U-Net with a ResNet43 backbone achieved the highest Dice score and IoU. Multi-scale models like PAN and DeepLabV3+ likely provided no advantage because random resizing already enforced scale invariance, and the ulnar nerve’s boundaries are not complex enough to benefit from such processing. Additionally, newer state-of-the-art models like Segformer did not show a noticeable improvement over U-Net, suggesting that appropriate model capacity and effective data augmentation strategies are sufficient for achieving state-of-the-art results in ulnar nerve segmentation. However, DeepLabV3+ with EfficientNet backbones performed comparably to U-Net at a lower computational cost, making them a viable alternative for compute-constrained applications despite a slight performance trade-off.

3.2. Qualitative

For our qualitative analysis, we use predictions from the best-performing model identified in our architecture search: a U-Net with a ResNet34 backbone. Figure 3 presents general segmentation results for ulnar nerve images captured by the miniSONO device. Ground-truth regions are outlined in green, while predicted areas are shown in red. Failure cases fall into three main categories. First, as shown in Figure 4a, some predictions only partially overlap with the ground truth. This may occur when the model misinterprets an internal nerve structure, possibly mistaking part of the perineurium (the sheath surrounding a nerve fascicle) for the outer nerve boundary. As a result, the predicted segmentation may erroneously pass through the middle of the nerve.
Second, as shown in Figure 4b, some segmentations deviate from the expected elliptical shape. In clinical practice, the ulnar nerve typically appears as an ellipse, and uneven margins should still approximate this form. However, in some cases, the model over-segments, including adjacent structures such as small vessels.
Third, Figure 4c illustrates a rare but significant failure: a total miss where the predicted segmentation does not overlap with the ground truth. In such cases, both precision and recall drop to zero. Although infrequent, only accounting for 0.6% of all results on the test set, these failures highlight potential limitations that could impact clinical applications.

4. Discussion

This study evaluated deep learning-based segmentation methods for ulnar nerve detection in US images, addressing key challenges in neuromuscular medicine. Using a robust dataset of 4789 US images from 545 patients, we compared various segmentation architectures and data augmentation techniques in Section 2.1 and Section 2.2, respectively. Through these experiments, we show that a sufficient dataset size, data augmentations, and traditional segmentation architectures, such as U-Net, provide reasonable performance in the segmentation of the ulnar nerve.
Since its development, U-Net [11] has been widely adopted for ultrasound (US) image segmentation in various applications, including the left ventricle [28], endometrium [29], breast [30], fetal skull [31], cerebral ventricles [32], and prostate [33,34]. U-Net has also been extensively applied to US nerve segmentation, particularly in regional anesthesia. Baby and Jereesh [35] used U-Net for brachial plexus nerve detection, while Kakade and Dumbali [36] combined Linear Gabor Binary Patterns for preprocessing with U-Net and Principal Component Analysis (PCA) for postprocessing. Other studies have introduced architectural modifications to enhance the U-Net’s performance. Wang et al. [37] introduced ResUNet, which incorporates residual modules, dense atrous convolutions, and multi-kernel pooling. Weng et al. [38] applied Neural Architecture Search (NAS) to optimize the U-Net’s backbone for US nerve segmentation. Horng et al. [19] developed DeepNerve, which integrates an LSTM into the U-Net’s bottom layer, using a MaskTrack scheme to improve median nerve tracking in sequential US video frames. Beyond these adaptations, various U-Net variants have been proposed for broader biomedical segmentation tasks. U-Net++ [21] incorporates dense blocks to enhance skip connections. Attention U-Net [39] introduces attention gates to focus on salient features while suppressing irrelevant regions. Scale-attentional U-Net [40] adds gating to cascaded convolutions without non-linearities or pooling at each block, allowing for dynamic adjustment of the receptive field throughout the network. Despite these advances, few studies have systematically compared segmentation models for ulnar nerve detection, particularly with respect to computational efficiency and data augmentation.
Rather than proposing a specialized architecture, this study systematically analyzed common data augmentations and segmentation networks on a large, expert-labeled US dataset. Data augmentation provided modest but consistent improvements. Random rotation augmentation yielded the most statistically significant performance improvement over the no-data-augmentation baseline (in a paired t-test), likely because it simulates common variations in probe orientation during ultrasound examinations. Random resizing and random shearing emerged as the second-most-effective augmentations, potentially aiding generalization across patients—a factor independent of operator technique or equipment. Contrast and brightness adjustments showed no significant effect in our study. Since all data were acquired using a single device, the miniSONO, the limited impact of this augmentation is unsurprising. Flipping augmentation had a larger effect than contrast and brightness augmentation. Still, the difference was not statistically significant, possibly due to factors that were consistent across operator techniques and device characteristics. As expected, augmentation effects were moderate given the large dataset, consistent with prior research showing diminishing returns with increased data [41]. Despite testing modern architectures, the original U-Net with a ResNet34 backbone achieved the highest performance (Dice: 0.8808, IoU: 0.8063), suggesting that traditional models remain competitive when well implemented. Shown in Figure 4 are some representative cases of poor predictions, including partial overlaps, misinterpretation of internal nerve structures, non-elliptical segmentations, and rare total misses. While our dataset was larger than most prior US studies, all images were captured on a single device (miniSONO). Future work should assess performance across multiple US machines for broader generalizability and explore shape priors or anatomical constraints to mitigate failure modes.
This study has significant merit in applying deep learning techniques to the ulnar nerve, one of the most frequently affected peripheral nerves in the upper extremity, for the treatment of neuropathy. We conducted the first comprehensive research using a large volume of high-quality peripheral nerve ultrasound data, demonstrating the successful application of deep learning to the ulnar nerve and achieving high accuracy in nerve identification while complementing previous work on the median nerve. The encouraging results suggest promising implications for developing sophisticated software capable of peripheral nerve identification. We expect that establishing a robust methodological framework will lay the critical groundwork for advanced neurological imaging technologies. Future research should expand our approach to additional clinically significant nerves in both the upper and lower extremities and systematically develop algorithms to distinguish between normal and abnormal neural structures, which will be essential for translating these findings into clinical diagnostic tools.
Most previous deep learning studies on peripheral nerve ultrasound segmentation have focused on the median nerve, with reported Dice coefficients generally ranging from 0.80 to 0.90, depending on the model architecture and dataset characteristics [18,19,35,36,40]. These studies, which include approaches such as U-Net, ResUNet, NAS-optimized U-Net, and DeepNerve, collectively highlight that the median nerve has been the predominant focus of segmentation research. In our study, the best-performing model achieved a Dice score of 0.8808 and an IoU of 0.8063, placing its performance within the upper range of these prior median nerve studies. Notably, despite the ulnar nerve exhibiting greater anatomical variability and a more heterogeneous echotexture than the median nerve, our model achieved a comparable level of accuracy. This suggests that, given sufficient dataset size and effective augmentation strategies, ulnar nerve segmentation can achieve performance on par with that reported for the median nerve. Such contextual comparison underscores the contribution of our work in addressing the relative scarcity of ulnar nerve segmentation research. It supports the potential development of more generalized multi-nerve segmentation frameworks in future studies.
The accuracy achieved by our model (Dice 0.8808) is comparable to, and in the upper range of, the inter-observer variability reported in neuromuscular ultrasound studies. Prior work has shown that manual delineation of nerve boundaries by different clinicians typically yields Dice scores of approximately 0.80–0.88, indicating that a degree of variability is inherent to human interpretation. In this context, the segmentation performance of our model reaches the level generally considered sufficient to support clinically meaningful tasks, such as automated CSA estimation or detection of focal nerve enlargement. Nevertheless, segmentation accuracy alone does not ensure diagnostic reliability, and future studies should evaluate the clinical performance of CSA measurements derived from automated segmentation.
Despite the promising performance of our segmentation model, several intrinsic characteristics of peripheral nerve US must be acknowledged when interpreting the results. In clinical practice, distinguishing the precise number and arrangement of fascicles on ultrasound is often tricky, as the echotexture of peripheral nerves varies considerably across individuals and anatomical regions. The boundaries between the epineurium and perineurium are frequently indistinct, and the transition from neural tissue to surrounding connective tissue is often gradual rather than sharply demarcated. These ambiguities can make it difficult to delineate the nerve’s outer contour accurately. As illustrated in the first example of Figure 4, the model occasionally misinterprets portions of the perineurium or adjacent hyperechoic structures as part of the nerve margin, resulting in segmentation that partially bisects the nerve or deviates from its expected elliptical morphology. Such cases highlight the fundamental limitations of ultrasound imaging in resolving delicate intraneural structures.
Recent multimodal studies integrating histology with high-resolution ultrasound have shown that many structures appearing as single fascicles on ultrasound actually correspond to histologically identified clusters of multiple fascicles. That epineurial or perineurial boundaries may shift or blend depending on scanning angle, probe pressure, and tissue composition [42]. These findings suggest that leveraging histological ground truth improves the interpretability and robustness of deep-learning models applied to nerve imaging. Incorporating multimodal supervision—such as histologic cross-sections, MR microscopy–derived fascicular maps, or other microstructural references—could enable future models to distinguish actual fascicular boundaries better, avoid erroneous perineurial segmentation, and more closely approximate the microanatomy of peripheral nerves.
Accordingly, future work should prioritize multimodal fusion approaches that bridge histological insights with ultrasound-based AI segmentation. Such studies have the potential not only to overcome the intrinsic resolution limits of ultrasound but also to establish biologically meaningful constraints that guide model predictions. Ultimately, integrating histologic validation into deep-learning development may yield segmentation systems capable of supporting fascicle-level analysis and more accurate assessment of neuropathic changes, thereby expanding the clinical relevance of automated peripheral nerve imaging.
Although our model demonstrated high segmentation accuracy, several considerations remain before it can be deployed in real-world clinical environments. Ultrasound imaging varies substantially across devices, transducer types, and operator techniques, which may affect model performance outside of the controlled setting of this study, where all images were acquired using a single device and a single examiner. Additionally, real-time inference capability and integration into existing clinical workflows must be evaluated for practical use. Further validation in diverse patient populations, particularly in those with neuropathies that alter nerve morphology, is essential to determine whether the model can reliably support clinical decision-making. Establishing such evidence will be a critical step toward translating automated ulnar nerve segmentation into routine neuromuscular ultrasound practice.
Despite the strengths of this study, several limitations must be acknowledged. First, the dataset was acquired exclusively using a single ultrasound device (miniSONO) and by a single expert examiner. Although this consistency reduces variability and improves annotation quality, it may limit the model’s generalizability to other devices or operators. Because a single examiner generated all annotations, the ground truth does not capture inter-observer variability, which is inherent to neuromuscular ultrasound interpretation. Future studies should incorporate multiple annotators or consensus labeling to establish more robust and generalizable reference standards. Second, the dataset included only patients with normal ulnar nerve conduction studies. Hence, the model’s ability to detect or segment pathological nerves, such as in cases of ulnar neuropathy or entrapment syndromes, remains untested. Finally, the study did not assess clinical endpoints, such as the accuracy of cross-sectional area (CSA) measurements or diagnostic utility, which are critical for real-world deployment in neuromuscular practice. Although Dice performance was high, segmentation accuracy does not necessarily translate to accurate CSA estimation, and future validation studies are required to assess the clinical reliability of automated CSA measurements.

5. Conclusions

Our comprehensive evaluation of deep learning approaches for ulnar nerve segmentation provides valuable insights for optimizing automated US-based nerve imaging. The findings suggest that traditional architectures with appropriate data augmentation can achieve robust performance. As these techniques mature, they have the potential to enhance clinical workflow efficiency, improve diagnostic accuracy, and ultimately benefit patient care in neuromuscular medicine.

Author Contributions

Conceptualization, K.E.K., Y.J.N., J.L. and B.S.K.; methodology, M.B.W.; software, M.B.W.; validation, J.L. and B.S.K.; formal analysis, M.B.W.; investigation, M.B.W.; resources, J.L. and B.S.K.; data curation, Y.J.N., J.L. and B.S.K.; writing—original draft preparation, M.B.W., K.E.K., J.L. and B.S.K.; writing—review and editing, M.B.W., K.E.K., Y.J.N., J.L. and B.S.K.; visualization, M.B.W. and K.E.K.; supervision, K.E.K., Y.J.N., J.L. and B.S.K.; project administration, K.E.K., J.L. and B.S.K.; funding acquisition, J.L. and B.S.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Chung-Ang University Research Grants in 2022. This works was supported by Korea University Guro Hospital (KOREA RESEARCH-DRIVEN HOSPITAL) and grant funded by Korea University Medicine (No. O2516191).

Institutional Review Board Statement

This study was approved by the Institutional Review Boards of Korea University Guro Hospital (code: 2020GR0389, date: 19 August 2020).

Informed Consent Statement

Patient consent was waived because the study used de-identified ultrasound images, and both IRBs determined that obtaining individual consent was not required.

Data Availability Statement

The datasets generated and/or analyzed during the current study are not publicly available due to a lack of IRB approval for public release, but are available from the corresponding author on reasonable request.

Conflicts of Interest

Matthew Webster and Joonnyong Lee were employed by Mellowing Factory Co. Ltd., a company specializing in biosignal based sleep research. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Beekman, R.; Visser, L.H. Sonography in the diagnosis of carpal tunnel syndrome: A critical review of the literature. Muscle Nerve 2003, 27, 26–33. [Google Scholar] [CrossRef]
  2. Cartwright, M.S.; Walker, F.O. Neuromuscular ultrasound in common entrapment neuropathies. Muscle Nerve 2013, 48, 696–704. [Google Scholar] [CrossRef]
  3. Walker, F.O.; Cartwright, M.S.; Alter, K.E.; Visser, L.H.; Hobson-Webb, L.D.; Padua, L.; Strakowski, J.A.; Preston, D.C.; Boon, A.J.; Axer, H.; et al. Indications for neuromuscular ultrasound: Expert opinion and review of the literature. Clin. Neurophysiol. 2018, 129, 2658–2679. [Google Scholar] [CrossRef]
  4. Cartwright, M.S.; Hobson-Webb, L.D.; Boon, A.J.; Alter, K.E.; Hunt, C.H.; Flores, V.H.; Werner, R.A.; Shook, S.J.; Thomas, T.D.; Primack, S.J.; et al. Evidence-based guideline: Neuromuscular ultrasound for the diagnosis of carpal tunnel syndrome. Muscle Nerve 2012, 46, 287–293. [Google Scholar] [CrossRef] [PubMed]
  5. Bianchi, S. Ultrasound of the peripheral nerves. Jt. Bone Spine 2008, 75, 643–649. [Google Scholar] [CrossRef]
  6. Martinoli, C.; Bianchi, S.; Dahmane, M.; Pugliese, F.; Bianchi-Zamorani, M.P.; Valle, M. Ultrasound of tendons and nerves. Eur. Radiol. 2002, 12, 44–55. [Google Scholar] [CrossRef]
  7. Silvestri, E.; Martinoli, C.; Derchi, L.E.; Bertolotto, M.; Chiaramondia, M.; Rosenberg, I. Echotexture of peripheral nerves: Correlation between US and histologic findings and criteria to differentiate tendons. Radiology 1995, 197, 291–296. [Google Scholar] [CrossRef] [PubMed]
  8. Kowalska, B.; Sudoł-Szopińska, I. Normal and sonographic anatomy of selected peripheral nerves. Part I: Sonohistology and general principles of examination, following the example of the median nerve. J. Ultrason. 2012, 12, 120–130. [Google Scholar] [CrossRef] [PubMed]
  9. Aljabri, M.; AlGhamdi, M. A review on the use of deep learning for medical image segmentation. Neurocomputing 2022, 506, 311–335. [Google Scholar] [CrossRef]
  10. Haque, I.R.I.; Neubert, J. Deep learning approaches to biomedical image segmentation. Inform. Med. Unlocked 2020, 18, 100297. [Google Scholar] [CrossRef]
  11. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241, ISBN 978-3-319-24574-4. [Google Scholar]
  12. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. Segformer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  13. Li, H.; Xiong, P.; An, J.; Wang, L. Pyramid attention network for semantic segmentation. In Proceedings of the British Machine Vision Conference 2018 (BMVC 2018), Newcastle, UK, 3–6 September 2018; BMVA Press: Durham, UK, 2018; Volume 285. [Google Scholar]
  14. Al-Dhabyani, W.; Gomaa, M.; Khaled, H.; Fahmy, A. Dataset of breast ultrasound images. Data Brief. 2020, 28, 104863. [Google Scholar] [CrossRef] [PubMed]
  15. Xu, Y.; Wang, Y.; Yuan, J.; Cheng, Q.; Wang, X.; Carson, P.L. Medical breast ultrasound image segmentation by machine learning. Ultrasonics 2019, 91, 1–9. [Google Scholar] [CrossRef]
  16. Jiang, B.; Chen, A.; Bharat, S.; Zheng, M. Automatic ultrasound vessel segmentation with deep spatiotemporal context learning. In Proceedings of the International Workshop on Advances in Simplifying Medical Ultrasound, Strasbourg, France, 27 September 2021; Springer International Publishing: Cham, Switzerland, 2021; pp. 3–13, ISBN 978-3-030-87583-1. [Google Scholar]
  17. Xie, M.; Li, Y.; Xue, Y.; Huntress, L.; Beckerman, W.; Rahimi, S.; Ady, J.; Roshan, U. Vessel lumen segmentation in carotid artery ultrasounds with the U-net convolutional neural network. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Republic of Korea, 16–19 December 2020; pp. 2680–2684. [Google Scholar] [CrossRef]
  18. Wang, J.C.; Shu, Y.C.; Lin, C.Y.; Wu, W.T.; Chen, L.R.; Lo, Y.C.; Chiu, H.C.; Ozçakar, L.; Chang, K.V. Application of deep learning algorithms in automatic sonographic localization and segmentation of the median nerve: A systematic review and meta-analysis. Artif. Intell. Med. 2023, 137, 102496. [Google Scholar] [CrossRef]
  19. Horng, M.H.; Yang, C.W.; Sun, Y.N.; Yang, T.H. Deepnerve: A new convolutional neural network for the localization and segmentation of the median nerve in ultrasound image sequences. Ultrasound Med. Biol. 2020, 46, 2439–2452. [Google Scholar] [CrossRef]
  20. Chaurasia, A.; Culurciello, E. Linknet: Exploiting encoder representations for efficient semantic segmentation. In Proceedings of the 2017 IEEE Visual Communications and Image Processing (VCIP), St. Petersburg, FL, USA, 10–13 December 2017; pp. 1–4. [Google Scholar] [CrossRef]
  21. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: A nested U-net architecture for medical image segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; Springer: Cham, Switzerland, 2018; pp. 3–11, ISBN 978-3-030-00889-5. [Google Scholar]
  22. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar] [CrossRef]
  23. Dai, Q.; Lin, J.; Li, W.; Wang, L. A real-time network for fast breast lesion detection in ultrasound videos. In Proceedings of the Chinese Conference on Pattern Recognition and Computer Vision, Urumqi, China, 18–20 October 2024; Springer: Singapore, 2024; pp. 40–50, ISBN 978-981-99-8558-6. [Google Scholar]
  24. Hu, Z.; Nasute Fauerbach, P.V.; Yeung, C.; Ungi, T.; Rudan, J.; Engel, C.J.; Mousavi, P.; Fichtinger, G.; Jabs, D. Real-time automatic tumor segmentation for ultrasound-guided breast-conserving surgery navigation. Int. J. Comput. Assist. Radiol. Surg. 2022, 17, 1663–1672. [Google Scholar] [CrossRef] [PubMed]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  26. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 10–15 June 2019; PMLR: Cambridge, MA, USA, 2019; pp. 6105–6114. [Google Scholar]
  27. Iakubovskii, P. Segmentation Models PyTorch [Internet]. 2019. Available online: https://github.com/qubvel/segmentation_models.pytorch (accessed on 11 April 2025).
  28. Smistad, E.; Østvik, A.; Haugen, B.O.; Lvstakken, L. 2D left ventricle segmentation using deep learning. In Proceedings of the 2017 IEEE International Ultrasonics Symposium, Washington, DC, USA, 6–9 September 2017; pp. 1–4. [Google Scholar] [CrossRef]
  29. Singhal, N.; Mukherjee, S.; Perrey, C. Automated assessment of endometrium from transvaginal ultrasound using deep learned snake. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI), Melbourne, VIC, Australia, 18–21 April 2017; pp. 283–286. [Google Scholar] [CrossRef]
  30. Almajalid, R.; Shan, J.; Du, Y.; Zhang, M. Development of a deep-learning-based method for breast ultrasound image segmentation. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Application (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 1103–1108. [Google Scholar] [CrossRef]
  31. Cerrolaza, J.J.; Sinclair, M.; Li, Y.; Gomez, A.; Ferrante, E.; Matthew, J.; Gupta, C.; Knight, C.L.; Rueckert, D. Deep learning with ultrasound physics for fetal skull segmentation. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI), Washington, DC, USA, 4–7 April 2018; pp. 564–567. [Google Scholar] [CrossRef]
  32. Martin, M.; Sciolla, B.; Sdika, M.; Wang, X.; Quetin, P.; Delachartre, P. Automatic segmentation of the cerebral ventricle in neonates using deep learning with 3D reconstructed freehand ultrasound imaging. In Proceedings of the 2018 IEEE International Ultrasonics Symposium, Kobe, Japan, 22–25 October 2018; pp. 1–4. [Google Scholar] [CrossRef]
  33. van Sloun, R.J.G.; Wildeboer, R.R.; Mannaerts, C.K.; Postema, A.W.; Gayet, M.; Beerlage, H.P.; Salomon, G.; Wijkstra, H.; Mischi, M. Deep learning for real-time, automatic, and scanner-adapted prostate (zone) segmentation of transrectal ultrasound, for example, magnetic resonance imaging-transrectal ultrasound fusion prostate biopsy. Eur. Urol. Focus 2021, 7, 78–85. [Google Scholar] [CrossRef]
  34. Orlando, N.; Gillies, D.J.; Gyacskov, I.; Romagnoli, C.; D’Souza, D.; Fenster, A. Automatic prostate segmentation using deep learning on clinically diverse 3D transrectal ultrasound images. Med. Phys. 2020, 47, 2413–2426. [Google Scholar] [CrossRef]
  35. Baby, M.; Jereesh, A.S. Automatic nerve segmentation of ultrasound images. In Proceedings of the 2017 International Conference of Electronics, Communication and Aerospace Technology (ICECA), Coimbatore, India, 20–22 April 2017; pp. 107–112. [Google Scholar] [CrossRef]
  36. Kakade, A.; Dumbali, J. Identification of nerve in ultrasound images using U-net architecture. In Proceedings of the 2018 International Conference on Communication Information and Computing Technology (ICCICT), Mumbai, India, 2–3 February 2018; pp. 1–6. [Google Scholar] [CrossRef]
  37. Wang, R.; Shen, H.; Zhou, M. Ultrasound nerve segmentation of brachial plexus based on optimized ResU-Net. In Proceedings of the 2019 IEEE International Conference on Imaging Systems and Techniques (IST), Abu Dhabi, United Arab Emirates, 9–10 December 2019; pp. 1–6. [Google Scholar] [CrossRef]
  38. Weng, Y.; Zhou, T.; Li, Y.; Qiu, X. Nas-unet: Neural architecture search for medical image segmentation. IEEE Access 2019, 7, 44247–44257. [Google Scholar] [CrossRef]
  39. Oktay, O.; Schlemper, J.; Le Folgoc, L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  40. Kim, B.S.; Yu, M.; Kim, S.; Yoon, J.S.; Baek, S. Scale-attentional U-Net for the segmentation of the median nerve in ultrasound images. Ultrasonography 2022, 41, 706–717. [Google Scholar] [CrossRef] [PubMed]
  41. Alomar, K.; Aysel, H.I.; Cai, X. Data augmentation in classification and segmentation: A survey and new strategies. J. Imaging 2023, 9, 46. [Google Scholar] [CrossRef] [PubMed]
  42. Pušnik, L.; Radochová, B.; Janáček, J.; Saudek, F.; Serša, I.; Cvetko, E.; Umek, N.; Snoj, Ž. Fascicle differentiation of upper extremity nerves on high-resolution ultrasound with multimodal microscopic verification. Sci. Rep. 2025, 15, 557. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Comparison of the test Dice score when no data augmentations are applied, individual data augmentations are used, and all data augmentations are applied. Results are produced from 6 training runs for each scenario, totaling 42 trained U-Net models with a ResNet34 backbone.
Figure 1. Comparison of the test Dice score when no data augmentations are applied, individual data augmentations are used, and all data augmentations are applied. Results are produced from 6 training runs for each scenario, totaling 42 trained U-Net models with a ResNet34 backbone.
Medicina 62 00113 g001
Figure 2. Average test Dice score by the number of data augmentations used. Results are averaged across three training sweeps for all 32 data augmentation combinations, yielding a total of 96 trained models, with the U-Net architecture using a ResNet34 backbone.
Figure 2. Average test Dice score by the number of data augmentations used. Results are averaged across three training sweeps for all 32 data augmentation combinations, yielding a total of 96 trained models, with the U-Net architecture using a ResNet34 backbone.
Medicina 62 00113 g002
Figure 3. (ac) General segmentation predictions of the ulnar nerve in US captured by miniSONO. Predictions are shown in red, and ground-truth labels are shown in green. Here the green segmentation represents the ground truth and the red segmentation represents the prediction made by the model.
Figure 3. (ac) General segmentation predictions of the ulnar nerve in US captured by miniSONO. Predictions are shown in red, and ground-truth labels are shown in green. Here the green segmentation represents the ground truth and the red segmentation represents the prediction made by the model.
Medicina 62 00113 g003
Figure 4. (ac) Examples of segmentation predictions where the nerve was not identified well or at all. Here the green segmentation represents the ground truth and the red segmentation represents the prediction made by the model. Here the green segmentation represents the ground truth and the red segmentation represents the prediction made by the model.
Figure 4. (ac) Examples of segmentation predictions where the nerve was not identified well or at all. Here the green segmentation represents the ground truth and the red segmentation represents the prediction made by the model. Here the green segmentation represents the ground truth and the red segmentation represents the prediction made by the model.
Medicina 62 00113 g004
Table 1. The computational complexity of each model tested in this study for an input size of 512 × 512.
Table 1. The computational complexity of each model tested in this study for an input size of 512 × 512.
ArchitectureResNet18ResNet34MiT-b0MiT-b1Efficientnet-b0Efficientnet-b1
U-Net21 G, 14 M31 G, 24 M10 G, 2 M10 G, 2 M
U-Net++64 G, 16 M73 G, 26 M20 G, 3 M20 G, 3 M
DeepLabV3+18 G, 12 M31 G, 22 M2 G, 939 K2 G, 959 K
LinkNet12 G, 12 M21 G, 22 M823 M, 199 K878 M, 219 K
PAN16 G, 11 M29 G, 21 M392 M, 131 K467 M, 151 K
Segformer7 G, 4 M13 G, 14 M
MACs given in giga (G) and mega (M), and parameter size given in millions (M) and thousands (K).
Table 2. Comparison of segmentation performance across the different architectures considered in this study. Bold text indicates the best value for a given metric.
Table 2. Comparison of segmentation performance across the different architectures considered in this study. Bold text indicates the best value for a given metric.
ArchitectureBackboneRecallPrecisionF1DiceIoU
U-NetResNet180.84470.8970.87010.86790.7917
U-NetResNet340.8690.89190.88030.88080.8063
U-NetEfficientNet-b00.86090.88460.87250.86980.7951
U-NetEfficientNet-b10.84660.90210.87350.86780.7898
U-Net++ResNet180.8460.90030.87230.86650.7925
U-Net++ResNet340.86090.89120.87580.87460.8015
U-Net++EfficientNet-b00.8420.89480.86760.8680.7913
U-Net++EfficientNet-b10.83220.90020.86480.85590.7796
DeepLabV3+ResNet180.84820.89460.87080.86590.7889
DeepLabV3+ResNet340.86190.88820.87490.87370.7987
DeepLabV3+EfficientNet-b00.85510.87730.8660.8630.7804
DeepLabV3+EfficientNet-b10.84850.88640.8670.86450.7834
LinknetResNet180.85170.89260.87170.86990.792
LinknetResNet340.84970.88530.86720.8680.7913
LinknetEfficientNet-b00.84370.87760.86030.85980.7772
LinknetEfficientNet-b10.8410.87350.85690.85810.7767
PANResNet180.81630.91120.86120.84830.7658
PANResNet340.85490.89090.87250.87270.7936
PANEfficientNet-b00.80570.90610.85290.84440.7545
PANEfficientNet-b10.83390.8930.86250.85650.7722
SegformerMIT-b00.85940.85150.85540.85780.7754
SegformerMIT-b10.86590.87170.86880.86850.7906
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Webster, M.B.; Kim, K.E.; Na, Y.J.; Lee, J.; Kim, B.S. Deep Learning-Based Segmentation of the Ulnar Nerve in Ultrasound Images. Medicina 2026, 62, 113. https://doi.org/10.3390/medicina62010113

AMA Style

Webster MB, Kim KE, Na YJ, Lee J, Kim BS. Deep Learning-Based Segmentation of the Ulnar Nerve in Ultrasound Images. Medicina. 2026; 62(1):113. https://doi.org/10.3390/medicina62010113

Chicago/Turabian Style

Webster, Matthew Bailey, Ko Eun Kim, Yong Jae Na, Joonnyong Lee, and Beom Suk Kim. 2026. "Deep Learning-Based Segmentation of the Ulnar Nerve in Ultrasound Images" Medicina 62, no. 1: 113. https://doi.org/10.3390/medicina62010113

APA Style

Webster, M. B., Kim, K. E., Na, Y. J., Lee, J., & Kim, B. S. (2026). Deep Learning-Based Segmentation of the Ulnar Nerve in Ultrasound Images. Medicina, 62(1), 113. https://doi.org/10.3390/medicina62010113

Article Metrics

Back to TopTop