Next Article in Journal
Asymmetric Dimethylarginine Vibrational Spectroscopy Spectra and Density Functional Theory Model
Previous Article in Journal
Dual-Electrode Glass Ribbons Intended for Use in Microplasma-Based Sensors
Previous Article in Special Issue
Towards AI-Based Strep Throat Detection and Interpretation for Remote Australian Indigenous Communities
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhanced Deep Neural Network for Prostate Segmentation in Micro-Ultrasound Images

by
Ahmed AL-Qurri
1,
Asem Thaher
2 and
Mohamed Khaled Almekkawy
1,*
1
The School of Electrical Engineering and Computer Science, Pennsylvania State University, University Park, PA 16802, USA
2
Independent Researcher, State College, PA 16803, USA
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(22), 6815; https://doi.org/10.3390/s25226815
Submission received: 8 September 2025 / Revised: 29 October 2025 / Accepted: 4 November 2025 / Published: 7 November 2025

Abstract

Prostate cancer is a global health concern, and early diagnosis plays a vital role in improving the survival rate. Accurate segmentation is a key step in the automated diagnosis of prostate cancer; however, manual segmentation remains time-consuming and challenging. Micro-Ultrasound (US) is particularly well-suited for prostate cancer detection, offering real-time imaging with a resolution comparable to that of MRI. This enables improved spatial resolution and detailed visualization of small anatomical structures. With recent advances in deep learning for medical image segmentation, precise prostate segmentation has become critical for biopsy guidance, disease diagnosis, and follow-up. However, segmentation of the prostate in micro-US images remains challenging due to indistinct boundaries between the prostate and surrounding tissue. In this work, we propose a model for precise micro-ultrasound image segmentation. The model employs a dual-encoder architecture that integrates Convolutional Neural Networks (CNN) and Transformer-based encoders in parallel, combined with a fusion module to capture both global dependencies and low-level spatial details. More importantly, we introduce a decoder based on Mamba v2 to enhance segmentation accuracy. A Hypergraph Neural Network (HGNN) is employed as a bridge between the dual encoders and Mamba decoder to model correlations among non-pairwise connections. Experimental results on micro-US datasets demonstrated that our model achieved superior or comparable performance to state-of-the-art methods, with a Dice score of 0.9416 and an HD95 of 1.93.

1. Introduction

Prostate cancer is the second most prevalent cancer worldwide, with over 1.2 million new cases reported in 2020 [1]. Accurate imaging is critical for an early diagnosis. Traditionally, lesions in the prostate are identified using multiparametric MRI (mpMRI), which provides high-resolution anatomical and functional information about the lesions. However, MRI is expensive, time-consuming, and often inaccessible in many clinical settings, limiting its widespread adoption [2]. Ultrasound has become a widely adopted imaging modality owing to its low cost and high accessibility [3]. Micro-Ultrasound (micro-US) has recently emerged as a promising alternative, operating at substantially higher frequencies (typically 29 MHz or greater) than conventional ultrasound, thereby offering improved spatial resolution for prostate imaging [4]. Medical image segmentation plays a critical role in clinical practice by enabling the automatic localization of the prostate capsule. Although manual segmentation is feasible, it remains a time-consuming and labor-intensive task that is often inadequate for capturing the broader pathological context [5,6,7]. Recently, deep learning has been successfully applied to medical image segmentation. Nevertheless, despite advances in neural network architectures, segmentation remains a challenging task owing to noisy and scattered features, low resolution, weak boundaries, and irregular shapes of human organs [6,8,9].
In deep learning architectures, the convolution operation is the fundamental building block of U-Net [10,11]; however, its inherently limited receptive field constrains the network’s ability to capture global context [5]. To overcome this limitation, attention mechanisms have been introduced into deep learning. Attention allocates computational resources to regions containing the most relevant information, mimicking the functionality of the human visual system [6,12]. Consequently, several attention mechanisms have been proposed that can be categorized according to various criteria. One common categorization is based on the dimensions along which the attention feature map functions. For instance, channel attention exploits inter-channel dependencies, which are employed in Squeeze-and-Excitation (SE) networks [12]. In contrast, spatial attention emphasizes regions within feature maps that contain critical information while suppressing less relevant areas, as demonstrated in Non-Local Neural Networks [13]. Attention mechanisms can also be categorized based on the operation of the attention map [5]. For example, global attention enhances interactions across both channels and spatial locations, as in DANet [5,14], whereas local attention, such as in CBAM [15], focuses on specific spatial regions and their relationships with the channels.
Recently, Transformer-based models have gained prominence, initially introduced for sequence-to-sequence tasks in natural language processing (NLP) [16]. Transformers have been adopted across various domains due to their powerful attention mechanisms. Unlike CNN-based models, they overcome the limitation of restricted receptive fields by employing a self-attention (SA) mechanism [17]. The SA captures the internal correlations among all input tokens, enabling the modeling of long-range dependencies [18]. The Transformer architecture incorporates multiple self-attention blocks, known as multi-head self-attention (MHSA), which operate in parallel to generate diverse feature representations [17,18]. Building on this, Dosovitskiy et al. [19] introduced the Vision Transformer (ViT) for computer vision tasks, such as image classification, achieving state-of-the-art results on ImageNet. Moreover, Cao et al. [20] introduced Swin-Unet, the first U-shaped segmentation network based entirely on a Transformer architecture [21]. Swin-Unet integrates Swin Transformer blocks, where Window Multi-Head Self-Attention (W-MSA) captures fine-grained details within each window, and Shifted Window Self-Attention (SW-MSA) models cross-window interactions [22]. Unlike U-Net, Swin-Unet replaces conventional upsampling with patch-expanding layers, eliminating the need for convolutions or interpolation [6,23,24]. Nevertheless, despite its strong capability to capture the global context, the Transformer remains limited in modeling fine-grained details due to its lack of spatial inductive bias for local information. This limitation is particularly evident in medical image segmentation tasks [25].
To leverage the strengths of both Transformers and U-Net, Chen et al. [26] proposed the TransUNet. This hybrid architecture employs Convolutional Neural Networks (CNNs) to extract low-level features and a Transformer to capture global information [25]. Furthermore, hybrid models, such as TransFuse, adopt parallel encoder branches. In TransFuse [25], both a CNN-based spatial branch and a transformer-based global branch are fused using a BiFusion module [24,27]. In the BiFusion module, Transformer features are refined using a channel attention SE block [12], whereas the CNN features are enhanced using a spatial attention module inspired by CBAM [15]. The features from both branches were then combined using a Hadamard product to model their interactions. Similarly, CoTrFuse [28] follows a dual-branch strategy but incorporates Swin Transformer blocks for the global branch and EfficientNet blocks for the spatial branch, with feature fusion achieved through a specialized STCF module. Other researchers have focused on improving decoder performance. For instance, TransNorm [29] employs spatial normalization outputs from the Transformer to construct a two-level attention gate, where channel attention normalizes the feature representation to emphasize more informative channels and Transformer-produced spatial coefficients to amplify the relevant areas of the feature map. Other architectures, such as MS-TransUNet++ [30] and CoT-UNet++ [31], enhance CNN-Transformer hybrid architectures by introducing dense skip connections between encoders and decoders at multiple levels, similar to U-Net++ [6]. Some works have explored the frequency domain. For example, Discrete Fourier Transform (DFT) modules have been integrated to capture long-range dependencies more effectively [32]. Another line of research utilized deep supervision, which was first introduced in Google’s Inception architecture [5,33]. U-Net++ adopts deep supervision to help intermediate layers learn discriminative features and mitigate vanishing gradients [34]. Other network architectures incorporate side outputs for deep supervision. For example, BASNet [35] employs deep supervision via side outputs and was proposed for salient-object detection. One network relevant to this research is MicroSegNet, which employs multi-scale deep supervision for prostate segmentation from micro-ultrasound images [4].
In this paper, we propose an enhanced network architecture that utilizes a dual encoder, comprising both CNN and Transformer encoders, to effectively capture both local and global contextual information. The decoder is designed based on Mamba v2 in combination with CNN layers to improve the segmentation accuracy. Additionally, a Hypergraph Neural Network (HGNN) is integrated into the skip connections to capture non-pairwise correlations. To mitigate overfitting, we employed a tailored ultrasound-specific augmentation scheme incorporating depth attenuation, Gaussian shadows, haze artifacts, and speckle reduction.The experimental results demonstrate that the proposed architecture outperforms state-of-the-art segmentation models, achieving a superior segmentation accuracy. Our contributions can be summarized as follows:
  • We propose a model for precise Micro-Ultrasound medical image segmentation composed of dual encoders with a fusion module designed to capture both local details and long-range dependencies. A Hypergraph Neural Network (HGNN) is integrated into the skip connections to model non-pairwise correlations.
  • To further enhance segmentation accuracy, a Mamba-based decoder is incorporated, utilizing VSSD blocks built upon Mamba-2 and NC-SSD.
  • Experimental results demonstrate that our method achieves superior performance on the Micro-Ultrasound (US) prostate medical image segmentation dataset.

2. Materials and Methods

2.1. Overall Architecture

Inspired by CoTrFuse [28], we designed a network comprising two parallel branches. One branch employs a Transformer (Swin blocks) to capture the global context, whereas the other branch utilizes EfficientNet blocks to extract local details. The outputs from both branches are fused using a module based on BiFusion [25]. The overall architecture of the model is illustrated in Figure 1. On the decoder side, a combination of CNN layers and Visual State Space Duality (VSSD) blocks based on Mamba v2 was employed. The VSSD incorporates Non-Causal SSD (NC-SSD) modules [36] to enhance feature representation. Furthermore, inspired by BASNet [35], we integrated side outputs from the decoder for deep supervision, leveraging complementary information across intermediate prediction maps [6]. These outputs were upsampled and incorporated into the training to provide guidance at multiple levels. This approach accelerates convergence and mitigates the vanishing gradient problems. A more detailed discussion of the network architecture is provided in the following section.

2.1.1. Dual Encoder

The dual-encoder architecture employs EfficientNet to extract local features [37]. EfficientNet was developed using a neural architecture search (NAS) to systematically scale convolutional networks, resulting in a family of highly efficient models that balance network width, depth, and resolution. Each EfficientNet model is constructed from a series of Inverted Residual Blocks, also referred to as MBConv blocks [38]. The other branch employs Swin Transformer blocks with patch-merging layers, similar to [20,28]. As mentioned in the introduction section, Cao et al. [20] introduced Swin-Unet, a Transformer-based architecture that incorporates a hierarchical structure and a window-shifting mechanism to enhance feature representation. Swin Transformer blocks perform feature learning through a window-based multi-head self-attention (W-MSA) module, followed by a shifted window multi-head self-attention (SW-MSA) module [20], as illustrated in Figure 2. Specifically, the W-MSA captures the local dependencies among pixels within each window, whereas the SW-MSA models the global interactions across windows, thereby enabling cross-contextual attention for feature recalibration. The mathematical formulation of the Swin Transformer blocks is shown in Equations (1)–(4):
Z n = WMSA ( L N ( Z n 1 ) ) + Z n 1
Z n = MLP ( L N ( Z n ) ) + Z n
Z n + 1 = SWMSA ( L N ( Z n ) ) + Z n
Z n + 1 = MLP ( L N ( Z n + 1 ) ) + Z n + 1
where Z n and Z n represent the outputs of the WMSA/SWMSA module and the MLP module of the n t h block.
The self-attention (SA) mechanism is the core component of the Transformer’s block, as shown in Equation (5):
A t t e n t i o n ( Q , K , V ) = s o f t m a x ( Q K T d + B ) V
where Q , K , V R P 2 ×   d represent the query, key, and value matrices, respectively. P 2 and d donate the number of patches in a window and the dimension of either the query and key, respectively. B is a relative position bias, and it is based on the bias matrix B R ( 2 P 1 ) × ( 2 P + 1 ) .
Similar to Swin-Unet, the first step in the Swin Transformer branch of our model is a linear embedding layer that projects the input features into an arbitrary dimension. This is followed by a patch-merging layer that performs downsampling while increasing the feature dimension before the Transformer blocks process the data. Skip connections carrying multi-scale features from the Transformer encoder are later fused with the feature maps from the CNN encoder in the fusion module, as discussed in the subsequent section. For simplicity, the linear embedding layer is omitted from Figure 1.
The two branches are merged using a fusion module as illustrated in the next section.

2.1.2. Fusion Module

TransFuse [25] and CoTrFuse [28] both leverage dual encoders that are combined through fusion modules. Although the two fusion modules share similar concepts, they differ in their implementations. Both modules incorporate channel and spatial attention mechanisms to enhance feature representation.
For example, the BiFusion module [25] in TransFuse uses an SE block [12] as channel attention to emphasize global information from the Transformer branch, while a CBAM block [15] is used as spatial attention to capture fine-grained details from the CNN branch.
The STCF fusion module in CoTrFuse adopts a similar attention-based approach. However, unlike BiFusion, where channel attention is applied to the Transformer feature map and spatial attention is applied to the CNN feature map, STCF applies both attention mechanisms to both feature maps before summing them. In our model, we adopted the STCF fusion strategy, as shown in Figure 3.

2.1.3. Hyper GNN

In an ordinary graph, each edge connects only two nodes, representing the pairwise correlations. In contrast, a hypergraph employs hyperedges that can connect more than two nodes, thereby enabling the modeling of non-pairwise correlations. This makes hypergraphs more effective in capturing complex relationships. Han et al. [39] introduced a Vision Hypergraph Neural Network (ViHGNN). Subsequently, Feng et al. [40] proposed a Hypergraph Neural Network (HGNN), which constructs a hypergraph from an image using K-Nearest Neighbors (KNN). In the context of medical image segmentation, Peng et al. [41] integrated an HGNN into a U-Net for MRI segmentation. More recent work has focused on adaptive hypergraph construction. For example, unlike earlier methods that relied on a fixed number of neighbors per node for graph construction, Chai et al. [42] proposed an adaptive strategy for hyperedge formation known as Adaptive Hypergraph Construction. This method utilizes the K-Nearest Neighbors (KNN) algorithm to generate a matrix used to compute the degree of each node, which in turn guides the construction of hyperedges. Readers are referred to [42] for further details regarding the Adaptive Hypergraph Construction procedure. Their Adaptive Hypergraph Construction method models the shape attributes more accurately, particularly in medical imaging. Furthermore, Chai et al. extended the concept of convolutional sliding windows to include hypergraphs. In this approach, the sliding-window-based convolution mechanism employs fixed-size kernels to convolve the feature maps, enabling interaction with local neighboring pixels across the entire image through stride operations.
The Adaptive Hypergraph Construction is illustrated in Figure 4, with hypergraph convolution denoted as HGC.

2.1.4. Mamba Decoder and VSSD Block

Motivated by the success of Transformers in capturing long-range dependencies, Mamba with State Space Models (SSMs) has been proposed to capture the global context while avoiding the high computational cost of Vision Transformers (ViTs). The Mamba S6 model improves upon the S4 model by introducing a selective mechanism and hardware optimization. To adapt Mamba for computer vision, VMamba [36] converts non-causal visual images into sequences of ordered patches using a Cross-Scan Module (CSM). The S6 block was further enhanced in Mamba2 [43] through the introduction of State Space Duality (SSD) [44]. Recently, Xu et al. proposed a Non-Causal SSD (NC-SSD) that eliminates the need for a causal mask and specialized scanning paths. Building on this, they introduced Visual State Space Duality (VSSD) to replace the VSS vision block [45]. The VSSD block based on Mamba2 is shown in Figure 5.
The VSSD block exhibits a design similar to that of the Mamba block [46]; however, it employs the NC-SSD module instead of the standard Mamba SSD. Vision Mamba models, such as Vision Mamba [47] and VMamba [36] typically rely on image traversal mechanisms that use different scanning routes to flatten a 2D image into a 1D sequence. In contrast, NC-SSD introduces an enhanced algorithm by leveraging the fact that the matrix A in Mamba 2 [43] is reduced to a scalar [45]. This design eliminates the need for a specific scanning route and removes causal mask requirements. For more detailed information on the NC-SSD, readers are referred to [45]. As shown in Figure 1, our decoder employs four VSSD blocks, with convolution and ReLU activation layers inserted between them, each of which is repeated twice. The configuration of the VSSD blocks is as follows: channel dimensions [256, 384, 448, 480] and number of heads [2, 4, 8, 16]. At each stage, each block was applied once without repetition.

2.2. Dataset

For this study, we utilized a prostate segmentation dataset based on micro-ultrasound (micro-US) images from [4]. The dataset is publicly available at https://zenodo.org/records/10475293 and was accessed on 2 February 2025. To the best of our knowledge, this is the only publicly available micro-US dataset. Micro-ultrasound is a 29-MHz imaging technology that provides a resolution that is 3–4 times higher than that of conventional ultrasound. A key advantage of micro-US over other modalities, such as MRI and standard ultrasound, is its real-time visualization capability, eliminating the need for image fusion, which can account for up to 50% of diagnostic errors. The dataset was collected from 75 men who underwent micro-US-guided prostate biopsy at the University of Florida [4]. Each patient underwent a prostate scan from left to right, capturing approximately 200–300 micro-US images. The images were converted from B-mode to a DICOM series with embedded pixel spacing information. Ground-truth prostate annotations were performed by two non-expert annotators and one expert annotator for all 75 patients. Additionally, for evaluation purposes, three annotators manually segmented the prostate capsule in 20 test cases. For model training, 2060 micro-US images from 55 patients were used, and 758 images from 20 patients were reserved for testing. The dataset collection study was approved by the Institutional Review Board at the University of Florida.

2.3. Augmentation

Ultrasound-specific data augmentation techniques were employed to prevent overfitting. Four different augmentations have been utilized: Depth Attenuation, Gaussian Shadow, Haze Artifact, and Speckle Reduction using the USAugment (version 1.0.1) [48] (https://github.com/adamtupper/ultrasound-augmentation, access date (7 May 2025)). Depth Attenuation simulates the gradual loss of ultrasound wave energy as it propagates through the tissue, causing the intensity to decrease with increasing distance from the probe. Haze Artifact models semi-static noise bands that occasionally appear in the ultrasound images. Gaussian shadows replicate the acoustic shadows caused by air or tissue obstructing wave propagation, generating two-dimensional Gaussian-shaped shadows with randomly selected parameters. Speckle noise, arising from interference among ultrasound waves, is mitigated using Speckle Reduction, which applies a bilateral filter with randomly sampled parameters to reduce speckle patterns. Empirically, we found that applying these augmentations with a probability of 0.2 yields the best result.

2.4. Loss Function and Evaluation

Following [4], the AG-BCE loss function was employed. The AG-BCE loss is based on the BCE loss but accounts for the characteristics of prostate segmentation in micro-US images by penalizing prediction errors more heavily in challenging regions, particularly along the borders between the prostate and bladder. Hence, it assigns different weights to each pixel.
L AG - BCE = 1 N i = 1 N w i y i log ( p i ) + ( 1 y i ) log ( 1 p i )
where
  • N is the total number of pixels,
  • w i { 0 , 1 } is the is the weight of assigned to pixel i,
  • y i { 0 , 1 } is the ground truth label for pixel i,
  • p i [ 0 , 1 ] is the predicted probability for pixel i.
The AG-BCE loss function differentiates between hard and easy regions. Hard regions are defined as areas where expert and non-expert annotations disagree, whereas easy regions are defined as those where both annotations coincide. AG-BCE assigns a weight (denoted by Equation (6) as w of four hard regions and one easy region, following the implementation in MicroSegNet [4]. For a more detailed explanation of the AG-BCE loss formulation, readers are referred to [4].
The average Dice Similarity Coefficient (DSC) and average Hausdorff Distance (HD95) were used as evaluation metrics. HD95 measures the Hausdorff distance by considering the 95th percentile of the distances between the boundary points of the two sets. Unlike the standard Hausdorff distance, which reports the maximum distance, HD95 reduces the influence of outliers by ignoring the extreme values [6]. The Dice Similarity Coefficient (DSC) is defined as
DSC ( G , P ) = 2 | G P | | G | + | P | ,
where G denotes the set of ground-truth pixels, and P denotes the set of predicted pixels. The Hausdorff distance between two sets G and P is defined as
d H ( G , P ) = max d h ( G , P ) , d h ( P , G ) ,
where
d h ( G , P ) = max g G min p P g p ,
and
d h ( P , G ) = max p P min g G p g .
Here, G and P denote the ground truth and the predicted outcome, respectively, while · represents the Euclidean distance. Intuitively, d h ( G , P ) is the maximum distance from any point in G to its nearest neighbor in P, and d h ( P , G ) is defined analogously.
In order to account for image resizing, the HD95 value is adjusted by multiplying it with a spacing parameter, which was specified as 0.033586 in the implementation described by [4].

2.5. Implementation Details

As a preprocessing step, all images were resized to 224 × 224 . Following the setup in [4], training was conducted using an image patch size of 16 and a batch size of eight. Stochastic Gradient Descent (SGD) optimizer and momentum, with a decaying learning rate, is used as model optimizer. The initial learning rate, momentum, and weight decay were set to 0.01, 0.9, and 0.0001, respectively. An adaptive learning rate strategy was employed, which reduced the learning rate after a certain number of iterations. The number of epochs was set to 150. Both training and testing were implemented and conducted using PyTorch (Version: 2.7.1+cu118) on an NVIDIA RTX A4500 GPU with 20 GB of memory and Python 3.12.7.

3. Results and Discussion

Table 1 presents a comparison between our proposed model and the state-of-the-art (SOTA) methods, with Dice Similarity Coefficient (DSC) and 95th percentile Hausdorff Distance (HD95) used as evaluation metrics. The results in Table 1 represent the average performance across several runs. Our reported results are based on our own training, which was conducted for both 150 and 10 epochs to ensure a fair comparison. The only exception is the MicroSegNet scores, that is based on 10 epochs. This score is reported from [4]. As noted by H. Jiang et al. [4], MicroSegNet training was limited to 10 epochs to avoid overfitting. In contrast, overfitting in our model was mitigated through data augmentation, as discussed earlier. However, to ensure a fair comparison under the same conditions (with augmentation), Table 2 presents results for our model and an augmented version of MicroSegNet. As shown in Table 1, our model achieved its best performance when trained for 150 epochs. For comparison, we incorporated several models, some of which are recent, such as RWKV-UNet [49] and SegU-KAN [50]. For all evaluated models, we used AG-BCE as the loss function, similar to [4], to focus more on challenging regions over easily segmented ones during training.
We employed pretraining for certain models, such as Swin-UNet [20], but not all models in Table 1 were pretrained. As shown in Table 1, our model achieved the highest performance, consistently outperforming the competing approaches. Specifically, it attained an average Dice Similarity Coefficient (DSC) of 94.16% and a Hausdorff Distance (HD95) of 1.93 mm. For example, compared with Swin-UNet, the average DSC increased from 0.932 to 0.9416, corresponding to an improvement of nearly 1%, whereas the HD95 decreased from 2.04 mm to 1.93 mm, representing a reduction of approximately 5%. These experimental results demonstrate that the proposed network architecture outperforms state-of-the-art segmentation models, achieving a significantly improved segmentation accuracy. Moreover, Table 1 also reports the number of FLOPs, model parameters, GPU memory usage, and inference time for each model. Notice, the number of FLOPs and the number of parameters are calculated using the calflops library (https://github.com/MrYxJ/calculate-flops.pytorch, access date (11 March 2024)).
Figure 6 presents the segmentation outcomes of the various approaches, namely, SwinUNet, MicroSegNet and our model on the Micro-ultrasound dataset from three images for qualitative comparison. To evaluate the impact of different optimizers, we conducted experiments on the Micro-Ultrasound dataset. Figure 7 and Figure 8 illustrate the results for four optimizers: SGD [57,58], AdamW [59], SAM [60,61], and Adan [62]. For clarity, the training losses and validation scores are presented in separate figures. SGD, which we used for our model, updates the gradient in each iteration using a randomly selected sample rather than computing the exact gradient. The AdamW optimizer is similar to Adam, but while the Adam optimizer adaptively adjusts the learning rate for each parameter based on first and second order moment estimates of the gradient [6] using the geometry curvature of the objective function [62], AdamW decouples the weight decay from the gradient update.
Sharpness-Aware Minimization (SAM) attempts to find parameter neighborhoods that have uniformly low loss, making it more robust to noisy labels in the training set by perturbing the loss landscape. SAM is particularly beneficial when applied to Vision Transformers (ViTs), as these architectures are more prone to end up in local minima than CNN-based architectures [63,64]. However, SAM requires forward and backward passes twice for each iteration. Adaptive Nesterov Momentum (Adan) introduces a new Nesterov Momentum Estimation (NME) method based on Nesterov acceleration. Its efficiency arises from the elimination of the overhead of computing the gradient at the extrapolation point. Figure 7 illustrates the behavior of each optimizer during the training. Both AdamW and Adan exhibited lower loss curves, indicating better performance than SGD and SAM. This trend is also confirmed by the evaluation curves shown in Figure 8. Unlike in Figure 7, the curves in Figure 8 are intertwined; however, careful examination reveals that AdamW and Adan show better accuracy than the others.
Figure 9 and Figure 10 depict the loss landscapes of different neural networks as 3D surface plots and contour maps, respectively. These visualizations were generated using the filter-normalization method [65], with the 3D surface renderings in Figure 9 produced using ParaView (Version 5.12) (http://paraview.org/, access date (5 March 2024)) software. The filter-normalization approach evaluates the loss landscape along filter-normalized directions, where the geometry of the landscape provides insights into model trainability and generalizability [64]. In particular, smoother and more convex landscapes are typically correlated with lower error values, whereas irregular or chaotic landscapes tend to yield higher training errors.

4. Ablation Study

Ablation studies were performed to assess the contribution of each component in the proposed method and to support the design decisions based on the performance outcomes of the Micro-Ultrasound dataset. These results reflect the average scores computed over multiple runs. The results for the five different configurations are summarized in Table 3. Notice that DS denotes deep supervision. Our analysis demonstrates the effectiveness of the proposed enhancements compared with the baseline model. As shown in the table, integrating the Mamba blocks into the decoder improved both the DSC and the 95th percentile HD95. The addition of deep supervision further enhanced the performance of both metrics. Table 3 also lists the number of FLOPs and parameters. It is noteworthy that the majority of components contribute minimally to the increase in FLOPs and parameters; the only exception is the integration of Mamba blocks within the decoder (VSSD), which leads to an increase in both metrics. Also, notice that adding the hyperGraph module slightly improved the DSC score, but it worsened the HD score. Hence, to verify the effectiveness of integrating HyperGraph, Table 4 presents a comparison of the model with and without HyperGraph. As shown, HyperGraph indeed enhances accuracy based on both metrics. Finally, the combination of Mamba, deep supervision, and ultrasound-specific data augmentation achieved the best average results for both HD95 and DSC scores.
Table 5 presents a comparison of the model performance based on different Swin block configurations. The Tiny (Swin-T) configuration has channel dimensions of [96, 192, 384, 768] and numbers of heads [3, 6, 12, 24], with the blocks repeated [2, 2, 6, 2]. The Small (Swin-S) configuration uses the same channel dimensions and numbers of heads but repeats the blocks [2, 2, 18, 2] [66]. Note that we did not test the larger Swin variants (Swin-B and Swin-L) due to their higher GPU memory requirements.

5. Conclusions

This work presents an improved model for ultrasound medical image segmentation, with particular emphasis on micro-US. The proposed architecture features a dual-encoder design: a Swin Transformer branch for capturing global context and a CNN-based branch for extracting fine local details. To further enhance feature representation, a hypergraph network is incorporated to model non-pairwise correlations. The decoder leverages the Mamba 2 architecture, built upon VSSD blocks, to improve localization accuracy. Furthermore, ultrasound-specific augmentation techniques, such as depth attenuation, were employed during training to mitigate overfitting and enhance robustness. Experimental results demonstrate that the proposed method achieves superior performance across multiple evaluation metrics, underscoring its potential to advance segmentation accuracy in ultrasound medical imaging. Future research could focus on mitigating the scarcity of Micro-Ultrasound (US) image datasets using semi-supervised or weakly supervised learning. These approaches can effectively leverage large volumes of unlabeled data to reduce annotation costs while preserving the segmentation boundary accuracy. In particular, weakly supervised methods may utilize alternative supervisory cues such as image-level tags, bounding boxes, or scribbles [67].

Author Contributions

Conceptualization, A.A.-Q., M.K.A.; methodology, A.A.-Q., M.K.A.; investigation, A.A.-Q., M.K.A.; writing—original draft preparation, A.A.-Q.; writing—review and editing, A.T., M.K.A.; supervision, M.K.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Micro-ultrasound (micro-US) images segmentation dataset is publicly available at https://zenodo.org/records/10475293 and was accessed on 2 February 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, J.; Xu, C.; Lee, H.J.; Ren, S.; Zi, X.; Zhang, Z.; Wang, H.; Yu, Y.; Yang, C.; Gao, X.; et al. A genomic and epigenomic atlas of prostate cancer in Asian populations. Nature 2020, 580, 93–99. [Google Scholar] [CrossRef]
  2. Imran, M.; Nguyen, B.; Pensa, J.; Falzarano, S.M.; Sisk, A.E.; Liang, M.; DiBianco, J.M.; Su, L.M.; Zhou, Y.; Joseph, J.P.; et al. Image registration of in vivo micro-ultrasound and ex vivo pseudo-whole mount histopathology images of the prostate: A proof-of-concept study. Biomed. Signal Process. Control 2024, 96, 106657. [Google Scholar] [CrossRef]
  3. Wasih, M.; Ahmad, S.; Almekkawy, M. A robust cascaded deep neural network for image reconstruction of single plane wave ultrasound RF data. Ultrasonics 2023, 132, 106981. [Google Scholar] [CrossRef]
  4. Jiang, H.; Imran, M.; Muralidharan, P.; Patel, A.; Pensa, J.; Liang, M.; Benidir, T.; Grajo, J.R.; Joseph, J.P.; Terry, R.; et al. MicroSegNet: A deep learning approach for prostate segmentation on micro-ultrasound images. Comput. Med. Imaging Graph. 2024, 112, 102326. [Google Scholar] [CrossRef]
  5. Sun, Y.; Dai, D.; Zhang, Q.; Wang, Y.; Xu, S.; Lian, C. MSCA-Net: Multi-scale contextual attention network for skin lesion segmentation. Pattern Recognit. 2023, 139, 109524. [Google Scholar] [CrossRef]
  6. Al Qurri, A.; Almekkawy, M. Improved UNet with Attention for Medical Image Segmentation. Sensors 2023, 23, 8589. [Google Scholar] [CrossRef]
  7. Wang, C.; Xu, R.; Xu, S.; Meng, W.; Zhang, X. Automatic polyp segmentation via image-level and surrounding-level context fusion deep neural network. Eng. Appl. Artif. Intell. 2023, 123, 106168. [Google Scholar] [CrossRef]
  8. Huo, X.; Sun, G.; Tian, S.; Wang, Y.; Yu, L.; Long, J.; Zhang, W.; Li, A. HiFuse: Hierarchical multi-scale feature fusion network for medical image classification. Biomed. Signal Process. Control 2024, 87, 105534. [Google Scholar] [CrossRef]
  9. Gao, Q.; Almekkawy, M. ASU-Net++: A nested U-Net with adaptive feature extractions for liver tumor segmentation. Comput. Biol. Med. 2021, 136, 104688. [Google Scholar] [CrossRef] [PubMed]
  10. Ahmed, A.Q.; Almekkawy, M. Improved UNet++ Based on Kolmogorov-Arnold Convolutions. In Proceedings of the 2025 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 14–18 September 2025; pp. 905–910. [Google Scholar]
  11. Kang, J.; Al-Qurri, A.; Almekkawy, M. Fast and Resource-Efficient Ultrasound Segmentation Using FPGAs. In Proceedings of the 2025 IEEE International Ultrasonics Symposium (IUS), Utrecht, The Netherlands, 15–18 September 2025; pp. 1–5. [Google Scholar]
  12. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  13. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-local neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  14. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3146–3154. [Google Scholar]
  15. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Computer Vision—ECCV 2018, Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–19. [Google Scholar]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems 30 (NIPS 2017), Proceedings of the Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., Von Luxburg, U., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates Inc.: Red Hook, NY, USA, 2017. [Google Scholar]
  17. Wang, H.; Xie, S.; Lin, L.; Iwamoto, Y.; Han, X.H.; Chen, Y.W.; Tong, R. Mixed transformer u-net for medical image segmentation. In Proceedings of the ICASSP 2022–2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 2390–2394. [Google Scholar]
  18. Li, J.; Chen, J.; Tang, Y.; Wang, C.; Landman, B.A.; Zhou, S.K. Transforming medical imaging with Transformers? A comparative review of key properties, current progresses, and future perspectives. Med. Image Anal. 2023, 85, 102762. [Google Scholar] [CrossRef] [PubMed]
  19. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  20. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-like pure transformer for medical image segmentation. In Computer Vision—ECCV 2022 Workshops, Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October, 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 205–218. [Google Scholar]
  21. Zuo, S.; Xiao, Y.; Chang, X.; Wang, X. Vision transformers for dense prediction: A survey. Knowl.-Based Syst. 2022, 253, 109552. [Google Scholar] [CrossRef]
  22. Gao, B.B.; Huang, Z. CSTrans: Correlation-guided Self-Activation Transformer for Counting Everything. Pattern Recognit. 2024, 153, 110556. [Google Scholar] [CrossRef]
  23. He, D.; Zhang, Y.; Huang, H.; Si, Y.; Wang, Z.; Li, Y. Dual-branch hybrid network for lesion segmentation in gastric cancer images. Sci. Rep. 2023, 13, 6377. [Google Scholar] [CrossRef]
  24. Azad, R.; Kazerouni, A.; Heidari, M.; Aghdam, E.K.; Molaei, A.; Jia, Y.; Jose, A.; Roy, R.; Merhof, D. Advances in medical image analysis with vision transformers: A comprehensive review. Med. Image Anal. 2023, 91, 103000. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, Y.; Liu, H.; Hu, Q. Transfuse: Fusing transformers and cnns for medical image segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2021: 24th International Conference, Strasbourg, France, 27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 14–24. [Google Scholar]
  26. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  27. Ahmed, A.Q.; Alqarni, A.; Almekkawy, M. Trifuse: Triplet Encoders Network for Medical Image Segmentation. In Proceedings of the 2025 IEEE 22nd International Symposium on Biomedical Imaging (ISBI), Houston, TX, USA, 14–17 April 2025; pp. 1–5. [Google Scholar]
  28. Chen, Y.; Wang, T.; Tang, H.; Zhao, L.; Zhang, X.; Tan, T.; Gao, Q.; Du, M.; Tong, T. CoTrFuse: A novel framework by fusing CNN and transformer for medical image segmentation. Phys. Med. Biol. 2023, 68, 175027. [Google Scholar] [CrossRef]
  29. Azad, R.; Al-Antary, M.T.; Heidari, M.; Merhof, D. Transnorm: Transformer provides a strong spatial normalization mechanism for a deep segmentation model. IEEE Access 2022, 10, 108205–108215. [Google Scholar] [CrossRef]
  30. Wang, B.; Wang, F.; Dong, P.; Li, C. Multiscale transunet++: Dense hybrid U-Net with transformer for medical image segmentation. Signal, Image Video Process. 2022, 16, 1607–1614. [Google Scholar] [CrossRef]
  31. Yin, Y.; Xu, W.; Chen, L.; Wu, H. CoT-UNet++: A medical image segmentation method based on contextual transformer and dense connection. Math. Biosci. Eng. 2023, 20, 8320–8336. [Google Scholar] [CrossRef]
  32. Al-Qurri, A.; Almekkawy, M. Ultrasound Image Segmentation using a Model of Transformer and DFT. In Proceedings of the 2024 IEEE UFFC Latin America Ultrasonics Symposium (LAUS), Montevideo, Uruguay, 8–10 May 2024; pp. 1–4. [Google Scholar]
  33. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  34. Zeng, Z.; Hu, Q.; Xie, Z.; Zhou, J.; Xu, Y. Small but Mighty: Enhancing 3D Point Clouds Semantic Segmentation with U-Next Framework. arXiv 2023, arXiv:2304.00749. [Google Scholar] [CrossRef]
  35. Qin, X.; Zhang, Z.; Huang, C.; Gao, C.; Dehghan, M.; Jagersand, M. Basnet: Boundary-aware salient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7479–7489. [Google Scholar]
  36. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Liu, Y. Vmamba: Visual state space model. arXiv 2024, arXiv:2401.10166. [Google Scholar] [PubMed]
  37. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  38. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  39. Han, Y.; Wang, P.; Kundu, S.; Ding, Y.; Wang, Z. Vision hgnn: An image is more than a graph of nodes. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 19878–19888. [Google Scholar]
  40. Feng, Y.; You, H.; Zhang, Z.; Ji, R.; Gao, Y. Hypergraph neural networks. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu, HI, USA, 27 January–1 February 2019; pp. 3558–3565. [Google Scholar]
  41. Peng, J.; Yang, J.; Xia, C.; Li, X.; Guo, Y.; Fu, Y.; Chen, X.; Cui, Z. Make U-Net Greater: An Easy-to-Embed Approach to Improve Segmentation Performance Using Hypergraph. Comput. Syst. Sci. Eng. 2022, 42, 319–333. [Google Scholar] [CrossRef]
  42. Chai, S.; Jain, R.K.; Mo, S.; Liu, J.; Yang, Y.; Li, Y.; Tateyama, T.; Lin, L.; Chen, Y.W. A Novel Adaptive Hypergraph Neural Network for Enhancing Medical Image Segmentation. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2024, Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 23–33. [Google Scholar]
  43. Dao, T.; Gu, A. Transformers are ssms: Generalized models and efficient algorithms through structured state space duality. arXiv 2024, arXiv:2405.21060. [Google Scholar] [CrossRef]
  44. Al-Qurri, A.; Almekkawy, M. Enhancing Medical Image Segmentation with Mamba and UNet++. In Proceedings of the 2025 IEEE 22nd International Symposium on Biomedical Imaging (ISBI), Houston, TX, USA, 14–17 April 2025; pp. 1–5. [Google Scholar]
  45. Shi, Y.; Dong, M.; Li, M.; Xu, C. VSSD: Vision Mamba with Non-Casual State Space Duality. arXiv 2024, arXiv:2407.18559. [Google Scholar] [CrossRef]
  46. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. In Proceedings of the First Conference on Language Modeling, Philadelphia, PA, USA, 7–9 October 2024. [Google Scholar]
  47. Zhu, L.; Liao, B.; Zhang, Q.; Wang, X.; Liu, W.; Wang, X. Vision mamba: Efficient visual representation learning with bidirectional state space model. arXiv 2024, arXiv:2401.09417. [Google Scholar] [CrossRef]
  48. Tupper, A.; Gagné, C. Revisiting Data Augmentation for Ultrasound Images. arXiv 2025, arXiv:2501.13193. [Google Scholar] [CrossRef]
  49. Jiang, J.; Zhang, J.; Liu, W.; Gao, M.; Hu, X.; Yan, X.; Huang, F.; Liu, Y. Rwkv-unet: Improving unet with long-range cooperation for effective medical image segmentation. arXiv 2025, arXiv:2501.08458. [Google Scholar]
  50. Li, C.; Liu, X.; Li, W.; Wang, C.; Liu, H.; Liu, Y.; Chen, Z.; Yuan, Y. U-kan makes strong backbone for medical image segmentation and generation. In Proceedings of the AAAI Conference on Artificial Intelligence, Philadelphia, PA, USA, 25 February–4 March 2025; pp. 4652–4660. [Google Scholar]
  51. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Proceedings of the 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Granada, Spain, 20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  52. Heidari, M.; Kazerouni, A.; Soltany, M.; Azad, R.; Aghdam, E.K.; Cohen-Adad, J.; Merhof, D. Hiformer: Hierarchical multi-scale representations using transformers for medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–7 January 2023; pp. 6202–6212. [Google Scholar]
  53. Azad, R.; Niggemeier, L.; Hüttemann, M.; Kazerouni, A.; Aghdam, E.K.; Velichko, Y.; Bagci, U.; Merhof, D. Beyond self-attention: Deformable large kernel attention for medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2024; pp. 1287–1297. [Google Scholar]
  54. Wu, R.; Liu, Y.; Liang, P.; Chang, Q. H-vmunet: High-order vision mamba unet for medical image segmentation. Neurocomputing 2025, 624, 129447. [Google Scholar] [CrossRef]
  55. Zhang, M.; Yu, Y.; Gu, L.; Lin, T.; Tao, X. VM-UNET-V2 Rethinking Vision Mamba UNet for Medical Image Segmentation. arXiv 2024, arXiv:2403.09157. [Google Scholar]
  56. Ruan, J.; Li, J.; Xiang, S. VM-UNet: Vision Mamba UNet for Medical Image Segmentation. arXiv 2024, arXiv:2402.02491. [Google Scholar] [CrossRef]
  57. Sun, S.; Cao, Z.; Zhu, H.; Zhao, J. A survey of optimization methods from a machine learning perspective. IEEE Trans. Cybern. 2019, 50, 3668–3681. [Google Scholar] [CrossRef]
  58. Li, Q.; Xiong, D.; Shang, M. Adjusted stochastic gradient descent for latent factor analysis. Inf. Sci. 2022, 588, 196–213. [Google Scholar] [CrossRef]
  59. Loshchilov, I.; Hutter, F. Decoupled weight decay regularization. arXiv 2017, arXiv:1711.05101. [Google Scholar]
  60. Roux, N.; Schmidt, M.; Bach, F. A stochastic gradient method with an exponential convergence _rate for finite training sets. In Advances in Neural Information Processing Systems 25 (NIPS 2012), Proceedings of the Annual Conference on Neural Information Processing Systems 2012, Lake Tahoe, NV, USA, 3–6 December 2012; Pereira, F., Burges, C.J., Bottou, L., Weinberger, K.Q., Eds.; Curran Associates Inc.: Red Hook, NY, USA, 2012. [Google Scholar]
  61. Foret, P.; Kleiner, A.; Mobahi, H.; Neyshabur, B. Sharpness-aware minimization for efficiently improving generalization. arXiv 2020, arXiv:2010.01412. [Google Scholar]
  62. Xie, X.; Zhou, P.; Li, H.; Lin, Z.; Yan, S. Adan: Adaptive nesterov momentum algorithm for faster optimizing deep models. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 9508–9520. [Google Scholar] [CrossRef]
  63. Chen, X.; Hsieh, C.J.; Gong, B. When vision transformers outperform resnets without pre-training or strong data augmentations. arXiv 2021, arXiv:2106.01548. [Google Scholar]
  64. Qurri, A.A.; Almekkawy, M. Hybrid MultiResUNet with transformers for medical image segmentation. Biomed. Signal Process. Control 2025, 110, 108056. [Google Scholar] [CrossRef]
  65. Li, H.; Xu, Z.; Taylor, G.; Studer, C.; Goldstein, T. Visualizing the loss landscape of neural nets. In Advances in Neural Information Processing Systems 31 (NeurIPS 2018), Proceedings of Annual Conference on Neural Information Processing Systems 2018, Montréal, QC, Canada, 3–8 December 2018; Bengio, S., Wallach, H., Larochelle, H., Grauman, K., Cesa-Bianchi, N., Garnett, R., Eds.; Curran Associates Inc.: Red Hook, NY, USA, 2018. [Google Scholar]
  66. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  67. Chen, J.; Duan, H.; Zhang, X.; Gao, B.; Grau, V.; Han, J. From gaze to insight: Bridging human visual attention and vision language model explanation for weakly-supervised medical image segmentation. IEEE Trans. Med. Imaging 2025. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overall network architecture with dual encoders, Hyper GNN, and Mamba decoder.
Figure 1. Overall network architecture with dual encoders, Hyper GNN, and Mamba decoder.
Sensors 25 06815 g001
Figure 2. Diagram of the Swin-Transformer Block based on Swin-Unet. The Swin transformer employs a window-based multi-head self-attention (W-MSA) module and a shifted window-based multi-head self-attention (SW-MSA) module.
Figure 2. Diagram of the Swin-Transformer Block based on Swin-Unet. The Swin transformer employs a window-based multi-head self-attention (W-MSA) module and a shifted window-based multi-head self-attention (SW-MSA) module.
Sensors 25 06815 g002
Figure 3. Architecture of the proposed fusion module block, adapted from the STCF fusion module [28]. The SE block and CBAM are used to perform channel and spatial attention, respectively.
Figure 3. Architecture of the proposed fusion module block, adapted from the STCF fusion module [28]. The SE block and CBAM are used to perform channel and spatial attention, respectively.
Sensors 25 06815 g003
Figure 4. The Adaptive Hypergraph Construction module, implemented following the method proposed by Chai et al. [42]. This module constructs hyper-edges adaptively based on node degree, utilizing the K-Nearest Neighbors (KNN) algorithm to form the adjacency matrix for hypergraph generation.
Figure 4. The Adaptive Hypergraph Construction module, implemented following the method proposed by Chai et al. [42]. This module constructs hyper-edges adaptively based on node degree, utilizing the K-Nearest Neighbors (KNN) algorithm to form the adjacency matrix for hypergraph generation.
Sensors 25 06815 g004
Figure 5. The VSSD block based on [45] that employ Mamba2.
Figure 5. The VSSD block based on [45] that employ Mamba2.
Sensors 25 06815 g005
Figure 6. Qualitative comparison of CT slices from the Micro-US dataset, showing prostate boundaries donated as a green circle. From left to right: (a) Ground Truth, (b) SwinUNet, (c) MicroSegNet, and (d) our model’s prediction.
Figure 6. Qualitative comparison of CT slices from the Micro-US dataset, showing prostate boundaries donated as a green circle. From left to right: (a) Ground Truth, (b) SwinUNet, (c) MicroSegNet, and (d) our model’s prediction.
Sensors 25 06815 g006
Figure 7. Dice Loss for different optimizers during training.
Figure 7. Dice Loss for different optimizers during training.
Sensors 25 06815 g007
Figure 8. Dice Score During Evaluation.
Figure 8. Dice Score During Evaluation.
Sensors 25 06815 g008
Figure 9. Loss surface visualization of different neural networks using the filter-normalization method. Top-left: UNet++; top-right: Swin-UNet; bottom-left: MicroSegNet; bottom-right: our model.
Figure 9. Loss surface visualization of different neural networks using the filter-normalization method. Top-left: UNet++; top-right: Swin-UNet; bottom-left: MicroSegNet; bottom-right: our model.
Sensors 25 06815 g009
Figure 10. Contour maps of the 2D loss landscape for different neural networks using the filter-normalization method. Top left: U-Net++; top right: Swin-UNet; bottom left: MicroSegNet; bottom right: proposed model.
Figure 10. Contour maps of the 2D loss landscape for different neural networks using the filter-normalization method. Top left: U-Net++; top right: Swin-UNet; bottom left: MicroSegNet; bottom right: proposed model.
Sensors 25 06815 g010
Table 1. The segmentation accuracy on the micro-US dataset was evaluated using the average Dice score and 95% Hausdorff distance (HD95). The best results are indicated in bold. The Dice scores marked with * are reported in [4], and their values are provided with a precision of three decimal places. All results are based on training for 150 epochs and 10 epochs to ensure a fair comparison. FLOPs (G) denote the number of floating-point operations (in billions). #Params (M) indicates the number of parameters (in millions). GPU Mem (G) denotes GPU memory usage (in gigabytes). Inference (S) represents the inference time (in seconds).
Table 1. The segmentation accuracy on the micro-US dataset was evaluated using the average Dice score and 95% Hausdorff distance (HD95). The best results are indicated in bold. The Dice scores marked with * are reported in [4], and their values are provided with a precision of three decimal places. All results are based on training for 150 epochs and 10 epochs to ensure a fair comparison. FLOPs (G) denote the number of floating-point operations (in billions). #Params (M) indicates the number of parameters (in millions). GPU Mem (G) denotes GPU memory usage (in gigabytes). Inference (S) represents the inference time (in seconds).
Model150 Epochs10 Epochs Inference (S)
DSC↑HD↓DSC↑HD↓FLOPs (G)#Params (M)GPU Mem (G)
Unet [26]0.88975.940.84207.6221.377.855.93120
UNet++ [51]0.90813.81 10.88944.8253.19.1611.39134
TransUNet [26]0.92932.390.93032.2058.49105.2810.57133
Swin-UNet [20]0.93272.040.92182.4917.441.3819.46165
TransNorm [29]0.92142.630.92322.4562.18117.6316.54130
HiFormer-B [52,53]0.89674.400.87845.348.04525.5111.40141
CoTrFuse [28]0.92662.740.90653.4233.0756.1915.30166
RWKV-UNet [49]0.88664.620.79585.8557.44120.2410.96122
H-vmunet [54]0.88173.340.87464.391140.68.973.79167
Seg. U-KAN [50]0.89185.090.84096.8114.026.3515.01130
VM-UNet [55,56]0.90422.890.86305.307.5634.6218.97126
MicroSegNet [4]0.93412.230.939 *2.0258.49105.2810.96131
Ours0.94161.930.93802.0270.7293.1311.93202
Table 2. Quantitative comparison between our model and MicroSegNet when data augmentation is applied to MicroSegNet.
Table 2. Quantitative comparison between our model and MicroSegNet when data augmentation is applied to MicroSegNet.
MethodsDSC↑HD95↓
MicroSegNet with Augmentation (150 Epochs)0.93212.29
MicroSegNet with Augmentation (10 Epochs)0.93562.04
Ours0.94161.93
Table 3. Contribution of each component to the overall performance on the Micro-Ultrasound dataset. FLOPs (G) denote the number of floating-point operations (in billions). #Params (M) indicates the number of parameters (in millions).
Table 3. Contribution of each component to the overall performance on the Micro-Ultrasound dataset. FLOPs (G) denote the number of floating-point operations (in billions). #Params (M) indicates the number of parameters (in millions).
Mamba (VSSD)DSHyperGraphAugment.FLOPs (G)#Params (M)DSC↑HD↓
Baseline 42.466.7992.842.28
68.593.0593.032.19
68.593.0593.731.95
70.7493.1393.751.99
Proposed Model70.7293.1394.161.93
Table 4. Comparison of the model with and without HyperGraph on the Micro-Ultrasound Dataset.
Table 4. Comparison of the model with and without HyperGraph on the Micro-Ultrasound Dataset.
MethodsDSC↑HD95↓
Without HyperGraph93.961.95
With HyperGraph (Overall)94.161.93
Table 5. Ablation study on the model scale for the Swin Blocks.
Table 5. Ablation study on the model scale for the Swin Blocks.
MethodsDSC↑HD95↓
Swin-T94.161.93
Swin-S93.961.95
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

AL-Qurri, A.; Thaher, A.; Almekkawy, M.K. Enhanced Deep Neural Network for Prostate Segmentation in Micro-Ultrasound Images. Sensors 2025, 25, 6815. https://doi.org/10.3390/s25226815

AMA Style

AL-Qurri A, Thaher A, Almekkawy MK. Enhanced Deep Neural Network for Prostate Segmentation in Micro-Ultrasound Images. Sensors. 2025; 25(22):6815. https://doi.org/10.3390/s25226815

Chicago/Turabian Style

AL-Qurri, Ahmed, Asem Thaher, and Mohamed Khaled Almekkawy. 2025. "Enhanced Deep Neural Network for Prostate Segmentation in Micro-Ultrasound Images" Sensors 25, no. 22: 6815. https://doi.org/10.3390/s25226815

APA Style

AL-Qurri, A., Thaher, A., & Almekkawy, M. K. (2025). Enhanced Deep Neural Network for Prostate Segmentation in Micro-Ultrasound Images. Sensors, 25(22), 6815. https://doi.org/10.3390/s25226815

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop