Next Article in Journal
Physics-Guided Multi-Task Learning for Small-Sample Soft Sensing: Simultaneous Prediction of Kappa Number and Viscosity in Continuous Kraft Pulping
Previous Article in Journal
Dosimetry of a Thermoregulated TEM Cell for 5G 700 MHz and 3.5 GHz Band Frequencies for Bioelectromagnetic Investigations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

RPFeaNet: Rethinking Deep Progressive Prompt-Guided Feature Interaction Fusion Network for Medical Ultrasound Image Segmentation

1
School of Mechanics and Engineering Science, Peking University, Beijing 100871, China
2
Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2026, 26(8), 2394; https://doi.org/10.3390/s26082394
Submission received: 9 March 2026 / Revised: 10 April 2026 / Accepted: 11 April 2026 / Published: 14 April 2026
(This article belongs to the Section Biomedical Sensors)

Abstract

Although ultrasound image segmentation has advanced significantly with deep learning, existing methods still suffer from a lack of prior knowledge guidance, partly due to the low-contrast, speckle-noise-corrupted nature from clinical ultrasound sensors. This paper proposes a novel ultrasound segmentation framework (RPFeaNet) that extracts progressive prompts from a low-to-high level prompt generation mechanism. Furthermore, the high-level prompt-guided feature interaction module (HPGFIM) fuses progressive prompt via Mamba blocks and stage-wise condition injection. The dynamic selective-frequency decoder (DSFD) combines dynamically selecting a strategy with the fusion of high-frequency details to suppress noise and refine edge details. Extensive experiments on six datasets demonstrate that RPFeaNet achieves state-of-the-art performance compared to existing methods, validating its strong generalization and robustness across diverse clinical ultrasound scenarios.

1. Introduction

Medical imaging sensors, particularly medical ultrasound sensors, have been increasingly advanced by the rapid development of image segmentation [1] and deep learning [2], effectively improving the diagnostic value of data acquired by medical ultrasound sensors. Ultrasound image segmentation has emerged as a pivotal technology in medical image processing, leveraging automatically segmented boundary information for the diagnosis of various organs and tissues such as carotid plaques [3], carotid arteries [4], breast lesions [5], and nerves [6], among others. Through image segmentation, sensors can more effectively identify regions of interest, and this capability can be significantly enhanced to achieve superior accuracy and reliability in clinical diagnosis. However, ultrasound images are inherently degraded by speckle noise [7], low contrast, and artifacts such as acoustic shadows, which lead to ambiguous lesion boundaries and greatly increase segmentation difficulty [8]. Regions with drastic grayscale changes, such as edges and shadowed areas, are particularly prone to segmentation errors. Unlike X-ray or magnetic resonance imaging, ultrasound imaging is based on the reflection and scattering of ultrasonic waves. This distinct imaging mechanism exacerbates the aforementioned issues, making it harder to distinguish tissue boundaries. Additionally, the lack of large-scale, high-quality annotated data [9] further hinders the development of accurate segmentation models. Thus, these factors make ultrasound image segmentation a highly challenging task in medical image analysis.
Many existing deep learning-based methods have been adopted for various ultrasound image segmentation tasks. Various network architectures, including convolutional neural networks (CNNs) [10], vision Transformers (ViTs) [11], visual state space models (VSS) [12], generative adversarial networks (GANs) [13], and other advanced designs have been applied to ultrasound image segmentation. However, these representative architectures only focus on general principles of feature extraction rather than the inherent characteristics of ultrasound medical images themselves. Fully connected networks, such as UNet, and their various variants [14] have achieved remarkable success in ultrasound image segmentation through feature-level or attention-level improvements. Swin Transformers [15] have been applied to breast tumor ultrasound images due to their ability to model long-range dependencies. Deepak et al. [16] proposed using an attention mechanism to handle broken boundaries, which are common in ultrasound images. Lyu et al. [17] exploited spatial and channel attention to help the model focus on edge information. While these methods follow the methodology of advanced feature learning, they mainly rely on generic deep learning paradigms (e.g., multi-scale feature fusion [18], self-attention aggregation [19]) rather than integrating domain-specific knowledge of ultrasound images. Specifically, such methods are usually derived from statistical characteristics of training data or heuristic structural designs, lacking the ability to adapt to the inherent properties of ultrasound images.
In order to discover the inherent guidance of medical images, increasingly more works have adopted the prior-guided prompt network for medical image segmentation tasks. Prompt–domain representation learning has been incorporated into not only convolutional neural networks (CNNs) but also Transformers, serving as a core module to enhance feature discriminability in ultrasound image segmentation. However, simple text-driven prompt methods using contrast language–image pretraining (CLIP) have actually exacerbated the domain gap, especially in ultrasound image. Single text guidance has been applied into ultrasound image segmentation; Chen et al. [20] used text prompt-guided multiple organ segmentation, and Spiegler et al. [21] explored the text-driven adaptation of segment anything model (SAM) without manual geometric prompts. Gowda et al. [22] designed different adapters used to refine the representation of low-contrast ultrasound images. To enhance boundary perception, Zhang et al. [23] used the DINOv3 pretrained foundation model with multi-scale frequency extraction and alignment to enrich spatial coherence. Despite the progress of prompt–domain knowledge (edge, boundary, spatial information, etc.) in the current model, these methods rely on explicit modeling manners by fusing conventional network architectures. However, pervasive speckle noise, low contrast, and shadow artifacts often cause target boundaries to become blurred, broken, or even indistinguishable from the background in ultrasound imaging. Thus, their predictions tend to be unstable for low-saliency objects, especially small targets and weak boundaries. This raises a critical question to discuss: Manually designed prompts struggle to effectively guide and optimize feature extraction. This inherent limitation has driven the emergence of numerous automatic prompt generation methods for medical image segmentation [21], aiming to enhance the model’s ability to capture high-level semantic features of objects. Without explicit prompts, they [24,25] generated automatic prompts for ultrasound image segmentation. Various prompts [26] have been proposed for object segmentation refinement. However, they lack a structured way to integrate multi-level prior information, leading to unstable performance on diverse ultrasound datasets.
To address the above critical issue, we aim to design a method that can enrich multi-level spatial prior information for low-contrast ultrasound images by a progressive prompt-driven perspective shown in Figure 1. We design a progressive prompt generation module to capture low-to-high level information for ultrasound images. This type of progressive prompt is expected to guide the model to gradually understand the characteristics of low-contrast ultrasound images, rather than relying on simple text descriptions. Compared with the existing works, it is not a simple combination of multiple discrete prompts, but a logically hierarchical elevation that mines multi-level spatial priors from the SAM2 Hiera [27], patch-level encoder and CLIP in a low-to-high progressive manner for ultrasound images. Low-level structural priors from the frozen CLIP encoder, middle-level dense priors from the patch-level encoder, and high-level semantic priors from SAM2’s hierarchical feature encoding. This forms a coherent prior system for ultrasound images instead of a disjointed prompt ensemble. Furthermore, the high-level prompt-guided feature interaction module (HPGFIM) is designed for guiding the model to focus on the essential shape and texture feature. To this end, we design this module to take progressive prompts as explicit conditioning signals and implement high-level stage-aware guidance via conditioning injection. This integration tailors Mamba/SS2D for capturing the shape and texture features of ultrasound targets, rather than merely performing a naive feature concatenation of prompts and backbone features. Finally, to bridge the gap between an encoder and decoder, the dynamic selective-frequency decoder is proposed based on selective-frequency decomposition and a multi-level frequency gating mechanism to enhance the ability of the spatial-frequency conversion pattern between high-level semantic features and low-level details. Overall, the main contributions can be summarized as follows:
  • The progressive prompt generation module is used to effectively enrich multi-level spatial prior information for low-contrast ultrasound images from a progressive prompt-driven perspective.
  • The high-level prompt-guided feature interaction fusion module is designed to achieve progressive prompt interaction fusion via visual mamba network and stage-wise conditioning injection.
  • The extensive experiments on six ultrasound image benchmark datasets demonstrate that our proposed method, RPFeaNet, achieves superior performance over comparative state-of-the-art methods.
The remaining sections of this paper are organized as follows. Section 2 reviews relevant research in vision Transformer-based ultrasound image segmentation and prompt-based ultrasound image segmentation. Section 3 describes the proposed RPFeaNet framework in detail. Section 4 presents the experimental setup, evaluation metrics, and comparative results against state-of-the-art approaches, followed by ablation studies and qualitative analyses. Finally, Section 5 concludes this paper and discusses future work.

2. Related Works

2.1. CNN/Transformer for Ultrasound Image Segmentation

Early attempts to introduce Transformers into ultrasound image segmentation primarily focused on integrating Transformer modules into convolutional neural network (CNN) architectures to compensate for the limited receptive field of convolution operations. Chen et al. [28] were among the first to demonstrate that incorporating Transformer components into CNN backbones can effectively enhance global context modeling, leading to improved structural consistency compared with purely convolutional approaches. Following this line of research, more structured hybrid CNN–Transformer architectures were proposed to better exploit the complementary strengths of both paradigms. He et al. [29] introduced Transformer encoder blocks (TEBlocks) into the encoder to explicitly model long-range dependencies, while interleaving convolutional layers and Transformer modules in a hybrid stem. Subsequent works further refined hybrid designs by enhancing feature representation and fusion strategies. For example, HAU-Net [30] embedded Transformer blocks into U-shaped skip connections to strengthen cross-layer semantic interaction. While the aforementioned hybrid CNN–Transformer methods alleviate some limitations of purely convolutional networks, they still rely heavily on convolutional structures for feature extraction. To further explore the potential of Transformer-based models in vision tasks, researchers began to investigate vision Transformers as standalone backbones. However, directly adapting Transformers from natural language processing to vision tasks poses significant challenges, mainly due to the high spatial resolution of images and large-scale variations of visual entities.
Motivated by the success of the Swin Transformer in general vision tasks, subsequent studies extended hierarchical Transformers to medical image segmentation. In particular, Swin-Unet [31] introduced a pure Transformer-based U-shaped encoder–decoder architecture that replaces convolutional operations with hierarchical Swin Transformer blocks. In ultrasound image segmentation, hierarchical Transformers were further combined with CNNs to enhance both local detail preservation and global semantic modeling. Yang et al. [32] extended this idea by introducing a CNN–Swin pyramid architecture to better capture multi-scale lesion characteristics. More recent studies have explored parallel dual-branch architectures, where CNN and Swin Transformer encoders operate concurrently. These developments mark a transition from sequential hybrid designs to parallel and hierarchical CNN–Transformer frameworks in ultrasound segmentation.

2.2. Prompt-Guided Ultrasound Image Segmentation

While text-based prompting technology is common in medical image segmentation tasks, the lack of image/label and anatomical priors degrades the performance of prompt-based methods when applied to ultrasound image segmentation. Ref. [21] explored the text-driven adaptation of the segment anything model (SAM) without the need for manual geometric prompts. Gowda et al. [22] designed different adapters to refine the representation of low-contrast ultrasound images. Without explicit prompts, Yin et al. [24] generated automatic prompts for breast lesion segmentation. Tian et al. [33] proposed a self-prompting mechanism to capture temporal features for carotid artery diagnosis. Zhao et al. [34] avoided reliance on manual prior guidance and proposed an uncertainty-driven edge prompt to enrich medical image features. However, all the aforementioned existing works relied only on a single prompt mechanism to enhance the model’s learning ability. A single prior is insufficient to overcome the limitations of low-contrast ultrasound images. Thus, we further address this gap by adopting a progressive prompt mechanism to further boost the learning capability for ultrasound image segmentation. The Table 1 shows the comparison of RPFeaNet with representative ultrasound image methods from different perspectives.

3. The Proposed RPFeaNet

3.1. The Overview of RPFeaNet

Given an ultrasound image I R H × W × 3 , we propose RPFeaNet, a progressive prompt-guided feature interaction fusion segmentation network designed to mitigate speckle noise, low contrast, and boundary ambiguity. The key idea is to (1) generate reliable progressive prompts from global-to-local perspectives, (2) perform high-level prompt-guided interaction and fusion on multi-scale features, and (3) decode features by dynamic selective-frequency modeling to recover sharp boundaries. As shown in Figure 1, RPFeaNet follows an encoder–decoder paradigm and consists of three major components: (i) the progressive prompt generation module (PPGM), (ii) the high-level prompt-guided feature interaction module (HPGFIM), and (iii) the dynamic selective-frequency decoder (DSFD). Specifically, the progressive prompt generation module (PPGM) is elaborated in Section 3.2, which serves as the core component for generating high-quality dense prompts. PPGM comprises three collaborative branches, namely the pretrained prompt teacher branch (PPTB), the prior prompt synthesis branch (PPSB), and the lightweight feature extractor (LFE), corresponding to low-level, middle-level, and high-level knowledge, respectively. To clarify their relationships, PPTB provides stable structural priors via a frozen Transformer backbone, acting as foundational low-level guidance. PPSB builds upon this prior by synthesizing patch-level cues and dense embeddings, forming the middle-level knowledge. Subsequently, LFE extracts high-level semantic representations from a lightweight encoder. These three branches are not independent but interact through a progressive fusion strategy, in which knowledge is gradually integrated from low-level to high-level representations. The final fused dense prompts generated by PPGM are then fed into the high-level prompt-guided feature interaction module (HPGFIM), where they guide the interaction of multi-scale features to further enhance segmentation performance.
These PPTB, PPSB, LFE branches in PPGM enable a coarse-to-fine prompting strategy that significantly enhances model generalization across low-contrast datasets. Here, we describe a data flow of the overall pipeline.
  • PPTB: Given input I, a frozen Transformer backbone outputs stable structural prompts P tea .
  • PPSB: Based on I, patch embeddings are synthesized into dense prompts P dense .
  • LFE: A lightweight encoder extracts semantic features X;
The three types of knowledge are integrated as P fusion = Fuse ( P tea , P dense , X ) .
The final fused dense prompts P fusion generated by PPGM are then fed into the high-level prompt-guided feature interaction module (HPGFIM), where they guide the interaction between multi-scale features:
F enhanced = HPGFIM ( X , P fusion ) ,
to enhance segmentation performance. The details can be found in the following sections.

3.2. Progressive Prompt Generation Module (PPGM)

Unlike natural or general medical imaging, feature extraction from ultrasound images is fundamentally constrained by intrinsic low contrast and speckle noise. Consequently, effective representation learning for ultrasound image must go beyond standard visual encoding. From our insights, relying solely on pixel representation does not facilitate subsequent tasks; it must actively compensate for the structural degradation caused by low contrast. To address this challenge, we introduce a progressive prompting strategy, which refines structural priors to overcome the inherent limitations of medical ultrasound images.
We propose the progressive prompt generation module (PPGM), designed to generate high-fidelity dense prompts for medical image segmentation without requiring manual intervention.
  • Lightweight Feature Extractor (LFE): A lightweight encoder produces the knowledge from a pretrained network termed “High-Level Knowledge”.
  • Prior Prompt Synthesis Branch (PPSB): An encoder–generator pipeline that synthesizes patch-level cues and dense embedding termed “Middle-Level Knowledge”.
  • Pretrained Prompt Teacher Branch (PPTB): A frozen Transformer backbone that utilizes trainable prompt tokens to provide stable representation evolution and strong structural priors, termed “Low-Level Knowledge”.

3.2.1. Lightweight Feature Extractor (LFE)

Given an input image I R H × W × 3 , we adopt a hierarchical visual Transformer encoder called Hiera [27] that outputs a four-stage feature pyramid:
{ X 1 , X 2 , X 3 , X 4 } = Φ enc ( I ) ,
where X k R C k × H 2 k × W 2 k denotes the feature map at the k-th stage (with progressively reduced spatial resolution and enriched global context).
To enable parameter-efficient domain adaptation, we freeze the pretrained encoder trunk and insert lightweight adapters before each block:
X ^ k = X k + A k ( X k ) , k { 1 , 2 , 3 , 4 } ,
where A k ( · ) denotes a stage-wise adapter. Since the hierarchical encoder contains a substantial number of parameters, we freeze the encoder trunk and insert adapters before each multi-scale block for parameter-efficient fine-tuning. Each adapter is implemented as a two-layer MLP with GeLU nonlinearity:
A k ( X ) = GeLU W ( k ) GeLU W ( k ) X ,
where W ( k ) and W ( k ) are the down-projection and up-projection linear mappings, respectively.

3.2.2. Pretrained Prompt Teacher Branch (PPTB)

To provide stable structural guidance and high-level semantic priors, we introduce the pretrained prompt teacher branch (PPTB), which utilizes a fine-tuned enhanced CLIP visual encoder [36]. While standard foundation models are pretrained on natural images, their direct application to ultrasound (US) imaging is hindered by the domain gap and low-contrast characteristics. We keep the CLIP-ViT backbone Φ C L I P frozen and insert a sequence of learnable visual prompts P = { p i } i = 1 N to bridge this gap.
The PPTB processes it through the prompt-tuned CLIP encoder to generate a stable teacher representation P t e a . The interaction within the Transformer layers is defined as follows:
P t e a = Φ CLIP ( [ x i l 1 , P l 1 ] )
where T l denotes the l-th Transformer layer, and x i represents the image features by optimizing only the prompt tokens P while keeping the backbone weights fixed.

3.2.3. Prior Prompt Synthesis Branch (PPSB)

The prior prompt synthesis branch (PPSB) refines images into actionable prompts in a coarse-to-fine manner. It first processes initial patch-level cues and progressively transforms them into dense prompt embeddings aligned with the latent image space, enabling robust prompt learning without the need for manual point or box annotations. The details are presents as follows.
x i = PatchEmbed ( I ) R N × D ,
where N denotes the number of patches and D is the embedding dimension.
The patch embedding is processed by a hybrid backbone with interleaved convolutional and Transformer components. Specifically, local spatial features are extracted through convolutional layers, normalization and linear projection. Moreover, a multi-head self-attention (MSA) module is subsequently applied:
h = MSA ( Linear LayerNorm ( Conv 2 Conv 1 ( x i ) ) ) ,
where the resulting embedding h simultaneously encodes fine-grained local details and global contextual information.
Specifically, within the given image embedding h , the function δ ( · ) performs patch merging to produce patch-level logits p ^ p a t :
p ^ p a t = δ ( h )
δ ( · ) is implemented using lightweight 2 × 2 stride-2 convolutions with a batch normalization layer, followed by a 1 × 1 convolution for channel compression. This stage provides a coarse but robust prompt prior. Furthermore, to obtain dense prompts, we convert patch logits p ^ p a t to probability function σ ( · ) and progressively lift them to the feature resolution:
p p a t = σ ( p ^ p a t )
p p a t = Interpolate ( p p a t )
P d e n s e = ψ ( p p a t )
where P d e n s e denotes the feature of PPSB, and ψ ( · ) denotes the prompt encoder, implemented by interpolation (for memory efficiency) and 1 × 1 convolutions to align the prompt embedding dimension to that of h .

3.3. High-Level Prompt-Guided Feature Interaction Module (HPGFIM)

In this way, PPGM progressively synthesizes dense prompt embedding P d e n s e , which can be directly fused with P t e a and injected into subsequent HPGFIM. Given the lightweight feature extractor defined by Equation (1) along with its output X i , and the progressive prompt ( F d e n s e , F t e a ) , HPGFIM aligns them to a shared embedding space and performs prompt-guided interaction fusion based on 2d selective space (SS2D) [37]. Different from common strategies in which existing works treat X i as the main stream and treat other prompts as guidance, without concatenation or attention, we use X i as an explicit conditioning injection throughout the interaction block, repeatedly modulating both F d e n s e and F t e a prompts before and after SS2D, as well as during residual refinement.
Let C i denote the transformed conditioning feature from X i . We first inject C i into both branches before the SS2D interaction:
U = F d e n s e + C i
V = F t e a + C i .
Then, we apply a bidirectional SS2D operator where the scan parameters of one branch are generated from the other branch, realizing a prompt-guided cross interaction:
Y d e n s e = SS 2 D U ; θ ( V )
Y t e a = SS 2 D V ; θ ( U )
where θ ( · ) denotes the SS2D parameterization function (e.g., producing Δ , A , B , C , D used by a selective scan).
After SS2D, C i is injected again to reinforce stage-aware conditioning:
Y d e n s e Y d e n s e + C i
Y t e a Y t e a + C i
We compute branch-specific gates to modulate the interacted responses:
g d e n s e = σ SiLU Linear F d e n s e
g t e a = σ SiLU Linear F t e a
and we apply gating to obtain the outputs of the interaction core:
O d e n s e = Linear Y d e n s e g d e n s e
O t e a = Linear Y t e a g t e a
Finally, we apply residual updates and an MLP refinement, where C i is injected multiple times to stabilize the fusion:
H d e n s e = O d e n s e + F d e n s e
H t e a = O t e a + F t e a
F d e n s e = H d e n s e + C i + MLP LN ( H d e n s e + C i ) + C i
F t e a = H t e a + C i + MLP LN ( H t e a + C i ) + C i
The refined features F d e n s e and F t e a are then concatenated and projected to form the fused representation F for decoding.

3.4. Dynamic Selective-Frequency Decoder (DSFD)

Accurate segmentation requires both low-frequency global consistency (shape/region) and high-frequency boundary precision. Static frequency operations may fail under varying noise levels and target morphologies. We thus design DSFD to dynamically select frequency transformations via a frequency-decomposed gating mechanism and explicitly exploit the complementarity between low and high frequencies.
Let X i R C × H i × W i be the input to DSD i . We first normalize and reshape it into spatial tokens T i R N i × C with N i = H i W i , and we then transform the tokens into the frequency domain as shown in Figure 2:
T ¯ i = Norm ( T i )
F i = F ( T ¯ i )
where F ( · ) denotes the FFT operation applied along the token axis. As illustrated, the frequency tokens F i are fed into K parallel MLP branches:
Z i ( k ) = ϕ k ( F i ) , k = 1 , , K
where each ϕ k ( · ) represents an MLP operating in the frequency domain, providing diverse learnable frequency transformations. A lightweight gating network takes the same frequency representation F i and generates selection weights over all branches:
α i = softmax g ( F i )
Each branch output is modulated by its corresponding gate weight (the ⊗ blocks in the figure) and then aggregated via the ⊕ node:
F ˜ i = k = 1 K α i ( k ) Z i ( k )
The aggregated frequency tokens are converted back to the token domain via IFFT and reshaped into final feature maps:
T ˜ i = F 1 ( F ˜ i )
R i = reshape 1 ( T ˜ i )
After obtaining R 1 , R 2 , R 3 , R 4 , we align R 2 , R 3 , R 4 to the spatial size of R 1 by interpolation and concatenate them along the channel dimension to form the middle representation:
M res = Concat Up ( R 2 ) , Up ( R 3 ) , Up ( R 4 ) ,
F = Fuse R 1 , M res .
We then combine middleres with the finest-scale decoded feature R 1 to obtain the final decoded feature F. Finally, DSFD predicts a mask and upsamples it to the original resolution:
Y ^ = Up SegHead ( F )

3.5. Training Objective

The overall training objective consists of a standard segmentation loss, including IoU loss and BCE loss.
L = L I o U + λ L B C E ,
where λ balances loss consistency. The IoU increases the contribution of pixels that are difficult to predict by assigning them larger weights, encouraging the model to concentrate on structurally important regions in challenging cases. In addition, the BCE provides complementary pixel-wise supervision with the same weighting principle. Together, these two terms promote both region-level overlap quality and fine-grained discrimination.

4. Experiment

4.1. Experiment Settings

The setting of pretrained models of the LFE and PPTB module we use are SAM2 Hiera [27] and [38], respectively. All the experiments are conducted in Ubuntu 20.04, NVIDIA 4090 GPU, Pytorch 2.1. Overall, six benchmark datasets are used for evaluations, including common carotid artery ultrasound images (CCAUI) [39], 2D Cardiac ultrasound dataset (CAMUS) [40], thyroid ultrasound dataset (DDTI) [41], fetal head ultrasound dataset (HC-18) [42], fetal head–pubic symphysis (JNU-LFM) [43], and thyroid nodule image dataset (TN3K) [44]. To comprehensively evaluate the effectiveness of our method, all experiments employ 5-fold cross-validation. To ensure experimental integrity, we re-run all comparative experiments in the same experimental environment and report them as mean ± standard deviation (SD). Code is available at https://github.com/Ezrealone/RPFeaNet (accessed on 2 March 2026). In detail, all images among the dataset we use are resized into 256 × 256. For transformation, we only use vertical and horizontal flips. The number of epochs is 70, and batch size is 12. The AdamW optimizer is selected with 1 × 10 4 . λ used in the loss function is 0.1, and a IoU loss is weighted by a focal parameter of 2. These basic settings are uniformly applied to all comparative methods. To verify the statistical significance of the performance improvement of our method over comparative baselines, paired t-tests are conducted based on the results of 5-fold cross-validation. The null hypothesis is that there is no significant difference between the performance of our proposed RPFeaNet method and the comparison method. A p-value of less than 0.05 (p < 0.05) is considered statistically significant, which is marked with an asterisk (∗) in the table to indicate that the baseline method is significantly inferior to our proposed method.

4.2. Comparative Methods

Extensive experiments are conducted on these six benchmark datasets to validate the effectiveness of the proposed RPFeaNet method. To comprehensively evaluate its segmentation performance, we compare our approach with nine state-of-the-art (SOTA) ultrasound image segmentation methods.
  • U-Net [45] enabled accurate pixel-wise segmentation by combining contextual information and precise localization via multi-scale skip connections.
  • UNet++ [46] designed a method for medical image semantic and instance segmentation based on redesigned skip connections and multi-depth network ensemble.
  • AttUNet [47] boosted medical image segmentation accuracy by suppressing irrelevant background features and highlighting salient target organ features.
  • nnU-Net [48] integrated fixed parameters, interdependent rules and empirical decisions to realize automated preprocessing, network construction, training and post-processing.
  • Swin-Unet [49] integrated Swin Transformer blocks and symmetric encoder-decoder structure to model global contextual and multi-scale hierarchical features.
  • TransFuse [50] captured global dependencies and low-level spatial details via parallel CNN and Transformer branches.
  • H2Former [51] integrated CNNs, multi-scale channel attention and Transformers to capture long-range dependencies, fuse multi-scale features and ensure computational efficiency for medical image segmentation.
  • ScribFormer [52] enabled high-performance scribble-based medical image segmentation by capturing global shape information and fusing local and global features via a triple-branch structure comprising a CNN, Transformer and attention-guided class activation map.
  • LGFFM [35] for ultrasound image segmentation integrated a Parallel Bi-Encoder, frequency-domain mapping module and multi-domain fusion to capture local–global and frequency-domain features, addressing low resolution, noise and poor generalization of existing methods.
To comprehensively evaluate the segmentation performance of the proposed method on the seven evaluation metrics, including Jaccard similarity coefficient (Jaccard), Dice similarity coefficient (Dice), weighted F-measure [53], S-measure [54], mean E-measure [55], average symmetric surface distance (ASSD), and mean square error (MSE), they were adopted to quantify region overlap, structural similarity, boundary proximity, and pixel-wise consistency from multiple perspectives. Notably, F β w represents the weighted F-measure, S α represents the S-measure, and E ϕ represents the mean E-measure.

4.3. Result Analysis

4.3.1. The Results of CCAUI and CAMUS

To evaluate the superiority of our proposed method RPFeaNet with states-of-the-art methods, the results of the CCAUI and CAMUS datasets can be found in Table 2. The visualized result can be found in Figure 3, which clearly demonstrates that our proposed method achieves more accurate and visually smooth segmentation contours on the CAMUS dataset with three classes. Both datasets focus on cardiac ultrasound imaging, where the segmented targets exhibit similar statistical shape and size characteristics. Quantitative results clearly demonstrate that our proposed RPFeaNet achieves an average performance gain of 10% across the Jaccard, Dice, ASSD and MSE metrics.
In the CCAUI dataset, the model delivers a substantial improvement over the state-of-the-art LGFFM and other Transformer-based methods with the Jaccard and Dice coefficients rising by 16.7% and 8.3%, respectively, the ASSD metric dropping by 58.9%, and the MSE metric decreasing by 62.2%. This significant enhancement fully validates the model’s superior ability in capturing target regions and fitting boundary details for ultrasound images with severe noise and blurred edges. Notably, RPFeaNet maintains the smallest standard deviation (SD) across all metrics (Jaccard SD = 0.0062), demonstrating superior stability in handling noisy ultrasound data. This significant enhancement fully validates the model’s ability to accurately capture target regions and fit boundary details under adverse imaging conditions. For the CAMUS dataset, even the basic U-Net model can achieve a Dice coefficient of 91%. Our RPFeaNet achieves steady and effective optimization. It outperforms the second one with slight increases in the Jaccard and Dice coefficients, and achieves the optimal value of the mean E-measure metric among all comparison methods. RPFeaNet maintains competitive performance with significantly higher core segmentation metrics. The model’s ability to break through the performance benefits from its targeted architectural design for the inherent characteristics of ultrasound images. The progressive prompt generation module provides hierarchical spatial prior guidance, making up for the deficiency of traditional models that rely solely on pixel-level supervision. The high-level prompt-guided feature interaction module strengthens the capture of target shape and texture features through cross-interaction fusion. The dynamic selective-frequency decoder achieves precise separation and fusion of high- and low-frequency information, effectively improving the accuracy of boundary segmentation.

4.3.2. The Results of DDTI and TN3K

The DDTI and TN3K datasets are analyzed together as they are both thyroid ultrasound image benchmarks. Specifically, DDTI focuses on larger targets, while TN3K targets small nodules. TN3K presents greater challenges due to its more blurred nodule boundaries, making it a critical benchmark for verifying the robustness of segmentation models. Quantitative results demonstrate that our proposed RPFeaNet maintains a consistent and significant performance advantage over state-of-the-art comparison methods on both thyroid ultrasound datasets, as shown in Table 3 and Figure 4.
On the DDTI dataset, where most comparison methods have already achieved relatively satisfactory segmentation performance, RPFeaNet still achieves steady and refined optimization compared with the state-of-the-art (SOTA) method. The Jaccard and Dice coefficients are slightly improved, the ASSD metric is reduced by 11.0% (from 0.4760 to 0.4235), and the MSE metric drops by 2.8% (from 0.0461 to 0.0448). The segmentation results on the TN3K dataset more prominently reflect the core advantages of RPFeaNet: it achieves an all-round performance breakthrough compared with other methods. The Jaccard and Dice coefficients are increased by 1.4% and 1.2%, respectively, maintaining the stability of regional segmentation in complex scenarios. The ASSD metric is drastically reduced by 21.1% (from 2.4380 to 1.9237). This fully demonstrates that the dynamic selective-frequency decoder of RPFeaNet can accurately capture the high-frequency information of blurred small nodule boundaries and effectively suppress the interference of ultrasound noise on boundary segmentation.

4.3.3. The Results of HC-18 and JNU-IFM

The HC-18 and JNU-IFM datasets are both designed for fetal ultrasound image segmentation, suffering from severe ultrasound artifacts, low contrast between adjacent tissues, and irregular target boundaries, leading to significantly increased segmentation difficulty. Quantitative results in Table 4 and visualization results in Figure 5 show that our proposed RPFeaNet achieves consistent and outstanding performance advantages over all comparison methods on both fetal ultrasound datasets.
On the HC-18 dataset, most comparison methods already achieve high overall segmentation performance due to the regular fetal head shape. For Jaccard and Dice coefficients, RPFeaNet outperforms others by 0.09 to 0.1. The S α and E ϕ metrics are also slightly improved, and the MSE metric is optimized to 0.0458, maintaining competitive segmentation accuracy. Notably, the ASSD metric of RPFeaNet is only slightly higher than that of other methods. The segmentation results on the JNU-IFM dataset demonstrate the core advantages of RPFeaNet in handling complex fetal ultrasound structures. While all comparison methods experience significant performance degradation on these datasets, RPFeaNet achieves comprehensive improvements. The Jaccard and Dice coefficients are increased by 5.4% and 3.6%, respectively, representing a substantial leap in regional segmentation accuracy for complex anatomical structures. The ASSD metric also shows remarkable improvement. This fully demonstrates that the progressive prompt generation module of RPFeaNet can effectively capture spatial prior information of complex fetal head–pubic symphysis structures via progressive guidance. Furthermore, the S α metric of RPFeaNet reaches 0.8947, and the MSE metric is optimized to 0.0527, achieving the best performance among all comparison methods. These results fully validate the model’s superior capability in preserving structural similarity and reducing pixel-level errors for complex fetal ultrasound segmentation.

4.3.4. Complexity Analysis

To provide a more comprehensive assessment of our method, we perform a complexity analysis, as summarized in Table 5. The table reports key metrics, including inference speed (frames per second, FPS), number of parameters (M), and floating-point operations (FLOPs). Compared with lightweight Transformer-based models, our FPS is relatively lower, which is reasonable since our framework used SS2D blocks in HPGFIM rely on Mamba’s selective scan, which involves sequential feature processing (less parallelizable than CNN convolutions/Transformer attention). Our RPFeaNet maintains competitive FPS, indicating strong practical applicability. This advantage primarily comes from the adopted lightweight architectural design. In addition, our model achieves balanced computational efficiency, making it suitable for real-world deployment.

4.4. Ablation Studies

To thoroughly validate the contribution of each component in RPFeaNet, we conducted ablation studies on the CCAUI and CAMUS datasets. We establish a baseline model that utilizes the same backbone and segmentation head but replaces our proposed modules with a simple U-Net decoder. Based on this, we progressively incorporated the progressive prompt generation module (PPGM), high-level prompt-guided feature interaction module (HPGFIM), and dynamic selective-frequency decoder (DSFD) to analyze their individual and combined impacts compared with our RPFeaNet.
The quantitative results are reported in Table 6 and Table 7. Comparing the first row, the inclusion of the PPGM yields a substantial improvement. This indicates that the progressive prompt guidance strategy effectively suppresses ultrasound-specific noise and compacts the representation into cleaner feature manifolds, providing a more robust initial feature basis for the decoder. The second row further boosts performance. This significant gain validates that the prompt-guided feature interaction mechanism successfully mitigates the deficiency of insufficient feature discriminability in ultrasound images by dynamically recalibrating the weights of multi-level prompts and backbone features, ensuring that complementary spatial information is fully exploited. The inclusion of the DSFD refines the segmentation results in last row, reaching an optimal Dice value of 93.0% on CCAUI and 95.7% on CAMUS. This demonstrates that the our final model precisely delineating the target boundaries in ultrasound images.
Furthermore, we provide the ablation of sub-branches within core module. The ablation results in Table 8 clearly demonstrate the effectiveness of each sub-branch (LEF, PPTB, PPSB) and their effects on the CCAUI dataset. Specifically, among the single sub-branch configurations, the PPTB sub-branch achieves the best performance with a Dice coefficient of 0.9083 and an ASSD of 0.2089, outperforming LEF (Dice: 0.9015, ASSD: 0.2217) and PPSB (Dice: 0.8957, ASSD: 0.2356) individually. When combining two sub-branches, the PPTB + PPSB combination yields better results than LEF + PPTB and LEF + PPSB, indicating a stronger complementary effect between PPTB and PPSB. Notably, the full configuration integrating all three sub-branches (LEF + PPTB + PPSB) achieves the optimal performance across all metrics, with the highest Dice (0.9300) and Jaccard (0.8783) coefficients, as well as the lowest ASSD (0.1507). These results confirm that each sub-branch contributes positively to the model’s performance.

4.5. Limitations

Here, we discuss the inherent limitations of our work. In terms of generalization, while the model is trained and validated on mainstream ultrasound modalities (including carotid, cardiac, fetal, and thyroid imaging), its performance remains insufficiently verified for niche clinical use cases and underrepresented imaging scenarios. Additionally, the fixed 256 × 256 input size requires resizing of clinical high-resolution ultrasound images, potentially losing fine-grained details of small lesions critical for clinical diagnosis. The Mamba-based SS2D blocks in HPGFIM use sequential selective scan operations, which are parameter-efficient but less parallelizable than CNN convolutions or Transformer self-attention. These limitations highlight critical directions for future work to enhance the model’s practical value for clinical deployment and real-world applicability.

5. Conclusions

This paper presents a medical ultrasound image segmentation framework (RPFeaNet). This model effectively mitigates the impact of ultrasound-specific noise and low contrast and significantly improves segmentation accuracy and boundary fitting by progressive prompt generation, high-level feature interaction fusion, and a dynamic frequency decoder. Experimental results on multiple datasets show that RPFeaNet outperforms existing SOTA methods. In future work, we will further apply the model for real clinical cases and extend it to multi-modal ultrasound image segmentation tasks.

Author Contributions

Conceptualization, L.Z. and Y.D.; methodology, L.Z. and Y.D.; software, L.Z. and Y.D.; validation, L.Z. and Y.D.; formal analysis, Y.D.; investigation, L.Z.; resources, L.Z.; data curation, L.Z. and Y.D.; writing—original draft preparation, L.Z. and Y.D.; writing—review and editing, L.Z. and Y.D.; visualization, L.Z. and Y.D.; supervision, Y.D.; project administration, L.Z.; funding acquisition, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under grant number 62474539.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The CCAUI dataset can be accessed at https://data.mendeley.com/datasets/d4xt63mgjm/1 (accessed on 24 December 2025). CAMUS can be downloaded from https://www.creatis.insa-lyon.fr/Challenge/camus/databases.html (accessed on 24 December 2025). TN3K can be downloaded from https://www.kaggle.com/datasets/tjahan/tn3k-thyroid-nodule-region-segmentation-dataset (accessed on 24 December 2025). DDTI can be downloaded from https://www.kaggle.com/datasets/dasmehdixtr/ddti-thyroid-ultrasound-images/data (accessed on 24 December 2025). HC-18 can be downloaded from https://hc18.grand-challenge.org (accessed on 24 December 2025). JNU-IFM can be downloaded from https://figshare.com/articles/dataset/JNU-IFM/14371652/2 (accessed on 24 December 2025).

Acknowledgments

The authors would like to thank the anonymous reviewers for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Xiao, X.; Zhang, J.; Shao, Y.; Liu, J.; Shi, K.; He, C.; Kong, D. Deep learning-based medical ultrasound image and video segmentation methods: Overview, frontiers, and challenges. Sensors 2025, 25, 2361. [Google Scholar] [CrossRef]
  2. Hussain, S.I.; Toscano, E. Optimized deep learning for mammography: Augmentation and tailored architectures. Information 2025, 16, 359. [Google Scholar] [CrossRef]
  3. Zhu, X.; Cheng, B.; Shen, Y.; Xia, B.; Yue, G. Boundary-Supplementary Network for Carotid Plaque Segmentation in Ultrasound Images. IEEE Signal Process. Lett. 2025, 32, 1765–1769. [Google Scholar] [CrossRef]
  4. Huang, Q.; Tian, H.; Jia, L.; Li, Z.; Zhou, Z. A review of deep learning segmentation methods for carotid artery ultrasound images. Neurocomputing 2023, 545, 126298. [Google Scholar] [CrossRef]
  5. Huang, Q.; Luo, Y.; Zhang, Q. Breast ultrasound image segmentation: A survey. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 493–507. [Google Scholar] [CrossRef] [PubMed]
  6. Horng, M.H.; Yang, C.W.; Sun, Y.N.; Yang, T.H. DeepNerve: A new convolutional neural network for the localization and segmentation of the median nerve in ultrasound image sequences. Ultrasound Med. Biol. 2020, 46, 2439–2452. [Google Scholar] [CrossRef] [PubMed]
  7. Li, H.; Xu, X.; Liu, Z.; Xia, Q.; Xia, M. Low-quality sensor data-based semi-supervised learning for medical image segmentation. Sensors 2024, 24, 7799. [Google Scholar] [CrossRef]
  8. Ansari, M.Y.; Mangalote, I.A.C.; Meher, P.K.; Aboumarzouk, O.; Al-Ansari, A.; Halabi, O.; Dakua, S.P. Advancements in deep learning for B-mode ultrasound segmentation: A comprehensive review. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 2126–2149. [Google Scholar] [CrossRef]
  9. Meiburger, K.M.; Acharya, U.R.; Molinari, F. Automated localization and segmentation techniques for B-mode ultrasound images: A review. Comput. Biol. Med. 2018, 92, 210–235. [Google Scholar] [CrossRef]
  10. Ma, J.; Wu, F.; Jiang, T.; Zhao, Q.; Kong, D. Ultrasound image-based thyroid nodule automatic segmentation using convolutional neural networks. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 1895–1910. [Google Scholar] [CrossRef]
  11. Jiang, T.; Xing, W.; Yu, M.; Ta, D. A hybrid enhanced attention transformer network for medical ultrasound image segmentation. Biomed. Signal Process. Control 2023, 86, 105329. [Google Scholar] [CrossRef]
  12. Xiong, Y.; Shu, X.; Liu, Q.; Yuan, D. HCMNet: A Hybrid CNN-Mamba Network for Breast Ultrasound Segmentation for Consumer Assisted Diagnosis. IEEE Trans. Consum. Electron. 2025, 71, 8045–8054. [Google Scholar] [CrossRef]
  13. Zhai, D.; Hu, B.; Gong, X.; Zou, H.; Luo, J. ASS-GAN: Asymmetric semi-supervised GAN for breast ultrasound image segmentation. Neurocomputing 2022, 493, 204–216. [Google Scholar] [CrossRef]
  14. Chen, G.; Li, L.; Zhang, J.; Dai, Y. Rethinking the unpretentious U-net for medical ultrasound image segmentation. Pattern Recognit. 2023, 142, 109728. [Google Scholar] [CrossRef]
  15. Zhu, C.; Chai, X.; Xiao, Y.; Liu, X.; Zhang, R.; Yang, Z.; Wang, Z. Swin-Net: A swin-transformer-based network combing with multi-scale features for segmentation of breast tumor ultrasound images. Diagnostics 2024, 14, 269. [Google Scholar] [CrossRef]
  16. Mishra, D.; Chaudhury, S.; Sarkar, M.; Soin, A.S. Ultrasound image segmentation: A deeply supervised network with attention to boundaries. IEEE Trans. Biomed. Eng. 2018, 66, 1637–1648. [Google Scholar] [CrossRef]
  17. Lyu, Y.; Xu, Y.; Jiang, X.; Liu, J.; Zhao, X.; Zhu, X. AMS-PAN: Breast ultrasound image segmentation model combining attention mechanism and multi-scale features. Biomed. Signal Process. Control 2023, 81, 104425. [Google Scholar] [CrossRef]
  18. Ma, L.; Lin, X.; Liang, W.; Liu, W. EAMSHA-UNet: A breast cancer ultrasound image segmentation model based on edge-aware multi-scale hybrid attention. Biomed. Signal Process. Control 2026, 120, 110160. [Google Scholar] [CrossRef]
  19. Chen, R.; Zheng, X.; Su, H.; Wu, K. DeNAS-ViT: Data Efficient NAS-Optimized Vision Transformer for Ultrasound Image Segmentation. In Proceedings of the AAAI Conference on Artificial Intelligence; Association for the Advancement of Artificial Intelligence: Washington, DC, USA, 2026; Volume 40, pp. 3002–3010. [Google Scholar]
  20. Chen, H.; Cai, Y.; Wang, C.; Chen, L.; Zhang, B.; Han, H.; Guo, Y.; Ding, H.; Zhang, Q. Multi-Organ Foundation Model for Universal Ultrasound Image Segmentation With Task Prompt and Anatomical Prior. IEEE Trans. Med. Imaging 2025, 44, 1005–1018. [Google Scholar] [CrossRef] [PubMed]
  21. Spiegler, P.; Koleilat, T.; Harirpoush, A.; Miller, C.S.; Rivaz, H.; Kersten-Oertel, M.; Xiao, Y. Textsam-eus: Text prompt learning for sam to accurately segment pancreatic tumor in endoscopic ultrasound. In Proceedings of the IEEE/CVF International Conference on Computer Vision; IEEE: New York, NY, USA, 2025; pp. 948–957. [Google Scholar]
  22. Gowda, S.N.; Clifton, D.A. Cc-sam: Sam with cross-feature attention and context for ultrasound image segmentation. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2024; pp. 108–124. [Google Scholar]
  23. Zhang, Y.; Xu, Q.; Li, Y.; He, X.; Zhang, Q.; Haque, M.; Qu, R.; Duan, W.; Chen, Z. FreqDINO: Frequency-Guided Adaptation for Generalized Boundary-Aware Ultrasound Image Segmentation. arXiv 2025, arXiv:2512.11335. [Google Scholar]
  24. Yin, D.; Zheng, Q.; Chen, L.; Hu, Y.; Wang, Q. APG-SAM: Automatic prompt generation for SAM-based breast lesion segmentation with boundary-aware optimization. Expert Syst. Appl. 2025, 276, 127048. [Google Scholar] [CrossRef]
  25. Lin, X.; Xiang, Y.; Yu, L.; Yan, Z. Beyond adapting SAM: Towards end-to-end ultrasound image segmentation via auto prompting. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2024; pp. 24–34. [Google Scholar]
  26. Guo, A.; Fei, G.; Pasupuleti, H.; Wang, J. ClickSAM: Fine-tuning Segment Anything Model using click prompts for ultrasound image segmentation. In Proceedings of the Medical Imaging 2024: Ultrasonic Imaging and Tomography; SPIE: Bellingham, WA, USA, 2024; Volume 12932, pp. 240–244. [Google Scholar]
  27. Ryali, C.; Hu, Y.T.; Bolya, D.; Wei, C.; Fan, H.; Huang, P.Y.; Aggarwal, V.; Chowdhury, A.; Poursaeed, O.; Hoffman, J.; et al. Hiera: A hierarchical vision transformer without the bells-and-whistles. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; PMLR 202, pp. 29441–29454. [Google Scholar]
  28. Chen, J.; Mei, J.; Li, X.; Lu, Y.; Yu, Q.; Wei, Q.; Luo, X.; Xie, Y.; Adeli, E.; Wang, Y.; et al. TransUNet: Rethinking the U-Net architecture design for medical image segmentation through the lens of transformers. Med. Image Anal. 2024, 97, 103280. [Google Scholar] [CrossRef] [PubMed]
  29. He, Q.; Yang, Q.; Xie, M. HCTNet: A hybrid CNN-transformer network for breast ultrasound image segmentation. Comput. Biol. Med. 2023, 155, 106629. [Google Scholar] [CrossRef]
  30. Zhang, H.; Lian, J.; Yi, Z.; Wu, R.; Lu, X.; Ma, P.; Ma, Y. HAU-Net: Hybrid CNN-transformer for breast ultrasound image segmentation. Biomed. Signal Process. Control 2024, 87, 105427. [Google Scholar] [CrossRef]
  31. Dialameh, M.; Rajabzadeh, H.; Sadeghi-Goughari, M.; Sim, J.S.; Kwon, H.J. DualSwinUnet++: An enhanced Swin-Unet architecture with dual decoders for PTMC segmentation. Comput. Biol. Med. 2025, 196, 110716. [Google Scholar] [CrossRef]
  32. Yang, H.; Yang, D. CSwin-PNet: A CNN-Swin Transformer combined pyramid network for breast lesion segmentation in ultrasound images. Expert Syst. Appl. 2023, 213, 119024. [Google Scholar] [CrossRef]
  33. Tian, C.; Hu, Y.; Zhang, M.; Liao, X.; Lv, J.; Si, W. Self-prompt contextual learning with AxialMamba for multi-label segmentation in carotid ultrasound. Expert Syst. Appl. 2025, 274, 126749. [Google Scholar] [CrossRef]
  34. Zhao, J.; Sun, L.; Fan, D.; Wang, K.; Si, H.; Fu, H.; Zhang, D. Uncertainty-driven edge prompt generation network for medical image segmentation. IEEE Trans. Med. Imaging 2025, 44, 3950–3961. [Google Scholar] [CrossRef]
  35. Luo, X.; Wang, Y.; Ou-Yang, L. Lgffm: A localized and globalized frequency fusion model for ultrasound image segmentation. IEEE Trans. Med. Imaging 2025, 45, 515–527. [Google Scholar] [CrossRef] [PubMed]
  36. Jia, M.; Tang, L.; Chen, B.C.; Cardie, C.; Belongie, S.; Hariharan, B.; Lim, S.N. Visual prompt tuning. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2022; pp. 709–727. [Google Scholar]
  37. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Jiao, J.; Liu, Y. Vmamba: Visual state space model. Adv. Neural Inf. Process. Syst. 2024, 37, 103031–103063. [Google Scholar]
  38. Zhai, X.; Mustafa, B.; Kolesnikov, A.; Beyer, L. Sigmoid loss for language image pre-training. In Proceedings of the IEEE/CVF International Conference on Computer Vision; IEEE: New York, NY, USA, 2023; pp. 11975–11986. [Google Scholar]
  39. Gagan, J.; Shirsat, H.S.; Mathias, G.P.; Mallya, B.V.; Andrade, J.; Rajagopal, K.; Kumar, J.H. Automated segmentation of common carotid artery in ultrasound images. IEEE Access 2022, 10, 58419–58430. [Google Scholar] [CrossRef]
  40. Leclerc, S.; Smistad, E.; Pedrosa, J.; Østvik, A.; Cervenansky, F.; Espinosa, F.; Espeland, T.; Berg, E.A.R.; Jodoin, P.M.; Grenier, T.; et al. Deep learning for segmentation using an open large-scale dataset in 2D echocardiography. IEEE Trans. Med. Imaging 2019, 38, 2198–2210. [Google Scholar] [CrossRef]
  41. Pedraza, L.; Vargas, C.; Narváez, F.; Durán, O.; Muñoz, E.; Romero, E. An open access thyroid ultrasound image database. In Proceedings of the 10th International Symposium on Medical Information Processing and Analysis; SPIE: Bellingham, WA, USA, 2015; Volume 9287, pp. 188–193. [Google Scholar]
  42. van den Heuvel, T.L.; de Bruijn, D.; de Korte, C.L.; Ginneken, B.v. Automated measurement of fetal head circumference using 2D ultrasound images. PLoS ONE 2018, 13, e0200412. [Google Scholar] [CrossRef]
  43. Lu, Y.; Zhou, M.; Zhi, D.; Zhou, M.; Jiang, X.; Qiu, R.; Ou, Z.; Wang, H.; Qiu, D.; Zhong, M.; et al. The JNU-IFM dataset for segmenting pubic symphysis-fetal head. Data Brief 2022, 41, 107904. [Google Scholar] [CrossRef]
  44. Gong, H.; Chen, G.; Wang, R.; Xie, X.; Mao, M.; Yu, Y.; Chen, F.; Li, G. Multi-task learning for thyroid nodule segmentation with thyroid region prior. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI); IEEE: New York, NY, USA, 2021; pp. 257–261. [Google Scholar]
  45. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  46. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. Unet++: Redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 2019, 39, 1856–1867. [Google Scholar] [CrossRef] [PubMed]
  47. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  48. Isensee, F.; Jaeger, P.F.; Kohl, S.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef]
  49. Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-unet: Unet-like pure transformer for medical image segmentation. In Proceedings of the European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2022; pp. 205–218. [Google Scholar]
  50. Zhang, Y.; Liu, H.; Hu, Q. Transfuse: Fusing transformers and cnns for medical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2021; pp. 14–24. [Google Scholar]
  51. He, A.; Wang, K.; Li, T.; Du, C.; Xia, S.; Fu, H. H2former: An efficient hierarchical hybrid transformer for medical image segmentation. IEEE Trans. Med. Imaging 2023, 42, 2763–2775. [Google Scholar] [CrossRef]
  52. Li, Z.; Zheng, Y.; Shan, D.; Yang, S.; Li, Q.; Wang, B.; Zhang, Y.; Hong, Q.; Shen, D. Scribformer: Transformer makes cnn work better for scribble-based medical image segmentation. IEEE Trans. Med. Imaging 2024, 43, 2254–2265. [Google Scholar] [CrossRef]
  53. Margolin, R.; Zelnik-Manor, L.; Tal, A. How to evaluate foreground maps? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition; IEEE: New York, NY, USA, 2014; pp. 248–255. [Google Scholar]
  54. Fan, D.P.; Cheng, M.M.; Liu, Y.; Li, T.; Borji, A. Structure-measure: A new way to evaluate foreground maps. In Proceedings of the IEEE International Conference on Computer Vision; IEEE: New York, NY, USA, 2017; pp. 4548–4557. [Google Scholar]
  55. Fan, D.P.; Ji, G.P.; Qin, X.; Cheng, M.M. Cognitive vision inspired object segmentation metric and loss function. Sci. Sin. Informationis 2021, 6, 5. [Google Scholar]
Figure 1. Overall architecture of the proposed RPFeaNet for medical ultrasound image segmentation. It consists of three core components: (1) The progressive prompt generation module (PPGM) with frozen backbone stages and trainable adapters, which generates multi-level spatial prior prompts P i from the input image. (2) The high-level prompt-guided feature interaction module (HPGFIM), which fuses the backbone feature X i with the prompts P i via Selective State Space 2D (SS2D) blocks to enhance target-aware feature representation. (3) The dynamic selective-frequency (DSE) decoder stack, which refines the fused features in the frequency domain and produces the final segmentation map through a concatenation head.
Figure 1. Overall architecture of the proposed RPFeaNet for medical ultrasound image segmentation. It consists of three core components: (1) The progressive prompt generation module (PPGM) with frozen backbone stages and trainable adapters, which generates multi-level spatial prior prompts P i from the input image. (2) The high-level prompt-guided feature interaction module (HPGFIM), which fuses the backbone feature X i with the prompts P i via Selective State Space 2D (SS2D) blocks to enhance target-aware feature representation. (3) The dynamic selective-frequency (DSE) decoder stack, which refines the fused features in the frequency domain and produces the final segmentation map through a concatenation head.
Sensors 26 02394 g001
Figure 2. Architecture of the dynamic selective-frequency decoder (DSFD). The input feature map is first normalized and transformed to the frequency domain via FFT. Three parallel MLP branches extract multi-scale frequency features, which are then fused with learnable gates to suppress redundant noise. Finally, the fused frequency features are converted back to the spatial domain via IFFT to generate the refined segmentation output.
Figure 2. Architecture of the dynamic selective-frequency decoder (DSFD). The input feature map is first normalized and transformed to the frequency domain via FFT. Three parallel MLP branches extract multi-scale frequency features, which are then fused with learnable gates to suppress redundant noise. Finally, the fused frequency features are converted back to the spatial domain via IFFT to generate the refined segmentation output.
Sensors 26 02394 g002
Figure 3. Visual segmentation results on CCAUI and CAMUS datasets. Top two rows belong to the CAMUS dataset, and the bottom two rows belong to the CCAUI dataset. The proposed method demonstrates improved segmentation accuracy and boundary consistency compared to state-of-the-art methods, including UNet, AttUNet, nnUNet, SwinUNet, TransFuse, H2Former, ScribFormer, and LGFFM.
Figure 3. Visual segmentation results on CCAUI and CAMUS datasets. Top two rows belong to the CAMUS dataset, and the bottom two rows belong to the CCAUI dataset. The proposed method demonstrates improved segmentation accuracy and boundary consistency compared to state-of-the-art methods, including UNet, AttUNet, nnUNet, SwinUNet, TransFuse, H2Former, ScribFormer, and LGFFM.
Sensors 26 02394 g003
Figure 4. Visual segmentation results on DDTI and TN3K datasets. The top two rows belong to the DDTI dataset, and the bottom two rows belong to the TN3K dataset. The proposed method demonstrates improved segmentation accuracy and boundary consistency compared to state-of-the-art methods, including UNet, AttUNet, nnUNet, SwinUNet, TransFuse, H2Former, ScribFormer, and LGFFM.
Figure 4. Visual segmentation results on DDTI and TN3K datasets. The top two rows belong to the DDTI dataset, and the bottom two rows belong to the TN3K dataset. The proposed method demonstrates improved segmentation accuracy and boundary consistency compared to state-of-the-art methods, including UNet, AttUNet, nnUNet, SwinUNet, TransFuse, H2Former, ScribFormer, and LGFFM.
Sensors 26 02394 g004
Figure 5. Visual segmentation results on HC-18 and JNU-IFM datasets. The top two rows belong to the HC-18 dataset, and the bottom two rows belong to the JNU-IFM dataset. The proposed method demonstrates improved segmentation accuracy and boundary consistency compared to state-of-the-art methods, including UNet, AttUNet, nnUNet, SwinUNet, TransFuse, H2Former, ScribFormer, and LGFFM.
Figure 5. Visual segmentation results on HC-18 and JNU-IFM datasets. The top two rows belong to the HC-18 dataset, and the bottom two rows belong to the JNU-IFM dataset. The proposed method demonstrates improved segmentation accuracy and boundary consistency compared to state-of-the-art methods, including UNet, AttUNet, nnUNet, SwinUNet, TransFuse, H2Former, ScribFormer, and LGFFM.
Sensors 26 02394 g005
Table 1. Comparison of RPFeaNet with representative state-of-the-art ultrasound segmentation methods (focused on prompt application, prior modeling and frequency processing).
Table 1. Comparison of RPFeaNet with representative state-of-the-art ultrasound segmentation methods (focused on prompt application, prior modeling and frequency processing).
MethodVenuePrompt Application PositionPrior Modeling StrategyFrequency-Domain Processing
APG-SAM [24]ESWA 2025Box Prompt, Point PromptYolov9 Detection-
CC-SAM [22]ECCV 2024LLM Text PromptGrounding DINONo frequency-domain processing
FreqDINO [23]ArXiv 2025Spatial Semantic PromptDINOv3Fixed high-/low-frequency fusion
LGFFM [35]IEEE TMI 2025-SAM2Static Wavelet frequency fusion
RPFeaNet (Ours)-Low-to-high level progressive promptStructural (CLIP), patch-level encoder, semantic (SAM2)Prompt-guided dynamic selective-frequency fusion
Table 2. Quantitative comparison of different segmentation methods on CCAUI and CAMUS datasets (mean ± SD, *: p < 0.05).
Table 2. Quantitative comparison of different segmentation methods on CCAUI and CAMUS datasets (mean ± SD, *: p < 0.05).
DatasetMethodJaccardDice F β w S α E ϕ ASSDMSE
CCAUIU-Net [45]0.7025 ± 0.0123 *0.8214 ± 0.0095 *0.9856 ± 0.0032 *0.9512 ± 0.0076 *0.8023 ± 0.0142 *0.4258 ± 0.0215 *0.0095 ± 0.0012 *
UNet++ [46]0.7218 ± 0.0108 *0.8386 ± 0.0087 *0.9879 ± 0.0028 *0.9587 ± 0.0069 *0.8195 ± 0.0131 *0.3964 ± 0.0198 *0.0088 ± 0.0010 *
AttUNet [47]0.7352 ± 0.0095 *0.8491 ± 0.0076 *0.9892 ± 0.0025 *0.9623 ± 0.0062 *0.8317 ± 0.0118 *0.3725 ± 0.0182 *0.0081 ± 0.0009 *
nnU-Net [48]0.7486 ± 0.0089 *0.8593 ± 0.0071 *0.9905 ± 0.0022 *0.9668 ± 0.0058 *0.8429 ± 0.0109 *0.3682 ± 0.0175 *0.0083 ± 0.0009 *
Swin-Unet [49]0.7154 ± 0.0112 *0.8327 ± 0.0092 *0.9871 ± 0.0030 *0.9564 ± 0.0072 *0.8136 ± 0.0135 *0.4059 ± 0.0203 *0.0091 ± 0.0011 *
TransFuse [50]0.7289 ± 0.0102 *0.8435 ± 0.0082 *0.9885 ± 0.0026 *0.9605 ± 0.0065 *0.8258 ± 0.0125 *0.3847 ± 0.0190 *0.0085 ± 0.0010 *
H2Former [51]0.7395 ± 0.0091 *0.8526 ± 0.0074 *0.9898 ± 0.0024 *0.9641 ± 0.0060 *0.8362 ± 0.0115 *0.3659 ± 0.0172 *0.0080 ± 0.0008 *
ScribFormer [52]0.7451 ± 0.0087 *0.8572 ± 0.0070 *0.9912 ± 0.0021 *0.9675 ± 0.0057 *0.8405 ± 0.0110 *0.3671 ± 0.0173 *0.0082 ± 0.0009 *
LGFFM [35]0.7525 ± 0.0085 *0.8586 ± 0.0068 *0.9923 ± 0.0020 *0.9736 ± 0.0052 *0.8586 ± 0.0105 *0.3666 ± 0.0170 *0.0082 ± 0.0009 *
RPFeaNet (Ours)0.8783 ± 0.00620.9300 ± 0.00450.9967 ± 0.00150.9879 ± 0.00380.9300 ± 0.00650.1507 ± 0.00850.0031 ± 0.0005
CAMUSU-Net [45]0.9093 ± 0.0026 *0.9185 ± 0.0021 *0.8512 ± 0.0030 *0.9215 ± 0.0019 *0.2216 ± 0.0041 *0.2854 ± 0.0053 *0.0356 ± 0.0012 *
UNet++ [46]0.9101 ± 0.0017 *0.9324 ± 0.0015 *0.8726 ± 0.0022 *0.9308 ± 0.0014 *0.2305 ± 0.0035 *0.2469 ± 0.0044 *0.0321 ± 0.0010 *
AttUNet [47]0.9079 ± 0.0032 *0.9432 ± 0.0028 *0.8915 ± 0.0035 *0.9386 ± 0.0026 *0.2387 ± 0.0039 *0.2158 ± 0.0038 *0.0298 ± 0.0009 *
nnU-Net [48]0.9207 ± 0.0056 *0.9506 ± 0.0049 *0.9058 ± 0.0051 *0.9459 ± 0.0045 *0.2423 ± 0.0042 *0.1987 ± 0.0035 *0.0292 ± 0.0008 *
Swin-Unet [49]0.8751 ± 0.0017 *0.9397 ± 0.0020 *0.8864 ± 0.0024 *0.9412 ± 0.0018 *0.2365 ± 0.0037 *0.2241 ± 0.0040 *0.0305 ± 0.0009 *
TransFuse [50]0.8589 ± 0.0109 *0.9461 ± 0.0063 *0.8967 ± 0.0071 *0.9438 ± 0.0058 *0.2401 ± 0.0045 *0.2095 ± 0.0039 *0.0296 ± 0.0009 *
H2Former [51]0.9182 ± 0.0076 *0.9528 ± 0.0052 *0.9102 ± 0.0058 *0.9486 ± 0.0048 *0.2445 ± 0.0043 *0.1896 ± 0.0033 *0.0289 ± 0.0008 *
ScribFormer [52]0.9055 ± 0.0024 *0.9552 ± 0.0026 *0.9185 ± 0.0029 *0.9507 ± 0.0024 *0.2469 ± 0.0040 *0.1789 ± 0.0031 *0.0287 ± 0.0007 *
LGFFM [35]0.9204 ± 0.0041 *0.9559 ± 0.0023 *0.9204 ± 0.0025 *0.9518 ± 0.0021 *0.2482 ± 0.0038 *0.1716 ± 0.0029 *0.0288 ± 0.0006 *
RPFeaNet (Ours)0.9216 ± 0.00250.9565 ± 0.00180.9217 ± 0.00200.9519 ± 0.00170.2488 ± 0.00320.1778± 0.00300.0291 ± 0.0007
Table 3. Quantitative comparison of different segmentation methods on DDTI and TN3K datasets (mean ± SD, *: p < 0.05).
Table 3. Quantitative comparison of different segmentation methods on DDTI and TN3K datasets (mean ± SD, *: p < 0.05).
DatasetMethodJaccardDice F β w S α E ϕ ASSDMSE
DDTIU-Net [45]0.7625 ± 0.0123 *0.8614 ± 0.0095 *0.9456 ± 0.0032 *0.8712 ± 0.0076 *0.8123 ± 0.0142 *0.5258 ± 0.0215 *0.0495 ± 0.0012 *
UNet++ [46]0.7818 ± 0.0108 *0.8786 ± 0.0087 *0.9579 ± 0.0028 *0.8887 ± 0.0069 *0.8295 ± 0.0131 *0.4964 ± 0.0198 *0.0488 ± 0.0010 *
AttUNet [47]0.7952 ± 0.0095 *0.8891 ± 0.0076 *0.9692 ± 0.0025 *0.8923 ± 0.0062 *0.8417 ± 0.0118 *0.4725 ± 0.0182 *0.0481 ± 0.0009 *
nnU-Net [48]0.8086 ± 0.0089 *0.8993 ± 0.0071 *0.9705 ± 0.0022 *0.8968 ± 0.0058 *0.8529 ± 0.0109 *0.4682 ± 0.0175 *0.0483 ± 0.0009 *
Swin-Unet [49]0.7754 ± 0.0112 *0.8727 ± 0.0092 *0.9571 ± 0.0030 *0.8864 ± 0.0072 *0.8236 ± 0.0135 *0.5059 ± 0.0203 *0.0491 ± 0.0011 *
TransFuse [50]0.7889 ± 0.0102 *0.8835 ± 0.0082 *0.9685 ± 0.0026 *0.8905 ± 0.0065 *0.8358 ± 0.0125 *0.4847 ± 0.0190 *0.0485 ± 0.0010 *
H2Former [51]0.8095 ± 0.0091 *0.8926 ± 0.0074 *0.9798 ± 0.0024 *0.8941 ± 0.0060 *0.8462 ± 0.0115 *0.4659 ± 0.0172 *0.0480 ± 0.0008 *
ScribFormer [52]0.8151 ± 0.0087 *0.8972 ± 0.0070 *0.9812 ± 0.0021 *0.8975 ± 0.0057 *0.8505 ± 0.0110 *0.4671 ± 0.0173 *0.0482 ± 0.0009 *
LGFFM [35]0.8502 ± 0.0085 *0.9174 ± 0.0068 *0.9539 ± 0.0020 *0.9007 ± 0.0052 *0.9174 ± 0.0105 *0.4760 ± 0.0170 *0.0461 ± 0.0009 *
RPFeaNet (Ours)0.8536 ± 0.00620.9198 ± 0.00450.9555 ± 0.00150.9021 ± 0.00380.9198 ± 0.00650.4235 ± 0.00850.0448 ± 0.0005
TN3KU-Net [45]0.6276 ± 0.0171 *0.7298 ± 0.0164 *0.9285 ± 0.0032 *0.8815 ± 0.0076 *0.7716 ± 0.0142 *0.3880 ± 0.0026 *0.0356 ± 0.0012 *
UNet++ [46]0.6424 ± 0.0027 *0.7487 ± 0.0042 *0.9324 ± 0.0028 *0.8908 ± 0.0069 *0.7805 ± 0.0131 *0.4070 ± 0.0058 *0.0321 ± 0.0010 *
AttUNet [47]0.6359 ± 0.0161 *0.7393 ± 0.0164 *0.9432 ± 0.0025 *0.8986 ± 0.0062 *0.7887 ± 0.0118 *0.4100 ± 0.0070 *0.0298 ± 0.0009 *
nnU-Net [48]0.7102 ± 0.0148 *0.8062 ± 0.0122 *0.9506 ± 0.0022 *0.9059 ± 0.0058 *0.7923 ± 0.0109 *0.3700 ± 0.0068 *0.0292 ± 0.0009 *
Swin-Unet [49]0.6599 ± 0.0117 *0.7576 ± 0.0112 *0.9397 ± 0.0030 *0.9012 ± 0.0072 *0.7865 ± 0.0135 *0.3420 ± 0.0092 *0.0305 ± 0.0011 *
TransFuse [50]0.7118 ± 0.01180.8163 ± 0.00890.9461 ± 0.00260.9038 ± 0.00650.7901 ± 0.01250.2820 ± 0.00090.0296 ± 0.0010
H2Former [51]0.7111 ± 0.01220.8072 ± 0.01250.9528 ± 0.00240.9086 ± 0.00600.7945 ± 0.01150.2840 ± 0.00670.0289 ± 0.0008
ScribFormer [52]0.6730 ± 0.0147 *0.7757 ± 0.0129 *0.9552 ± 0.0021 *0.9107 ± 0.0057 *0.7969 ± 0.0110 *0.3360 ± 0.0056 *0.0287 ± 0.0009 *
LGFFM [35]0.7243 ± 0.0085 *0.8210 ± 0.0068 *0.9693 ± 0.0020 *0.9331 ± 0.0052 *0.8265 ± 0.0105 *0.2438 ± 0.0170 *0.0297 ± 0.0009 *
RPFeaNet (Ours)0.7342 ± 0.00750.8306 ± 0.00650.9703 ± 0.00150.9344 ± 0.00380.8142 ± 0.00650.1924 ± 0.00270.0289 ± 0.0005
Table 4. Quantitative comparison of different segmentation methods on HC-18 and JNU-IFM datasets (mean ± SD, *: p < 0.05).
Table 4. Quantitative comparison of different segmentation methods on HC-18 and JNU-IFM datasets (mean ± SD, *: p < 0.05).
DatasetMethodJaccardDice F β w S α E ϕ ASSDMSE
U-Net [45]0.7815 ± 0.0128 *0.8752 ± 0.0106 *0.9512 ± 0.0085 *0.8825 ± 0.0092 *0.8217 ± 0.0114 *0.6128 ± 0.0412 *0.0523 ± 0.0035 *
UNet++ [46]0.8028 ± 0.0114 *0.8906 ± 0.0098 *0.9635 ± 0.0076 *0.8947 ± 0.0083 *0.8389 ± 0.0106 *0.5814 ± 0.0386 *0.0501 ± 0.0032 *
AttUNet [47]0.8165 ± 0.0107 *0.8987 ± 0.0092 *0.9718 ± 0.0069 *0.9012 ± 0.0078 *0.8495 ± 0.0099 *0.5578 ± 0.0364 *0.0489 ± 0.0030 *
nnU-Net [48]0.8302 ± 0.0099 *0.9075 ± 0.0085 *0.9786 ± 0.0063 *0.9085 ± 0.0072 *0.8603 ± 0.0092 *0.5492 ± 0.0348 *0.0485 ± 0.0028 *
Swin-Unet [49]0.7946 ± 0.0119 *0.8834 ± 0.0102 *0.9589 ± 0.0080 *0.8893 ± 0.0087 *0.8312 ± 0.0111 *0.5947 ± 0.0398 *0.0512 ± 0.0033 *
TransFuse [50]0.8098 ± 0.0103 *0.8942 ± 0.0088 *0.9674 ± 0.0071 *0.8978 ± 0.0079 *0.8426 ± 0.0098 *0.5735 ± 0.0357 *0.0496 ± 0.0029 *
H2Former [51]0.8215 ± 0.0094 *0.9028 ± 0.0081 *0.9759 ± 0.0060 *0.9051 ± 0.0068 *0.8538 ± 0.0087 *0.5467 ± 0.0331 *0.0482 ± 0.0027 *
ScribFormer [52]0.8279 ± 0.0089 *0.9061 ± 0.0077 *0.9803 ± 0.0057 *0.9097 ± 0.0065 *0.8584 ± 0.0083 *0.5481 ± 0.0324 *0.0484 ± 0.0026 *
LGFFM [35]0.8427 ± 0.0076 *0.9158 ± 0.0064 *0.9618 ± 0.0051 *0.9125 ± 0.0058 *0.8716 ± 0.0072 *0.5398 ± 0.0298 *0.0472 ± 0.0023 *
RPFeaNet (Ours)0.8519 ± 0.00420.9214 ± 0.00350.9645 ± 0.00280.9158 ± 0.00310.8789 ± 0.00390.4876 ± 0.01860.0458 ± 0.0017
JNU-IFMU-Net [45]0.7128 ± 0.0156 *0.8315 ± 0.0123 *0.9326 ± 0.0092 *0.8517 ± 0.0105 *0.7824 ± 0.0131 *1.8254 ± 0.1286 *0.0615 ± 0.0042 *
UNet++ [46]0.7345 ± 0.0142 *0.8478 ± 0.0115 *0.9418 ± 0.0085 *0.8639 ± 0.0097 *0.7987 ± 0.0124 *1.6872 ± 0.1175 *0.0589 ± 0.0039 *
AttUNet [47]0.7519 ± 0.0133 *0.8592 ± 0.0108 *0.9503 ± 0.0078 *0.8725 ± 0.0091 *0.8094 ± 0.0118 *1.5786 ± 0.1063 *0.0567 ± 0.0036 *
nnU-Net [48]0.7684 ± 0.0125 *0.8687 ± 0.0101 *0.9578 ± 0.0072 *0.8802 ± 0.0085 *0.8189 ± 0.0112 *1.4963 ± 0.0984 *0.0559 ± 0.0034 *
Swin-Unet [49]0.7427 ± 0.0147 *0.8536 ± 0.0112 *0.9465 ± 0.0089 *0.8684 ± 0.0099 *0.8031 ± 0.0127 *1.6348 ± 0.1132 *0.0578 ± 0.0038 *
TransFuse [50]0.7586 ± 0.0137 *0.8631 ± 0.0105 *0.9537 ± 0.0075 *0.8758 ± 0.0088 *0.8146 ± 0.0121 *1.5279 ± 0.1035 *0.0562 ± 0.0037 *
H2Former [51]0.7725 ± 0.0121 *0.8714 ± 0.0097 *0.9605 ± 0.0069 *0.8827 ± 0.0082 *0.8225 ± 0.0109 *1.4785 ± 0.0957 *0.0554 ± 0.0033 *
ScribFormer [52]0.7798 ± 0.0116 *0.8759 ± 0.0093 *0.9632 ± 0.0066 *0.8859 ± 0.0079 *0.8271 ± 0.0105 *1.4692 ± 0.0942 *0.0551 ± 0.0032 *
LGFFM [35]0.7865 ± 0.0108 *0.8807 ± 0.0089 *0.9714 ± 0.0059 *0.8913 ± 0.0074 *0.8358 ± 0.0098 *1.4287 ± 0.0896 *0.0543 ± 0.0030 *
RPFeaNet (Ours)0.7982 ± 0.00530.8889 ± 0.00410.9738 ± 0.00320.8947 ± 0.00380.8426 ± 0.00461.2854 ± 0.05720.0527 ± 0.0021
Table 5. Complexity analysis of different segmentation methods.
Table 5. Complexity analysis of different segmentation methods.
MethodFPSParams (M)FLOPs (G)
U-Net89.4534.5365.52
UNet++51.7436.63138.66
AttUNet87.2834.8866.63
nnU-Net89.6334.5665.72
Swin-UNet148.9927.155.92
TransFuse115.6326.178.65
H2Former107.0333.6332.25
ScribFormer59.7150.4254.46
LGFFM32.2215.0264.42
RPFeaNet (Ours)10.5429.31108.12
Table 6. Performance comparison of ablation study on CCAUI dataset.
Table 6. Performance comparison of ablation study on CCAUI dataset.
PPGMHPGFIMDSFDDice ↑Jaccard ↑ASSD ↓
0.8520.7410.321
0.8670.7630.305
0.8810.7850.289
0.9050.8270.242
0.9120.8390.231
0.9180.8480.224
0.9300.8780.151
Table 7. Performance comparison of ablation study on CAMUS dataset.
Table 7. Performance comparison of ablation study on CAMUS dataset.
PPGMHPGFIMDSFDDice ↑Jaccard ↑ASSD ↓
0.9210.8530.285
0.9340.8720.268
0.9420.8890.251
0.9480.9010.224
0.9510.9070.216
0.9530.9120.209
0.9570.9220.178
Table 8. Ablation analysis of sub-branches (LEF, PPTB, PPSB) on CCAUI dataset.
Table 8. Ablation analysis of sub-branches (LEF, PPTB, PPSB) on CCAUI dataset.
Sub-Branch SetupDice ↑Jaccard ↑ASSD ↓
w/LEF only0.90150.82070.2217
w/PPTB only0.90830.83240.2089
w/PPSB only0.89570.81160.2356
w/LEF + PPTB0.92460.85980.1783
w/LEF + PPSB0.91920.85030.1895
w/PPTB + PPSB0.92970.86850.1672
w/LEF + PPTB + PPSB (Full)0.93000.87830.1507
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, L.; Du, Y. RPFeaNet: Rethinking Deep Progressive Prompt-Guided Feature Interaction Fusion Network for Medical Ultrasound Image Segmentation. Sensors 2026, 26, 2394. https://doi.org/10.3390/s26082394

AMA Style

Zhu L, Du Y. RPFeaNet: Rethinking Deep Progressive Prompt-Guided Feature Interaction Fusion Network for Medical Ultrasound Image Segmentation. Sensors. 2026; 26(8):2394. https://doi.org/10.3390/s26082394

Chicago/Turabian Style

Zhu, Lei, and Yuqing Du. 2026. "RPFeaNet: Rethinking Deep Progressive Prompt-Guided Feature Interaction Fusion Network for Medical Ultrasound Image Segmentation" Sensors 26, no. 8: 2394. https://doi.org/10.3390/s26082394

APA Style

Zhu, L., & Du, Y. (2026). RPFeaNet: Rethinking Deep Progressive Prompt-Guided Feature Interaction Fusion Network for Medical Ultrasound Image Segmentation. Sensors, 26(8), 2394. https://doi.org/10.3390/s26082394

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop