Next Article in Journal
Advanced Fault Diagnosis and Health Monitoring Techniques for Complex Engineering Systems: 2nd Edition
Previous Article in Journal
Balancing Cost and Precision: An Experimental Evaluation of Sensors for Monitoring in Electrical Generation Systems
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DPM-UNet: A Mamba-Based Network with Dynamic Perception Feature Enhancement for Medical Image Segmentation

1
Key Laboratory of Opto-Electronic Information Processing, Chinese Academy of Sciences, Shenyang 110016, China
2
Shenyang Institute of Automation, Chinese Academy of Sciences, Shenyang 110016, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(22), 7053; https://doi.org/10.3390/s25227053
Submission received: 24 October 2025 / Revised: 11 November 2025 / Accepted: 14 November 2025 / Published: 19 November 2025
(This article belongs to the Section Biomedical Sensors)

Abstract

In medical image segmentation, effective integration of global and local features is crucial. Current methods struggle to simultaneously model long-range dependencies and fine local details. Convolutional Neural Networks (CNNs) excel at extracting local features but are limited by their local receptive fields for capturing long-range dependencies. While global self-attention mechanisms (e.g., in Transformers) can capture long-range spatial relationships, their quadratic computational complexity incurs high costs for high-resolution medical images. To address these limitations, State Space Models (SSMs), which maintain linear complexity while effectively establishing long-range dependencies, have been introduced to visual tasks. Leveraging the advantages of SSMs, this paper proposes DPM-UNet. The network employs a Dual-path Residual Fusion Module (DRFM) at shallow layers to extract local detailed features and a DPMamba Module at deep layers to model global semantic information, achieving effective local global feature fusion. A Multi-scale Aggregation Attention Network (MAAN) is further incorporated to enhance multi-scale representations. The proposed method collaboratively captures local details, long-range dependencies, and multi-scale information in medical images. Experiments on three public datasets demonstrate that DPM-UNet outperforms existing methods across multiple evaluation metrics.

1. Introduction

Medical image segmentation plays a pivotal role in modern clinical practice and scientific research, with its value extending across the entire workflow from auxiliary diagnosis and treatment planning to therapeutic outcome evaluation [1,2,3,4]. Traditional segmentation methods heavily rely on manual annotation by physicians, which are time-consuming, labor-intensive, inefficient, and suffer from significant inter-observer variability, directly impacting diagnostic consistency and result reproducibility [5,6]. Against this backdrop, deep learning-based automated segmentation techniques have emerged and have been successfully applied to various imaging segmentation tasks, including magnetic resonance imaging (MRI), nuclei segmentation in microscopic images [7,8], and multi-organ segmentation in computed tomography (CT) [9,10]. These techniques, by virtue of their high efficiency, high accuracy, and robust stability [11,12], lay the groundwork for translating precision medicine from concept to clinical application.
Despite significant progress in deep learning for medical image segmentation, effectively integrating local features with global long-range dependencies to further enhance segmentation accuracy remains a key research challenge. Convolutional neural networks (CNNs) [13,14], exemplified by U-Net [15], SegResNet [16], and nnU-Net [17], excel at extracting local detailed features. However, their inherent local receptive fields limit the model’s capacity to capture information from distant regions within an image, thereby constraining further improvements in segmentation accuracy. In contrast, Vision Transformers (ViTs) [18,19], through their global self-attention mechanism, empower each image patch (token) to attend to all other patches, demonstrating superior performance in modeling long-range feature interactions and capturing global context [20,21]. Nevertheless, the computational complexity of the attention mechanism in ViTs grows quadratically with the number of image patches [22], imposing a substantial computational burden when processing high-resolution medical images.
In recent years, State Space Sequence Models (SSMs) [23,24] have garnered significant attention in the field of computer vision due to their efficiency in processing long sequences. Among them, the Mamba model [22], with its linear computational complexity and advantages in global modeling [25,26], has emerged as an effective solution for capturing long-range dependencies in visual tasks. Compared to Transformer architectures, which exhibit quadratic computational complexity, Mamba-based vision models can process long sequences with near-linear computational overhead, significantly enhancing their scalability and practicality in high-resolution image tasks. Vmamba [25] effectively addressed the mismatch between the 1D sequential processing of traditional SSMs and the 2D spatial structure of images by introducing strategies such as image patching, sequence flattening, and cross-scanning, thereby achieving efficient modeling of the global image context. This progress has spurred the development of a series of Mamba-based medical image segmentation models, including U-Mamba [27], SegMamba [28], and SwinUMamba [29]. However, existing models, while achieving global perception, fail to adequately integrate local detailed features and spatial contextual information, which limits their expressive power in complex medical image segmentation tasks.
To fully integrate local and global features, this paper proposes a Mamba-based DPM-UNet network. The network employs a Dual-path Residual Fusion Module (DRFM) at shallow layers to extract low-level visual features, while a DPMamba Module is introduced at deep layers. The DPMamba Module utilizes Mamba to flatten the image into a sequence, capturing long-range dependencies and global contextual information. To further enhance the model’s ability to extract and fuse image features at a low computational cost, a Dynamic Perception Feature Enhancement Block (DPFE) is applied to the globally aware feature maps generated by Mamba. Additionally, we design a Multi-scale Aggregation Attention Network (MAAN) to extract multi-scale information from the outputs of the DRFM and optimize feature transmission along the encoder-to-decoder path via skip connections. The main contributions of this paper are summarized as follows:
(1)
We propose a novel segmentation network named DPM-UNet, which integrates the local feature extraction capability of CNNs with the global information aggregation ability of Mamba, aiming to achieve precise medical image segmentation.
(2)
We design three key components: the Dual-path Residual Fusion Module (DRFM), the DPMamba Module, and the Multi-scale Aggregation Attention Network (MAAN). The DRFM enhances local feature extraction by fusing features from standard and dilated convolutions. The DPMamba Module leverages Mamba to generate global features and further enhances feature representation in critical channels through a Dynamic Perception Feature Enhancement Block (DPFE). Additionally, the MAAN is embedded in skip connection paths to optimize the transmission and fusion of multi-scale information.
(3)
Experimental results on three public medical image segmentation datasets demonstrate that DPM-UNet achieves state-of-the-art segmentation performance compared to existing methods, fully validating the effectiveness and strong generalization capability of our approach for medical image segmentation tasks.

2. Methods

The overall architecture of DPM-UNet is illustrated in Figure 1a. In contrast to the nearly pure VSS block design of Swin-UMamba, DPM-UNet adopts a hybrid structure: its shallow stages (1–3) use convolution-based DRFMs to focus on extracting local detailed features, while the deeper stages (4–5) incorporate DPMamba modules (VSS-based) to capture global semantic dependencies. The network follows a U-shaped symmetric layout, with max pooling for downsampling and bilinear interpolation for upsampling. Skip connections are incorporated between each corresponding encoder and decoder stage to fuse features across different hierarchies. Additionally, a Multi-scale Aggregation Attention Network (MAAN) is embedded in the skip connections from Stages 1 to 3 to extract multi-scale features and optimize feature transmission during the skip-connection process. This hybrid design balances local and global feature learning, improving performance while reducing complexity compared to pure VSS-based approaches.

2.1. DPMamba Module

As shown in Figure 1b, the DPMamba Module first utilizes a VSS Block at the input stage to capture global feature information, and then employs a DPFE Block to further enhance its feature representation capability. Assuming the input feature X l has a shape of R C × H × W , we have:
  X l + 1 = V S S B a t c h N o r m X l + X l
X o u t = D P F E B a t c h N o r m X l + 1 + X l + 1
DPMamba can be decomposed into two independent functional components, VSS(·) and DPFE (·), dedicated to global spatial information extraction and feature refinement, respectively.

2.1.1. VSS Block

Traditional attention mechanisms struggle with efficient long-sequence modeling due to their quadratic computational complexity, posing significant challenges for processing large-scale medical images. The Mamba architecture, based on State Space Sequence Models (SSMs), reduces computational complexity to linear while demonstrating remarkable performance in natural language processing. Leveraging this advantage of Mamba, we introduce its vision variant—the Visual State Space (VSS) model—into the field of medical image segmentation. As illustrated in Figure 1c, the VSS Block employed in this work is designed following the methodology presented in reference [25]. The processing pipeline is as follows: the input features first pass through a linear layer, whose output is split evenly along the channel dimension into two tensors. One branch undergoes processing through depthwise separable convolution, a SiLU activation function, 2D Selective Scanning (SS2D), and layer normalization. The other branch is processed only by a SiLU activation function. Finally, the outputs of the two branches are multiplied element-wise, and the result is fed into another linear layer to produce the final output.

2.1.2. Dynamic Perception Feature Enhancement Block (DPFE)

Recent studies have demonstrated that the gated multilayer perceptron (gated MLP) architecture delivers remarkable performance in natural language processing tasks [30]. We posit that the gating mechanism introduced in this architecture holds significant potential for application in visual tasks as well. Building upon this rationale, we propose the Dynamic Perception Feature Enhancement Block (DPFE), designed to further enhance the model’s capability for image feature extraction and fusion at a low computational cost. The structure of the DPFE module is illustrated in Figure 1d. The output features X l from the VSS Block are first processed by an SE Block [31] for channel-wise attention calibration, yielding features X l + 1 to emphasize important feature channels while suppressing less significant ones. Subsequently, the features X l + 1 are split into X 1 l + 1 and X 2 l + 1 , which are fed into two parallel branches. X 1 l + 1 passes through a 1 × 1 convolutional layer, followed by a 3 × 3 depthwise separable convolution with residual connections to extract local spatial features, and undergoes non-linear transformation via a GELU activation function, resulting in features X 1 l + 2 . Meanwhile, X 2 l + 1 is processed by a 1 × 1 convolutional layer to produce features X 2 l + 2 . The features X 1 l + 2 and X 2 l + 2 are then multiplied element-wise to achieve dynamic weighting based on feature importance. Finally, the weighted result is passed through a 1 × 1 convolutional layer to produce the output X o u t , thereby completing an adaptive feature enhancement process that progresses from channel attention calibration to gated dynamic modulation. The mathematical formulation of the DPFE is as follows:
  X l + 1 = S E X l
  X 1 l + 2 = σ f 3 × 3 d e c f 1 × 1 X 1 l + 1 + f 1 × 1 X 1 l + 1
  X 2 l + 2 = f 1 × 1
  X o u t = f 1 × 1 X 1 l + 2 X 2 l + 2
Among these, SE denotes the Squeeze-and-Excitation module, f 1 × 1 represents a convolutional layer with a kernel size of 1 × 1, f 3 × 3 d e c denotes a depthwise separable convolutional layer with a kernel size of 3 × 3, and σ(·) refers to the GELU activation function.

2.2. Dual-Path Residual Fusion Module (DRFM)

In medical image segmentation tasks, effectively capturing local features while expanding the receptive field to improve segmentation accuracy remains a core challenge. Existing methods often employ multi-scale convolutional kernels to address this issue. While capable of capturing broader contextual information, these approaches typically incur high computational costs and fail to adequately model the correlations between features under different receptive fields. To overcome these limitations, we propose the Dual-path Residual Fusion Module (DRFM), whose structure is illustrated in Figure 1e. This module adopts a dual-path parallel residual design, synergistically integrating features extracted by standard convolutions and dilated convolutions to generate more informative and robust feature representations. Specifically, the input feature X l first undergoes a 1 × 1 convolution for channel dimension adjustment, yielding feature X l + 1 . It is then fed into two parallel paths: one path uses a 3 × 3 standard convolution to capture local detailed textures, while the other employs a 3 × 3 dilated convolution with a dilation rate of 2 to acquire wide-range contextual information at a lower computational cost. The outputs from both paths are fused to obtain X l + 2 . This feature X l + 2 subsequently undergoes deeper processing via a parallel set of standard and dilated convolutions, producing X l + 3 . Finally, a 1 × 1 convolution integrates all learned features, and the result is added to the residual connection-provided X l + 1 to produce the final output feature X o u t . The mathematical formulation of the DRFM is as follows:
X l + 1 = f 1 × 1 X l
  X l + 2 = C o n c a t B a t c h N o r m f 3 × 3 X l + 1 ,   B a t c h N o r m f 3 × 3 2 X l + 1
  X l + 3 = C o n c a t B a t c h N o r m f 3 × 3 X l + 2 ,   B a t c h N o r m f 3 × 3 2 X l + 2
  X o u t = X l + 1 + f 1 × 1 X l + 3
Among these, f 1 × 1 and f 3 × 3 denote convolutional layers with kernel sizes of 1 × 1 and 3 × 3, respectively, f 3 × 3 2 represents a dilated convolutional layer with a dilation rate of 2 and a kernel size of 3 × 3, and C o n c a t refers to concatenation along the channel dimension.

2.3. Multi-Scale Aggregation Attention Network (MAAN)

In medical image segmentation tasks, significant size variations among target structures impose high demands on multi-scale feature extraction capabilities. To integrate multi-scale semantic information and enhance feature representation capacity for targets of different sizes, we design the Multi-scale Aggregation Attention Network (MAAN), whose structure is illustrated in Figure 1f. This module adopts a parallel dual-branch architecture to enhance input features from both channel and spatial dimensions, respectively. The feature map X output by the DRFM encoder is fed into the MAAN module for refinement. In the channel branch, the spatial dimensions are compressed into channel descriptors of size C × 1 × 1 through average pooling and max pooling. These descriptors then sequentially pass through a 1 × 1 convolution, a ReLU activation function, another 1 × 1 convolution, and a Sigmoid activation function to generate channel attention weights. These weights are multiplied element-wise with the input feature X , yielding the channel-enhanced feature X c h a n n e l . In the spatial branch, the feature X undergoes a serial cascaded structure composed of 3 × 3, 5 × 5, and 7 × 7 convolutions, with 1 × 1 convolutions used for feature aggregation between layers, producing the multi-scale fused feature X M S . Subsequently, both average pooling and max pooling are applied, and the resulting features are concatenated along the channel dimension. This is followed by a 7 × 7 convolution and a Sigmoid activation function to generate a spatial attention map. Finally, this map is weighted and fused with X M S to produce the spatially enhanced feature X s p a t i a l . The channel feature X c h a n n e l and the spatial feature X s p a t i a l are combined with the original input feature X via residual connection to produce the output feature X o u t , thereby enhancing the feature representation capacity in both spatial and channel dimensions. This cross-scale information interaction mechanism effectively improves the model’s ability to represent multi-target structures with significant scale variations. The mathematical formulation of the MAAN is as follows:
X c h a n n e l = σ f 1 × 1 R e L U f 1 × 1 M P s X + A P s X X
  X M S = f 7 × 7 f 1 × 1 f 5 × 5 f 1 × 1 f 3 × 3 X
  X s p a t i a l = σ f 7 × 7 C o n c a t M P c X M S , A P c X M S X M S
  X o u t = X + X c h a n n e l + X s p a t i a l
Among these, f 1 × 1 , f 3 × 3 , f 5 × 5 and f 7 × 7 denote convolutional layers with kernel sizes of 1 × 1, 3 × 3, 5 × 5 and 7 × 7; M P c and A P c indicate max pooling and average pooling operations along the channel dimension; M P s and A P s represent max pooling and average pooling operations along the spatial dimensions; σ(·) denotes the Sigmoid activation function; and C o n c a t refers to concatenation along the channel dimension.

3. Experiments

3.1. Datasets

(1) The Abdomen MRI dataset: This dataset used in this study is sourced from the publicly available MICCAI 2022 AMOS Challenge resources [29]. It is designed for segmenting 13 abdominal organs, such as the liver, spleen, pancreas, kidneys, stomach, gallbladder, esophagus, aorta, inferior vena cava, adrenal glands, and duodenum. To ensure statistical reliability, the data partition (60 scans for training, 50 for testing) follows the established benchmark protocol in reference [27], which introduced additional scans to overcome limitations of a smaller validation set. These test scans are fully independent, with no patient overlap against the training set, and were annotated by professional radiologists to ensure quality. Throughout the experiments, all images were preprocessed to a resolution of 320 × 320 pixels.
(2) The Microscopy dataset: This dataset focuses on cell instance segmentation tasks, with image data originating from the publicly available NeurIPS 2022 Cell Segmentation Challenge dataset [32]. Its specific composition includes 1000 images for training and 101 images for testing. To standardize the input size, all images were cropped to 512 × 512 pixels before training and testing. The data processing methodology follows the scheme proposed in reference [27].
(3) The ACDC dataset: This dataset comprises cardiac MRI scans from 150 patients, with each patient containing scans from different physiological phases, such as systole and diastole. Sourced from the Automatic Cardiac Diagnosis Challenge [33], its core task is to segment the left ventricle, right ventricle, and myocardium. Our experiment utilizes data from 100 patients in this dataset. We divided these 100 cases into training and test sets in an 8:2 ratio. The training set contains 160 scans from 80 patients, and the test set contains 40 scans from 20 patients. All images used for training and testing were preprocessed and uniformly resized to 256 × 256 pixels.

3.2. Evaluation Metrics and Baselines

For MRI-based datasets, Abdomen MRI and ACDC, the Dice Similarity Coefficient (DSC) was used to measure volumetric overlap between segmentations and ground truth, while the Normalized Surface Distance (NSD) evaluated boundary accuracy. For the Microscopy dataset, which involves instance-level cell identification, the F1-Score served as the primary metric to assess detection performance at the object level. A prediction was counted as a True Positive only if the IoU with the ground truth exceeded 0.5. Additionally, the DSC metric was applied to complement this by evaluating pixel-level segmentation accuracy within each correctly detected instance. The calculation formulas for all metrics are as follows:
D S C = 2 T P T P + F N + T P + F P
  N S D = S p r e d S g t , τ + S g t S p r e d , τ S p r e d + S g t
I o U = T P T P + F P + F N
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l ,   i f   I o U > 0.5
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
Here, TP (True Positive) denotes the number of samples correctly predicted as positive; FP (False Positive) denotes the number of samples incorrectly predicted as positive; and FN (False Negative) denotes the number of samples incorrectly predicted as negative. Spred and Sgt represent the sets of surface points corresponding to the predicted segmentation and the ground truth segmentation, respectively. Spred,τ is defined as the set of all points in the predicted surface whose distance to the ground truth surface is less than the threshold τ; similarly, Sgt,τ refers to the set of all points in the ground truth surface whose distance to the predicted surface is within τ.
We compared five medical image segmentation models under uniform experimental conditions, covering three major architecture types: The traditional CNN architectures are represented by nnU-Net [17] and SegResNet [16]. The Transformer-based architectures are represented by UNETR [34] and SwinUNETR [35]. Swin-UMamba [29] was selected as the representative Mamba-based architecture. All models were rigorously reproduced within the nnU-Net framework, strictly adhering to the hyperparameter configurations reported in their respective original publications.

3.3. Implementation Details

In the architecture of the DPM-UNet network, each of stages 4 to 5 is configured with four consecutive DPMamba modules, forming a [4, 4] structure. All models are implemented within the nnU-Net framework without using any pre-trained weights, and all parameters are optimized through training from scratch. This design enables a focused investigation into network architecture innovation while maintaining consistent experimental conditions—such as image preprocessing and data augmentation—across all compared methods. Consequently, DPM-UNet is evaluated under uniform settings, ensuring that the network architecture remains the sole differentiating factor. During training, patch size, batch size, and network configuration adhere to the standard nnU-Net settings. The training process employs the AdamW optimizer with a weight decay coefficient of 0.05. The initial learning rate is set to 2 × 10−4 and decays to a minimum of 1 × 10−6 using a cosine annealing scheduling strategy. The loss function is defined as the unweighted sum of Dice loss and cross-entropy loss. A five-fold cross-validation strategy is applied on the training set for all three datasets. Since the Dice Similarity Coefficient (DSC) is the most critical evaluation metric in our task, the model checkpoint achieving the highest DSC on the validation set is selected as the optimal model for each training run. The models are trained for 1000 epochs in total without adopting an early stopping strategy, leveraging the cosine annealing learning rate scheduler for full-cycle optimization. The batch size is set to 4 for the microscopy dataset and 8 for the other two datasets. The training curves of three datasets are shown in Figure 2. All experiments are implemented using the PyTorch (torch = 2.1.1) framework on hardware equipped with an NVIDIA RTX 4070 GPU.

3.4. Experimental Results

Table 1 presents the quantitative results of multi-organ segmentation on the Abdomen MRI dataset. The DPM-UNet model achieved a DSC of 78.15%, representing a 1.52 percentage point improvement over the second-best performing nnU-Net, while its NSD reached 84.67%, with an improvement of 1.16 percentage points. As shown in Figure 3, regarding organ-level segmentation accuracy, DPM-UNet ranked first in 10 organs including the liver, left and right kidneys, spleen, pancreas, inferior vena cava, left adrenal gland, esophagus, stomach, and duodenum. The overall segmentation visualization in Figure 4 and the enlarged local details in Figure 5 consistently demonstrate that DPM-UNet not only accurately captures the main structures of organs but also produces more continuous and smoother segmentation contours along complex anatomical boundaries. These results fully validate the effectiveness of the proposed global–local feature collaboration mechanism in representing multi-scale anatomical structures.
The experimental results on the Microscopy dataset are presented in Table 2. The DPM-UNet model achieved a DSC of 73.25%, representing a 1.56 percentage point improvement over the second-best performing UNETR, while its F1-score reached 60.23%, with an improvement of 5.58 percentage points. As shown in the visual comparison in Figure 6, DPM-UNet accurately captures the main structures of cells. The enlarged local details in Figure 7 further demonstrate that DPM-UNet effectively identifies and separates small-sized or adherent nuclei with more precise edge delineation. These results underscore the advantages of the model’s local feature extraction mechanism, which enables it to capture subtle differential features and fine boundary information between nuclei, thereby significantly improving segmentation accuracy.
The experimental results on the ACDC dataset are presented in Table 3. While the DPM-UNet model achieved optimal performance, the overall performance gap among different models was relatively small. Specifically, DPM-UNet improved the DSC metric by 0.14% compared to nnU-Net and enhanced the NSD metric by 0.05% compared to Swin-UMamba. In the substructure segmentation tasks, DPM-UNet delivered optimal performance for myocardial (Myo) and left ventricular (LV) segmentation, while ranking third in right ventricular (RV) segmentation. The overall comparison in Figure 8 and the enlarged local details in Figure 9 demonstrate that DPM-UNet generates more accurate and sharper segmentation contours, further validating its local–global feature interaction capability in effectively capturing critical subtle structural features along the edges.

3.5. Further Analysis

3.5.1. Ablation Study

The ablation experiments on the abdominal MRI dataset evaluated the effectiveness of the proposed modules, with results summarized in Table 4. The introduction of the DRFM alone improved performance, increasing the Dice Similarity Coefficient (DSC) by 0.45% and the Normalized Surface Distance (NSD) by 0.88%. After integrating both the DRFM and DPMamba modules, the model performance was further enhanced, achieving additional gains of 0.40% in DSC and 0.41% in NSD. This demonstrates the capability of the DPMamba module to effectively improve segmentation accuracy. Subsequently, incorporating the CBAM module led to further improvements, with DSC and NSD increasing by an additional 0.82% and 0.73%, respectively. When the CBAM module was replaced with the MAAN module, the model achieved its optimal performance, reaching a DSC of 78.15% and an NSD of 84.67%. MAAN outperforms CBAM due to its structural advantage; a dual-path design that concurrently models’ channel and spatial relationships, combined with multi-scale feature aggregation, which enables more effective context capture than CBAM’s sequential design. These results fully validate the effectiveness of the designed modules and their synergistic contribution to enhancing the model’s overall segmentation capability. (Note: In the following table, a check mark (✔) or cross mark (✘) indicates the presence or absence of the corresponding module. An upward (↓) or downward (↑) arrow indicates that a higher or lower value denotes better performance, respectively.)
To further investigate the impact of the DRFM structure on the model’s feature extraction capability, we conducted additional ablation experiments, with the results summarized in Table 5. The corresponding schematic diagrams of each experimental setup can be found in Figure 10a–d. The experimental results demonstrate that introducing dilated convolutions in the second path effectively enhances model performance. Furthermore, fusing features extracted by standard convolutions and dilated convolutions yielded more substantial performance improvements. Additionally, the incorporation of residual connections also contributed significantly to the overall performance enhancement. These results indicate that the DRFM can effectively strengthen the model’s feature representation capability.

3.5.2. Model Complexity

As shown in Table 6, this study employs floating-point operations (FLOPs), number of parameters (Params), and the total training time over 1000 epochs as key metrics to evaluate model complexity and practical training efficiency. In terms of computational and storage complexity, DPM-UNet demonstrates a moderate level among the compared models. However, its training time is the second longest, indicating a higher computational cost during the training phase. Critically, the superior performance of DPM-UNet justifies this increased training cost. It achieves the highest Dice Similarity Coefficient (DSC) and Normalized Surface Dice (NSD), demonstrating that the method achieves a favorable balance between computational efficiency and segmentation accuracy.

4. Conclusions and Future Work

This study proposes DPM-UNet, a medical image segmentation network based on the Mamba architecture. The method adopts a U-shaped structure, where a Dual-path Residual Fusion Module (DRFM) is introduced in the shallow layers to enhance local detail feature extraction, while a DPMamba module capable of modeling long-range dependencies is employed in the deep layers to capture global contextual information. Furthermore, the Multi-scale Aggregation Attention Network (MAAN) is incorporated to strengthen the model’s ability to perceive and fuse multi-scale features. Experimental results on three public medical image segmentation datasets demonstrate the superior segmentation performance of the proposed method. While the current study has established a solid foundation, certain aspects, such as multi-run performance validation, statistical significance testing, and detailed efficiency profiling, were not fully explored within the scope of this work. These areas present meaningful opportunities for further investigation. Future efforts will focus on developing more lightweight architectures and efficient training strategies to reduce computational costs, alongside incorporating rigorous benchmarking and efficiency analysis to enhance the robustness and practicality of the model under resource constraints.

Author Contributions

Conceptualization, S.X., X.L. and B.H.; formal analysis, S.X.; investigation, S.X.; methodology, S.X.; resources, S.X.; software, S.X.; supervision, B.H.; validation, S.X., X.L. and H.L.; visualization, S.X., X.L. and H.L.; writing—original draft, S.X.; writing—review and editing, S.X., X.L., H.L. and B.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

You can download the AbdomenMRI/Microscopy/ACDC dataset at https://drive.google.com/drive/folders/1CH2OWQpd4Sa-BES6oFLRC469gTxf6QUO (accessed on 11 November 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Dai, L.; Sheng, B.; Chen, T.; Wu, Q.; Liu, R.; Cai, C.; Wu, L.; Yang, D.; Hamzah, H.; Liu, Y.; et al. A deep learning system for predicting time to progression of diabetic retinopathy. Nat. Med. 2024, 30, 584–594. [Google Scholar] [CrossRef] [PubMed]
  2. Chen, X.; Wang, X.; Zhang, K.; Fung, K.M.; Thai, T.C.; Moore, K.; Mannel, R.S.; Liu, H.; Zheng, B.; Qiu, Y. Recent advances and clinical applications of deep learning in medical image analysis. Med. Image Anal. 2022, 79, 102444. [Google Scholar] [CrossRef] [PubMed]
  3. Bai, W.; Suzuki, H.; Huang, J.; Francis, C.; Wang, S.; Tarroni, G.; Guitton, F.; Aung, N.; Fung, K.; Petersen, S.E.; et al. A population-based phenome-wide association study of cardiac and aortic structure and function. Nat. Med. 2020, 26, 1654–1662. [Google Scholar] [CrossRef]
  4. Mei, X.; Lee, H.-C.; Diao, K.-y.; Huang, M.; Lin, B.; Liu, C.; Xie, Z.; Ma, Y.; Robson, P.M.; Chung, M.; et al. Artificial intelligence–enabled rapid diagnosis of patients with COVID-19. Nat. Med. 2020, 26, 1224–1228. [Google Scholar] [CrossRef] [PubMed]
  5. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  6. Jungo, A.; Meier, R.; Ermis, E.; Blatti-Moreno, M.; Herrmann, E.; Wiest, R.; Reyes, M. On the Effect of Inter-Observer Variability for a Reliable Estimation of Uncertainty of Medical Image Segmentation; Springer: Cham, Switzerland, 2018; pp. 682–690. [Google Scholar]
  7. Graham, S.; Vu, Q.D.; Raza, S.E.A.; Azam, A.; Tsang, Y.W.; Kwak, J.T.; Rajpoot, N. Hover-Net: Simultaneous segmentation and classification of nuclei in multi-tissue histology images. Med. Image Anal. 2019, 58, 101563. [Google Scholar] [CrossRef]
  8. Sun, M.; Zou, W.; Wang, Z.; Wang, S.; Sun, Z. An Automated Framework for Histopathological Nucleus Segmentation with Deep Attention Integrated Networks. IEEE ACM Trans. Comput. Biol. Bioinform. 2024, 21, 995–1006. [Google Scholar] [CrossRef]
  9. Gibson, E.; Giganti, F.; Hu, Y.; Bonmati, E.; Bandula, S.; Gurusamy, K.; Davidson, B.; Pereira, S.P.; Clarkson, M.J.; Barratt, D.C. Automatic Multi-Organ Segmentation on Abdominal CT with Dense V-Networks. IEEE Trans. Med. Imaging 2018, 37, 1822–1834. [Google Scholar] [CrossRef]
  10. Qi, X.; Wu, Z.; Zou, W.; Ren, M.; Gao, Y.; Sun, M.; Zhang, S.; Shan, C.; Sun, Z. Exploring Generalizable Distillation for Efficient Medical Image Segmentation. IEEE J. Biomed. Health Inform. 2024, 28, 4170–4183. [Google Scholar] [CrossRef]
  11. Khened, M.; Kollerathu, V.A.; Krishnamurthi, G. Fully convolutional multi-scale residual DenseNets for cardiac segmentation and automated cardiac diagnosis using ensemble of classifiers. Med. Image Anal. 2019, 51, 21–45. [Google Scholar] [CrossRef]
  12. Jiang, X.; Hoffmeister, M.; Brenner, H.; Muti, H.S.; Yuan, T.; Foersch, S.; West, N.P.; Brobeil, A.; Jonnagaddala, J.; Hawkins, N.; et al. End-to-end prognostication in colorectal cancer by deep learning: A retrospective, multicentre study. Lancet Digit. Health 2024, 6, e33–e43. [Google Scholar] [CrossRef] [PubMed]
  13. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  14. Wang, W.; Dai, J.; Chen, Z.; Huang, Z.; Li, Z.; Zhu, X.; Hu, X.; Lu, T.; Lu, L.; Li, H.; et al. InternImage: Exploring Large-Scale Vision Foundation Models with Deformable Convolutions. arXiv 2022, arXiv:2211.05778. [Google Scholar]
  15. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  16. Myronenko, A. 3D MRI Brain Tumor Segmentation Using Autoencoder Regularization; Springer: Cham, Switzerland, 2019; pp. 311–320. [Google Scholar]
  17. Isensee, F.; Jaeger, P.F.; Kohl, S.A.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef] [PubMed]
  18. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  19. Zhang, D.; Zhang, L.; Tang, J. Augmented FCN: Rethinking context modeling for semantic segmentation. Sci. China Inf. Sci. 2023, 66, 142105. [Google Scholar] [CrossRef]
  20. Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do Vision Transformers See Like Convolutional Neural Networks? In Proceedings of the Neural Information Processing Systems, Virtual, 6–10 December 2021. [Google Scholar]
  21. Hatamizadeh, A.; Yin, H.; Kautz, J.; Molchanov, P. Global Context Vision Transformers. In Proceedings of the International Conference on Machine Learning, Baltimore, MD, USA, 17–23 July 2022. [Google Scholar]
  22. Gu, A.; Dao, T. Mamba: Linear-Time Sequence Modeling with Selective State Spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar] [CrossRef]
  23. Gu, A.; Goel, K.; Ré, C. Efficiently Modeling Long Sequences with Structured State Spaces. arXiv 2021, arXiv:2111.00396. [Google Scholar]
  24. Gu, A.; Johnson, I.; Goel, K.; Saab, K.K.; Dao, T.; Rudra, A.; Ré, C. Combining Recurrent, Convolutional, and Continuous-time Models with Linear State-Space Layers. In Proceedings of the Neural Information Processing Systems, Virtual, 6–10 December 2021. [Google Scholar]
  25. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Liu, Y. VMamba: Visual State Space Model. arXiv 2024, arXiv:2401.10166. [Google Scholar]
  26. Huang, T.; Pei, X.; You, S.; Wang, F.; Qian, C.; Xu, C. LocalMamba: Visual State Space Model with Windowed Selective Scan. In Proceedings of the ECCV Workshops, Milan, Italy, 29 September–4 October 2024. [Google Scholar]
  27. Ma, J.; Li, F.; Wang, B.J.A. U-Mamba: Enhancing Long-range Dependency for Biomedical Image Segmentation. arXiv 2024, arXiv:2401.04722. [Google Scholar]
  28. Xing, Z.; Ye, T.; Yang, Y.; Liu, G.; Zhu, L. SegMamba: Long-range Sequential Modeling Mamba For 3D Medical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024. [Google Scholar]
  29. Liu, J.; Yang, H.; Zhou, H.-Y.; Xi, Y.; Yu, L.; Yu, Y.; Liang, Y.; Shi, G.; Zhang, S.; Zheng, H.; et al. Swin-UMamba: Mamba-based UNet with ImageNet-based pretraining. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakesh, Morocco, 6–10 October 2024. [Google Scholar]
  30. Rajagopal, A.; Nirmala, V. Convolutional Gated MLP: Combining Convolutions & gMLP. arXiv 2021, arXiv:2111.03940. [Google Scholar] [CrossRef]
  31. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  32. Ma, J.; Xie, R.; Ayyadhury, S.; Ge, C.; Gupta, A.; Gupta, R.; Gu, S.; Zhang, Y.; Lee, G.; Kim, J.; et al. The Multi-modality Cell Segmentation Challenge: Towards Universal Solutions. arXiv 2023, arXiv:2308.05864. [Google Scholar]
  33. Bernard, O.; Lalande, A.; Zotti, C.; Cervenansky, F.; Yang, X.; Heng, P.A.; Cetin, I.; Lekadir, K.; Camara, O.; Ballester, M.A.G.; et al. Deep Learning Techniques for Automatic MRI Cardiac Multi-Structures Segmentation and Diagnosis: Is the Problem Solved? IEEE Trans. Med. Imaging 2018, 37, 2514–2525. [Google Scholar] [CrossRef] [PubMed]
  34. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.; Xu, D. UNETR: Transformers for 3D Medical Image Segmentation. arXiv 2021, arXiv:2103.10504. [Google Scholar] [CrossRef]
  35. Hatamizadeh, A.; Nath, V.; Tang, Y.; Yang, D.; Roth, H.; Xu, D. Swin UNETR: Swin Transformers for Semantic Segmentation of Brain Tumors in MRI Images. arXiv 2022, arXiv:2201.01266. [Google Scholar] [CrossRef]
Figure 1. The overall architecture of the proposed Mamba-based U-Net with Dynamic Perception Feature Enhancement (DPM-UNet) for medical image segmentation (DPM-Unet). (a) The main framework of DPM-Unet. (b) The detailed design of the DPMamba Module, which internally contains a (c) VSS Block and a (d) Dynamic Perception Feature Enhancement (DPFE) Block. (e) The Dual-path Residual Fusion Module (DRFM). (f) The Multi-scale Aggregation Attention Network (MAAN).
Figure 1. The overall architecture of the proposed Mamba-based U-Net with Dynamic Perception Feature Enhancement (DPM-UNet) for medical image segmentation (DPM-Unet). (a) The main framework of DPM-Unet. (b) The detailed design of the DPMamba Module, which internally contains a (c) VSS Block and a (d) Dynamic Perception Feature Enhancement (DPFE) Block. (e) The Dual-path Residual Fusion Module (DRFM). (f) The Multi-scale Aggregation Attention Network (MAAN).
Sensors 25 07053 g001
Figure 2. Training loss curves for the three datasets.
Figure 2. Training loss curves for the three datasets.
Sensors 25 07053 g002
Figure 3. Segmentation Results (DSC) for Various Organs on the Abdomen MRI dataset.
Figure 3. Segmentation Results (DSC) for Various Organs on the Abdomen MRI dataset.
Sensors 25 07053 g003
Figure 4. Visualizations on the Abdomen MRI dataset.
Figure 4. Visualizations on the Abdomen MRI dataset.
Sensors 25 07053 g004
Figure 5. Enlarged view of the region indicated by the red box in the Abdomen MRI visualization.
Figure 5. Enlarged view of the region indicated by the red box in the Abdomen MRI visualization.
Sensors 25 07053 g005
Figure 6. Visualizations on the Microscopy dataset.
Figure 6. Visualizations on the Microscopy dataset.
Sensors 25 07053 g006
Figure 7. Enlarged view of the region indicated by the red box in the Microscopy visualization.
Figure 7. Enlarged view of the region indicated by the red box in the Microscopy visualization.
Sensors 25 07053 g007
Figure 8. Visualizations on the ACDC dataset.
Figure 8. Visualizations on the ACDC dataset.
Sensors 25 07053 g008
Figure 9. Enlarged view of the region indicated by the red box in the ACDC visualization.
Figure 9. Enlarged view of the region indicated by the red box in the ACDC visualization.
Sensors 25 07053 g009
Figure 10. Comparison of convolutional module structures. (a) Standard convolution block, (b) Dual-path convolution block, (c) Dual-path convolution block with feature mixing, (d) Our proposed DRFM block.
Figure 10. Comparison of convolutional module structures. (a) Standard convolution block, (b) Dual-path convolution block, (c) Dual-path convolution block with feature mixing, (d) Our proposed DRFM block.
Sensors 25 07053 g010
Table 1. Segmentation accuracy of different models on the Abdomen MRI dataset. The best results are displayed in bold, and the second-best results are indicated with an underscore.
Table 1. Segmentation accuracy of different models on the Abdomen MRI dataset. The best results are displayed in bold, and the second-best results are indicated with an underscore.
MethodsDSCNSD
nnU-Net76.6383.51
SegResNet73.8480.13
UNETR57.4662.82
SwinUNETR69.0874.73
Swin-UMamba74.5681.15
DPM-UNet78.1584.67
Table 2. Segmentation accuracy of different models on the Microscopy dataset. The best results are displayed in bold, and the second-best results are indicated with an underscore.
Table 2. Segmentation accuracy of different models on the Microscopy dataset. The best results are displayed in bold, and the second-best results are indicated with an underscore.
MethodsDSCF1
nnU-Net69.5554.65
SegResNet68.5154.02
UNETR71.6940.57
SwinUNETR66.6937.08
Swin-UMamba67.6049.00
DPM-UNet73.2560.23
Table 3. Segmentation accuracy of different models on the ACDC dataset. The best results are displayed in bold, and the second-best results are indicated with an underscore.
Table 3. Segmentation accuracy of different models on the ACDC dataset. The best results are displayed in bold, and the second-best results are indicated with an underscore.
MethodsDSCNSDRVMyoLV
nnU-Net91.8197.8889.4490.6095.38
SegResNet91.7197.9989.6490.2795.23
UNETR89.3495.4986.7687.4693.80
SwinUNETR91.5097.6189.4190.0395.06
Swin-UMamba91.6998.0490.0089.7895.30
DPM-UNet91.9598.0989.4690.7695.64
Table 4. Ablation study of the designed blocks on the Abdomen MRI dataset.
Table 4. Ablation study of the designed blocks on the Abdomen MRI dataset.
DRFMDPMambaCBAMMAANDSC ↑NSD ↑
75.1781.52
75.6282.40
76.0282.81
76.8483.54
78.1584.67
Table 5. Ablation study of DRFM structure on the Abdomen MRI dataset.
Table 5. Ablation study of DRFM structure on the Abdomen MRI dataset.
Dilated ConvolutionFusionResidualDSC ↑NSD ↑
76.0482.80
76.3583.06
77.5384.32
78.1584.67
Table 6. Comparison of FLOPs and Params on the Abdomen MRI dataset using our method with other models.
Table 6. Comparison of FLOPs and Params on the Abdomen MRI dataset using our method with other models.
MethodsFLOPs (G) ↓Param. (M) ↓Training Time (H)DSC ↑NSD ↑
nnU-Net2333676.6383.51
SegResNet246873.8480.13
UNETR41871757.4662.82
SwinUNETR29252069.0874.73
Swin-UMamba63593074.5681.15
DPM-UNet31382478.1584.67
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, S.; Liu, X.; Lei, H.; Hui, B. DPM-UNet: A Mamba-Based Network with Dynamic Perception Feature Enhancement for Medical Image Segmentation. Sensors 2025, 25, 7053. https://doi.org/10.3390/s25227053

AMA Style

Xu S, Liu X, Lei H, Hui B. DPM-UNet: A Mamba-Based Network with Dynamic Perception Feature Enhancement for Medical Image Segmentation. Sensors. 2025; 25(22):7053. https://doi.org/10.3390/s25227053

Chicago/Turabian Style

Xu, Shangyu, Xiaohang Liu, Hongsheng Lei, and Bin Hui. 2025. "DPM-UNet: A Mamba-Based Network with Dynamic Perception Feature Enhancement for Medical Image Segmentation" Sensors 25, no. 22: 7053. https://doi.org/10.3390/s25227053

APA Style

Xu, S., Liu, X., Lei, H., & Hui, B. (2025). DPM-UNet: A Mamba-Based Network with Dynamic Perception Feature Enhancement for Medical Image Segmentation. Sensors, 25(22), 7053. https://doi.org/10.3390/s25227053

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop