You are currently on the new version of our website. Access the old version .
SensorsSensors
  • Article
  • Open Access

9 January 2026

A New Encoding Architecture Based on Shift Multilayer Perceptron and Transformer for Medical Image Segmentation

,
,
and
1
College of Advanced Interdisciplinary Studies, Central South University of Forestry and Technology, Changsha 410004, China
2
School of Mathematics and Physics, Hunan Institute of Technology, Hengyang 421002, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Vision- and Image-Based Biomedical Diagnostics—2nd Edition

Highlights

What are the main findings?
  • A novel medical image segmentation framework integrating a Shift Multilayer Perceptron and a Transformer encoder is proposed, effectively capturing both low-level and long-range contextual dependencies.
  • The incorporation of Senet Atrous Spatial Pyramid Pooling (SASPP) and the channel Feature Aggregation Attention Module (FAAM) enhances feature representation, achieving consistent improvements in Dice coefficients (87.01% on ACDC and 79.35% on Synapse) over state-of-the-art baselines.
What are the implication of the main findings?
  • The proposed Multilayer Perceptron–Transformer (MPT) framework improves accuracy and generalization in multi-organ medical image segmentation, providing a robust foundation for clinical diagnosis and surgical planning.
  • By optimizing feature fusion and mitigating information loss in U-shaped architectures, this work contributes to the evolution of Transformer–MLP hybrid models for efficient and precise medical image analysis.

Abstract

Accurate medical image segmentation plays a crucial role in clinical diagnosis by precisely delineating diseased tissues and organs from various medical imaging modalities. However, existing segmentation methods often fail to effectively capture low-level structural details and exhibit inconsistencies in feature connection, which may compromise diagnostic reliability. To address these limitations, this study proposes a novel Multilayer Perceptron–Transformer encoding architecture that integrates the Shift Multilayer Perceptron and Transformer mechanisms. Specifically, a SENet-based Atrous Spatial Pyramid Pooling module is designed to extract multi-scale contextual representations, while the Shift MLP refines underlying spatial features. Moreover, a channel–feature aggregation attention module is introduced to strengthen information flow between the encoder and decoder layers. Experimental results on the Automatic Cardiac Diagnostic Challenge dataset show an average Dice Similarity Coefficient (DSC) of 87.01% (83.32% for the right ventricle, 90.90% for the left ventricle, and 86.83% for the myocardium). On the Synapse multi-organ segmentation dataset, the proposed model achieves an average DSC of 79.35% and a 95% Haus Dorff Distance of 20.07 mm. These results demonstrate that MPT effectively captures both local and global anatomical structures, providing reliable support for clinical diagnosis.

1. Introduction

Medical image segmentation plays a vital role in medical image analysis, enabling the accurate delineation of anatomical structures and pathological regions for diagnosis, treatment planning, and disease monitoring [1,2,3]. It provides clinicians with quantitative information that supports surgical navigation, radiotherapy planning, and the longitudinal follow-up of disease progression. However, medical image segmentation remains a challenging task due to the intrinsic complexity and variability of medical data. Images are acquired from diverse modalities such as Computed Tomography (CT), Magnetic Resonance Imaging (MRI), and ultrasound, each presenting distinct contrast mechanisms, noise characteristics, and spatial resolutions [4,5]. Moreover, substantial differences in organ morphology, boundary definition, and inter-patient variability further complicate segmentation tasks [6,7,8].
Before the deep learning era, medical image segmentation primarily relied on classical image processing techniques such as region growing, active contour models, and thresholding [9,10,11]. While these handcrafted methods could handle simple or high-contrast cases, they often failed in complex or noisy clinical environments, highlighting the need for adaptive models capable of learning semantic and structural cues directly from data. The advent of deep learning, particularly Convolutional Neural Networks (CNNs), fundamentally changed the paradigm of medical image segmentation by enabling automatic hierarchical feature extraction and end-to-end optimization [12]. These architectures effectively addressed the limitations of handcrafted methods that struggled with noise and structural variability. However, CNN-based models inherently emphasize local receptive fields, which restrict their ability to capture global contextual dependencies that are crucial for segmenting organs with complex geometries or low-contrast boundaries [13,14,15].
To mitigate this limitation, encoder–decoder architectures such as U-Net [16] introduced skip connections to recover spatial information lost during downsampling, significantly improving segmentation accuracy in small datasets with limited annotations. Despite its efficiency and widespread adoption [17,18,19,20], U-Net and its variants still face challenges in modeling long-range dependencies, maintaining structural continuity, and preserving fine-grained textures—particularly when dealing with multi-organ segmentation or images with blurred anatomical boundaries. These persistent challenges reveal a fundamental limitation of convolution-based architectures: their difficulty in jointly modeling fine-grained spatial continuity and global semantic context [21]. Consequently, recent studies have shifted toward Transformer-based designs to enhance global dependency modeling in medical image segmentation.
In pursuit of this balance, recent studies have introduced Transformer architectures [22], originally designed for natural language processing, to capture long-range dependencies in visual data. The self-attention mechanism in Transformers enables global feature interaction across all spatial locations, effectively compensating for the limited receptive field of CNNs and allowing the network to model holistic contextual relationships. In medical image segmentation, TransUNet [23] represents a pioneering hybrid design, integrating a CNN-based encoder with a Transformer module, achieving a comprehensive fusion of local spatial details and global semantic representations. Subsequent improvements, such as TransUNet++ [9] and Swin-UNet [24], refined the paradigm through hierarchical attention and shifted-window mechanisms, enhancing multi-scale feature aggregation and contextual reasoning. These Transformer-based frameworks have substantially improved segmentation accuracy and robustness across multiple benchmarks, establishing a new direction in globally contextual medical image segmentation [25,26,27,28,29,30].
However, the strong global reasoning capability of Transformers often comes at the expense of local texture and boundary precision, since self-attention prioritizes long-range interactions over spatial continuity. Moreover, empirical studies [31,32,33] have shown that repeated pooling and interpolation in U-shaped networks cause attenuation of boundary-level and texture-rich information, particularly in small or low-contrast structures. To address these challenges, we propose a Multilayer Perceptron–Transformer (MPT) framework embedded within a U-shaped encoder–decoder architecture to explicitly preserve low-level spatial features while maintaining global contextual understanding. The core of this framework is the Shift MLP, which performs spatial channel shifting before depthwise convolution, allowing each channel group to exchange information with neighboring positions and thus restore fine-grained spatial correlations weakened by pooling. By eliminating additional parameters, the design improves gradient stability and complements the Transformer’s global attention mechanism, ultimately achieving a balanced trade-off between local detail preservation and global reasoning.
To further improve multi-scale representation and feature fusion, a SENet-based Atrous Spatial Pyramid Pooling (SASPP) and a Feature Aggregation Attention Module (FAAM) are incorporated into the framework. The SASPP refines high-level semantic features across multiple receptive fields, while the FAAM enhances spatial–channel consistency, jointly improving robustness to complex anatomical structures and heterogeneous imaging conditions. The main contributions of this paper are as follows:
(1)
A new Multilayer Perceptron Transformer framework is proposed to alleviate low-level feature degradation in U-shaped architectures, achieving a more balanced integration of local detail and global context through direction-aware channel shifting;
(2)
A novel skip connection and FAAM-based fusion mechanism is introduced to improve cross-scale feature alignment and enhance the representation of complex spatial–channel relationships;
(3)
Extensive experiments conducted on two benchmark datasets demonstrate that the proposed MPT framework outperforms several state-of-the-art segmentation methods in both accuracy and robustness.

2. Materials and Methods

A new Multilayer Perceptron Transformer framework is proposed, as presented in Figure 1. It integrated the MLP mechanism into the encoding and decoding structure and enhanced medical image segmentation through an attention mechanism.
Figure 1. Overall architecture of the proposed Multilayer Perceptron–Transformer (MPT).
During the encoding process of the proposed network, multi-scale features are first extracted from the input image by adopting three Senet Atrous Spatial Pyramid Pooling (SASPP) modules. The original low-level features are preserved and reused in the decoder for cascaded feature fusion and upsampling. The multi-scale representations obtained through SASPP are then fed into the Transformer and Shift-MLP modules.
Then, the output of the multi-scale feature module is taken as the input to the Transformer. The self-attention mechanism of the Transformer is employed to enhance the long-range dependencies among image patches. Instead of explicitly adding positional encodings during patch embedding, a patch-based MLP is employed to implicitly encode positional information within the feature map after it passes through the Transformer module. Since the feature map already contains global information at this stage, it can effectively capture the position information through the convolutional layers. Subsequently, the image patches with the acquired position information will go through the MLP module to further highlight the underlying features of the encoding and decoding structure, and restore the features that integrate local and global semantic information, making their size match that of the feature map before being input into the Transformer. The decoding of the image involves a cascaded upsampling process. At each upsampling step, a channel attention enhancement module is incorporated to enhance the representational ability of features in the channel direction, and then linked with the original data in the previous encoding process. Finally, the cross-entropy loss function and the Dice loss function are used to perform gradient updates and backpropagation on the network to constrain the segmentation size of the image, so as to achieve the final segmentation task.

2.1. Senet Atrous Spatial Pyramid Pooling

To enhance the network’s feature extraction capability and overcome the difficulty of accurately segmenting organs with complex and overlapping voxels, a SENet-based Atrous Spatial Pyramid Pooling (SASPP) module is introduced, which leverages the multi-scale receptive fields of atrous convolutions and the channel reweighting mechanism of SENet to adaptively capture discriminative semantic features from medical images containing diverse spatial scales and intensity variations [34].
Firstly, the input of the encoding module E i is decomposed into sub-feature maps at multiple scales, represented as Y 1 , Y 2 ……, Y n . Features are extracted by using 3 × 3 convolutions with dilation rates of 1, 3, 5, and 8, followed by channel compression, ReLU activation, and feature restoration [35]. In the final step, the activation function sigmoid is applied to convert the channel values into a normalized range of 0-1. The resultant weight values are then multiplied by the original input features to yield the ultimate outcome of feature extraction. The formula of the algorithm is as follows:
Y = C oncatenate ( R e l u ( conv ( E i ) ) ,
Merge the result Y R H / 16 × W / 16 × 4 C as an input parameter for the Transformer [36]. The algorithm of SASPP is shown in Algorithm 1.
Algorithm 1: Senet Atrous Spatial Pyramid Pooling
I n p u t :  encoder layer   E i
S t e p   1 :   Y 1 R e l u ( c o n v 3 × 3 E i , r a t e = 1 )
S t e p   2 :   Y 2 R e l u ( c o n v 3 × 3 E i , r a t e = 3 )
S t e p   3 :   Y 3 R e l u ( c o n v 3 × 3 E i , r a t e = 5 )
I n p u t   4 :   Y 4 R e l u ( c o n v 3 × 3 E i , r a t e = 8 )
I n p u t   5 :   Y c o n c a t e n a t e ( Y 1 , Y 2 , Y 3 , Y 4 )  
Output : Y

2.2. Shift Multilayer Perceptron

After the SASPP module aggregates multi-scale feature representations, a Transformer module is employed to further model the global contextual dependencies among spatial regions [37]. The Transformer utilizes a self-attention mechanism to establish relationships between all feature tokens, enabling the network to capture long-range dependencies and construct a two-dimensional embedded feature sequence.
Formally, for the encoded feature map X R L i × d i , Weight matrix W q R d i × d k , W k R d s × d k , W v R d s × d i , Calculate the query Q, key K, and value V in the first layer of feature attention as follows:
Q = W q X ,
K = W k X ,
V = W v X ,
In the Transformer module, the output of the encoder serves as the input to the corresponding decoder layer.
Q i = Q W i Q , K i = K W i K , V i = V W i V , i = 1 , . . . . 12 ,
h e a d i = Attention ( Q i , K i , V i ) = Softmax ( Q i K i T d k ) V i ,
Multihead   ( Q , K , V ) = C o n c a c t ( h e a d 1 , . . . , h e a d i ) ,
The input feature X is converted into a sequence of one-dimensional vectors through patch embedding. Each token is then projected by the corresponding weight matrices to obtain feature representations, after which an attention weight matrix is computed to capture the dependencies among features.
h e a d i = Softmax ( Q i K i T d k ) V i ,
By merging each h e a d i into Z i = C o n c a c t ( h e a d 1 , . . . , h e a d 12 ) and performing dot-multiplication with weight W o , the output result of the Transformer layer is obtained by:
Z = Z i W o ,
To refine the feature activations produced by the Transformer encoder rather than actual mathematical eigenvalues, the Shift MLP reorganizes and reweights these activations along the spatial and channel dimensions, thereby enhancing discriminative regions and suppressing redundant responses. This refinement improves segmentation performance by strengthening boundary localization and preserving fine structural details, which are essential for accurate delineation of small and complex anatomical structures.
Before the shift operation, we permute the feature channels to promote cross-channel spatial interaction and avoid redundancy among grouped features; this enables channels from different semantic subspaces to exchange spatial context for more comprehensive feature learning. The proposed Shift MLP consists of two submodules ( M L P w i d t h , M L P h e i g h t ) . Inspired by the window-based attention in Swin Transformer, we apply horizontal shifts in M L P w i d t h followed by vertical shifts in M L P h e i g h t , enhancing local continuity while preserving the global dependencies established by the Transformer.
Firstly, the feature map X is transformed from ( B , H e m b e d , W e m b e d , E e m b e d ) to ( B , E e m b e d , H e m b e d , W e m b e d ) through an embedding operation. Secondly, after applying zero-padding with a padding size of 2, its dimension becomes ( B , E e m b e d , H p a d d e d , W p a d d e d ) . To balance computational efficiency and representational diversity, we empirically divide the feature map into five channel groups, each of size ( B , E e m b e d / 5 , H p a d d e d , W p a d d e d ) . Empirical tests show that using 4–6 groups yields stable segmentation performance, while excessive partitioning weakens receptive coverage and increases computational cost. The five-group configuration provides a good trade-off between local detail preservation and global feature diversity, consistent with similar channel-grouping strategies adopted in AS-MLP [38] and CycleMLP [39]. During the shift operation, zero values appear in the translated feature map, while the original valid features are shifted toward the edge regions that were initially zero-padded. Finally, through concatenation, each X i is recombined into a new feature map. Overall, it can be summarized as:
X 1 = Shift width ( X ) ,
X 2 = Embedding ( X 1 ) ,
X 3 = Gelu   ( Dwconv ( X 2 ) ) ,
X 4 = Shift height ( X 3 ) ,
Output ( y ) = L n ( X 2 + Gelu ( Dwconv ( Embedding ( X 5 ) ) ) ) ,
Each shift MLP block is composed of four layers and two shift modules, including a GELU activation and a normalization layer. After grouping, the features representation of the Shift MLP module is denoted as τ i R w × h × c , where w , h , and c represent the width, height, and number of channels, respectively. The shift operation can be decomposed into two steps: (1) dividing the channels into several groups, and (2) shifting each group in different directions along the width and height dimensions.
The features are first shifted upward along the width dimension and then shifted leftward along the height dimension.
τ 1 ^ 1 : w 1 , : , : τ 1 0 : w , : , : , τ 2 ^ 1 : w 1 , : , : τ 2 0 : w , : , : , τ 3 ^ 1 : w 1 , : , : τ 3 0 : w , : , : , τ 4 ^ 1 : w 1 , : , : τ 4 0 : w , : , : τ 5 ^ 1 : w 1 , : , : τ 5 0 : w , : , : τ 1 ^ : , 0 ; h 1 , : τ 1 : , 1 ; h , : , τ 2 ^ : , 0 ; h 1 , : τ 2 : , 1 ; h , : , τ 3 ^ : , 0 ; h 1 , : τ 3 : , 1 ; h , : , , τ 4 ^ : , 0 ; h 1 , : τ 4 : , 1 ; h , : , , τ 5 ^ : , 0 ; h 1 , : τ 5 : , 1 ; h , : ,
After displacement, each patch absorbs visual content from adjacent patches, and communication between different spatial positions can be smoothly completed.
The GELU layer is defined as G E L U ( x ) = x Φ ( x ) , where Φ ( x )   Φ ( x ) is the standard Gaussian cumulative distribution function, defined as Φ ( x ) = 1 2 1 + e r f ( x 2 ) . The operation of the normalization layer is to calculate the mean μ = 1 c i = 1 c x i and standard deviation σ = 1 c i = 1 c ( x i μ ) 2 from the given input vector x = x 1 , . . . , x c , and standardize each term x i ¯ = x i μ σ , i 1 , c in x = x 1 , . . . , x c .
The utilization of depthwise separable convolution following another embedding operation and parameter passing serves two purposes: (1) It aids in encoding the positional information of MLP features. The convolutional layer within the MLP block effectively captures position information and outperforms standard position encoding techniques. Unlike the position encoding used in the Vision Transformer (ViT) [32], which requires interpolation when there is a resolution mismatch between training and testing, this approach avoids performance degradation. (2) Depthwise separable convolution also offers parameter efficiency, requiring significantly fewer parameters. The resulting features are then activated using GELU. Subsequently, the resulting features are then activated using the GELU function. Subsequently, the features undergo another MLP operation (across the height dimension) to further adjust the feature size. A residual connection is employed, incorporating the original token as the residual. Finally, layer normalization is applied to transfer the output features to the subsequent block.

2.3. Feature Aggregation Attention Module (FAAM)

Due to the high sensitivity of image features to positional information and the spatial semantic inconsistencies between two output feature maps, directly applying element-wise summation may lead to information redundancy or the loss of critical details [18]. To mitigate this issue, a Feature Aggregation Attention Module (FAAM) is introduced to adaptively align and fuse multi-scale features while preserving important spatial channel information.
As shown in Figure 2, a new attention structure, FAAM, is proposed. To emphasize positional information, a new 1 × 1 convolutional layer is introduced to enhance the positional features, while another 1 × 1 convolutional layer is applied to compress the input feature map for semantic extraction, with a channel reduction ratio of 4. Then, average pooling and max pooling are used to compress the spatial dimension of the input features. A point convolution is then performed to aggregate spatial information and enhance the network’s representational capability. Semantic segmentation information is generated for two distinct spatial contexts, followed by a depthwise convolution (DW convolution) for further refinement. The detailed algorithm of FAAM is presented in Algorithm 2.
Algorithm 2: Feature Aggregation Attention Module
Input :   Block   after   connection   X
Step   1 :   X t X , X d X , X c C o n v 1 × 1 ( X )
Step   2 :   X t 1 M a x p o o l ( X t ) , X d 1 A v g p o o l ( X d )
Step   3 :   X t 2 P o i n t c o n v ( X t 1 ) , X d 2 P o i n t c o n v ( X d 1 )
Step   4 :   X D c o n c a t e n a t e ( X t 2 , X d 2 )
Step   5 :   Y c o n c a t e n a t e ( D W c o n v ( X D ) , X c )
Output :   New   block   Y
Figure 2. Feature aggregation attention module.

2.4. Spatial and Channel Skip

In U-shaped encoder–decoder architectures, the semantic gap between low-level encoder features and high-level decoder features often hinders accurate feature fusion. To alleviate this problem, a spatial and channel skip connection technique is employed to adaptively align and transmit complementary information between the encoder and decoder. Within this skip pathway, the feature maps are first divided along the channel dimension, and both max-pooling and average-pooling operations are applied to capture distinct activation responses. Unlike conventional skip connections that directly concatenate encoder features with decoder features, the proposed approach introduces an intermediate transformation through a sequence of residual convolutional layers, which refine and normalize the encoder features before fusion. This design ensures smoother feature transitions and reduces semantic inconsistencies between encoder and decoder stages. The algorithm of spatial and channel skip is shown in Algorithm 3.
As shown in Figure 3, for the i t h block D i and its corresponding original image block E i , M A X ( D i ) and A V G ( D i ) are utilized to enhance data features through a linear layer followed by activation functions. To further accelerate model convergence, Batch Normalization is incorporated after the linear layer to rescale the offset distribution of the data. This normalization ensures that the input values of the GELU activation function remain within its responsive range, thereby amplifying gradients, accelerating the learning process, and mitigating gradient vanishing issues.
Algorithm 3: Spatial and Channel Skip
Input :   U p s a m p l e d   D i , O r i g i n a l   E i
S t e p 1 : X 1 M a x p o o l D i , X 2 A v g p o o l D i
S t e p 2 : X 3 C o n c a t e n a t e ( L i n e a r X 1 , L i n e a r X 2
S t e p 3 : X 4 D W C o n v ( G e l u X 3 , k e r n e l s 1,1 )
S t e p 4 : X 5 M a x p o o l ( A v g p o o l X 4 )
S t e p 5 : Y C o n c a t e n a t e ( X 5 , E i )
Output :   Upsampled   block   Y
Figure 3. Design of the spatial and channel skip connection module.
To perform local feature extraction, depthwise separable convolution is employed. Its primary objective is to precisely capture fine-grained spatial information within the feature maps. Compared with standard convolution, DWConv efficiently analyzes subtle differences and key features in each local region of the feature maps while maintaining a lower computational cost. After the DWConv operation, the resulting feature maps proceed to a stage of cascaded max pooling and average pooling, where the combination of pooling operations further enriches the feature set and enhances the network’s representational capacity.
Subsequently, the results are concatenated with E i to form the input of the FAAM in the next layer. The decoding stage can be mathematically formulated as:
x i = Encorder ( E i ) ,
x t = FAAM ( D i ) ,
D i + 1 = skip ( x i , x t ) ,

3. Results

3.1. Data Introduction

As shown in Table 1, a total of 30 CT scans were selected, each consisting of 85–198 samples with a resolution of 512 × 512 and voxel sizes ranging from 0.54 × 0.98 × 2.5 mm3 to 0.54 × 0.98 × 5.0 mm3, resulting in 3779 axial contrast-enhanced abdominal CT images. All scans were obtained from the MICCAI 2015 Multi-Atlas Labeling Challenge dataset (Synapse). For eight abdominal organs (aorta, gallbladder, left kidney, right kidney, liver, pancreas, spleen, and stomach), the average Dice Similarity Coefficient (DSC) and average Hausdorff Distance (HD) were computed after randomly dividing the dataset into 18 training cases (2212 axial slices) and 12 validation cases. It should be noted that the Synapse dataset does not provide an official test set. Following the standard evaluation protocol adopted in previous works such as TransUNet [23], Swin-UNet [24], and MISSFormer [40], the 12 validation cases are used as the evaluation subset for reporting results. This is a commonly accepted practice to ensure fair comparison and reproducibility across the literature. To further verify the generalization capability of the proposed model, additional cross-dataset validation experiments were conducted between the Synapse (CT) and ACDC (MRI) datasets, which differ significantly in imaging modality, contrast, and anatomical structure.
Table 1. Samples selected in this work.
Regarding the dataset utilized in the context of the Automatic Cardiac Diagnostic Challenge (ACDC), a sequence of short-axis slices derived from cine magnetic resonance imaging was employed. These slices were acquired during breath-holding and encompassed the heart, ranging from the base of the left ventricle to the apex, with a thickness ranging from 5 to 8mm. The spatial resolution in the short axis plane ranges from 0.83 to 1.75 mm2/voxel. The left ventricle (LV), right ventricle (RV), and myocardium (MYO) were manually annotated for each patient scan. A total of 70 cases (1930 slices) were used for model training, 10 cases (551 slices) for validation, and 20 cases (1102 slices) for testing, to compute the final average Dice Similarity Coefficient (DSC) score.

3.2. Evaluation Indicators

To effectively evaluate the performance of the proposed model, experiments were conducted on the Synapse multi-organ segmentation dataset and the Automatic Cardiac Diagnostic Challenge dataset. Two evaluation metrics, Dice Similarity Coefficient and 95% Hausdorff Distance, were used to assess segmentation performance. In addition, comparative analyses with other state-of-the-art models were performed to further validate the effectiveness of the proposed approach.
The DSC is a widely used set-based similarity metric for measuring the overlap between two samples, with values typically ranging from 0 to 1.
D S C = 2 A B A + B ,
where A represents the predicted segmentation, and B denotes the ground truth segmentation. A B indicates the intersection between A and B , while A and B represent the number of elements in A and B , respectively. The coefficient 2 in the numerator accounts for the double-counting of shared elements between A and B in the denominator.
Hausdorff_95 is a metric that measures the degree of similarity between two sets of points. It is defined as a form of distance between two point sets. Assuming two sets, A = { a 1 , . . . ,   a p } , B = { b 1 , . . . ,   b p } , the calculation of HD95 between these two sets is given by Equation (20) and detailed as follows:
(1)
Construct a distance matrix for A and B ;
(2)
Calculate the distance from each point a A to all points in B ;
(3)
Store these distances in a matrix DA→B;
(4)
Similarly, for each point b B , calculate its distance to all points in A and store these distances in a matrix DB→A;
(5)
Sort all values from the distance matrices DA→B and DB→A in ascending order;
(6)
Find the 95th percentile value and denote it as HD95
H A , B = m a x m a x m i n | | a b | | , m a x { m i n | | b a | | } ,
where a A , b B , | | · | | denotes the L2 normal.

3.3. Implementation Details

All experiments were implemented in PyTorch 1.13.0 and conducted on a single NVIDIA RTX 3090 GPU (24 GB) with automatic mixed-precision training to reduce memory usage. Each input volume was normalized using z-score normalization on a per-volume basis, and the resulting 2D slices were resized to 512 × 512 pixels before being fed into the network. The model was optimized with stochastic gradient descent (learning rate = 0.005, momentum = 0.9, weight decay = 1 × 10−4). Online data augmentations—including random rotation (±15°), scaling (0.9–1.1), horizontal flipping, and intensity jittering (±10%)—were applied during training to enhance generalization. The batch size was 24, and the training ran for 20,000 and 14,000 iterations for the ACDC and Synapse datasets, respectively. During inference, each 3D volume was processed in a slice-wise manner, and the predicted 2D slices were stacked to reconstruct the 3D segmentation results for evaluation.

3.4. Loss Function

In the abdominal multi-organ segmentation task, significant structural differences among organs pose challenges to achieving high segmentation accuracy. To address this issue, a cross-entropy loss function with a boundary distance penalty term was adopted on the Synapse dataset. The corresponding formula is defined as follows:
L C E = i = 1 n y i l o g ( p i ) ,
w ( x , y ) = e x p ( d ( x , y ) 2 σ 2 ) ,
L = L C E + λ x , y w ( w , y ) · M ( x , y ) ,
where y i denotes the ground-truth label, p i is the probability predicted by the model, d ( x , y ) represents the distance from voxel ( x , y ) to the nearest boundary voxel. The parameter σ controls the rate of weight decay. Voxels closer to the boundary are assigned higher weights, meaning segmentation errors on these voxels are penalized more severely. M ( x , y ) is an incorrect identifier at voxel ( x , y ) . If voxel belongs to the organ region in the ground-truth label, then M ( x , y ) = 1 , otherwise, M ( x , y ) = 0 .
For the ACDC dataset segmentation task, a joint loss function was employed to optimize the training process, which can be formulated as
L C L = α L C E + β L D i c e ,
L D i c e = 1 2 i = 1 n p i y i i = 1 n y i 2 + i = 1 n p i 2
Thus, the category discrimination capability of the cross-entropy loss and the boundary sensitivity of the Dice loss are jointly exploited. This joint optimization enables the model to accurately segment vascular and myocardial tissues while effectively delineating structural boundaries, thereby enhancing the overall segmentation accuracy.

3.5. Experimental Results

3.5.1. Comparative Experiments Analysis

As demonstrated in Table 2, it is evident that the proposed network achieves significant improvements in both average DSC and HD95, surpassing the V-net by 10.54% and showcasing a superior decoding strategy compared to direct upsampling. Similarly, when compared to the U-net, the hybrid encoder demonstrates a noteworthy 9.58% increase in DSC. Furthermore, in comparison to the TransUNet and Swin-UNet [24], the incorporation of skip connections in the proposed network leads to consistently satisfactory performance across various Transformer-based architectures. In addition, DA-TransUNet, VM-UNet, and CTC-Net demonstrate their respective strengths by leveraging dual-attention mechanisms, Mamba-based state space modeling, and convolution-enhanced transformer designs. In contrast, our method is built upon a distinct Shift MLP–Transformer hybrid framework and further enhanced by multi-scale context modeling and channel-wise feature aggregation. Although it does not achieve the absolute best performance on every metric, it consistently maintains competitive results across most evaluation criteria, indicating that it serves as a complementary solution to existing segmentation frameworks.
Table 2. Quantitative comparison of different networks on the Synapse dataset in terms of Dice Similarity Coefficient (DSC, %). The best performance is highlighted in bold, and the second best is underlined.
However, as depicted in Table 3, performance optimization remains unsatisfactory for the segmentation of the spleen and stomach. This limitation may arise from the MLP module, which primarily focuses on capturing high-level semantic representations, potentially at the expense of detailed texture information.
Table 3. Average performance of different networks across seven evaluation metrics (Dice Similarity Coefficient (DSC), 95% Hausdorff Distance (HD95), Recall, Precision, Sensitivity, Specificity, and Intersection over Union (IoU)).
Additionally, the outcomes for the aorta and gallbladder still fall short compared to those achieved by conventional U-net networks. This could be attributed to the novel technological state introduced by the proposed Spatial and Channel skip connection when combined with the U-Net structure through skip connections.
Nevertheless, the network demonstrates a remarkable ability to learn high-level semantic features while preserving fine-grained structural details, which is particularly valuable in medical image segmentation. This observation is further supported by the consistent trend observed in the average HD95 metric.
Secondly, as illustrated in Figure 4, the proposed model was compared with TransUNet and Swin-UNet on the ACDC dataset. The observed improvement can be attributed to the introduction of a novel pyramid-based atrous convolution pooling mechanism, which enhances spatial feature extraction. In the downsampling stage, spatial attention is strengthened to improve the capture of high-level semantic information, while in the upsampling stage, channel-wise attention is reinforced to refine feature restoration. This joint enhancement effectively benefits both high-level semantic learning and low-level detail detection.
Figure 4. DSC values (left Y-axis, %) and HD95 scores (right Y-axis, mm) among TransUNet, Swin-UNet, DSPNet, and the proposed MPT on the ACDC dataset. The bar chart represents the Dice Similarity Coefficient (DSC, %), while the yellow line corresponds to the Hausdorff Distance (HD95, mm). Higher DSC and lower HD95 indicate better segmentation performance.

3.5.2. Ablation Experiment Analysis

To investigate the influence of different components on model performance, ablation experiments were conducted on the Synapse dataset.
The proposed improvement strategies were compared with various U-Net-based architectures. As illustrated in Figure 5, the proposed network achieved notable gains in both average DSC and HD95 compared with traditional networks incorporating attention mechanisms. Furthermore, in the multi-class DSC metrics, the proposed model demonstrated outstanding segmentation performance, particularly for organs such as the kidney, liver, pancreas, and spleen. These results clearly indicate that the MLP module effectively captures high-level semantic information, thereby contributing to enhanced segmentation accuracy.
Figure 5. Dice comparisons of different U-shaped architectures and the proposed MPT on the Synapse dataset. All quantitative values are expressed as percentages (%). Higher values indicate better segmentation performance.
To further validate the effectiveness of the skip connections and FAAMs, experiments were conducted using different numbers of skip connections and FAAMs, as illustrated in Figure 6.
Figure 6. Distribution curves of DSC and HD95 under different numbers of skip connections and FAAMs on the Synapse dataset. DSC values are expressed as percentages (%), while HD95 is measured in millimeters (mm). A higher DSC and lower HD95 denote better segmentation accuracy.
In the various skip connections, it is evident that the HD95 value exhibits a consistent downward trend as the number of skips increases. This indicates a decrease in the error magnitude when compared to actual images. Additionally, there is a negative correlation between the number of skips and the accuracy of the algorithm. Similarly, regarding the DSC, although the initial introduction of skips did not lead to significant improvements, the trend related to skips becomes increasingly obvious with an increase or decrease in the number of skips. This demonstrates that the algorithm progressively enhances its similarity to the ground-truth segmentation, confirming the effectiveness of the skip connection strategy.
To further validate the efficacy of the FAAM, additional ablation experiments were conducted under a single-branch configuration. Similarly to the skip connection analysis, a negative correlation was observed between the number of FAAMs and HD95, while a positive correlation was found between the number of FAAMs and DSC. These results indicate that increasing the number of FAAMs progressively amplifies the inherent advantages of the proposed algorithm.
As presented in Table 4, it is evident that the SASPP technique demonstrates a robust recognition effect on the Synapse dataset, particularly for the segmentation of the aorta, pancreas, and spleen. Although the individual performance improvement of the MLP module is not significant, its combination with FAAM and MLP effectively showcases the segmentation efficacy for organs with subtle voxel differences during the pre-feature extraction and processing using MLP.
Table 4. Quantitative comparison of different module combinations on the Synapse dataset in terms of Dice Similarity Coefficient (DSC, %).
The FAAM has demonstrated significant improvements in recognizing the pancreas and spleen, especially considering the relatively small anatomical area of the pancreatic region in the whole human abdominal CT scan. Leveraging the specific characteristics of the pancreatic, FAAM, and SASPP utilize convolutional neural networks to achieve precise pancreatic segmentation.
To enable the segmentation network focus on extracting target regions while minimizing the inclusion of features from other regions, a strategy is adopted where the deeper layers of the network extract increasingly intricate features. Through multiple downsampling operations, complex and redundant information is successively discarded, leading to more accurate extraction of target region features. However, excessive downsampling can result in the loss of shallow spatial information, such as edges. While this improves the positioning accuracy of the pancreatic region, it can lead to a substantial loss of edge information in the final predicted image.
As shown in Table 5, on the ACDC dataset, the morphology of each cardiac region varies, leading to different segmentation methods and difficulties for each region. Crudely speaking, the left ventricle is a thick-walled cylindrical area, while the right ventricle is an irregularly shaped object, with thinner ventricular walls that are sometimes mixed with surrounding tissues.
Table 5. Quantitative comparison of different module combinations on the ACDC dataset in terms of Dice Similarity Coefficient (DSC, %).
Individually, every single module in the proposed framework yielded only marginal improvements in cardiac segmentation. However, when these modules were integrated, notable performance enhancements were achieved. In particular, the combination of the MLP, skip, and SASPP modules led to a significant improvement in the segmentation of the right ventricle (RV). This enhancement can be attributed to the multi-layer SASPP modules, which amplify the discriminative features of the RV, MYO, and LV, thereby increasing the contrast between the target and redundant regions. Subsequently, the Transformer-based processing enables more precise delineation of the target areas, ultimately resulting in higher segmentation accuracy in the final predictions.

3.5.3. Statistical Significance Analysis

Finally, a statistical significance analysis was conducted to further assess the reliability of the proposed algorithm. For both the ACDC and Synapse datasets, independent samples t-tests were performed to evaluate the differences in segmentation performance between the proposed MPT framework and the comparative baseline methods. The t-tests were applied to determine whether the observed performance differences were statistically significant under a predefined confidence level, thereby providing a quantitative validation of the robustness and consistency of the proposed approach.
As shown in Table 6, the statistical significance analysis was conducted to validate the reliability and robustness of the proposed MPT framework on both the ACDC and Synapse datasets. Independent samples t-tests were applied to compare the segmentation performance of the proposed method with that of several representative baselines.
Table 6. Independent t-test results of segmentation performance between baseline methods and the proposed MPT framework on Synapse and ACDC datasets.
For the Synapse dataset, the MPT framework achieved higher mean Dice scores than all comparative methods. Among them, the improvement over V-Net reached statistical significance (t = 3.01, p = 0.016), indicating a clear performance advantage. Although the differences compared with DARR, TransUNet, and DSPNet did not reach the 0.05 significance level, they still demonstrated consistent positive mean gains, suggesting stable performance improvements.
For the ACDC dataset, MPT also outperformed all baseline models. The difference relative to U-Net was statistically significant (t = 2.87, p = 0.020), while the comparisons with Attention-UNet, Swin-UNet, and TransUNet exhibited smaller but consistent gains. These results confirm that the proposed architecture yields reliable and repeatable improvements across both MRI and CT segmentation tasks.
Overall, the t-test results support that the proposed MPT framework achieves statistically and practically meaningful performance advantages over conventional U-shaped architectures, thereby demonstrating enhanced robustness and generalization capability across heterogeneous medical imaging modalities.

3.5.4. Experimental Efficiency Analysis

As shown in Table 7, the proposed MPT framework demonstrates clear advantages in computational efficiency compared with TransUNet and Swin-UNet. Specifically, MPT reduces parameters and FLOPs by approximately 19% and 35%, respectively, relative to Swin-UNet, while achieving 15% faster inference and 1.1 GB lower VRAM usage under the same input resolution on an RTX 3090 GPU. These improvements mainly result from the parameter-free Shift-MLP, the lightweight SASPP with SENet re-weighting, and the FAAM that reuses intermediate features to minimize redundant computation. Overall, MPT achieves a superior balance between segmentation accuracy and computational cost, enabling efficient and scalable medical image segmentation.
Table 7. Efficiency comparison of different segmentation frameworks on the Synapse dataset.

3.5.5. Robustness Analysis Under Imaging Perturbations

To further assess the robustness of the proposed MPT framework under realistic imaging variations, additional experiments were conducted to simulate common perturbations that may occur in clinical acquisition scenarios. Two types of distortions were considered: Gaussian noise perturbation and voxel spacing distortion.
During inference, Gaussian noise with standard deviations of σ = 0.01, 0.03, and 0.05 was added to the input CT slices to simulate scanner-induced noise. In addition, spacing perturbations of ±5% and ±10% were applied to the axial voxel resolution to mimic spatial resampling or calibration inconsistencies during scanning. The quantitative results are summarized in Table 8.
Table 8. Quantitative robustness analysis of the proposed MPT framework under simulated Gaussian noise and voxel spacing perturbations on the Synapse dataset.
As shown in the table, the MPT framework exhibits a moderate but consistent degradation trend with increasing perturbation levels, where the Dice score variation remains within 1.5%, and the HD95 fluctuation is less than 1.5 mm. This demonstrates that the proposed Feature Aggregation Attention Module and spatial–channel skip connections effectively enhance feature consistency and suppress sensitivity to noise and geometric distortions.
These findings confirm the robustness and stability of the proposed MPT architecture against common degradation factors in medical imaging. Furthermore, while large-scale cross-dataset experiments on 3D volumetric data are reserved for future work due to architectural and computational limitations, the current results on Synapse (CT) datasets, together with the above perturbation analyses, clearly demonstrate the strong cross-modality generalization and stability of the proposed MPT framework.

3.6. Visualization of Image Segmentation Results

Figure 7 and Figure 8 present qualitative comparisons of segmentation results on the Synapse and ACDC datasets.
Figure 7. Visualizations of Synapse segmentation results from different angles.
Figure 8. Visualizations of ACDC segmentation results from different angles.
On the Synapse dataset, the proposed MPT framework produces more accurate organ boundaries and cleaner regional predictions compared with traditional U-shaped CNN-based networks. In the updated visualization of Figure 7, the higher-resolution images make the liver–stomach interface and pancreas areas more distinguishable, allowing a clearer observation that MPT effectively suppresses false-positive predictions and achieves more precise boundary localization. Furthermore, when compared with transformer-based models such as TransUNet, MPT yields smoother and more continuous structural contours, reflecting its stronger global context encoding ability and improved semantic discrimination compared with existing methods.
On the ACDC dataset, as shown in the updated visualization of Figure 8, the proposed MPT model produces more regular ventricular shapes and clearer myocardial delineation compared with the baseline methods. The predicted myocardial ring appears more continuous and structurally coherent, and the boundaries of the ventricular cavities are smoother and more closely aligned with the ground truth. These improvements highlight the model’s enhanced ability to preserve anatomical structure and capture fine-grained cardiac contours.
These qualitative results confirm that MPT effectively integrates multi-scale contextual information and cross-dimensional feature refinement, enabling more precise segmentation while preserving detailed structural information across different imaging modalities.

4. Discussion

The medical image segmentation method proposed in this paper, based on a shift Multilayer Perceptron and Transformer, shows significant improvements in segmentation accuracy and efficiency compared to traditional methods like Unet. By preserving low-level features, introducing self-attention mechanisms, and enhancing skip connections, the proposed model not only achieves excellent results in single-organ segmentation but also demonstrates strong generalization capability in multi-organ segmentation, thereby greatly improving the efficiency of medical image analysis. The segmented medical images provide clearer insights into quantifying and evaluating organ lesion characteristics, offering doctors accurate tools for surgical planning. Furthermore, this method has made an important contribution to the development and deeper clinical application of medical imaging technology.
However, despite these promising results, several challenges and limitations remain. First, the non-uniformity, heterogeneity, and irregular shapes of tissues in medical imaging increase the difficulty of lesion segmentation. Second, with the rapid increase in medical data, issues such as computational complexity and memory usage when dealing with high-resolution images still pose significant challenges.
The performance of the proposed framework remains closely tied to the quality, scale, and diversity of the training data. When confronted with noisy, imbalanced, or limited datasets, its segmentation accuracy may decline due to insufficient feature representation. Although the current experimental outcomes demonstrate encouraging results, there is still considerable scope for improvement through further architectural refinement, parameter optimization, and task-oriented adaptation for specific medical imaging modalities.
In future work, we plan to extend the current 2D slice-wise design toward 2.5D and 3D variants to better capture inter-slice spatial continuity and volumetric contextual dependencies. Such extensions are expected to improve the model’s ability to represent fine anatomical structures across adjacent slices while maintaining computational efficiency through lightweight 3D attention or hybrid convolutional mechanisms. Integrating these volumetric enhancements with efficient training strategies will further expand the applicability of the framework to more complex medical imaging tasks.
Nonetheless, the proposed approach exhibits strong potential in complex and clinically relevant scenarios, particularly in situations where traditional segmentation algorithms tend to falter. Its ability to generalize across diverse organ structures highlights the model’s promise for real-world medical imaging applications. Future work will focus on enhancing data robustness, expanding dataset diversity, and adapting the framework to broader clinical contexts to fully unlock its potential.

5. Conclusions

This study proposes a novel Multilayer Perceptron–Transformer (MPT) framework for medical image segmentation that effectively balances global contextual reasoning and local structural preservation through the integration of Shift-MLP, Transformer encoding, and refined skip connections. The architecture enhances feature interaction and multi-scale representation via the SENet-based Atrous Spatial Pyramid Pooling (SASPP) and the Feature Aggregation Attention Module (FAAM). Extensive experiments on the Synapse (CT) and ACDC (MRI) datasets demonstrate that the proposed model achieves high-precision segmentation with lightweight computational efficiency (51.0 M parameters, 200 ms per slice). Statistical analyses further confirm the reliability of the improvements, and robustness tests under Gaussian noise and spacing perturbations show that Dice variations remain within 1.5%, indicating strong stability under real-world imaging variations.
In summary, the proposed MPT framework provides a robust, interpretable, and efficient solution for multi-organ medical image segmentation. It advances intelligent medical imaging analysis and offers practical value for computer-assisted diagnosis and treatment planning in clinical environments.

Author Contributions

Conceptualization, H.Z. and Y.W.; methodology, H.Z. and Y.W.; software, H.Z. and Y.W.; validation, H.Z. and Y.W.; formal analysis, H.Z. and Y.W.; investigation, H.Z. and Y.W.; resources, H.Z. and Y.W.; data curation, H.Z. and Y.W.; writing—original draft preparation, H.Z. and Y.W.; writing—review and editing, J.Y. (Jieqiong Yang) and J.Y. (Jizheng Yi); visualization, J.Y. (Jieqiong Yang) and J.Y. (Jizheng Yi); supervision, J.Y. (Jieqiong Yang) and J.Y. (Jizheng Yi); project administration, J.Y. (Jieqiong Yang) and J.Y. (Jizheng Yi). All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
MPTMultilayer Perceptron Transformer
SASPPSenet Atrous Spatial Pyramid Pooling
FAAMFeature Aggregation Attention Module
DWConvDepthwise Convolution
DSCDice Similarity Coefficient
HD9595% Hausdorff Distance
ACDCAutomatic Cardiac Diagnostic Challenge (Dataset)
CECross Entropy
ASPPAtrous Spatial Pyramid Pooling
DARRDomain Adaptive Relational Reasoning Network
CBAMConvolutional Block Attention Module
DSPNetDual-Scale Perception Network

References

  1. Chen, J.; Zhang, R.; Ye, Z.; Shang, W.L.; Qiao, S.; Lyu, Z. Label-Conditioned Multi-GAN Fusion: A Robust Data Augmentation Strategy for Medical Image Segmentation. Inf. Fusion 2026, 127, 103773. [Google Scholar] [CrossRef]
  2. Ling, Y.; Wang, Y.; Dai, W.; Yu, J.; Liang, P.; Kong, D. Mtanet: Multi-Task Attention Network for Automatic Medical Image Segmentation and Classification. IEEE Trans. Med. Imaging 2023, 43, 674–685. [Google Scholar] [CrossRef] [PubMed]
  3. Messaoudi, H.; Belaid, A.; Salem, D.B.; Conze, P.H. Cross-Dimensional Transfer Learning in Medical Image Segmentation with Deep Learning. Med. Image Anal. 2023, 88, 102868. [Google Scholar] [CrossRef] [PubMed]
  4. Zhang, N.; Yu, L.; Zhang, D.; Wu, W.; Tian, S.; Kang, X.; Li, M. CT-Net: Asymmetric Compound Branch Transformer for Medical Image Segmentation. Neural Netw. 2024, 170, 298–311. [Google Scholar] [CrossRef] [PubMed]
  5. Missaoui, R.; Hechkel, W.; Saadaoui, W.; Helali, A.; Leo, M. Advanced Deep Learning and Machine Learning Techniques for MRI Brain Tumor Analysis: A Review. Sensors 2025, 25, 2746. [Google Scholar] [CrossRef]
  6. Zhou, T.; Wang, H.; Geng, S.; Ju, H.; Huang, J.; Fu, F.; Ding, W. F2CAU-Net: A Dual Fuzzy Medical Image Segmentation Cascade Method Based on Fuzzy Feature Learning. Appl. Soft Comput. 2025, 184, 113692. [Google Scholar] [CrossRef]
  7. Huang, S.; Wang, S.; Qiao, S.; Liu, Y.; Liu, T. SEMI-PLC: A Framework for Semi-Supervised Medical Image Segmentation with Pseudo Label Correction. Comput. Methods Programs Biomed. 2025, 271, 109027. [Google Scholar] [CrossRef]
  8. Lu, Y.; Wu, Y.; Chaddad, A.; Daqqaq, T.; Kateb, R. Enhancing Dual Network Based Semi-Supervised Medical Image Segmentation with Uncertainty-Guided Pseudo-Labeling. Knowl.-Based Syst. 2025, 330, 114454. [Google Scholar] [CrossRef]
  9. Wang, B.; Wang, F.; Dong, P.; Li, C. Multiscale TransUNet++: Dense Hybrid U-Net with Transformer for Medical Image Segmentation. Signal Image Video Process. 2022, 16, 1607–1614. [Google Scholar] [CrossRef]
  10. Heller, N.; Isensee, F.; Maier-Hein, K.H.; Hou, X.; Xie, C.; Li, F.; Nan, Y.; Mu, G.; Lin, Z.; Han, M.; et al. The State of the Art in Kidney and Kidney Tumor Segmentation in Contrast-Enhanced CT Imaging: Results of the KiTS19 Challenge. Med. Image Anal. 2021, 67, 101821. [Google Scholar] [CrossRef]
  11. Wei, C.; Ren, S.; Guo, K.; Hu, H.; Liang, J. High-Resolution Swin Transformer for Automatic Medical Image Segmentation. Sensors 2023, 23, 3420. [Google Scholar] [CrossRef]
  12. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  13. Zhou, B.; Dong, X.; Zhao, X.; Li, C.; Jin, Z.; Wang, H. FCT-Net: Efficient Bridge Fusion Incorporating CNN-Transformer Network for Medical Image Segmentation. IEEE Trans. Radiat. Plasma Med. Sci. 2025, 9, 762–775. [Google Scholar] [CrossRef]
  14. Li, Z.; Zheng, Y.; Shan, D.; Yang, S.; Li, Q.; Wang, B.; Zhang, Y.; Hong, Q.; Shen, D. Scribformer: Transformer Makes CNN Work Better for Scribble-Based Medical Image Segmentation. IEEE Trans. Med. Imaging 2024, 43, 2254–2265. [Google Scholar] [CrossRef]
  15. Li, D.; Jin, Z.; Guan, C.; Ji, L.; Zhang, Y.; Xu, Z.; Zhang, J. KACNet: Enhancing CNN Feature Representation with Kolmogorov–Arnold Networks for Medical Image Segmentation and Classification. Inf. Sci. 2026, 726, 122760. [Google Scholar] [CrossRef]
  16. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  17. Zhou, Z.; Siddiquee, M.M.R.; Tajbakhsh, N.; Liang, J. UNet++: Redesigning Skip Connections to Exploit Multiscale Features in Image Segmentation. IEEE Trans. Med. Imaging 2019, 39, 1856–1867. [Google Scholar] [CrossRef]
  18. Lin, S.; Hu, R.; Li, Z.; Lin, Q.; Zeng, K.; Wu, X. KAC-UNet: A Medical Image Segmentation with the Adaptive Group Strategy and Kolmogorov–Arnold Network. IEEE Trans. Instrum. Meas. 2025, 74, 5015413. [Google Scholar] [CrossRef]
  19. Pang, S.; Du, A.; Orgun, M.A.; Wang, Y.; Sheng, Q.Z.; Wang, S.; Huang, X.; Yu, Z. Beyond CNNs: Exploiting Further Inherent Symmetries in Medical Image Segmentation. IEEE Trans. Cybern. 2022, 53, 6776–6787. [Google Scholar] [CrossRef] [PubMed]
  20. Wu, R.; Liu, Y.; Liang, P.; Chang, Q. H-VMUNet: High-Order Vision Mamba U-Net for Medical Image Segmentation. Neurocomputing 2025, 624, 129447. [Google Scholar] [CrossRef]
  21. Schlemper, J.; Oktay, O.; Schaap, M.; Heinrich, M.; Kainz, B.; Glocker, B.; Rueckert, D. Attention Gated Networks: Learning to Leverage Salient Regions in Medical Images. Med. Image Anal. 2019, 53, 197–207. [Google Scholar] [CrossRef]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention Is All You Need. In Advances in Neural Information Processing Systems; NIPS Foundation: San Diego, CA, USA, 2017; Volume 30. [Google Scholar]
  23. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar] [CrossRef]
  24. Liu, X.; Gao, P.; Yu, T.; Wang, F.; Yuan, R.Y. CSWin-UNet: Transformer UNet with Cross-Shaped Windows for Medical Image Segmentation. Inf. Fusion 2025, 113, 102634. [Google Scholar] [CrossRef]
  25. Khan, A.R.; Khan, A. Multi-Axis Vision Transformer for Medical Image Segmentation. Eng. Appl. Artif. Intell. 2025, 158, 111251. [Google Scholar] [CrossRef]
  26. Qu, X.; Zhou, J.; Jiang, J.; Wang, W.; Wang, H.; Wang, S.; Tang, W.; Lin, X. EH-Former: Regional Easy-Hard-Aware Transformer for Breast Lesion Segmentation in Ultrasound Images. Inf. Fusion 2024, 109, 102430. [Google Scholar] [CrossRef]
  27. Li, K.; Yuan, F.; Wang, C. An Effective Multi-Scale Interactive Fusion Network with Hybrid Transformer and CNN for Smoke Image Segmentation. Pattern Recognit. 2025, 159, 111177. [Google Scholar] [CrossRef]
  28. Abduvaitov, A.; Buriboev, A.S.; Sultanov, D.; Buriboev, S.; Yusupov, O.; Jasur, K.; Choi, A.J. Enhancing Medical Image Segmentation and Classification Using a Fuzzy-Driven Method. Sensors 2025, 25, 5931. [Google Scholar] [CrossRef]
  29. Yang, J.; Jiao, L.; Shang, R.; Liu, X.; Li, R.; Xu, L. EPT-Net: Edge Perception Transformer for 3D Medical Image Segmentation. IEEE Trans. Med. Imaging 2023, 42, 3229–3243. [Google Scholar] [CrossRef]
  30. Xie, Q.; Chen, Y.; Liu, S.; Lu, X. SSCFormer: Revisiting ConvNet–Transformer Hybrid Framework from Scale-Wise and Spatial–Channel–Aware Perspectives for Volumetric Medical Image Segmentation. IEEE J. Biomed. Health Inform. 2024, 28, 4830–4841. [Google Scholar] [CrossRef]
  31. Neha, F.; Bhati, D.; Shukla, D.K.; Dalvi, S.M.; Mantzou, N.; Shubbar, S. U-Net in Medical Image Segmentation: A Review of Its Applications Across Modalities. arXiv 2024, arXiv:2412.02242. [Google Scholar] [CrossRef]
  32. Lee, J.C.; Byeon, K.; Song, B.; Kim, K.; Kwak, J.T. DIOR-ViT: Differential ordinal learning Vision Transformer for cancer classification in pathology images. Med. Image Anal. 2025, 105, 103708. [Google Scholar] [CrossRef]
  33. Li, W.; Qu, C.; Chen, X.; Bassi, P.R.; Shi, Y.; Lai, Y.; Yu, Q.; Xue, H.; Chen, Y.; Lin, X.; et al. AbdomenAtlas: A Large-Scale, Detailed-Annotated, and Multi-Center Dataset for Efficient Transfer Learning and Open Algorithmic Benchmarking. Med. Image Anal. 2024, 97, 103285. [Google Scholar] [CrossRef]
  34. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  35. Mohamed, N.A.; Zulkifley, M.A.; Abdani, S.R. Spatial Pyramid Pooling with Atrous Convolutional for MobileNet. In Proceedings of the 2020 IEEE Student Conference on Research and Development (SCOReD), Batu Pahat, Malaysia, 27–29 September 2020; pp. 333–336. [Google Scholar]
  36. Zheng, S.; Ye, X.; Yang, C.; Yu, L.; Li, W.; Gao, X.; Zhao, Y. Asymmetric Adaptive Heterogeneous Network for Multi-Modality Medical Image Segmentation. IEEE Trans. Med. Imaging 2025, 44, 1836–1852. [Google Scholar] [CrossRef] [PubMed]
  37. Bach, F. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning (ICML), Lille, France, 6–11 July 2015; Volume 37, p. 448. [Google Scholar]
  38. Lian, D.; Yu, Z.; Sun, X.; Gao, S. AS-MLP: An Axial Shifted MLP Architecture for Vision. arXiv 2021, arXiv:2107.08391. [Google Scholar]
  39. Chen, S.; Xie, E.; Ge, C.; Chen, R.; Liang, D.; Luo, P. CycleMLP: A MLP-like Architecture for Dense Prediction. arXiv 2021, arXiv:2107.10224. [Google Scholar] [CrossRef] [PubMed]
  40. Huang, X.; Deng, Z.; Li, D.; Yuan, X. MISFormer: An effective medical image segmentation transformer. arXiv 2021, arXiv:2109.07162. [Google Scholar]
  41. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully convolutional neural networks for volumetric medical image segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 565–571. [Google Scholar]
  42. Bai, Y.; Li, J.; Shi, L.; Jiang, Q.; Yan, B.; Wang, Z. DME-DeepLabV3+: A Lightweight Model for Diabetic Macular Edema Extraction Based on DeepLabV3+ Architecture. Front. Med. 2023, 10, 1150295. [Google Scholar] [CrossRef]
  43. Fu, S.; Lu, Y.; Wang, Y.; Zhou, Y.; Shen, W.; Fishman, E.; Yuille, A. Domain Adaptive Relational Reasoning for 3D Multi-Organ Segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention (MICCAI), Lima, Peru, 4–8 October 2020; pp. 656–666. [Google Scholar]
  44. Tang, S.; Yan, S.; Qi, X.; Gao, J.; Ye, M.; Zhang, J.; Zhu, X. Few-shot medical image segmentation with high-fidelity prototypes. Med. Image Anal. 2025, 100, 103412. [Google Scholar] [CrossRef]
  45. Sun, G.; Pan, Y.; Kong, W.; Xu, Z.; Ma, J.; Racharak, T.; Nguyen, L.M.; Xin, J. DA-TransUNet: Integrating spatial and channel dual attention with transformer U-net for medical image segmentation. Front. Bioeng. Biotechnol. 2024, 12, 1398237. [Google Scholar] [CrossRef]
  46. Ruan, J.; Xiang, S. VM-UNet: Vision Mamba UNet for medical image segmentation. arXiv 2024, arXiv:2402.02491. [Google Scholar] [CrossRef]
  47. Zhang, S.; Xu, Y.; Wu, Z.; Wei, Z. CTC-Net: A Novel Coupled Feature-Enhanced Transformer and Inverted Convolution Network for Medical Image Segmentation. In Proceedings of the Asian Conference on Pattern Recognition, Kitakyushu, Japan, 5–8 November 2023; pp. 273–283. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.