Abstract
Accurate coronary artery segmentation from two-dimensional Digital Subtraction Angiography (DSA) images is paramount for robot-assisted percutaneous coronary intervention (PCI). Still, it is severely challenged by complex background artifacts, the intricate morphology of fine vascular branches, and frequent discontinuities in segmentation. These inherent difficulties often render conventional segmentation approaches inadequate for the stringent precision demands of surgical navigation. To address these limitations, we propose a novel deep learning framework incorporating a Composite Feature Extraction Module (CFEM) and a Multi-scale Composite Attention Module (MCAM) within a U-shaped architecture. The CFEM is meticulously designed to capture tubular vascular characteristics and adapt to diverse vessel scales. In contrast, the MCAM, strategically embedded in skip connections, synergistically integrates multi-scale convolutions, spatial attention, and lightweight channel attention to enhance the perception of fine branches and model long-range dependencies, thereby improving topological connectivity. Additionally, a combined Dice-Focal loss function is employed to optimize segmentation boundary accuracy and mitigate class imbalance jointly. Extensive experiments on the public ARCADE dataset demonstrate that our method significantly outperforms state-of-the-art approaches, achieving a Dice coefficient of 76.74%, a clDice of 50.30%, and an HD95 of 57.84 pixels. These quantitative improvements in segmentation accuracy, vascular connectivity, and edge precision underscore its substantial clinical potential for providing robust vascular structure information in robot-assisted interventional surgery.
1. Introduction
Cardiovascular diseases (CVDs) remain the leading cause of global mortality, accounting for approximately 9.1 million deaths annually according to the World Health Organization’s (WHO) Global Health Estimates [,]. This pervasive health crisis underscores the urgent need for advanced diagnostic and interventional strategies. Percutaneous coronary intervention (PCI), as the primary treatment method for coronary heart disease, has been widely adopted in clinical practice. With the advancement of medical robotics, robot-assisted PCI procedures have emerged as a research hotspot in interventional therapy due to their high operational precision and reduced radiation exposure for physicians [,,,,].
In robot-assisted PCI surgeries, accurate coronary artery segmentation from two-dimensional DSA images is critical, as it serves as the core input for the surgical navigation system. Precise vessel segmentation not only facilitates the accurate planning of interventional instruments (such as guidewires and stents), helping to avoid interference from vascular branches, but also optimizes the imaging acquisition process, thereby reducing the use of contrast agents and minimizing radiation exposure for patients [,,,,]. However, the imaging characteristics of two-dimensional DSA images pose significant challenges for segmenting coronary arteries. First, the background of the photos contains artifacts from bone and soft tissues, which overlap with the grayscale intensity of blood vessels, often leading to false detections or missed regions. Second, the coronary arteries exhibit a tree-like topology, with vessel diameters varying by more than an order of magnitude between the main trunks and the delicate branches. Traditional segmentation methods struggle to extract vessels across such scales consistently. Third, intersections and curved regions of vessels are particularly susceptible to noise and artifacts, leading to segmentation errors and compromised vascular connectivity.
In recent years, methods for segmenting coronary arteries have evolved from traditional threshold segmentation and morphological operations to deep learning-based segmentation techniques. U-Net and its variants, with their symmetric encoder-decoder architecture and skip connection design, have been widely applied in the field of medical image segmentation. However, these models exhibit limited adaptability to multi-scale vascular structures and struggle to extract fine vessel branches [] accurately; AttUNet enhances the focus on vascular regions by introducing an attention mechanism into U-Net, but does not specifically optimize for the tubular morphology and branching structure of blood vessels [,,]; Although Transformer-based segmentation models are capable of capturing long-range dependencies, their high computational cost and large data requirements hinder their application in real-time clinical settings [].
Despite these advancements, current coronary artery segmentation methodologies still contend with three critical limitations: (1) Existing feature extraction modules often inadequately capture the inherent tubular morphology of blood vessels, exhibiting limited generalization across varying vessel thicknesses. (2) The capacity for modeling long-range dependencies remains insufficient, impeding the preservation of vascular connectivity, particularly within complex branching structures. (3) Loss function designs frequently fail to concurrently optimize for segmentation boundary accuracy and effectively address class imbalance, thereby constraining overall segmentation performance.
This paper proposes a multi-scale feature integration method for coronary artery segmentation, with the following key contributions:
- Composite Feature Extraction Module (CFEM): Combines dynamic snake convolution with dual-path scaling to effectively capture tubular vascular features through directional convolutions, while employing expansion-contraction operations to enhance multi-scale vessel adaptability.
- Multi-scale Composite Attention Module (MCAM): Integrates multi-scale convolution with spatial and lightweight channel attention mechanisms to improve perception of intricate vascular branches and model long-range dependencies, thereby enhancing topological connectivity.
- Hybrid Dice-Focal Loss: Strategically combines Dice and Focal losses to optimize region overlap while addressing class imbalance, particularly benefiting challenging cases like fine branches and vascular intersections.
- Comprehensive Validation: Extensive experiments on the public ARCADE dataset demonstrate statistically significant improvements in segmentation accuracy, topological connectivity, and boundary precision, validating the method’s clinical potential.
2. Related Works
2.1. Analysis of Morphological Characteristics of Coronary Vessels
The morphological characteristics of coronary vessels are the core basis for designing segmentation methods. They mainly include two types: tubular features and branching features, and their characteristics directly influence the design direction of the segmentation methods.
Cylindrical Feature
The coronary vessels (especially the aortic segment) present a distinct tubular morphology characterized by thinness, length, continuity, and directional consistency (as shown in the yellow box in Figure 1). The tubular morphology of coronary vessels has the following differences: (1) The diameter of blood vessels shows a uniform change trend with the increase in branching levels. The diameter of the main trunk can reach 2–3 mm, while that of the terminal branches is only 0.2–0.3 mm. (2) The gray value of the blood vessel wall shows a gradient change, which is prone to being confused with background artifacts. The above characteristics require the segmentation model to have strong capabilities in extracting tubular features and adapting to different scales. It must accurately capture the main vessels while avoiding missing the delicate branches.
Figure 1.
The image in panel (a) presents a coronary artery angiogram depicting the vascular structure, which includes both the main and fine branches. The yellow box highlights the aortic segment, and the blue box indicates a branch with complex morphology. The segmentation target in panel (b) captures the coronary vessels with the mask annotations clearly delineating the vascular structures. The background artifacts, including overlapping tissues and bones, were handled effectively by our model, reducing false positive detections. Methodologically, the CFEM played a key role in enhancing the extraction of tubular features, especially in the fine branches, across varying vessel scales.
Coronary vessels have a tree-like topological structure. The distal end of the aorta can be divided into the left anterior descending branch, the left circumflex branch, the right coronary artery, etc., as the main branches, and each branch is further divided into multiple levels of fine branches (as shown in Figure 1, the blue box). The projection characteristics of two-dimensional DSA images cause the superimposition of different depths of vascular branches, which alters the uniform change trend of vascular diameters and increases the difficulty of segmentation [,,,,]. Moreover, areas with vascular intersections and bends are prone to sudden drops in gray values, making it difficult for traditional models to maintain segmentation connectivity. Therefore, a specially designed long-distance dependent modeling module is required to restore the dendritic structure of the blood vessels.
2.2. Key Issues in Coronary Vessel Segmentation
2.2.1. Accuracy Issue
The accuracy issue primarily manifests as false detections and missed detections in pixel-level segmentation, making it challenging to restore the proper anatomical shape of coronary vessels accurately. The main reasons include: (1) The gray values of blood vessels are close to those of the background tissues (such as myocardium and veins), resulting in low contrast; (2) The proportion of pixels of small branches (diameter < 0.5 mm) is low, and the model easily ignores them; (3) The gray values of the lesion areas (such as stenosis and calcification) are abnormal, causing blurred segmentation boundaries. As shown in Figure 2, there are obvious mis-segmentation (the background area is identified as a blood vessel) and missed-segmentation (the blood vessel area is not recognized) phenomena in the segmentation results, which seriously affect the reliability of surgical navigation.
Figure 2.
Panel (a) presents the original image with visible segmentation challenges, such as missed and mis-segmented regions. The label in panel (b) highlights the ideal segmentation mask, while panel (c) shows the segmentation result from our method. Notably, our method addresses accuracy issues, specifically in small vessels and intersections, where conventional methods often fail due to the low contrast and fine details. The CFEM ensures that fine branches are retained, significantly reducing false positives and false negatives by focusing on tubular morphology and fine-scale features.
2.2.2. Connectivity Issues
The connectivity problem manifests as the breakage or incorrect connection between the main vessels and the branch vessels in the segmentation results, thereby disrupting the topological structure of the vascular network (Figure 3). The main reasons include: (1) The receptive field of traditional convolutional networks is limited, making it challenging to capture long-distance dependencies across branches; (2) The vascular crossing areas are affected by projection superposition, resulting in complex gray value distribution, which leads to model misjudgment of vascular continuity; (3) Noise and motion artifacts interfere, causing discontinuous segmentation of delicate branches. The connectivity problem can lead to misjudgment of vascular paths by the surgical navigation system, increasing surgical risks. Therefore, it is a problem that the segmentation method needs to address.
Figure 3.
In panel (a), the original DSA image depicts complex vascular intersections, where the model often struggles with connectivity. The label in panel (b) shows the correct connectivity of vascular branches, and panel (c) illustrates the segmentation result. Our method resolves the breakage in vascular networks, especially at crossing points and sharp bends. The Multi-scale Composite Attention Module (MCAM) is instrumental in improving connectivity by modeling long-range dependencies and integrating multi-scale features, which is particularly critical for vascular branches that span across distant regions.
2.3. Classification of Existing Partitioning Methods
Depending on the different technical routes, the existing methods for segmenting coronary vessels can be classified into three categories:
- 1.
- Traditional segmentation methods: The traditional methods mainly rely on low-level features such as image gray values and textures for segmentation, including threshold segmentation, edge detection, and morphological operations. For instance, Hassouna et al. [] achieved brain vessel segmentation based on a random model, but it was sensitive to noise; Sun et al. [,,] extracted vascular trees by combining morphological multi-scale enhancement with the watershed algorithm, but it was difficult to handle complex background artifacts.
- 2.
- Segmentation method based on deep learning: Based on deep learning methods, which possess the ability of end-to-end learning [], they have become the mainstream technology for coronary artery segmentation at present. U-Net [], the benchmark model for medical image segmentation, integrates high- and low-level features through skip connections; however, its adaptability to multi-scale blood vessels is insufficient. MedNeXt [] employs a scaling architecture to expand the receptive field, thereby enhancing its ability to segment complex structures; however, it has not been optimized for vascular tubular features. DSCNet [] combines dynamic snake convolution to capture tubular features and has a relatively high segmentation accuracy. However, it cannot model long-distance dependencies and is prone to connectivity issues.
- 3.
- Segmentation method based on Transformer: The Transformer architecture achieves long-range dependency modeling through the self-attention mechanism, providing a new approach to solving the vascular connectivity problem. TransUNet [] combines the Transformer and U-Net architectures, enhancing its ability to perceive the overall structure. Swin-Unet [] reduces computational costs through hierarchical window attention; however, it still faces challenges such as high data requirements and slow inference speed, which make it difficult to meet the real-time navigation requirements in clinical settings.
2.4. U-Net in Broader Biomedical Segmentation Context
The U-Net architecture and its variants have demonstrated remarkable versatility beyond human medical image analysis, proving effective in diverse segmentation tasks involving tubular and fine structures. These advancements in adjacent fields often provide valuable insights for coronary artery segmentation.
For instance, in the field of agricultural biometrics, U-Net-based approaches have been successfully employed for tasks such as cattle identification and recognition []. This application highlights the model’s capability to capture unique, intricate patterns from visual data, which parallels the challenge of discerning individual vascular branching patterns in DSA images.
More directly relevant are advancements in retinal vessel segmentation, which shares the core challenges of segmenting thin, tree-like structures from a complex background. Recent work, such as the Dense U-Net based on patch-based learning [], addresses issues like class imbalance and the precise delineation of micro-vessels. The patch-based learning strategy is particularly effective in focusing on local details of fine vessels, a common challenge in segmenting distal coronary branches. While our method is tailored to the specific artifacts and projection characteristics of DSA imaging, the successes of these approaches in other domains underscore the potential of innovative U-Net adaptations to overcome challenges in vascular morphology capture and connectivity preservation.
Our proposed CFEM and MCAM contribute to this ongoing evolution by offering a specialized solution that simultaneously addresses the tubular feature extraction, multi-scale adaptability, and long-range connectivity challenges inherent to 2D coronary artery DSA images.
3. Methods
3.1. Overall Network Architecture
The proposed coronary artery segmentation network is architected upon a U-shaped symmetrical encoder-decoder framework, fundamentally enhanced by the integration of our novel Composite Feature Extraction Module (CFEM) and Multi-scale Composite Attention Module (MCAM) (Figure 4). This architecture is characterized by:
Figure 4.
This figure presents the overall architecture of the proposed coronary artery segmentation framework. The input DSA images first undergo feature extraction through the Composite Feature Extraction Module (CFEM), which captures tubular vascular features and adapts to varying vessel scales. The extracted features are then processed through the Multi-scale Composite Attention Module (MCAM) embedded in the skip connections to enhance fine-branch perception and long-range dependency modeling. Down-sampling and up-sampling blocks progressively extract hierarchical representations and restore spatial resolution. The final segmentation output delineates coronary vessels of different calibers with improved continuity and boundary accuracy.
- 1.
- Encoder: Comprising four sequential stages, each consisting of a CFEM followed by a downsampling block. Each downsampling block, implemented via a convolutional operation with a stride of 2, progressively reduces the feature map resolution by half while doubling the channel count, thereby expanding the receptive field and extracting hierarchical global features.
- 2.
- Decoder: Structured with four corresponding stages, each featuring an upsampling block followed by a CFEM. Upsampling is achieved through transposed convolution, which restores the feature map resolution. Crucially, multi-scale features from the encoder are integrated into the decoder via skip connections, where the MCAM plays a pivotal role in refining these features before they are fused with the upsampled features.
- 3.
- Bottleneck: A single CFEM is positioned at the network’s bottleneck, responsible for extracting the highest-level global features and providing essential contextual information to the decoder.
- 4.
- Output Layer: The final segmentation probability map is generated by a 1 × 1 convolutional layer, which maps the feature channels to two (representing foreground vessel and background), followed by a sigmoid activation function.
The input and output dimensions of each module in the network follow the following rules: (1) The downsampling block reduces the feature map size by 1/2 and doubles the number of channels; (2) The upsampling block expands the feature map size by 2 times and halves the number of channels; (3) The CFEM and MCAM maintain the feature map size and channel number unchanged to ensure the compatibility of feature fusion between the modules.
3.2. Composite Feature Extraction Module (CFEM)
The CFEM is designed based on the characteristics of vascular tube shapes and scale variations (Figure 5). It consists of a tube-shaped feature extraction layer and a dual-path scaling architecture, enabling the precise extraction of vascular features and scale generalization [,,].
Figure 5.
This figure illustrates the design of the CFEM. Panel (a) shows the Combined Snake Convolution Block, followed by the Feature Expansion Block and Depthwise Convolutional Layer, which enhance feature extraction. Panel (b) demonstrates the expansion-contraction path using 1 × 1 convolution layers, with CELOU activation, adapting to varying vessel scales. Panel (c) depicts the directional dynamic snake convolutions along the x and y axes, integrated with group normalization and a 3 × 3 convolutional layer. The CFEM improves the representation of vascular morphology and scale consistency, significantly boosting segmentation accuracy and robustness.
3.2.1. Cylindrical Feature Extraction Layer
The tubular feature extraction layer captures the tubular morphological characteristics of blood vessels directionally through dynamic snake convolution in the x-axis and y-axis directions. The specific steps are as follows:
- 1.
- Group Normalization: Apply Group Normalization (GN) to the input feature maps to reduce the impact of batch size on training and stabilize the training process.
- 2.
- Directional convolution: The normalized feature maps are respectively input into the dynamic snake-shaped convolutions along the x-axis (9 × 1 convolution kernel) and y-axis (1 × 9 convolution kernel), resulting in two directional tubular feature maps.
- 3.
- Feature Integration: Concatenate Fx and Fy along the channel dimension, and integrate the features from both directions through a 3 × 3 ordinary convolution to achieve comprehensive capture of tubular features.
- 4.
- Activation function: Perform group normalization on the integrated feature map, combine with the SiLU activation function to enhance the non-linear expression ability of the features, and output tubular feature maps.
The above process can be expressed by Formula (1).
3.2.2. Dual-Path Scaling Architecture
The dual-path scaling architecture expands and contracts convolution operations to expand the feature receptive field, enhancing the model’s adaptability to different scales of blood vessels. The specific steps are as follows:
- 1.
- Feature routing: The tubular feature map (Fconcat) is divided into two paths, which are respectively input into the expansion convolution (Exp) layer.
- 2.
- Extended convolution: Utilizing a 1 × 1 convolution kernel with a 4-fold expansion rate, the feature receptive field is expanded by a factor of 4, enabling the capture of large-scale vascular features.
- 3.
- Feature Interaction: Apply the GELU activation function to the output of one path, and perform matrix multiplication with the production of another path to achieve feature interaction and scale adaptation.
- 4.
- Contraction convolution: Utilizing a 1 × 1 convolution kernel with a 4-fold contraction rate, the receptive field is restored to its original size, ensuring that the output feature map is consistent with the input size.
- 5.
- Deep convolution: By using 7 × 7 deep convolution to integrate features, the number of parameters is reduced and the computational efficiency is improved;
- 6.
- Residual Connection: Before the output of the module, a residual connection is introduced, where the input feature map (Fin) is added to the processed feature map, enhancing the robustness of the network and preventing the vanishing gradient.
The above process can be expressed by Formula (2):
3.2.3. Multi-Scale Composite Attention Module (MCAM)
The Multi-scale Composite Attention Module (MCAM) is strategically embedded within the skip connection path of the U-shaped network. Its primary function is to refine the features transmitted from the encoder to the decoder by leveraging a synergistic combination of multi-scale feature processing, spatial attention, and lightweight channel attention. Upon receiving the feature map from the skip connection, the MCAM performs internal multi-scale feature extraction. This process involves: 1. Channel Compression: Initially, a 1 × 1 convolution is applied to the input feature map to reduce its channel dimension to 1/4, thereby optimizing computational efficiency. 2. Multi-scale Convolution: The compressed feature map is then branched into multiple parallel paths. Specifically, three distinct paths are processed by convolutional layers employing 1 × 1, 3 × 3, and 5 × 5 kernels, respectively, enabling the extraction of features at varying receptive field scales. 3. Feature Fusion: The outputs from these multi-scale convolutional paths are subsequently concatenated along the channel dimension. This aggregated multi-scale feature representation is then further integrated through group normalization and another 1 × 1 convolution, yielding a rich, fused feature map (Fms) that captures diverse spatial contexts. This Fms then serves as the input for subsequent spatial and channel attention mechanisms, enhancing the perception of vascular branch structures and ultimately improving segmentation connectivity.
Spatial attention enhances the spatial weights of vascular regions and suppresses interference from background artifacts. The specific steps are as follows: 1. Pooling operation: Perform average pooling and maximum pooling on the fused feature maps () respectively, obtaining two spatial mapping images, and . 2. Feature concatenation: Concatenate and along the channel dimension to obtain a 2-channel feature map. 3. Convolutional activation: Through a convolutional layer, spatially dependent features are extracted. Combined with the SiLU activation function and the sigmoid function, a spatial attention map is generated. 4. Feature weighting: Multiply the spatial attention map element-wise with the fused feature map to obtain the spatially weighted feature map .
The above process can be expressed by Formula (3):
The channel attention module adopts the lightweight SimAM. By calculating the energy function of pixels within the channel, it generates channel weights that enhance the key feature channels. The specific steps are as follows: 1. Statistical calculation: For each channel of the feature map input to the MCAM, calculate the pixel mean and variance; 2. Energy function calculation: Based on the mean and variance, the energy value Ec for each channel is calculated using Formula (4). The smaller the energy value, the higher the importance of that channel for segmentation.
Weight Generation: The energy values are mapped to channel weights through the sigmoid function, with the weight range being [0, 1]. Feature weighting: Multiply the channel weights () with the input feature map () channel by channel to obtain the channel-weighted feature map ().
The above process can be expressed by Formula (5):
4. Results
4.1. Dataset Introduction
The experiment utilized the ARCADE public dataset [], a comprehensive collection of 1500 pairs of coronary artery Digital Subtraction Angiography (DSA) images meticulously labeled by medical experts []. Each image, initially sized at 512 × 512 pixels, encompasses diverse patient cases and various types of vascular lesions, making it highly suitable for the intricate task of coronary artery segmentation. To rigorously assess the method’s generalization capability, the dataset was systematically partitioned into a training set, validation set, and test set, adhering to a ratio of 1000:200:300, respectively. Furthermore, to augment data diversity and bolster the model’s robustness against variations, the training set underwent a three-stage image enhancement protocol, incorporating brightness and contrast adjustments, CLAHE equalization, and a suite of geometric transformations, including rotation, flipping, blurring, and affine transformations.
4.2. Experimental Environment and Parameter Settings
The experiment was implemented based on the PyTorch 1.12 deep learning framework. The hardware environment consisted of an Intel Xeon Gold 6338 CPU, 64 GB RAM, and an Nvidia RTX A6000 GPU (with 48 GB VRAM). The detailed configuration of the model training hyperparameters is summarized in Table 1.
Table 1.
Summary of training hyperparameters.
4.2.1. Evaluation Index
To comprehensively evaluate the performance of the proposed segmentation method, five widely recognized evaluation metrics were adopted, each addressing a distinct aspect of segmentation quality:
- 1.
- Dice Similarity Coefficient (Dice): Measures the degree of overlap between the segmented area and the actual area, with a range of . The higher the value, the higher the segmentation accuracy [].
- 2.
- Precision: Measures the proportion of pixels predicted as blood vessels among those that are actually blood vessels. The value ranges from 0 to 1. The higher the value, the lower the false detection rate.
- 3.
- Recall Rate: Measures the proportion of pixels that are actually blood vessels and are correctly predicted. The value ranges from 0 to 1, with a higher value indicating a lower rate of missed detections.
- 4.
- Central Line Dice (clDice): Measures the overlap degree between the segmented vessel centerline and the true centerline. The value ranges from 0 to 1, with a higher value indicating better connectivity.
- 5.
- 95% Hausdorff Distance (HD95): Measures the maximum distance between the segmentation boundary and the actual boundary, with the unit being pixels. The smaller the value, the higher the boundary accuracy.
4.2.2. Comparison of Experimental Results and Analysis
To rigorously validate the efficacy of the proposed methodology, five state-of-the-art segmentation methods were meticulously selected for comparative analysis. These include U-Net [], widely regarded as a foundational benchmark in medical image segmentation; AttUNet [], an attention-enhanced variant of U-Net designed to improve feature selectivity; MedNeXt [], a scaling architecture segmentation model known for its ability to capture multi-scale contexts; DSCNet [], which incorporates dynamic snake convolutions to better delineate tubular structures; and SPNet [], a model leveraging strip pooling for efficient spatial context aggregation. The comprehensive results of this comparative experimental evaluation are presented in Table 2.
Table 2.
The results of comparison experiment.
As can be seen from Table 2, the method proposed in this paper outperforms the comparison methods in all evaluation indicators:
- Dice coefficient: The method in this paper reaches 76.74%, which is 4.48%, 4.23%, 1.64%, 3.26%, and 20.84% higher than U-Net (72.26%), AttUNet (72.51%), MedNeXt (75.10%), DSCNet (73.48%), and SPNet (55.90%) respectively. This indicates that the proposed method has a significant advantage in the overlap degree of the segmented regions.
- clDice coefficient: The method in this paper achieves 50.30%, which is an improvement of 1.54% compared to the best-performing method in the comparison (MedNeXt, 48.76%), indicating that the multi-scale composite attention module effectively improves the segmentation connectivity.
- HD95 distance: The method in this paper achieves 57.8358 pixels, which is 18.18 pixels lower than U-Net (76.0132), 14.18 pixels lower than AttUNet (72.0205), 4.12 pixels lower than MedNeXt (61.9603), 11.69 pixels lower than DSCNet (69.5269), and 21.56 pixels lower than SPNet (79.3913). This indicates that the combined loss function effectively improves the accuracy of the segmentation boundary.
- Precision and Recall: The Precision of the method proposed in this paper reaches 80.66% and the Recall reaches 74.87%, both of which are superior to the comparison methods. This indicates that the proposed method has a significant effect in reducing the false detection rate and missed detection rate.
The results of the ablation experiments (in Table 2, Our Method without Attention Module called Ours-Att) show that the MCAM plays a significant role in improving the segmentation performance: after adding the MCAM, the Dice coefficient increased by 1.35%, the clDice coefficient increased by 0.28%, and the HD95 distance decreased by 4.04 pixels, verifying the effectiveness of the multi-scale composite attention module in improving the connectivity and boundary accuracy of the segmentation.
To assess the statistical significance of the performance improvements observed in our experiments, we conducted a comprehensive statistical analysis based on the aggregated Dice scores across all 300 test images. The results of this analysis, presented in Table 3, demonstrate that the improvements of our method over all baseline approaches are statistically significant. The consistent and substantial performance gaps are quantified through both mean Dice differences and effect size measures (Cohen’s d). Notably, all comparisons yield estimated p-values below 0.001, indicating high levels of statistical significance. The exceptionally large effect sizes, particularly against SPNet (Cohen’s –), provide strong evidence that the observed improvements are substantial and unlikely to occur by chance.
Table 3.
Estimated Statistical Significance Analysis Based on Aggregated Performance Data.
The exceptionally low p-values (all far below the 0.01 threshold) provide strong statistical evidence that the superior performance of our method, as reflected by the Dice coefficient, is not a random occurrence but a consistent and significant improvement across the test dataset.
Figure 6 presents a visual comparison of the segmentation results obtained by different methods. As can be seen from the figure:
Figure 6.
This figure compares the segmentation results from various models, including U-Net, AttUNet, DSCNet, MedNeXt, and our approach. The first row displays the original image, with the second row showing the ground truth (Real). Subsequent rows present the segmentation results for each method. The yellow boxes highlight fine branches that are missed or poorly segmented by other models, while our method accurately segments fine branches at the aortic end. Additionally, our model excels in suppressing background artifacts and improving connectivity at vascular crossings, as demonstrated in the red boxes. This performance is attributed to the synergistic effect of the CFEM and MCAM.
The segmentation comparison in Figure 6 demonstrates the superior performance of the proposed method across multiple models, including U-Net, AttUNet, DSCNet, and MedNeXt [,]. The figure presents the original DSA images and corresponding ground-truth labels (Real), followed by the segmentation results obtained by each model. The comparison methods generally have segmentation breakage phenomena, especially SPNet []. Due to its strip-shaped pooling, it is only applicable to linear objects. It is difficult to handle the dendritic vascular structure, resulting in the poorest connectivity of the segmentation results. Compared with the reference methods, our network shows notably better delineation of fine vascular branches (highlighted by the yellow boxes), where U-Net and AttUNet exhibit omissions and MedNeXt and DSCNet produce partial discontinuities. Moreover, our method maintains stronger connectivity in regions of vascular intersections and curvature (marked by the red boxes), effectively suppressing background artifacts and false detections. These improvements confirm that the synergistic integration of the CFEM and MCAM enable more precise modeling of vessel morphology, resulting in enhanced continuity and topological integrity of the segmented coronary arteries.
Figure 7 presents the visualization results of the ablation experiment for the MCAM. After the addition of the MCAM, the integrity and connectivity of the vascular branches have significantly improved (as shown in the red box in Figure 7), especially in the areas of vascular intersections, where the occurrence of fractures has been markedly reduced, thereby verifying the effectiveness of the MCAM in long-distance dependent modeling.
Figure 7.
This figure compares the segmentation results of our method with the ground truth (Real) and the results from other models (Our work). The first row shows the original images. The second and third rows display the segmentation outputs, where the red boxes highlight areas of interest. Our method (shown in the second and third rows) provides more accurate and continuous segmentation, particularly in complex regions, compared to the ground truth and other models. This demonstrates the effectiveness of our approach in handling fine branches and vascular intersections.
To verify the effectiveness of the proposed module, three sets of ablation experiments were designed: (1) The basic model (U-shaped architecture + ordinary convolution + Dice loss); (2) The basic model + CFEM; (3) The basic model + CFEM + MCAM + combined loss function (the complete method in this paper).
The function of the CFEM: After adding the CFEM, the Dice coefficient increased from 71.52% to 75.39%, the clDice coefficient rose from 46.83% to 50.02%, and the HD95 distance decreased from 68.32 pixels to 61.88 pixels. This indicates that the CFEM effectively improved segmentation accuracy and connectivity by extracting tubular features and applying scale generalization.
The function of the MCAM: After adding the MCAM, the Dice coefficient further increased to 76.74%, the clDice coefficient rose to 50.30%, and the HD95 distance decreased to 57.84 pixels. This indicates that the MCAM further improved the segmentation connectivity and boundary accuracy through long-distance dependency modeling.
The function of the combined loss function: Compared with using the Dice loss alone, the combined loss function reduced the HD95 distance by 3.21 pixels. This indicates that the Focal loss played a crucial role in learning complex samples, thereby improving the accuracy of the segmentation boundary.
To assess the feasibility of the clinical application of the method, the computational complexity of different methods was analyzed, including the number of parameters (Params) and inference time (Inference Time).
Parameter quantity: The parameter quantity of the method proposed in this paper is 12.8 M, which is lower than AttUNet (15.6 M), MedNeXt (18.2 M), DSCNet (14.5 M), and only higher than U-Net (8.5 M) and SPNet (10.3 M). This suggests that the proposed module enhances performance without significantly increasing the model’s parameter count.
Inference time: The single-image inference time of this method is 0.12 s, meeting the clinical real-time navigation requirements (requiring < 0.5 s). It is lower than AttUNet (0.18 s), MedNeXt (0.21 s), and DSCNet (0.15 s), and only higher than U-Net (0.08 s). This shows that the proposed method has advantages in computational efficiency.
The results above indicate that the method proposed in this paper achieves a good balance between performance and computational complexity, and has potential for clinical applications.
5. Discussion
5.1. Methodological Performance Advantages and Clinical Significance
The proposed multi-scale feature-integrated coronary artery segmentation method demonstrably overcomes the inherent limitations of conventional approaches in terms of segmentation accuracy, connectivity, and boundary precision. This superior performance is attributable to the synergistic interplay of the CFEM, the MCAM, and the judiciously applied combined loss function. The primary performance advantages are multifaceted:
Enhanced Multi-scale Vascular Adaptability: The CFEM, through its innovative combination of dynamic snake convolution and a dual-path scaling architecture, achieves unparalleled precision in extracting blood vessels across a broad spectrum of diameters. Notably, for fine branches with diameters less than 0.5mm, a critical challenge in coronary angiography, our method exhibits a significantly improved segmentation recall rate, as evidenced by the ablation studies. This capability is crucial for comprehensive vascular mapping.
Robust Long-range Dependency Modeling: The MCAM effectively captures the intricate long-range dependency relationships characteristic of dendritic vascular branches. By integrating multi-scale feature fusion with sophisticated spatial and channel attention mechanisms, it significantly enhances the topological connectivity of segmentation results. Quantitatively, the clDice coefficient, a direct measure of connectivity, demonstrates an improvement of over 1.5% compared to the best-performing existing methods, underscoring its efficacy in maintaining vascular integrity.
Optimized Boundary Accuracy: The combined Dice-Focal loss function, achieved through a weighted fusion strategy, concurrently optimizes the overlap measure of segmented regions and prioritizes the learning of challenging samples. This dual optimization leads to a substantial reduction in boundary inaccuracies, with the HD95 metric decreasing by 4 to 18 pixels compared to various existing methods. This precision is vital for accurate surgical planning and intervention.
The precise segmentation achieved by our method provides a crucial foundation for detecting arterial stenosis and guiding therapeutic interventions. By accurately delineating vessel boundaries and preserving topological connectivity, our approach enables reliable diameter measurement along coronary segments, facilitating the identification of localized narrowings indicative of stenosis. The improved segmentation of fine branches and complex bifurcations is particularly valuable for detecting distal lesions that are challenging to visualize in conventional angiography. Furthermore, the detailed vascular roadmap generated by our method offers essential guidance for stent placement planning, allowing clinicians to precisely determine optimal stent sizes, lengths, and deployment locations. The real-time segmentation capability (0.12 s per frame) ensures that this guidance can be integrated into live intervention workflows, potentially reducing procedure time and improving the accuracy of device positioning.
From a clinical perspective, the segmentation results of the method presented in this paper can provide the following support for robot-assisted PCI surgeries: Precise path planning: Accurate vascular segmentation results can provide detailed vascular path information for surgical instruments, helping to avoid branch interference and reducing the risk of surgical complications. Radiation exposure control: By optimizing the imaging acquisition process, reducing the amount of contrast agent used, and the number of imaging acquisitions, the radiation exposure for patients and doctors is decreased. Improvement of surgical efficiency: Real-time segmentation capability (0.12 s per frame) can reduce surgical time, enhance surgical efficiency, and alleviate patient pain.
5.2. Methodological Limitations
Although the method presented in this paper has achieved good segmentation performance, it still has the following limitations:
- Insufficient model lightweighting: Although the computational complexity of this method is lower than that of most comparison methods, for resource-constrained embedded devices (such as local computing units of surgical robots), the model parameter size and computational load still need to be further reduced.
- Lack of multimodal data fusion: The current method only performs segmentation based on two-dimensional DSA images, lacking information on the microscopic structure of the vessel wall (such as plaque properties), making it challenging to meet the precise segmentation requirements of the lesion area.
- Generalization ability needs to be improved: The performance of the method depends on the distribution characteristics of the ARCADE dataset. On DSA images collected from different hospitals and different devices, the generalization ability may decline.
Furthermore, upon detailed visual inspection of the segmentation results (Figure 6), two specific limitations were observed, which align with the inherent challenges in coronary artery segmentation.
First, in certain cases, the proposed model exhibits a slight shortening of detected vessel branches (e.g., the second image under ‘Our method’ in Figure 6). This typically occurs at the distal ends of very fine vessels where the contrast-to-noise ratio is extremely low, causing the model’s confidence to drop below the segmentation threshold.
Second, in complex regions with overlapping or adjacent vessels (e.g., the third image under ‘Our method’ in Figure 6), the segmentation boundaries can become blurred, reducing the fidelity of individual vessel delineation. This is primarily due to the projection nature of 2D DSA, where vessels from different depths superimpose, creating ambiguous image features that are challenging even for expert annotation.
While our method shows overall superior performance, these edge cases highlight opportunities for further refinement.
5.3. Future Research Directions
In response to the aforementioned limitations, future research will be conducted in the following directions:
- Model lightweight optimization: Utilize techniques such as model pruning, quantization, and knowledge distillation, combined with lightweight network architectures (such as MobileNet, EfficientNet), to further reduce the model’s parameters and inference time while maintaining accuracy, in order to meet the requirements of embedded devices.
- Multi-modal data fusion: Combine modal data such as intravascular ultrasound (IVUS) and optical coherence tomography (OCT), integrate macrovascular structure (DSA) and microvascular wall information (IVUS/OCT), to improve the segmentation accuracy of lesion areas (such as stenosis, calcification).
- Generalized segmentation model adaptation: Based on large-scale medical image pre-training generalized segmentation models (such as SAM-Med2D), through transfer learning, adapt the coronary artery segmentation task, reduce the reliance on labeled data, and enhance the model’s generalization ability on different datasets.
- Multi-task joint learning: Combine coronary artery segmentation with functions such as stenosis detection and surgical path planning, design a multi-task joint learning model, achieve an end-to-end surgical navigation system, and further improve surgical efficiency and safety.
6. Conclusions
This paper addresses the issues of complex background, large-scale differences of blood vessels, and poor connectivity in the segmentation of coronary vessels in two-dimensional DSA images. It proposes a deep learning segmentation method that integrates multi-scale features. By designing a composite feature extraction module (CFEM), a multi-scale composite attention module (MCAM), and a Dice-Focal combined loss function, it achieves precise extraction of vessels of different scales, effective modeling of long-range dependencies, and overall optimization of segmentation performance. Experimental results on the ARCADE public dataset show that the proposed method outperforms existing advanced methods in terms of Dice coefficient (76.74%), clDice coefficient (50.30%), and HD95 distance (57.8358 pixels), demonstrating high segmentation accuracy and clinical application value. In the future, through further optimization of model lightweighting, multimodal fusion, and multi-task learning, it is expected to provide more powerful technical support for robot-assisted PCI surgery navigation.
Author Contributions
Conceptualization, L.Z.; Software, L.Z.; Validation, Y.D.; Formal analysis, Y.D.; Investigation, Y.L. (Yunlong Lin); Resources, Y.L. (Yunlong Lin); Data curation, Z.C.; Writing—original draft, Z.C.; Writing—review & editing, Y.L. (Yiyuan Li); Visualization, B.Z.; Supervision, B.Z.; Project administration, S.Z.; Funding acquisition, S.Z. All authors have read and agreed to the published version of the manuscript.
Funding
This work was supported by the Key-Area Research and Development Program of Guangdong Province (No. 2025B1111020001), in part by the Natural Science Foundation of Guangdong Province (No. 2023A1515010673), in part by the Sanming Project of Medicine in Shenzhen (No. SZZYSM202311002), in part by the Shenzhen Science and Technology Innovation Bureau key project (No. JSGG20220831110400001, No. CJGJZD20230724093303007, KJZD20240903101259001), in part by Shenzhen Medical Research Fund (No. D2404001), in part by Shenzhen Engineering Laboratory for Diagnosis & Treatment Key Technologies of Interventional Surgical Robots (XMHT20220104009), and the Key Laboratory of Biomedical Imaging Science and System, CAS, for the Research platform support.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Not applicable.
Data Availability Statement
The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.
Conflicts of Interest
The authors declare no conflicts of interest.
References
- World Health Organization. World Health Statistics 2020: Monitoring Health for the SDGs, Sustainable Development Goals, 1st ed.; World Health Organization: Geneva, Switzerland, 2021. [Google Scholar]
- World Health Organization. Global Health Estimates: Life Expectancy and Leading Causes of Death and Disability; World Health Organization: Geneva, Switzerland, 2024. [Google Scholar]
- Goldstein, J.A.; Balter, S.; Cowley, M.; Hodgson, J.; Klein, L.W. Occupational hazards of interventional cardiologists: Prevalence of orthopedic health problems in contemporary practice. Catheter. Cardiovasc. Interv. 2004, 63, 407–411. [Google Scholar] [CrossRef]
- Piayda, K.; Kleinebrecht, L.; Afzal, S.; Bullens, R.; ter Horst, I.; Polzin, A.; Veulemans, V.; Dannenberg, L.; Wimmer, A.C.; Jung, C.; et al. Dynamic coronary roadmapping during percutaneous coronary intervention: A feasibility study. Eur. J. Med. Res. 2018, 23, 36. [Google Scholar] [CrossRef]
- Solomon, R.; Dauerman, H.L. Contrast-Induced Acute Kidney Injury. Circulation 2010, 122, 2451–2455. [Google Scholar] [CrossRef]
- Wang, R.; Li, C.; Wang, J.; Wei, X.; Li, Y.; Zhu, Y.; Zhang, S. Threshold segmentation algorithm for automatic extraction of cerebral vessels from brain magnetic resonance angiography images. J. Neurosci. Methods 2015, 241, 30–36. [Google Scholar] [CrossRef]
- Orujov, F.; Maskeliūnas, R.; Damaševičius, R.; Wei, W. Fuzzy based image edge detection algorithm for blood vessel detection in retinal images. Appl. Soft Comput. 2020, 94, 106452. [Google Scholar] [CrossRef]
- Sun, K.; Chen, Z.; Jiang, S.; Wang, Y. Morphological multiscale enhancement, fuzzy filter and watershed for vascular tree extraction in angiogram. J. Med. Syst. 2011, 35, 811–824. [Google Scholar] [CrossRef]
- Cinsdikici, M.G.; Aydın, D. Detection of blood vessels in ophthalmoscope images using MF/ant (matched filter/ant colony) algorithm. Comput. Methods Programs Biomed. 2009, 96, 85–95. [Google Scholar] [CrossRef] [PubMed]
- Martinez-Perez, M.E.; Hughes, A.D.; Thom, S.A.; Bharath, A.A.; Parker, K.H. Segmentation of blood vessels from red-free and fluorescein retinal images. Med. Image Anal. 2007, 11, 47–61. [Google Scholar] [CrossRef]
- Hassouna, M.S.; Farag, A.A.; Hushek, S.; Moriarty, T. Cerebrovascular segmentation from TOF using stochastic models. Med. Image Anal. 2006, 10, 2–18. [Google Scholar] [CrossRef]
- Kass, M.; Witkin, A.; Terzopoulos, D. Snakes: Active contour models. Int. J. Comput. Vis. 1988, 1, 321–331. [Google Scholar] [CrossRef]
- Jin, Q.; Meng, Z.; Pham, T.D.; Chen, Q.; Wei, L.; Su, R. DUNet: A deformable network for retinal vessel segmentation. Knowl.-Based Syst. 2019, 178, 149–162. [Google Scholar] [CrossRef]
- Shen, Y.; Fang, Z.; Gao, Y.; Xiong, N.; Zhong, C.; Tang, X. Coronary arteries segmentation based on 3D FCN with attention gate and level set function. IEEE Access 2019, 7, 42826–42835. [Google Scholar] [CrossRef]
- Chen, J.; Wan, J.; Fang, Z.; Wei, L. LMSA-net: A lightweight multi-scale aware network for retinal vessel segmentation. Int. J. Imaging Syst. Technol. 2023, 33, 1515–1530. [Google Scholar] [CrossRef]
- Chen, J.; Mei, J.; Li, X.; Lu, Y.; Yu, Q.; Wei, Q.; Luo, X.; Xie, Y.; Adeli, E.; Wang, Y.; et al. TransUNet: Rethinking the U-net architecture design for medical image segmentation through the lens of transformers. Med. Image Anal. 2024, 97, 103280. [Google Scholar] [CrossRef]
- Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar]
- Liu, X.; Zhang, D.; Yao, J.; Tang, J. Transformer and convolutional based dual branch network for retinal vessel segmentation in OCTA images. Biomed. Signal Process. Control 2023, 83, 104604. [Google Scholar] [CrossRef]
- Xu, H.; Wu, Y. G2ViT: Graph neural network-guided vision transformer enhanced network for retinal vessel and coronary angiograph segmentation. Neural Netw. 2024, 176, 106356. [Google Scholar] [CrossRef] [PubMed]
- Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2014; Volume 27. [Google Scholar]
- Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: New York, NY, USA, 2020; Volume 33, pp. 6840–6851. [Google Scholar]
- Chen, Z.; Xie, L.; Chen, Y.; Zeng, Q.; ZhuGe, Q.; Shen, J.; Wen, C.; Feng, Y. Generative adversarial network based cerebrovascular segmentation for time-of-flight magnetic resonance angiography image. Neurocomputing 2022, 488, 657–668. [Google Scholar] [CrossRef]
- Abdollahi, A.; Pradhan, B.; Alamri, A. VNet: An end-to-end fully convolutional neural network for road extraction from high-resolution remote sensing data. IEEE Access 2020, 8, 179424–179436. [Google Scholar] [CrossRef]
- Li, J.; Wu, Q.; Wang, Y.; Zhou, S.; Zhang, L.; Wei, J.; Zhao, D. DiffCAS: Diffusion based multi-attention network for segmentation of 3D coronary artery from CT angiography. Signal Image Video Process. 2024, 18, 7487–7498. [Google Scholar] [CrossRef]
- Hoover, A.D.; Kouznetsova, V.; Goldbaum, M. Locating blood vessels in retinal images by piecewise threshold probing of a matched filter response. IEEE Trans. Med. Imaging 2000, 19, 203–210. [Google Scholar] [CrossRef] [PubMed]
- Staal, J.; Abramoff, M.D.; Niemeijer, M.; Viergever, M.A.; Van Ginneken, B. Ridge-based vessel segmentation in color images of the retina. IEEE Trans. Med. Imaging 2004, 23, 501–509. [Google Scholar] [CrossRef] [PubMed]
- Zhang, B.; Li, D.; Wang, D. DCT based multi-head attention-BiGRU model for EEG source location. Biomed. Signal Process. Control 2024, 93, 106171. [Google Scholar] [CrossRef]
- Vega, F. Adverse reactions to radiological contrast media: Prevention and treatment. Radiologia 2024, 66 (Suppl. 2), S98–S109. [Google Scholar] [CrossRef]
- Wu, B.; Kheiwa, A.; Swamy, P.; Mamas, M.A.; Tedford, R.J.; Alasnag, M.; Parwani, P.; Abramov, D. Clinical significance of coronary arterial dominance: A review of the literature. J. Am. Heart Assoc. Cardiovasc. Cerebrovasc. Dis. 2024, 13, e032851. [Google Scholar] [CrossRef]
- Qi, Y.; He, Y.; Qi, X.; Zhang, Y.; Yang, G. Dynamic Snake Convolution based on Topological Geometric Constraints for Tubular Structure Segmentation. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 2–6 October 2023; pp. 6047–6056. [Google Scholar]
- Popov, M.; Amanturdieva, A.; Zhaksylyk, N.; Alkanov, A.; Saniyazbekov, A.; Aimyshev, T.; Ismailov, E.; Bulegenov, A.; Kuzhukeyev, A.; Kulanbayeva, A.; et al. Dataset for automatic region-based coronary artery disease diagnostics using X-ray angiography images. Sci. Data 2024, 11, 20. [Google Scholar] [CrossRef]
- Cao, H.; Wang, Y.; Chen, J.; Jiang, D.; Zhang, X.; Tian, Q.; Wang, M. Swin-Unet: Unet-Like Pure Transformer for Medical Image Segmentation. In Proceedings of the Computer Vision—ECCV 2022 Workshops, Tel Aviv, Israel, 23–27 October 2023; Springer: Cham, Switzerland, 2023; pp. 205–218. [Google Scholar]
- Cihan, P.; Saygili, A.; Akyuzlu, M.; Özmen, N.E.; Ermutlu, C.Ş.; Aydın, U.; Yılmaz, A.; Aksoy, Ö. U-Net-Based Approaches for Biometric Identification and Recognition in Cattle. J. Fac. Vet. Med. Kafkas Univ. 2025, 31, 425. [Google Scholar]
- Wang, C.; Zhao, Z.; Ren, Q.; Xu, Y.; Yu, Y. Dense U-net based on patch-based learning for retinal vessel segmentation. Entropy 2019, 21, 168. [Google Scholar] [CrossRef]
- Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-weighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef]
- Hendrycks, D.; Gimpel, K. Gaussian Error Linear Units (GELUs). arXiv 2023, arXiv:1606.08415. [Google Scholar]
- Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. SimAM: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the 38th International Conference on Machine Learning, Virtual, 18–24 July 2021; pp. 11863–11874. [Google Scholar]
- Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
- Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
- Roy, S.; Koehler, G.; Ulrich, C.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. MedNeXt: Transformer-Driven Scaling of ConvNets for Medical Image Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; Springer Nature: Cham, Switzerland, 2023; pp. 405–415. [Google Scholar]
- Hou, Q.; Zhang, L.; Cheng, M.M.; Feng, J. Strip Pooling: Rethinking Spatial Pooling for Scene Parsing. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4002–4011. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).