Next Article in Journal
A Fractional-Order Spatiotemporal Unified Energy Framework for Non-Repetitive LiDAR Point Cloud Registration
Previous Article in Journal
Transfer Learning Fractional-Order Recurrent Neural Network for MPPT Under Weak PV Generation Conditions
Previous Article in Special Issue
When RNN Meets CNN and ViT: The Development of a Hybrid U-Net for Medical Image Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HR-UMamba++: A High-Resolution Multi-Directional Mamba Framework for Coronary Artery Segmentation in X-Ray Coronary Angiography

by
Xiuhan Zhang
1,2,
Peng Lu
2,*,
Zongsheng Zheng
2 and
Wenhui Li
3,*
1
AIEN Institute, Shanghai Ocean University, Shanghai 201306, China
2
College of Information Technology, Shanghai Ocean University, Shanghai 201306, China
3
Modern Educational Technology Center, Shanghai Maritime University, Shanghai 201306, China
*
Authors to whom correspondence should be addressed.
Fractal Fract. 2026, 10(1), 43; https://doi.org/10.3390/fractalfract10010043
Submission received: 25 November 2025 / Revised: 1 January 2026 / Accepted: 6 January 2026 / Published: 9 January 2026

Abstract

Coronary artery disease (CAD) remains a leading cause of mortality worldwide, and accurate coronary artery segmentation in X-ray coronary angiography (XCA) is challenged by low contrast, structural ambiguity, and anisotropic vessel trajectories, which hinder quantitative coronary angiography. We propose HR-UMamba++, a U-Mamba-based framework centered on a rotation-aligned multi-directional state-space scan for modeling long-range vessel continuity across multiple orientations. To preserve thin distal branches, the framework is equipped with (i) a persistent high-resolution bypass that injects undownsampled structural details and (ii) a UNet++-style dense decoder topology for cross-scale topological fusion. On an in-house dataset of 739 XCA images from 374 patients, HR-UMamba++ is evaluated using eight segmentation metrics, fractal-geometry descriptors, and multi-view expert scoring. Compared with U-Net, Attention U-Net, HRNet, U-Mamba, DeepLabv3+, and YOLO11-seg, HR-UMamba++ achieves the best performance (Dice 0.8706, IoU 0.7794, HD95 16.99), yielding a relative Dice improvement of 6.0% over U-Mamba and reducing the deviation in fractal dimension by up to 57% relative to U-Net. Expert evaluation across eight angiographic views yields a mean score of 4.24 ± 0.49/5 with high inter-rater agreement. These results indicate that HR-UMamba++ produces anatomically faithful coronary trees and clinically useful segmentations that can serve as robust structural priors for downstream quantitative coronary analysis.

1. Introduction

Coronary artery disease (CAD) is one of the most prevalent cardiovascular disorders and remains a leading cause of death worldwide [1,2]. Epidemiological evidence indicates that CAD accounts for around nine million deaths annually and affects more than 300 million individuals, with its global burden projected to increase further driven by population growth and aging [3,4,5,6,7]. CAD arises primarily from the gradual accumulation of atherosclerotic plaques in the coronary arteries, resulting in luminal narrowing and impaired myocardial perfusion, which in turn leads to varying degrees of myocardial ischemia and functional impairment [8]. Clinically, CAD manifests as stable or unstable angina, silent myocardial ischemia, or acute myocardial infarction [9,10], with acute events frequently triggered by thrombosis following rupture or erosion of thin-cap fibroatheromas [11,12].
In current clinical practice, imaging modalities used for the diagnosis and assessment of CAD include X-ray coronary angiography (XCA) [13], coronary computed tomography angiography (CCTA) [14], optical coherence tomography (OCT) [15], and near-infrared spectroscopy (NIRS) [16]. Among these, XCA is widely regarded as the gold standard for diagnosing CAD [13]. However, visual interpretation of XCA depends heavily on operator experience and exhibits substantial subjectivity and inter-observer variability. Prior studies have reported inter-clinician discrepancies in lesion-boundary identification and stenosis-severity assessment of up to 15–45% [17,18]. Furthermore, manual delineation of coronary vessels is time-consuming, which can compromise diagnostic efficiency and delay clinical decision-making [19,20].
To mitigate the subjectivity and inconsistency inherent in visual assessment, quantitative coronary angiography (QCA) has been developed with the aim of providing objective vascular evaluation in an automated and reproducible manner [21]. QCA enables precise measurement of parameters such as vessel diameter, lesion length, and reference lumen diameter, demonstrating higher consistency compared to traditional visual estimation [21,22,23]. In this workflow, accurately extracting the coronary artery tree from XCA images is a critical prerequisite for QCA implementation [24]. However, this task remains challenging due to inherently low vessel-to-background contrast, overlap caused by 2D projection, and confounders such as noise, motion artifacts, and illumination variability [24,25,26,27]. Figure 1 illustrates the common imaging challenges encountered in clinical XCA images, including insufficient contrast, speckle noise interference, and structural artifacts caused by guidewires and bony structures. Consequently, achieving robust and high-precision automatic segmentation of coronary arteries in XCA images remains a major technical bottleneck in both QCA workflows and medical image analysis.
Despite substantial recent progress, existing coronary-vessel segmentation methods still face several important limitations. First, the structural complexity of coronary vasculature in XCA makes it difficult for traditional unsupervised methods to reliably preserve the full vessel morphology, especially at bifurcations and in distal small-caliber vessels [28,29]. Although deeply supervised methods have achieved state-of-the-art accuracy and robustness, most existing models primarily target the extraction of major epicardial vessels and provide only limited characterization of the full coronary artery tree [30,31]. In addition, many approaches rely on multi-frame sequences or three-dimensional imaging conditions [32,33,34], which constrains their applicability to the clinically more common XCA setting. Finally, most deep learning models remain suboptimal in modeling slender tubular structures and global vascular topology, often yielding prediction masks with broken centerlines, missing details, and discontinuous vascular structures [31,35].
These challenges underscore the need for a segmentation framework that can simultaneously preserve thin distal branches, capture global vascular topology, and remain robust to noise and imaging artifacts in single-frame XCA images. Accordingly, we propose HR-UMamba++, a novel coronary artery segmentation framework built on U-Mamba and equipped with a rotation-aligned multi-directional scanning mechanism that explicitly models long-range vessel continuity in horizontal, vertical, and diagonal directions. To make this long-range directional modeling effective for faint coronary branches, we further introduce two architectural designs:
  • Inspired by the segregated processing of magnocellular and parvocellular pathways in the primate retina [36], we construct a persistent high-resolution branch that continuously injects non-downsampled structural details along the entire decoding path, thereby markedly enhancing the spatial discriminability of fine vessels and distal terminal branches.
  • We introduce a UNet++-style densely connected decoding topology that enables cross-level and multi-scale semantic feature fusion, strengthening the model’s ability to represent multi-level vascular bifurcations and the complete coronary artery tree.
Through this architectural design, HR-UMamba++ achieves high-resolution detail preservation, cross-scale vascular topology modeling, and robust performance under noise contaminated conditions in XCA images. The proposed model demonstrates clear advantages across multiple quantitative metrics and clinical expert evaluations, providing a reliable segmentation backbone for building automated and clinically applicable QCA workflows. Furthermore, we develop a vascular tree topological evaluation framework based on box-counting fractal dimension, leveraging fractal descriptors as a complementary measure to conventional segmentation metrics and establishing a methodological basis for the exploration of potential fractal-based biomarkers.

2. Related Work

A variety of frameworks have been proposed for automatic segmentation of coronary arteries, which can broadly be categorized into unsupervised and supervised methods [37]. Unsupervised methods can further be divided into three classes: filter-based, tracking-based, and model-based approaches [38]. Filter-based approaches [24,39,40,41,42,43] assume that local vessel cross-sections can be approximated by line- or tube-like intensity profiles. Such methods typically employ spatial- or frequency-domain filters to enhance vessel-like structures and suppress background noise.
Tracking-based approaches [28,44,45] select initial seed points along vessel centerlines or boundaries. Starting from these seeds, the methods predict the propagation direction and radius of the vessels by analyzing local patches of gray-level intensities or gradient features. New center points are then iteratively generated along the inferred direction, gradually tracking and reconstructing the full vascular tree topology. Model-based approaches [46,47,48,49,50,51] likewise require an initial contour or seed region as the starting point for curve evolution. These methods model the dynamic evolution of vessel structures using an energy functional or partial differential equation, and update the contour via energy minimization so that it progressively approaches the true vessel boundary.
Although unsupervised methods avoid labor-intensive manual annotation and often offer advantages in computational efficiency [52,53], their limitations are substantial. Filter-based approaches tend to respond poorly at complex bifurcations and are vulnerable to tubular artifacts such as catheters and the spine, leading to spurious vessel detections [54,55,56]. Tracking-based and model-based methods are highly sensitive to the quality and placement of initial seeds, and suboptimal seed initialization can cause the tracking procedure to deviate from the true lumen or drift into the background [28]. In addition, when neighboring high-intensity structures such as myocardial tissue are present, both tracking- and model-based approaches frequently exhibit boundary leakage into adjacent regions [37,57], further compromising segmentation stability and accuracy.
In contrast, supervised methods rely on large collections of annotated vessel masks for training. By learning discriminative feature representations from labeled data, these models establish decision boundaries in feature space that separate vascular from non-vascular regions [58,59]. He et al. [60] proposed a segmentation framework based on multi-frame invasive coronary angiography (ICA) sequences, in which temporal information was treated as a third dimension. They employed a 3D encoder–2D decoder architecture to fuse spatiotemporal features and introduced an elastic energy-based loss to encourage vessel connectivity. Han et al. [32] developed a topologically constrained Multiscale Attention Network (MAN) that integrated a Multiscale Squeeze Attention (MSA) mechanism into a U-Net backbone to extract multi-scale features, and employed a topology-aware loss function to preserve the continuity of the coronary tree. Zhao et al. [61] proposed a Hyper Association Graph Matching Network for semantic labeling of coronary arteries. In this framework, coronary semantic annotation was formulated as a hypergraph matching problem, wherein higher-order topological relationships were explicitly exploited to enhance labeling accuracy. He et al. [29] further proposed a coronary segmentation model that incorporated geometric priors and state-space modeling. Their framework integrated a Multiscale Vessel Structure-Aware (MVSA) module and a Kolmogorov–Arnold State Space (KASS) component within a U-shaped architecture to jointly encode vessel morphology and long-range dependencies, and replaced conventional skip connections with cross-stage feature fusion to improve detail preservation. Nannini et al. [33] employed a two-stage U-Net pipeline to segment coronary arteries and quantitatively assess coronary artery calcium and tortuosity from coronary CT angiography (CCTA). A multi-view 2.5D network performed an initial segmentation in the first stage, and a 3D U-Net refined the result in the second stage. Geometric postprocessing was then applied to extract vessel centerlines and morphological descriptors. Iyer et al. [62] introduced an Angiographic Processing Network (APN) module into an end-to-end model, enabling automatic learning of optimal pre-processing filters that sharpen vessel boundaries and enhance local contrast. Xu et al. [34] proposed a hybrid framework that couples spatial attention with a diffusion model for automatic segmentation of CCTA images, effectively addressing the challenges posed by complex vessel morphology, imaging noise, and substantial shape variability.
A synthesis of these recent supervised methods reveals three major limitations: (1) Most supervised coronary segmentation studies focus on delineating the major coronary arteries, with limited ability to model the full coronary artery tree and its fine distal branches. (2) Many frameworks are specifically tailored to CCTA or temporal angiographic sequences, limiting their applicability to the clinically common scenario of X-ray coronary angiography (XCA). (3) Conventional convolutional neural networks struggle to fully capture thin, elongated tubular structures. Their isotropic convolution kernels provide limited ability to encode complex vascular topology and long-range connectivity, which often leads to fragmented or discontinuous vessel predictions in the segmentation masks.
Motivated by these limitations, we propose HR-UMamba++, a novel segmentation framework designed to recover thin distal vessels and maintain topological continuity of the coronary artery tree from XCA images.

3. Materials and Methods

3.1. Dataset Description

The research team collaborated with the Affiliated Hospital of Shandong University of Traditional Chinese Medicine (College Town Branch) and retrospectively assembled a dataset of 739 XCA images from 374 patients who underwent routine clinical examinations between February 2023 and October 2024. The images cover the major epicardial coronary arteries, including the left anterior descending (LAD), left circumflex (LCX), and right coronary artery (RCA), and encompass a broad spectrum of vascular morphologies. To increase diversity and robustness for model training, the dataset includes normal arteries, mild stenosis, moderate-to-severe stenosis, and complete occlusions, reflecting clinically relevant pathophysiological variation.
All subjects underwent standardized X-ray coronary angiography via a transradial approach. Image acquisition was performed using a digital subtraction angiography (DSA) system (Optima IGS 330, GE Healthcare, Beijing, China). Prior to angiography, patients received weight-adjusted heparin for anticoagulation, and during the procedure an iodinated contrast agent (iohexol; 40 ± 10 mL ; 30 g I / 100 mL ) was administered. From each DSA sequence, we extracted representative frames and stored them as single-channel grayscale images with a spatial resolution of 512 × 512 pixels. Ground-truth vessel masks were annotated by two experienced interventional cardiologists using the LabelMe image annotation platform. Key acquisition and sampling parameters of the dataset are summarized in Table 1.
To protect patient privacy, all images underwent a rigorous de-identification process before use. Any personal information present in the original images was completely removed through standardized masking and cropping procedures, ensuring that individual subjects could not be re-identified. The study relied exclusively on existing clinical imaging records and did not involve additional interventions or prospective data collection. All procedures were conducted in accordance with the principles of the Declaration of Helsinki and relevant biomedical ethics guidelines. The Academic Ethics and Morality Committee of Shanghai Ocean University reviewed the research protocol and, given the use of fully anonymized retrospective data, granted a waiver of formal ethical approval and informed consent requirements.
Table 2 compares our dataset with those used in several representative recent studies. Our in-house dataset includes a larger number of patients than most recent coronary angiography segmentation datasets, and the number of annotated frames is also among the highest reported. In addition, our dataset follows the clinically more common single-frame input setting, and the spatial resolution is consistent with common practice in the literature. Therefore, this dataset provides a solid foundation for validating the effectiveness of the proposed method.

3.2. Overall Architecture of HR-UMamba++

In recent years, vision architectures based on state space models (SSMs) have attracted increasing attention in medical image analysis owing to their ability to efficiently capture long-range dependencies [68]. Building on this paradigm, Ma et al. [69] proposed U-Mamba, which combines CNN-based local feature extraction with Mamba-based sequence modeling in a U-shaped encoder-decoder. Empirical studies have demonstrated that U-Mamba and its derivatives achieve superior performance compared with mainstream CNN- and Transformer-based models across a variety of tasks, including abdominal MRI/CT segmentation, 2D endoscopic imaging, microscopy images, and brain tumor segmentation on MRI [69,70,71,72]. These methods exhibit fewer segmentation artifacts and more stable predictions, suggesting that the U-Mamba paradigm also holds substantial potential for vascular tree segmentation, where both fine-grained local textures and global structural consistency must be preserved.
The proposed HR-UMamba++ architecture is constructed upon the U-Mamba framework and is specifically tailored to the highly filamentous, multi-branching geometry of coronary artery trees. To this end, three key modifications are introduced: (i) a persistent high-resolution feature branch ( H R i ) that continuously injects fine structural details into the decoding pathway; (ii) a UNet++-style dense decoder topology that performs multi-level aggregation across both columns and scales to strengthen semantic fusion; and (iii) a rotation-aligned multi-directional Mamba scanning mechanism designed to capture vessel continuity along multiple spatial orientations. As illustrated in Figure 2, the overall architecture follows an encoder-decoder paradigm while incorporating dense lateral aggregation and a dedicated high-resolution pathway to better accommodate the geometric complexity of coronary vasculature.

3.3. Persistent High-Resolution Branch in HR-UMamba++

In the conventional U-Net architecture, high-resolution features are only transmitted once from encoder to decoder via a single skip connection at each resolution level [73]. As decoding proceeds through repeated concatenation and fusion, fine details from shallow layers can be diluted by deeper semantic features and multi-scale aggregation [74]. This effect is particularly detrimental for thin vessels, small distal branches, and boundary regions, where subtle structural cues are easily overwhelmed, leading to broken or missing segments in the final prediction [75].
To alleviate this limitation, we draw inspiration from the parallel processing mechanisms of the magnocellular and parvocellular pathways in the primate retinal visual system [36] and introduce a persistent high-resolution branch ( H R i ). This branch maintains an undownsampled, high-resolution stream of structural information and continuously injects it into the decoder, thereby providing a stable source of fine-detail compensation throughout the multi-scale integration process.
Let the encoder feature map at the i-th resolution level be denoted as
X i , 0 R B × C i × H i × W i ,
the persistent high-resolution branch refines local structures via a lightweight residual block D ( · ) to obtain
HR i = D ( X i , 0 ) , i = 0 , , S 2 .
where S denotes the total number of resolution levels in the encoder.
The refinement block D ( · ) is implemented as
D ( X ) = LeakyReLU Norm 2 Conv 2 ( LeakyReLU ( Norm 1 ( Conv 1 ( X ) ) ) ) + X .
By maintaining this persistent high-resolution stream, HR-UMamba++ ensures that the decoder consistently receives high-fidelity spatial cues. In contrast to the fully parallel multi-resolution branch design exemplified by HRNet [76], our architecture maintains only a lightweight H R i branch, thereby substantially reducing computational cost. Meanwhile, H R i serves as a stable high-resolution detail bypass, mitigating the degradation of fine vessels caused by repeated fusion of high- and low-resolution features, and thereby making the architecture particularly suitable for the faithful reconstruction of coronary arteries with slender trunks and densely packed distal branches.

3.4. Encoder Module with Residual Convolutions and Multi-Directional Mamba

As illustrated in Figure 2, the encoder of HR-UMamba++ adopts a multi-stage down-sampling design. At each stage, residual convolutional units are employed to extract multi-scale semantic representations, while rotation-aligned multi-directional Mamba modules are inserted at selected resolutions to enhance the modeling of long-range dependencies within the same spatial scale. An anatomical analysis of common clinical XCA projections [77] indicates that the major coronary arteries predominantly course along the atrioventricular groove and thus appear approximately horizontal in the image; in contrast, vessels such as the LAD extend toward the cardiac apex along the anterior interventricular groove, exhibiting an approximately vertical orientation. Diagonal and lateral branches are often distributed obliquely. To cover these principal orientations while avoiding unnecessary computational overhead, we fix the scanning directions to the four angles { 0 , 90 , + 45 , 45 } . Let the input image be
I R B × C in × H × W .
The encoder consists of a stem block followed by S resolution stages, which sequentially produce feature maps
X 0 , 0 , X 1 , 0 , , X S 1 , 0 .

3.4.1. Stem Residual Block

In the stem stage, the input image is first processed by a residual alignment block:
X 0 , 0 = R 0 ( I ) ,
where the residual alignment block R 0 ( · ) is instantiated as
R 0 ( X ) = LeakyReLU Norm 2 Conv 2 LeakyReLU Norm 1 Conv 1 ( X ) + Conv 1 × 1 ( X ) ,
to perform channel remapping and initial feature compression.

3.4.2. Multi-Stage Residual Down-Sampling

The s-th encoder stage is composed of one strided residual block for down-sampling and multiple non-strided residual refinement blocks for deep feature enrichment. Specifically, we first obtain
Z s ( 0 ) = R s ( X s 1 , 0 ) ,
and then recursively refine the features via
Z s ( k ) = D s Z s ( k 1 ) , k = 1 , , n s 1 .
where n s denotes the number of residual refinement blocks at stage s. The output of the s-th stage is thus given by
X s , 0 = Z s ( n s 1 ) .
The strided residual down-sampling block and the same-scale residual refinement block are defined as
R s ( X ) = LeakyReLU Norm 2 Conv 2 LeakyReLU Norm 1 Conv 1 ( X ) + Conv 1 × 1 ( X ) , D s ( X ) = LeakyReLU Norm 2 Conv 2 LeakyReLU ( Norm 1 ( Conv 1 ( X ) ) ) + X .
The spatial resolution of the encoder features is progressively reduced according to
( H s , W s ) = H i = 0 s k i , W i = 0 s k i
where k i denotes the down-sampling factor of the i-th stage.

3.4.3. Rotation-Aligned Multi-Directional Mamba Scanning Mechanism

In clinical XCA images, coronary vessels often extend over long distances along horizontal, vertical, and oblique directions. To enhance the encoder’s ability to capture long-range dependencies and orientation-specific continuity, HR-UMamba++ integrates a rotation-aligned multi-directional Mamba module at selected encoder stages. The core idea is to rotate 2D feature maps to align different orientations and flatten them into 1D sequences. These sequences are processed by Mamba units for direction-aware state-space modeling, and the resulting features are rotated back and fused in the original coordinate system. As illustrated in Figure 3, four orientations (horizontal, vertical, and two diagonals) are handled by rotating the feature map to 0 , 90 , 45 , and 45 , respectively, followed by independent Mamba scans to capture vessel structures that predominantly extend along each direction.
Formally, let the input feature at the s-th encoder stage be
X s , 0 R B × C × H × W .
To avoid boundary cropping during rotation, X s , 0 is first reflect-padded to the minimal enclosing square:
L = H 2 + W 2 , X s , 0 pad = P ( X s , 0 ) .
Subsequently, geometric rotations are applied at four orientation angles Θ = { 0 , 90 , 45 , 45 } :
X s , 0 ( θ ) = T θ ( X s , 0 pad ) ,
and each rotated 2D feature map is flattened into a sequence for state-space modeling:
Z s ( θ ) = Flatten X s , 0 ( θ ) R B × C × ( L 2 ) , Z ˜ s ( θ ) = LN ( Z s ( θ ) ) .
The orientation-specific sequences are then fed into a Mamba unit to perform selective state evolution and recurrent convolution, yielding direction-aware responses:
Z ^ s ( θ ) = Mamba Z ˜ s ( θ ) .
Each sequence is subsequently reshaped back to a 2D feature map, inverse-rotated to the original coordinate system, and normalized by the corresponding interpolation weights to ensure consistent spatial alignment:
X ^ s ( θ ) = C T θ 1 reshape 1 ( Z ^ s ( θ ) ) W ( θ ) + ε ,
where W ( θ ) denotes the interpolation weight map associated with rotation angle θ , and C ( · ) denotes center cropping used to recover the original spatial resolution ( H , W ) . Finally, the four directional feature maps are concatenated along the channel dimension and projected back to the original channel size C:
X s , mamba = BN Conv 1 × 1 ( BN ( [ X ^ s ( 0 ) , X ^ s ( 90 ) , X ^ s ( 45 ) , X ^ s ( 45 ) ] ) )
thereby producing orientation-enhanced features at the same spatial scale as the original encoder stage.

3.5. Decoder and Dense Connection Module in HR-UMamba++

The decoder of HR-UMamba++ adopts a UNet++-style dense topology [78] to facilitate lateral aggregation across decoder columns and cross-scale semantic fusion. Let the feature map from the i-th encoder layer be
X i , 0 R B × C × H × W .
Each decoder node X i , j ( j 1 ) receives a concatenated input F i , j obtained by fusing all preceding nodes at the same depth, the upsampled feature from the next deeper level, and the corresponding high-resolution branch H R i :
F i , j = X i , 0 , X i , 1 , , X i , j 1 , Up ( X i + 1 , j 1 ) , H R i ,
where the upsampling operator is defined as
Up ( X ) = Conv 1 × 1 Interp × 2 ( X ) ,
and simultaneously restores the spatial resolution and aligns the channel dimension.
To ensure that the outputs of all nodes share a consistent dimensionality compatible with the UNet++ grid, each decoder node is implemented as a two-stage convolutional block comprising a channel alignment block A i , j and a feature refinement block F i , j , defined as
A i , j ( X ) = LeakyReLU Norm 2 Conv 2 LeakyReLU Norm 1 Conv 1 ( X ) + Conv 1 × 1 ( X ) , F i , j ( X ) = LeakyReLU Norm 2 Conv 2 LeakyReLU ( Norm 1 ( Conv 1 ( X ) ) ) + X .
The final output of each decoder node is obtained by sequentially passing the concatenated feature F i , j through these two blocks:
X i , j = F i , j A i , j ( F i , j ) .

4. Experiments and Results

4.1. Implementation Details

This study is implemented within the nnU-Net v2 framework [79], into which we integrate a custom HR-UMamba++ model while preserving the standardized automated training pipeline. The backbone follows a 2D U-Net architecture with eight resolution stages, using input patches of size 512 × 512 and a batch size of 3. Each encoder stage contains two 3 × 3 convolutional layers, with the number of feature channels progressively increasing from 32 at the highest resolution to 512 at the bottleneck. Spatial down-sampling is performed by seven 2 × 2 pooling operations with stride 2, reducing the feature-map resolution to 4 × 4 . The decoder is composed of seven symmetric up-sampling stages, where 2 × 2 bilinear interpolation is used at each level to double the spatial resolution while decreasing the channel dimensionality from 512 back to 32. Each decoder stage likewise includes two 3 × 3 convolutional layers, and after the final up-sampling step the output feature maps are restored to the original spatial resolution of 512 × 512 .
To mitigate the low contrast and noise corruption commonly observed in XCA images, we introduce a lightweight fractional-order Laplacian operator into the preprocessing pipeline. In the frequency domain, the fractional Laplacian of an image I can be expressed as
F { ( ) α / 2 I } ( ω x , ω y ) = | ω | α F { I } ( ω x , ω y ) ,
where F { I } ( ω x , ω y ) denotes the two-dimensional Fourier transform of I, | ω | = ω x 2 + ω y 2 is the magnitude of the frequency vector, and the parameter 0 < α < 2 controls the response to high-frequency components. Based on this operator, the enhanced image is given by
I enh = I + λ ( ) α / 2 I ,
where λ is an enhancement coefficient. This operation selectively amplifies high-frequency vascular structures, such as vessel boundaries and distal branches, thereby improving the visibility of fine vessels and regions with low contrast. Additional preprocessing and data augmentation strategies are summarized in Table 3.
The training objective combines a pixel-wise cross-entropy loss with a soft Dice loss. The cross-entropy term is defined as
L C E = 1 B H W b , i , j c y b , c , i , j log p b , c , i , j ,
where p b , c , i , j and y b , c , i , j denote the predicted probability and one-hot encoded ground-truth label of class c at spatial location ( i , j ) in sample b. The Dice loss measures the overlap between the predicted segmentation and the ground truth, and is given by
L D i c e = 1 2 p y + ϵ p + y + ϵ .
The final training objective is an equally weighted combination of the two terms:
L T o t a l = L C E + L D i c e .
The model is trained for 1000 epochs using a 5-fold cross-validation protocol for data partitioning. We adopt stochastic gradient descent (SGD) with Nesterov momentum of 0.99 as the optimizer, and set the weight decay to 3 × 10 5 . The initial learning rate is set to 1 × 10 2 and progressively decreased according to a polynomial learning rate schedule. Automatic mixed precision (AMP) is enabled throughout training to reduce GPU memory consumption and improve computational efficiency.
All experiments are conducted on a workstation running Ubuntu 22.04.3 LTS equipped with a single NVIDIA GeForce RTX 4090 GPU (24 GB). The CUDA runtime version is 11.8, and both training and inference are implemented in PyTorch 2.1.2+cu118 with Python 3.10. All experiments are performed under this unified software and hardware environment to ensure the comparability and reproducibility of the reported results.

4.2. Evaluation Metrics

To comprehensively evaluate model performance on the vessel segmentation task, we employ eight quantitative metrics: Accuracy, Sensitivity, Specificity, Positive Predictive Value (PPV), Negative Predictive Value (NPV), Dice coefficient, Intersection over Union (IoU), and the 95th-percentile Hausdorff distance (HD95). All metrics are computed from pixel-wise comparisons between the predicted segmentation masks and the manual annotations, where TP , FP , TN , and FN denote the numbers of true-positive, false-positive, true-negative, and false-negative pixels, respectively.
Accuracy quantifies the overall correctness of the predictions and is defined as
Accuracy = TP + TN TP + TN + FP + FN .
Sensitivity (recall) measures the ability of the model to correctly identify vessel pixels:
Sensitivity = TP TP + FN .
Specificity measures the capability of the model to correctly recognize background (non-vessel) pixels:
Specificity = TN TN + FP .
PPV quantifies the proportion of correctly predicted vessel pixels among all pixels predicted as vessel:
PPV = TP TP + FP .
NPV quantifies the proportion of correctly predicted background pixels among all pixels predicted as background:
NPV = TN TN + FN .
The Dice coefficient measures the overlap between the predicted and ground-truth vessel regions and is one of the most widely used metrics in medical image segmentation:
Dice = 2 TP 2 TP + FP + FN .
IoU (Jaccard index) measures the ratio between the intersection and the union of the predicted and ground-truth regions:
IoU = TP TP + FP + FN .
In addition, HD95 is used to assess the spatial consistency of boundary predictions. The Hausdorff distance measures the maximum deviation between the predicted and ground-truth boundaries in the spatial domain, whereas HD95 takes the 95th percentile of the bidirectional boundary distances to reduce the influence of outliers and annotation noise. Let P and G denote the sets of boundary points extracted from the prediction and the ground truth, respectively; HD95 is defined as
HD 95 = max percentile 95 min g G p g , percentile 95 min p P g p .
Among these metrics, Accuracy, Sensitivity, Specificity, PPV, NPV, Dice, and IoU take values within the interval [ 0 , 1 ] , with values closer to 1 indicating stronger agreement between the predicted segmentation and the reference annotations. In contrast, HD95 is a distance-based measure, where smaller values approaching 0 indicate more precise localization of the vessel boundaries.

4.3. Ablation Studies

To disentangle the contribution of each core design choice in HR-UMamba++, we conduct ablation studies on three key components: (i) the UNet++-style dense connections, (ii) the persistent high-resolution (HR) branch, and (iii) the rotation-aligned multi-directional Mamba scanning mechanism. We perform both single-module and pairwise ablation experiments. All variants are trained under identical training protocols and data configurations, and the corresponding quantitative results are summarized in Table 4.
As shown in Table 4, the complete HR-UMamba++ model achieves the best performance across all eight evaluation metrics. Removing any individual module consistently leads to a deterioration in segmentation performance, indicating that these components collectively contribute to the overall performance improvements in coronary vessel segmentation.

4.3.1. Effect of Dense Connections

As reported in Table 4, removing the dense connections yields a pronounced decline in the model’s ability to capture vessel connectivity and thin filamentous structures. Specifically, Sensitivity and IoU decrease by 5.3% and 6.3%, respectively, while HD95 increases by approximately 19.6% compared with the full model. These changes highlight the critical role of dense connections in maintaining the continuity and spatial coherence of coronary vasculature.
Figure 4 illustrates three representative cases. In Case 1, faint vessel boundaries in a low-perfusion region are misclassified as background by the ablated model, resulting in local discontinuities. In Case 2, the continuity of small-caliber vessels is substantially compromised, and gaps emerge at the junctions between the main trunk and its side branches. In Case 3, the ablated model fails to preserve the complex bifurcation topology, with several thin branches completely missing. In contrast, the full HR-UMamba++ model accurately reconstructs the global vascular topology in all cases, confirming that dense connections are essential for effective multi-scale feature aggregation and fine-grained vessel structure recovery.

4.3.2. Effect of the Persistent High-Resolution Branch

The persistent high-resolution (HR) branch maintains a dedicated high-resolution feature stream throughout the feature extraction process, enabling the network to preserve fine spatial details and vessel boundaries and thereby providing more discriminative cues for thin, filamentous coronary structures. As reported in Table 4, removing the HR branch reduces the Dice coefficient and IoU by approximately 3.4% and 5.4%, respectively, while increasing HD95 by about 10.2%, underscoring its importance for accurate and anatomically coherent vessel delineation. To further quantify the role of the persistent HR branch in preserving multi-scale vascular structures, we perform a frequency-domain analysis of the outputs produced by the full HR-UMamba++ model and its ablated counterpart. Specifically, the predicted segmentation masks from both models are first apodized with a Hanning window and have their DC components removed, after which a 2D Fourier transform is applied to obtain the corresponding power spectra. We then compute the proportion of mid-to-high frequency energy within the normalized radial frequency range of 0.25–0.9.
The experimental results show that the complete model exhibits, on average, a 4.10% increase in mid-to-high frequency energy, indicating that the HR branch substantially enhances the preservation of fine vascular filaments and directional continuity. As illustrated in Figure 5, the log-scaled power spectra reveal that the model without the HR branch suffers from a marked loss of high-frequency energy, leading to degraded texture representation and reduced connectivity of thin vessel segments. In contrast, the additional energy in the full model is primarily concentrated in radial streaks aligned with the dominant coronary directions, suggesting that the HR branch selectively amplifies structure-related high-frequency components rather than noise, thereby improving the continuity and completeness of the reconstructed coronary artery tree.

4.3.3. Effect of the Rotation-Aligned Multi-Directional Mamba Module

Removing the rotation-aligned multi-directional Mamba scanning mechanism reduces the Dice coefficient from 0.8706 to 0.8478 and the IoU from 0.7794 to 0.7470, indicating that this component contributes meaningfully to the overall segmentation performance. To further assess its role in maintaining directional coherence and global connectivity of the coronary vasculature, we additionally include 200 angiograms with visually normal coronary anatomy and define a binary structural-integrity indicator based on whether the predicted mask contains exactly one foreground connected component. Ideally, all 200 predictions should exhibit a single connected vascular tree.
Experimental results show that the complete model preserves single-component connectivity in 187 out of 200 images, whereas the ablated variant achieves this in only 175 cases. Because both models generate segmentation masks on the same image set, these outcomes constitute a paired binary dataset, for which we employ McNemar’s test to assess the statistical significance of the difference [80]; the corresponding paired outcomes are summarized in the contingency table in Table 5. A two-sided exact binomial test yields a p-value of 0.0075, demonstrating that removal of the rotation-aligned multi-directional Mamba module significantly degrades the model’s ability to preserve global vascular connectivity. Consistently, the paired risk difference increases by 6.0 percentage points, indicating that this mechanism provides a stable improvement in the structural consistency of the predicted vessel trees.

4.4. Performance Comparison with Canonical Baselines

To further assess the overall performance of HR-UMamba++ on the coronary vessel segmentation task, we compare it against five representative semantic segmentation baselines: U-Net [81], Attention U-Net [82], HRNet [76], U-Mamba [69], and DeepLabv3+ [83]. Given the competitive performance of the YOLO family reported in coronary vessel segmentation [84,85,86], we additionally include YOLO11-seg as an instance segmentation baseline. As YOLO targets instance segmentation [87], we convert each binary vessel mask into instance labels via connected-component decomposition, and merge the predicted instances back into a semantic vessel map for metric evaluation. YOLO11-seg is initialized from pretrained YOLO11m-seg weights and trained with the official default hyperparameters. All models are trained and evaluated under identical data configurations and training protocols to ensure the fairness and reproducibility of the comparison. Table 6 summarizes the quantitative results across eight evaluation metrics, while Figure 6 provides a visual comparison of the predicted segmentation masks.
The CNN-based U-Net and Attention U-Net exhibit stable performance on large and moderately sized structures; however, their limited capacity for rich high-resolution feature representation leads to suboptimal detection of fine vessels. Consequently, their Sensitivity scores are restricted to 0.7189 and 0.7153, with Dice coefficients of 0.7758 and 0.7777, respectively. In the visual results, both models frequently produce interruptions in slender vessels and miss small peripheral branches. This issue is particularly evident in the complex tree-like region of Case 2, where multiple low-contrast bifurcations are incorrectly collapsed into discontinuous fragments, highlighting the limitations of purely convolutional encoders in modeling delicate vascular details.
By maintaining high-resolution representations via parallel multi-resolution streams with repeated feature fusion, HRNet provides a stronger baseline for vessel delineation. It improves Sensitivity to 0.7684 and Dice to 0.8184, outperforming U-Net and Attention U-Net. Qualitatively, HRNet recovers more secondary branches in Case 2 and produces fewer breaks along medium-caliber segments. However, its boundary accuracy and topological completeness remain limited: distal vessels with rapidly changing orientations can still appear locally discontinuous, and small peripheral branches are prone to truncation. These observations suggest that preserving high-resolution features alone is insufficient to maintain directional coherence and long-range continuity across the full coronary tree, and that explicitly modeling long-range dependencies remains necessary.
Leveraging state space modeling, U-Mamba substantially enhances the capture of long-range dependencies and achieves noticeable improvements over conventional CNN-based architectures, with Sensitivity increasing to 0.7768 and the Dice coefficient to 0.8215. The visual comparison indicates that U-Mamba maintains more consistent continuity along the main trunk and medium-caliber segments. Nevertheless, connection gaps remain in distal branches with rapidly changing orientations. For instance, U-Mamba fails to correctly recover small vessels such as the acute marginal branch and right ventricular branch in Case 1, indicating that its modeling of directional consistency is still insufficient to fully preserve the complete topology of thin coronary vessels.
DeepLabv3+ benefits from its strong multi-scale feature extraction capability, achieving an HD95 of 17.22, which reflects more accurate boundary localization. Visual comparison shows that its vessel contours are relatively smooth. However, in the fine distal branches of Case 1, it still exhibits terminal branch truncation and local structural adhesion, implying that atrous convolutions alone are not fully adequate for capturing high-frequency vascular texture details. Although YOLO11-seg demonstrates strong performance in object detection and region-level segmentation, its multi-task design introduces trade-offs when modeling subtle vessel patterns, and its overall performance does not surpass that of more specialized semantic segmentation architectures.
In contrast, the proposed HR-UMamba++ achieves the best performance across all eight evaluation metrics. It attains an Accuracy of 0.9877, while Sensitivity is further improved to 0.8179, and the Dice coefficient and IoU reach 0.8706 and 0.7794, respectively. In the visual comparison of the three representative cases in Figure 6, the proposed model consistently reconstructs the most complete coronary artery tree topology, preserving the continuity and directional coherence of thin bifurcations. It effectively avoids the common issues observed in other models, such as small-branch breakage, trajectory deviation, and local omissions, thereby demonstrating its clear overall advantage in modeling complex coronary vessel structures.

4.5. Fractal-Geometry-Based Topological Assessment of Coronary Arterial Trees

The coronary arterial tree exhibits a characteristic branching morphology with pronounced multiscale and self-similar patterns [88,89]. Its geometric complexity and hierarchical branching architecture can be described in a scale-invariant manner by the fractal dimension. To further assess the ability of different models to preserve the topological structure of the coronary arterial tree, we complement conventional overlap- and distance-based metrics with a box-counting-based fractal-dimension analysis.
Specifically, we first apply skeletonization to the binary vessel masks obtained from the manual annotations and model predictions to extract vessel centerlines, so that the fractal dimension primarily reflects the branching hierarchy and spatial spread of the vascular tree. Let M { 0 , 1 } H × W denote a binary vessel mask, where M ( x , y ) = 1 indicates that pixel ( x , y ) belongs to the vessel foreground. We apply the Zhang-Suen thinning algorithm to iteratively remove boundary pixels that do not affect connectivity, thereby shrinking the vessels into a one-pixel-wide centerline. Let { M ( t ) } t 0 be the thinning sequence, defined by
M ( 0 ) = M , M ( t + 1 ) = M ( t ) R M ( t ) ,
where R ( M ( t ) ) denotes the set of pixels removable at iteration t. We first define the foreground set and its boundary at iteration t as
F ( t ) = { p Ω M ( t ) ( p ) = 1 } , F ( t ) = { p F ( t ) | q N 8 ( p ) , M ( t ) ( q ) = 0 } ,
where Ω is the image lattice and N 8 ( p ) denotes the 8-neighborhood of pixel p. For any candidate pixel p 1 and its clockwise-ordered neighbors { p 2 , , p 9 } , we define the number of foreground neighbors B ( p 1 ) and the number of 0 1 transitions A ( p 1 ) as
B ( p 1 ) = i = 2 9 M ( t ) ( p i ) ,
A ( p 1 ) = i = 2 9 1 M ( t ) ( p i ) M ( t ) ( p i + 1 ) , ( p 10 p 2 ) .
To avoid directional bias that may cause skeleton breakage, we adopt a two-subiteration scheme with parallel deletions. The removable sets for the two subiterations are given by
R ( t , 1 ) = p 1 F ( t ) | 2 B ( p 1 ) 6 , A ( p 1 ) = 1 , M ( t ) ( p 2 ) M ( t ) ( p 4 ) M ( t ) ( p 6 ) = 0 , M ( t ) ( p 4 ) M ( t ) ( p 6 ) M ( t ) ( p 8 ) = 0 ,
M ˜ ( t ) = M ( t ) R ( t , 1 ) ,
R ( t , 2 ) = p 1 F ˜ ( t ) | 2 B M ˜ ( p 1 ) 6 , A M ˜ ( p 1 ) = 1 , M ˜ ( t ) ( p 2 ) M ˜ ( t ) ( p 4 ) M ˜ ( t ) ( p 8 ) = 0 , M ˜ ( t ) ( p 2 ) M ˜ ( t ) ( p 6 ) M ˜ ( t ) ( p 8 ) = 0 ,
M ( t + 1 ) = M ˜ ( t ) R ( t , 2 ) .
F ˜ ( t ) is the boundary recomputed from M ˜ ( t ) , and B M ˜ , A M ˜ denote the corresponding quantities B and A evaluated on M ˜ ( t ) .
The iterations terminate when R ( T , 1 ) = and R ( T , 2 ) = , and the final skeleton is defined as
S = S ( M ) = M ( T ) { 0 , 1 } H × W .
The resulting S is a one-pixel-wide centerline skeleton that preserves the hierarchical branching structure of the vascular tree, enabling subsequent fractal-dimension analysis to focus on topological complexity rather than variations in vessel caliber.
For a set of spatial scales { ϵ k } , we perform box-counting on S by tiling the image with square boxes of side length ϵ k and counting the number of boxes N ( ϵ k ) that contain at least one foreground pixel. Figure 7 illustrates the box-counting procedure applied to the ground truth (GT) and the DeepLabv3+ prediction for the same image. Repeating this counting across multiple scales yields the paired observations log ( 1 / ϵ k ) , log N ( ϵ k ) . We then fit the following linear model in the log-log domain:
log N ( ϵ k ) = D log ϵ k + C ,
where the slope D provides an estimate of the fractal dimension of the vessel-tree skeleton. For each image, we compute the fractal dimensions of the ground-truth skeleton D GT and the predicted skeleton D pred , and define
D = D pred D GT
as a measure of the preservation of fractal topological complexity. Smaller values of D indicate that the model more faithfully preserves the geometric characteristics of the true coronary arterial tree. The final score for each model is obtained by averaging D over all test images.
Table 7 summarizes the fractal-geometry-based evaluation results for all representative methods. The mean fractal dimension of the ground-truth coronary arterial trees is approximately D GT = 1.1076 , which is comparable in magnitude to values reported for other two-dimensional vascular networks [90]. This indicates that the test set exhibits a realistic level of topological complexity.
Conventional encoder-decoder architectures exhibit notable limitations in preserving fractal complexity. The predicted fractal dimensions of U-Net and Attention U-Net are D pred = 1.0258 and 1.0327 , respectively, with Mean ( D ) values around 0.08 . This suggests that, despite achieving relatively high overlap metrics, these models tend to over-simplify fine distal branches and bifurcation patterns, leading to a substantial underestimation of the global fractal complexity. By maintaining high-resolution representations via parallel multi-resolution streams, HRNet increases the predicted fractal dimension to D pred = 1.0697 and reduces the deviation to Mean ( D ) = 0.0468 , corresponding to an approximate 43% reduction in fractal-dimension deviation relative to U-Net. Nevertheless, its predicted topology remains simplified compared with the ground truth, suggesting that preserving high-resolution features alone may be insufficient to fully recover the finest distal arborization.
DeepLabv3+, which incorporates multiscale atrous convolutions, partially alleviates this issue. Its predicted fractal dimension increases to D pred = 1.0714 , and Mean ( D ) decreases to 0.0531 , corresponding to an approximate 35% reduction in fractal-dimension deviation compared with U-Net. YOLO11-seg yields intermediate performance, with D pred = 1.0506 and Mean ( D ) = 0.0614 , indicating that a noticeable loss of fractal complexity persists.
The state-space-based U-Mamba and the proposed HR-UMamba++ model achieve the closest agreement with the ground-truth fractal geometry. U-Mamba attains D pred = 1.0712 with Mean ( D ) = 0.0377 . Building on this, HR-UMamba++ further improves the preservation of vascular topology by incorporating a persistent high-resolution branch and a rotation-aligned multi-directional Mamba scanning mechanism, yielding an average predicted fractal dimension of D pred = 1.0721 and the lowest Mean ( D ) = 0.0355 among all methods. Relative to U-Net, HR-UMamba++ reduces the fractal-dimension deviation by approximately 57%, and it also achieves reductions of about 42% and 33% compared with YOLO11-seg and DeepLabv3+, respectively, supporting improved preservation of the coronary arterial tree’s topological complexity.
To assess the statistical reliability of the observed improvements, we compared per-image fractal-dimension deviations ( D ) using paired nonparametric tests. A Friedman test across all methods indicated a significant overall difference ( χ F 2 ( 6 ) = 63.11 , p = 1.05 × 10 11 ; N = 50 , k = 7 ). Post-hoc comparisons between HR-UMamba++ and each baseline were then performed using two-sided Wilcoxon signed-rank tests with Holm–Bonferroni adjustment to control the family-wise error rate at α = 0.05 . The proposed method achieved significantly lower D than U-Net ( p adj = 3.75 × 10 5 ), Attention U-Net ( p adj = 3.75 × 10 5 ), HRNet ( p adj = 2.08 × 10 4 ), U-Mamba ( p adj = 0.041 ), DeepLabv3+ ( p adj = 1.05 × 10 5 ), and YOLO11-seg ( p adj = 4.18 × 10 9 ), supporting a statistically reliable improvement in fractal-topology consistency.

4.6. Clinical Usability Evaluation

To further evaluate the clinical usability of the model across different coronary angiographic projections, we randomly selected 20 representative frames from each of eight commonly used views: LAO view for RCA, AP caudal ( 20 ), LAO 45 + CAU 20 (spider view), RAO 30 + CAU 20 , RCA cranial view, RAO 30 + CRA 20 , AP cranial ( 30 ), and LAO 45 + CRA 20 . These frames were processed using the HR-UMamba++ model and independently evaluated by three cardiovascular specialists with more than five years of clinical experience, none of whom were involved in model development.
A five-point Likert scale was employed to rate the segmentation quality. Each physician assigned a single composite score from 1 to 5 after holistically considering the delineation completeness of the main trunk and major branches, the preservation of small vessels and bifurcations, and the overall diagnostic usability of the segmentation. The scoring criteria and their corresponding clinical interpretations are summarized in Table 8.
In total, 160 angiographic frames (20 per view) were evaluated by three physicians, yielding 480 individual ratings. The overall mean score across all views was 4.24 ± 0.49 , and the three raters exhibited highly similar means and standard deviations (Doctor 1: 4.24 ± 0.53 ; Doctor 2: 4.24 ± 0.52 ; Doctor 3: 4.24 ± 0.52 ), suggesting no evident systematic rater bias. As summarized in Table 9, the mean scores for individual views ranged from 3.92 to 4.52. Among them, RCA cranial view and LAO view for RCA achieved the highest ratings ( 4.52 ± 0.48 and 4.48 ± 0.48 , respectively), whereas LAO 45 + CRA 20 and RAO 30 + CAU 20 received slightly lower, yet still favorable, scores ( 3.92 ± 0.47 and 4.08 ± 0.51 , respectively). The remaining projections consistently maintained mean ratings around 4.2.
With respect to inter-rater reliability, the overall intraclass correlation coefficients were ICC(2,1) = 0.83 and ICC(2,3) = 0.94. Further view-specific analysis showed that the single-rater consistency ICC(2,1) ranged from 0.80 to 0.83, whereas the average-measures consistency ICC(2,3) ranged from 0.92 to 0.94 across the eight views (Table 9). According to established benchmarks [91], these values indicate good-to-excellent agreement among the three experts in their subjective assessment of the model outputs.
To further visualize the distribution of scores across different projections, we plotted violin diagrams of the 5-point Likert ratings for all three raters and eight views (Figure 8). The distributions exhibit a strong concentration of scores at 4 and 5 for all views, with only approximately 4.6% of ratings equal to 3 and no ratings of 1 or 2. These results indicate that HR-UMamba++ provides robust and clinically useful coronary vessel segmentations across diverse angiographic angles, supporting its potential for practical deployment in routine clinical workflows.

4.7. Computational Complexity Analysis

To quantify the computational cost of each design component, we report the number of trainable parameters (Params), floating-point operations per forward pass (FLOPs), peak GPU memory footprint during inference (Memory), and inference latency (Inference) under a single-image setting with an input size of 512 × 512 (batch size = 1). Each variant adds one component to the baseline U-Mamba, whereas HR-UMamba++ combines all components.
As shown in Table 10, introducing the multi-directional SSM increases the parameter count by only about 4.9%. However, the redundant computation incurred by multi-branch directional modeling leads to a 26.2% increase in FLOPs and a 39.1% increase in GPU memory, with the inference latency rising to 24.45 ms/img. The HR branch remains computationally efficient while continuously injecting high-resolution details, increasing the per-image inference time by only 3.23 ms.
Dense connections require cross-scale feature concatenation and fusion convolutions, resulting in approximately doubled parameters and GPU memory compared with the baseline, while the FLOPs surge to 2117.77G. Although dense connections constitute the dominant source of complexity growth, they also deliver the most pronounced gain among single-module variants, as adding dense connections improves IoU by 4.3% and reduces HD95 by 7.7% (Table 4).
Overall, the proposed HR-UMamba++ reaches 184.02M parameters (+187.5%), 2557.98G FLOPs (+522.2%), and a GPU memory footprint of 2.33 GB (+167.8%), with an inference latency of 48.86 ms/img (+277.0%). Despite the substantial increase in computational cost, it is noteworthy that clinical XCA sequences are typically acquired at 15 fps (i.e., a frame interval of approximately 66.7 ms) [63]. Since the proposed model’s per-frame latency is below this interval, it is able to complete inference before the next frame arrives, thereby achieving clinically meaningful real-time performance. Meanwhile, the 2.33 GB memory footprint is far below the capacity limits of mainstream GPUs, indicating that the model retains strong feasibility for clinical deployment.

5. Discussion

5.1. Summary of Main Findings

This study addresses long-standing challenges in single-frame coronary angiography-based vessel segmentation, including low contrast, structural ambiguity, and pronounced directional variability. Building upon the U-Mamba paradigm, we propose HR-UMamba++, centered on a rotation-aligned multi-directional Mamba module and incorporating a persistent high-resolution branch and a UNet++-style dense decoder topology.
Compared with existing approaches, traditional unsupervised methods struggle to preserve vessel connectivity and distal branches under low-contrast conditions and overlapping artifacts, and they are easily confounded by guidewires and bony structures [28,54,55,56,57]. Although recent supervised methods achieve substantial improvements in overall segmentation accuracy, most of them primarily focus on delineating major coronary trunks and remain limited in capturing the complete coronary arterial tree and higher-order bifurcation patterns [30,92,93,94]. State space model-based architectures such as Mamba and its variants have shown promising performance across various medical imaging tasks, yet they are still suboptimal when dealing with highly filamentous, strongly anisotropic vascular networks [95,96]. To address these limitations, we adapt the U-Mamba framework to the specific characteristics of XCA imaging and develop the HR-UMamba++ architecture. In a systematic comparison with representative methods, HR-UMamba++ achieves the best performance across all eight evaluation metrics (Table 6).
The ablation experiments clarify the functional contributions of each architectural component. Removing the UNet++-style dense decoding module leads to a decrease in Sensitivity and IoU by approximately 5.3% and 6.3%, respectively, and an increase in HD95 by nearly 20%. Qualitative results further reveal broken vessel continuity in low-perfusion regions and disruptions of complex bifurcation topology (Figure 4).
The ablation of the persistent high-resolution branch primarily results in deterioration in fine structural details. After removing this branch, Dice and IoU decrease by about 3.4% and 5.4%, respectively, whereas HD95 increases by approximately 10.2%. Frequency-domain power-spectrum analysis shows that the complete model exhibits, on average, a 4.10% increase in mid-to-high frequency energy within the normalized radial frequency range 0.25–0.9. This energy enhancement is concentrated in radial frequency bands aligned with the dominant vessel orientations (Figure 5), indicating that the high-resolution branch selectively preserves and amplifies high-frequency components associated with the geometric structure of the coronary vasculature.
The rotation-aligned multi-directional Mamba scanning mechanism mainly strengthens directional coherence and global connectivity of the vessel tree. When this module is removed, the Dice coefficient and IoU drop from 0.8706 and 0.7794 to 0.8478 and 0.7470, respectively. Furthermore, a single-component connectivity analysis on 200 angiograms with visually normal coronary anatomy (Table 5) shows that the complete model maintains a single foreground connected component in 93.5% of cases, whereas the ablated variant achieves this in only 87.5% of images. This difference in topological consistency is statistically significant, suggesting that rotating the original projection images into orientation-aligned views and feeding them into separate Mamba units for state-space evolution facilitates continuous vessel tracking along multiple directions.
In addition, we incorporate fractal-geometry-based descriptors to quantitatively analyze the fractal dimension of skeletonized coronary trees (Table 7). The fractal-dimension analysis demonstrates that the predictions of HR-UMamba++ exhibit the closest agreement with the ground-truth coronary trees, achieving the lowest Mean ( D ) among all evaluated methods. This finding indicates that HR-UMamba++ better preserves the self-similar geometric complexity of the coronary arterial tree in terms of fractal geometry.
Importantly, beyond algorithmic metrics, we also conduct a systematic evaluation of the clinical usability of HR-UMamba++ using expert ratings. The overall clinical score of the model is 4.24 ± 0.49 , with mean scores across different angiographic views ranging from 3.92 to 4.52. No ratings of 1 or 2 (completely unusable or potentially misleading) are observed, and only approximately 4.6% of the ratings correspond to a score of 3 (Table 9, Figure 8). These results indicate that HR-UMamba++ yields clinically usable vessel maps across diverse angiographic views, which is critical for downstream quantitative coronary analysis (QCA). In standard QCA workflows, centerlines and lumen contours are typically extracted along the entire coronary tree to compute key indices such as minimal lumen diameter (MLD), reference vessel diameter (RVD), diameter stenosis (DS), and lesion length (LL) [22,97]. Topological distortion in the segmented vessel tree can propagate to unstable centerline extraction and contour tracking, thereby biasing stenosis quantification and lesion-length measurements. By better preserving the global continuity of the vessel tree, as well as the connectivity of distal branches and bifurcation regions, HR-UMamba++ can provide a reliable structural prior for automated or semi-automated QCA measurements, reducing downstream manual editing and correction.
Overall, the proposed framework demonstrates stable and consistent advantages across quantitative evaluation metrics, structural integrity analyses, and subjective clinical assessments, confirming its suitability for complex coronary scenarios and its promising potential clinical value.

5.2. Limitations and Future Work

Despite the substantial performance gains achieved by HR-UMamba++, several limitations remain. First, the dataset used in this study is collected from a single medical center, where imaging equipment, exposure parameters, and operational protocols are relatively fixed. For algorithms intended to be deployed in multi-center and cross-device settings, it is necessary to incorporate data from different hospitals, vendors, and patient populations to more comprehensively assess the cross-domain robustness and generalizability of the model.
Second, in a small minority of cases, as illustrated in Figure 9, the model may still misclassify interventional devices as vascular structures in images containing artifacts such as catheters, guidewires, balloons, and pacing electrodes. Although many existing studies on XCA image analysis choose to exclude such abnormal frames during pre-processing [98,99], from a clinical perspective it is more desirable to improve the robustness of the algorithm while retaining these images. Enhancing artifact robustness is crucial for improving segmentation quality and generalization under realistic operating conditions. Therefore, future work will focus on constructing an extended dataset that includes a broader spectrum of artifact types and on incorporating a dedicated artifact-detection branch to locally suppress these regions. This strategy is expected to further improve the overall stability and reliability of HR-UMamba++ in real clinical environments.

6. Conclusions

This study targets the pervasive challenges of low contrast, structural ambiguity, and pronounced directional variability that hinder coronary arterial tree segmentation in single-frame X-ray coronary angiography. Building on the U-Mamba framework, we propose HR-UMamba++, a model specifically tailored for coronary vessel-tree segmentation. HR-UMamba++ is centered on a rotation-aligned multi-directional Mamba scan, complemented by a persistent high-resolution branch and a UNet++-style dense decoder. Together, these components enhance the representation of complex vascular networks from three complementary perspectives: fine-grained detail preservation, cross-scale topological modeling, and multi-directional long-range dependency capture.
Comprehensive baseline comparisons and ablation studies demonstrate that HR-UMamba++ achieves superior performance across all quantitative metrics, while exhibiting more reliable preservation of thin vessels, distal branch continuity, and global topological integrity. A box-counting-based fractal-dimension analysis of the coronary arterial tree further shows that HR-UMamba++ achieves closer agreement between the predicted and reference fractal dimensions and yields smaller deviations D than competing models, reflecting a better preservation of the self-similar geometric characteristics of the coronary arterial tree. In addition, expert subjective assessments across eight commonly used angiographic projections show that HR-UMamba++ attains high reliability in terms of clinical interpretability and usability, suggesting its potential to provide high-quality structural priors for quantitative coronary angiography, stenosis severity assessment, and interventional decision support.
Future work will extend the validation of HR-UMamba++ to multi-center datasets to further assess its cross-domain generalization capability, and will focus on designing more robust structural modeling strategies tailored to challenging clinical scenarios. These efforts are expected to facilitate the integration of HR-UMamba++ into routine clinical workflows.

Author Contributions

Conceptualization, X.Z., P.L. and Z.Z.; methodology, X.Z., P.L. and Z.Z.; software, X.Z. and Z.Z.; validation, X.Z. and P.L.; formal analysis, X.Z.; investigation, X.Z. and W.L.; resources, P.L. and W.L.; data curation, X.Z. and W.L.; writing—original draft preparation, X.Z.; writing—review and editing, P.L., Z.Z. and W.L.; visualization, X.Z.; supervision, P.L., Z.Z. and W.L.; project administration, P.L. and W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki. Ethical review and approval were waived by the Academic Ethics and Morality Committee of Shanghai Ocean University, which issued a formal waiver statement for this retrospective study on 12 November 2025.

Informed Consent Statement

Patient consent was waived by the Academic Ethics and Morality Committee of Shanghai Ocean University due to the retrospective design of the study and the exclusive use of fully anonymized clinical images.

Data Availability Statement

The data presented in this study are not publicly available due to institutional regulations and patient privacy restrictions. De-identified data may be made available from the corresponding author upon reasonable request and with permission from the data-providing institution.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Golub, I.S.; Termeie, O.G.; Kristo, S.; Schroeder, L.P.; Lakshmanan, S.; Shafter, A.M.; Hussein, L.; Verghese, D.; Aldana-Bitar, J.; Manubolu, V.S.; et al. Major Global Coronary Artery Calcium Guidelines. JACC Cardiovasc. Imaging 2023, 16, 98–117. [Google Scholar] [CrossRef] [PubMed]
  2. Shehu, M.N.; Adamu, U.G.; Ojji, D.B.; Ogah, O.S.; Sani, M.U. The Pandemic of Coronary Artery Disease in the Sub-Saharan Africa: What Clinicians Need to Know. Curr. Atheroscler. Rep. 2023, 25, 571–578. [Google Scholar] [CrossRef] [PubMed]
  3. Khan, M.A.; Hashim, M.J.; Mustafa, H.; Baniyas, M.Y.; Al Suwaidi, S.K.B.M.; AlKatheeri, R.; Alblooshi, F.M.K.; Almatrooshi, M.E.A.H.; Alzaabi, M.E.H.; Al Darmaki, R.S.; et al. Global Epidemiology of Ischemic Heart Disease: Results from the Global Burden of Disease Study. Cureus 2020, 12, e9349. [Google Scholar] [CrossRef] [PubMed]
  4. Stark, B.; Johnson, C.; Roth, G.A. Global prevalence of coronary artery disease: An update from the global burden of disease study. J. Am. Coll. Cardiol. 2024, 83, 2320. [Google Scholar] [CrossRef]
  5. Hay, S.I.; Ong, K.L.; Santomauro, D.F.; A, B.; Aalipour, M.A.; Aalruz, H.; Ababneh, H.S.; Abaraogu, U.O.; Abate, B.B.; Abbafati, C.; et al. Burden of 375 diseases and injuries, risk-attributable burden of 88 risk factors, and healthy life expectancy in 204 countries and territories, including 660 subnational locations, 1990–2023: A systematic analysis for the Global Burden of Disease Study 2023. Lancet 2025, 406, 1873–1922. [Google Scholar] [CrossRef]
  6. Vos, T.; Allen, C.; Arora, M.; Barber, R.M.; Bhutta, Z.A.; Brown, A.; Carter, A.; Casey, D.C.; Charlson, F.J.; Chen, A.Z.; et al. Global, regional, and national incidence, prevalence, and years lived with disability for 310 diseases and injuries, 1990–2015: A systematic analysis for the Global Burden of Disease Study 2015. Lancet 2016, 388, 1545–1602. [Google Scholar] [CrossRef]
  7. Wang, H.; Naghavi, M.; Allen, C.; Barber, R.M.; Bhutta, Z.A.; Carter, A.; Casey, D.C.; Charlson, F.J.; Chen, A.Z.; Coates, M.M.; et al. Global, regional, and national life expectancy, all-cause mortality, and cause-specific mortality for 249 causes of death, 1980–2015: A systematic analysis for the Global Burden of Disease Study 2015. Lancet 2016, 388, 1459–1544. [Google Scholar] [CrossRef]
  8. Gurgoglione, F.L.; Denegri, A.; Russo, M.; Calvieri, C.; Benatti, G.; Niccoli, G. Intracoronary Imaging of Coronary Atherosclerotic Plaque: From Assessment of Pathophysiological Mechanisms to Therapeutic Implication. Int. J. Mol. Sci. 2023, 24, 5155. [Google Scholar] [CrossRef]
  9. Nammour, A.Y.; Tabl, M.A.; Elrabat, K.E.; Wadie, M.T. Distal left main stenting versus coronary artery bypass graft in intermediate syntax score. Int. J. Cardiol. Sci. 2024, 6, 124–128. [Google Scholar] [CrossRef]
  10. Lubrano, V.; Balzan, S. Status of biomarkers for the identification of stable or vulnerable plaques in atherosclerosis. Clin. Sci. 2021, 135, 1981–1997. [Google Scholar] [CrossRef]
  11. Li, C.; Deng, C.; Shi, B.; Zhao, R. Thin-cap fibroatheroma in acute coronary syndrome: Implication for intravascular imaging assessment. Int. J. Cardiol. 2024, 405, 131965. [Google Scholar] [CrossRef] [PubMed]
  12. Byrne, R.A.; Rossello, X.; Coughlan, J.J.; Barbato, E.; Berry, C.; Chieffo, A.; Claeys, M.J.; Dan, G.A.; Dweck, M.R.; Galbraith, M.; et al. 2023 ESC Guidelines for the management of acute coronary syndromes: Developed by the task force on the management of acute coronary syndromes of the European Society of Cardiology (ESC). Eur. Heart J. Acute Cardiovasc. Care 2023, 13, 55–161. [Google Scholar] [CrossRef]
  13. Athreyas, A.G.; Nayak, S.R.; Chimmalagi, S.V.; Chachadi, K. Detection of CAD using XCA images(CADICA). In Proceedings of the 2025 International Conference on Biomedical Engineering and Sustainable Healthcare (ICBMESH), Manipal, India, 8–9 August 2025; pp. 1–5. [Google Scholar] [CrossRef]
  14. Marano, R.; Rovere, G.; Savino, G.; Flammia, F.C.; Carafa, M.R.P.; Steri, L.; Merlino, B.; Natale, L. CCTA in the diagnosis of coronary artery disease. La Radiol. Medica 2020, 125, 1102–1113. [Google Scholar] [CrossRef] [PubMed]
  15. Almagal, N.; Leahy, N.; Alqahtani, F.; Alsubai, S.; Elzomor, H.; Del Sole, P.A.; Sharif, R.; Sharif, F. Review of Optical Imaging in Coronary Artery Disease Diagnosis. J. Cardiovasc. Dev. Dis. 2025, 12, 288. [Google Scholar] [CrossRef]
  16. Mintz, G.S.; Guagliumi, G. Intravascular imaging in coronary artery disease. Lancet 2017, 390, 793–809. [Google Scholar] [CrossRef]
  17. Shivaie, S.; Tohidi, H.; Loganathan, P.; Kar, M.; Hashemy, H.; Shafiee, M.A. Interobserver Variability of Coronary Stenosis Characterized by Coronary Angiography: A Single-Center (Toronto General Hospital) Retrospective Chart Review by Staff Cardiologists. Vasc. Health Risk Manag. 2024, 20, 359–368. [Google Scholar] [CrossRef]
  18. Généreux, P.; Palmerini, T.; Caixeta, A.; Cristea, E.; Mehran, R.; Sanchez, R.; Lazar, D.; Jankovic, I.; Corral, M.D.; Dressler, O.; et al. SYNTAX Score Reproducibility and Variability Between Interventional Cardiologists, Core Laboratory Technicians, and Quantitative Coronary Measurements. Circ. Cardiovasc. Interv. 2011, 4, 553–561. [Google Scholar] [CrossRef]
  19. Serrano-Antón, B.; Villa, M.I.; Pendón-Minguillón, S.; Paramés-Estévez, S.; Otero-Cacho, A.; López-Otero, D.; Díaz-Fernández, B.; Bastos-Fernández, M.; González-Juanatey, J.R.; Muñuzuri, A.P. Unsupervised clustering based coronary artery segmentation. BioData Min. 2025, 18, 21. [Google Scholar] [CrossRef]
  20. Tran, D.S.; Huynh, A.K.; Huynh, A.D.; Nguyen-Thoi, T. Anatomy-specific two-stage YOLOv8 approach for improved coronary segmentation using the ARCADE dataset. Opt. Contin. 2025, 4, 303–317. [Google Scholar] [CrossRef]
  21. Zhang, H.; Mu, L.; Hu, S.; Nallamothu, B.K.; Lansky, A.J.; Xu, B.; Bouras, G.; Cohen, D.J.; Spertus, J.A.; Masoudi, F.A.; et al. Comparison of Physician Visual Assessment With Quantitative Coronary Angiography in Assessment of Stenosis Severity in China. JAMA Intern. Med. 2018, 178, 239–247. [Google Scholar] [CrossRef]
  22. Suzuki, N.; Asano, T.; Nakazawa, G.; Aoki, J.; Tanabe, K.; Hibi, K.; Ikari, Y.; Kozuma, K. Clinical expert consensus document on quantitative coronary angiography from the Japanese Association of Cardiovascular Intervention and Therapeutics. Cardiovasc. Interv. Ther. 2020, 35, 105–116. [Google Scholar] [CrossRef] [PubMed]
  23. Sen, T.; Kilit, C.; Astarcioglu, M.A.; Asarcikli, L.D.; Aksu, T.; Kafes, H.; Parspur, A.; Gozubuyuk, G.; Amasyali, B. Comparison of quantitative and qualitative coronary angiography: Computer versus the eye. Cardiovasc. J. Afr. 2018, 29, 278–282. [Google Scholar] [CrossRef] [PubMed]
  24. Gao, Z.; Wang, L.; Soroushmehr, R.; Wood, A.; Gryak, J.; Nallamothu, B.; Najarian, K. Vessel segmentation for X-ray coronary angiography using ensemble methods with deep learning and filter-based features. BMC Med. Imaging 2022, 22, 10. [Google Scholar] [CrossRef] [PubMed]
  25. Zhou, X.; Wang, H.; Zhang, L. A multi-backbone cascade and morphology-aware segmentation network for complex morphological X-ray coronary artery images. Comput. Med. Imaging Graph. 2025, 125, 102629. [Google Scholar] [CrossRef]
  26. Shakeri, S.; Almasganj, F. Online tree-structure-constrained RPCA for background subtraction of X-ray coronary angiography images. Comput. Methods Programs Biomed. 2025, 258, 108463. [Google Scholar] [CrossRef]
  27. Tu, S.; Koning, G.; Jukema, W.; Reiber, J.H.C. Assessment of obstruction length and optimal viewing angle from biplane X-ray angiograms. Int. J. Cardiovasc. Imaging 2010, 26, 5–17. [Google Scholar] [CrossRef]
  28. Lalinia, M.; Sahafi, A. Coronary Vessel Segmentation in X-ray Angiography Images Using Edge-Based Tracking Method. Sens. Imaging 2024, 25, 32. [Google Scholar] [CrossRef]
  29. He, Y.; Lyu, Z.; Mai, Y.; Li, S.; Hu, C.k. VM-CAGSeg: A vessel structure-aware state space model for coronary artery segmentation in angiography images. Front. Med. 2025, 12, 1661680. [Google Scholar] [CrossRef]
  30. Yang, S.; Kweon, J.; Roh, J.H.; Lee, J.H.; Kang, H.; Park, L.J.; Kim, D.J.; Yang, H.; Hur, J.; Kang, D.Y.; et al. Deep learning segmentation of major vessels in X-ray coronary angiography. Sci. Rep. 2019, 9, 16897. [Google Scholar] [CrossRef]
  31. Thuy, L.N.L.; Trinh, T.D.; Anh, L.H.; Kim, J.Y.; Hieu, H.T.; Bao, P.T. Coronary Vessel Segmentation by Coarse-to-Fine Strategy Using U-nets. BioMed Res. Int. 2021, 2021, 5548517. [Google Scholar] [CrossRef]
  32. Han, T.; Xiong, J.; Lin, T.; An, T.; Wang, C.; Zhu, J.; Li, Z.; Lu, L.; Zhang, Y.; Teng, G.J. Multiscale attention network via topology learning for cerebral vessel segmentation in angiography images. EngMedicine 2024, 1, 100004. [Google Scholar] [CrossRef]
  33. Nannini, G.; Saitta, S.; Baggiano, A.; Maragna, R.; Mushtaq, S.; Pontone, G.; Redaelli, A. A fully automated deep learning approach for coronary artery segmentation and comprehensive characterization. APL Bioeng. 2024, 8, 016103. [Google Scholar] [CrossRef] [PubMed]
  34. Xu, R.; Dai, L.; Wang, J.; Zhang, L.; Wang, Y. SADiff: Coronary Artery Segmentation in CT Angiography Using Spatial Attention and Diffusion Model. J. Imaging 2025, 11, 192. [Google Scholar] [CrossRef] [PubMed]
  35. Shit, S.; Paetzold, J.C.; Sekuboyina, A.; Ezhov, I.; Unger, A.; Zhylka, A.; Pluim, J.P.W.; Bauer, U.; Menze, B.H. clDice-a Novel Topology-Preserving Loss Function for Tubular Structure Segmentation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16555–16564. [Google Scholar] [CrossRef]
  36. Edwards, M.; Goodhew, S.C.; Badcock, D.R. Using perceptual tasks to selectively measure magnocellular and parvocellular performance: Rationale and a user’s guide. Psychon. Bull. Rev. 2021, 28, 1029–1050. [Google Scholar] [CrossRef] [PubMed]
  37. Gharleghi, R.; Chen, N.; Sowmya, A.; Beier, S. Towards automated coronary artery segmentation: A systematic review. Comput. Methods Programs Biomed. 2022, 225, 107015. [Google Scholar] [CrossRef]
  38. Qin, B.; Jin, M.; Hao, D.; Lv, Y.; Liu, Q.; Zhu, Y.; Ding, S.; Zhao, J.; Fei, B. Accurate vessel extraction via tensor completion of background layer in X-ray coronary angiograms. Pattern Recognit. 2019, 87, 38–54. [Google Scholar] [CrossRef]
  39. Cruz-Aceves, I.; Cervantes-Sanchez, F.; Hernandez-Aguirre, A.; Perez-Rodriguez, R.; Ochoa-Zezzatti, A. A novel Gaussian matched filter based on entropy minimization for automatic segmentation of coronary angiograms. Comput. Electr. Eng. 2016, 53, 263–275. [Google Scholar] [CrossRef]
  40. Vakilkandi, M.J.; Ahmadzadeh, M.R.; Amirfattahi, R.; Sadri, S. Accurate Coronary Vessel Extraction Using Fast Directional Filter Bank. In Proceedings of the 2011 7th Iranian Conference on Machine Vision and Image Processing, Tehran, Iran, 16–17 November 2011; pp. 1–5. [Google Scholar] [CrossRef]
  41. Yang, G.; Kitslaar, P.; Frenay, M.; Broersen, A.; Boogers, M.J.; Bax, J.J.; Reiber, J.H.C.; Dijkstra, J. Automatic centerline extraction of coronary arteries in coronary computed tomographic angiography. Int. J. Cardiovasc. Imaging 2012, 28, 921–933. [Google Scholar] [CrossRef]
  42. Sheng, X.; Fan, T.; Jin, X.; Jin, J.; Chen, Z.; Zheng, G.; Lu, M.; Zhu, Z. Extraction Method of Coronary Artery Blood Vessel Centerline in CT Coronary Angiography. IEEE Access 2019, 7, 170690–170702. [Google Scholar] [CrossRef]
  43. Huang, Y.; Yang, J.; Sun, Q.; Ma, S.; Yuan, Y.; Tan, W.; Cao, P.; Feng, C. Vessel filtering and segmentation of coronary CT angiographic images. Int. J. Comput. Assist. Radiol. Surg. 2022, 17, 1879–1890. [Google Scholar] [CrossRef]
  44. Chen, L.; Liu, W.; Balu, N.; Mossa-Basha, M.; Hatsukami, T.S.; Hwang, J.N.; Yuan, C. Deep Open Snake Tracker for Vessel Tracing. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2021, Strasbourg, France, 27 September–1 October 2021; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer: Cham, Switzerland, 2021; pp. 579–589. [Google Scholar] [CrossRef]
  45. Jeon, B. Deep Recursive Bayesian Tracking for Fully Automatic Centerline Extraction of Coronary Arteries in CT Images. Sensors 2021, 21, 6087. [Google Scholar] [CrossRef]
  46. Jawaid, M.M.; Liatsis, P.; Narejo, S. Automated Extraction of the Coronary Tree by Integrating Localized Aorta-Based Intensity Distribution Statistics in Active Contour Segmentation. In Proceedings of the 2015 International Conference on Developments of E-Systems Engineering (DeSE), Dubai, United Arab Emirates, 13–14 December 2015; pp. 83–87. [Google Scholar] [CrossRef]
  47. Chen, X.; Jiang, J.; Zhang, X. Automatic 3D coronary artery segmentation based on local region active contour model. J. Thorac. Dis. 2024, 16, 2563. [Google Scholar] [CrossRef] [PubMed]
  48. Gao, R.; Hou, Z.; Li, J.; Han, H.; Lu, B.; Zhou, S.K. Joint Coronary Centerline Extraction And Lumen Segmentation From Ccta Using Cnntracker And Vascular Graph Convolutional Network. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), Nice, France, 13–16 April 2021; pp. 1897–1901. [Google Scholar] [CrossRef]
  49. Lv, T.; Yang, G.; Zhang, Y.; Yang, J.; Chen, Y.; Shu, H.; Luo, L. Vessel segmentation using centerline constrained level set method. Multimed. Tools Appl. 2019, 78, 17051–17075. [Google Scholar] [CrossRef]
  50. Ma, G.; Yang, J.; Zhao, H. A coronary artery segmentation method based on region growing with variable sector search area. Technol. Health Care 2020, 28, 463–472. [Google Scholar] [CrossRef] [PubMed]
  51. Xia, S.; Zhu, H.; Liu, X.; Gong, M.; Huang, X.; Xu, L.; Zhang, H.; Guo, J. Vessel Segmentation of X-Ray Coronary Angiographic Image Sequence. IEEE Trans. Biomed. Eng. 2020, 67, 1338–1348. [Google Scholar] [CrossRef]
  52. Imran, A.; Li, J.; Pei, Y.; Yang, J.J.; Wang, Q. Comparative Analysis of Vessel Segmentation Techniques in Retinal Images. IEEE Access 2019, 7, 114862–114887. [Google Scholar] [CrossRef]
  53. L Srinidhi, C.; Aparna, P.; Rajan, J. Recent Advancements in Retinal Vessel Segmentation. J. Med. Syst. 2017, 41, 70. [Google Scholar] [CrossRef]
  54. Zai, S.; Abbas, A. An Effective Enhancement and Segmentation of Coronary Arteries in 2D Angiograms. In Proceedings of the 2018 International Conference on Smart Computing and Electronic Enterprise (ICSCEE), Shah Alam, Malaysia, 11–12 July 2018; pp. 1–4. [Google Scholar] [CrossRef]
  55. Longo, A.; Morscher, S.; Najafababdi, J.M.; Jüstel, D.; Zakian, C.; Ntziachristos, V. Assessment of hessian-based Frangi vesselness filter in optoacoustic imaging. Photoacoustics 2020, 20, 100200. [Google Scholar] [CrossRef]
  56. Fazlali, H.R.; Karimi, N.; Soroushmehr, S.M.R.; Shirani, S.; Nallamothu, B.K.; Ward, K.R.; Samavi, S.; Najarian, K. Vessel segmentation and catheter detection in X-ray angiograms using superpixels. Med. Biol. Eng. Comput. 2018, 56, 1515–1530. [Google Scholar] [CrossRef]
  57. Zai, S.; Ansari, M.A.; Moon, Y.S. Automatic and Effective Delineation of Coronary Arteries from CTA Data Using Two-Way Active Contour Model. IEICE Trans. Inf. 2017, 100, 901–909. [Google Scholar] [CrossRef]
  58. Aljuaid, A.; Anwar, M. Survey of Supervised Learning for Medical Image Processing. SN Comput. Sci. 2022, 3, 292. [Google Scholar] [CrossRef]
  59. Aslani, S.; Sarnel, H. A new supervised retinal vessel segmentation method based on robust hybrid features. Biomed. Signal Process. Control 2016, 30, 1–12. [Google Scholar] [CrossRef]
  60. He, H.; Banerjee, A.; Choudhury, R.P.; Grau, V. Deep learning based coronary vessels segmentation in X-ray angiography using temporal information. Med. Image Anal. 2025, 102, 103496. [Google Scholar] [CrossRef] [PubMed]
  61. Zhao, C.; Esposito, M.; Xu, Z.; Zhou, W. HAGMN-UQ: Hyper association graph matching network with uncertainty quantification for coronary artery semantic labeling. Med. Image Anal. 2025, 99, 103374. [Google Scholar] [CrossRef] [PubMed]
  62. Iyer, K.; Najarian, C.P.; Fattah, A.A.; Arthurs, C.J.; Soroushmehr, S.M.R.; Subban, V.; Sankardas, M.A.; Nadakuditi, R.R.; Nallamothu, B.K.; Figueroa, C.A. AngioNet: A convolutional neural network for vessel segmentation in X-ray angiography. Sci. Rep. 2021, 11, 18066. [Google Scholar] [CrossRef] [PubMed]
  63. Lee, C.K.; Hong, J.W.; Wu, C.L.; Hou, J.M.; Lin, Y.A.; Huang, K.C.; Tseng, P.H. Real-time coronary artery segmentation in CAG images: A semi-supervised deep learning strategy. Artif. Intell. Med. 2024, 153, 102888. [Google Scholar] [CrossRef]
  64. Chang, S.S.; Lin, C.T.; Wang, W.C.; Hsu, K.C.; Wu, Y.L.; Liu, C.H.; Fann, Y.C. Optimizing ensemble U-Net architectures for robust coronary vessel segmentation in angiographic images. Sci. Rep. 2024, 14, 6640. [Google Scholar] [CrossRef]
  65. Gao, Y.; Ai, D.; Wang, Y.; Cao, K.; Song, H.; Fan, J.; Xiao, D.; Zhang, T.; Wang, Y.; Yang, J. Spatio-Temporal correspondence attention network for vessel segmentation in X-ray coronary angiography. Biomed. Signal Process. Control 2025, 99, 106792. [Google Scholar] [CrossRef]
  66. Ramos-Cortez, J.S.; Alvarado-Carrillo, D.E.; Ovalle-Magallanes, E.; Avina-Cervantes, J.G. Lightweight U-Net for Blood Vessels Segmentation in X-Ray Coronary Angiography. J. Imaging 2025, 11, 106. [Google Scholar] [CrossRef]
  67. Wei, G.; Zeng, X.; Zhang, Q. FlowVM-Net: Enhanced Vessel Segmentation in X-Ray Coronary Angiography Using Temporal Information Fusion. J. Imaging Inform. Med. 2025. [Google Scholar] [CrossRef]
  68. Wang, J.; Lei, D.; Zhang, Y.; Yuan, J.; Liu, C.; Luo, B.; Liu, Q.; Wang, G. MedMamba: Multi-scale deformable attention via state space models for robust medical image segmentation. Biomed. Signal Process. Control 2026, 112, 108363. [Google Scholar] [CrossRef]
  69. Ma, J.; Li, F.; Wang, B. U-Mamba: Enhancing Long-range Dependency for Biomedical Image Segmentation. arXiv 2024, arXiv:2401.04722. [Google Scholar] [CrossRef]
  70. Liu, J.; Yang, H.; Zhou, H.Y.; Xi, Y.; Yu, L.; Li, C.; Liang, Y.; Shi, G.; Yu, Y.; Zhang, S.; et al. Swin-UMamba: Mamba-Based UNet with ImageNet-Based Pretraining. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2024, Marrakesh, Morocco, 6–10 October 2024; Linguraru, M.G., Dou, Q., Feragen, A., Giannarou, S., Glocker, B., Lekadir, K., Schnabel, J.A., Eds.; Springer: Cham, Switzerland, 2024; pp. 615–625. [Google Scholar] [CrossRef]
  71. Liu, J.; Yang, H.; Zhou, H.Y.; Yu, L.; Liang, Y.; Yu, Y.; Zhang, S.; Zheng, H.; Wang, S. Swin-UMamba†: Adapting Mamba-Based Vision Foundation Models for Medical Image Segmentation. IEEE Trans. Med. Imaging 2025, 44, 3898–3908. [Google Scholar] [CrossRef] [PubMed]
  72. Dang, S.; Matsumoto, T.; Takeuchi, Y.; Kudo, H. U-Mamba-Net: A highly efficient Mamba-based U-net style network for noisy and reverberant speech separation. In Proceedings of the 2024 Asia Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Macau, Macao, 3–6 December 2024; pp. 1–5. [Google Scholar] [CrossRef]
  73. Yuan, Y.; Cheng, Y. Medical image segmentation with UNet-based multi-scale context fusion. Sci. Rep. 2024, 14, 15687. [Google Scholar] [CrossRef] [PubMed]
  74. Ouyang, J.; Liu, S.; Peng, H.; Garg, H.; Thanh, D.N.H. LEA U-Net: A U-Net-based deep learning framework with local feature enhancement and attention for retinal vessel segmentation. Complex Intell. Syst. 2023, 9, 6753–6766. [Google Scholar] [CrossRef]
  75. Wang, R.; Lei, T.; Cui, R.; Zhang, B.; Meng, H.; Nandi, A.K. Medical image segmentation using deep learning: A survey. IET Image Process. 2022, 16, 1243–1267. [Google Scholar] [CrossRef]
  76. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep High-Resolution Representation Learning for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3349–3364. [Google Scholar] [CrossRef]
  77. Kesieme, E.B.; Omoregbee, B.; Ngaage, D.L.; Danton, M.H.D. Comprehensive Review of Coronary Artery Anatomy Relevant to Cardiac Surgery. Curr. Cardiol. Rev. 2025, 21, 27–41. [Google Scholar] [CrossRef]
  78. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. UNet++: A Nested U-Net Architecture for Medical Image Segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support, Granada, Spain, 20 September 2018; Stoyanov, D., Taylor, Z., Carneiro, G., Syeda-Mahmood, T., Martel, A., Maier-Hein, L., Tavares, J.M.R., Bradley, A., Papa, J.P., Belagiannis, V., et al., Eds.; Springer: Cham, Switzerland, 2018; pp. 3–11. [Google Scholar] [CrossRef]
  79. Isensee, F.; Jaeger, P.F.; Kohl, S.A.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef]
  80. Fagerland, M.W.; Lydersen, S.; Laake, P. The McNemar test for binary matched-pairs data: Mid-p and asymptotic are better than exact conditional. BMC Med. Res. Methodol. 2013, 13, 91. [Google Scholar] [CrossRef]
  81. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar] [CrossRef]
  82. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar] [CrossRef]
  83. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer: Cham, Switzerland, 2018; pp. 833–851. [Google Scholar] [CrossRef]
  84. Pokhrel, S.; Bhandari, S.; Vazquez, E.; Shrestha, Y.R.; Bhattarai, B. Cross-Task Data Augmentation by Pseudo-Label Generation for Region Based Coronary Artery Instance Segmentation. In Proceedings of the Data Engineering in Medical Imaging, Vancouver, BC, Canada, 8 October 2023; Bhattarai, B., Ali, S., Rau, A., Caramalau, R., Nguyen, A., Gyawali, P., Namburete, A., Stoyanov, D., Eds.; Springer: Cham, Switzerland, 2025; pp. 166–175. [Google Scholar] [CrossRef]
  85. Liu, T.; Lin, H.; Katsaggelos, A.K.; Kline, A. YOLO-Angio: An Algorithm for Coronary Anatomy Segmentation. arXiv 2023, arXiv:2310.15898. [Google Scholar] [CrossRef]
  86. Díaz-Gaxiola, E.; Yee-Rendon, A.; Vega-Lopez, I.F.; Campos-Leal, J.A.; García-Aguilar, I.; López-Rubio, E.; Luque-Baena, R.M. Experimental Assessment of YOLO Variants for Coronary Artery Disease Segmentation from Angiograms. Electronics 2025, 14, 2683. [Google Scholar] [CrossRef]
  87. Jocher, G.; Qiu, J. Ultralytics YOLO11. 2024. Available online: https://yolo11.com/ (accessed on 5 January 2026).
  88. Dérimay, F.; Kassab, G.S.; Finet, G. Chapter 24-Structure-function relation in the coronary artery tree: Theory and applications in interventional cardiology. In Biomechanics of Coronary Atherosclerotic Plaque; Ohayon, J., Finet, G., Pettigrew, R.I., Eds.; Academic Press: Cambridge, MA, USA, 2021; Volume 4, pp. 545–554. [Google Scholar] [CrossRef]
  89. Finet, G.; Gilard, M.; Perrenot, B.; Rioufol, G.; Motreff, P.; Gavit, L.; Prost, R. Fractal geometry of arterial coronary bifurcations: A quantitative coronary angiography and intravascular ultrasound analysis. EuroIntervention 2008, 3, 490–498. [Google Scholar] [CrossRef]
  90. Liew, G.; Mitchell, P.; Rochtchina, E.; Wong, T.Y.; Hsu, W.; Lee, M.L.; Wainwright, A.; Wang, J.J. Fractal analysis of retinal microvasculature and coronary heart disease mortality. Eur. Heart J. 2010, 32, 422–429. [Google Scholar] [CrossRef]
  91. Koo, T.K.; Li, M.Y. A Guideline of Selecting and Reporting Intraclass Correlation Coefficients for Reliability Research. J. Chiropr. Med. 2016, 15, 155–163. [Google Scholar] [CrossRef]
  92. Kaba, Ş.; Haci, H.; Isin, A.; Ilhan, A.; Conkbayir, C. The Application of Deep Learning for the Segmentation and Classification of Coronary Arteries. Diagnostics 2023, 13, 2274. [Google Scholar] [CrossRef] [PubMed]
  93. Zhang, X.; Sun, K.; Wu, D.; Xiong, X.; Liu, J.; Yao, L.; Li, S.; Wang, Y.; Feng, J.; Shen, D. An Anatomy- and Topology-Preserving Framework for Coronary Artery Segmentation. IEEE Trans. Med. Imaging 2024, 43, 723–733. [Google Scholar] [CrossRef] [PubMed]
  94. Park, J.; Kweon, J.; Kim, Y.I.; Back, I.; Chae, J.; Roh, J.H.; Kang, D.Y.; Lee, P.H.; Ahn, J.M.; Kang, S.J.; et al. Selective ensemble methods for deep learning segmentation of major vessels in invasive coronary angiography. Med. Phys. 2023, 50, 7822–7839. [Google Scholar] [CrossRef] [PubMed]
  95. Wang, H.; Chen, Y.; Chen, W.; Xu, H.; Zhao, H.; Sheng, B.; Fu, H.; Yang, G.; Zhu, L. Serp-Mamba: Advancing High-Resolution Retinal Vessel Segmentation with Selective State-Space Model. IEEE Trans. Med. Imaging 2025, 44, 4811–4825. [Google Scholar] [CrossRef]
  96. Shao, D.; Xu, R.; Ma, L.; Yi, S. Tubular-aware mamba for accurate retinal vessel segmentation: Preserving fine details and topological connectivity. Multimed. Syst. 2025, 31, 182. [Google Scholar] [CrossRef]
  97. Collet, C.; Onuma, Y.; Cavalcante, R.; Grundeken, M.J.; Généreux, P.; Popma, J.J.; Costa, R.A.; Stankovic, G.; Tu, S.; Reiber, J.H.; et al. Quantitative angiography methods for bifurcation lesions: A consensus statement update from the European Bifurcation Club. EuroIntervention 2017, 13, 115–123. [Google Scholar] [CrossRef]
  98. Avram, R.; Olgin, J.E.; Ahmed, Z.; Verreault-Julien, L.; Wan, A.; Barrios, J.; Abreau, S.; Wan, D.; Gonzalez, J.E.; Tardif, J.C.; et al. CathAI: Fully automated coronary angiography interpretation and stenosis estimation. NPJ Digit. Med. 2023, 6, 142. [Google Scholar] [CrossRef]
  99. Kruzhilov, I.; Mazanov, G.; Ponomarchuk, A.; Zubkova, G.; Shadrin, A.; Utegenov, R.; Blinov, P.; Bessonov, I. CoronaryDominance: Angiogram dataset for coronary dominance classification. Sci. Data 2025, 12, 341. [Google Scholar] [CrossRef]
Figure 1. Illustrative examples of common quality-limiting factors in XCA imaging, such as reduced contrast, noise contamination, and artifacts associated with guidewires and bony anatomy.
Figure 1. Illustrative examples of common quality-limiting factors in XCA imaging, such as reduced contrast, noise contamination, and artifacts associated with guidewires and bony anatomy.
Fractalfract 10 00043 g001
Figure 2. Overall architecture of the proposed HR-UMamba++ framework, together with implementation details of the High-Resolution Refinement Block, Dense Decoder Node, and Encoder Skip Node.
Figure 2. Overall architecture of the proposed HR-UMamba++ framework, together with implementation details of the High-Resolution Refinement Block, Dense Decoder Node, and Encoder Skip Node.
Fractalfract 10 00043 g002
Figure 3. Schematic illustration of the rotation-aligned multi-directional Mamba scanning module. The feature map is rotated by 0 , 90 , 45 , and 45 to align different orientations, and an independent Mamba scan is performed along each direction to capture long-range continuity of vessels in horizontal, vertical, and diagonal orientations.
Figure 3. Schematic illustration of the rotation-aligned multi-directional Mamba scanning module. The feature map is rotated by 0 , 90 , 45 , and 45 to align different orientations, and an independent Mamba scan is performed along each direction to capture long-range continuity of vessels in horizontal, vertical, and diagonal orientations.
Fractalfract 10 00043 g003
Figure 4. Visual comparison of segmentation results in the ablation study of dense connections. The figure presents three representative cases: Case 1 illustrates a low-blood-flow region with faint vascular contrast, Case 2 depicts thin-vessel structures, and Case 3 shows a complex bifurcation area.
Figure 4. Visual comparison of segmentation results in the ablation study of dense connections. The figure presents three representative cases: Case 1 illustrates a low-blood-flow region with faint vascular contrast, Case 2 depicts thin-vessel structures, and Case 3 shows a complex bifurcation area.
Fractalfract 10 00043 g004
Figure 5. Comparison of log-scaled power spectra between the proposed HR-UMamba++ and its variant without the persistent high-resolution (HR) branch. The first two columns display the logarithmically scaled 2D power spectra of the reconstructed vessel maps, while the third column shows the log-ratio spectrum ( log 10 ratio) between the two models. Red regions indicate frequency components where the complete model exhibits stronger responses, whereas blue regions highlight frequencies with relatively higher energy in the ablated variant.
Figure 5. Comparison of log-scaled power spectra between the proposed HR-UMamba++ and its variant without the persistent high-resolution (HR) branch. The first two columns display the logarithmically scaled 2D power spectra of the reconstructed vessel maps, while the third column shows the log-ratio spectrum ( log 10 ratio) between the two models. Red regions indicate frequency components where the complete model exhibits stronger responses, whereas blue regions highlight frequencies with relatively higher energy in the ablated variant.
Fractalfract 10 00043 g005
Figure 6. Qualitative visualization results of the proposed HR-UMamba++ model and six representative vessel-segmentation baselines.
Figure 6. Qualitative visualization results of the proposed HR-UMamba++ model and six representative vessel-segmentation baselines.
Fractalfract 10 00043 g006
Figure 7. Illustration of the box-counting procedure used to estimate the fractal dimension of coronary vessel centerline skeletons. (a,b) Skeletonized masks from the ground truth (GT) and the DeepLabv3+ prediction for the same image. (cf) Example square-grid tilings at two spatial scales ( ϵ k = 8 and 32 pixels), where N ( ϵ k ) denotes the number of boxes that contain at least one skeleton pixel. (g) Log-log plots of log N ( ϵ k ) versus log ( 1 / ϵ k ) over multiple scales with least-squares linear fits.
Figure 7. Illustration of the box-counting procedure used to estimate the fractal dimension of coronary vessel centerline skeletons. (a,b) Skeletonized masks from the ground truth (GT) and the DeepLabv3+ prediction for the same image. (cf) Example square-grid tilings at two spatial scales ( ϵ k = 8 and 32 pixels), where N ( ϵ k ) denotes the number of boxes that contain at least one skeleton pixel. (g) Log-log plots of log N ( ϵ k ) versus log ( 1 / ϵ k ) over multiple scales with least-squares linear fits.
Fractalfract 10 00043 g007
Figure 8. Distribution of Likert scores across the eight coronary angiography views. Violin widths reflect score densities; black points denote individual ratings; blue points represent mean scores.
Figure 8. Distribution of Likert scores across the eight coronary angiography views. Violin widths reflect score densities; black points denote individual ratings; blue points represent mean scores.
Fractalfract 10 00043 g008
Figure 9. Examples of common device-related artifacts in coronary angiography.
Figure 9. Examples of common device-related artifacts in coronary angiography.
Fractalfract 10 00043 g009
Table 1. Imaging and acquisition parameters of the coronary angiography dataset.
Table 1. Imaging and acquisition parameters of the coronary angiography dataset.
ItemValue
Number of images739
Number of patients374
Number of male patients211
Number of female patients163
Mean patient age (years)59.7
Image resolution 512 × 512
Tube voltage75–110 kVp
Tube current4–7 mA
Table 2. Comparison of our in-house dataset with published coronary angiography segmentation datasets.
Table 2. Comparison of our in-house dataset with published coronary angiography segmentation datasets.
StudyPatients (n)Annotated
Frames
Input FormImage Size
Lee et al., 2024 [63]213500Single frame512 × 512
Chang et al., 2024 [64]100619Single frame512 × 512
Gao et al., 2025 [65]1802-frame sequence512 × 512
Ramos-Cortez et al., 2025 [66]750Single frame300 × 300
512 × 512
Wei et al., 2025 [67]1105423-frame sequence800 × 800
Ours374739Single frame512 × 512
Table 3. Preprocessing and data augmentation parameters.
Table 3. Preprocessing and data augmentation parameters.
ItemParameter
Intensity normalizationZ-score
Multiplicative brightness perturbation 0.7 1.3
Additive brightness perturbation ± 0.1
Contrast perturbation 0.65 1.5
Gamma correction 0.7 1.5
Gaussian noise σ [ 0 , 0.1 ]
Gaussian blurkernel size = 3 , σ [ 1.0 , 3.0 ]
Random scaling 0.85 1.25
Random rotation ± 180
Bidirectional mirroringaxes = ( 0 , 1 )
Low-resolution simulationdown-sampling factor 0.5 1.0
Table 4. Ablation Analysis of the Proposed HR-UMamba++ Framework. The best results for each metric are highlighted in bold.
Table 4. Ablation Analysis of the Proposed HR-UMamba++ Framework. The best results for each metric are highlighted in bold.
Model VariantAccuracySensitivitySpecificityPPVNPVDiceIoUHD95
Proposed Model 0.9877 0.8179 0.9975 0.9461 0.9895 0.8706 0.7794 16.9943
Single-module ablations
w/o Dense conn.0.98430.77460.99680.92930.98670.83640.730020.3191
w/o HR branch0.98440.80030.99550.90020.98800.84130.737518.7250
w/o MD SSM0.98520.79540.99690.92590.98760.84780.747017.2209
Double-module ablations
w/o Dense conn. & HR Branch0.98420.76770.99630.91560.98710.82940.716420.3458
w/o Dense conn. & MD SSM0.98370.76720.99640.91740.98650.82750.716119.5036
w/o HR branch & MD SSM0.98420.78520.99600.91390.98730.83790.730318.7881
U-Mamba (baseline)0.98330.77680.99520.88520.98720.82150.700520.3530
Table 5. McNemar contingency table for single-component connectivity preservation.
Table 5. McNemar contingency table for single-component connectivity preservation.
Ablation CorrectAblation Wrong
Full Model Correct17215
Full Model Wrong310
Table 6. Quantitative comparison between the proposed HR-UMamba++ model and six representative vessel-segmentation baselines. The best results for each metric are highlighted in bold.
Table 6. Quantitative comparison between the proposed HR-UMamba++ model and six representative vessel-segmentation baselines. The best results for each metric are highlighted in bold.
Model VariantAccuracySensitivitySpecificityPPVNPVDiceIoUHD95
U-Net0.98040.71890.99380.85300.98550.77580.636227.6258
Attention U-Net0.98020.71530.99410.86160.98510.77770.638923.0382
HRNet0.98270.76840.99420.87900.98740.81840.696223.1386
U-Mamba0.98330.77680.99520.88520.98720.82150.700520.3530
DeepLabv3+0.98040.76920.99200.82900.98720.79370.662417.2217
YOLO11-seg0.98020.74900.99160.83640.98740.78820.651918.7155
Proposed Model0.98770.81790.99750.94610.98950.87060.779416.9943
Table 7. Fractal-geometry-based evaluation of coronary arterial tree topology. D pred denotes the mean fractal dimension of the predicted vessel skeleton, and D = | D pred D GT | denotes the deviation in fractal complexity. The best results for each metric are highlighted in bold.
Table 7. Fractal-geometry-based evaluation of coronary arterial tree topology. D pred denotes the mean fractal dimension of the predicted vessel skeleton, and D = | D pred D GT | denotes the deviation in fractal complexity. The best results for each metric are highlighted in bold.
Method D pred Mean ( D ) ± SD
U-Net1.02580.0818 ± 0.0390
Attention U-Net1.03270.0811 ± 0.0367
HRNet1.06970.0468 ± 0.0245
U-Mamba1.07120.0377 ± 0.0193
DeepLabv3+1.07140.0531 ± 0.0301
YOLO11-seg1.05060.0614 ± 0.0346
Proposed Model1.07210.0355 ± 0.0178
Table 8. Five-point Likert scale for expert ratings and corresponding clinical interpretations.
Table 8. Five-point Likert scale for expert ratings and corresponding clinical interpretations.
ScoreClinical Interpretation
1Completely unusable; segmentation is severely erroneous and cannot be used for any diagnostic reference.
2Partially usable, but with substantial errors that may mislead clinical judgment.
3Basically usable; segmentation contains noticeable errors but can still assist decision-making when interpreted with caution.
4Good; segmentation is reliable in most regions and can serve as stable auxiliary diagnostic information.
5Excellent; segmentation closely matches visual assessment and can be directly used as a reference for clinical decision-making.
Table 9. Expert rating statistics and inter-rater reliability for different coronary angiographic projections. Bold values indicate the overall summary across all views.
Table 9. Expert rating statistics and inter-rater reliability for different coronary angiographic projections. Bold values indicate the overall summary across all views.
ViewMean Score ± SDICC (2,1)ICC (2,3)
LAO view for RCA4.48 ± 0.480.8080.927
AP caudal ( 20 )4.18 ± 0.480.8080.927
LAO 45 + CAU 20 (spider view)4.22 ± 0.390.8120.928
RAO 30 + CAU 20 4.08 ± 0.510.8270.935
RCA cranial view4.52 ± 0.480.8090.927
RAO 30 + CRA 20 4.30 ± 0.470.8030.925
AP cranial ( 30 )4.23 ± 0.470.8060.926
LAO 45 + CRA 20 3.92 ± 0.470.8040.925
Overall4.24 ± 0.490.8280.935
Table 10. Complexity comparison of HR-UMamba++ and variants.
Table 10. Complexity comparison of HR-UMamba++ and variants.
MethodParams
(M)
FLOPs
(G)
Memory
(GB)
Inference
(ms/img)
Baseline U-Mamba64.01411.130.8712.96
+ Multi-directional SSM67.17518.681.2124.45
+ HR Branch88.88639.631.1616.19
+ Dense Connections131.872117.771.7832.04
HR-UMamba++184.022557.982.33
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Lu, P.; Zheng, Z.; Li, W. HR-UMamba++: A High-Resolution Multi-Directional Mamba Framework for Coronary Artery Segmentation in X-Ray Coronary Angiography. Fractal Fract. 2026, 10, 43. https://doi.org/10.3390/fractalfract10010043

AMA Style

Zhang X, Lu P, Zheng Z, Li W. HR-UMamba++: A High-Resolution Multi-Directional Mamba Framework for Coronary Artery Segmentation in X-Ray Coronary Angiography. Fractal and Fractional. 2026; 10(1):43. https://doi.org/10.3390/fractalfract10010043

Chicago/Turabian Style

Zhang, Xiuhan, Peng Lu, Zongsheng Zheng, and Wenhui Li. 2026. "HR-UMamba++: A High-Resolution Multi-Directional Mamba Framework for Coronary Artery Segmentation in X-Ray Coronary Angiography" Fractal and Fractional 10, no. 1: 43. https://doi.org/10.3390/fractalfract10010043

APA Style

Zhang, X., Lu, P., Zheng, Z., & Li, W. (2026). HR-UMamba++: A High-Resolution Multi-Directional Mamba Framework for Coronary Artery Segmentation in X-Ray Coronary Angiography. Fractal and Fractional, 10(1), 43. https://doi.org/10.3390/fractalfract10010043

Article Metrics

Back to TopTop