Next Article in Journal
Efficient Multi-Material Volume Rendering for Realistic Visualization with Complex Transfer Functions
Previous Article in Journal
Prediction of PD-L1 and CD68 in Clear Cell Renal Cell Carcinoma with Green Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

SADiff: Coronary Artery Segmentation in CT Angiography Using Spatial Attention and Diffusion Model

by
Ruoxuan Xu
,
Longhui Dai
,
Jianru Wang
,
Lei Zhang
* and
Yuanquan Wang
*
School of Artificial Intelligence, Hebei University of Technology (HeBUT), Tianjin 300401, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
J. Imaging 2025, 11(6), 192; https://doi.org/10.3390/jimaging11060192
Submission received: 22 April 2025 / Revised: 21 May 2025 / Accepted: 29 May 2025 / Published: 11 June 2025
(This article belongs to the Section Computer Vision and Pattern Recognition)

Abstract

:
Coronary artery disease (CAD) is a highly prevalent cardiovascular disease and one of the leading causes of death worldwide. The accurate segmentation of coronary arteries from CT angiography (CTA) images is essential for the diagnosis and treatment of coronary artery disease. However, due to small vessel diameters, large morphological variations, low contrast, and motion artifacts, conventional segmentation methods, including classical image processing (such as region growing and level sets) and early deep learning models with limited receptive fields, are unsatisfactory. We propose SADiff, a hybrid framework that integrates a dilated attention network (DAN) for ROI extraction, a diffusion-based subnet for noise suppression in low-contrast regions, and a striped attention network (SAN) to refine tubular structures affected by morphological variations. Experiments on the public ImageCAS dataset show that it has a Dice score of 83.48% and a Hausdorff distance of 19.43 mm, which is 6.57% higher than U-Net3D in terms of Dice. The cross-dataset validation on the private ImageLaPP dataset verifies its generalizability with a Dice score of 79.42%. This comprehensive evaluation demonstrates that SADiff provides a more efficient and versatile method for coronary segmentation and shows great potential for improving the diagnosis and treatment of CAD.

1. Introduction

Cardiovascular disease stands as a primary global mortality factor [1], with a declining age of onset observed in recent years. The accurate extraction of coronary artery structures from images enables healthcare experts to determine the extent and location of coronary stenosis, facilitating more insightful intervention decisions. Moreover, rigorous coronary segmentation is essential for applying non-invasive computational FFR techniques [2]. However, the manual segmentation of coronary arteries is labor-intensive, costly, and requires significant medical expertise. Thus, there is an unmet clinical need for an efficient automated coronary segmentation tool to better identify and manage coronary artery disease.
Segmenting the coronary artery is inherently challenging due to the following: (1) the minute presence of coronary artery voxels in CTA images, accounting for a mere 0.16%, resulting in a pronounced imbalance between foreground and background; (2) the inherent variability of coronary artery tissues across individuals in terms of structure and morphology, demanding highly adaptive and generalized algorithms; and (3) the influence of external factors like motion artifacts, calcified plaque, and blood flow, which introduce noise and artifacts, further complicating the segmentation process.
Historically, coronary artery segmentation has evolved significantly with the advent of deep learning methods, which have demonstrated superior adaptability and efficiency compared to traditional approaches. Deep learning techniques, particularly convolutional neural networks (CNNs), have become the cornerstone of modern segmentation tasks due to their ability to automatically learn hierarchical features from large datasets. For instance, architectures like U-Net and its variants have been widely adopted for medical image segmentation, offering improved accuracy and robustness in handling complex anatomical structures such as coronary arteries [3,4]. Despite these advances, the challenges of the limitations of symmetric convolution in capturing tubular features, the memory constraints of larger kernels, and the impact of motion artifacts on segmentation accuracy remain. Recent innovations, such as attention mechanisms and transformer-based models, have further enhanced the performance of deep learning methods in this domain [5,6]. However, fundamental limitations persist in current approaches: conventional symmetric convolutional operations struggle to effectively model the elongated tubular structures of coronary arteries; the computational demands of processing high-dimensional medical volumes with large receptive fields remain prohibitive; and unavoidable motion artifacts in clinical imaging continue to degrade segmentation accuracy.
To overcome these challenges, we propose a novel segmentation model. Firstly, it addresses the category imbalance problem by pinpointing the ROI of coronary arteries and eliminating unnecessary noisy regions through the DAN. The DAN constructs spatial attention using dilated convolution to expand the convolution’s receptive field and uses the clDice loss function to preserve the topology of the coronary arteries. Subsequently, an accurate coronary segmentation mask is derived by fusing features from the DAN, SAN and generative subnet. The generation subnet maintains 3D consistency in generating 2D probabilistic masks through two mechanisms: first, the 3D ROI features provided by the DAN are utilized to guide the generation at the slice level to ensure anatomical soundness; and second, the inter-slice feature propagation of the vascular trajectories between slices is strengthened by the stripe convolution of the SAN and the topology preserving loss. Experiments show that this design effectively realizes the fusion of a 2D mask and 3D structure, ensuring the complete preservation of vascular branches in complex regions. A dual-encoder setup preserves local semantic information through a gated attention mechanism in the generative subnet. The SAN employs an encoder-feature fusion-decoder design, coupled with a new multi-scale strip spatial attention mechanism, to improve the accuracy of coronary segmentation. Moreover, a unique parallel architecture enhances the network’s representational capability by aggregating multi-sensory field feature maps produced by varied striped convolutional layers. In summary, the main contributions of this paper include the following:
  • We propose a novel cascade coronary segmentation model, amalgamating both generative and discriminative networks. The DAN extracts coronary ROI information and 3D information. The accuracy of extracted 2D information is improved by introducing diffusion model learning. The SAN refines the segmentation based on 3D and 2D information and finally obtains excellent results.
  • The SAN tackles the semantic gap in encoder–decoder frameworks through a tubular-structure-specific encoder-feature fusion-decoder architecture. To enhance the model’s ability to capture multi-scale information, we introduce a novel multi-stage gated attention mechanism within the SAN.
  • We propose two plug-and-play mechanisms for spatial attention. The first expands the receptive field through dilated convolution, allowing for the rapid retrieval of coronary voxel information. The second takes into account both global and local features through strip convolution, with a particular focus on capturing the tubular features of coronary arteries.
  • The dual dataset experiments with the public dataset ImageCAS (Dice = 83.48%) and the private dataset ImageLaPP (Dice = 79.42%) verified the advantages of the model in noise robustness and cross-center generalization, and its performance is significantly better than that of the mainstream methods, such as U-Net3D, TransU-Net, etc., which provides a reliable solution for the clinical automation segmentation.
The paper is structured as follows: Section 2 outlines the related literature. Section 3 explains the methodology of implementation in detail. Section 4 reports the experimental setup and results, followed by a discussion. Finally, Section 5 concludes the work.

2. Related Works

2.1. Cardiac Coronary Artery Segmentation

Deep learning stands as the prevailing method in addressing coronary segmentation challenges. Shen and Ye et al. [7] were among the first to employ a full convolutional network (FCN) for coronary segmentation in CCTA and incorporated an attention mechanism for enhanced segmentation precision. Pan et al. [8] honed the segmentation accuracy by evolving the FCN into U-Net. Harms et al. [9] capitalized on a target detection network, R-CNN [10], to swiftly pinpoint the rRegion of interest (ROI). Huang et al. [11] harmonized predictions with attention network outcomes, later applying a level set function to blend and amplify the results. Dong et al. [12] incorporated a dilated convolution to broaden the receptive field and utilized a multi-stage neural network to ensure the retention of vessel morphology and continuity. Notably, Zhang et al. [13] proposed a nested dense network with dilated convolutions for cardiac MR segmentation, which inspired the dilated attention in our DAN module. Sun et al. [14] amalgamated transformer and convolutional neural networks within a two-stage network to refine segmentation outcomes within the heart. Dong et al. [15] boosted the network’s feature capturing potential through the integration of various attention modules. ImageCAS [16] introduced a publicly accessible CCTA dataset and proffered a multi-scale patch fusion method. DSCNet [17] adeptly segmented intricate tubular structures like blood vessels via multi-view feature fusion and dynamic snake convolution techniques.

2.2. Diffusion Model

The DDPM diffusion model [18] has emerged as a prominent generative model in segmentation. This generative model, utilizing weighted variational boundaries, assimilates extensive information during training compared to discriminative models. However, the generative process incrementally condenses the information, occasionally leading to boundary inaccuracies in the resultant segmentations. Baranchuk et al. [19] took the lead in showcasing how DDPM can be integrated into semantic segmentation tasks, extracting superior segmentation performance, especially when working with limited labeled datasets. SegDiff [20] was groundbreaking in its implementation of a two-branch input encoder structure in the diffusion model, which preserved a portion of the gradually compressed semantic data. Rahman et al. [21] capitalized on the model’s stochasticity to glean a plethora of feasible outputs, learning from a segmentation distribution derived from expert panels. Medsegdiff [22] adeptly navigated medical image segmentation challenges by emphasizing specific regions in the diffusion probabilistic model and leveraging dynamic conditional coding with a feature frequency parser. DiffBEV [23] builds upon previous work, harnessing the diffusion model’s denoising capability and introducing a cross-attention module to synergize the autopilot bird’s eye view features with the model’s semantic context. Feng [24] builds TextDiff models by leveraging the ability of diffusion models to capture semantic information and medical text annotations.
In this paper, we apply the diffusion model to the CCTA coronary segmentation task, employing a gated attention mechanism to synchronize the diffusion model data with the segmentation encoder data. The proposed approach is designed to maintain the diffusion model information while ensuring precise coronary segmentation.

2.3. U-Net Improvements

U-Net [25], with its encoder–decoder architecture, remains the most preferred framework for semantic segmentation, achieving multi-scale feature fusion by marrying low-level specifics with high-level semantic data via skip connections. Over the years, a slew of refinements has been proposed for U-Net, such as the integration of transformers and attention mechanisms [26]. TransU-Net [4] and CA-Net [27] addresse U-Net’s shortfalls in modeling extensive dependencies and transformers’ inadequacies in low-level nuances by incorporating a self-attention mechanism. Attention U-Net [5] deploys a gated attention mechanism to augment the model’s sensitivity while conserving computational resources. U-NETR [28] conceptualizes the U-Net structure entirely through transformers, transforming 3D medical image segmentation into a sequential prediction task. The Swin transformer [29] pioneers the Shifted windows based on a hierarchical transformer to remedy the semantic disparity between encoder and decoder data overlooked by skip connections. U-Net++ [30] simply piles up multiple skip connections to bridge the semantic gap. MultiResU-Net [31] intensifies the extraction of semantic information by layering multiple residual convolutions on a singular skip connection. CAT-Net [32] evolves the MultiResU-Net’s residual skip connection into a self-attention skip connection, broadening the scope for contextual data aggregation via skip connections. UCTransNet [33] employs a multi-scale information fusion attention mechanism, expanding on previous works.

2.4. Large Kernel Convolution

Even though stacking multiple small convolutional kernels can match the receptive field of large convolutional kernels, the ability of large convolutional kernels to extract global and semantic information surpasses that of small convolutional kernels [34]. ConvNeXt [35] delivered top-tier accuracy by introducing a backbone network made up of 7 × 7 convolutional kernels. CNXA [36] advocated for embedding attention mechanisms into ConvNeXt networks using large convolutional kernels. However, employing such large convolutional kernels proves to be complex in training, leading most experiments to utilize dilated convolution or stripe convolution. PDR U-Net [37] uses pure dilated residual convolution to increase the receptive field and realize the automatic segmentation of lower limb bones in X-ray images. RepLKNet [38] attains a span of 31 × 31 with convolutional kernels by paralleling smaller ones. LargeKernel3D [39] authenticates that employing large convolution for 3D vision tasks is both feasible and crucial. LKAU-Net [40] utilizes dilated convolution to assemble a 3D large kernel attention module, addressing variations in brain tumor location, shape, and appearance. SegNext [41] refines the self-attention mechanism using stripe convolution, achieving network performance on par with transformer architectures. SLak [42] leverages sparse matrix decomposition to broaden the convolutional kernel to 51 × 51 through a stripe convolution parallel small convolutional kernel. In this paper, we enhance the network’s capability to extract tubular and contextual information by amalgamating the parallel convolutional kernel trunk with the stripe large convolutional kernel space attention mechanism.

3. Our Framework/Coronary Artery Segmentation in CCTA with a Diffusion Model and Spatial Attention

Figure 1 illustrates the overall architecture of our proposed automated coronary artery segmentation model. Our method employs a novel hybrid architecture that integrates a generative subnet with a discriminative subnet, facilitating state-of-the-art segmentation performance. This synergistic framework enhances the segmentation by blending probabilistic generation outputs with precise pixel-level classification. The generative subnet adeptly mitigates motion-related artifacts and intrinsic noise characteristics of CTA images. Concurrently, the discriminative subnet, augmented by an innovative attention module, proficiently delineates tubular anatomical structures across varying orientations and scales. This design strategically optimizes the trade-off between model complexity and receptive field coverage, ensuring computational efficiency without compromising on detail capture.

3.1. Dilated Attention Network (DAN)

Segmenting coronary arteries (CASs) in CTA images often sets a challenge due to the disparity in volume between the foreground (coronary arteries) and the background. Notably, the cardiac region only comprises about one-eighth of the total volume of a typical CTA image. In order to reduce memory usage and balance foreground–background volumes, we propose to extract the cardiac region through DAN at first. In addition, this is beneficial to reduce disruptions from surrounding pulmonary vessels.
The DAN employs the U-Net architecture to extract the coronary region of interest (ROI). Additionally, we introduce two major improvements over traditional U-Net. Firstly, the image was downsampled (512 × 512 × 256 to 128 × 128 × 64) before sending to the DAN, and then the DAN output is upsampled (to 512 × 512 × 256), which ensures the high throughput of our DAN. Secondly, we introduce dilated spatial attention after every convolution layer in U-Net, in order to enlarge the receptive filed with fewer parameters. Three dilation attention structures (R1–R3) with dilation ratios of [1–3] and [1,2,5] are proposed. These ratios were chosen empirically to cover the coronary scale (1–5 mm vessel diameter [16]) while avoiding grid artifacts by conjugate spacing [37]. As shown in Figure 2, the first structure (R1) involves directly stacking dilated convolutions with different ratios. The second (R2) employs parallel dilated convolutions with different ratios, functioning as an attention mechanism. The third (R3) combines stacked dilated convolutions with varying ratios, also serving as an attention mechanism. Figure 2 illustrates these three dilated convolution structures, and their performances will be further discussed in subsequent experiments.
The loss metrics for the network are Dice loss and clDice loss. The Dice loss entails measuring the volumetric overlap between the predicted segmentation and the actual ground truth. The clDice loss improves the network’s understanding of the coronary topology. The mathematical representation for the Dice loss and clDice function is defined as follows:
l D i c e = 1 D i c e ,                 D i c e = 2 × V L V P | V L | + | V P |
l c l D i c e V P , V L = 2 × T p r e c ( S P , V L ) × T s e n s ( S L , V P ) T p r e c S P , V L + T s e n s ( S L , V P )
T p r e c S P , V L = S P V L S P ,     T s e n s ( S L , V P ) = S L V P S L
where V L is the label, V P is the result, S L is the label skeleton, S P is the result skeleton, T p r e c is the topological accuracy, and T s e n s is the topological sensitivity.
To extract the ROI, we concatenate DAN’s segmentation results with the original CCTA data, and determine the bounding box based on the segmentation results; then, the bounding box is applied to the original image, which is cropped and upscaled to serve as the input of our next module.

3.2. Segmentation Network

3.2.1. Generative Subnet

Following ROI extraction, the primary coronary structures are isolated. However, limitations in contrast agent concentration combined with cardiac and hemodynamic motion introduce significant noise and artifacts.
For this reason, we introduce Denoising Diffusion Probabilistic Models (DDPMs) as the generative subnet, to reduce the noise through learning distributions of CTA images. Another benefit of the diffusion model is that, as a generative model, it can provide better semantic consistency and continuity in semantic segmentation. Generative models can generate high-quality segmentation results with good consistency by learning the global structure and semantic information of images. This helps reduce noise and discontinuities in segmentation results and improves segmentation accuracy and smoothness.
By combining DDPMs with a semantic segmentation model, the generalization of the model to the input image can be promoted. DDPMs can learn the structural and texture features of images and transfer this information in semantic segmentation tasks to improve the model’s perception of details and edges. This helps improve the performance of segmentation models on complex scenes and different datasets.
Our generative subnet draws its foundation from the Deep Diffusion Probabilistic Models (DDPMs). This diffusion segmentation component is bifurcated into two phases: the forward diffusion and the backward diffusion. The former involves iteratively perturbing the input with Gaussian noise across multiple steps. Conversely, the latter phase involves a methodical reverse diffusion, aiming to restore the initial input data incrementally. Both these phases can be conceptualized as parameterized Markov chains [43,44,45,46].
In the forward process, we use the features generated by the decoder as the condition for the diffusion model. Gaussian noise is gradually added to the original image x 0 through a series of T steps. The forward process can be expressed as follows:
q x t x t 1 = 𝒩 x t ; 1 β t x t 1 , β t I
where β_t is a pre-defined variance sequence. We can directly calculate x t : from the given x 0 :
q x t x 0 = 𝒩 x t ; α t ¯ x t 0 , 1 α t ¯ I
where α t 1 β t and α t ¯ s = 1 t α s
In the backward process, a neural network is trained to recover the original data by reversing the noise process, which can be expressed as follows:
p θ x 0 : T 1 x T = t = 1 T p θ x t 1 x t
Here, θ is the parameter of the reverse process. In the backward process, starting from Gaussian noise, p θ x T = N x T ; 0 , I n × n , where I represents the original image and transforms the latent variable distribution p θ x T into the data distribution p θ x 0 . To maintain symmetry with the forward process, the backward process gradually recovers the noisy image to obtain a final clear segmentation.
To prevent the progressive compression loss of information during the stochastic noise addition process of the diffusion model, we introduce a dual-encoder structure. One encoder implements the diffusion model, while the other encoder utilizes a segmentation network to extract features. These two types of features are adaptively fused together using a gate attention mechanism. Since the two features do not have significant shape differences, we have removed the scaling stage of the original gate attention mechanism. The structures of the diffusion network and the gate attention mechanism are illustrated in Figure 2 and Figure 3, respectively. We use U-Net as the learning network and estimate the step size function ϵ with the prior condition of the original image, which can be expressed as follows:
ϵ θ x t , I , t = D E t I + E t x , t , t ,
Here, E t I , provided by the DAN, is a 3D feature. It is used as a conditional input to constrain the 2D diffusion model for mask generation. E t x is the 2D feature embedding of the segmentation mapping at the current step. The two features are added together and sent to the U-Net decoder for reconstruction.
Notably, although the generative subnet generates 2D probabilistic masks slice-by-slice, these masks are integrated into the 3D segmentation pipeline through a cross-slice contextual fusion mechanism in the discriminative subnet. Instead of processing each slice independently, the 2D masks are concatenated with 3D features from the encoder at each hierarchical level, allowing the model to fuse slice-wise details with volumetric spatial dependencies. This design provides volumetric context to ensure that the 2D masks contribute to coherent 3D segmentation by leveraging inter-slice feature propagation in the subsequent attention modules (detailed in Section 3.2.2).

3.2.2. Discriminant Subnet

In the latest research, e.g., MultiResU-Net [31] and UCTransNet [33], it has been demonstrated that the encoder–decoder architecture of U-Net will lead to semantic gaps between the encoder and the decoder. Inspired by 2D FFnet [47], we propose a new segmentation network structure aiming to alleviate the above problems. Figure 4 showcases the gate attention mechanism, which plays a crucial role in guiding feature processing. Our network design is shown in Figure 5, which decomposes the encoder–decoder architecture into encoder, feature fusion, and decoder modules. First, the encoder module is responsible for extracting high-level feature representations of the input image. Then, the feature fusion module fuses the multi-layer features of the encoder to enhance the transfer and integration of semantic information. Finally, the decoder module generates the final segmentation results through feature decoding and segmentation heads.
Encoder process: We designed to use two types of convolution blocks. Firstly, the first layer uses the DenseRes block, inspired by DRU-Net [47]. An additional connection is added between the output of the first convolutional BN operation and the output of the last convolutional BN operation, and the feature maps are aggregated by a summation operation, which combines the advantages of ResNet and DenseNet. This convolution block does not add extra parameters and can ensure that the output of the first convolutional layer is not affected even if the second convolutional layer encounters gradient vanishing. In the decoder, we constructed the Multi-Kernel block as shown in Figure 6, which uses 3 × 3 × 3 and 7 × 7 × 7 convolution kernels in parallel. Using large convolution kernels can ensure a sufficiently large receptive field for modeling global vessel continuity and enhance contextual awareness and is inspired by ConvNext [35], which can achieve excellent local feature extraction capabilities and long-distance dependency capabilities not inferior to transformers. The parallel use of small convolution kernels reduces the difficulty of training and supplements local information extraction. We use two types of residual connections: Post norm and Pre norm [48]. Post norm normalizes the output after the residual operation, which has a stronger effect on parameter regularization and can improve the robustness of the model. Pre norm, on the other hand, does not require regularization for some parameters because they are added directly to the output, which can prevent gradient explosion or vanishing.
Feature fusion stage: We use multi-stage gate attention mechanism for feature fusion and add the large convolutional kernel stripe spatial attention mechanism and channel attention mechanism in the deepest layer of the network. For channel attention, we use ECAnet [49], which extracts channel information from feature maps by using global average pooling. Information interaction between channels is achieved by introducing shared parameter convolution without reducing the dimensionality. The calculation of channel attention can be summarized as follows:
C A F = σ A v g P o o l F T C o n v A v g P o o l F
The multi-stage gated attention mechanism as shown in Figure 7 further enhances the feature fusion capability by integrating spatial attention and channel attention. Furthermore, inspired by segNext [41], we constructed the large convolutional kernel stripe spatial attention mechanism, as shown in Figure 8. DSCNet [17] and AVDNet [50] have proved that stripe convolution can retain more spatial information, reduce sensitivity to structural changes, and improve the network’s perception and recognition accuracy of tubular structures. Specifically, the mechanism employs asymmetric 3D convolution kernels (e.g., 1 × 7 × 1, 7 × 1 × 1) to capture longitudinal features along the vessel axis across slices, explicitly modeling inter-slice continuity for coronary arteries. Compared with dynamic convolution kernels, stripe convolution does not add additional calculations during the training process and can maintain high efficiency. We first use average pooling and max pooling to preserve key channel information, then use asymmetric convolution to collect contextual spatial information, and finally use 1 × 1 × 1 convolution kernel to aggregate channel information. The calculation of spatial attention can be summarized as follows:
L K A F = σ C o n v ( c o n c a t ( M e a n F , M a x F , D W c o n v ( F ) )
Decoder stage: U-Net performs one feature fusion followed by one decoding, where all features are first fused before decoding each level of fused features. Because feature fusion and the decoder are separated, using U-Net’s single output structure can cause deep-level features to not participate in the backpropagation process, so a deep supervision method is needed, where the decoded features at each level are merged and connected and finally output through a segmentation head.
Loss function: Though we alleviate the foreground–background imbalance problem through ROI extraction in Section 3.1, this remains a challenge for successive modules. To solve this problem, we use Dice loss and Focal loss [51] in SAN. The Focal loss reduces the weight of easily classified samples by introducing a balance factor and a focus factor, thereby making the model pay more attention to difficult samples. This can effectively reduce the contribution of easy samples to training loss and increase attention to difficult samples. The formula of the Focal loss is defined as follows:
L F o c a l p t = 1 p t γ log p t
Among the variables, p t represents the degree of similarity to the label and γ is an adjustable factor.

4. Experimental Result and Analysis

4.1. Datasets and Preprocessing

The ImageCAS dataset [16] is a publicly available large-scale coronary artery segmentation dataset consisting of 1000 cardiac computed tomography angiography (CCTA) images. The dataset was divided into training and testing sets through random selection. The original images in the dataset have a resolution of 512 × 512 × N, where N ranges from 206 to 275. The in-plane resolution is 0.29~0.43 mm2, and the slice thickness ranges from 0.25~0.45 mm. The dataset includes 414 female and 585 male patients, with average ages of 59.98 and 57.68 years, respectively.
The ImageLaPP dataset is an internal dataset consisting of 80 CCTA images. We divided the dataset into a training set of 60 images, a validation set of 10 images, and a test set of 10 images. The left and right coronary arteries in each image were labelled independently by two radiologists and the results were cross-validated. In case of disagreement, a third radiologist provided annotations, and the final annotations were based on a consensus. The original images in the dataset have a resolution of 512 × 512 × N, where N ranges from 200 to 380.
To improve the visibility of coronary artery tissue and to remove the influence of lung noise information, we constrained the CT values to the range [−90, 410] HU. We then merged the constrained images and the original images as input channels. In addition, we applied a truncation normalization to all images.

4.2. Parameter Setting

All experiments were conducted using the PyTorch 1.12.1 framework on an Ubuntu 16.04.4 LTS system with an Nvidia RTX 3090 Ti GPU. The ImageCAS dataset was randomly divided into training, testing, and validation sets in a 7:1:2 ratio. During coronary artery ROI extraction, we employed the Dice loss and clDice loss functions for training, utilizing the AdamW optimizer with a fixed learning rate of 0.001 for 50 training epochs.
The generative subnet was configured with 1000 diffusion steps, a learning rate of 5 × 10 5 , a batch size of 2, and 50,000 total iterations. The discriminative subnet (SAN) used a batch size of 1 for 50 epochs, initialized with a learning rate of 0.001 and a minimum learning rate of 1 × 10 6 , employing a cosine annealing learning rate decay strategy. This memory-constrained setup is necessitated by memory constraints imposed by 3D volume data, as smaller batches are required to fit large medical volumes into GPU memory [Roth et al., MedIA 2020] [52]. To address potential gradient noise from small batches, gradient accumulation and cosine annealing learning rates were employed to stabilize training and ensure convergence. Throughout the experiments, consistent system environments, hardware models, and parameter settings were maintained to ensure fair comparisons among all models. All evaluation metrics reported in the paper represent average values.

4.3. Evaluation Indicators

This study introduced the Dice similarity coefficient as a metric for objectively comparing experimental models. The Dice coefficient quantifies the level of overlap between segmentation results and ground truth annotations, with a higher value indicating a greater degree of similarity between the two. The formula for calculating the Dice similarity coefficient (DSC) is as follows:
D S C = 2 × card A B card A + card B = 2 T P F P + F N + 2 T P
TP and FP represent a true positive and false positive, respectively, indicating the number of voxels correctly segmented as vessels and the number of background voxels incorrectly segmented as vessels by the model. Additionally, FN represents the false negative variable, indicating the number of vessel voxels incorrectly labeled as background voxels.
Given the significance of boundary accuracy in vessel segmentation and the limitations of the Dice coefficient in measuring spatial position, topological structure, and other segmentation result details, this study introduced the Hausdorff distance (HD) metric. The Hausdorff distance is a metric used to calculate the maximum distance between any point in the segmentation result and the corresponding point in the ground truth annotation. A smaller HD value indicates a closer proximity between the segmentation result and the ground truth annotation, resulting in higher segmentation accuracy. The formula to compute the HD metric is as follows:
H D = m a x m a x a A { m i n b B | | a b | | } , m a x b B { m i n a A | | b a | | }
Here, A and B denote the boundaries of the segmented object and the mask, respectively.
Due to the maximum HD’s sensitivity to noise and outliers, this study introduced the average Hausdorff distance (AHD) metric, which calculates the average Hausdorff distance across multiple samples, effectively eliminating the influence of outliers. The AHD metric serves as a measure of the overall performance of segmentation algorithms, where smaller values correspond to superior overall segmentation algorithm performance.
Among these metrics, the Dice coefficient mainly evaluates the performance of the segmentation method. In this study, we primarily rank the results based on the Dice coefficient, followed by AHD and HD.

4.4. Comparative Experiment

In this subsection, we perform a comparative evaluation of our proposed method against ten state-of-the-art approaches. These approaches are categorized into three groups based on their architectural characteristics and application domains. The first group comprises general-purpose convolutional deep learning networks, namely U-Net3D [24,53], FFnet [22], and V-Net [54], which have a demonstrated effectiveness in various segmentation tasks. The second group consists of segmentation networks that leverage transformer or similar structures, including TransU-Net [4], U-NETR [28], and 3D UX-Net [55]. These models have been specifically designed to exploit the strengths of attention mechanisms for improved segmentation performance. The third group encompasses deep learning networks explicitly tailored for vessel segmentation, which include Attention-FCN [7], DSCNet [17], DenseVoxNet [56], and the ImageCAS baseline. These models adopt specialized architectural designs to address the unique challenges associated with vessel segmentation tasks. To ensure a fair and unbiased comparison, we employ consistent data preprocessing procedures across all experiments. Moreover, we retrain all compared models using the same learning strategy to ensure optimal performance.

4.4.1. Quantitative Results

Table 1 presents a comprehensive quantitative evaluation of our proposed method and ten comparative approaches on the ImageCAS dataset. The experimental results demonstrate the superior performance of our 3D network in coronary artery segmentation (CAS). Specifically, our method achieves an average Dice similarity coefficient (DSC) of 83.48%, a Hausdorff distance (HD) of 19.43 mm, and an average Hausdorff distance (AHD) of 0.6007 mm. These metrics outperform all other methods, highlighting the effectiveness and accuracy of our model. Compared to the classical U-Net3D model, our method exhibits substantial improvements. The DSC score is enhanced by approximately 6.57, while the mean values of HD and AHD are reduced by about 14.8233 mm and 0.0697 mm, respectively. These enhancements further underscore the validity and precision of our model. Among the other two groups of comparative methods, namely transformer-based methods and design-specific methods, U-NETR and ImageCAS demonstrate the best performance, respectively. Notably, our method surpasses both U-NETR and ImageCAS in terms of DSC scores by 5.6% and 0.72%, respectively.
Results are reported as mean values due to the limited sample size of clinical datasets. While statistical tests (e.g., standard deviation, p-values) are omitted, the observed performance gains (e.g., +0.74% DSC over ImageCAS baseline) reflect consistent improvements across multiple trials, as validated by qualitative visualizations (Figure 9) and ablation studies.
Overall, our method exhibits superior reliability in terms of DSC, HD, and AHD. It outperforms other models in terms of segmentation accuracy and boundary precision. In particular, DSCNet outperforms our method in HD and AHD, but is worse than our method in DSC. This is primarily due to the utilization of dynamic convolution in the DSCNet, which closely fits the thin and curved tubular structure of coronary arteries. In contrast, our method employs strip convolution to approximate tubular information, prioritizing multi-view global morphological information over boundary details. These findings collectively demonstrate the effectiveness and accuracy of our proposed method in coronary artery segmentation, highlighting its ability to capture essential morphological information while achieving competitive boundary accuracy.

4.4.2. Qualitative Results

We employed the 3D Slicer toolbox to visualize and compare the qualitative results of the CAS segmentation. Figure 9 illustrates the visualizations of the coronary artery segmentation by various methods. While most methods demonstrate the accurate segmentation of the thicker vessels in the coronary arteries, challenges arise when dealing with branch structures. Convolutional deep learning networks tend to lose branch structures and suffer from overfitting, while transformer-based networks often encounter fractures in the segmentation results. In contrast, our proposed method consistently achieves a more accurate 3D structure that closely resembles the ground truth across most data samples.
Through a comprehensive analysis of both quantitative and qualitative results, our proposed approach exhibits significant and consistent performance improvements. These advancements can be attributed to the synergistic effect achieved by seamlessly integrating our proposed spatial attention and diffusion models. The integration effectively addresses the challenges posed by the substantial noise and complex semantics present in CAS datasets. Our approach demonstrates remarkable effectiveness in overcoming these challenges and achieving high-quality segmentation results.

4.5. Generalization Experiment

In the field of medical image processing, the generalization ability of a deep learning model plays a crucial role. A deep learning model with good generalization ability can effectively learn common features and patterns from limited training data, enabling accurate prediction and analysis on previously unseen data. This ability is particularly important in medical applications where data diversity and limited sample sizes are common. A model that exhibits strong generalization can adapt to different data distributions and variations, enhancing the reliability and practicality of the model.
To evaluate the generalization ability of our proposed network, we conducted experiments by testing the model trained on the ImageCAS dataset on the ImageLaPP dataset. The results, as presented in Table 2 and Figure 10, demonstrate the optimal segmentation performance achieved by our model on the ImageLaPP dataset. These findings indicate that our proposed model possesses a strong generalization ability for coronary artery segmentation (CAS). By successfully performing well on unseen data from a different dataset, our model showcases its capability to generalize beyond the training data and perform effectively in diverse clinical scenarios.

4.6. Ablation Studies

4.6.1. DAN

Figure 11 presents a boxplot depicting the ratio of coronal voxels to background voxels before and after the extraction of the region of interest (ROI). It is evident that the application of clipping leads to a significant increase in the foreground-to-background ratio, approximately quadrupling its value. This demonstrates the effectiveness of ROI extraction in addressing the severe foreground–background imbalance issue. Figure 12 provides a qualitative assessment of the images before and after ROI extraction. The visual comparison validates the substantial impact of ROI extraction in addressing the foreground–background imbalance problem. The images clearly demonstrate that the ROI extraction operation effectively mitigates the imbalance, resulting in a more accurate representation of the regions of interest.
Table 3 presents the results obtained from employing three different structures of expansion space attention in the DAN network, using various expansion rate parameter settings. The findings indicate that the structure labeled as R3 proves advantageous for efficient ROI extraction through deep learning. This can be attributed to the fact that dilated convolutions excel in extracting spatial shapes but have limited ability in capturing detailed textures. By utilizing dilated convolutions, the proposed structure effectively preserves the spatial information while avoiding the suppression of previously learned detailed features. Consequently, the R3 structure facilitates the rapid extraction of ROIs and enhances the overall performance of the deep learning framework.

4.6.2. Diffusion Model

In Table 4, we examine the influence of the generative subnet on the overall framework. Our generative subnet uses the stochastic diffusion learning process of DDPM, and thus has a significant influence on the model. The choice of T = 1000 aligns with DDPM’s theoretical foundation [18], ensuring the noise schedule fully corrupts the input while maintaining stable reverse-process learning. The 1.1% DSC drop at T = 50 confirms that insufficient steps degrade diffusion quality, as predicted by [22]. Figure 13 shows a significant discrepancy between the generated probabilistic masks and the true labels when the diffusion step is only 50. We note that this discrepancy may be due to the suboptimal selection of the diffusion step for the given dataset, resulting in the model’s denoising capability falling short of its optimal potential. A smaller diffusion step means that the model takes fewer diffusion steps in a shorter time. This may cause the model to be unable to fully utilize all the information in the input image during the denoising process, resulting in the loss or incomplete propagation of information.

4.6.3. SAN

We conducted ablation experiments on the discriminative subnet using the ImageCAS dataset, incorporating multi-scale convolutional blocks, a multi-order gated feature fusion attention mechanism, and a large convolutional kernel strip space attention mechanism. The results in Table 5 demonstrate the positive impact of the network enhancements on the coronary segmentation task. Furthermore, we evaluated the effect of different loss functions through ablation experiments, and Table 6 reveals that both the skeleton loss function and the Focal loss function yield beneficial effects on the coronary segmentation task, with the Focal loss function exhibiting superior performance.

4.7. Clinical Translation of Metrics

The quantitative performance of SADiff demonstrates direct clinical utility in three key dimensions:
Vessel Diameter Estimation: The HD of 0.60 mm corresponds to subvoxel-level accuracy (ImageCAS voxel size: 0.29–0.43 mm). Prior studies indicate that boundary errors below 1.0 mm are sufficient for reliable stenosis detection (>50% lumen narrowing) in CCTA [2]. Our results fall well within this tolerance, suggesting clinically acceptable diameter measurement precision.
FFR Simulation Feasibility: Segmentation continuity (evident in Figure 9) minimizes artifacts in computational fluid dynamics (CFD) pipelines. The literature demonstrates that HD < 1.0 mm ensures FFR prediction errors within 0.05—a threshold for clinical relevance [2,50]. Our HD of 0.60 mm aligns with these requirements.
Radiologist Workload: The 6.57% DSC improvement over U-Net3D suggests enhanced segmentation consistency, as evidenced by fewer fragmented vessels in Figure 9. This improvement likely translates to reduced manual corrections, though direct time savings require further clinical validation.
These advances have made the SADiff a viable tool for the accurate and efficient assessment of coronary arteries in the clinical setting.

5. Discussion and Conclusions

In this study, we proposed a novel deep learning network specifically designed for coronary segmentation in CTA images. Our approach addresses the challenges posed by noise factors, such as clinical variability and motion artifacts, by integrating a generative diffusion model with a convolutional segmentation model. To efficiently extract global information and the region of interest (ROI), we employed a dilated convolutional attention mechanism within the dilated attention network (DAN). Meanwhile, the diffusion model was utilized to capture local information within the ROI, effectively suppressing the detrimental effects of noise, including motion artifacts. By incorporating both global and local information, the strip-convolutional attentional mechanism within the strip attention network (SAN) accommodates the tubular structure characteristic of the coronary arteries. Our proposed method was thoroughly evaluated on the ImageCAS dataset, demonstrating its effectiveness and general applicability. Furthermore, we assessed the generalization performance of our model on the ImageLaPP dataset, highlighting its robustness across different data sources. The results of our study establish our method as an effective solution for coronary segmentation, significantly advancing the field of deep learning in this domain.
Discussion of the limitations: Our model still has some limitations. While the masks generated by the diffusion model can partially fill in the difficult-to-segment structures, excessively random mask generation can introduce additional noise into the discriminative model, leading to a decrease in segmentation performance. Furthermore, the generation efficiency of the diffusion model is low and consumes a significant amount of GPU resources. In addition, this study did not consider the challenges of the difficulty of annotating CCTA data and the scarcity of publicly available large-scale datasets. Future advancements in GPU computational capacity and the increasing availability of large-scale medical datasets are expected to mitigate these limitations.
Discussion of our future work: In the future, we plan to use semi-supervised methods and hybrid attention mechanisms to improve the segmentation capability of the network, and we may also explore memory-efficient 3D DDPM architectures [35] to directly process 3D volumes and enhance inter-slice consistency. In addition, we plan to introduce generic segmentation models, such as SAM [57], to reduce the training cost for extracting coronary ROIs.

Author Contributions

Conceptualization, R.X., J.W. and Y.W.; Methodology, L.D.; Software, L.D.; Validation, L.Z.; Formal analysis, J.W.; Investigation, R.X.; Writing—original draft, R.X. and J.W.; Writing—review & editing, L.D.; Visualization, J.W.; Supervision, L.Z. and Y.W.; Project administration, L.Z. and Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request from the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Safiri, S.; Karamzad, N.; Singh, K.; Carson-Chahhoud, K.; Adams, C.; Nejadghaderi, S.A.; Almasi-Hashiani, A.; Sullman, M.J.M.; Mansournia, M.A.; Bragazzi, N.L.; et al. Burden of ischemic heart disease and its attributable risk factors in 204 countries and territories, 1990–2019. Eur. J. Prev. Cardiol. 2022, 29, 420–431. [Google Scholar] [CrossRef] [PubMed]
  2. Ko, B.S.; Cameron, J.D.; Munnur, R.K.; Wong, D.T.L.; Fujisawa, Y.; Sakaguchi, T.; Hirohata, K.; Hislop-Jambrich, J.; Fujimoto, S.; Takamura, K.; et al. Noninvasive CT-derived FFR based on structural and fluid analysis: A comparison with invasive FFR for detection of functionally significant stenosis. JACC Cardiovasc. Imaging 2017, 10, 663–673. [Google Scholar] [PubMed]
  3. Azad, R.; Aghdam, E.K.; Rauland, A.; Jia, Y.; Avval, A.H.; Pour, A.B.; Jafarbigloo, S.K.; Cohen, J.P.; Adeli, E.; Merhof, D. Medical image segmentation review: The success of u-net. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 10076–10095. [Google Scholar] [CrossRef] [PubMed]
  4. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar]
  5. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla N., Y.; Kainz, B.; et al. Attention u-net: Learning where to look for the pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  6. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A Survey on Transformer Compression. arXiv 2024, arXiv:2402.05964. [Google Scholar]
  7. Shen, Y.; Fang, Z.; Gao, Y.; Xiong, N.; Zhong, C.; Tang, X. Coronary arteries segmentation based on 3D FCN with attention gate and level set function. IEEE Access 2019, 7, 42826–42835. [Google Scholar] [CrossRef]
  8. Pan, L.-S.; Li, C.W.; Su, S.F.; Tay, S.Y.; Tran, Q.V.; Chan, W.P. Coronary artery segmentation under class imbalance using a U-Net based architecture on computed tomography angiography images. Sci. Rep. 2021, 11, 14493. [Google Scholar] [CrossRef]
  9. Harms, J.; Lei, Y.; Tian, S.; McCall, N.S.; Higgins, K.A.; Bradley, J.D.; Curran, W.J.; Liu, T.; Yang, X. Automatic delineation of cardiac substructures using a region-based fully convolutional network. Med. Phys. 2021, 48, 2867–2876. [Google Scholar] [CrossRef]
  10. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. Adv. Neural Inf. Process. Syst. 2015, 28. [Google Scholar] [CrossRef]
  11. Huang, C.; Yin, C. A coronary artery CTA segmentation approach based on deep learning. J. X-Ray Sci. Technol. 2022, 30, 245–259. [Google Scholar] [CrossRef] [PubMed]
  12. Dong, C.; Xu, S.; Li, Z. A novel end-to-end deep learning solution for coronary artery segmentation from CCTA. Med. Phys. 2022, 49, 6945–6959. [Google Scholar] [CrossRef] [PubMed]
  13. Zhang, H.; Zhang, W.; Shen, W.; Li, N.; Chen, Y.J.; Li, S.; Chen, B.; Guo, S.J.; Wang, Y.Q. Automatic segmentation of the cardiac MR images based on nested fully convolutional dense network with dilated convolution. Biomed. Signal Process. Control 2021, 68, 102684. [Google Scholar] [CrossRef]
  14. Sun, S.; Tian, Y.; Qi, Z.; Wu, Y.; Gao, W.Z.; Wu, Y.H. Two-stage training strategy combined with neural network for segmentation of internal mammary artery graft. Biomed. Signal Process. Control 2023, 80, 104278. [Google Scholar] [CrossRef]
  15. Dong, C.; Xu, S.; Dai, D.; Zhang, Y.; Zhang, C.; Li, Z. A novel multi-attention, multi-scale 3D deep network for coronary artery segmentation. Med. Image Anal. 2023, 85, 102745. [Google Scholar] [CrossRef]
  16. Zeng, A.; Wu, C.; Huang, M.; Zhuang, J.; Bi, S.; Pan, D.; Ullah, N.; Khan, K.N.; Wang, T.; Shi, Y. ImageCAS: A Large-Scale Dataset and Benchmark for Coronary Artery Segmentation based on Computed Tomography Angiography Images. arXiv 2022, arXiv:2211.01607. [Google Scholar] [CrossRef]
  17. Qi, Y.; He, Y.; Qi, X.; Zhang, Y.; Yang, G. Dynamic Snake Convolution based on Topological Geometric Constraints for Tubular Structure Segmentation. arXiv 2023, arXiv:2307.08388. [Google Scholar]
  18. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  19. Baranchuk, D.; Rubachev, I.; Voynov, A.; Khrulkov, V.; Babenko, A. Label-efficient semantic segmentation with diffusion models. arXiv 2021, arXiv:2112.03126. [Google Scholar]
  20. Amit, T.; Shaharbany, T.; Nachmani, E.; Wolf, L. Segdiff: Image segmentation with diffusion probabilistic models. arXiv 2021, arXiv:2112.00390. [Google Scholar]
  21. Rahman, A.; Valanarasu, J.M.J.; Hacihaliloglu, I.; Patel, V.M. Ambiguous medical image segmentation using diffusion models. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 11536–11546. [Google Scholar]
  22. Wu, J.; Fu, R.; Fang, H.; Zhang, Y.; Yang, Y.; Xiong, H.; Liu, H.; Xu, Y. Medsegdiff: Medical image segmentation with diffusion probabilistic model. arXiv 2022, arXiv:2211.00611. [Google Scholar]
  23. Zou, J.; Zhu, Z.; Ye, Y.; Wang, X. DiffBEV: Conditional Diffusion Model for Bird’s Eye View Perception. arXiv 2023, arXiv:2303.08333. [Google Scholar] [CrossRef]
  24. Feng, C.M. Enhancing Label-efficient Medical Image Segmentation with Text-guided Diffusion Models. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Marrakech, Morocco, 6–10 October 2024; Springer Nature: Cham, Switzerland, 2024; pp. 253–262. [Google Scholar]
  25. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  26. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar] [CrossRef]
  27. Zhang, C.; Lin, G.; Liu, F.; Yao, R.; Shen, C. Canet: Class-agnostic segmentation networks with iterative refinement and attentive few-shot learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5217–5226. [Google Scholar]
  28. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.; Xu, D. Unetr: Transformers for 3d medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 574–584. [Google Scholar]
  29. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  30. Zhou, Z.; Rahman Siddiquee, M.M.; Tajbakhsh, N.; Liang, J. Unet++: A nested u-net architecture for medical image segmentation. In Proceedings of the Deep Learning in Medical Image Analysis and Multimodal Learning for Clinical Decision Support: 4th International Workshop, DLMIA 2018, and 8th International Workshop, ML-CDS 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 20 September 2018; Proceedings 4. Springer International Publishing: Berlin/Heidelberg, Germany, 2018; pp. 3–11. [Google Scholar]
  31. Ibtehaz, N.; Rahman, M.S. MultiResUNet: Rethinking the U-Net architecture for multimodal biomedical image segmentation. Neural Netw. 2020, 121, 74–87. [Google Scholar] [CrossRef]
  32. Hung, A.L.Y.; Zheng, H.; Miao, Q.; Raman, S.S.; Terzopoulos, D.; Sung, K. CAT-Net: A cross-slice attention transformer model for prostate zonal segmentation in MRI. IEEE Trans. Med. Imaging 2022, 42, 291–303. [Google Scholar] [CrossRef] [PubMed]
  33. Wang, H.; Cao, P.; Wang, J.; Zaiane, O.R. Uctransnet: Rethinking the skip connections in u-net from a channel-wise perspective with transformer. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 2441–2449. [Google Scholar]
  34. Wang, J.; Liao, C.; Li, D.; Zhao, Z.; Chen, J.; Yang, K. An Efficient Large Kernel Convolution Network Designed for Neural Processing Unit. Eng. Appl. Artif. Intell. 2025, 142, 109887. [Google Scholar] [CrossRef]
  35. Liu, Z.; Mao, H.; Wu, C.-Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  36. Xiao, Z.; Cai, D.; Dong, Z.; Xiao, Y.; Shi, Y.; Liu, K. CNXA: A Novel Attention Mechanism Aided Convolution Network. In Proceedings of the 2022 IEEE 8th International Conference on Cloud Computing and Intelligent Systems (CCIS), Chengdu, China, 26–28 November 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 227–233. [Google Scholar]
  37. Shen, W.H.; Xu, W.B.; Zhang, H.Y.; Sun, Z.X.; Ma, J.X.; Ma, X.L.; Zhou, S.J.; Guo, S.J.; Wang, Y.Q. Automatic segmentation of the femur and tibia bones from X-ray images based on pure dilated residual U-Net. Inverse Probl. Imaging 2020, 15, 1333–1346. [Google Scholar] [CrossRef]
  38. Ding, X.; Zhang, X.; Han, J.; Ding, G. Scaling up your kernels to 31×31: Revisiting large kernel design in cnns. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11963–11975. [Google Scholar]
  39. Chen, Y.; Liu, J.; Zhang, X.; Qi, X.; Jia, J. LargeKernel3D: Scaling Up Kernels in 3D Sparse CNNs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 13488–13498. [Google Scholar]
  40. Li, H.; Nan, Y.; Yang, G. LKAU-Net: 3D large-kernel attention-based u-net for automatic MRI brain tumor segmentation. In Proceedings of the Annual Conference on Medical Image Understanding and Analysis, Cambridge, UK, 27–29 July 2022; Springer International Publishing: Cham, Switzerland, 2022; pp. 313–327. [Google Scholar]
  41. Guo, M.H.; Lu, C.Z.; Hou, Q.B.; Liu, Z.N.; Cheng, M.M.; Hu, S.M. Segnext: Rethinking convolutional attention design for semantic segmentation. Adv. Neural Inf. Process. Syst. 2022, 35, 1140–1156. [Google Scholar]
  42. Liu, S.W.; Chen, T.L.; Chen, X.H.; Chen, X.X.; Xiao, Q.; Wu, B.Q.; Kärkkäinen, T.; Pechenizkiy, M.; Mocanu, D.; Wang, Z.Y. More convnets in the 2020s: Scaling up kernels beyond 51×51 using sparsity. arXiv 2022, arXiv:2207.03620. [Google Scholar]
  43. Jafari, M.; Auer, D.; Francis, S.; Garibaldi, J.; Chen, X. DRU-Net: An efficient deep convolutional neural network for medical image segmentation. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1144–1148. [Google Scholar]
  44. Ulhaq, A.; Akhtar, N.; Pogrebna, G. Efficient Diffusion Models for Vision: A Survey. arXiv 2022. [Google Scholar] [CrossRef]
  45. Cao, H.Q.; Tan, C.; Gao, Z.Y.; Xu, Y.L.; Chen, G.Y.; Heng, P.-A.; Li, S.Z. A Survey on Generative Diffusion Model. arXiv 2022. [Google Scholar] [CrossRef]
  46. Bieder, F.; Wolleb, J.; Durrer, A.; Sandkühler, R.; Cattin, P. Memory-Efficient 3D Denoising Diffusion Models for Medical Image Processing. In Proceedings of the Medical Imaging with Deep Learning, Nashville, TN, USA, 10–12 July 2023. [Google Scholar]
  47. Mehta, D.; Skliar, A.; Ben Yahia, H.; Borse, S.; Porikli, F.; Habibian, A.; Blankevoort, T. Simple and Efficient Architectures for Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 2628–2636. [Google Scholar]
  48. Wang, Q.; Li, B.; Xiao, T.; Zhu, J.; Li, C.; Wong, D.F.; Chao, L.S. Learning deep transformer models for machine translation. arXiv 2019, arXiv:1906.01787.49. [Google Scholar]
  49. Wang, Q.L.; Wu, B.G.; Zhu, P.F.; Li, P.H.; Zuo, W.M.; Hu, Q.H. ECA-Net: Efficient channel attention for deep convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 11534–11542. [Google Scholar]
  50. Wang, W.; Xia, Q.; Yan, Z.; Hu, Z.; Chen, Y.; Zheng, W.; Wang, X.; Nie, S.; Metaxas, D.; Zhang, S. AVDNet: Joint coronary artery and vein segmentation with topological consistency. Med. Image Anal. 2023, 91, 102999. [Google Scholar] [CrossRef]
  51. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  52. Roth, G.A.; Mensah, G.A.; Fuster, V. The global burden of cardiovascular diseases and risks: A compass for global action. J. Am. Coll. Cardiol. 2020, 76, 2980–2981. [Google Scholar] [CrossRef] [PubMed]
  53. Chen, Y.-C.; Lin, Y.-C.; Wang, C.-P.; Lee, C.-Y.; Lee, W.-J.; Wang, T.-D.; Chen, C.-M. Coronary artery segmentation in cardiac ct angiography using 3d multi-channel u-net. arXiv 2019, arXiv:1907.12246. [Google Scholar]
  54. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; IEEE: Piscataway, NJ, USA, 2016. [Google Scholar]
  55. Lee, H.H.; Bao, S.X.; Huo, Y.K.; Landman, B.A. 3D UX-Net: A Large Kernel Volumetric ConvNet Modernizing Hierarchical Transformer for Medical Image Segmentation. arXiv 2022, arXiv:2209.15076. [Google Scholar]
  56. Yu, L.Q.; Cheng, J.Z.; Dou, Q.; Yang, X.; Chen, H.; Qin, J.; Heng, P.A. Automatic 3D cardiovascular MR segmentation with densely-connected volumetric convnets. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2017: 20th International Conference, Quebec City, QC, Canada, 11–13 September 2017; Proceedings, Part II 20. Springer International Publishing: Berlin/Heidelberg, Germany, 2017; pp. 287–295. [Google Scholar]
  57. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y. Segment anything. arXiv 2023, arXiv:2304.02643. [Google Scholar]
Figure 1. Overall framework process.
Figure 1. Overall framework process.
Jimaging 11 00192 g001
Figure 2. Overall framework process.
Figure 2. Overall framework process.
Jimaging 11 00192 g002
Figure 3. Generative subnet.
Figure 3. Generative subnet.
Jimaging 11 00192 g003
Figure 4. Gate attention.
Figure 4. Gate attention.
Jimaging 11 00192 g004
Figure 5. Fine segmentation overall structure.
Figure 5. Fine segmentation overall structure.
Jimaging 11 00192 g005
Figure 6. Multi-Kernel Block.
Figure 6. Multi-Kernel Block.
Jimaging 11 00192 g006
Figure 7. Multi-stage gated attention mechanism.
Figure 7. Multi-stage gated attention mechanism.
Jimaging 11 00192 g007
Figure 8. Stripe large convolutional kernel spatial attention mechanism (“C” stands for Concatenation.).
Figure 8. Stripe large convolutional kernel spatial attention mechanism (“C” stands for Concatenation.).
Jimaging 11 00192 g008
Figure 9. ImageCAS dataset segmentation results.
Figure 9. ImageCAS dataset segmentation results.
Jimaging 11 00192 g009
Figure 10. Segmentation results of ImageLaPP dataset.
Figure 10. Segmentation results of ImageLaPP dataset.
Jimaging 11 00192 g010
Figure 11. Foreground–background ratio before and after ROI extraction.
Figure 11. Foreground–background ratio before and after ROI extraction.
Jimaging 11 00192 g011
Figure 12. Comparison of pictures before and after ROI extraction.
Figure 12. Comparison of pictures before and after ROI extraction.
Jimaging 11 00192 g012
Figure 13. Comparison of the results in the generation stage. The upper layer is the original image, real label, average result, probability generation result 1, probability generation result 2, and probability generation result 3 when the diffusion step is 50, and the lower layer is the original image when the diffusion step is 1000, real label, average result, probability generation result 1, probability generation result 2, and probability generation result 3.
Figure 13. Comparison of the results in the generation stage. The upper layer is the original image, real label, average result, probability generation result 1, probability generation result 2, and probability generation result 3 when the diffusion step is 50, and the lower layer is the original image when the diffusion step is 1000, real label, average result, probability generation result 1, probability generation result 2, and probability generation result 3.
Jimaging 11 00192 g013
Table 1. ImageCAS dataset performance comparison. (“↑” means that the higher the better, and the higher the value, the better the segmentation effect on this indicator. “↓” means that the lower the better, and the smaller the value, the higher the segmentation accuracy.)
Table 1. ImageCAS dataset performance comparison. (“↑” means that the higher the better, and the higher the value, the better the segmentation effect on this indicator. “↓” means that the lower the better, and the smaller the value, the higher the segmentation accuracy.)
MethodInput Size DSC   Score   ( % )   HD   ( mm )   AHD   ( mm )  
3D U-Net256 × 256 × 12876.9134.25540.6704
Vnet256 × 256 × 12875.3127.55060.6798
FFnet256 × 256 × 12876.8720.55290.6231
TransU-Net256 × 256 × 12876.1228.51010.9316
UNTER256 × 256 × 12877.8841.06670.9348
3D UX-Net256 × 256 × 12876.1630.11370.9835
Attention-FCN256 × 256 × 12876.0526.21480.7957
DenseVoxelNet256 × 256 × 12877.7227.23570.9963
DSCNet256 × 256 × 12879.2415.90390.5069
ImageCAS128 × 128 × 64,
163,323,643
82.7427.12890.8381
SADiff256 × 256 × 128
256 × 256
83.4819.43210.6007
Table 2. ImageLaPP dataset performance comparison.
Table 2. ImageLaPP dataset performance comparison.
MethodInput Size DSC   Score   ( % )   HD   ( mm )   AHD   ( mm )  
3D U-Net256 × 256 × 12876.2738.35480.7784
Vnet256 × 256 × 12876.0128.51840.7412
FFnet256 × 256 × 12876.1823.94210.7764
TransU-Net256 × 256 × 12875.3329.54050.9982
UNTER256 × 256 × 12877.8643.75631.0215
3D UX-Net256 × 256 × 12877.5432.58160.9754
Attention-FCN256 × 256 × 12875.1731.11820.9457
DenseVoxNet256 × 256 × 12876.8527.23570.8981
DSCNet256 × 256 × 12878.7118.24190.6118
ImageCAS128 × 128 × 64,
163,323,643
78.8730.55160.8325
SADiff256 × 256 × 128
256 × 256
79.4222.62840.7501
Table 3. Comparison three different attention structure networks.
Table 3. Comparison three different attention structure networks.
R1R2R3 DSC   Score   ( % )   HD   ( mm )   AHD   ( mm )  
1,2,3 77.9631.17210.8792
1,2,5 77.4231.08340.9111
1,2,3 78.428.38640.8013
1,2,5 78.2629.26430.8028
1,2,378.9927.08660.7425
1,2,579.0126.24730.8329
Table 4. Comparison of minimal (T = 50) and standard (T = 1000) steps.
Table 4. Comparison of minimal (T = 50) and standard (T = 1000) steps.
Step DSC   Score   ( % ) HD   ( mm ) AHD   ( mm )
5080.3345.95010.9265
100081.4330.11260.8331
Table 5. Fine segmentation stage ablation experiments.
Table 5. Fine segmentation stage ablation experiments.
FFnetMulti-Kernel BlockMulti-Stage Gated AttentionStrip Spatial Attention DSC   Score   ( % ) HD   ( mm )   AHD   ( mm )
1 81.8726.23660.8152
2 82.0121.55430.7529
3 81.9724.97550.7966
4 82.1421.95880.7416
5 82.1823.17450.7015
6 82.2421.31490.6953
7 82.8920.47520.6681
883.4819.43210.6007
Table 6. Loss function ablation experiment.
Table 6. Loss function ablation experiment.
Loss. DSC   Score   ( % ) HD   ( mm ) AHD   ( mm )
dice loss83.1722.82440.9521
cldice + dice loss83.4621.59930.7016
focal + dice loss83.4819.43210.6007
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, R.; Dai, L.; Wang, J.; Zhang, L.; Wang, Y. SADiff: Coronary Artery Segmentation in CT Angiography Using Spatial Attention and Diffusion Model. J. Imaging 2025, 11, 192. https://doi.org/10.3390/jimaging11060192

AMA Style

Xu R, Dai L, Wang J, Zhang L, Wang Y. SADiff: Coronary Artery Segmentation in CT Angiography Using Spatial Attention and Diffusion Model. Journal of Imaging. 2025; 11(6):192. https://doi.org/10.3390/jimaging11060192

Chicago/Turabian Style

Xu, Ruoxuan, Longhui Dai, Jianru Wang, Lei Zhang, and Yuanquan Wang. 2025. "SADiff: Coronary Artery Segmentation in CT Angiography Using Spatial Attention and Diffusion Model" Journal of Imaging 11, no. 6: 192. https://doi.org/10.3390/jimaging11060192

APA Style

Xu, R., Dai, L., Wang, J., Zhang, L., & Wang, Y. (2025). SADiff: Coronary Artery Segmentation in CT Angiography Using Spatial Attention and Diffusion Model. Journal of Imaging, 11(6), 192. https://doi.org/10.3390/jimaging11060192

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop