Next Article in Journal
Medical Image Segmentation: A Comprehensive Review of Deep Learning-Based Methods
Previous Article in Journal
Real-Time Detection of Meningiomas by Image Segmentation: A Very Deep Transfer Learning Convolutional Neural Network Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Performance Evaluation of Image Segmentation Using Dual-Energy Spectral CT Images with Deep Learning Image Reconstruction: A Phantom Study

1
Department of Radiology, Beijing Children’s Hospital, Capital Medical University, National Center for Children’s Health, No. 56, Nanlishi Road, Xicheng District, Beijing 100045, China
2
Qingdao Academy of Chinese Medical Sciences, Shandong University of Traditional Chinese Medicine, Qingdao 266112, China
3
Children’s Hospital of Xinjiang Uygur Autonomous Region, Xinjiang Hospital of Beijing Children’ s Hospital, Urumqi 830054, China
*
Authors to whom correspondence should be addressed.
Tomography 2025, 11(5), 51; https://doi.org/10.3390/tomography11050051
Submission received: 9 March 2025 / Revised: 18 April 2025 / Accepted: 24 April 2025 / Published: 27 April 2025
(This article belongs to the Section Artificial Intelligence in Medical Imaging)

Abstract

:
Objectives: To evaluate the medical image segmentation performance of monochromatic images in various energy levels. Methods: The low-density module (25 mm in diameter, 6 Hounsfield Unit (HU) in density difference from background) from the ACR464 phantom was scanned at both 10 mGy and 5 mGy dose levels. Virtual monoenergetic images (VMIs) at different energy levels of 40, 50, 60, 68, 74, and 100 keV were generated. The images at 10 mGy reconstructed with 50% adaptive statistical iterative reconstruction veo (ASIR-V50%) were used to train an image segmentation model based on U-Net. The evaluation set used 5 mGy VMIs reconstructed with various reconstruction algorithms: FBP, ASIR-V50%, ASIR-V100%, deep learning image reconstruction (DLIR) with low (DLIR-L), medium (DLIR-M), and high (DLIR-H) strength levels. U-Net was employed as a tool to compare algorithm performance. Image noise and segmentation metrics, such as the DICE coefficient, intersection over union (IOU), sensitivity, and Hausdorff distance, were calculated to assess both image quality and segmentation performance. Results: DLIR-M and DLIR-H consistently achieved lower image noise and better segmentation performance, with the highest results observed at 60 keV, and DLIR-H had the lowest image noise across all energy levels. The performance metrics, including IOU, DICE, and sensitivity, were ranked in descending order with energy levels of 60 keV, 68 keV, 50 keV, 74 keV, 40 keV, and 100 keV. Specifically, at 60 keV, the average IOU values for each reconstruction method were 0.60 for FBP, 0.67 for ASIR-V50%, 0.68 for ASIR-V100%, 0.72 for DLIR-L, 0.75 for DLIR-M, and 0.75 for DLIR-H. The average DICE values were 0.75, 0.80, 0.82, 0.83, 0.85, and 0.86. The sensitivity values were 0.93, 0.91, 0.96, 0.95, 0.98, and 0.98. Conclusions: For low-density, non-enhancing objects under a low dose, the 60 keV VMIs performed better in automatic segmentation. DLIR-M and DLIR-H algorithms delivered the best results, whereas DLIR-H provided the lowest image noise and highest sensitivity.

1. Introduction

Automatic segmentation technology for medical images is widely used in clinical diagnosis, lesion measurement, radiotherapy, and other medical fields and plays a crucial role in accurately segmenting, measuring lesions, and planning radiation therapy. The quality of the CT image used for segmentation affects the results of automatic segmentation [1,2,3,4,5,6,7,8]. Dual-energy spectral computed tomography (DEsCT) offers expanded diagnostic capabilities by providing virtual monochromatic images (VMIs) of various photon energy levels, such as low kilo-electron volt (keV) images, to enhance the detection of small iodine contrast-enhancing lesions and improve image contrast compared to traditional CT [9,10,11,12,13]. This improved contrast of low-keV images in contrast-enhanced DEsCT can also be converted to reduce the radiation dose or contrast agent usage. In general, image noise increases as the energy level decreases in DEsCT, and iterative reconstruction (IR) algorithms such as the adaptive statistical iterative reconstruction veo (ASIR-V) have been used to reduce image noise. However, research indicates that the use of high-strength IR algorithms may lead to alterations in image texture, and the image texture change, together with the different image noises and CT values at various energy levels, could potentially impact the efficiency of artificial intelligence (AI) in automatic segmentation. This influence extends to the broader application of AI in CT imaging, where both efficiency and accuracy are critical [6,14,15,16]. Recently, a deep learning image reconstruction (DLIR) algorithm (TrueFidelityTM, GE HealthCare, Milwaukee, WI, USA) was introduced and incorporated into DEsCT and has demonstrated its ability to significantly reduce image noise without significantly altering the image texture compared to IR algorithms, resulting in improved image quality even under reduced radiation dose conditions [17,18]. Integrating DLIR with spectral CT images, particularly in low-dose scenarios, has the potential to improve the accuracy of automatic medical image segmentation further [19,20,21].
DEsCT was widely used in pediatrics, providing more diagnostic information, including tumors, lung perfusion, and metal implants [22,23,24,25]. Low-keV imaging reduced the usage of contrast agents in pediatric CT [26], enhanced the sensitivity in detecting children’s pulmonary embolism [27], and minimized the metal implant-based artifacts [28]. In the field of cardiology, low-keV imaging demonstrates the potential to reduce radiation doses and contrast agent usage while simultaneously offering the capability to differentiate materials with varying atomic numbers, which holds significant clinical implications [29,30]. Several powerful reconstruction algorithms adopted DEsCT to further improve the segmentation accuracy of radiotherapy images [31]. The iterative reconstruction algorithm improved clinical image segmentation results using model-based iterative reconstruction (MBIR) [32]. However, different VMI performances with DLIR on the low-density object’s segmentation accuracy, particularly in reduced radiation dose conditions, have not been comprehensively investigated. Although there is no significant statistical difference between DEsCT and single-energy CT, a series of slight differences, including CT attenuation, may potentially impact segmentation accuracy. Additionally, different energy levels and algorithms also influence the results [32,33,34]. Previously, segmentation methods demonstrated that deep learning reconstruction algorithms combined with spectral CT can reduce the radiation doses. Sixty keV is an effective value to maintain image quality while decreasing radiation exposure in pediatric CT [35]. Accurate segmentation plays a crucial role in measuring organ residual volume and informing clinical decisions [5,36,37].
The U-Net architecture has significantly advanced the domain of medical image segmentation by introducing an efficient and robust methodology for pixel-level image prediction [38,39]. A primary innovation of the architecture is its encoder-decoder structure, which facilitates the model’s ability to capture both high-level semantic features and detailed spatial information. The incorporation of skip connections between corresponding encoder and decoder layers ensures the preservation of high-resolution details, which is essential for tasks such as segmenting small or complex structures in medical images. This architecture has demonstrated high efficacy in applications such as tumor detection, organ segmentation, and other medical imaging tasks, yielding accurate segmentation results even with limited training data. The design of U-Net also supports training on smaller datasets, rendering it particularly valuable in the medical field, where annotated data are often scarce and costly to obtain. Its capability to produce precise segmentation maps has substantially enhanced the analysis and diagnosis of medical conditions from imaging data.
In this study, a low-resolution, non-enhancing object was used, and a U-Net deep learning algorithm was employed for automatic segmentation. The accuracies of the DLIR algorithm of low (DLIR-L), medium (DLIR-M), and high (DLIR-H) strength levels in scans with 50% of the standard radiation dose were compared with those of traditional CT reconstruction algorithms (filtered back-projection: FBP and ASIR-V) in segmenting low-density lesions. This method is derived from pediatric clinical application. When the radiation dose is reduced by half, segmentation tasks become more challenging, as noted in some studies examining model accuracy in the low-dose domain [40]. Utilizing phantom models addresses the ethical concerns associated with scanning pediatric patients [41]. Moreover, deep learning segmentation models for pediatric applications require parameter adjustments relative to adult models to achieve optimal performance [42]. The VMIs of various energy levels in DEsCT were also introduced to study the impact of different energy levels on segmentation accuracy. This approach highlights the potential of integrating DEsCT with DLIR to enhance the precision and effectiveness of medical imaging in clinical settings [43,44,45].

2. Materials and Methods

2.1. Phantom

This study involved phantom experiments that did not require ethical approval. An ACR 464 phantom (Phantom Laboratory Inc., Gammex, WI, USA) was employed in this study, as depicted in Figure 1a. Module 2, with an object of 6HU in density difference (DCT) and 25 mm in diameter, was utilized in this study. And Raw, ROI, and Maskmask were shown in Figure 1b.

2.2. CT Systems, Parameters for Acquisition and Reconstruction

All scans were performed on a 256-row CT (Revolution Apex, GE Healthcare, Milwaukee, WI, USA) utilizing the DEsCT imaging mode of fast tube voltage switching at two different dosage levels (volume CT dose indexes: CTDIvol of 5 mGy and 10 mGy determined on a 32 cm diameter reference phantom). For the 10 mGy scan, the tube current was set to 405 mA with a pitch of 0.992:1. For the 5 mGy scan, the tube current was set to 280 mA with a pitch of 1.375:1 [46]. The device was operated at a rotation speed of 0.5 s, captured images were captured with a pixel matrix size of 512 × 512. All the reconstructions were performed using a standard kernel with a slice thickness of 0.625 mm. The detector width along the z-axis was 40 mm. Six sets of virtual monochromatic images (VMIs) at energy levels of 40, 50, 60, 68, 74, and 100 keV were obtained. For the 10 mGy data acquisition, images were reconstructed using the standard ASIR-V at 50% strength level (ASIR-V50%), the currently commonly used reconstruction in clinical practice, and the default reconstruction weighting factor used in our center. For the 5 mGy data acquisition, images were reconstructed using the following algorithms: FBP, ASIR-V50% and 100% (ASIR-V100%) strength, and DLIR at low (DLIR-L), medium (DLIR-M), and high (DLIR-H) settings.

2.3. Deep Learning Model Construction

Model Training: The U-Net model training involved patching, which entailed cropping the original images into smaller sections containing low-density regions to augment the dataset for segmentation model training. The gold standard region of interest (ROI) and mask for low-density objects were delineated by two radiologists with ten years and three years of experience in radiology. Prior to the formal outlining process, all participating radiologists underwent standardized training, which included the use of consistent image display settings (e.g., window width and position) and outlining tools. The gold standard segmentation images were manually drawn to highlight regions with a slightly higher density. Masks were carefully outlined within the circular regions of the phantom, excluding the phantom frame and bed, to prevent interference in segmentation. As shown in Figure 1b. The original images were used for segmentation in the final test. A lower radiation dose was applied to amplify minor differences between the algorithms, helping to establish the applicable range and accuracy of each algorithm.
Training Set: The training set consisted of 10 mGy images reconstructed with ASIR-V50%, and the training set included eight images per energy level. In total, 48 images from 6 different VMIs were selected in this study; 80% of the images were utilized for model training, and 20% of the images were adopted for model evaluation.
Test Set: The 5 mGy images were reconstructed using the FBP, ASIR-V50%, ASIR-V100%, DLIR-L, DLIR-M, and DLIR-H algorithms across energy levels of 40, 50, 60, 68, 74, and 100 keV, resulting in 36 sets of images. Each set includes three consecutive slices of an object, with 108 sets of images in total. Four diagnostic doctors determined the optimal window width and window level for each energy level based on their clinical experience, yielding 12 segmentation results per set.
The pipeline of our performance evaluation method and the distribution of the dataset are shown in Figure 2.

2.4. Metrics for Deep Learning Automatic Segmentation Evaluation

To evaluate the performance of the deep learning automatic segmentation, several metrics were used, including the Jaccard Index (intersection over union, IOU), DICE Coefficient (DICE), sensitivity, and Hausdorff distance (consisting of Manhattan, Euclidean, and Cosine distances). The calculation formula could be found in Appendix A.

2.5. Deep Learning Segmentation Model

To evaluate the performance of the deep learning automatic segmentation, several metrics were used, including the Jaccard Index (intersection over union, IOU), DICE Coefficient (DICE), sensitivity, and Hausdorff distance (consisting of Manhattan, Euclidean, and Cosine distances).
These metrics were utilized to assess the precision and effectiveness of the deep learning image reconstruction algorithm (TrueFidelityTM, GE Healthcare, Milwaukee, WI, USA) compared to traditional CT reconstruction algorithms for segmenting low-density lesions.
The deep learning segmentation model was developed using Python 3.8 and PyTorch 1.9.1. The results were visualized using the Matplotlib 3.6 library (NumFOCUS, Austin, TX, USA). The hardware setup included an Core i7-10700K (for USA) (Intel, Santa Clara, CA, USA) processor and an RTX 2080 Super graphics card (NVIDIA, CA, USA). The specific parameters of this study are as follows.
We developed a convolutional neural network based on the U-Net architecture for CT image segmentation tasks. The model mainly consists of an encoder and a decoder. The encoder part included an adaptation layer and four consecutive down-sampling layers. The adaptation layer is a convolutional module used to adapt to different channel inputs, with input and output channels of 1 and 64 channels, respectively. The down-sampling layers consist of convolutional, max-pooling, and activation modules. The convolutional modules output 64, 128, 256, and 512 channels, respectively, for local feature extraction without changing the feature dimensions. The max-pooling layers have a window size of two to extract the most distinctive features. The activation layer uses the ReLU to provide nonlinearity to the model. Stacking multiple down-sampling layers increases the depth of the model and effectively extracts global features from the CT images for subsequent segmentation tasks.
The decoder consists of four consecutive up-sampling layers. Bilinear interpolation layers were used to decode the global features and restore the original image size. The skip connection mechanism effectively mitigates gradient dispersion and accelerates model convergence; therefore, we symmetrically incorporated convolutional modules identical to those in the down-sampling layers into the up-sampling layers and skip connections of the output features. The final convolutional layer with two channels is used for foreground–background classification at each pixel, providing the segmentation results.
All trainable parameters were randomly initialized in a standard normal distribution and optimized using standard backpropagation with stochastic gradient descent by minimizing the loss of cross-entropy. The loss can be formulated as follows:
L c e = 1 N i L i = 1 N i M c y i c log ( p i c )
where c is the total number of classes (in this study, foreground and background). y i c is the one-hot encoding of the pixel’s actual label. For class i , y i c = 1 indicates that the pixel belongs to class i ; otherwise, y i c = 0, p i c is the probability of the pixel belonging to class i as predicted by the model. The Adadelta gradient descent algorithm (Adam) [47,48] with a batch size of 4 was used to train our model. Integrated L2 regularization (weight decay coefficient 0.0001) in Adam was used to stabilize training.

2.6. Measurement of CT Attenuation and Noise (Standard Deviation, SD)

CT attenuation and image noise of the low-contrast resolution phantom were gauged using ImageJ 1.54 (National Institutes of Health, Bethesda, MD, USA). ImageJ is an open-source image processing and analysis software widely used in scientific research for visualizing, quantifying (such as length, area, and pixel intensity), and segmenting images, with a modular architecture that supports user-defined plugin extensions for multidisciplinary applications [49]. The CT values and standard deviations (SD) were obtained by positioning a circular region of interest (ROI) with an area equal to half of the low-density resolution insert and a layer thickness of 5 mm. Subsequently, a comparably sized circular ROI was placed in the background to measure the CT and SD values. Measurements were taken ten times, and the average values were recorded. A subjective evaluation of medical images was conducted by three physicians who reviewed the images and provided a subjective quality assessment of the VMIs.

3. Results

The segmentation model achieved a reliable training performance, with an accuracy of 99.6% on the training set. The loss curves are shown in Figure 3. The segmentation results with different algorithms (VMls) and photon energy levels are shown in Figure 4.

3.1. Metrics for Deep Learning Automatic Segmentation in Validation Set (5 mGy)

3.1.1. Performance Metrics (IOU, DICE, and Sensitivity)

The specific IOU values as a function of the energy levels for FBP, ASIR-V50%, ASIR-V100%, DLIR-L, DLIR-M, and DLIR-H are as follows: 60 keV: 0.60, 0.67, 0.68, 0.72, 0.75, and 0.75; 68 keV: 0.62, 0.67, 0.70, 0.71, 0.72, and 0.72; 74 keV: 0.62, 0.65, 0.67, 0.69, 0.69, and 0.69, respectively. The specific DICE values for FBP, ASIR-V50%, ASIR-V100%, DLIR-L, DLIR-M, and DLIR-H were as follows: 60 keV: 0.75, 0.80, 0.82, 0.83, 0.85 and 0.86; 68 keV: 0.76, 0.80, 0.80, 0.83, 0.84 and 0.83; 74 keV: 0.75, 0.78, 0.80, 0.81, 0.81 and 0.81, respectively. The specific sensitivity values for FBP, ASIR-V50%, ASIR-V100%, DLIR-L, DLIR-M, and DLIR-H were as follows: 60 keV: 0.93, 0.91, 0.96, 0.95, 0.98, and 0.98; 68 keV: 0.84 0.85, 0.92, 0.93, 0.93, and 0.95; 74 keV: 0.83, 0.88, 0.93, 0.96, 0.97, and 0.98, respectively.
With energy levels (keV), the performance metrics (IOU, DICE, and sensitivity) ranked from the highest to the lowest as a function of the energy level for different reconstruction algorithms were, for the FBP and 100% ASIR-V algorithms, 60 keV, 68 keV, 74 keV, 50 keV, 100 keV, and 40 keV; for DLIR-H, DLIR-M, DLIR-L, and 50% ASIR-V, they were 60 keV, 68 keV, 74 keV, 50 keV, 40 keV, and 100 keV. For reconstruction algorithms, the performance metrics were ranked from the highest to the lowest for different reconstruction algorithms, and over all energy levels were DLIR-H, DLIR-M, DLIR-L, 100% ASIR-V, 50% ASIR-V, and FBP. All the sum performance metrics results are shown in Figure 5a and Figure 6a. Detailed values are listed in the Appendix B (Table A1).

3.1.2. Hausdorff Distance

The specific cosine distance values for FBP, ASIR-V50%, ASIR-V100%, DLIR-L, DLIR-M, and DLIR-H were as follows: 60 keV: 0.52, 0.49, 0.52, 0.51, 0.35, and 0.34; 68 keV: 0.48, 0.49, 0.58, 0.52, 0.43, and 0.42; 74 keV: 0.57, 0.54, 0.57, 0.41, 0.44, and 0.38. The specific Manhattan_distance values for FBP, ASIR-V50%, ASIR-V100%, DLIR-L, DLIR-M, and DLIR-H were as follows: 60 keV: 14.08, 13.17, 12.58, 11.42, 10.92, and 9.08; 68 keV: 13.83, 12.92, 12.50, 11.35, 11.31, and 9.5; 74 keV: 15.08, 13.52, 13.53, 12.08, 11.76, and 11.08. The specific Euclidean_distance values for FBP, ASIR-V50%, ASIR-V100%, DLIR-L, DLIR-M, and DLIR-H were as follows: 60 keV: 3.73, 3.53, 3.51, 3.25, 3.19, and 3.01; 68 keV: 3.68, 3.27, 3.49, 3.42, 3.33, and 3.03; 74 keV: 3.79, 3.64, 3.53, 3.45, 3.40, and 3.31.
With respect to energy levels: The Hausdorff distance ranked from the lowest to the highest as a function of energy level for different reconstruction algorithms were as follows: for FBP, 50% ASIR-V, and 100% ASIR-V algorithms, they were 68 keV, 60 keV, 74 keV, 50 keV, 40 keV, and 100 keV; for DLIR-H, DLIR-M, and DLIR-L, they were 60 keV, 68 keV, 74 keV, 50 keV, 40 keV, and 100 keV. Regarding the reconstruction algorithms, the Hausdorff distance ranked from the lowest to the highest for different reconstruction algorithms over all energy levels were DLIR-H, DLIR-M, DLIR-L, ASIR 100%, 50% ASIR, and FBP. All the Hausdorff results are shown in Figure 5b and Figure 6b. Detailed results are listed in the Appendix B (Table A1).

3.2. CT Attenuation and Standard Deviations (SDs) of the Dual-Energy Spectral CT Image

The CT attenuation value changed with the energy level. At each energy level, the CT attenuation values were statistically the same across all reconstruction algorithms; however, the image noise values (SD) showed statistically significant differences. Table 1 shows the CT attenuation values for the low-resolution phantom and background under different algorithms, and Table 2 details the corresponding image noise levels. Table 2 demonstrates that as the keV increases, the standard deviation (SD) values of the low-density phantom in the images continuously decrease. The DLIR-H algorithm exhibits the lowest standard deviation (SD) value among all the algorithms. The differences observed in these data are statistically significant (p < 0.05). Table 3 shows the D-values between the low-resolution phantom and the background. Table 4 presents the subjective evaluation results of the different monochromatic images (validation sets) with different reconstruction algorithms. Specific charts are displayed in the Appendix C (Figure A1, Figure A2 and Figure A3).
As the energy level (keV) increased, the CT attenuation for the ASIR-V50% low-contrast resolution insert gradually increased with attenuation values of 10.70, 53.59, 80.73, 94.85, 102.03, and 119.28. In contrast, the noise levels decreased, with values of 14.86, 11.26, 8.68, 7.23, 6.52, and 4.78, respectively. Similarly, the CT attenuation of the background also increased, with values of 4.09, 47.42, 73.99, 89.07, 96.41, and 113.78, whereas the noise levels decreased to 16.75, 12.44, 9.27, 7.66, 7.19, and 5.14, respectively. The differences between the low-resolution insert and background across different energy levels (keV) and the reconstruction algorithms were not statistically significant. At each energy level, the DLIR-H provided the strongest ability to reduce image noise.
The specific IOU, DICE, sensitive average value was presented in Appendix B, Table A1. The average value of all the segmentation metrics is also provided.
Finally, the average results for the performance metric, Hausdorff distance, and image quality measurements for the reconstruction algorithms across all energy levels are listed in Table 5.

4. Discussion

This study comprehensively discussed the impact of different energy levels and different reconstruction algorithms in low-dose (5 mGy) DEsCT on the accuracy of automatic segmentation of low-density objects using an objective in DICEs. DEsCT is increasingly being adopted in clinical practice for imaging pediatric patients [22,23,50,51]. Integrating low-dose DEsCT imaging with deep-learning image reconstruction algorithms is a fundamental approach in clinical imaging, allowing the identification of the most effective combinations of reconstruction algorithms and energy levels [52]. Research involving pediatric phantoms has shown that low-dose, high-end CT devices can effectively guide ear surgery [53]. Our findings suggest that by using phantom experiments and objective indicators, deep learning image reconstruction algorithms can significantly reduce image noise and improve image quality in low radiation dose DEsCT for pediatric patients. Children are not simply smaller versions of adults. Different radiation doses and distinct histopathological features require consideration, and phantom model research will further guide pediatric clinical practice [42]. According to the literature, the optimal single energy level is 65 keV for the gray matter–white matter boundary in the brain, which aligns with the soft tissue boundary findings in our study, where 60 keV was found to generate the best segmentation accuracy for low-density, non-enhancing objects [54].
The accurate segmentation of lesions is crucial in clinical practice, particularly in pediatric patients. In the surgical evaluation of hepatoblastoma, precise segmentation techniques are essential to accurately delineate tumor boundaries, enabling a more accurate assessment of the extent of the tumor and its anatomical relationships with adjacent structures. Additionally, the postoperative evaluation of residual liver tissue was significantly improved using these segmentation methods. By accurately measuring the volume and distribution of the remaining liver parenchyma, clinicians can better predict a patient’s recovery trajectory and potential complications, ultimately contributing to more informed surgical planning and improved patient outcomes.
Our comprehensive evaluation highlights the effectiveness of integrating low-dose dual-energy spectral CT imaging with deep learning image reconstruction algorithms. The results demonstrated significant improvements in segmentation accuracy and image quality, particularly with the DLIR-M and DLIR-H algorithms, across all energy levels. These findings underscore the potential of advanced techniques to enhance clinical imaging and diagnostic precision.
We averaged the values across all six energy levels for the different reconstruction algorithms to evaluate their overall performance. The results indicated that DLIR-M and DLIR-H had similar values in IOU and DICE, which focused on internal filling and clinical indicators of the phantom and were among the best among all the reconstruction algorithms. We believe this is because of their ability to provide low image noise while maintaining image textures to achieve the most balanced segmentation and to correctly identify the majority of the segmented regions. In contrast, DLIR-H performed better in terms of sensitivity and the Hausdorff distance metric, which emphasizes the number and distance of misclassified points in the segmentation results, indicating that it had fewer segmentation errors.
DLIR-H provided images with significantly less noise than did DLIR-M. However, according to earlier research by Greffier, J, while DLIR-M noise reduction is not as effective as DLIR-H, the noise power spectrum (NPS) and image texture of DLIR-M are more similar to those of FBP [55]. In the field of image segmentation, the effectiveness of segmentation is closely related to both the image noise and texture. This also explains why, despite the much lower noise levels in ASIR-V100%, its segmentation performance was inferior to that of the DLIR-L series.
Low-energy images in dEsCT, such as 40 keV images, are most effective in imaging objects enhanced with iodine because of the closer proximity in energy value of 40 keV photons to the k-edge of iodine. However, in our study, a non-enhancing object simulating low-density lesions was used in clinical applications. From the measured CT attenuation, the poor segmentation accuracy at 40 keV can be attributed to the small differences between the low-contrast object and background. In our study, the contrast of the low-density object was not improved by using the 40 keV photons, whereas the image noise was significantly increased compared with that of the 60 keV images. In addition, the CT values of the object itself at 40 keV were significantly different from those of the 68 keV images on which the segmentation model was trained. All of these factors make lesion segmentation challenging. Similarly, at 100 keV, the significant changes in the image texture (image noise and CT value) likely contributed to the reduced segmentation accuracy. The potential value of this study lies in its ability to guide clinical practice. For example, the application of DLIR-H can enhance segmentation accuracy when targeting liver and other soft tissue lesions with varying densities. Additionally, the use of low-dose energy-spectrum CT, particularly at 60 keV, is more effective for pediatric cases. Moreover, body-modeling studies can help avoid ethical concerns. We have initiated trials of the model with clinical data, and we plan to report the results of this phase in future studies.
This study had several limitations. First, the data were collected from a single center using a specific low-resolution phantom, which did not represent all clinical scenarios, particularly contrast-enhanced lesions. It is necessary to expand the dataset to validate the robustness of the model further. Additionally, a specific ACR464 phantom was used in this study only to assess the success rate of automatic segmentation under the same CNR conditions without considering the extent to which the radiation dose and contrast agent usage could be reduced. Based on previous research, 50 keV has been shown to reduce both the radiation dose and the contrast agent usage more effectively. Finally, this study focused only on non-enhancing, low-density objects, without evaluating segmentation performance in the context of contrast-enhanced CT in actual clinical settings. Further validation is required, either with contrast phantoms containing different iodine concentrations or through clinical experiments.

5. Conclusions

In conclusion, a comprehensive analysis suggests that the combination of 60 keV and deep learning image reconstruction algorithms with medium or high strengths (DLIR-M and DLIR-H) delivers the best results for segmenting low-density objects, whereas DLIR-H provides images with the highest sensitivity and lowest image noise across all energy levels in dEsCT. In future medical applications, we will further fine-tune model parameters and test the applicability of applying DLIR-H in combination with 60 keV to improve the segmentation accuracy.

Author Contributions

Conceptualization, J.S. and H.L.; methodology, Z.C.; software, H.L. and Z.C.; validation, H.L., Z.Y. and S.G.; formal analysis, J.H. and Z.Y.; investigation, H.L. and J.H.; resources, J.S.; data curation, S.G.; writing—original draft preparation, H.L.; writing—review and editing, J.S. and H.L.; visualization, H.L.; supervision, J.S. and Y.P.; project administration, Y.P.; funding acquisition, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Xinjiang Uygur Autonomous Region, grant number: 2022D01A306 (receiver: Jihang Sun), and State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai 201210, China, grant number: YGSKL-SHTech-2025-KF01 (receiver: Yun Peng), and Xinjiang Hospital of Beijing Children’s Hospital science program: 2022005 (receiver: Jihang Sun).

Institutional Review Board Statement

Not applicable. Ethical review and approval were waived for this study due to phantom research, which did not involve humans or animals.

Informed Consent Statement

Not applicable. Patient consent was waived due to phantom research. These studies do not involve humans.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We express our gratitude to Wenhuan Li, for her assistance in enhancing the comprehension of dual-energy techniques.

Conflicts of Interest

The authors have no conflicts of interest to disclose.

Abbreviations

The following abbreviations are used in this manuscript:
ASIR-VAdaptive statistical iterative reconstruction-veo
CTDIvolCT dose index
DEsCTDual-energy spectral computed tomography
DICEDICE coefficient
DLIRDeep learning image reconstruction
FBPFiltered back-projection
HUHounsfield Unit
IOUIntersection over union
IRIterative reconstruction
keVKilo-electron volts
NPSNoise power spectrum
ROIRegion of interest
SDStandard deviation
U-NetConvolutional Networks for Biomedical Image Segmentation
VMIsVirtual monochromatic images

Appendix A. Evaluation Index of the Accuracy of Deep Learning Segmentation Model

Jaccard Index (intersection over union, IOU):
The Jaccard Index is a statistical measure that compares the similarity and diversity between two sets. It is defined as the ratio of the size of the intersection of the two sets to the size of their union. In this study, the Jaccard Index is primarily used to evaluate how closely the segmentation results from the U-Net model align with the gold standard delineated by physicians. An IOU value closer to 1 indicates a higher degree of similarity. The formula for IOU is as follows:
I O U = A B A B = T P F P + T P + F N
The DICE coefficient is a metric used to measure the overlap between two sets, similar to the Jaccard Index but giving more weight to the size of the overlapping area. It is particularly useful in segmentation tasks where precise overlap is crucial. The formula for the DICE coefficient is as follows:
D i c e = 2 A B A + B = 2 T P F P + 2 T P + F N
Sensitivity, also known as recall, measures the proportion of actual positives that the model correctly identifies. The formula for sensitivity is as follows:
S e n s i t i v i t y = T P T P + F N
PPV, also known as precision, measures the proportion of positive results that are true positives. The formula for PPV is as follows:
P P V = T P T P + F P
In these metrics: True Positive (TP): The model predicts a positive sample, and the actual result is also a positive sample. False Positive (FP): The model predicts a positive sample, but the actual result is a negative sample. True Negative (TN): The model predicts a negative sample, and the actual result is also a negative sample. False Negative (FN): The model predicts a negative sample, but the actual result is a positive sample.
The Hausdorff distance measures the degree of resemblance between two sets of points. It calculates the maximum distance from a point in one set to the nearest point in the other set. Variants include the following:
d H X , Y = m a x max xϵX min yϵY d x , y , max yϵY min xϵX d x , y
Manhattan distance: The sum of the absolute differences between the Cartesian coordinates of two points.
Euclidean distance: The straight-line distance between two points in Euclidean space.
Cosine distance: Measures the cosine of the angle between two vectors.

Appendix B

Table A1. The average value of the metrics related to the segmentation results (All).
Table A1. The average value of the metrics related to the segmentation results (All).
IOUDICESensitivityManhattan_
Distance
Euclidean_
Distance
Cosine_
Distance
40 keVFBP0.480.630.8824.174.840.77
50% ASIR0.570.720.9320.254.420.68
100% ASIR0.590.740.9214.923.820.55
DLIR-L0.630.770.9713.753.690.51
DLIR-M0.610.770.9912.673.530.49
DLIR-H0.630.760.9912.833.460.39
50 keVFBP0.590.740.8516.924.070.61
50% ASIR0.620.760.8515.253.850.56
100% ASIR0.650.780.9113.753.700.42
DLIR-L0.690.810.9313.423.640.46
DLIR-M0.70.830.9412.253.490.42
DLIR-H0.720.820.9511.923.420.43
60 keVFBP0.600.750.9314.083.730.52
50% ASIR0.670.800.9113.173.530.49
100% ASIR0.680.820.9612.583.510.52
DLIR-L0.720.830.9511.423.250.51
DLIR-M0.750.850.9810.923.190.35
DLIR-H0.750.860.989.083.010.34
68 keVFBP0.620.760.8413.833.680.48
50% ASIR0.670.800.8512.923.270.49
100% ASIR0.700.800.9212.53.490.58
DLIR-L0.710.830.9311.353.420.52
DLIR-M0.720.840.9311.313.330.43
DLIR-H0.720.830.959.503.030.42
74 keVFBP0.620.750.8315.083.790.57
50% ASIR0.650.780.8813.523.640.54
100% ASIR0.670.800.9313.533.530.57
DLIR-L0.690.810.9612.083.450.41
DLIR-M0.690.810.9711.763.400.44
DLIR-H0.690.810.9811.083.310.38
100 keVFBP0.550.700.8427.755.420.82
50% ASIR0.560.710.9527.525.300.81
100% ASIR0.610.760.9626.675.060.78
DLIR-L0.600.750.9821.584.540.76
DLIR-M0.620.760.9719.674.340.73
DLIR-H0.630.770.9718.424.280.69

Appendix C

Figure A1. CT attenuation of the low-resolution phantom and background under different algorithms.
Figure A1. CT attenuation of the low-resolution phantom and background under different algorithms.
Tomography 11 00051 g0a1
Figure A2. Image noise of the low-resolution phantom and background under different algorithms.
Figure A2. Image noise of the low-resolution phantom and background under different algorithms.
Tomography 11 00051 g0a2
Figure A3. The difference between the low-resolution phantom and the background. Note: Low-res, low-resolution insert. Back-ground, back-ground of the insert.
Figure A3. The difference between the low-resolution phantom and the background. Note: Low-res, low-resolution insert. Back-ground, back-ground of the insert.
Tomography 11 00051 g0a3

References

  1. Lin, H.; Xiao, H.; Dong, L.; Teo, K.B.; Zou, W.; Cai, J.; Li, T. Deep learning for automatic target volume segmentation in radiation therapy: A review. Quant. Imaging Med. Surg. 2021, 11, 4847–4858. [Google Scholar] [CrossRef]
  2. Lastrucci, A.; Wandael, Y.; Ricci, R.; Maccioni, G.; Giansanti, D. The Integration of Deep Learning in Radiotherapy: Exploring Challenges, Opportunities, and Future Directions through an Umbrella Review. Diagnostics 2024, 14, 939. [Google Scholar] [CrossRef]
  3. Wang, J.; Zhang, X.; Lv, P.; Wang, H.; Cheng, Y. Automatic Liver Segmentation Using EfficientNet and Attention-Based Residual U-Net in CT. J. Digit. Imaging 2022, 35, 1479–1493. [Google Scholar] [CrossRef] [PubMed]
  4. Walter, A.; Hoegen-Saßmannshausen, P.; Stanic, G.; Rodrigues, J.P.; Adeberg, S.; Jäkel, O.; Frank, M.; Giske, K. Segmentation of 71 Anatomical Structures Necessary for the Evaluation of Guideline-Conforming Clinical Target Volumes in Head and Neck Cancers. Cancers 2024, 16, 415. [Google Scholar] [CrossRef]
  5. Li, W.; Sun, Y.; Zhang, G.; Yang, Q.; Wang, B.; Ma, X.; Zhang, H. Automated segmentation and volume prediction in pediatric Wilms’ tumor CT using nnu-net. BMC Pediatr. 2024, 24, 321. [Google Scholar] [CrossRef] [PubMed]
  6. de Margerie-Mellon, C.; Chassagnon, G. Artificial intelligence: A critical review of applications for lung nodule and lung cancer. Diagn. Interv. Imaging 2023, 104, 11–17. [Google Scholar] [CrossRef] [PubMed]
  7. Lin, Z.; Cui, Y.; Liu, J.; Sun, Z.; Ma, S.; Zhang, X.; Wang, X. Automated segmentation of kidney and renal mass and automated detection of renal mass in CT urography using 3D U-Net-based deep convolutional neural network. Eur. Radiol. 2021, 31, 5021–5031. [Google Scholar] [CrossRef]
  8. Hild, O.; Berriet, P.; Nallet, J.; Salvi, L.; Lenoir, M.; Henriet, J.; Thiran, J.P.; Auber, F.; Chaussy, Y. Automation of Wilms’ tumor segmentation by artificial intelligence. Cancer Imaging Off. Publ. Int. Cancer Imaging Soc. 2024, 24, 83. [Google Scholar] [CrossRef]
  9. Bousse, A.; Kandarpa, V.S.S.; Rit, S.; Perelli, A.; Li, M.; Wang, G.; Zhou, J.; Wang, G. Systematic Review on Learning-based Spectral CT. IEEE Trans. Radiat. Plasma Med. Sci. 2024, 8, 113–137. [Google Scholar] [CrossRef]
  10. Dabli, D.; Loisy, M.; Frandon, J.; de Oliveira, F.; Meerun, A.M.; Guiu, B.; Beregi, J.P.; Greffier, J. Comparison of image quality of two versions of deep-learning image reconstruction algorithm on a rapid kV-switching CT: A phantom study. Eur. Radiol. Exp. 2023, 7, 1. [Google Scholar] [CrossRef]
  11. Shapira, N.; Mei, K.; Noël, P.B. Spectral CT quantification stability and accuracy for pediatric patients: A phantom study. J. Appl. Clin. Med. Phys. 2021, 22, 16–26. [Google Scholar] [CrossRef] [PubMed]
  12. Fernández-Pérez, G.C.; Fraga Piñeiro, C.; Oñate Miranda, M.; Díez Blanco, M.; Mato Chaín, J.; Collazos Martínez, M.A. Dual-energy CT: Technical considerations and clinical applications. Radiologia 2022, 64, 445–455. [Google Scholar] [CrossRef] [PubMed]
  13. McCollough, C.H.; Leng, S.; Yu, L.; Fletcher, J.G. Dual- and Multi-Energy CT: Principles, Technical Approaches, and Clinical Applications. Radiology 2015, 276, 637–653. [Google Scholar] [CrossRef]
  14. Koetzier, L.R.; Mastrodicasa, D.; Szczykutowicz, T.P.; van der Werf, N.R.; Wang, A.S.; Sandfort, V.; van der Molen, A.J.; Fleischmann, D.; Willemink, M.J. Deep Learning Image Reconstruction for CT: Technical Principles and Clinical Prospects. Radiology 2023, 306, e221257. [Google Scholar] [CrossRef]
  15. Ghasemi Shayan, R.; Oladghaffari, M.; Sajjadian, F.; Fazel Ghaziyani, M. Image Quality and Dose Comparison of Single-Energy CT (SECT) and Dual-Energy CT (DECT). Radiol. Res. Pract. 2020, 2020, 1403957. [Google Scholar] [CrossRef] [PubMed]
  16. Nagayama, Y.; Sakabe, D.; Goto, M.; Emoto, T.; Oda, S.; Nakaura, T.; Kidoh, M.; Uetani, H.; Funama, Y.; Hirai, T. Deep Learning-based Reconstruction for Lower-Dose Pediatric CT: Technical Principles, Image Characteristics, and Clinical Implementations. Radiogr. A Rev. Publ. Radiol. Soc. N. Am. Inc. 2021, 41, 1936–1953. [Google Scholar] [CrossRef]
  17. Clark, D.P.; Schwartz, F.R.; Marin, D.; Ramirez-Giraldo, J.C.; Badea, C.T. Deep learning based spectral extrapolation for dual-source, dual-energy X-ray computed tomography. Med. Phys. 2020, 47, 4150–4163. [Google Scholar] [CrossRef]
  18. Schwartz, F.R.; Clark, D.P.; Ding, Y.; Ramirez-Giraldo, J.C.; Badea, C.T.; Marin, D. Evaluating renal lesions using deep-learning based extension of dual-energy FoV in dual-source CT-A retrospective pilot study. Eur. J. Radiol. 2021, 139, 109734. [Google Scholar] [CrossRef]
  19. Hlouschek, J.; König, B.; Bos, D.; Santiago, A.; Zensen, S.; Haubold, J.; Pöttgen, C.; Herz, A.; Opitz, M.; Wetter, A.; et al. Experimental Examination of Conventional, Semi-Automatic, and Automatic Volumetry Tools for Segmentation of Pulmonary Nodules in a Phantom Study. Diagnostics 2023, 14, 28. [Google Scholar] [CrossRef]
  20. Hardie, R.C.; Trout, A.T.; Dillman, J.R.; Narayanan, B.N.; Tanimoto, A.A. Performance of Lung-Nodule Computer-Aided Detection Systems on Standard-Dose and Low-Dose Pediatric CT Scans: An Intraindividual Comparison. AJR Am. J. Roentgenol. 2024, 223, e2431972. [Google Scholar] [CrossRef]
  21. Groendahl, A.R.; Huynh, B.N.; Tomic, O.; Søvik, Å.; Dale, E.; Malinen, E.; Skogmo, H.K.; Futsaether, C.M. Automatic gross tumor segmentation of canine head and neck cancer using deep learning and cross-species transfer learning. Front. Vet. Sci. 2023, 10, 1143986. [Google Scholar] [CrossRef] [PubMed]
  22. Siegel, M.J.; Ramirez-Giraldo, J.C. Dual-Energy CT in Children: Imaging Algorithms and Clinical Applications. Radiology 2019, 291, 286–297. [Google Scholar] [CrossRef] [PubMed]
  23. Siegel, M.J.; Bhalla, S.; Cullinane, M. Dual-Energy CT Material Decomposition in Pediatric Thoracic Oncology. Radiology. Imaging Cancer 2021, 3, e200097. [Google Scholar] [CrossRef]
  24. Kamps, S.E.; Otjen, J.P.; Stanescu, A.L.; Mileto, A.; Lee, E.Y.; Phillips, G.S. Dual-Energy CT of Pediatric Abdominal Oncology Imaging: Private Tour of New Applications of CT Technology. AJR Am. J. Roentgenol. 2020, 214, 967–975. [Google Scholar] [CrossRef]
  25. Gallo-Bernal, S.; Peña-Trujillo, V.; Gee, M.S. Dual-energy computed tomography: Pediatric considerations. Pediatr. Radiol. 2024, 54, 2112–2126. [Google Scholar] [CrossRef] [PubMed]
  26. Tabari, A.; Gee, M.S.; Singh, R.; Lim, R.; Nimkin, K.; Primak, A.; Schmidt, B.; Kalra, M.K. Reducing Radiation Dose and Contrast Medium Volume with Application of Dual-Energy CT in Children and Young Adults. AJR Am. J. Roentgenol. 2020, 214, 1199–1205. [Google Scholar] [CrossRef]
  27. Yang, L.; Sun, J.; Li, J.; Peng, Y. Dual-energy spectral CT imaging of pulmonary embolism with Mycoplasma pneumoniae pneumonia in children. Radiol. Med. 2022, 127, 154–161. [Google Scholar] [CrossRef]
  28. Sun, J.; Li, H.; Yu, T.; Huo, A.; Hua, S.; Zhou, Z.; Peng, Y. Application of metal artifact reduction algorithm in reducing metal artifacts in post-surgery pediatric low radiation dose spine computed tomography (CT) images. Quant. Imaging Med. Surg. 2024, 14, 4648–4658. [Google Scholar] [CrossRef]
  29. Xie, M.; Wang, H.; Tang, S.; Chen, M.; Li, T.; He, L. Application of dual-energy CT with prospective ECG-gating in cardiac CT angiography for children: Radiation and contrast agent dose. Eur. J. Radiol. 2024, 170, 111229. [Google Scholar] [CrossRef]
  30. Schicchi, N.; Fogante, M.; Esposto Pirani, P.; Agliata, G.; Basile, M.C.; Oliva, M.; Agostini, A.; Giovagnoni, A. Third-generation dual-source dual-energy CT in pediatric congenital heart disease patients: State-of-the-art. Radiol. Med. 2019, 124, 1238–1252. [Google Scholar] [CrossRef]
  31. Hudobivnik, N.; Schwarz, F.; Johnson, T.; Agolli, L.; Dedes, G.; Tessonnier, T.; Verhaegen, F.; Thieke, C.; Belka, C.; Sommer, W.H.; et al. Comparison of proton therapy treatment planning for head tumors with a pencil beam algorithm on dual and single energy CT images. Med. Phys. 2016, 43, 495. [Google Scholar] [CrossRef] [PubMed]
  32. Bruns, S.; Wolterink, J.M.; Takx, R.A.P.; van Hamersvelt, R.W.; Suchá, D.; Viergever, M.A.; Leiner, T.; Išgum, I. Deep learning from dual-energy information for whole-heart segmentation in dual-energy and single-energy non-contrast-enhanced cardiac CT. Med. Phys. 2020, 47, 5048–5060. [Google Scholar] [CrossRef] [PubMed]
  33. Miller, C.; Mittelstaedt, D.; Black, N.; Klahr, P.; Nejad-Davarani, S.; Schulz, H.; Goshen, L.; Han, X.; Ghanem, A.I.; Morris, E.D.; et al. Impact of CT reconstruction algorithm on auto-segmentation performance. J. Appl. Clin. Med. Phys. 2019, 20, 95–103. [Google Scholar] [CrossRef] [PubMed]
  34. Siegel, M.J.; Mhlanga, J.C.; Salter, A.; Ramirez-Giraldo, J.C. Comparison of radiation dose and image quality between contrast-enhanced single- and dual-energy abdominopelvic computed tomography in children as a function of patient size. Pediatr. Radiol. 2021, 51, 2000–2008. [Google Scholar] [CrossRef]
  35. Lee, S.; Choi, Y.H.; Cho, Y.J.; Lee, S.B.; Cheon, J.E.; Kim, W.S.; Ahn, C.K.; Kim, J.H. Noise reduction approach in pediatric abdominal CT combining deep learning and dual-energy technique. Eur. Radiol. 2021, 31, 2218–2226. [Google Scholar] [CrossRef]
  36. Somasundaram, E.; Taylor, Z.; Alves, V.V.; Qiu, L.; Fortson, B.L.; Mahalingam, N.; Dudley, J.A.; Li, H.; Brady, S.L.; Trout, A.T.; et al. Deep Learning Models for Abdominal CT Organ Segmentation in Children: Development and Validation in Internal and Heterogeneous Public Datasets. AJR Am. J. Roentgenol. 2024, 223, e2430931. [Google Scholar] [CrossRef] [PubMed]
  37. Bachanek, S.; Wuerzberg, P.; Biggemann, L.; Janssen, T.Y.; Nietert, M.; Lotz, J.; Zeuschner, P.; Maßmann, A.; Uhlig, A.; Uhlig, J. Renal tumor segmentation, visualization, and segmentation confidence using ensembles of neural networks in patients undergoing surgical resection. Eur. Radiol. 2024, 35, 2147–2156. [Google Scholar] [CrossRef]
  38. Liu, S.; Liang, S.; Huang, X.; Yuan, X.; Zhong, T.; Zhang, Y. Graph-enhanced U-Net for semi-supervised segmentation of pancreas from abdomen CT scan. Phys. Med. Biol. 2022, 67, 155017. [Google Scholar] [CrossRef]
  39. Delmoral, J.C.; JM, R.S.T. Semantic Segmentation of CT Liver Structures: A Systematic Review of Recent Trends and Bibliometric Analysis: Neural Network-based Methods for Liver Semantic Segmentation. J. Med. Syst. 2024, 48, 97. [Google Scholar] [CrossRef]
  40. Nadeem, S.A.; Hoffman, E.A.; Sieren, J.C.; Comellas, A.P.; Bhatt, S.P.; Barjaktarevic, I.Z.; Abtin, F.; Saha, P.K. A CT-Based Automated Algorithm for Airway Segmentation Using Freeze-and-Grow Propagation and Deep Learning. IEEE Trans. Med. Imaging 2021, 40, 405–418. [Google Scholar] [CrossRef]
  41. Ntoufas, N.; Raissaki, M.; Damilakis, J.; Perisinakis, K. Comparison of radiation exposure from dual- and single-energy CT imaging protocols resulting in equivalent contrast-to-noise ratio of lesions for adults and children: A phantom study. Eur. Radiol. 2024. online ahead of print. [Google Scholar] [CrossRef] [PubMed]
  42. Chatterjee, D.; Kanhere, A.; Doo, F.X.; Zhao, J.; Chan, A.; Welsh, A.; Kulkarni, P.; Trang, A.; Parekh, V.S.; Yi, P.H. Children Are Not Small Adults: Addressing Limited Generalizability of an Adult Deep Learning CT Organ Segmentation Model to the Pediatric Population. J. Imaging Inform. Med. 2024. online ahead of print. [Google Scholar] [CrossRef] [PubMed]
  43. Holz, J.A.; Alkadhi, H.; Laukamp, K.R.; Lennartz, S.; Heneweer, C.; Püsken, M.; Persigehl, T.; Maintz, D.; Große Hokamp, N. Quantitative accuracy of virtual non-contrast images derived from spectral detector computed tomography: An abdominal phantom study. Sci. Rep. 2020, 10, 21575. [Google Scholar] [CrossRef] [PubMed]
  44. Ikeda, R.; Kadoya, N.; Nakajima, Y.; Ishii, S.; Shibu, T.; Jingu, K. Impact of CT scan parameters on deformable image registration accuracy using deformable thorax phantom. J. Appl. Clin. Med. Phys. 2023, 24, e13917. [Google Scholar] [CrossRef]
  45. Jiang, B.; Li, N.; Shi, X.; Zhang, S.; Li, J.; de Bock, G.H.; Vliegenthart, R.; Xie, X. Deep Learning Reconstruction Shows Better Lung Nodule Detection for Ultra-Low-Dose Chest CT. Radiology 2022, 303, 202–212. [Google Scholar] [CrossRef]
  46. Greffier, J.; Villani, N.; Defez, D.; Dabli, D.; Si-Mohamed, S. Spectral CT imaging: Technical principles of dual-energy CT and multi-energy photon-counting CT. Diagn. Interv. Imaging 2023, 104, 167–177. [Google Scholar] [CrossRef]
  47. Zeiler, M.D.J.A. ADADELTA: An Adaptive Learning Rate Method. arXiv 2012, arXiv:1212.5701. [Google Scholar]
  48. Kingma, D.P.; Ba, J.J.C. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  49. Schneider, C.A.; Rasband, W.S.; Eliceiri, K.W. NIH Image to ImageJ: 25 years of image analysis. Nat. Methods 2012, 9, 671–675. [Google Scholar] [CrossRef]
  50. Rajiah, P.; Parakh, A.; Kay, F.; Baruah, D.; Kambadakone, A.R.; Leng, S. Update on Multienergy CT: Physics, Principles, and Applications. Radiogr. A Rev. Publ. Radiol. Soc. N. Am. Inc. 2020, 40, 1284–1308. [Google Scholar] [CrossRef]
  51. Rapp, J.B.; Biko, D.M.; Siegel, M.J. Dual-Energy CT for Pediatric Thoracic Imaging: A Review. AJR Am. J. Roentgenol. 2023, 221, 526–538. [Google Scholar] [CrossRef] [PubMed]
  52. Li, H.; Li, Z.; Gao, S.; Hu, J.; Yang, Z.; Peng, Y.; Sun, J. Performance evaluation of deep learning image reconstruction algorithm for dual-energy spectral CT imaging: A phantom study. J. Xray Sci. Technol. 2024, 32, 513–528. [Google Scholar] [CrossRef]
  53. Steiniger, B.; Lechel, U.; Reichenbach, J.R.; Fiebich, M.; Aschenbach, R.; Schegerer, A.; Waginger, M.; Bobeva, A.; Teichgräber, U.; Mentzel, H.J. In vitro measurements of radiation exposure with different modalities (computed tomography, cone beam computed tomography) for imaging the petrous bone with a pediatric anthropomorphic phantom. Pediatr. Radiol. 2022, 52, 1125–1133. [Google Scholar] [CrossRef] [PubMed]
  54. McCollough, C.H.; Boedeker, K.; Cody, D.; Duan, X.; Flohr, T.; Halliburton, S.S.; Hsieh, J.; Layman, R.R.; Pelc, N.J. Principles and applications of multienergy CT: Report of AAPM Task Group 291. Med. Phys. 2020, 47, e881–e912. [Google Scholar] [CrossRef]
  55. Greffier, J.; Hamard, A.; Pereira, F.; Barrau, C.; Pasquier, H.; Beregi, J.P.; Frandon, J. Image quality and dose reduction opportunity of deep learning image reconstruction algorithm for CT: A phantom study. Eur. Radiol. 2020, 30, 3951–3959. [Google Scholar] [CrossRef]
Figure 1. (a) The scan image of phantom, (b) The gold standard region of interest (ROI) of image and mask for low-density object, respectively.
Figure 1. (a) The scan image of phantom, (b) The gold standard region of interest (ROI) of image and mask for low-density object, respectively.
Tomography 11 00051 g001
Figure 2. (a) The pipeline of our performance evaluation method. (b) The distribution of the dataset. Note: (b) shows the Kernel Density Estimation (KDE) curves of the training set and test set. The left figure represents the distribution of CT value in the ROI region, the middle figure represents the distribution of CT value in the mask region, and the right figure represents the pixel number in the ROI region.
Figure 2. (a) The pipeline of our performance evaluation method. (b) The distribution of the dataset. Note: (b) shows the Kernel Density Estimation (KDE) curves of the training set and test set. The left figure represents the distribution of CT value in the ROI region, the middle figure represents the distribution of CT value in the mask region, and the right figure represents the pixel number in the ROI region.
Tomography 11 00051 g002
Figure 3. Training loss and Val loss.
Figure 3. Training loss and Val loss.
Tomography 11 00051 g003
Figure 4. Segmentation results of low contrast resolution (1 case). Note: The areas indicated by the red arrows indicate that the phantom processed using the FBP algorithm exhibited numerous erroneous segmentation points in the lower-left corner. After processing with the DLIR-H algorithm, segmentation errors were eliminated. The positions indicated by the green arrows demonstrate that the ROI circle of the phantom was highly irregular when the 50% ASIR algorithm was used. In contrast, after processing using the DLIR-H algorithm, the ROI circle became more regular. The displayed segmentation results of the phantom reflect the actual segmentation performance more intuitively.
Figure 4. Segmentation results of low contrast resolution (1 case). Note: The areas indicated by the red arrows indicate that the phantom processed using the FBP algorithm exhibited numerous erroneous segmentation points in the lower-left corner. After processing with the DLIR-H algorithm, segmentation errors were eliminated. The positions indicated by the green arrows demonstrate that the ROI circle of the phantom was highly irregular when the 50% ASIR algorithm was used. In contrast, after processing using the DLIR-H algorithm, the ROI circle became more regular. The displayed segmentation results of the phantom reflect the actual segmentation performance more intuitively.
Tomography 11 00051 g004
Figure 5. (a) The sum of IOU, DICE, and sensitivity in different algorithms at different energy levels. (b) The sum of the Hausdorff distance in different algorithms at different energy levels.
Figure 5. (a) The sum of IOU, DICE, and sensitivity in different algorithms at different energy levels. (b) The sum of the Hausdorff distance in different algorithms at different energy levels.
Tomography 11 00051 g005
Figure 6. (a) The sum of IOU, DICE, and sensitivity in different algorithms at different energy levels. (b) The sum of the Hausdorff distance in different algorithms at different energy levels. Note: The DLIR-H reconstruction algorithm can significantly improve the segmentation accuracy at 50 keV, which is superior to all other Asir-V algorithms and keV, making it possible to further reduce the amount of contrast agent.
Figure 6. (a) The sum of IOU, DICE, and sensitivity in different algorithms at different energy levels. (b) The sum of the Hausdorff distance in different algorithms at different energy levels. Note: The DLIR-H reconstruction algorithm can significantly improve the segmentation accuracy at 50 keV, which is superior to all other Asir-V algorithms and keV, making it possible to further reduce the amount of contrast agent.
Tomography 11 00051 g006
Table 1. CT attenuation of the low-resolution phantom and background under different algorithms.
Table 1. CT attenuation of the low-resolution phantom and background under different algorithms.
ReconFBP50% ASIR100% ASIRDLIR-LDLIR-MDLIR-H
Low-ResBack-GroundLow-ResBack-GroundLow-ResBack-GroundLow-ResBack-GroundLow-ResBack-GroundLow-ResBack-Ground
40 keV11.394.7210.704.0910.685.3012.986.1712.475.6810.084.75
50 keV54.1848.1053.5947.4254.1047.6855.2049.6455.3149.2355.6548.68
60 keV80.1074.2980.7373.9983.3776.9381.9075.6182.1876.0282.6775.69
68 keV94.1088.7894.8589.0795.2788.8196.3990.4795.9790.0396.1390.35
74 keV102.2896.32102.0396.41101.6896.38103.6397.46103.7597.26103.5997.33
100 keV119.80113.25119.28113.78120.24114.87121.06115.47121.74115.90120.96114.99
Table 2. Standard deviations (SDs) of the low-resolution phantom and background under different algorithms.
Table 2. Standard deviations (SDs) of the low-resolution phantom and background under different algorithms.
ReconFBP50% ASIR100% ASIRDLIR-LDLIR-MDLIR-H
Low-ResBack-GroundLow-ResBack-GroundLow-ResBack-GroundLow-ResBack-GroundLow-ResBack-GroundLow-ResBack-Ground
40 keV21.5123.5114.8616.759.6812.3214.7016.8712.3415.0710.0612.04
50 keV16.1218.2011.2612.447.948.9810.9512.679.5010.537.848.84
60 keV12.1813.918.6759.277.006.418.499.677.348.125.616.64
68 keV10.3111.757.237.665.967.677.328.176.006.534.895.66
74 keV9.5910.656.527.194.485.046.317.145.436.184.604.95
100 keV6.517.534.785.143.223.554.603.553.834.353.193.45
Table 3. Difference (D-value) of the low-resolution phantom and background.
Table 3. Difference (D-value) of the low-resolution phantom and background.
ReconFBP50% ASIR100% ASIRDLIR-LDLIR-MDLIR-H
40 keV6.676.615.386.816.795.33
50 keV6.096.176.425.566.086.97
60 keV5.816.746.456.296.176.99
68 keV5.315.786.475.925.945.78
74 keV5.965.635.306.166.486.26
100 keV6.555.505.375.595.845.97
Table 4. Subjective evaluation results of images in the validation group at different energy levels and with different reconstruction algorithms.
Table 4. Subjective evaluation results of images in the validation group at different energy levels and with different reconstruction algorithms.
ReconFBP50% ASIR100% ASIRDLIR-LDLIR-MDLIR-H
40 keV1.002.001.672.002.002.33
50 keV1.502.502.333.003.333.33
60 keV2.003.002.673.003.333.33
68 keV2.003.002.673.003.333.33
74 keV2.003.003.003.003.333.67
100 keV1.672.002.002.332.673.00
Table 5. Average values (over six energy levels) of metrics related to the segmentation results (ALL 40–100 keV).
Table 5. Average values (over six energy levels) of metrics related to the segmentation results (ALL 40–100 keV).
ReconIOUDICESensitivityManhattan DistanceEuclidean DistanceCosine DistanceBKG NoiseQuality Score
FBP0.580.720.8618.644.260.6314.261.70
50% ASIR0.620.760.9017.114.000.609.742.58
100% ASIR0.650.780.9315.663.850.577.332.39
DLIR-L0.670.800.9513.933.670.5310.012.72
DLIR-M0.680.810.9613.103.550.488.463.00
DLIR-H0.690.810.9712.143.420.446.933.17
Note: IOU, intersection over union. DICE, DICE similarity coefficient. Hausdorff distance includes Manhattan_distance, Euclidean_distance, and Cosine_distance.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, H.; Chen, Z.; Gao, S.; Hu, J.; Yang, Z.; Peng, Y.; Sun, J. Performance Evaluation of Image Segmentation Using Dual-Energy Spectral CT Images with Deep Learning Image Reconstruction: A Phantom Study. Tomography 2025, 11, 51. https://doi.org/10.3390/tomography11050051

AMA Style

Li H, Chen Z, Gao S, Hu J, Yang Z, Peng Y, Sun J. Performance Evaluation of Image Segmentation Using Dual-Energy Spectral CT Images with Deep Learning Image Reconstruction: A Phantom Study. Tomography. 2025; 11(5):51. https://doi.org/10.3390/tomography11050051

Chicago/Turabian Style

Li, Haoyan, Zhenpeng Chen, Shuaiyi Gao, Jiaqi Hu, Zhihao Yang, Yun Peng, and Jihang Sun. 2025. "Performance Evaluation of Image Segmentation Using Dual-Energy Spectral CT Images with Deep Learning Image Reconstruction: A Phantom Study" Tomography 11, no. 5: 51. https://doi.org/10.3390/tomography11050051

APA Style

Li, H., Chen, Z., Gao, S., Hu, J., Yang, Z., Peng, Y., & Sun, J. (2025). Performance Evaluation of Image Segmentation Using Dual-Energy Spectral CT Images with Deep Learning Image Reconstruction: A Phantom Study. Tomography, 11(5), 51. https://doi.org/10.3390/tomography11050051

Article Metrics

Back to TopTop