You are currently viewing a new version of our website. To view the old version click .
Electronics
  • Article
  • Open Access

11 December 2020

Medical Prior Knowledge Guided Automatic Detection of Coronary Arteries Calcified Plaque with Cardiac CT

,
,
and
College of Computer Science and Technology, Jilin University, Changchun 130012, China
*
Author to whom correspondence should be addressed.
This article belongs to the Section Artificial Intelligence

Abstract

Calcified plaque in coronary arteries is one major cause and prediction of future coronary artery disease risk. Therefore, the detection of calcified plaque in coronary arteries is exceptionally significant in clinical for slowing coronary artery disease progression. At present, the Convolutional Neural Network (CNN) is exceedingly popular in natural images’ object detection field. Therefore, CNN in the object detection field of medical images also has a wide range of applications. However, many current calcified plaque detection methods in medical images are based on improving the CNN model algorithm, not on the characteristics of medical images. In response, we propose an automatic calcified plaque detection method in non-contrast-enhanced cardiac CT by adding medical prior knowledge. The training data merging with medical prior knowledge through data augmentation makes the object detection algorithm achieve a better detection result. In terms of algorithm, we employ a deep learning tool knows as Faster R-CNN in our method for locating calcified plaque in coronary arteries. To reduce the generation of redundant anchor boxes, Region Proposal Networks is replaced with guided anchoring. Experimental results show that the proposed method achieved a decent detection performance.

1. Introduction

Coronary artery disease (CAD) is a typical type of cardiovascular disease that seriously threatens people’s health, and the fall in mortality from cardiovascular disease can increase life expectancy [1]. Calcified plaque in the coronary arteries predicts cardiovascular disease and all-cause mortality and is essential for making an early diagnosis of CAD [2]. Calcified plaques are usually found in the coronary arteries of patients with coronary artery disease [3]. Therefore, calcified plaque detection in coronary arteries is incredibly significant for slowing CAD progression.
The application of Deep Learning has extensively promoted the development of computer vision. Convolutional Neural Network (CNN) has achieved or even exceeded the human level in image classification, image and video recognition, natural language processing, etc. With the development of Deep Learning, more and more Deep Learning methods have also appeared in medical image processing. Especially in the novel corona-virus disease (COVID-19) pandemic, Deep Learning is used to detect lung infections in chest CT images. In [4], a weakly supervised Deep Learning method was proposed to detect infection and distinguish COVID-19 from non-COVID-19 cases. In [5], COVID TV-UNet was proposed to segment the regions infected by COVID-19 in CT images.
However, there are some problems in using Deep Learning to process medical images. First, it is challenging to label medical images. Most previous methods relied on segmentation as the necessary step in the calcified plaque detection processing. In the actual calcified plaques labeling, human operators use commercial software to manually label calcified plaques point by point in each Computed Tomography (CT) scan. Labeling medical images is a time-consuming and labor-intensive process and must be done by professionals.
The second is insufficient medical image data. Due to the privacy protection requirements, it is necessary to eliminate patient identification information and erase sensitive information, that is, de-identification, before using Deep Learning to process medical images. Therefore, medical image data is not as easy to obtain as natural image data. Meanwhile, Deep learning methods require a more considerable amount of image data to achieve better performance.
The third is various problems arising from different medical imaging tasks. We found redundant anchors to one object during the experiment when using the Region Proposal Network (RPN) method to generate anchor boxes. Calcified plaques were set an area ≥1 mm2 where the computed tomographic density of every pixel was ≥130 Hounsfield units [6]. According to the definition, calcified plaques are impossible overlapping, although objects’ centers may overlap in natural images.
Many methods just transplant natural image processing methods into medical image processing. Our solutions to the above problems are personalized to the calcified plaque detection task and different from natural image processing methods. In particular, the details of our solutions to the above issues are as follows.
First, in the clinical diagnosis, the fundamental problem is the existence of calcified plaque in coronary arteries. Therefore, the goal of our method is to ensure that no calcified plaque remains undetected. Moreover, for a large number of medical images, lesion-level labeling is easier and faster than pixel-level. We transform a public pixel-level segmentation dataset into a lesion-level detection dataset. More labeled medical image data can be obtained for detection training by the Deep Learning model.
Second, the medical image dataset’s data augmentation is quite necessary to use a limited number of data to achieve the desired result. Current traditional data augmentation solutions of natural images, such as flipping, cropping, etc., are not for medical images’ characteristics. We propose a data augmentation technique for calcified plaque detection, a fusion of medical prior knowledge. Using prior knowledge improves the detection accuracy while increases the amount of data.
Finally, although Not Maximum Suppression (NMS) avoids redundant anchor boxes by setting IoU-threshold in object detection networks, the IoU-threshold needs to be manually set through the experience of object detection. If the IoU-threshold is not set correctly, there are missing detection or redundant detection boxes inevitably. The idea of guided anchoring [7] predicts whether the feature map’s pixel is an object’s center (the probability of an object’s center). Therefore, there can only be one anchor for each pixel location. This character of guided anchoring makes it very suitable for detecting calcified plaques.
In this paper, we merge medical prior knowledge in the data augmentation stage to improve calcified plaque detection accuracy. In terms of model, we chose guided anchoring to avoid redundant anchor boxes.

3. Methods

In this section, we introduce our method of calcified plaques detection of coronary arteries. Figure 1 is the framework of the proposed automatic method. Firstly, the train set of the model is obtained by data augmentation with prior knowledge. Secondly, we get feature maps of input images in the train set by the backbone network. Thirdly, the guided anchoring module uses the extracted feature map as an input and outputs generated anchors and the adapted new feature map. Finally, in the detection network, classify and regress the anchors on the new feature map to get the final detection result.
Figure 1. The framework of the proposed automatic method.

3.1. Data Augmentation

Deep learning training requires a lot of labeled data. It is difficult to obtain medical images. Labeled medical images are more difficult to obtain because they require professionals to the label. The data augmentation method used in natural images may affect the detection accuracy in medical images. To increase the number of images without affecting the accuracy of calcified plaque detection, we propose a data augmentation method with merging the prior knowledge of calcified plaque.
The train set we used is expressed as follow:
I pri_k   =   I original ,   I threshold ,   I equalization
where I o r i g i n a l is a collection of original labeled CT images. I t h r e s h o l d is I o r i g i n a l after threshold processing. Calcium appears bright on a CT image, that is, it has a high CT Hounsfield unit. In [6], the threshold is set to 130 HU. A calcified plaque is defined as an area (≥1 mm2) in which all pixels are with a CT density above 130 HU. We use the calcified plaque definition as prior knowledge of the detection model. The prior knowledge is fused into the training process of the model by adding I t h r o u h o l d in the train set. Every image in I t h r o u h o l d is retaining the pixels with CT density greater than or equal to 130 HU and resetting pixels to 0 with CT density less than 130 HU, as follow:
θ t h r e s h o l d = 0 θ o r i g i n a l < 130   h u θ o r i g i n a l o t h e r w i s e ,
where θ t h r e s h o l d and θ o r i g i n a l is the CT value of the pixel in I t h r e s h o l d image and I o r i g i n a l image respectively.
I e q u a l i z a t i o n is the set of images after histogram equalization in I t h r e s h o l d Histogram equalization can increase image contrast and better detail in the images. The intensities can be better distributed on the histogram through histogram equalization. The conversion formula is as follow:
v e q u a l i z a i t o n k = f v t h r e s h o l d k = L 1 · i = 0 k n t h r e s h o l d i N t h r e s h o l d ,
where k , k 0 , L 1 , L is the total number of gray levels in the image. v k is the gray value of image at k level. n i is the number of occurrences at gray level i . N is the total number of pixels in the image. f is the histogram equalization function.

3.2. Detection Framework

Until now, there are two types of object detection CNN models, one-stage learning and two-stage learning. The Faster R-CNN [24] we adopted is a widespread object detection CNN framework and belongs to two-stage learning. The one-stage detector is generally faster but less accurate than the two-stage detector. Medical images require higher precision, not faster speed. Consequently, we adopt Faster R-CNN to detect calcified plaque in CT images.
Faster R-CNN model workflow includes the backbone for feature extraction, RPN for region proposal generation, and detection network for classification and regression. In feature extraction, Faster R-CNN uses a backbone network to extract feature maps of input images. The feature maps are shared for subsequent RPN and RoIHead. RPN generated nine anchors centered on every pixel of the feature map. For a convolutional feature map of size W × H, there are 9 × W × H anchors in total. N c l s out of 9 × W × H anchors were selected to participate in RPN training. After RPN training, a set of region proposals is obtained. The detection network classifies the region proposals and regresses the bounding boxes of region proposals.

3.3. Region Proposal Generation

In a two-stage object detection framework, the anchors generated by region proposal generation are used as the input of the detection network for classification and regression. Therefore, the quality of anchors has a significant influence on the final detection result. RPN in Faster R-CNN generates a large set of densely distributed anchors with sliding windows. Only a few anchors are related to the regions of interest. A large of useless anchors caused computational cost. Therefore, Wang, J et al. [7] proposed guided anchoring, a more effective region proposal generation method.
The guided anchoring method generated sparse anchors that may contain objects instead of roughly generating even anchors. Firstly, the guided anchoring method used an image pyramid to build a feature pyramid. Secondly, for each output feature map of the feature pyramid, the anchor location and shape were predicted by two sub-networks branches. A feature adaption module applied deformable convolution to the original feature map to obtain an adapted feature map. Finally, classification and bounding-box regression were performed on the adapted feature map. There are three steps to generate anchors as follows.
Anchor location prediction. For each output feature map F I of the feature pyramid, anchor location prediction branch using a sub-network to obtain the probability map p ( i ,   j |   F I ) , where i , j is the coordinate on F I .
Anchor shape prediction. For each output feature map F I of the feature pyramid, anchor shape prediction branch using a sub-network to predict the best shape w ,   h for each location, where w ,   h is the width and height of anchor. The branch outputs d w ,   d h and uses the transformation w = σ · s · e d w ,   h = σ · s · e d h to obtain w ,   h , where s is the stride and σ is an empirical scale factor.
Anchor-guided feature adaption. Since each anchor has a different shape at the corresponding location, the original feature map cannot be used to obtain the region proposal. Anchor-guided feature adaption branch applies a 3 × 3 deformable convolutional layer to transform the feature f i based at the i -th location f i = N T f i ,   w i , h i , where f i is the transformed feature and ( w i , h i ) is the corresponding anchor shape.

4. Experiments and Results

4.1. Dataset

CT is commonly used in clinical examination of the coronary arteries, which allows for noninvasive detection and characterization of calcified plaques. Coronary CT is a widely used imaging examination, and it is an essential basis for doctors to confirm the diagnosis. In clinical, calcified plaque is typically measured on ECG-gated 3 mm CT images [29].
OrCaScore [30] is a public dataset of calcified plaque scores. There is a non-contrast-enhanced calcium scoring CT (CSCT) and a corresponding contrast-enhanced cardiac computed tomography angiography (CCTA) of each patient. As described in [30], the train set includes 32 patients CSCT and CCTA, and the test set includes 40 patients’ images. Images in the dataset were acquired on four different CT scanners from four different vendors in four different hospitals. Patients were scanned on a multidetector-row CT scanner from one of the four different vendors (Lightspeed VCT, software version gmp vct. 26, GE Healthcare, Milwaukee, WI, USA; Brilliance iCT, software version 3.2.0, Philips Healthcare, Best, The Netherlands; SOMATOM Definition Flash, software version syngo ct 2010A, Siemens Healthcare, Forchheim, Germany; and Aquilion ONE, software version 4.93, Toshiba Medical Systems, Otawara, Japan).
In clinical practice, not every patient can undergo a CCTA scan. Many patients only receive CSCT with a lower radiation dose if the patient cannot receive the iodine-containing contrast material as an intravenous (IV) injection. Therefore, we used only the CSCT of the OrCaScore train set.
Unprocessed data is CT volume data in DICOM format. Digital Imaging and Communications in Medicine (DICOM) is the standard for the communication and management of medical imaging information and related data. Each patient corresponds to a DICOM file. A DICOM file consists of a number of attributes, including items such as name, ID, pixel information, etc. We extract axial slices from the CT volume data. We convert the slices to grayscale images with 8-bit. Images with no calcified plaques are eliminated. The dataset I o r i g i n a l consists of 180 images that contain calcified plaques. After data augmentation in Section 2.1, there are 432 images in the train set and 36 images in the test set. The size of the image is 512 × 512. The number of images in each set of the original dataset I o r i g i n a l and dataset with prior knowledge I p r i _ k .is shown in Table 1.
Table 1. The number of images in each set.
orCaScore is a voxel-level segmentation dataset, and we transform it into a lesion-level detection dataset by adding a bounding box outside of every lesion in the reference standard. Figure 2 shows an example of an image and corresponding ground truth.
Figure 2. An example of an image and corresponding ground truth. (a) An image of one patient in orCaScore. (b) The corresponding ground truth, the red bounding box points out the calcified plaques in this image.

4.2. Performance Metrics

We use a confusion matrix to present a visualization of algorithm performance. Each column represents the predicted value, and each row represents the actual category. Such as Table 2.
Table 2. Performance confusion matrix.
The conditions for correct prediction are below:
  • The category is correct, and the confidence is more significant than a certain threshold. In our experiments, the threshold is 0.5.
  • The IoU of the prediction box and the ground truth box is more significant than a certain threshold, IoU_threshold which is set as 0.5 in the experiments. The intersection ratio IoU measures the degree of overlap between the two regions and is the overlap of the two regions occupying the total area.
I o U A , B =   A B A B ,
where I o U A ,   B is the intersection ratio IoU between bounding box A and ground truth B .
In the evaluation, there are three classifications of every bounding box predicted by our model. The bounding box which correctly identified calcified plaque is classified as Ture Positive (TP). The bounding box which identifies calcified plaque but is not in reference standard is classified as False Positive (FP). The bounding box which is calcified plaque but not identified is classified as False Negative (FN).
In the lesion objection, we want to reduce the probability of missed diagnosis as much as possible in clinical. We need to increase the true positives and decrease the false negatives. Therefore, we choose Average Precision (AP) and recall as the evaluation standard. AP measures the overall performance of the model, and recall measures the missed diagnosis of the model. The higher the AP and recall, the better the performance of the model.
Recall and AP are defined as the following formulas:
R e c a l l =   T r u e   P o s i t i v e T u r e   P o s i t i v e + F a l s e   N e g a t i v e ,
A P = R n R n 1 P r e p l a c e R n ,
where P and R are Precision and Recall, respectively. R n is the Recall value at the n-th threshold. P R is the Precision value at the corresponding recall. The precision and P r e p l a c e R are defined as follows.
P r e c i s i o n = T r u e   P o s i t i v e T u r e   P o s i t i v e + F a l s e   P o s t i v e ,
P r e p l a c e R = max R R P R .

4.3. Experiment Implementation Details

4.3.1. Training Hyper-Parameters

All experiments are run on Ubuntu 18.04 with Python 3.7. Our experiments are implemented with PyTorch. We trained on 4 NVIDIA Geforce 2080ti GPUs with 2 images per GPU. We use the ResNet50 [31] with Feature Pyramid Network(FPN) [32] as the backbone architecture. The learning rate is initialized to 0.01 and decreased by 0.1 at epoch 8 and 11. The coding is implemented based on a public platform, MMDetection [33]. The Hyper-parameters detail is as Table 3.
Table 3. The hyper-parameters detail.

4.3.2. Loss Function

The loss function in the end-to-end training is a multi-task loss. There are two parts to the detection framework, region proposal generation, and detection network. In region proposal generation, since the center region of the object usually occupies a small part of a whole picture, the loss L l o c for anchor location prediction branch is Focal Loss [34]:
L l o c = α 1 p γ log p i f   y = 1 1 α p γ log 1 p o t h e r w i s e ,
where y ± 1 is the label representing whether the pixel belongs to the center region of the object and p 0 , 1 is the probability for the class with label y = 1 . α and γ are weighting factors, and the values are 0.25 and 2.0 in the experiments, respectively.
The loss L s h a p e for anchor shape prediction branch is bounded iou loss which is defined as:
L s h a p e =   L 1 1 min w w g ,   w g w + L 1 1 min h h g ,   h g h ,
where L 1 is the smooth L1 loss, w ,   h and w g ,   h g are the predicted anchor shape and the corresponding ground truth bounding box shape.
In the detection network, the loss L c l s and L r e g are cross-entropy loss and smooth L1 loss, respectively.
The final joint training loss is as follows:
L =   λ 1 L l o c + λ 2 L s h a p e + L c l s + L r e g .
where both λ 1 and λ 2 are 1 in the experiments.

4.4. Calcified Plaque Detection Result

We compare the six methods in the experiment, Faster R-CNN with RPN, K. Chellamuthu et al. [22] which is introduced in Section 2.2, DetectoRS [35], Libra R-CNN [36], CARAFE [37], and Faster R-CNN with guided anchoring employed in this paper. We train and test the six methods with two datasets, respectively, the original dataset I o r i g i n a l and dataset with prior knowledge I p r i _ k .
Table 4 shows that the performance of Faster R-CNN with guided anchoring trained on the dataset I p r i _ k is superior to that of the other methods. Of the methods considered, the method proposed in this study has the highest recall and AP. As pointed out in Table 4, guided anchoring increases recall to 0.814 and AP to 0.724 instead of RPN, although two sequential Faster R-CNN increase recall from 0.744 to 0.756 and AP from 0.554 to 0.567. Guided anchoring uses fewer parameters and faster speed to get better performance. Although the recall of Libra R-CNN and CARAFE is higher than Faster R-CNN with guided anchoring, we can see from the lower AP of Libra R-CNN and CARAFE that the high recall is obtained through the generation of a large number of predicted bounding boxes to prevent missed detection. Therefore, Faster R-CNN with guided anchoring performs better from the better AP, which measures the model’s overall performance.
Table 4. The comparison results of Faster R-CNN with RPN, K. Chellamuthu et al., DetectoRS, Libra R-CNN, CARAFE, and Faster R-CNN with guided anchoring trained on original dataset and dataset with prior knowledge, respectively. CI = confidence interval.
The role of data augmentation with prior knowledge is also reflected in Table 4. From the experiment result, we can see that the results of the three methods trained on the dataset I p r i _ k greatly surpassed that trained on the dataset I o r i g i n a l . Especially in Faster R-CNN with guided anchoring, the recall increases from 0.576 to 0.814, and AP increases from 0.379 to 0.724. The data augmentation with prior knowledge can significantly improve the calcified plaque detection performance.
Figure 3 and Figure 4 are the detection results comparing Faster R-CNN and Faster R-CNN with guided anchoring of the same image. The result images are cropped to see clearly. Figure 3 shows the situation of missed detection in the experiment of Faster R-CNN. The false-negative of Faster R-CNN (Figure 4a) is detected by Faster R-CNN with guided anchoring (Figure 4b). Figure 4 shows two examples of reducing redundant detection of Faster R-CNN with guided anchoring. Moreover, both in the two examples of Figure 4, the calcified plaques detected by Faster R-CNN with Guided Anchoring have a higher classification score than by Faster R-CNN.
Figure 3. Missed detection of Faster R-CNN. (a) The detection result of Faster R-CNN. The calcified plaque is not detected. (b) The detection result of Faster R-CNN with guided anchoring. (c) The ground truth of the image. The ground truth is labeled by the red bounding box and pointed out by a red arrow.
Figure 4. Two examples of the redundant detection of Faster R-CNN. (a) The detection results of Faster R-CNN. (b) The detection results of faster R-CNN with guided anchoring. (c) The ground truths of the images.

4.5. Analysis of Improvement

Table 5 lists the comparison result of different backbone in the detection framework. In the experiment of our proposed method, we choose four popular CNN architectures as backbone, ResNet50, ResNet101, ResNeXt101_32 × 4d and ResNeXt101_64 × 4d. As pointed as Table 5, ResNet101 as the backbone network has the best result in both recall and AP. Therefore, it is not that the more complex the network has better results. Over-fitting caused by an overly complex network makes recall and AP drop.
Table 5. Results of different backbones.
Table 6 shows the comparison result of a pre-trained network and a network trained from scratch. Here we use Faster R-CNN with guided anchoring trained on PASCAL VOC dataset as the pre-trained network. Table 6 results show that the pre-trained network has improved recall and AP compared to the network trained from scratch.
Table 6. Comparison of a pre-trained network and a network trained from scratch.
Table 7 shows the impact of the hyper-parameter scale value at 2, 4, 6, 8. The scale is a hyper-parameter representing the size range of the anchor, detail in Section 3.3. Since the size of the calcified plaques is smaller than most objects in natural images, the model has the best performance when the scale value is 2 as pointed out in Table 7.
Table 7. The impact of the scale on the model performance.

5. Conclusions

We have presented a method with prior knowledge to automatically detect calcified plaque of coronary arteries in non-contrast-enhanced cardiac CT. Our proposed method expands the training data by adding prior knowledge and improves the performance of the calcified plaque detection task. The method employs guided anchoring as the region proposal generation network in Faster R-CNN. In the comparing of the RPN, guided anchoring can solve the problem of anchor box redundancy. Compared with other methods, our method gets better results.

Author Contributions

Conceptualization, M.Z.; methodology, M.Z.; software, M.Z. and H.L.; formal analysis, M.Z.; investigation, M.Z.; resources, X.C.; data curation, M.Z.; writing—original draft preparation, M.Z.; writing—review and editing, X.C. and Q.L.; supervision, X.C.; project administration, M.Z.; funding acquisition, X.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 61672260, by Science and Technology Development Plan of Jilin Province of China under Grant 20200401077GX, 20200201292JC.

Acknowledgments

We thank Yahui Shao for his help in data analysis.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Naghavi, M.; Wang, H.; Lozano, R.; Davis, A.; Liang, X.; Zhou, M.; Vollset, S.E.; Abbasoglu Ozgoren, A.; Abdalla, S.; Abd-Allah, F.; et al. Global, regional, and national age-sex specific all-cause and cause-specific mortality for 240 causes of death, 1990–2013: A systematic analysis for the Global Burden of Disease Study 2013. Lancet 2015, 385, 117–171. [Google Scholar]
  2. Detrano, R.; Guerci, A.D.; Carr, J.J.; Bild, D.E.; Burke, G.; Folsom, A.R.; Liu, K.; Shea, S.; Szklo, M.; Bluemke, D.A.; et al. Coronary calcium as a predictor of coronary events in four racial or ethnic groups. N. Engl. J. Med. 2008, 358, 1336–1345. [Google Scholar] [CrossRef]
  3. Liu, W.; Zhang, Y.; Yu, C.M.; Ji, Q.W.; Cai, M.; Zhao, Y.X.; Zhou, Y.J. Current understanding of coronary artery calcification. J. Geriatr. Cardiol. 2015, 12, 668–675. [Google Scholar] [PubMed]
  4. Hu, S.; Gao, Y.; Niu, Z.; Jiang, Y.; Li, L.; Xiao, X.; Wang, M.; Fang, E.F.; Menpes-Smith, W.; Xia, J.; et al. Weakly Supervised Deep Learning for COVID-19 Infection Detection and Classification from CT Images. IEEE Access 2020, 8, 118869–118883. [Google Scholar] [CrossRef]
  5. Saeedizadeh, N.; Minaee, S.; Kafieh, R.; Yazdani, S.; Sonka, M. COVID TV-UNet: Segmenting COVID-19 chest CT images using connectivity imposed U-Net. arXiv 2020, arXiv:2007.12303. [Google Scholar]
  6. Agatston, A.S.; Janowitz, W.R.; Hildner, F.J.; Zusmer, N.R.; Viamonte, M.; Detrano, R. Quantification of coronary artery calcium using ultrafast computed tomography. J. Am. Coll. Cardiol. 1990, 15, 827–832. [Google Scholar] [CrossRef]
  7. Wang, J.; Chen, K.; Yang, S.; Loy, C.C.; Lin, D. Region proposal by guided anchoring. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2019, 2019, 2960–2969. [Google Scholar]
  8. Išgum, I.; Rutten, A.; Prokop, M.; Staring, M.; Klein, S.; Pluim, J.P.; Viergever, M.A.; van Ginneken, B. Automated aortic calcium scoring on low-dose chest computed tomography. Med. Phys. 2010, 37, 714–723. [Google Scholar] [CrossRef]
  9. Išgum, I.; van Ginneken, B.; Viergever, M.A. Automatic detection of calcifications in the aorta from abdominal CT scans. Int. Congr. Ser. 2003, 1256, 1037–1042. [Google Scholar] [CrossRef]
  10. Išgum, I.; Prokop, M.; Niemeijer, M.; Viergever, M.A.; Van Ginneken, B. Automatic coronary calcium scoring in low-dose chest computed tomography. IEEE Trans. Med. Imaging 2012, 31, 2322–2334. [Google Scholar] [CrossRef]
  11. Wolterink, J.M.; Leiner, T.; Takx, R.A.P.; Viergever, M.A.; Išgum, I. Automatic Coronary Calcium Scoring in Non-Contrast-Enhanced ECG-Triggered Cardiac CT with Ambiguity Detection. IEEE Trans. Med. Imaging 2015, 34, 1867–1878. [Google Scholar] [CrossRef] [PubMed]
  12. Yang, G.; Chen, Y.; Ning, X.; Sun, Q.; Shu, H.; Coatrieux, J.L. Automatic coronary calcium scoring using noncontrast and contrast CT images. Med. Phys. 2016, 43, 2174–2186. [Google Scholar] [CrossRef] [PubMed]
  13. Durlak, F.; Wels, M.; Schwemmer, C.; Sühling, M.; Steidl, S.; Maier, A. Growing a random forest with Fuzzy spatial features for fully automatic artery-specific coronary calcium scoring. Lect. Notes Comput. Sci. (Incl. Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinform.) 2017, 10541, 27–35. [Google Scholar]
  14. Acharya, U.R.; Meiburger, K.M.; Koh, J.E.W.; Ciaccio, E.J.; Vicnesh, J.; Tan, S.K.; Wong, J.H.D.; Aman, R.R.A.R.; Ng, K.H. Automated detection of calcified plaque using higher-order spectra cumulant technique in computer tomography angiography images. Int. J. Imaging Syst. Technol. 2019, 30, 285–297. [Google Scholar] [CrossRef]
  15. Mirsky, Y.; Mahler, T.; Shelef, I.; Elovici, Y. CT-GAN: Malicious tampering of 3D medical imagery using deep learning. In Proceedings of the 28 USENIX Security Symposium, Santa Clara, CA, USA, 14–16 August 2019; pp. 461–478. [Google Scholar]
  16. Prasoon, A.; Petersen, K.; Igel, C.; Lauze, F.; Dam, E.; Nielsen, M. Deep feature learning for knee cartilage segmentation using a triplanar convolutional neural network. Lect. Notes Comput. Sci. (Incl. Subser. Lect. Notes Artif. Intell. Lect. Notes Bioinform.) 2013, 8150, 246–253. [Google Scholar]
  17. Lessmann, N.; Išgum, I.; Setio, A.A.A.; de Vos, B.D.; Ciompi, F.; de Jong, P.A.; Oudkerk, M.; Mali, W.P.T.M.; Viergever, M.A.; van Ginneken, B. Deep convolutional neural networks for automatic coronary calcium scoring in a screening study with low-dose chest CT. In Medical Imaging 2016: Computer-Aided Diagnosis; International Society for Optics and Photonics: San Diego, CA, USA, 2016; Volume 9785, p. 978511. [Google Scholar]
  18. Wolterink, J.M.; Leiner, T.; de Vos, B.D.; van Hamersvelt, R.W.; Viergever, M.A.; Išgum, I. Automatic coronary artery calcium scoring in cardiac CT angiography using paired convolutional neural networks. Med. Image Anal. 2016, 34, 123–136. [Google Scholar] [CrossRef]
  19. Lessmann, N.; Van Ginneken, B.; Zreik, M.; De Jong, P.A.; De Vos, B.D.; Viergever, M.A.; Išgum, I. Automatic Calcium Scoring in Low-Dose Chest CT Using Deep Neural Networks with Dilated Convolutions. IEEE Trans. Med. Imaging 2018, 37, 615–625. [Google Scholar] [CrossRef]
  20. Gernaat, S.A.M.; van Velzen, S.G.M.; Koh, V.; Emaus, M.J.; Išgum, I.; Lessmann, N.; Moes, S.; Jacobson, A.; Tan, P.W.; Grobbee, D.E.; et al. Automatic quantification of calcifications in the coronary arteries and thoracic aorta on radiotherapy planning CT scans of Western and Asian breast cancer patients. Radiother. Oncol. 2018, 127, 487–492. [Google Scholar] [CrossRef]
  21. Liu, J.; Lu, L.; Yao, J.; Bagheri, M.; Summers, R.M. Pelvic artery calcification detection on CT scans using convolutional neural networks. In Medical Imaging 2017: Computer-Aided Diagnosis; International Society for Optics and Photonics: San Diego, CA, USA, 2017; Volume 10134, p. 101341A. [Google Scholar]
  22. Chellamuthu, K.; Liu, J.; Yao, J.; Bagheri, M.; Lu, L.; Sandfort, V.; Summers, R.M. Atherosclerotic vascular calcification detection and segmentation on low dose computed tomography scans using convolutional neural networks. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, Australia, 18–21 April 2017; pp. 388–391. [Google Scholar]
  23. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE conference on computer vision and pattern recognition 2014, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  24. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 2015, 91–99. [Google Scholar] [CrossRef]
  25. Zreik, M.; Van Hamersvelt, R.W.; Wolterink, J.M.; Leiner, T.; Viergever, M.A.; Išgum, I. A Recurrent CNN for Automatic Detection and Classification of Coronary Artery Plaque and Stenosis in Coronary CT Angiography. IEEE Trans. Med. Imaging 2019, 38, 1588–1598. [Google Scholar] [CrossRef]
  26. Fischer, A.M.; Eid, M.; De Cecco, C.N.; Gulsun, M.A.; Van Assen, M.; Nance, J.W.; Sahbaee, P.; De Santis, D.; Bauer, M.J.; Jacobs, B.E.; et al. Accuracy of an Artificial Intelligence Deep Learning Algorithm Implementing a Recurrent Neural Network with Long Short-term Memory for the Automated Detection of Calcified Plaques from Coronary Computed Tomography Angiography. J. Thorac. Imaging 2020, 35, S49–S57. [Google Scholar] [CrossRef] [PubMed]
  27. Liu, J.; Wang, D.; Lu, L.; Wei, Z.; Kim, L.; Turkbey, E.B.; Sahiner, B.; Petrick, N.A.; Summers, R.M. Detection and diagnosis of colitis on computed tomography using deep convolutional neural networks. Med. Phys. 2017, 44, 4630–4642. [Google Scholar] [CrossRef] [PubMed]
  28. Yazdani, S.; Minaee, S.; Kafieh, R.; Saeedizadeh, N.; Sonka, M. COVID CT-Net: Predicting Covid-19 from Chest CT Images Using Attentional Convolutional Network. arXiv 2020, arXiv:2009.05096. [Google Scholar]
  29. Hughes-Austin, J.M.; Dominguez, A.; Allison, M.A.; Wassel, C.L.; Rifkin, D.E.; Morgan, C.G.; Daniels, M.R.; Ikram, U.; Knox, J.B.; Wright, C.M.; et al. Relationship of Coronary Calcium on Standard Chest CT Scans With Mortality. JACC Cardiovasc. Imaging 2016, 9, 152–159. [Google Scholar] [CrossRef]
  30. Wolterink, J.M.; Leiner, T.; De Vos, B.D.; Coatrieux, J.L.; Kelm, B.M.; Kondo, S.; Salgado, R.A.; Shahzad, R.; Shu, H.; Snoeren, M.; et al. An evaluation of automatic coronary artery calcium scoring methods with cardiac CT using the orCaScore framework. Med. Phys. 2016, 43, 2361–2373. [Google Scholar] [CrossRef] [PubMed]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2016, 2016, 770–778. [Google Scholar]
  32. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  33. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open MMLab Detection Toolbox and Benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  34. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef]
  35. Qiao, S.; Chen, L.C.; Yuille, A. DetectoRS: Detecting objects with recursive feature pyramid and Switchable Atrous Convolution. arXiv 2020, arXiv:2006.02334. [Google Scholar]
  36. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra R-CNN: Towards balanced learning for object detection. Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. 2019, 2019, 821–830. [Google Scholar]
  37. Wang, J.; Chen, K.; Xu, R.; Liu, Z.; Loy, C.C.; Lin, D. CARAFE: Content-aware reassembly of features. Proc. IEEE Int. Conf. Comput. Vis. 2019, 2019, 3007–3016. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.