Skip to Content
SensorsSensors
  • Article
  • Open Access

3 December 2021

Detection of Collaterals from Cone-Beam CT Images in Stroke

,
,
,
,
,
and
1
Department of Electrical and Electronic Engineering, Universiti Teknologi PETRONAS (UTP), Seri Iskandar 32610, Malaysia
2
Centre for Intelligent Signal and Imaging Research (CISIR), Department of Electrical and Electronic Engineering, Universiti Teknologi PETRONAS (UTP), Seri Iskandar 32610, Malaysia
3
School of Electrical Engineering, Faculty of Engineering, Universiti Teknologi Malaysia, Johor Bahru 81310, Malaysia
4
Department of Radiology, Faculty of Medicine and Health Sciences, Hospital Pengajar Universiti Putra Malaysia, Serdang 43400, Malaysia
This article belongs to the Section Sensing and Imaging

Abstract

Collateral vessels play an important role in the restoration of blood flow to the ischemic tissues of stroke patients, and the quality of collateral flow has major impact on reducing treatment delay and increasing the success rate of reperfusion. Due to high spatial resolution and rapid scan time, advance imaging using the cone-beam computed tomography (CBCT) is gaining more attention over the conventional angiography in acute stroke diagnosis. Detecting collateral vessels from CBCT images is a challenging task due to the presence of noises and artifacts, small-size and non-uniform structure of vessels. This paper presents a technique to objectively identify collateral vessels from non-collateral vessels. In our technique, several filters are used on the CBCT images of stroke patients to remove noises and artifacts, then multiscale top-hat transformation method is implemented on the pre-processed images to further enhance the vessels. Next, we applied three types of feature extraction methods which are gray level co-occurrence matrix (GLCM), moment invariant, and shape to explore which feature is best to classify the collateral vessels. These features are then used by the support vector machine (SVM), random forest, decision tree, and K-nearest neighbors (KNN) classifiers to classify vessels. Finally, the performance of these classifiers is evaluated in terms of accuracy, sensitivity, precision, recall, F-Measure, and area under the receiver operating characteristics curve. Our results show that all classifiers achieve promising classification accuracy above 90% and able to detect the collateral and non-collateral vessels from images.

1. Introduction

Stroke is the third leading cause of mortality in Malaysia, and the cases of stroke are growing every year [1]. Stroke can damage brain tissues, and impair our body functions leading to numbness, weakness, or paralysis on the side of the body opposite to the obstruction of the artery. Collateral vessels are vessels that connect the damaged tissue with other arteries and restore the blood flow to the tissues, thus preventing serious damage and offer better chance of recovery. Imaging of the brain vessels has shown that collateral flow can sustain brain tissue for hours after the occlusion of major arteries to the brain, and maintenance of collateral flow is therefore important. Several studies show that the outcome after stroke treatment is highly dependent on the extent of collateral flow [2,3]. The number of collateral vessels varies from person to person, contributing to different stroke outcomes. A patient with good collateral flow may have favorable outcomes and better clinical response to a recanalization therapy, even though the treatment is initiated at delayed hours after the onset of stroke [4,5], unlike a patient with poor collateral flow.
Imaging plays a critical role in the diagnosis of stroke patients before initiating treatment and for early intervention. There are various imaging techniques available for assessing vascular lesions and brain tissue status in acute stroke patients. Digital subtraction angiography (DSA) is gold standard imaging to identify collateral vessels in acute stroke. However, DSA only provides two-dimensional (2D) imaging, and it is an invasive technique requiring a special contrast medium or ‘dye’ to be injected into blood vessels to make the vessels visible. Imaging techniques based on computed tomography (CT) and magnetic resonance imaging (MRI) are able to provide non-invasive angiography and perfusion imaging for the identification of cerebral injury in the first few hours after arterial occlusion [6]. But these imaging techniques have low spatial information of collateral vessels.
Cone-beam computed tomography (CBCT), on the other hand, is an advanced imaging modality that provides accurate and three-dimensional (3D) imaging of hard tissues, soft tissues, and bones. Recently, CBCT is increasingly popular in acute stroke and neurovascular image-guided procedures due to its advantages over conventional CT, such as less x-ray beam limitation, high spatial resolution and rapid scan time depending on the manufacturer providers [7,8]. A newer study applying machine learning techniques to automate the process of estimating stroke tissues has shown a promising result in accuracy [9]. However, this work uses MRI images as supposed to CBCT images. To the best of our ability, we could not find any work on the detection of collateral vessels for stroke cases using CBCT imaging modality based on image processing approach. The only closely related work is about the vessel detection of CBCT images on prostatic arteries [10].
Despite many advantages of CBCT, there are a number of challenges in detecting collateral vessels from CBCT images which are: (1) low contrast, (2) existence of various noises and artifacts such as motion, beam hardening, aliasing and ring artifacts, and (3) non-uniform illumination caused by exponential edge gradient effect [11]. In addition, it is challenging to detect collateral vessels in brain due to their randomized and non-homogeneous structure, and small size (ranging from 0.4–0.6 mm [12]).
Image processing is a popular technique used for detection and classification of various diseases. The work in [13] uses the hyperspectral imaging technique to detect blood vessels and to differentiate the artery from vein during surgeries using MRI images. Several works are proposed for detecting blood vessels of retinal images based on fundus modality. The complete review on these works is available in [14]. In our work, we use image processing technique for the application of stroke and our images are based on CBCT image modality.
In our work, we develop a technique for accurate detection of collateral vessels from CBCT images to assist neuroradiologists to objectively visualize the collateral vessels. Typically, subjective assessment is used in diagnosis and depending on neuroradiologist’s expertise and experience, it may lead to different interpretation, consequently influencing the outcome of treatment received by patients. The proposed technique consists of: (1) image pre-processing, (2) feature extraction, and (3) classification processes. Firstly, the original images are processed and enhanced to improve their visibility. Then, we extract their texture and shape features. Finally, we investigate several types of classifiers on the extracted features to explore which classifier is the best to classify the collateral vessels from normal vessels. The performance of our detection technique is then validated using a 25% holdout cross-validation technique and measured using various performance metrics such as accuracy, sensitivity, precision, recall, F-Measure, and AUC values of the receiver operating characteristics (ROC) curves.
The paper outline is as follows. In Section 2, we provide the review on the related works for detecting the collateral vessels. Next, Section 3 presents the methodology of our work, and Section 4 discusses results and presents the analysis of the results. Lastly, Section 5 concludes the paper and suggests further work in this area.

3. Methodology

The chart in Figure 1 illustrates the collateral vessel detection methodology in our study.
Figure 1. The flowchart of the proposed collateral vessel detection.

3.1. Image Acquisition

In this study, we use anonymous data of the CBCT image validated by radiologists. The images are acquired from Hospital Universiti Kebangsaan Malaysia (HUKM) using a Philips X-ray scanner. CBCT image is a 3D volumetric image acquired using a flat detector of the angiography system, which reconstructed as axial CT like images. The images obtained of stroke patients used are in the DICOM format. The thickness of each image slice is 0.762841 mm, and the image size is 256 × 256.

3.2. Conversion of CBCT to CT Values

CT image is reconstructed based on the attenuation coefficient, whereby air, water, soft tissues, and bone have their attenuation value. The attenuation coefficient measures how easily the tissue can be penetrated by an X-ray beam. The gray value in CT image is called CT value and it is measured in the Hounsfield unit (HU). CBCT image has a larger amount of scattered X-rays that resulted in artifacts. Scatter is well known to further reduce soft-tissue contrast and it will also affect the density values of all other tissues [Artifact: the downturn of CBCT image]. and the Different from CT image, CBCT, as mention in the previous section, has a larger amount of scattered X-rays. Because of the scatter and resultant artifact, CT values in CBCT are not valid [61]. In the previous study [62], a linear relationship between gray values in CBCT and the attenuation coefficient of various materials in several CBCT scanners was reported [63]. It can be concluded that the conversion of gray values from CBCT to CT values could be performed. In another study, Chindasombatjaroen et al. [63] highlighted that the pixels value in CBCT is lower than CT values in CT. Therefore, we proposed two steps in the CT value conversion of CBCT image; (1) brightness adjustment and (2) CT value conversion.
Brightness adjustment: The aim is to adjust the tonal range of the image to increase the intensity of blood vessels in the image. We adjust the brightness of the input image by implementing a linear regression model, as described in Equation (1). The regression model in Equation (1) is designed using a quality factor, Q, as a predictor variable, and recognition performance, Y, as the response variable in the model [64],
Y = m + w Q
where w represents the corresponding regression coefficients, and m is the regression constant.
CT value conversion: During image acquisition, various rescale slope and rescale intercept are adopted in CT modality due to various manufacturing providers. It will affect the CT values in image reconstruction of CT or CBCT images. Therefore, CT value conversion aims to calculate the adjusted image, Y, in step 1 to obtain the CT value, HU, ensuring this converted image is capable of preserving as much information as possible, particularly the small vessel structures after the CT value conversion using Equation (2) [65]. The CT value conversion is performed as follows:
HU = Y R S + R I
where R S and R I are generally the rescale slope and rescale intercept, respectively. These values can be found in the DICOM file itself (i.e., image tag), numbered as and respectively. If the image tags are not present in the DICOM file, it means that the CT values are in the default unit, which is Hounsfield (HU) unit. The brightness adjustment and conversion of CT value must then be performed for each image slice.

3.3. Maximum Intensity Projection

Maximum intensity projection (MIP) is subsequently performed onto the converted images, HU, to extract pixels with the highest Hounsfield value along the Z-axis, as shown in the diagram in Figure 2a, so that in a single bi-dimensional image all dense structures in a given volume are observed. The aim is to create an image similar to the ground truth provided by the radiologist using a HOROS software since the MIP method tends to display bone and contrast material-filled structures preferentially. Therefore, the MIP is the most suitable approach compared with minimum and average intensity projection. Figure 2a is the example of 16 mm thick-stacked images and Figure 2b is the result of the MIP process. The size of thickness is selected to replicate the ground truth images prepared by the expert. Using stacked images in MIP, we can improve the visibility of the small vessels in the image compared to only relying on one single slice image. This process will allow us to extract as much information as possible from the image such as the visualization of a long segment of a vessel, decreased perceived image noise as well as providing a sufficiently high contrast between vessel of interest and surrounding structures. In medical practice, the clinicians use MIP images to differentiate the normal blood vessels from the collateral blood vessels.
Figure 2. The images from (a) are stacked, and the maximum pixel value is extracted to generate the image in (b).
The MIP step is performed in every slice. Assuming one subject has 256 sliced images, the first MIP image is from slice 1 to 20. With a gap of one slice, the second MIP image will be from slice 2 to 21. This process will go on until it reaches the last image slice. This is illustrated in Figure 3.
Figure 3. Steps for generating MIP images.

3.4. Skull Removal

The purpose of skull removal is to obtain an MIP image without the skull to facilitate collateral vessel enhancement and segmentation. A mask image of the region of interest (brain area without the skull) is obtained for the skull removal to avoid any false-positive results since the intensity range of the skull is quite close to the intensity range of blood vessels. Skull removal algorithm based on morphological reconstruction, which relies on the connected components and region properties, is used. Skull will be removed based on the labeling connected components and computation of region properties. Figure 4 shows the resultant images obtained throughout the skull removal process to obtain blood vessel structures from an MIP image.
Figure 4. The algorithm of skull removal. (a) MIP input image, (b) binary image, (c) removing object touching the binary image boundary, (d) mask image, and (e) region of interest, masked out from mask image.
The skull removal involves several processes. MIP image obtained from Section 3.3 is used as the input image for this purpose. The first process is to convert the input image into a binary image. To achieve this, we use a thresholding method ( I < x ) where x is a scalar value that is obtained from a mean value of input image, I. The result is as illustrated in Figure 4b. The second process is to eliminate unwanted region; in this case, the white region surrounding the outer area of the skull in Figure 4b. This area is also known as the area connected to the image border. Hence, to remove this, we use the method that removes/clear out all connected components or pixels connected to the image boundary which are the outer white pixels surrounding the skull region. The result image obtained is illustrated in Figure 4c.
The third process is to create a mask image with only white pixels in the foreground and black pixels in the background. The process involves a flood-fill operation that will change connected background pixels (0 s) to foreground pixels (1 s) and a measurement of area for each region. Then, by selecting the largest area, what left is the brain region as illustrated in Figure 4d; the mask image. Lastly, the final process is to obtain the MIP input image without the skull. By performing an element-by-element multiplication involving the MIP input image and the mask image, we obtain the final result as illustrated in Figure 4e.

3.5. Noise and Artifact Removal

There are many filtering methods available for noise and artifact removal in image processing. Before further processing and analysis, removing the noise and artifact is vital to obtain an accurate result. From previous studies in [17,20], the anisotropic diffusion filter has the ability to smoothen images while preserving the edges, thus we use this filter to remove noise in the CBCT images. In our case, the anisotropic diffusion filter may cause the loss of valuable information on the blood vessels’ structure during the smoothing process, as shown in Figure 8. The peak signal-to-noise ratio and structural similarity index are used to measure the image quality, calculated using Equations (3)–(5).
PSNR = 10 l o g ( R 2 M S E )
MSE = [ I 1 ( m , n ) I 2 ( m , n ) ] 2 M N
SSIM = ( 2 μ x μ y + c 1 ) ( 2 σ x y + c 2 ) ( μ x 2 + μ y 2 + c 1 ) ( σ x 2 + σ y 2 + c 2 )
where c 1 = ( K 1 L ) 2 and c 2 = ( K 2 L ) 2 . L is the dynamic range of the pixel value. The constant c 1 and c 2 are used to avoid instability for image regions where the local mean or standard deviation is close to zero. Therefore, a small non-zero value of constant c 1 and c 2 are calculated by setting K 1 = 0.01 , and K 2 = 0.03 . The parameter values can be referred to the work in [16]. The results of PSNR and SSIM values are depicted in Table 2.

3.6. Vessel Enhancement

The final preprocessing step consists of generating a new vessel enhanced image for feature extraction. Vessel enhancement is performed by applying the morphological top-hat transformation. It involves several morphological operations, which are dilation, erosion, and opening. Let f ( x , y ) and B ( u , v ) represent the grayscale image and the structuring element, respectively. The dilation and erosion of f ( x , y ) and B ( u , v ) are denoted by f B and f B respectively and they are defined by the following equations:
f B ( x , y ) = max u , v ( f ( x u , y v ) + B ( u , v ) )
f B ( x , y ) = max u , v ( f ( x u , y v ) + B ( u , v ) )
Based on dilation and erosion operations, opening and closing of f ( x , y ) and B ( u , v ) , indicated as f B and f B are defined as follows:
f B ( x , y ) = ( f B ) B
f B ( x , y ) = ( f B ) B
The opening operation is used to smoothen bright small regions of an image, while the closing operation eliminates small dark holes. There are white and black top-hat (called bottom hat) which are denoted by I W T H and I B T H ,
I W T H = f ( x , y ) ( f B ) ( x , y )
I B T H = ( f B ) ( x , y ) f ( x , y )
where γ is a morphological opening operation using a disc structuring element of ten pixels in radius. Thus, while bright vessel structures are removed, the darker structures remaining after the opening operation become enhanced, for example, blood vessels. This method can extract vessels in non-uniform illumination. The sample result of vessel enhancement operation is shown in Figure 8a. Then, a contrast adjustment is applied to increase the contrast of the blood vessel shown in Figure 8b.
The images used in this study have a bright foreground and dark background, hence the suitable method used is white top-hat transformation. The result of the top-hat transformation illustrated in Figure 8a.

3.7. Feature Extraction

In this study, we have selected three types of feature extractors for characterizing the non-collateral and collateral case as summarized in Table 1. The extracted features are evaluated and analysed to identify the features that can segregate normal and collateral vessels into their classes in the best manner. In total, 40 features per image were extracted.
Table 1. Feature Extraction Techniques.
Sliding window: Sliding window method is another name for block processing. There are two types of block processing approaches; non-overlapping and overlapping. The concept of non-overlapping is by means of dividing an image into a uniform size of blocks. The disadvantage of a non-overlapping block is that the object that we want to extract might be and might not be inside the block or the object separated into different blocks. It will lead to a misinterpretation of the object; for example, normal vessels are interpreted as collateral and otherwise.
The overlapping block processing method on the other hand is more suitable in this case to avoid misinterpretation. In this method, as for example, if the block size is 16 with a pixel shift by 1, then the first block will be the indices of 1:16 in x and y while the second block will be 2:17, and the third block will be 3:18, and so forth. This means that there is an overlap data between neighboring blocks. So, in this case, the number of blocks will be multiplied. However, the drawback of the overlapping block is it is time-consuming depending on the number of pixel shift selected. Figure 5 depicts an image of a sliding window for overlapping blocks. The red square is the window that will move until the very end of the pixel available in the image. In our work, all feature information ranging from texture and shape-based are extracted from each image patch obtained using the overlapping block sliding window method.
Figure 5. The red square represents the windowing slide.

3.8. Classification

Four classification algorithms which are support vector machine (SVM), random forest, decision tree (DET) and K-nearest neighbor (KNN) are used to classify collateral and normal vessels based on the forthy features extracted in the previous process. The SVM in our work uses polynomial kernel of 3 degree hyperparameter. For the random forest, half of input features for a logistic model training is used to sample at each split point. For the decision tree classifier, a standard model is used with maximum of 100 splits. As for the KNN, we use a standardized model consisting of Euclidean distance function with equal-distance weights. The number of neighbors is set to 2. We select the K value by running the KNN algorithm many times with different values of K, The K value that is best to reduce the number errors while maintaining the accuracy is chosen.
The 10-fold cross validation is used to measure the performance of each classifier and to evaluate how well the classifier generalizes the data. A total of 100 data was used for the 10-fold cross validation meanwhile additional 25 data were used for testing after the model is trained.

3.9. Performance Measures

The performance measures used to evaluate and analyze the features and classifiers such as sensitivity, specificity, accuracy, precision, F-Measure, AUC and ROC-AUC curve are presented in this section.
The statistical calculations for the first three performance measures are given as follows,
Sensitivity = T P T P + F N
Specificity = T N T N + F P
Accuracy = T P + T N T P + T N + F P + F N
where T P (True Positive) represents the number of pixels that are classified correctly as collateral vessels, T N (True Negative) represents the number of pixels classified correctly as non-collateral vessels, F P (False Positive) represents the number of pixels belonging to non-collateral vessels classified as collateral vessels, and F N (False Negative) represents the number of pixels belonging to collateral vessels classified as non-collateral vessels.
Another performance measure that is used which is the F-Measure, refers to the harmonic means of the precision and recall, which is calculated as below,
F - Measure = 2 p r e c i s i o n r e c a l l p r e c i s i o n + r e c a l l
where recall refers to the percentage of total relevant results that are correctly classified, and precision refers to the percentage of the relevant results. Both are calculated as follows.
Recall = T P T P + F N
Precision = T P T P + F P
Besides that, we also measured the performance by using the area under the curve—receiver operating characteristics (AUC-ROC). It is one of the most important evaluation metrics for checking any classification model’s performance. It is used in this study to evaluate the quality of classifiers in correctly classifying the vessels into vessel and non-collateral vessel classes. The AUC-ROC curve plots two parameters; TPR (true positive rate) against FPR (false positive rate), to show the performance of the classifiers at all classification thresholds. The TPR and FPR can be defined as follows,
TPR = T P T P + F N = S e n s i t i v i t y
FPR = T P T P + F P = 1 S p e c i f i c i t y

4. Results and Discussion

The results presented in this section are obtained using a dataset consisting of 125 MIP images obtained from CBCT images. Here, we also include results on feature evaluation for classification of vessels.

4.1. HU Conversion and MIP Projection

The regression model is used to estimate the regression coefficient to obtain the most promising image for further analysis. By implementing the linear equation in Equation (1), we can control the parameters of w and m to obtain the output that we desire. The linear function can be used to convert pixel values from the CBCT scanner to CT values using the Equation (2). As illustrated in Figure 6, if a CBCT image is converted to HU without adjusting the brightness component by multiplication with the output of Equation (20), we will lose the information of the blood vessels in MIP image.
Y = 1.0236 Q + 200
where Q indicates pixel value in CBCT.
Figure 6. MIP image (a) without brightness adjustment prior to HU conversion in comparison to (b) with brightness adjustment prior to HU conversion.
MIP image (a) without brightness adjustment prior to HU conversion in comparison to (b) with brightness adjustment prior to HU conversion. As shown in Figure 6b, some vessel information is preserved as compared to that of Figure 6a. Therefore, it is necessary to perform brightness adjustment before converting the image into HU. The conversion is applied for each slice before performing the maximum intensity projection. Then, skull removal will be carried out to obtain MIP images without the skull as shown in Figure 4e.

4.2. Artifact Removal

Five noise filtering methods as given in Table 2, have been applied to the skull removed MIP images to reduce unwanted features/artifacts while smoothing the image. Figure 7 shows the resultant images after performing the noise filtering methods. The performances of the methods are compared based on quality metrics such as peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) as given in Table 2. In the ideal case, the highest PSNR value indicates that the quality of the image is the best as the distortion is the smallest. Similarly for SSIM, when its value is higher or converges to 1, the smaller is the image distortion of the reconstructed image in terms of brightness, contrast and structure respectively. As shown in Figure 7c, the image after Wiener filtering has the least distortion of vessel structure and contains less noise and artifact. However, as tabulated in Table 2, the Wiener filtering method has the lowest PSNR unlike its SSIM value which is the highest. It can also be observed that some details were also lost as a result of filtering using the L 1 and L 2 total variation smoothing methods as shown in Figure 7d and Figure 7e respectively. From these findings, we select the Wiener filtering method that achieved the best SSIM value for smoothed and noise reduced image.
Table 2. Quality metrics.
Figure 7. Five noise reduction methods used, (a) anisotropic diffusion filter, (b) Median filter, (c) Wiener filter, (d) and (e) are L 1 and L 2 total variation smoothing, respectively.

4.3. Vessel Enhancement

Once image is smoothed and noise reduced by Wiener filtering method (see Figure 8a), vessel enhancement is performed to highlight further the vessel structures. To enhance further the vessels in non-uniform background, we employed top-hat transformation which is based on the residual; the difference between two or more common operators.
Figure 8. (a) The result of implementing top-hat transformation on the Wiener filtered image and (b) Contrast adjustment of top-hat transformed image.

4.4. Feature Extraction

Once vessels are further enhanced, features are extracted from the vessels to be used as the input for classifiers to determine the type of vessels by the information given. The feature analysis in this study includes extraction of GLCM, shape, and invariant moments (refer Table 1). In feature extraction, first, we need to obtain image patches that contain either collateral or normal vessels from a filtered and enhanced MIP image e.g., Figure 8b, based on ground truth (labelled as collateral vessels). The aim is to create a library of features that stores all the feature information extracted from these patches.
Figure 9 is the zoomed-in version of the image of Figure 8b on the left hemisphere. The region of interest in our study is the blood vessels hence we marked several locations according to the ground truth image, as highlighted in red dots in Figure 9. These marked ROIs represent the location where image patches were obtained/cropped. The location of patches will be similar for all selected MIP images. In ground truth images, collateral vessels are delineated and labeled by the experts.
Figure 9. Zoomed-in image with marked locations in dotted red.
Samples of image patches obtained from three MIP images are depicted in Figure 10; the first two rows represent normal vessels, while the last two rows represent collateral vessels. It is observed that the brightest blood vessels represent normal vessels because it is denser compared to that of collateral vessels. However, some parts of collateral vessels have similar intensity as to that of the normal vessels. Furthermore, they may also have similar sizes. Hence, the challenge lies in the segregation of collateral and normal vessels that are similar in intensities and sizes. To do this, region descriptors or features need to be extracted from the vessels to provide description/information about the vessels according to its type. In this study, the features as given in Table 1 were extracted from the image patches obtained using sliding window approach (see Figure 5).
Figure 10. A sample of 24 image patches obtained from 3 MIP images.

4.5. Performance of Classifiers

Table 3 and Figure 11 show the results of four classifiers for vessel classification performed on GLCM, Shape and moment invariant features with respect to their performance measures and AUC-ROC curve. Based on the average accuracy results, SVM is the best classifier to classify GLCM feature with 99.92% accuracy, meanwhile random forest is found to be best to classify the moment invariant with 78.04% accuracy, and finally KNN is more suitable the shape feature. We can also see that GLCM is the best feature for the CBCT image modality compared to momentinvariant and shape features in which the accuracy of GLCM is above 97% (97.41–99.92%) whereas the accuracy of moment invariant and shape is much lower which is about 77% with the highest is 79.33%. In terms of sensitivity and specificity, all four classifiers that use GLCM feature have values above 99% except for KNN classifier which obtains slightly lower value about 98.46% sensitivity and 96.40% specificity respectively. Similar trend is also observed for F-measure and AUC value for all classifiers applied on GLCM feature. They perform well when applied to GLCM compared to moment invariant and shape features. Since the AUC is related to the performance of sensitivity and specificity, the SVM for GLCM feature is expected to be the highest among all other classifiers which is 0.9999. SVM also produces the highest AUC value under moment invariant feature although its accuracy is not the highest one under this feature category. In shape feature, KNN has the best AUC value of 0.7964.
Table 3. Classification performance of four feature types.
Figure 11. AUC-ROC curves for (a) Moment invariants, (b) GLCM, and (c) shape features.
We can conclude our finding that for GLCM feature, SVM can provide better predictive ability in generating hyperplane that can optimally separate the collateral from normal vessel classes and GLCM is the best feature for our images despite the small data size.
Figure 11 shows the AUC-ROC curves of each type of features plotted for all four classifiers. We can observe that all features achieved good measure of separability as the AUCs are near to 1. It shows that the classifier models are capable of classifying the collateral and normal vessels into their classes based on these features.
In general, the performance of the classification models varies according to the type of features applied to the models. GLCM is found to be the best feature in our study. We can say that all the classifier models obtain satisfying results in using each feature to distinguish between collateral vessels and non-collateral vessels with SVN outperforms the rest of classifier for GLCM feature, followed by random forest for moment invariant and KNN for the shape feature. The results of our work is also consistent to the No Free Lunch (NFL) Theorem [66] which claims that the performance of classifiers varies according to the types of data set used.The computation time of our proposed work measured from pre-processing until classification is 35 seconds and it is computationally fast in comparison with manual detection technique performed by the clinicians.

5. Conclusions

Detection of collateral vessels in the management of acute ischemic stroke is crucial to assist clinicians in determining the type of treatment suitable for patient. This paper presented the MIP image construction from the CBCT images followed by pre-processing of the MIP images to distinguish vessel network from the background. In noise removal and image smoothing, Wiener filtering method has the highest SSIM value and is found to outperform other filtering methods such as anisotropic diffusion filter, median filter and RTV. However, the PSNR value obtained by this method is the lowest compared to the PSNR value of other filtering methods. Despite the lowest PSNR value, the Wiener filtering can preserve the small vessel structures with low distortion and reduce the noise that causes the non-uniform background. Further enhancement and background removal is achieved by means of multiscale top-hat transformation. Then, important features from several feature descriptors such as GLCM, shape, and moment invariant were extracted to analyze and represent the characteristics of collateral and non-collateral vessels. Altogether, about 40 features were extracted from 125 images. These features serve as the input for the classifiers to determine the type of vessels. In this study, four different classifier models were evaluated and compared namely SVM, KNN, decision tree and random forest. The classifiers were validated by the 10-fold cross-validation strategy running on 100 images and 25 images for testing. The performance for each feature type and classifier model was evaluated in terms of accuracy, sensitivity (recall), specificity, precision, F-measure and AUC values. Our findings show that the best feature for our CBCT image modality is GLCM whereas the SVM is the best classifier for this feature.
Since our proposed detection technique is based on conventional image processing, we anticipate that our technique may work on other medical image modalities such as CT and Magnetic Resonance Angiogram (MRA). Further investigation on the application of the proposed collateral detection on these modalities is our future interest. For future work, we will gather more data and explore deep learning techniques on our new data.

Author Contributions

A.A.A.and A.A. contributed to the main idea and the methodology of the research. A.A. and A.A.A. designed the experiment, performed the simulations and wrote the original manuscript. A.A.A. and L.I.I. contributed significantly in improving the technical and grammatical contents of the manuscript. V.S.A., T.B.T., Z.O. and S.M. reviewed the manuscript and provided valuable suggestions to further refine the manuscript. S.M. has provided his service in identification and validation of the collateral vessels. All authors have read and agreed to the published version of the manuscript.

Funding

This research has been supported by the MURATA Grant under the cost centre 015ME0-185.

Institutional Review Board Statement

Not Applicable.

Data Availability Statement

Not Applicable.

Acknowledgments

The authors acknowledge the support of this research by the MURATA Grant under the cost centre 015ME0-185. This work was also partially supported by the Ministry of Higher Education, Malaysia, under Higher Institution Centre of Excellence (HICoE) Scheme awarded to Centre for Intelligent Signal and Imaging Research (CISIR). The authors would like to thank Sobri Muda from Hospital UPM for providing the medical images.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, Y.; Shafie, A.; Sidek, N.; Aziz, Z. Economic burden of stroke in malaysia: Results from national neurology registry. J. Neurol. Sci. 2017, 381, 167–168. [Google Scholar] [CrossRef]
  2. Galimanis, A.; Jung, S.; Mono, M.L.; Fischer, U.; Findling, O.; Weck, A.; Meier, N.; Marco De Marchis, G.; Brekenfeld, C.; El-Koussy, M.; et al. Endovascular therapy of 623 patients with anterior circulation stroke. Stroke 2012, 43, 1052–1057. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Kucinski, T.; Koch, C.; Eckert, B.; Becker, V.; Krömer, H.; Heesen, C.; Grzyska, U.; Freitag, H.; Röther, J.; Zeumer, H. Collateral circulation is an independent radiological predictor of outcome after thrombolysis in acute ischaemic stroke. Neuroradiology 2003, 45, 11–18. [Google Scholar] [CrossRef] [PubMed]
  4. Bang, O.Y.; Saver, J.L.; Kim, S.J.; Kim, G.M.; Chung, C.S.; Ovbiagele, B.; Lee, K.H.; Liebeskind, D.S. Collateral flow predicts response to endovascular therapy for acute ischemic stroke. Stroke 2011, 42, 693–699. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Bang, O.Y.; Saver, J.L.; Kim, S.J.; Kim, G.M.; Chung, C.S.; Ovbiagele, B.; Lee, K.H.; Liebeskind, D.S.; UCLA-Samsung Stroke Collaborators. Collateral flow averts hemorrhagic transformation after endovascular therapy for acute ischemic stroke. Stroke 2011, 42, 2235–2239. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Shuaib, A.; Butcher, K.; Mohammad, A.A.; Saqqur, M.; Liebeskind, D.S. Collateral blood vessels in acute ischaemic stroke: A potential therapeutic target. Lancet Neurol. 2011, 10, 909–921. [Google Scholar] [CrossRef]
  7. Ibraheem, I. Reduction of artifacts in dental cone beam CT images to improve the three dimensional image reconstruction. J. Biomed. Sci. Eng. 2012, 5, 409–415. [Google Scholar] [CrossRef] [Green Version]
  8. Scarfe, W.C.; Farman, A.G. What is cone-beam CT and how does it work? Dent. Clin. N. Am. 2008, 52, 707–730. [Google Scholar] [CrossRef]
  9. McKinley, R.; Häni, L.; Gralla, J.; El-Koussy, M.; Bauer, S.; Arnold, M.; Fischer, U.; Jung, S.; Mattmann, K.; Reyes, M.; et al. Fully automated stroke tissue estimation using random forest classifiers (FASTER). J. Cereb. Blood Flow Metab. 2017, 37, 2728–2741. [Google Scholar] [CrossRef] [Green Version]
  10. Chiaradia, M.; Radaelli, A.; Campeggi, A.; Bouanane, M.; De La Taille, A.; Kobeiter, H. Automatic three-dimensional detection of prostatic arteries using cone-beam CT during prostatic arterial embolization. J. Vasc. Interv. Radiol. 2015, 26, 413–417. [Google Scholar] [CrossRef]
  11. Schulze, R.; Heil, U.; Groβ, D.; Bruellmann, D.; Dranischnikow, E.; Schwanecke, U.; Schoemer, E. Artefacts in CBCT: A review. Dentomaxillofac. Radiol. 2011, 40, 265–273. [Google Scholar] [CrossRef] [Green Version]
  12. Iqbal, S. A comprehensive study of the anatomical variations of the circle of willis in adult human brains. J. Clin. Diagn. Res. JCDR 2013, 7, 2423. [Google Scholar] [CrossRef] [PubMed]
  13. Akbari, H.; Kosugi, Y.; Kojima, K.; Tanaka, N. Blood vessel detection and artery-vein differentiation using hyperspectral imaging. In Proceedings of the 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Minneapolis, MN, USA, 3–6 September 2009; pp. 1461–1464. [Google Scholar] [CrossRef]
  14. Preethi, M.; Vanithamani, R. Review of retinal blood vessel detection methods for automated diagnosis of Diabetic Retinopathy. In Proceedings of the IEE International Conference on Advances In Engineering, Science And Management (ICAESM-2012), Nagapattinam, India, 30–31 March 2012; pp. 262–265. [Google Scholar]
  15. Thakur, A.; Malik, M.; Phutela, N.; Khare, P.; Mor, P. CBCT image noise reduction and enhancement using Bi-Histogram method with bent activation function. In Proceedings of the International Conference on Information Technology (InCITe)—The Next Generation IT Summit on the Theme-Internet of Things: Connect your Worlds, Noida, India, 6–7 October 2016; pp. 242–246. [Google Scholar]
  16. Liang, X.; Zhang, Z.; Niu, T.; Yu, S.; Wu, S.; Li, Z.; Zhang, H.; Xie, Y. Iterative image-domain ring artifact removal in cone-beam CT. Phys. Med. Biol. 2017, 62, 5276. [Google Scholar] [CrossRef] [PubMed]
  17. Yilmaz, E.; Kayikcioglu, T.; Kayipmaz, S. Noise removal of CBCT images using an adaptive anisotropic diffusion filter. In Proceedings of the 40th International Conference on Telecommunications and Signal Processing (TSP), Barcelona, Spain, 5–7 July 2017; pp. 650–653. [Google Scholar]
  18. Zhang, Y.; Zhang, L.; Zhu, X.R.; Lee, A.K.; Chambers, M.; Dong, L. Reducing metal artifacts in cone-beam CT images by preprocessing projection data. Int. J. Radiat. Oncol. Biol. Phys. 2007, 67, 924–932. [Google Scholar] [CrossRef]
  19. Chen, Y.W.; Duan, G.; Fujita, A.; Hirooka, K.; Ueno, Y. Ring artifacts reduction in cone-beam CT images based on independent component analysis. In Proceedings of the Conference on Instrumentation and Measurement Technology, Singapore, 5–7 May 2009; pp. 1734–1737. [Google Scholar]
  20. Altunbas, C.; Lai, C.J.; Zhong, Y.; Shaw, C.C. Reduction of ring artifacts in CBCT: Detection and correction of pixel gain variations in flat panel detectors. Med. Phys. 2014, 41, 091913. [Google Scholar] [CrossRef]
  21. Yılmaz, E.; Kayıkçıoğlu, T.; Kayıpmaz, S. Experimental comparison of different noise reduction techniques on cone beam computed tomography images. In Proceedings of the 22nd Conference on Signal Processing and Communications Applications (SIU), Trabzon, Turkey, 23–25 April 2014; pp. 2086–2089. [Google Scholar]
  22. Perona, P.; Malik, J. Scale-space and edge detection using anisotropic diffusion. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 629–639. [Google Scholar] [CrossRef] [Green Version]
  23. Crete, F.; Dolmiere, T.; Ladret, P.; Nicolas, M. The blur effect: Perception and estimation with a new no-reference perceptual blur metric. In Human Vision and Electronic Imaging XII; International Society for Optics and Photonics: Bellingham, WA, USA, 2007; Volume 6492, p. 64920I. [Google Scholar]
  24. Huo, Q.; Li, J.; Lu, Y.; Yan, Z. Removing ring artifacts in CBCT images via smoothing. Int. J. Imaging Syst. Technol. 2016, 26, 284–294. [Google Scholar] [CrossRef]
  25. Xu, L.; Yan, Q.; Xia, Y.; Jia, J. Structure extraction from texture via relative total variation. ACM Trans. Graph. TOG 2012, 31, 1–10. [Google Scholar] [CrossRef]
  26. Chen, W.; Prell, D.; Kalender, W. Accelerating ring artifact correction for flat-detector CT using the CUDA framework. In Medical Imaging 2010: Physics of Medical Imaging; International Society for Optics and Photonics: Bellingham, WA, USA, 2010; Volume 7622, p. 76223A. [Google Scholar]
  27. Wei, Z.; Wiebe, S.; Chapman, D. Ring artifacts removal from synchrotron CT image slices. J. Instrum. 2013, 8, C06006. [Google Scholar] [CrossRef]
  28. Dufour, A.; Tankyevych, O.; Naegel, B.; Talbot, H.; Ronse, C.; Baruthio, J.; Dokládal, P.; Passat, N. Filtering and segmentation of 3D angiographic data: Advances based on mathematical morphology. Med. Image Anal. 2013, 17, 147–164. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  29. Truc, P.T.; Khan, M.A.; Lee, Y.K.; Lee, S.; Kim, T.S. Vessel enhancement filter using directional filter bank. Comput. Vis. Image Underst. 2009, 113, 101–112. [Google Scholar] [CrossRef]
  30. Hsu, C.Y.; Schneller, B.; Ghaffari, M.; Alaraj, A.; Linninger, A. Medical image processing for fully integrated subject specific whole brain mesh generation. Technologies 2015, 3, 126–141. [Google Scholar] [CrossRef] [Green Version]
  31. Frangi, A.F.; Niessen, W.J.; Vincken, K.L.; Viergever, M.A. Multiscale vessel enhancement filtering. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Cambridge, MA, USA, 11–13 October 1998; Springer: Berlin/Heidelberg, Germany, 1998; pp. 130–137. [Google Scholar]
  32. Alharbi, S.S.; Sazak, Ç.; Nelson, C.J.; Obara, B. Curvilinear structure enhancement by multiscale top-hat tensor in 2D/3D images. In Proceedings of the International Conference on Bioinformatics and Biomedicine (BIBM), Madrid, Spain, 3–6 December 2018; pp. 814–822. [Google Scholar]
  33. Sun, K.; Chen, Z.; Jiang, S.; Wang, Y. Morphological multiscale enhancement, fuzzy filter and watershed for vascular tree extraction in angiogram. J. Med. Syst. 2011, 35, 811–824. [Google Scholar] [CrossRef]
  34. Bai, X.; Zhou, F. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  35. Medjahed, S.A. A comparative study of feature extraction methods in images classification. Int. J. Image Graph. Signal Process. 2015, 7, 16. [Google Scholar] [CrossRef] [Green Version]
  36. Mishra, S.; Panda, M. Medical image retrieval using self-organising map on texture features. Future Comput. Inform. J. 2018, 3, 359–370. [Google Scholar] [CrossRef]
  37. Takala, V.; Ahonen, T.; Pietikäinen, M. Block-based methods for image retrieval using local binary patterns. In Proceedings of the Scandinavian Conference on Image Analysis, Joensuu, Finland, 19–22 June 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 882–891. [Google Scholar]
  38. Reed, T.R.; Dubuf, J.H. A review of recent texture segmentation and feature extraction techniques. CVGIP Image Underst. 1993, 57, 359–372. [Google Scholar] [CrossRef]
  39. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  40. Soh, L.K.; Tsatsoulis, C. Texture analysis of SAR sea ice imagery using gray level co-occurrence matrices. IEEE Trans. Geosci. Remote Sens. 1999, 37, 780–795. [Google Scholar] [CrossRef] [Green Version]
  41. Ojala, T.; Pietikainen, M.; Maenpaa, T. Multiresolution gray-scale and rotation invariant texture classification with local binary patterns. IEEE Trans. Pattern Anal. Mach. Intell. 2002, 24, 971–987. [Google Scholar] [CrossRef]
  42. Dornaika, F.; Moujahid, A.; El Merabet, Y.; Ruichek, Y. A comparative study of image segmentation algorithms and descriptors for building detection. In Handbook of Neural Computation; Elsevier: Amsterdam, The Netherlands, 2017; pp. 591–606. [Google Scholar]
  43. Gunay, A.; Nabiyev, V.V. Automatic age classification with LBP. In Proceedings of the 23rd International Symposium on Computer and Information Sciences, Istanbul, Turkey, 27–29 October 2008; pp. 1–4. [Google Scholar]
  44. Alhindi, T.J.; Kalra, S.; Ng, K.H.; Afrin, A.; Tizhoosh, H.R. Comparing LBP, HOG and deep features for classification of histopathology images. In Proceedings of the International Joint Conference on Neural Networks (IJCNN), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–7. [Google Scholar]
  45. Clausi, D.A. An analysis of co-occurrence texture statistics as a function of grey level quantization. Can. J. Remote Sens. 2002, 28, 45–62. [Google Scholar] [CrossRef]
  46. Mingqiang, Y.; Kidiyo, K.; Joseph, R. A survey of shape feature extraction techniques. Pattern Recognit. 2008, 15, 43–90. [Google Scholar]
  47. Ranjidha, A.; Ramesh Kumar, A.; Saranya, M. Survey on medical image retrieval based on shape features and relevance vector machine classification. Int. J. Emerg. Trends Technol. Comput. Sci. IJETTCS 2013, 2, 333–339. [Google Scholar]
  48. Chaugule, A.; Mali, S.N. Evaluation of texture and shape features for classification of four paddy varieties. J. Eng. 2014, 2014. [Google Scholar] [CrossRef] [Green Version]
  49. Hu, M.K. Visual pattern recognition by moment invariants. IRE Trans. Inf. Theory 1962, 8, 179–187. [Google Scholar]
  50. Nixon, M.; Aguado, A. Feature Extraction and Image Processing for Computer Vision; Academic Press: Cambridge, MA, USA, 2019. [Google Scholar]
  51. Jain, A. Fundamentals of Digital Image Processing; Prentice Hall of India Private Limited: Upper Saddle River, NJ, USA, 1995. [Google Scholar]
  52. Steven, L.E.; Rafael, C.G.; Richard, E.W. Digital Image Processing Using Matlab; Princeton Hall Pearson Education: Upper Saddle River, NJ, USA, 2004. [Google Scholar]
  53. Altman, N.S. An introduction to kernel and nearest-neighbor nonparametric regression. Am. Stat. 1992, 46, 175–185. [Google Scholar]
  54. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  55. Amelio, L.; Amelio, A. Classification methods in image analysis with a special focus on medical analytics. In Machine Learning Paradigms; Springer: Berlin/Heidelberg, Germany, 2019; pp. 31–69. [Google Scholar]
  56. Nardelli, P.; Jimenez-Carretero, D.; Bermejo-Pelaez, D.; Washko, G.R.; Rahaghi, F.N.; Ledesma-Carbayo, M.J.; San José Estépar, R. Pulmonary Artery–Vein Classification in CT Images Using Deep Learning. IEEE Trans. Med. Imaging 2018, 37, 2428–2440. [Google Scholar] [CrossRef]
  57. Girard, F.; Cheriet, F. Artery/vein classification in fundus images using CNN and likelihood score propagation. In Proceedings of the 2017 IEEE Global Conference on Signal and Information Processing (GlobalSIP), Montreal, QC, Canada, 14–16 November 2017; pp. 720–724. [Google Scholar] [CrossRef]
  58. Montagne, C.; Kodewitz, A.; Vigneron, V.; Giraud, V.; Lelandais, S. 3D Local Binary Pattern for PET image classification by SVM, Application to early Alzheimer disease diagnosis. In Proceedings of the 6th International Conference on Bio-Inspired Systems and Signal Processing (BIOSIGNALS 2013), Barcelona, Spain, 11–14 February 2013; pp. 145–150. [Google Scholar]
  59. Xiao, Z.; Ding, Y.; Lan, T.; Zhang, C.; Luo, C.; Qin, Z. Brain MR image classification for Alzheimer’s disease diagnosis based on multifeature fusion. Comput. Math. Methods Med. 2017, 2017, 1952373. [Google Scholar] [CrossRef] [PubMed]
  60. Yang, F.; Hamit, M.; Yan, C.B.; Yao, J.; Kutluk, A.; Kong, X.M.; Zhang, S.X. Feature extraction and classification on esophageal X-ray images of Xinjiang Kazak nationality. J. Healthc. Eng. 2017, 2017. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  61. Hua, Y.; Nackaerts, O.; Duyck, J.; Maes, F.; Jacobs, R. Bone quality assessment based on cone beam computed tomography imaging. Clin. Oral Implant. Res. 2009, 20, 767–771. [Google Scholar] [CrossRef] [PubMed]
  62. Mah, P.; Reeves, T.; McDavid, W. Deriving Hounsfield units using grey levels in cone beam computed tomography. Dentomaxillofac. Radiol. 2010, 39, 323–335. [Google Scholar] [CrossRef] [PubMed]
  63. Chindasombatjaroen, J.; Kakimoto, N.; Shimamoto, H.; Murakami, S.; Furukawa, S. Correlation between pixel values in a cone-beam computed tomographic scanner and the computed tomographic values in a multidetector row computed tomographic scanner. J. Comput. Assist. Tomogr. 2011, 35, 662–665. [Google Scholar] [CrossRef] [PubMed]
  64. Zohra, F.T.; Gavrilov, A.D.; Duran, O.Z.; Gavrilova, M. A linear regression model for estimating facial image quality. In Proceedings of the 16th International Conference on Cognitive Informatics & Cognitive Computing (ICCI*CC), Oxford, UK, 26–28 July 2017; pp. 130–138. [Google Scholar]
  65. Sharma, K.; Soni, H.; Agarwal, K. Lung Cancer Detection in CT Scans of Patients Using Image Processing and Machine Learning Technique. In Advanced Computational and Communication Paradigms; Springer: Berlin/Heidelberg, Germany, 2018; pp. 336–344. [Google Scholar]
  66. García-Martinez, C.; Rodriguez, F.J.; Lozano, M. Analysing the significance of no free lunch theorems on the set of real-world binary problems. In Proceedings of the 11th International Conference on Intelligent Systems Design and Applications, Cordoba, Spain, 22–24 November 2011; pp. 344–349. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.