1. Introduction
Coronary heart disease is the main cause of morbidity all over the world [
1]. Consequently, it is highly important for coronary stenosis to be detected and diagnosed by cardiologists and addressed in computational science. Nowadays, X-ray coronary angiography is the main source of decision making in stenosis diagnosis. In order to detect coronary stenosis, a specialist performs an exhaustive visual examination of the entire angiogram, and based on their knowledge, the stenosis regions are labeled. In order to illustrate the challenging and laborious task carried out by the specialist in terms of the visual examination of coronary angiograms, in
Figure 1, a set of X-ray angiograms along with manually detected stenosis regions is presented.
The main disadvantages of working with X-ray coronary angiograms are the high noise levels and low-contrast regions, which make automatic vessel identification, measurement and classification tasks difficult. Moreover, for the automatic stenosis classification problem, some approaches have been reported. The method proposed by Saad [
2] detects the presence of atherosclerosis in a coronary artery image using a vessel-width variation measure. The measurements are computed from a previously segmented image containing only vessel pixels and its corresponding skeleton in order to determine the vessel center line, from which the orthogonal line length of a fixed-size window is computed, moving through the image. Kishore and Jayanthi [
3] applied a manually fixed-size window from an enhanced image. The vessel pixels were measured, adding them to intensity values in order to obtain a coronary stenosis grading measure. Other approaches make use of the Hessian matrix properties to enhance or extract vessel trees at the first stage. For instance, the works of Wan et al. [
4], Sameh et al. [
5], and Cervantes-Sanchez et al. [
6] applied the Hessian matrix properties in order to enhance vessel pixels in coronary angiograms. The response image allows for the measurement and extraction of features related with vessel shapes that are used for the automatic classification and grading of coronary stenosis.
The use of classification techniques and search metaheuristics are additional approaches that have been used to address vessel disease problems. Cervantes-Sanchez et al. [
7] proposed a Bayesian-based method using a
feature vector that was extracted from the image histogram in order to classify stenosis cases. Taki et al. [
8] achieved a competitive result in the categorization of calcified and noncalcified coronary artery plaques using a Bayesian-based classifier. The proposal of Welikala et al. [
9] works with retinal vessels, applying a genetic algorithm to reduce the number of needed features that perform a correct classification of proliferative diabetic retinopathy cases. Sreng et al. [
10], proposed a hybrid simulated annealing method to select relevant features that are then used with an ensemble bagging classifier in order to produce a suitable screening of the eye. The method of Chen et al. [
11] works with a
feature vector related with the morphology of bifurcated vessels in order to detect coronary artery disease. A fuzzy criterion was used by Giannoglou et al. [
12] in order to select features in characterization of atherosclerotic plaques, and Wosiak and Zakrzewska [
13], proposed an automatic feature selection method by integrating correlation and clustering strategies in cardiovascular disease diagnosis.
On the other hand, emergent evolution of deep learning techniques such as the convolutional neural network (CNN) have made it possible for them to be applied to the coronary artery disease problem [
14]. CNN contains a set of layers (convolutional layers) focused on the automatic segmentation of the image in order to keep only the data that allow the CNN to achieve correct classification rates [
15,
16,
17]. Antczak and Liberadzki [
18], proposed a method that is able to generate synthetic coronary stenosis and nonstenosis patches in order to improve the CNN training rates. The strategy proposed by Ovalle et al. [
19] makes use of a transfer learning [
20,
21,
22] strategy in order to achieve correct training and classification rates with complex CNN architectures. One of the main drawbacks of CNN is the need for large-instance databases in order to achieve correct training rates. Data augmentation techniques [
23,
24,
25,
26,
27] are commonly used as a way to generate large amounts of instances that are used in the training and testing of the CNN. In addition, it is difficult for a CNN to identify which features are really useful for a correct classification and what they represent [
28,
29].
In the present paper, a novel method for the automatic classification of coronary stenosis based on feature selection and a hybrid evolutionary algorithm in X-ray angiograms is presented. The proposed method uses the evolutionary computation technique for addressing the high-dimensional problem of selecting an efficient subset of features from a bank of 49 features, where the problem is a computational complexity of O
. To select the best evolutionary technique, different population-based strategies were compared in terms of feature reduction and classification accuracy using a training set of coronary stenosis images. From the comparative analysis, the Boltzmann univariate marginal distribution algorithm (BUMDA) and simulated annealing (SA) were selected for further analysis. In the experiments, two different databases were used. The first database was provided by the Mexican Institute of Social Security (IMSS), and it contains 500 images. The second database corresponds to Antczak [
18], which is publicly available and contains 2700 patches. Finally, the proposed method was compared with different state-of-the-art classification methods in terms of classification accuracy and Jaccard coefficient, working with both databases in order to show the classification robustness achieved by the subset of 11 features, which were obtained from the feature selection step using the hybrid BUMDA-SA evolutionary technique.
The remaining of this paper is as follows. In
Section 2, the background methodology is described.
Section 3 presents the proposed method and the hybrid approach that performs the automatic feature selection task. In
Section 4, the experiment details and results are described, and finally, conclusions are given in
Section 5.
3. Proposed Method
The proposed method consists of the steps of feature extraction, automatic feature selection, feature subset testing and performance evaluation. The first stage is focused in the extraction of 49 features from the image database by considering texture, intensity and morphology feature types. The extracted texture features were those proposed by Haralik [
59], and the morphological features were based on Welikala [
9]. The bank of 49 features is described below.
- 1.
The minimum pixel intensity present in the patch.
- 2.
The maximum pixel intensity present in the patch.
- 3.
The mean pixel intensity in the patch.
- 4.
The standard deviation of the pixel intensities in the patch.
- 5–18.
Features 5 to 18 are composed of the Haralik features: angular second moment (energy), contrast, correlation, variance, inverse difference moment (homogeneity), sum average, sum variance, sum entropy, entropy, difference variance, difference entropy, information measure of correlation 1, information measure of correlation 2, maximum correlation coefficient.
- 19–32.
The Haralik features applied to the Radon transform response of the patch: angular second moment (energy), contrast, correlation, variance, inverse difference moment (homogeneity), sum average, sum variance, sum entropy, entropy, difference variance, difference entropy, information measure of correlation 1, information measure of correlation 2, maximum correlation coefficient.
- 33.
The Radon ratio-X measure.
- 34.
The Radon ratio-Y measure.
- 35.
The mean of pixel intensities from the Radon transform response of the patch.
- 36.
The standard deviation of the pixel intensities from the Radon transform response of the patch.
- 37.
The vessel pixel count in the patch.
- 38.
The vessel segment count in the patch.
- 39.
Vessel density. The rate of vessel pixels in the patch.
- 40.
Tortuosity 1. The tortuosity of each segment is calculated using the true length (measured with the chain code) divided by the Euclidean length. The mean tortuosity is calculated from all the segments within the patch.
- 41.
Sum of vessel lengths.
- 42.
Number of bifurcation points. The number of bifurcation points within the patch when vessel segments were extracted.
- 43.
Gray-level coefficient of variation. The ratio of the standard deviation to the mean of the gray level of all segment pixels within the patch.
- 44.
Gradient mean. The mean gradient magnitude along all segment pixels within the subwindow, calculated using the Sobel gradient operator applied on the preprocessed image.
- 45.
Gradient coefficient of variation. The ratio of the standard deviation to the mean of the gradient of all segment pixels within the subwindow.
- 46.
Mean vessel width. Skeletonization correlates to vessel center lines. The distance from the segment pixel to the closest boundary point of the vessel using the vessel map prior to skeletonization. This gives the half-width at that point, which is then multiplied by 2 to achieve the full vessel width. The mean is calculated for all segment pixels within the subwindow.
- 47.
The minimum standard deviation of the vessel length, based on the vessel segments present in the patch. The segments are obtained by the tortuosity points along the vessel.
- 48.
The maximum standard deviation of the vessel length, based on the vessel segments present in the patch. The segments are obtained by the tortuosity points along the vessel.
- 49.
The mean of the standard deviations of the vessel length, based on the vessel segments present in the patch. The segments are obtained by the tortuosity points along the vessel.
In
Figure 2, the overall hybrid evolutionary proposed method steps are illustrated.
In the second step, the automatic feature selection task is performed. It is driven by the proposed hybrid evolutionary strategy involving the BUMDA and SA techniques. In this stage, BUMDA is initialized and iterated until the maximum number of generations is reached. In the third step, the selected feature subset is tested using testing cases, and finally, the obtained classification results are measured based on the accuracy and Jaccard coefficient metrics in order to evaluate their performance.
In
Figure 3, the hybrid evolutionary strategy is described in detail. This stage of the proposed method is focused on the automatic feature selection task. It starts with the BUMDA initialization, requiring only the max number of generations and the population size. With BUMDA being a population-based technique, it produces a set of solutions on each iteration. Each solution indicates which features will be used and which will be discarded. Consequently, for each solution, a particular SVM is trained using only the feature subset expressed in the solution. On each BUMDA generation, different SVMs are trained according to each feature vector, which is represented by each individual in the BUMDA population. Based on the SVM training accuracy and the number of selected features, the best individual in each generation is selected. In the next step, the previously selected individual is improved by the SA strategy. Since SA is a single-solution technique, it is useful to improve the best solution produced by the BUMDA. If the SA-obtained result is higher than the best result obtained by BUMDA, its best individual is replaced by the individual improved by the SA. When the max number of BUMDA generations is reached, the individual with the highest fitness value over all generations is selected as the best solution achieved. This solution contains the selected feature subset, which will be directly applied on the test set of coronary stenosis images. In this stage, the use of a hybrid evolutionary strategy based on the BUMDA and SA techniques is relevant, because SA helps to further reduce the number of features represented in the best solution achieved in each BUMDA generation, at the same time keeping the training accuracy rate, or even improving it.
For the experiment, two different image databases were used. The first database was provided by the Mexican Social Security Institute (IMSS) and approved by a local committee under reference R-2019-1001-078. It contains 500 coronary image patches, with a proportion of
for positive and negative stenosis cases. From this database, 400 instances were used for the automatic feature selection stage and the remaining 100 instances were used for testing after this stage ends. All patch sizes were
pixels and were validated by a cardiologist.
Figure 4 illustrates sample patches of the IMSS database with their respective vessel segmentation response and skeleton, according to the Frangi method, from which the morphological-based feature extraction task was performed.
The second database was provided by Antczak [
18], which is publicly available. It contains 2700 instances, which are also balanced for positive and negative stenosis cases. From this database, 2160 instances
were used for training and the remaining 540 instances were used for testing.
Figure 5 illustrates sample patches of the Antczak database with their respective vessel segmentation response and skeleton, according with the Frangi method.
In order to evaluate the performance of the proposed method, the accuracy metric (
Acc) and the Jaccard similarity coefficient (
) were adopted. The accuracy metric considers the fraction of correct classified cases as positive or negative by defining four necessary measures: true-positive cases (
), true-negative cases (
), false-positive cases (
) and false-negative cases. The
value is the fraction of positive cases classified correctly. The
value is the fraction of negative cases classified correctly. The
cases is the fraction of negative cases classified as positive. The
value is the fraction of positive cases classified as negative. Based on this, the accuracy is computed as follows:
The
measures the similarity of two element sets. Applying this principle, it is possible to measure the accuracy of a classifier using only positive instances, as follows:
It is important to mention that only the IMSS database was used for the automatic feature selection stage. Furthermore, with the Antczak database, only the feature subset obtained by the proposed method was used in order to probe the method’s effectiveness. Additionally, classic search techniques from the literature, such as the Tabú search (TS) [
60] and the iterated local search (ILS) [
61], were also included in the experimentation. For the hybrid approaches, the simulated annealing strategy was used in all experiments in order to improve the best solution achieved for each particular technique.
4. Results and Discussion
In this section, the proposed method for feature selection and automatic classification is evaluated with different state-of-the-art methods using two databases of X-ray angiograms. All the experiments were performed using the Matlab software version 2018 on a computer with an Intel core i7 processor with 8 GB of RAM.
Table 1 describes the parameter settings of the compared methods used in the automatic feature selection stage, considering the same conditions for all of them in order to avoid biased measurements.
The SA strategy was configured with , and . In order to ensure the obtained results, the proposed method was performed with 30 independent trials. For the SVM, 1000 max iterations were established using a cross-validation with . The parameter values for all techniques described previously were set taking into account the tradeoff between the classification accuracy and the execution time required to achieve a solution.
In
Table 2, a comparative analysis related to the best results obtained by different strategies during the automatic feature selection stage is presented.
Based on the results described in
Table 2, the SVM training efficiency was improved in almost all cases when only a feature subset was used instead of the full set with 49 features. This behavior is because of the difficulty in projecting a high amount of overlapped data to dimensional orders higher than 49-
D. Consequently, the proposed method achieved the best result since only 11 of 49 features were selected. This means that
of the initial feature set was discriminated, achieving a training rate efficiency of
at the same time in terms of the accuracy metric. In addition, some of the compared methods presented important variations on the best solution achieved according to the standard deviation accuracy, which gives some evidence of possible local-optima falls in some of the trials. In contrast, the standard deviation for the accuracy of the proposed method was lower, and considering the tradeoff between all measured factors, such as number of selected features, max training accuracy, mean training accuracy and standard deviation accuracy, the proposed method achieved the highest score.
After the automatic feature selection process was performed, in the next stage, the best feature subset, which was achieved by the hybrid BUMDA-SA method, was tested using the test cases from the IMSS and the Antczak databases, separately.
Table 3 contains the corresponding confusion matrix, from which the accuracy and Jaccard coefficient metrics are computed.
In
Table 4, a comparative analysis between the proposed method and different state-of-the-art methods is presented, using the test set of 100 images of the IMSS database. The results of the proposed method are described based on the confusion matrix presented in
Table 3.
According to the data presented in
Table 4, the proposed method achieved the highest classification rate in terms of the accuracy and Jaccard coefficient metrics, whose values were
and
, respectively. By contrasting the accuracy in the training and testing stages, there is evidence of variation rates corresponding to the compared strategies. In
Figure 6, the variation differences in accuracy in the training and testing stages for the contrasted strategies are illustrated.
The values of the accuracy rates show how competitive the feature subset was at classifying stenosis cases. It is remarkable how some feature subsets give evidence of possible overfitting training, such as the Hybrid-TS, since the training accuracy was against when using testing cases. In contrast, the proposed method achieved a low difference in performance in the training and testing phases compared with the Hybrid-TS and the Hybrid-GA techniques, which was , indicating that the achieved subset with 11 features is highly suitable for the classification task.
In order to evaluate the subset of 11 features achieved by the proposed hybrid BUMDA-SA method in the automatic feature selection stage, the Antczak database was used. In
Table 5, the confusion matrix obtained from the proposed method using the Antczak database is presented.
On the other hand, in
Table 6, the results obtained by the proposed method and different strategies in the testing stage using the Antczak database are presented.
Based on the results presented in
Table 6, the highest accuracy and Jaccard coefficient rates were achieved with the proposed method, whose values were
and
respectively. This result is relevant to show that the feature subset found by the proposed hybrid BUMDA-SA method is suitable. Consequently, it is important to mention that according to the results presented in
Table 4 and
Table 6, the GA and UMDA techniques achieved very closed accuracy rates in contrast with the proposed method. However, the reduction rate, which is related to the number of selected features, was overcome by the proposed method when it was contrasted with the other methods. In addition, the proposed method also achieved the highest Jaccard coefficient rate compared with the others. It is important to mention these findings in order to show the importance of the use of a hybrid strategy in this multiobjective problem, where is required to keep or improve a high accuracy rate in the classification task, and at the same time, the use of a minimum number of features.
Finally, in
Table 7, the set of 11 features obtained by the proposed method is described, along with the frequency selection rate obtained from the statistical analysis of the 30 independent runs.
According to the results presented in
Table 7, the Min, Sum Average and Sum Variance features, which correspond to intensity and texture, have the highest frequency selection rates, followed by the Bifurcation Points and Radon-Sum features, which corresponds to morphology and Radon-based texture, respectively. This analysis is relevant since it allows us to remark on the importance of performing the automatic feature selection process with a high number of features involving different feature types such as texture, intensity and morphology.
5. Conclusions
In this paper, a novel method for the automatic classification of coronary stenosis in X-ray angiograms was introduced. The method is based on feature selection using a hybrid evolutionary algorithm and a support vector machine for classification. The hybrid method was used to explore the high-dimensional search space O of a bank of 49 features involving properties of intensity, texture, and morphology. To determine the best evolutionary method, a comparative analysis in terms of feature reduction rate and classification accuracy was performed using a training set of X-ray images. From the analysis, the method using BUMDA and SA achieved the best performance, selecting a subset of 11 features and achieving a feature reduction rate of , and a classification accuracy of . Moreover, two different databases of coronary stenosis were used; the first one was provided by the Mexican Social Security Institute (IMSS), containing 500 images; and the second database is publicly available, with 2700 patches. In the experimental results, the proposed method, using the set of 11 selected features, was compared with different state-of-the-art classification methods, achieving an accuracy and Jaccard coefficient of and in the first database and and in the second one, respectively. Finally, it is important to point out that considering the execution time obtained by the proposed method when testing images (0.02 s per image), the proposed method can be useful in assisting cardiologists in clinical practice or as part of a computer-aided diagnostic system.