Next Article in Journal
Life Aspirations, Generativity and Compulsive Buying in University Students
Next Article in Special Issue
Dynamical Analysis of Universal Masking on the Pandemic
Previous Article in Journal
Influence of Physical Activity and Socio-Economic Status on Depression and Anxiety Symptoms in Patients after Stroke
Previous Article in Special Issue
Automatic COVID-19 Detection Using Exemplar Hybrid Deep Features with X-ray Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fusion of Higher Order Spectra and Texture Extraction Methods for Automated Stroke Severity Classification with MRI Images

1
Department of Engineering and Mathematics, Sheffield Hallam University, Sheffield S1 1WB, UK
2
School of Electronics and Computer Engineering, Ngee Ann Polytechnic, Singapore 599489, Singapore
3
School of Electronics Engineering, Kalinga Institute of Industrial Technology, Bhubaneswar, Odisha 751024, India
4
Department of Medicine-Cardiology, Columbia University, New York, NY 10027, USA
5
Sheffield Institute for Translational Neuroscience, University of Sheffield, Sheffield S10 2HQ, UK
6
Sheffield Teaching Hospitals NIHR Biomedical Research Centre, Sheffield S10 2JF, UK
7
Liverpool Centre for Cardiovascular Science, University of Liverpool and Liverpool Heart & Chest Hospital, Liverpool L69 7TX, UK
8
Aalborg Thrombosis Research Unit, Department of Clinical Medicine, Aalborg University, 9000 Aalborg, Denmark
9
School of Science and Technology, Singapore University of Social Sciences, 463 Clementi Road, Singapore 599494, Singapore
10
Department of Bioinformatics and Medical Engineering, Asia University, Taichung 41354, Taiwan
11
International Research Organization for Advanced Science and Technology (IROAST), Kumamoto University, Kumamoto 860-8555, Japan
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2021, 18(15), 8059; https://doi.org/10.3390/ijerph18158059
Submission received: 18 May 2021 / Revised: 5 July 2021 / Accepted: 23 July 2021 / Published: 29 July 2021

Abstract

:
This paper presents a scientific foundation for automated stroke severity classification. We have constructed and assessed a system which extracts diagnostically relevant information from Magnetic Resonance Imaging (MRI) images. The design was based on 267 images that show the brain from individual subjects after stroke. They were labeled as either Lacunar Syndrome (LACS), Partial Anterior Circulation Syndrome (PACS), or Total Anterior Circulation Stroke (TACS). The labels indicate different physiological processes which manifest themselves in distinct image texture. The processing system was tasked with extracting texture information that could be used to classify a brain MRI image from a stroke survivor into either LACS, PACS, or TACS. We analyzed 6475 features that were obtained with Gray-Level Run Length Matrix (GLRLM), Higher Order Spectra (HOS), as well as a combination of Discrete Wavelet Transform (DWT) and Gray-Level Co-occurrence Matrix (GLCM) methods. The resulting features were ranked based on the p-value extracted with the Analysis Of Variance (ANOVA) algorithm. The ranked features were used to train and test four types of Support Vector Machine (SVM) classification algorithms according to the rules of 10-fold cross-validation. We found that SVM with Radial Basis Function (RBF) kernel achieves: Accuracy (ACC) = 93.62%, Specificity (SPE) = 95.91%, Sensitivity (SEN) = 92.44%, and Dice-score = 0.95. These results indicate that computer aided stroke severity diagnosis support is possible. Such systems might lead to progress in stroke diagnosis by enabling healthcare professionals to improve diagnosis and management of stroke patients with the same resources.

1. Introduction

Cerebrovascular accident, commonly known as stroke, is a major cause of death and chronic disability on a global scale [1,2,3]. Cerebral ischemia causes approximately 80% of strokes [4]. Energy depletion and cell death might lead to ischemic brain injuries [5]. These injuries cause diminished functionality of the affected nerve cells, which leads to serious long-term disability. Brain imaging, such as Computed Tomography (CT) and Magnetic Resonance Imaging (MRI), is an important tool for ischemic stroke assessment [6,7]. Trained professionals use brain images to determine the affected lesion and classify the stroke into one of three types: Partial Anterior Circulation Syndrome (PACS), Lacunar Syndrome (LACS), and Total Anterior Circulation Stroke (TACS) [8,9,10]. PACS is a cortical stroke and affects the middle/anterior cerebral artery. A blockage in blood vessels which supply deep brain structures leads to LACS. TACS affects the middle/anterior cerebral areas which are supplied by middle and anterior cerebral arteries [11]. Discriminating between PACS, LACS and TACS yields important information for prognosis and treatment planning [12,13,14,15,16]. Treatment for ischaemic stroke exists in the form of intravenous fibrinolysis, intra-arterial therapies, and mechanical revascularization [17]. Currently, human experts use their knowledge on how the physiology of different stroke types is reflected in image texture to classify stroke severity [10]. Cardiologists use textural information to identify extent, shape, and location of lesions. This information forms the basis for stroke severity classification. Unfortunately, manual image analysis is laborious [10], and prone to inter- as well as intra-operator variability [18,19]. Furthermore, expert analysis is time-consuming [16] which drives up the diagnosis cost. Automated lesion identification and subsequent stroke severity classification can considerably reduce the delineation times and detect lesions accurately [16]. Therefore, the development of computer aided diagnostic systems based on automated brain lesion detection after stroke is an active field of research.
Automated stroke severity classification systems are linked to specific data sources. To differentiate PACS, LACS and TACS in a practical setting, the choice comes down to CT or MRI imaging. Both methods yield a graphical representation of the human brain which contains distinct image objects. Identifying these objects through image segmentation is an important step to extract diagnostically important information. When compared to MRI, CT is more widely used because it is faster and less expensive. However, MRI has a much higher sensitivity for acute ischemic lesions [20]. In addition, MRI examinations can be extended by adding functional information to the anatomical data to form Diffusion-Weighted Imaging (DWI). Diffusion refers to the random Brownian motion of water molecules driven by thermal energy [21]. Acute cerebral ischemic lesions, even in their early stage, have a different water proportion when compared to normal brain tissue. Hence, DWI imaging is more sensitive than other MRI modalities to small water diffusion changes in the acute ischemic brain, especially within 48 hours of the ictus [22,23,24,25,26]. Hence, MRI images are a good data source for automated and semi-automated lesion delineation. Early attempts focused on identifying abnormal voxels in individual T1-weighted MRI scans through Voxel-Based Morphometry (VBM) [27]. Although VBM may be capable of highlighting structural abnormalities that can facilitate lesion delineation by expert raters, it lacks the spatial resolution and statistical power to provide a true replacement for manual lesion delineation [28]. Alternatively, several semi-automated and fully automated procedures for lesion identification have been proposed [28,29,30,31,32,33,34,35,36,37,38,39,40,41]. Rekik et al. have summarized and discussed current research on ischemic stroke lesions in terms of prediction and insights into dynamic evolution simulation models [42]. Lesion identification through segmentation is the first step towards stroke severity classification. However, questions remain as to whether there is sufficient understanding to automate the information extraction processes needed to establish stroke severity. With this question, we shift the focus from fundamental research towards more practical considerations which relate to computer aided stroke severity diagnosis.
In this paper, we test the hypothesis that MRI images of the brain hold computer extractible information to determine ischemic stroke severity. To do so, we extract, quantify, and assess diagnostically relevant information that can assist with automated stroke severity classification. This work is based on the idea that precise identification of structural brain damage in lesion areas is essential to quantify and indeed understand the brain abnormality caused by stroke. The precision requirement led us to choose MRI as the imaging modality. Our work is based on 267 MRI brain images taken from patients who suffered either LACS, PACS, or TACS. Adaptive Synthetic (AdaSyn) was used to balance the class specific feature sets. The information extraction was done with a combination of Discrete Wavelet Transform (DWT) and Gray-Level Co-occurrence Matrix (GLCM) as well as the standalone methods of Gray-Level Run Length Matrix (GLRLM), and Higher Order Spectra (HOS). The analysis methods resulted in 6526 parameters, used as features. The features were assessed in terms of their ability to discriminate LACS, PACS, and TACS. The assessment was done with statistical analysis and classification performance optimization. First, we used the Analysis Of Variance (ANOVA) method to rank the features in terms of their ability to discriminate the disease classes. With these results, the classification process uses 10-fold cross-validation to establish three assessment criteria: (1) classification algorithm, (2) performance, and (3) number of features. As classification algorithm we tested Support Vector Machine (SVM) with different kernel configurations. By averaging the performance across all folds, we found that SVM with Radial Basis Function (RBF) kernel showed the best result: 0.95 Dice score, 93.62% Accuracy (ACC), f1-score of 95.00%. Based on the observation space of 267 brain MRI images, we failed to reject our initial hypothesis. Hence, the results support the idea that automated stroke severity classification is possible. This scientific foundation justifies further investigations on a larger observation space. With such a study we might be able to determine the practical feasibility of automated stroke severity classification by focusing on human factor ergonomics during the diagnostic process. Careful analysis could lead to the understanding that there is scope for a stroke severity classification tool which creates an environment that allows human experts and machine algorithms to work cooperatively on the diagnostic task.
We have structured the remainder of the manuscript as follows. The Section 2 introduces both feature generation and classification methods. The features, extracted from MRI images, were used to train and test the classification methods. Section 3 details the test results. In the Discussion section, we compare our work to previous studies, and we state where we could improve and complement these studies. The conclusion section summarizes our efforts to support the thesis that it is possible to automate the extraction of diagnostically relevant information from MRI images.

2. Methods

This section describes the methods used to construct and test a system which extracts diagnostically relevant information from MRI images of the brain. More specifically, we are interested to establish the diagnostic support quality for differentiating LACS, PACS, and TACS. We have addressed this problem with a state-of-the-art design strategy which involved pre-processing, feature extraction, data augmentation, feature ranking, and classification. Figure 1 shows an overview block diagram which indicates how the data flows through the individual processing methods. We have distilled diagnostically relevant information from the pre-processed MRI images with three distinct feature extraction methods. In the first phase GLCM, GLRLM, and HOS methods were used directly on the pre-processed images. In the second phase, the same three methods were used to extract features from different levels of DWT decomposition. After feature ranking, four different implementations of the SVM classification algorithm were used to assess the ability of the features to discriminate LACS, PACS, and TACS. The classifiers differ in terms of their kernel methods; one RBF- and three polynomial-kernels were used. The following sections introduce the processing steps in more detail.

2.1. Image Dataset

We have considered 267 slices (PACS 222, LACS 18, and TACS 27) of brain MRI datasets collected from ischemic stroke patients at the Institute of Medical Science and SUM Hospital, Bhubaneswar, Odisha, India. The data was anonymised such that even basic demographics, like age and gender of the patients, are not available. Experts at the hospital selected a single diffusion weighted slice from all the MRI images that were taken during one imaging session from a stroke patient. The selected slices show the stroke lesion and they were used by human experts to support the diagnosis of PACS, LACS, and TACS, on the basis of the clinical features of the patients. The MRI images of the affected brain were acquired with the Signa HDxT 1.5 T Optima Edition machine (GE Healthcare, Waukesha, WI, USA). The first row of images in Figure 2 shows three example MRI images depicting LACS, PACS, and TACS.

2.2. Image Resizing and Contrast Limited Adaptive Histogram Equalization

In a first processing step, we resized all 267 MRI images to 200 × 200 pixels. Subsequently, CLAHE was used to increase the difference in luminance which improves visual feature detection. That was necessary because the brain tissue, shown in the MRI images, lacks contrast, which makes it more difficult to distinguish texture morphology. CLAHE is based on histogram equalization, which can help to increase the prominence of textural changes caused by lesions, via contrast enhancement [43]. Adaptive histogram equalization starts with partitioning an image into nonoverlapping 8 × 8 pixel segments known as tiles. Each tile is enhanced so that the resulting histogram matches a uniform distribution. That process is governed by the contrast enhancement limit (0.01), which prevents over-saturation in uniform image regions. Manipulating the tiles independently might introduce artificial boundaries between them. Therefore, in a final processing step, we used bilinear interpolation to smooth the boundaries. The second row of images in Figure 2 shows the example MRI images after CLAHE.

2.3. Adaptive Synthetic Sampling

The image acquisition process resulted in an imbalanced dataset where 83.15% of the images came from PACS, 10.11% came from LACS, and only 6.74% came from TACS. The imbalance is a problem for classification algorithms, because it introduces bias towards the larger set [44]. For example, a decision support system can score 83.15% accuracy by classifying all input as PACS. Even without that dump approach, it is difficult for the learning algorithm to focus on the two minority classes, because this conflicts with the generalization requirement. To be specific, an adequate representation of the minority classes results in overtraining the network, which causes a deterioration of the classification performance for unknown data. The ability to infer the correct label from unknown data is impeded when the training process focuses too much on the individual cases.
Obtaining more data from the minority classes would be the best way of solving that problem. Unfortunately, in this case it is not possible because the data was obtained from a medical study, and the results reflect the distribution of stroke severity in patients. Clearly, even when more MRI scans are conducted, it is expected that the distribution of PACS, LACS, and TACS is similar to the original dataset. Hence, the problem must be solved on the data processing level. There are two different approaches to accomplish that. We could introduce a cost factor and make it more ‘expensive’ for the machine learning algorithm to misclassify elements of either LACS or TACS [45,46,47]. This form of intervention requires specific classification algorithms which are aware of the cost associated with specific classes. The second method for reducing the bias is by balancing the dataset through over- and under-sampling [48,49,50]. The cost function requirement limits the choice of classifiers, because not all relevant algorithms come with that functionality. Therefore, we have chosen to resample the dataset with AdaSyn. The method was originally proposed by He et al. [51] and it aims to preserve all of the data while harvesting the benefits of balanced datasets.

2.4. Feature Extraction

The DWT [52] was employed on the balanced dataset, wherein the images were decomposed up to 3 levels. Feature extraction was done in 2 phases; (i) before DWT and (ii) after DWT. In (i), the texture features, GLCM [53], GLRLM [53] and HOS [54] were extracted directly from the balanced set of images. The same features were also extracted in (ii).

2.4.1. Discrete Wavelet Transform

DWT was used to decompose the pre-processed images to extract both location and frequency [55,56]. For this work we have used the first three levels of decomposition. Each level of decomposition yields a detailed (D) and an approximated (A) image. The block diagram in Figure 3 shows how an original MRI slice is decomposed to three DWT levels. Hence, after DWT we had seven images for each of the 267 MRI slices: one original and three detailed as well as three approximated images.

2.4.2. Gray-Level Co-Occurrence Matrix

GLCM features establish a relationship between pixels by analyzing images with second-order statistics. During the analysis, the permutation frequency of the pixel brightness is identified [57].
Fourteen linear and nonlinear GLCM features were extracted from the pre-processed MRI images. The linear feature extraction methods were: autocorrelation [57,58], Max Probability [57,59], Dissimilarity [57], Entropy [60], Cluster Shade [57], Sum Average [57], Sum Entropy [57], Sum Variance [57], Difference Variance [57], Difference Entropy [57], Information Correlation Measures 1 and 2 [57].

2.4.3. Gray-Level Run Length Matrix

GLRLM extracts spatial plane features per pixel, relative to the higher-order statistics, wherein a two-dimensional feature matrix is formed at the end [61].
11 GLRLM features, such as the Short Run Emphasis [62], Long Run Emphasis [63], Gray-Level Nonuniformity [63], Run Length Non-uniformity [63], Run Percentage [63], Low Gray-Level Run Emphasis [64], High Gray-Level Run Emphasis [62], Short Run Low Gray-Level Run Emphasis [64], Short Run High Gray-Level Run Emphasis [64], Long Run Low Gray-Level Run Emphasis [62] and Long Run High Gray-Level Run Emphasis were extracted [62].

2.4.4. Higher Order Spectra

HOS based feature extraction methods pick up small structural changes in physiological signals and medical images [65,66,67]. Some of these small structural changes might be due to natural variation, while others might result from pathophysiologic changes caused by a particular disease [68,69,70]. For the current study, we try to classify brain lesions caused by stroke into PACS, LACS and TACS. The classification reflects different pathophysiological processes which result in subtle changes in the MRI image texturing.
HOS feature extraction methods use the Fourier transform of higher order correlations to identify the presence of nonlinear coupling information [71]. These textural features are competent, and hence they were used in our study. To be specific, the following nonlinear features were extracted with HOS methods: entropies 1, 2, 3 as well as HOS phase entropy and HOS mean of magnitude [54]. These features were extracted with phase angels from 1 to 180 in steps of 1°.

2.5. Statistical Analysis

Statistical analysis was used for feature ranking, which is a prerequisite for classification-based feature selection. A particular feature can be considered more important if we can rank it among the other features based on some metric. Therefore, a higher-ranked feature is more valuable for classification than a lower-ranked feature. Moreover, ignoring features that have a rank lower than a specific threshold can also increase classification speed.
The Feature extraction yielded 6475 parameters. This number results from 14 GLCM and 11 GLRLM features that were extracted from the seven images (two for each level of DWT decomposition and one original). Overall that resulted in 175 features. In addition, the five HOS features were also extracted from each image. Evaluating these features in one degree phase steps for half a circle resulted in 6300 = 5 × 180 × 7 features. The 6475 computed features were ranked using the F-value obtained from ANOVA [72]. That allowed us to order the feature set from the largest F-value down to the lowest one.

2.6. Classification

Training and testing machine learning algorithms involves creating training and testing datasets. That data split might introduce bias and the results are less reliable because not all available data were used for testing. We have used 10-fold cross-validation to reduce bias and to improve test coverage. The method unfolds as follows. Initially, all available data i.e., the feature vectors, are split into 10 parts. From these 10 parts we constructed 10 folds by selecting each of the parts as test data and the remaining nine parts as training data. Sequential forward selection was used on each fold for feature selection [73]. Subsequently, these folds were used to train and test the SVM algorithms. Hence, for each of the four tested classification algorithms we obtained 10 sets of performance measures—one for each fold. Averaging the individual performance measures over the 10 folds serves to reduce bias, and it improves the test coverage [74].

2.6.1. Support Vector Machine Classifier

We have selected SVM algorithms [75] to classify the feature vectors into PACS, LACS, and TACS. A training step was used to construct two hyperplanes in a high-dimensional feature space which separate class-specific training data points [76]. During testing, the hyperplanes were used as decision borders, which facilitated the classification task. In this study, we have used four different methods to construct the hyperplanes. These methods differ from one another in terms of the kernel that was used to map the feature vectors into the high-dimensional space [77]. The first method was based on a first order polynomial kernel, also known as a linear kernel. Two more methods were constructed by increasing the polynomial order to two and three, respectively. The fourth method employs the RBF kernel function which is based on the squared Euclidean distance between the feature vectors.

2.6.2. Performance Measures

At a global level, the classification method is judged successful if it can identify the correct stroke type. This testing results in a 2 by 2 confusion matrix with True Positive (TP), True Negative (TN), False Positive (FP), and False Negative (FN) entries. These entries were used to calculate ACC, Positive Predictive Value (PPV), Sensitivity (SEN), and Specificity (SPE), as defined in the following equations:
ACC = ( TP + TN ) / ( TP + TN + FP + FN )
PPV = TP / ( TP + FP )
SEN = TP / ( TP + FN )
SPE = TN / ( TN + FP )
We also established Dice’s similarity index [78,79] using the following formula:
Dice = 2 TP / ( 2 TP + FP + FN )

3. Results

This section presents the 10-fold cross-validation performance results achieved by the four tested SVM classifiers. These results are instrumental to judge feature quality. As such, feature quality reflects the diagnostically relevant information that can be automatically extracted with signal processing algorithms. Hence, the classification results help us in our quest to determine if, and indeed to what extent, brain MRI imagery contain machine extractable information for stroke severity classification. Table 1, Table 2, Table 3 and Table 4 detail the classification performance for each fold and the average performance, overall folds, of the four tested SVM classifiers.
Table 5 shows the classification results of the SVM classifiers. From the results, it is clear that the SVM RBF classifier obtained the highest classification accuracy of 93.62% and a dice score of 0.95 for the classification of PACS, LACS and TACS images.

4. Discussion

Current medical imaging technology allows us to support human expertise during diagnosis and treatment monitoring [80]. With this study, we propose to extend existing technology with decision support algorithms to create a Computer-Aided-Diagnosis (CAD) system. We investigated ischemic stroke severity classification based on standard MRI imagery. Framing the study in such a way allowed us to set the aims and objectives for the proposed decision support algorithms. Within this framework, the main aim of our study is to bring about evolutionary change by providing an adjunct tool for stroke severity classification based on MRI image analysis. To achieve that aim, we must show that it is possible to automate the extraction of diagnostically relevant information from MRI imagery. Therefore, the main objective for this study was to evaluate a range of information extraction methods in terms of their ability to support a stroke severity diagnosis. The evaluation process was structured into statistical and classification performance assessment. The assessment results provide a measure that determines to what extent the main objective was met. Furthermore, these results feed into the wider research community because they provide a way to compare different studies.
Looking back, the initial computer support tools were capable of lesion segmentation in MRI scans [16,81,82]. These methods employed linear decision algorithms, such as thresholding, to accomplish the segmentation task. More recently, these linear methods have been replaced by machine classification methods, because these methods consider nonlinear relationships between information extracted from the MRI images. Studies which investigated lesion segmentation problems show that machine classification outperforms linear decision methodology [15,42,83]. The medical rationale behind these studies was that size, region, and density of the lesion are correlated with the damage caused to brain tissue, which correlates to the cognitive impairments of patients. Addressing this medical need with computer algorithms instead of human expertise, followed the same reasoning as we have put forward for our study. Manual lesion segmentation by a trained professional is referred to as the gold standard against which the automated systems are compared [8,9]. Having automated systems for lesion segmentation offers several advantages. Such a system can be used as an adjunct tool which provides a second opinion on a specific MRI scan. Having such a second opinion, from a tool that does not suffer from inter- and intra-observer variability [18], might lead to a better diagnosis. There is also the economic argument on progress which means to do more with the same resources. For CAD systems, progress implies a shift from human labor to machine work. Lesion segmentation follows a set of rules, and knowing these rules requires theoretical as well as practical training. That training process is a significant cost factor when it comes to health economics. CAD algorithms formalize the lesion segmentation rules such that they steer the execution of a computer program. That computer program needs to be coded only once and it can be deployed multiple times. Usually, this is more cost-effective when compared to training human experts. However, human experts are not limited to analyzing the images; they are also establishing a diagnosis through decision-making. Reaching these decisions requires not only reason but also perception, and indeed, intuition guided by experience. Hence, one way to make further progress is to shift work that requires higher mental capabilities from humans to machines.
Artificial Intelligence (AI) is a technology which can mimic higher mental functions of human experts. On a taxonomy level, machine classification is a subset of AI. However, machine classification methods tend to be referred to as AI when they perform higher mental functions for decision-making. Being capable of reaching that type of decision is distinct from the logical analysis tasks for image segmentation, discussed in the previous paragraph. Combining AI with analysis algorithms creates decision support systems which can execute more tasks on a higher mental level. Inevitably, that kind of progress means to shift more work and indeed responsibility from human experts to machines. Therefore, the performance of such systems is even more critical than it is for analysis support algorithms. When compared to analysis systems, the advice from decision support systems appears later in the diagnosis process, i.e., closer to a decision point; hence there is less room or time for human experts to falsify the machine result. For example, there is a distinct possibility that a decision support system provides the wrong advice, and a human expert will follow that advice, which could lead to negative outcomes for patients. Hence, systems which interpret lesion segmentation results must cope with additional performance requirements. The technology to meet these stricter requirements became available only in recent years; therefore, decision support systems, which provide a statement about the disease itself such as stroke severity classification, were created more recently as compared to analysis systems.
There is a direct relationship between the capability of medical support tools and the responsibility that they incur. Unfortunately, that requirement becomes harder to achieve when medical support tools mimic the higher order mental capabilities of human experts. Therefore, these tools tend to become more complex and less explainable. Looking over the history of MRI image analysis for lesion segmentation, linear image analysis results are straightforward to explain. For example, it might come down to comparing a specific parameter extracted from an MRI image to a specific threshold. From a consideration of nonlinear relationships, machine learning based image segmentation results are less explainable. Supervised machine learning, which was employed by all the reviewed lesion segmentation systems, requires a training step which extracts the parameters for the nonlinear model. Hence, the system performance depends on the training step, and that step might introduce bias and other limitations. When it is difficult to explain the machine results directly, human experts can move into a supervisory position to verify them. Decisions, done as part of an automated lesion segmentation, can be verified relatively quickly through visual inspection. In many cases, seeing the proposed lesion region superimposed over the MRI image is sufficient to accept or reject the segmentation result. Verification of decision support, which involves mimicking higher mental facilities, is more difficult to verify. For example, stroke-type classification to assess stroke severity requires MRI image analysis followed by a diagnostic decision. The verification process is similar to the diagnostic process; hence it takes about the same time. Consequently, there is little or no progress in terms of timesaving. In this case, having an automated stroke severity classification might serve as a second opinion which improves the quality of a diagnostic decision. Real progress is made when human experts can trust automated stroke severity classification systems sufficiently to base their diagnosis on a suggestion from these systems. However, current methods were not designed with sufficient formality to provide the required decision quality, and indeed they were not tested rigorously enough [84]. Simulating lesion fate prediction might help us to understand the physiology of acute stroke, and it could aide in the creation of stroke severity classification systems [85]. We aimed to support the creation of stroke severity classification systems by assessing task-specific feature extraction and AI methods. By doing so, we established that it is possible to automate the extraction of diagnostically relevant information for stroke severity classification from MRI images. This could add to the theoretical foundation for formally developed decision support systems that can bear the responsibility of deciding stroke severity.
Discussing lesion segmentation and automated stroke severity classification from a historical perspective sets the context for reviewing both processing methods and the quality assessment results of relevant research work. We start that discussion with a method developed by Seghier and colleagues [15]. They used probabilistic tissue segmentations obtained from healthy control subjects to estimate mean gray matter and white matter probabilities for each voxel. To identify lesions in patient scans, these estimates were used to compute a metric for quantifying each voxel’s membership to a fuzzy set within each tissue class. A linear threshold method was used to determine the lesion tissue class for an unknown voxel. Wilke and colleagues [16] developed a semi-automated algorithm that uses probabilistic tissue segmentation to construct a set of four feature maps which encode information about tissue composition, tissue homogeneity, shape, and laterality at each voxel. These feature maps were used to construct robust z-score maps, subjected to manual thresholds. A user must combine the results to establish the final lesion delineation. Mitra et al. [86] approached the problem of lesion segmentation with a combination of Bayesian–Markov random fields and random decision forests for voxel-wise classification in multi-spectral MRI volumes. Their system achieved a Dice coefficient of 0.60 ± 0.12. Karthik et al. [87] used transfer learning to segment ischemic lesions. Their deep supervised fully convolutional network achieved a mean segmentation accuracy of 70%. Acharya et al. [88] used HOS entropy features to build a system which is capable of classifying MRI images after stroke into LACS, PACS, and TACS. Their system achieved sensitivity, specificity, accuracy, and positive predictive values equal to 96.4%, 100%, 97.6% and 100%, respectively. Vupputuri et al. [89] investigated the symmetry between left and right brain hemisphere. Their hypothesis was that hemisphere asymmetry provides an important cue for abnormality estimation, and they tested their hypothesis by building an automated lesion segmentation system based on a novel superpixel cluster method. That system achieved a Dice similarity score of 0.704 ± 0.27. Our feature extraction method fuses texture analysis methods with HOS. The resulting features are more expressive when compared to the individual methods.

Limitations and Future Work

Our current work is an attempt to improve the results of our previous work [88]. We have tested a diverse range of features. However, the best performance was close to the one obtained previously. In the current effort, we have fused HOS based features with texture features extracted through GLCM and GLRLM. The fused feature extraction methods were applied to the image directly, as in our previous work, and the first three levels of DWT decomposition. Our results show that the additional texture feature extraction methods (GLCM and GLRLM) and decomposing an image with DWT fail to improve the classification performance. We have to conclude that HOS based features are most suitable for the classification of brain MRI images, because these methods are able to capture subtle changes in the pixel intensities efficiently. Having realized the discriminative power of HOS for stroke severity classification based on MRI images, there might be scope for investigating higher order cumulants as a feature extraction method. Our current work on HOS can serve as a benchmark with which any future method is compared.
Common to both studies is the fact that we had only 267 MRI images and there were significantly more PACS (222) images than TACS (27), and LACS (18). Having so few images limited our study to information extraction and assessment. We could only show that it is possible to extract information from the MRI images that can be used to differentiate the stroke severity. However, it was not possible to construct a system which can be used in a practical diagnosis support setting, because these images do not reflect depth and breadth of MRI images collected for stroke severity diagnosis. Breadth refers to the fact that every brain and indeed every stroke is different. More images are needed for the machine to learn what constitutes a normal brain and how it is affected by different types of stroke. With the current limited dataset, a system which could extract such deep knowledge would overfit. That means the system would extract relevant and irrelevant information during the training phase and thereby gain no knowledge on how a stroke alters the brain. The failure to gain that knowledge manifests itself in a poor performance when the classification system is tested with unknown MRI images. Significantly more data is needed to force such AI systems to learn only from relevant information which lead to an understanding on how the MRI technology captures the different types of stroke. Another limitation is that we had data from only one center in Bhubaneswar, Odisha, India. Hence, our study might be biased to detecting stroke in a specific ethnic group, because the measurements were carried out in one geographical region. In other words, for a different region, different features might be more relevant. Furthermore, having only 267 MRI images from one center is not deep enough to rule out observer bias. One final thought on the data and its limitations centers on the fact that the MRI slices were carefully selected to show clear-cut textbook examples for the different types of stroke. This limits our study in terms of automatization and generalization. Manual selection of MRI slices introduces inter- and intra-observer variability. Furthermore, it requires expert labor which is a major cost factor. Hence, the proposed automatization is still incomplete, and slice selection is an important topic for future research work. Indeed, looking at the continuum of slices, rather than just one, might reveal more information which can lead to more robust decision support results. The generalization limitation arises also from the expert selection of MRI slices. Having good examples is important for a learning situation because it avoids confusion among the learners. From this perspective, training AI is not different from training human experts. However, for human experts we have ample of evidence that they can extract transferable knowledge from limited and biased data sets, which helps them in a practical setting. There are significantly fewer examples showing that AI has that ability. In that specific sense, we did not investigate whether the proposed feature extraction methods can be transferred to a practical setting.
The classification into PACS, TACS, and LACS helps human experts to reach a diagnosis, which is the cornerstone for managing the disease with treatment and specialized care. In this study we show that AI systems can be used to formalize the identification of different stroke types. However, for future projects we need more data to address the limitations and biases inherent in the current image dataset. Data from more geographical regions are required to address ethnic bias. More data, even from the same center, is required to avoid observer bias and to reflect observable variations adequately. As a result, the AI algorithms for future studies must deal with significantly larger datasets. Experience shows that feature extraction coupled with classical machine learning is not capable of handling the training data needed to prepare AI systems for practical diagnosis support [90]. Therefore, future studies require different AI algorithms, such as Deep Learning (DL). DL is a concept which turns big training data into an advantage for practical decision support systems. This is done by allowing sufficient network depth to capture more relevant knowledge without the need for feature engineering. We suggest that using this type of algorithm will improve the practical relevance for future stroke severity classification systems. However, we are aware that even considering MRI scans from multiple centers of longer time periods might not yield sufficient data to train the most potent deep networks. One way of addressing that problem might be to use transfer learning, which is based on the idea of customizing pre-trained networks [91]. Customizing the network requires less labeled data when compared to training a deep network from scratch. Furthermore, the computational complexity is lower, which makes hyperparameter tuning less time-consuming.
Our final thought on limitations and future work is concerned with the practicalities of having an automated stroke severity classification system in the care pathway. Our initial assumption is that such a system will be useful whenever MRI images are captured to determine the stroke type. In the future, working with multiple centers might help us to determine which stage of the care pathway will benefit most from the application of automated stroke severity classification. This leads to focused research questions, such as: Should automated stroke severity classification be used in acute, sub-acute or chronic settings? Would such a system help in acute treatment with IV or IA thrombolysis or chronic disability management by predicting prognosis or something else?

5. Conclusions

With this paper we set-out to answer the question as to whether it is possible to automate ischemic stroke severity classification based on MRI imagery. To investigate that research question, we designed a system which uses HOS, GLRLM, as well as a combination of DWT and GLCM methods for feature extraction. The resulting features were assessed with statistical and classification methods. Statistical assessment allowed us to rank the features. The best features scored a very low p-value; therefore, we have to reject the hypothesis that the features were drawn from the same data class. In other words, the statistical analysis indicates that the features reflect different data classes. The classification assessment quantifies the ability of the best features to differentiate between PACS, LACS, and TACS. We found that SVM with RBF kernel achieves: ACC = 93.62%, SPE = 95.91%, SEN = 92.44%, and Dice = 0.95. These performance measures were established based on the rules of 10-fold cross-validation. Hence, the results are statistically robust, and they indicate that fusing texture analysis with HOS produces superior results when compared to individual methods. Furthermore, the established Dice score indicates that our feature extraction method outperforms all MRI based lesion segmentation and stroke severity classification algorithms published in the scientific literature. Therefore, fusing texture and HOS methods can serve as a scientific foundation which indicates that it is possible to differentiate PACS, LACS, and TACS based on MRI images. The proposed feature extraction system helped us to support our hypothesis that it is possible to automate stroke severity classification based on MRI images.
This paper documents a significant step towards automated ischemic stroke severity classification based on MRI imagery. The results provide the necessary confidence to propose further investigations. Data is the most crucial part of these investigations because we need datasets which reflect the diversity encountered during a medical diagnosis. We are confident that the necessary processing tools are available and MRI images contain diagnostically relevant information that can be extracted automatically. Hence, we can shift our research effort from information focused feature engineering to knowledge-based decision support systems. Ultimately, we would like to build a system which understands the MRI image texture for PACS, LACS and TACS.

Author Contributions

Conceptualization, U.R.A.; methodology, J.E.W.K. and V.J.; data curation, S.S.; validation, A.M. and G.Y.H.L.; writing—original draft preparation, O.F.; writing—review and editing, A.A., G.Y.H.L., E.J.C. and U.R.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

We conducted a retrospective study. Therefore, ethical approval was not required.

Informed Consent Statement

Informed consent was not required for this retrospective study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to ethical restrictions.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
ACCAccuracy
AdaSynAdaptive Synthetic
ANOVAAnalysis Of Variance
AIArtificial Intelligence
CADComputer-Aided-Diagnosis
CLAHEContrast Limited Adaptive Histogram Equalization
CTComputed Tomography
DLDeep Learning
DWIDiffusion-Weighted Imaging
DWTDiscrete Wavelet Transform
FNFalse Negative
FPFalse Positive
GLCMGray-Level Co-occurrence Matrix
GLRLMGray-Level Run Length Matrix
HOSHigher Order Spectra
LACSLacunar Syndrome
MRIMagnetic Resonance Imaging
PACSPartial Anterior Circulation Syndrome
PPVPositive Predictive Value
RBFRadial Basis Function
SENSensitivity
SPESpecificity
SVMSupport Vector Machine
TACSTotal Anterior Circulation Stroke
TNTrue Negative
TPTrue Positive
VBMVoxel-Based Morphometry

References

  1. Bonita, R. Epidemiology of stroke. Lancet 1992, 339, 342–344. [Google Scholar] [CrossRef]
  2. Lopez, A.D.; Mathers, C.D.; Ezzati, M.; Jamison, D.T.; Murray, C.J. Global and regional burden of disease and risk factors, 2001: Systematic analysis of population health data. Lancet 2006, 367, 1747–1757. [Google Scholar] [CrossRef]
  3. Smith, S.C., Jr. Reducing the global burden of ischemic heart disease and stroke: A challenge for the cardiovascular community and the United Nations. Circulation 2011, 124, 278–279. [Google Scholar] [CrossRef] [Green Version]
  4. Feigin, V.L.; Lawes, C.M.; Bennett, D.A.; Anderson, C.S. Stroke epidemiology: A review of population-based studies of incidence, prevalence, and case-fatality in the late 20th century. Lancet Neurol. 2003, 2, 43–53. [Google Scholar] [CrossRef]
  5. Dirnagl, U.; Iadecola, C.; Moskowitz, M.A. Pathobiology of ischaemic stroke: An integrated view. Trends Neurosci. 1999, 22, 391–397. [Google Scholar] [CrossRef]
  6. van der Worp, H.B.; van Gijn, J. Acute ischemic stroke. N. Engl. J. Med. 2007, 357, 572–579. [Google Scholar] [CrossRef] [PubMed]
  7. Latchaw, R.E.; Alberts, M.J.; Lev, M.H.; Connors, J.J.; Harbaugh, R.E.; Higashida, R.T.; Hobson, R.; Kidwell, C.S.; Koroshetz, W.J.; Mathews, V.; et al. Recommendations for imaging of acute ischemic stroke: A scientific statement from the American Heart Association. Stroke 2009, 40, 3646–3678. [Google Scholar] [CrossRef] [PubMed]
  8. Bates, E.; Wilson, S.M.; Saygin, A.P.; Dick, F.; Sereno, M.I.; Knight, R.T.; Dronkers, N.F. Voxel-based lesion–symptom mapping. Nat. Neurosci. 2003, 6, 448–450. [Google Scholar] [CrossRef] [PubMed]
  9. Dronkers, N.F.; Wilkins, D.P.; Van Valin, R.D., Jr.; Redfern, B.B.; Jaeger, J.J. Lesion analysis of the brain areas involved in language comprehension. Cognition 2004, 92, 145–177. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  10. Fiez, J.A.; Damasio, H.; Grabowski, T.J. Lesion segmentation and manual warping to a reference brain: Intra-and interobserver reliability. Hum. Brain Mapp. 2000, 9, 192–211. [Google Scholar] [CrossRef]
  11. Lindgren, A.; Norrving, B.; Rudling, O.; Johansson, B.B. Comparison of clinical and neuroradiological findings in first-ever stroke. A population-based study. Stroke 1994, 25, 1371–1377. [Google Scholar] [CrossRef] [Green Version]
  12. Zaidi, S.F.; Aghaebrahim, A.; Urra, X.; Jumaa, M.A.; Jankowitz, B.; Hammer, M.; Nogueira, R.; Horowitz, M.; Reddy, V.; Jovin, T.G. Final infarct volume is a stronger predictor of outcome than recanalization in patients with proximal middle cerebral artery occlusion treated with endovascular therapy. Stroke 2012, 43, 3238–3244. [Google Scholar] [CrossRef] [Green Version]
  13. Vogt, G.; Laage, R.; Shuaib, A.; Schneider, A. Initial lesion volume is an independent predictor of clinical stroke outcome at day 90: An analysis of the Virtual International Stroke Trials Archive (VISTA) database. Stroke 2012, 43, 1266–1272. [Google Scholar] [CrossRef] [Green Version]
  14. Rangaraju, S.; Owada, K.; Noorian, A.R.; Nogueira, R.G.; Nahab, F.; Glenn, B.A.; Belagaje, S.R.; Anderson, A.M.; Frankel, M.R.; Gupta, R. Comparison of final infarct volumes in patients who received endovascular therapy or intravenous thrombolysis for acute intracranial large-vessel occlusions. JAMA Neurol. 2013, 70, 831–836. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. Seghier, M.L.; Ramlackhansingh, A.; Crinion, J.; Leff, A.P.; Price, C.J. Lesion identification using unified segmentation-normalisation models and fuzzy clustering. Neuroimage 2008, 41, 1253–1266. [Google Scholar] [CrossRef] [Green Version]
  16. Wilke, M.; de Haan, B.; Juenger, H.; Karnath, H.O. Manual, semi-automated, and automated delineation of chronic brain lesions: A comparison of methods. NeuroImage 2011, 56, 2038–2046. [Google Scholar] [CrossRef] [PubMed]
  17. Meyers, P.M.; Schumacher, H.C.; Connolly, E.S., Jr.; Heyer, E.J.; Gray, W.A.; Higashida, R.T. Current status of endovascular stroke treatment. Circulation 2011, 123, 2591–2601. [Google Scholar] [CrossRef]
  18. Ashton, E.A.; Takahashi, C.; Berg, M.J.; Goodman, A.; Totterman, S.; Ekholm, S. Accuracy and reproducibility of manual and semiautomated quantification of MS lesions by MRI. J. Magn. Reson. Imaging Off. J. Int. Soc. Magn. Reson. Med. 2003, 17, 300–308. [Google Scholar] [CrossRef]
  19. Filippi, M.; Horsfield, M.; Bressi, S.; Martinelli, V.; Baratti, C.; Reganati, P.; Campi, A.; Miller, D.; Comi, G. Intra-and inter-observer agreement of brain MRI lesion volume measurements in multiple sclerosis: A comparison of techniques. Brain 1995, 118, 1593–1600. [Google Scholar] [CrossRef]
  20. Lansberg, M.G.; Albers, G.W.; Beaulieu, C.; Marks, M.P. Comparison of diffusion-weighted MRI and CT in acute stroke. Neurology 2000, 54, 1557–1561. [Google Scholar] [CrossRef] [PubMed]
  21. Baliyan, V.; Das, C.J.; Sharma, R.; Gupta, A.K. Diffusion weighted imaging: Technique and applications. World J. Radiol. 2016, 8, 785. [Google Scholar] [CrossRef]
  22. Al-Khaled, M.; Matthis, C.; Münte, T.F.; Eggers, J.; QugSS2-Study. The incidence and clinical predictors of acute infarction in patients with transient ischemic attack using MRI including DWI. Neuroradiology 2013, 55, 157–163. [Google Scholar] [CrossRef]
  23. Lettau, M.; Laible, M. 3-T high-b-value diffusion-weighted MR imaging in hyperacute ischemic stroke. J. Neuroradiol. 2013, 40, 149–157. [Google Scholar] [CrossRef]
  24. Lutsep, H.L.; Albers, G.; DeCrespigny, A.; Kamat, G.; Marks, M.; Moseley, M. Clinical utility of diffusion-weighted magnetic resonance imaging in the assessment of ischemic stroke. Ann. Neurol. Off. J. Am. Neurol. Assoc. Child Neurol. Soc. 1997, 41, 574–580. [Google Scholar] [CrossRef] [PubMed]
  25. Newcombe, V.; Das, T.; Cross, J. Diffusion imaging in neurological disease. J. Neurol. 2013, 260, 335–342. [Google Scholar] [CrossRef]
  26. De La Ossa, N.P.; Hernandez-Perez, M.; Domenech, S.; Cuadras, P.; Massuet, A.; Millan, M.; Gomis, M.; Lopez-Cancio, E.; Dorado, L.; Davalos, A. Hyperintensity of distal vessels on FLAIR is associated with slow progression of the infarction in acute ischemic stroke. Cerebrovasc. Dis. 2012, 34, 376–384. [Google Scholar] [CrossRef]
  27. Gitelman, D.R.; Ashburner, J.; Friston, K.J.; Tyler, L.K.; Price, C.J. Voxel-based morphometry of herpes simplex encephalitis. Neuroimage 2001, 13, 623–631. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Mehta, S.; Grabowski, T.J.; Trivedi, Y.; Damasio, H. Evaluation of voxel-based morphometry for focal lesion detection in individuals. Neuroimage 2003, 20, 1438–1454. [Google Scholar] [CrossRef]
  29. Anbeek, P.; Vincken, K.L.; van Osch, M.J.; Bisschops, R.H.; van der Grond, J. Automatic segmentation of different-sized white matter lesions by voxel probability estimation. Med. Image Anal. 2004, 8, 205–215. [Google Scholar] [CrossRef]
  30. Capelle, A.S.; Colot, O.; Fernandez-Maloigne, C. Evidential segmentation scheme of multi-echo MR images for the detection of brain tumors using neighborhood information. Inf. Fusion 2004, 5, 203–216. [Google Scholar] [CrossRef]
  31. Colliot, O.; Bernasconi, N.; Khalili, N.; Antel, S.B.; Naessens, V.; Bernasconi, A. Individual voxel-based analysis of gray matter in focal cortical dysplasia. Neuroimage 2006, 29, 162–171. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Colliot, O.; Mansi, T.; Bernasconi, N.; Naessens, V.; Klironomos, D.; Bernasconi, A. Segmentation of focal cortical dysplasia lesions on MRI using level set evolution. Neuroimage 2006, 32, 1621–1630. [Google Scholar] [CrossRef] [Green Version]
  33. Datta, S.; Sajja, B.R.; He, R.; Wolinsky, J.S.; Gupta, R.K.; Narayana, P.A. Segmentation and quantification of black holes in multiple sclerosis. Neuroimage 2006, 29, 467–474. [Google Scholar] [CrossRef] [Green Version]
  34. Fletcher-Heath, L.M.; Hall, L.O.; Goldgof, D.B.; Murtagh, F.R. Automatic segmentation of non-enhancing brain tumors in magnetic resonance images. Artif. Intell. Med. 2001, 21, 43–63. [Google Scholar] [CrossRef] [Green Version]
  35. Hojjatoleslami, S.; Kruggel, F. Segmentation of large brain lesions. IEEE Trans. Med. Imaging 2001, 20, 666–669. [Google Scholar] [CrossRef] [PubMed]
  36. Prastawa, M.; Bullitt, E.; Ho, S.; Gerig, G. A brain tumor segmentation framework based on outlier detection. Med. Image Anal. 2004, 8, 275–283. [Google Scholar] [CrossRef]
  37. Sajja, B.R.; Datta, S.; He, R.; Mehta, M.; Gupta, R.K.; Wolinsky, J.S.; Narayana, P.A. Unified approach for multiple sclerosis lesion segmentation on brain MRI. Ann. Biomed. Eng. 2006, 34, 142–151. [Google Scholar] [CrossRef] [PubMed]
  38. Stamatakis, E.A.; Tyler, L.K. Identifying lesions on structural brain images—Validation of the method and application to neuropsychological patients. Brain Lang. 2005, 94, 167–177. [Google Scholar] [CrossRef]
  39. Wu, Y.; Warfield, S.K.; Tan, I.L.; Wells, W.M., III; Meier, D.S.; van Schijndel, R.A.; Barkhof, F.; Guttmann, C.R. Automated segmentation of multiple sclerosis lesion subtypes with multichannel MRI. NeuroImage 2006, 32, 1205–1215. [Google Scholar] [CrossRef]
  40. Xie, K.; Yang, J.; Zhang, Z.; Zhu, Y. Semi-automated brain tumor and edema segmentation using MRI. Eur. J. Radiol. 2005, 56, 12–19. [Google Scholar] [CrossRef]
  41. Zhou, J.; Chan, K.; Chong, V.; Krishnan, S.M. Extraction of brain tumor from MR images using one-class support vector machine. In Proceedings of the 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, Shanghai, China, 17–18 January 2006; pp. 6411–6414. [Google Scholar]
  42. Rekik, I.; Allassonnière, S.; Carpenter, T.K.; Wardlaw, J.M. Medical image analysis methods in MR/CT-imaged acute-subacute ischemic stroke lesion: Segmentation, prediction and insights into dynamic evolution simulation models. A critical appraisal. Neuroimage Clin. 2012, 1, 164–178. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Zuiderveld, K. Contrast Limited Adaptive Histogram Equalization. In Graphics Gems IV; Academic Press Professional, Inc.: Cambridge, MA, USA, 1994; pp. 474–485. [Google Scholar]
  44. He, H.; Garcia, E.A. Learning from imbalanced data. IEEE Trans. Knowl. Data Eng. 2009, 21, 1263–1284. [Google Scholar]
  45. Pazzani, M.; Merz, C.; Murphy, P.; Ali, K.; Hume, T.; Brunk, C. Reducing misclassification costs. In Machine Learning Proceedings 1994; Elsevier: Amsterdam, The Netherlands, 1994; pp. 217–225. [Google Scholar]
  46. Domingos, P. MetaCost: A General Method for Making Classifiers Cost-Sensitive. In Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD ’99), New York, NY, USA, 1 August 1999; pp. 155–164. [Google Scholar] [CrossRef]
  47. Phankokkruad, M. Cost-Sensitive Extreme Gradient Boosting for Imbalanced Classification of Breast Cancer Diagnosis. In Proceedings of the 2020 10th IEEE International Conference on Control System, Computing and Engineering (ICCSCE), Penang, Malaysia, 21–22 August 2020; pp. 46–51. [Google Scholar]
  48. Kubat, M.; Matwin, S. Addressing the curse of imbalanced training sets: One-sided selection. ICML Citeseer 1997, 97, 179–186. [Google Scholar]
  49. Japkowicz, N. The class imbalance problem: Significance and strategies. In Proceedings of the 2000 International Conference on Artificial Intelligence (ICAI), Las Vegas, NV, USA, 1 July 2000; Volume 56. [Google Scholar]
  50. Susan, S.; Kumar, A. The balancing trick: Optimized sampling of imbalanced datasets—A brief survey of the recent State of the Art. Eng. Rep. 2020, 3, e12298. [Google Scholar] [CrossRef]
  51. He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive synthetic sampling approach for imbalanced learning. In Proceedings of the 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), Hong Kong, China, 1–8 June 2008; pp. 1322–1328. [Google Scholar]
  52. Mishra, H.O.S.; Bhatnagar, S. MRI and CT image fusion based on wavelet transform. Int. J. Inf. Comput. Technol. 2014, 4, 47–52. [Google Scholar]
  53. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural features for image classification. IEEE Trans. Syst. Man Cybern. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef] [Green Version]
  54. Ng, T.T.; Chang, S.F.; Sun, Q. Blind detection of photomontage using higher order statistics. In Proceedings of the 2004 IEEE International Symposium on Circuits and Systems (IEEE Cat. No. 04CH37512), Vancouver, BC, Canada, 23–26 May 2004; Volume 5. [Google Scholar] [CrossRef] [Green Version]
  55. Acharya, U.R.; Faust, O.; Sree, S.V.; Molinari, F.; Suri, J.S. ThyroScreen system: High resolution ultrasound thyroid image characterization into benign and malignant classes using novel combination of texture and discrete wavelet transform. Comput. Methods Programs Biomed. 2012, 107, 233–241. [Google Scholar] [CrossRef] [PubMed]
  56. Acharya, U.R.; Faust, O.; Sree, S.V.; Molinari, F.; Garberoglio, R.; Suri, J. Cost-effective and non-invasive automated benign & malignant thyroid lesion classification in 3D contrast-enhanced ultrasound using combination of wavelets and textures: A class of ThyroScan™ algorithms. Technol. Cancer Res. Treat. 2011, 10, 371–380. [Google Scholar]
  57. Öztürk, Ş.; Akdemir, B. Application of feature extraction and classification methods for histopathological image using GLCM, LBP, LBGLCM, GLRLM and SFTA. Procedia Comput. Sci. 2018, 132, 40–46. [Google Scholar] [CrossRef]
  58. Rimavičius, T.; Gelžinis, A.; Verikas, A.; Vaičiukynas, E.; Bačauskienė, M.; Šaškov, A. Automatic benthic imagery recognition using a hierarchical two-stage approach. Signal Image Video Process. 2018, 12, 1107–1114. [Google Scholar] [CrossRef]
  59. Soh, L.K.; Tsatsoulis, C. Texture analysis of SAR sea ice imagery using gray level co-occurrence matrices. IEEE Trans. Geosci. Remote Sens. 1999, 37, 780–795. [Google Scholar] [CrossRef] [Green Version]
  60. Kannathal, N.; Choo, M.L.; Acharya, U.R.; Sadasivan, P. Entropies for detection of epilepsy in EEG. Comput. Methods Programs Biomed. 2005, 80, 187–194. [Google Scholar] [CrossRef]
  61. Mohanty, A.K.; Beberta, S.; Lenka, S.K. Classifying benign and malignant mass using GLCM and GLRLM based texture features from mammogram. Int. J. Eng. Res. Appl. 2011, 1, 687–693. [Google Scholar]
  62. Tang, X. Texture information in run-length matrices. IEEE Trans. Image Process. 1998, 7, 1602–1609. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  63. Galloway, M.M. Texture analysis using grey level run lengths. NASA STI/Recon Tech. Rep. N 1974, 75, 18555. [Google Scholar]
  64. Xu, D.H.; Kurani, A.S.; Furst, J.D.; Raicu, D.S. Run-length encoding for volumetric texture. Heart 2004, 27, 452–458. [Google Scholar]
  65. Martis, R.J.; Acharya, U.R.; Mandana, K.; Ray, A.K.; Chakraborty, C. Cardiac decision making using higher order spectra. Biomed. Signal Process. Control 2013, 8, 193–203. [Google Scholar] [CrossRef]
  66. Noronha, K.P.; Acharya, U.R.; Nayak, K.P.; Martis, R.J.; Bhandary, S.V. Automated classification of glaucoma stages using higher order cumulant features. Biomed. Signal Process. Control 2014, 10, 174–183. [Google Scholar] [CrossRef]
  67. Swapna, G.; Rajendra Acharya, U.; VinithaSree, S.; Suri, J.S. Automated detection of diabetes using higher order spectral features extracted from heart rate signals. Intell. Data Anal. 2013, 17, 309–326. [Google Scholar] [CrossRef]
  68. Acharya, U.R.; Faust, O.; Molinari, F.; Sree, S.V.; Junnarkar, S.P.; Sudarshan, V. Ultrasound-based tissue characterization and classification of fatty liver disease: A screening and diagnostic paradigm. Knowl. Based Syst. 2015, 75, 66–77. [Google Scholar] [CrossRef]
  69. Acharya, U.R.; Raghavendra, U.; Fujita, H.; Hagiwara, Y.; Koh, J.E.; Hong, T.J.; Sudarshan, V.K.; Vijayananthan, A.; Yeong, C.H.; Gudigar, A.; et al. Automated characterization of fatty liver disease and cirrhosis using curvelet transform and entropy features extracted from ultrasound images. Comput. Biol. Med. 2016, 79, 250–258. [Google Scholar] [CrossRef] [PubMed]
  70. Raghavendra, U.; Gudigar, A.; Maithri, M.; Gertych, A.; Meiburger, K.M.; Yeong, C.H.; Madla, C.; Kongmebhol, P.; Molinari, F.; Ng, K.H.; et al. Optimized multi-level elongated quinary patterns for the assessment of thyroid nodules in ultrasound images. Comput. Biol. Med. 2018, 95, 55–62. [Google Scholar] [CrossRef] [PubMed]
  71. Assi, E.B.; Gagliano, L.; Rihana, S.; Nguyen, D.K.; Sawan, M. Bispectrum features and multilayer perceptron classifier to enhance seizure prediction. Sci. Rep. 2018, 8, 1–8. [Google Scholar]
  72. Hoaglin, D.C.; Welsch, R.E. The hat matrix in regression and ANOVA. Am. Stat. 1978, 32, 17–22. [Google Scholar]
  73. Last, M.; Kandel, A.; Maimon, O. Information-theoretic algorithm for feature selection. Pattern Recognit. Lett. 2001, 22, 799–811. [Google Scholar] [CrossRef]
  74. Berrar, D. Cross-validation. Encycl. Bioinform. Comput. Biol. 2019, 1, 542–545. [Google Scholar]
  75. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  76. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning: Data Mining, Inference, and Prediction; Springer Science & Business Media: Heidelberger/Berlin, Germany, 2009. [Google Scholar]
  77. Chang, Y.W.; Hsieh, C.J.; Chang, K.W.; Ringgaard, M.; Lin, C.J. Training and Testing Low-degree Polynomial Data Mappings via Linear SVM. J. Mach. Learn. Res. 2010, 11, 1471–1490. [Google Scholar]
  78. Ye, S.; Ye, J. Dice similarity measure between single valued neutrosophic multisets and its application in medical diagnosis. Neutrosophic Sets Syst. 2014, 6, 9. [Google Scholar]
  79. Dice, L.R. Measures of the amount of ecologic association between species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  80. Sarmento, R.M.; Vasconcelos, F.F.X.; Rebouças Filho, P.P.; Wu, W.; de Albuquerque, V.H.C. Automatic neuroimage processing and analysis in stroke—A systematic review. IEEE Rev. Biomed. Eng. 2019, 13, 130–155. [Google Scholar] [CrossRef] [PubMed]
  81. Kruggel, F.; Chalopin, C.; Descombes, X.; Kovalev, V.; Rajapakse, J. Segmentation of pathological features in MRI brain datasets. In Proceedings of the 9th International Conference on Neural Information Processing, 2002. ICONIP’02, Singapore, 18–22 November 2002; Volume 5, pp. 2673–2677. [Google Scholar]
  82. Agam, G.; Weiss, D.; Soman, M.; Arfanakis, K. Probabilistic brain lesion segmentation in DT-MRI. In Proceedings of the 2006 International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 89–92. [Google Scholar]
  83. Chen, L.; Bentley, P.; Rueckert, D. Fully automatic acute ischemic lesion segmentation in DWI using convolutional neural networks. NeuroImage: Clin. 2017, 15, 633–643. [Google Scholar] [CrossRef] [PubMed]
  84. Faust, O.; Acharya, U.R.; Tamura, T. Formal design methods for reliable computer-aided diagnosis: A review. IEEE Rev. Biomed. Eng. 2012, 5, 15–28. [Google Scholar] [CrossRef] [PubMed]
  85. Debs, N.; Rasti, P.; Victor, L.; Cho, T.H.; Frindel, C.; Rousseau, D. Simulated perfusion MRI data to boost training of convolutional neural networks for lesion fate prediction in acute stroke. Comput. Biol. Med. 2020, 116, 103579. [Google Scholar] [CrossRef]
  86. Mitra, J.; Bourgeat, P.; Fripp, J.; Ghose, S.; Rose, S.; Salvado, O.; Connelly, A.; Campbell, B.; Palmer, S.; Sharma, G.; et al. Lesion segmentation from multimodal MRI using random forest following ischemic stroke. NeuroImage 2014, 98, 324–335. [Google Scholar] [CrossRef]
  87. Karthik, R.; Gupta, U.; Jha, A.; Rajalakshmi, R.; Menaka, R. A deep supervised approach for ischemic lesion segmentation from multimodal MRI using fully convolutional network. Appl. Soft Comput. 2019, 84, 105685. [Google Scholar] [CrossRef]
  88. Acharya, U.R.; Meiburger, K.M.; Faust, O.; Koh, J.E.W.; Oh, S.L.; Ciaccio, E.J.; Subudhi, A.; Jahmunah, V.; Sabut, S. Automatic detection of ischemic stroke using higher order spectra features in brain MRI images. Cogn. Syst. Res. 2019, 58, 134–142. [Google Scholar] [CrossRef]
  89. Vupputuri, A.; Ashwal, S.; Tsao, B.; Ghosh, N. Ischemic stroke segmentation in multi-sequence MRI by symmetry determined superpixel based hierarchical clustering. Comput. Biol. Med. 2020, 116, 103536. [Google Scholar] [CrossRef]
  90. Faust, O.; Hagiwara, Y.; Hong, T.J.; Lih, O.S.; Acharya, U.R. Deep learning for healthcare applications based on physiological signals: A review. Comput. Methods Programs Biomed. 2018, 161, 1–13. [Google Scholar] [CrossRef]
  91. Baygin, M.; Dogan, S.; Tuncer, T.; Barua, P.D.; Faust, O.; Arunkumar, N.; Abdulhay, E.W.; Palmer, E.E.; Acharya, U.R. Automated ASD detection using hybrid deep lightweight features extracted from EEG signals. Comput. Biol. Med. 2021, 134, 104548. [Google Scholar] [CrossRef]
Figure 1. Overview block diagram.
Figure 1. Overview block diagram.
Ijerph 18 08059 g001
Figure 2. Class specific examples from the image dataset. The original images are shown in the first row: (ac). The images were created by processing the original images with contrast adaptive histogram equalization: (df).
Figure 2. Class specific examples from the image dataset. The original images are shown in the first row: (ac). The images were created by processing the original images with contrast adaptive histogram equalization: (df).
Ijerph 18 08059 g002
Figure 3. DWT analysis block diagram. DX indicates the detailed image at level X and AX indicates the approximated image at level X.
Figure 3. DWT analysis block diagram. DX indicates the detailed image at level X and AX indicates the approximated image at level X.
Ijerph 18 08059 g003
Table 1. SVM with linear kernel. The first 10 rows in the table document the performance results for each individual fold. The last row states the averaged performance results.
Table 1. SVM with linear kernel. The first 10 rows in the table document the performance results for each individual fold. The last row states the averaged performance results.
TPTNFPFNACC %PPV %SEN %SPE %Dice
37211689.2397.3786.0595.450.91
41194290.9191.1195.3582.610.93
38175683.3388.3786.3677.270.87
40211492.4297.5690.9195.450.94
39175584.8588.6488.6477.270.89
39184586.3690.7088.6481.820.90
40193489.3993.0290.9186.360.92
36202884.8594.7481.8290.910.88
38202687.8895.0086.3690.910.90
36211787.6997.3083.7295.450.90
384193285387.6993.3887.8887.350.90
Table 2. SVM with a second order polynomial kernel. The first 10 rows in the table document the performance results for each individual fold. The last row states the averaged performance results.
Table 2. SVM with a second order polynomial kernel. The first 10 rows in the table document the performance results for each individual fold. The last row states the averaged performance results.
TPTNFPFNACC %PPV %SEN %SPE %Dice
37220690.77100.0086.05100.000.93
39230493.94100.0090.70100.000.95
40202490.9195.2490.9190.910.93
39220592.42100.0088.64100.000.94
43211196.9797.7397.7395.450.98
40220493.94100.0090.91100.000.95
41211393.9497.6293.1895.450.95
36220887.88100.0081.82100.000.90
40220493.94100.0090.91100.000.95
40220395.38100.0093.02100.000.96
39521744293.0199.0690.3998.180.94
Table 3. SVM with a third order polynomial kernel. The first 10 rows in the table document the performance results for each individual fold. The last row states the averaged performance results.
Table 3. SVM with a third order polynomial kernel. The first 10 rows in the table document the performance results for each individual fold. The last row states the averaged performance results.
TPTNFPFNACC %PPV %SEN %SPE %Dice
38220592.31100.0088.37100.000.94
38230592.42100.0088.37100.000.94
40211492.4297.5690.9195.450.94
39220592.42100.0088.64100.000.94
42211295.4597.6795.4595.450.97
41220395.45100.0093.18100.000.96
38202687.8895.0086.3690.910.90
37220789.39100.0084.09100.000.91
37220789.39100.0084.09100.000.91
41220296.92100.0095.35100.000.98
39121744692.4199.0289.4898.180.94
Table 4. SVM with RBF kernel. The first 10 rows in the table document the performance results for each individual fold. The last row states the averaged performance results.
Table 4. SVM with RBF kernel. The first 10 rows in the table document the performance results for each individual fold. The last row states the averaged performance results.
TPTNFPFNACC %PPV %SEN %SPE %Dice
38211590.7797.4488.3795.450.93
40230395.45100.0093.02100.000.96
41184389.3991.1193.1881.820.92
40220493.94100.0090.91100.000.95
43202195.4595.5697.7390.910.97
40211492.4297.5690.9195.450.94
42211295.4597.6795.4595.450.97
39220592.42100.0088.64100.000.94
41220395.45100.0093.18100.000.96
40220395.38100.0093.02100.000.96
40421293393.6297.9392.4495.910.95
Table 5. Classification results of the SVM algorithms.
Table 5. Classification results of the SVM algorithms.
ClassifierAverage
SEN %
Average
SPE %
Average
PPV %
Average
ACC %
Average
Dice
SVM linear87.8887.3593.3887.690.90
SVM RBF92.4495.9197.9393.620.95
SVM polynomial 290.3998.1899.0693.010.94
SVM polynomial 389.4898.1899.0292.410.94
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Faust, O.; En Wei Koh, J.; Jahmunah, V.; Sabut, S.; Ciaccio, E.J.; Majid, A.; Ali, A.; Lip, G.Y.H.; Acharya, U.R. Fusion of Higher Order Spectra and Texture Extraction Methods for Automated Stroke Severity Classification with MRI Images. Int. J. Environ. Res. Public Health 2021, 18, 8059. https://doi.org/10.3390/ijerph18158059

AMA Style

Faust O, En Wei Koh J, Jahmunah V, Sabut S, Ciaccio EJ, Majid A, Ali A, Lip GYH, Acharya UR. Fusion of Higher Order Spectra and Texture Extraction Methods for Automated Stroke Severity Classification with MRI Images. International Journal of Environmental Research and Public Health. 2021; 18(15):8059. https://doi.org/10.3390/ijerph18158059

Chicago/Turabian Style

Faust, Oliver, Joel En Wei Koh, Vicnesh Jahmunah, Sukant Sabut, Edward J. Ciaccio, Arshad Majid, Ali Ali, Gregory Y. H. Lip, and U. Rajendra Acharya. 2021. "Fusion of Higher Order Spectra and Texture Extraction Methods for Automated Stroke Severity Classification with MRI Images" International Journal of Environmental Research and Public Health 18, no. 15: 8059. https://doi.org/10.3390/ijerph18158059

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop