Next Article in Journal
Development of a Short-Form Hwa-Byung Symptom Scale Using Machine Learning Approaches
Previous Article in Journal
Comparative Efficacy of Quadratus Lumborum Muscle Energy Technique with Gluteus Medius Strengthening Versus Gluteus Medius Strengthening Alone in Sacroiliac Joint Dysfunction: A Randomized Controlled Trial
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Subarachnoid Hemorrhage Using CNN with Dynamic Factor and Wandering Strategy-Based Feature Selection

by
Jewel Sengupta
1,*,
Robertas Alzbutas
1,2,
Tomas Iešmantas
1,
Vytautas Petkus
3,
Alina Barkauskienė
4,
Vytenis Ratkūnas
5,
Saulius Lukoševičius
5,
Aidanas Preikšaitis
6,
Indre Lapinskienė
7,
Mindaugas Šerpytis
7,
Edgaras Misiulis
2,
Gediminas Skarbalius
2,
Robertas Navakas
2 and
Algis Džiugys
2
1
Department of Mathematics and Natural Sciences, Kaunas University of Technology, K. Donelaičio st. 73, 44249 Kaunas, Lithuania
2
Laboratory of Heat-Equipment Research and Testing, Lithuanian Energy Institute, Breslaujos st. 3, 44403 Kaunas, Lithuania
3
Health Telematics Science Institute, Kaunas University of Technology, K. Donelaičio st. 73, 44249 Kaunas, Lithuania
4
Center for Radiology and Nuclear Medicine, Vilnius University Hospital Santaros Klinikos, Santariskiu st. 2, 08661 Vilnius, Lithuania
5
Department of Radiology, Lithuanian University of Health Sciences, Eiveniu st. 2, 50009 Kaunas, Lithuania
6
Clinic of Neurology and Neurosurgery, Faculty of Medicine, Vilnius University, M. K. Ciurlionio st. 21, 03101 Vilnius, Lithuania
7
Clinic of Anaesthesiology and Intensive Care, Faculty of Medicine, Vilnius University, M. K. Ciurlionio st. 21, 03101 Vilnius, Lithuania
*
Author to whom correspondence should be addressed.
Diagnostics 2024, 14(21), 2417; https://doi.org/10.3390/diagnostics14212417
Submission received: 16 September 2024 / Revised: 10 October 2024 / Accepted: 10 October 2024 / Published: 30 October 2024
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Objectives: Subarachnoid Hemorrhage (SAH) is a serious neurological emergency case with a higher mortality rate. An automatic SAH detection is needed to expedite and improve identification, aiding timely and efficient treatment pathways. The existence of noisy and dissimilar anatomical structures in NCCT images, limited availability of labeled SAH data, and ineffective training causes the issues of irrelevant features, overfitting, and vanishing gradient issues that make SAH detection a challenging task. Methods: In this work, the water waves dynamic factor and wandering strategy-based Sand Cat Swarm Optimization, namely DWSCSO, are proposed to ensure optimum feature selection while a Parametric Rectified Linear Unit with a Stacked Convolutional Neural Network, referred to as PRSCNN, is developed for classifying grades of SAH. The DWSCSO and PRSCNN surpass current practices in SAH detection by improving feature selection and classification accuracy. DWSCSO is proposed to ensure optimum feature selection, avoiding local optima issues with higher exploration capacity and avoiding the issue of overfitting in classification. Firstly, in this work, a modified region-growing method was employed on the patient Non-Contrast Computed Tomography (NCCT) images to segment the regions affected by SAH. From the segmented regions, the wide range of patterns and irregularities, fine-grained textures and details, and complex and abstract features were extracted from pre-trained models like GoogleNet, Visual Geometry Group (VGG)-16, and ResNet50. Next, the PRSCNN was developed for classifying grades of SAH which helped to avoid the vanishing gradient issue. Results: The DWSCSO-PRSCNN obtained a maximum accuracy of 99.48%, which is significant compared with other models. The DWSCSO-PRSCNN provides an improved accuracy of 99.62% in CT dataset compared with the DL-ICH and GoogLeNet + (GLCM and LBP), ResNet-50 + (GLCM and LBP), and AlexNet + (GLCM and LBP), which confirms that DWSCSO-PRSCNN effectively reduces false positives and false negatives. Conclusions: the complexity of DWSCSO-PRSCNN was acceptable in this research, for while simpler approaches appeared preferable, they failed to address problems like overfitting and vanishing gradients. Accordingly, the DWSCSO for optimized feature selection and PRSCNN for robust classification were essential for handling these challenges and enhancing the detection in different clinical settings.

1. Introduction

The sudden breaking of an aneurysm resulting in bleeding inside the subarachnoid space around the spinal cord and brain is called Subarachnoid Hemorrhage (SAH), where this space is generally filled with the colorless and clear fluid named cerebrospinal fluid [1,2]. Based on the cause, the SAH can be categorized into two types, traumatic and spontaneous. Spontaneous SAH causes 20% of acute cerebrovascular disorders due to hypertension, intracranial aneurysms, and spinal/brain arteriovenous malformations [3,4,5]. Spontaneous SAH is usually caused by the rupture of an intracranial aneurysm and by a diversity of other conditions like meningitis, malignant tumors, moyamoya disease, issues related to anticoagulation therapy, encephalitis, homological abnormalities, and brain vasculitis [6,7]. Traumatic SAH is the existence of blood in the subarachnoid space caused by a traumatic head/brain injury and is usually accompanied by a brain contusion. Arterial dissection, increase in intravascular pressure, damage in small arteries or veins, and direct blood extravasation are the major causes of traumatic SAH [8,9,10].
The computed tomography imaging modality is effective in the detection of SAH compared to other medical techniques like ultrasound, magnetic resonance imaging, positron emission tomography, etc. The NCCT images are effective in detecting bleeding in brain regions. In relation to other imaging techniques, the NCCT images are highly interactive and effectively find the bleeding, even when the leaked blood amount is small [11,12]. In SAH detection, the NCCT images include benefits like higher sensitivity to blood and less scanning time. The clinician’s diagnosis of SAH is based on the properties of lesions in the NCCT images [13,14,15]. However, it is difficult to obtain an accurate estimation of bleeding volume in the patients, particularly in the NCCT images. To overcome this problem, several automated models are designed related to machine learning and deep learning (DL) models [16,17]. The existing research faces many challenges such as failure to extract an appropriate feature, existence of irrelevant features, and ineffective training during the detection of the SAH grade. The addressing of aforementioned issues is considered a motivation of this research. Therefore, three pre-trained models such as GoogleNet, VGG-16, and ResNet50 were used for the effective acquisition of a wide range of patterns and irregularities, fine-grained textures and details, and complex and abstract features. Next, an optimum feature subset from an extracted feature was discovered using DWSCSO which aided in reducing the time and classifier’s complexity. The classification stability of PRSCNN was enhanced by avoiding the vanishing gradient issue based on the incorporation of PReLU.
In order to estimate the amount of leaked blood, i.e., the hemorrhage region, in a more acceptable time interval, a novel automated DL approach with feature selection was designed in the present article.
The major contributions are pointed out as follows:
  • Proposed a modified region-growing for segmenting the regions affected by SAH. Modified region-growing is a simple and efficient segmentation method. The seed points were not positioned precisely in the conventional region-growing method when the background had similar color and texture to the object of interest. It was overcome in the modified region-growing by selecting the seed points based on the orientation and intensity threshold values.
  • Integrated three pre-trained models (GoogleNet, VGG-16, and ResNet50) for extracting features from the segmented images. In that, GoogleNet, VGG-16, and ResNet50 were used to extract the wide range of patterns and irregularities, fine-grained textures and details, and complex and abstract features to enhance the SAH grade classification.
  • The use of high-level (semantic content and visual patterns) and low-level (color, texture, contrast, and edges) features decreased the semantic space between the vectors, which increased the success rate of SAH grade classification.
  • DWSCSO was developed for reducing the dimensions of extracted features; this enabled a significant reduction in the training time and complexity of the classifier. The water waves dynamic factor (WWDF) and wandering strategy were included in the DWSCSO for an effective adaptation to the complex operations and for increasing the exploration capacity.
  • The Parametric Rectified Linear Unit (PReLU) Stacked Convolution Neural Network model was used for stable SAH grade classification by avoiding the vanishing gradient issue. This PReLU was chosen because it offers smoother gradient flow by backpropagation that used to achieve stable training by avoiding the vanishing gradient issue.
The present article is organized in this manner: the papers related to the topic “SAH detection” are surveyed in Section 2. The detail about DWSCSO-PRSCNN-based grade classification of SAH, outcomes, and the conclusion are specified in Section 3, Section 4, and Section 5, correspondingly.

2. Literature Survey

The existing articles related to SAH detection are briefly reviewed in this section. Mansour, R.F and Aljehane, N.O [14] developed the DL-based intracranial hemorrhage (ICH) approach for optimal segmentation with the Inception Network. The segmentation over the CT images was accomplished by using Kapur’s thresholding with Elephant Herd Optimization (EHO). The features from the segmented portions were obtained using DL-based Inception v4 network while multilayer perceptron was used to perform the classification. The identification of optimum threshold using EHO enhanced the segmentation. The redundant features were included in the overall feature set subjected to affect the classification performances. Wang et al. [18] used XGBoost model with the ‘xgboost’ package from R version 3.6.1 to predict the outcomes of aneurysmal SAH patients. The early prediction performed by the XGBoost assists physicians in strengthening the clinical care and therapeutic strategies for aneurysmal SAH patients. In this study, the input data were collected from 351 aneurysmal SAH patients, who were admitted to a hospital in West China. From the collected data, 70% was utilized for model training and the residual 30% was utilized for model testing. When compared with logistic regression, the XGBoost obtained better results, but this model was quite sensitive to the outliers. Furthermore, Mohammed et al. [19] presented three systems for ICH detection with different methods and materials. In the first system, the low-dimensional vectors were extracted from the CT images by employing two global descriptors. The extracted low-dimensional vectors from the Gray Level Co-occurrence Matrix (GLCM) and Local Binary Pattern (LBP) were integrated with the high-level vectors from AlexNet, ResNet-50, and GoogleNet. Further, the feature dimensions were minimized by implementing the Principal Component Analysis (PCA) technique. In the second system, AlexNet, ResNet-50, and GoogleNet models were initially employed to extract vectors from the CT images, which were further given to the Support Vector Machine (SVM) to classify the feature maps. The third system was developed by employing only AlexNet, ResNet-50, and GoogleNet. Different evaluation measures revealed that the presented systems attained more promising results in ICH detection than the existing systems. However, the use of only pre-trained models increased the computational cost of the system.
In the context of SAH detection, Rau et al. [20] used a decision tree classifier for predicting isolated traumatic SAH patients with high mortality risk, but the traditional machine learning classifier (decision tree) includes two major issues, i.e., outliers and overfitting. Malik, P [21] presented the stacked deep model classifier which has numerous layers of neural networks where every layer was developed to obtain appropriate levels of features. These features were used to obtain an improved representation learning during the classification. However, an existence of inadequate features was subjected to affect the classification. Additionally, Lee et al. [22] developed an efficient deep learning classifier for detecting acute ICH and classifying the subtypes of ICH (no hemorrhage, epidural, subdural, intra-ventricular, and intra-parenchymal hemorrhages) in the CT images. The developed system not only showed comparable performance in ICH subtype classification but also in ICH detection (while aggregating various hemorrhages). Imran et al. [23] used U-Net for detecting the subtypes of ICH. In that literature, the U-Net model’s efficacy was validated utilizing evaluation measures like recall, F1-score, and precision. The results confirmed that the U-Net model achieved superior results in ICH subtypes classification compared with the comparative models (logistic regression, stacked CNN, and SVM). However, a huge number of parameters in the U-Net model leads to an overfitting problem.
A new architecture (combination of CNN with a Recurrent Neural Network (RNN)) was created by Liu et al. [24] for detecting and classifying the subtypes of ICH in the CT images. A new loss function was utilized in the developed architecture to neglect the label dependencies that help in resolving the problem of imbalanced data distribution. The experiments performed on a benchmark dataset showed the efficacy of the developed architecture over the traditional models. However, training the CNN-RNN model is a complex task and it leads to vanishing gradient problems while using a standard activation function. Kärkkäinen et al. [25] presented an unsupervised framework based on mixture models for classifying the healthy and hemorrhaged tissues. The performance of the presented framework was validated on a public dataset, which has different hemorrhage types with various intensities and sizes. In this study, the presented framework’s results were compared with different supervised and unsupervised models. The outcomes confirmed that the developed framework significantly improved more than the existing methods by means of dice score. Furthermore, Sage and Badura [26] implemented a double branch ResNet-50 model for detecting the subtypes of ICH. Compared with the random forest and SVM classifiers, the double-branch ResNet-50 model achieved higher classification results in light of detection accuracy and F1-score, but the double-branch ResNet-50 model had high complexity. Correspondingly, Barros et al. [27] presented a stacked CNN model for effective SAH segmentation in the CT images. Initially, trivial background voxels and trivial voxels were excluded from the collected CT images, and then new hemorrhage patches were generated with a sagittal plane by performing flipping. The generated hemorrhage patches were passed to the stacked CNN model to exclude the background patches with lower classification errors. One of the main drawbacks of the CNN model in SAH segmentation was the need for an enormous number of labeled images to effectively train the model, which was a time-consuming process and computationally expensive.
An automated model was implemented by Li et al. [28] for the effective segmentation and detection of SAH. Firstly, the posterior intersection points, brain boundaries, and anterior intersection points were computed from the CT images. Secondly, vectors were extracted concerning the computed points, and further, the feature vectors were passed to the Bayesian decision model to classify the grades of SAH. Generally, the Bayesian decision model generates an enormous number of parameters during the computationally expensive classification. In addition to this, Shahzad et al. [29] developed an effective deep-learning framework for detecting and segmenting aneurysms in SAH patients on the CT images. The developed framework has obtained only comparable segmentation performance on the CT images with artifacts. Sun et al. [30] presented a Deformable Attention U-Shaped Network (DAUN) for precise segmentation of SAH. A region boundary-aware loss optimizer was utilized in the DAUN to improve the accuracy of segmentation on the smaller lesions and uneven edges. Additionally, a supervised learning approach was used for training the DAUN to balance the position and semantic information of every pixel. The experiments carried out on the Monuseg dataset demonstrate the effectiveness of the DAUN over traditional models, but it has a high processing time.
Nagaraju, S et al. [31] implemented a Transfer Learning Ensemble (TLE) framework which was used to detect and classify the intracranial hemorrhage. The developed TLE was developed by incorporating different classifiers, such as XceptionNet, InceptionV3, Resnet50, VGG19, Desnet121 and VGG16. Subsequently, a voting classifier was utilized for identifying the best classifier with enhanced accuracy. However, appropriate features were required to be extracted for further enhancement of the classification performances. Korra, S et al. [32] developed a fully connected separable convolutional network which helped the clinician at the preliminary stage of the treatment process by obtaining the discriminative feature vectors from various layers. This research considered various data augmentation approaches such as shifting, brightness adjustment, scaling, horizontal flipping, and rotation for generating the numerous image variations that improve the generalization capacity during the classification. The optimum feature subset was essential for avoiding the irrelevant features, because it was subjected to cause misclassification.
SS [33] presented the deep learning classifier to perform the detection and classification of intracranial hemorrhage based on the stacked generalization ensemble approach. The ensemble approach was designed based on five transfer learning classifiers such as InceptionResNetV2, EfficientNetB0, NASNetMobile, InceptionV3, and MobileNetV2. Subsequently, soft voting was used to add the evaluated probabilities from the individual classifier. The developed ensemble approach used a meta-learning algorithm to learn how to effectively combine the estimations from the deep learning classifiers. The deep learning classifiers in the ensemble approach were used to learn the feature hierarchies, i.e., both the low- and high-level features. However, this stacked generalization ensemble approach failed to consider the optimum feature subset, because the features with redundant information were subject to affect the performances.
Taking into account the considered models and their features, a novel automated deep learning method/model is proposed in this article for comparably precise and efficient SAH segmentation and classification of SAH grades.
The issues found from the existing research are mentioned as follows: inadequate feature extraction, existence of irrelevant features, and inefficient training during the classification. The aforementioned issues are addressed in this research based on the following strategies: The combination of GoogleNet, VGG-16, and ResNet50 were used to obtain different level of features including a wide range of patterns and irregularities, fine-grained textures and details, and complex and abstract features. Irrelevant features were removed by DWSCSO, which helps to reduce dimensions of extracted features to enhance classification. Further, the smoother gradient flow obtained by backpropagation of PReLU in PRSCNN helped to avoid the vanishing gradient issue during classification.

3. Methods

The proposed automated SAH detection model includes five phases such as NCCT image collection (from clinical data sources, where various artifacts persist), region segmentation (applying modified region-growing method), feature extraction (using ResNet50, VGG-16, and GoogleNet), feature optimization (applying DWSCSO algorithm), and SAH grade classification (using the PRSCNN model). The automated SAH classification using DWSCSO and PRSCNN is represented in Figure 1.

3.1. Data Description

This research was analyzed in two different datasets: collected dataset and CT dataset [34].

3.1.1. Collected Dataset

The NCCT brain images of 49 SAH patients were acquired from two distinct university hospitals in Lithuania. Patient inclusion criteria were: older than 18 years, diagnosed cerebral aneurysm rupture, initial NCCT scans performed routinely at patient admission. In order to increase the sample size, the collected NCCT brain images are augmented and can be used to mark regions of interest, specifically highlighting areas indicative of SAH. The image augmentation was performed by employing flipping, shifting, and random rotation techniques to generate more training images, as performing calculations with limited data samples in the deep learning models leads to overfitting problems. The use of image augmentation techniques reduces data overfitting, prevents data scarcity, and improves the model’s efficiency in image segmentation and classification. In total, 1400 NCCT images were generated, of which 1120 images were used in a training set, and the remaining 280 images were used in a testing set. The sample-acquired NCCT images are given in Figure 2.

3.1.2. CT Dataset

The CT dataset for hemorrhages was acquired from the Near East Hospital in Cyprus. This dataset had 7032 CT images collected from 18 patients who had cerebral hemorrhages and 27 people who did not have cerebral hemorrhages. Accordingly, the dataset had 4343 healthy images and 2689 hemorrhagic images. The acquired images from the dataset were processed under the modified region-growing method to obtain the region of hemorrhage.

3.2. Region Segmentation

The region-growing method is one of the most efficient region-centric-based segmentation methods. It is also known as a pixel-based segmentation that involves seed point selection. In this case, the regions of seed pixels were grown by adding similar neighborhood pixels [35]. In conventional region-growing, the seed points are selected based on an intensity threshold value. The deviation in the intensity or noise causes over-segmentation or holes. To overcome the above-stated concern in the modified region-growing method, the seed points are selected based on the intensity and orientation threshold values [36,37]. In this segmentation, cross validation was utilized in the training set to fine-tune the threshold values. A various candidate threshold was analyzed by performing the regions of interest’s segmentation, and the segmentation performance was computed using the Dice Similarity Coefficient (DSC) and Jaccard Index (JI). The respective threshold values, which returned a higher DSC and JI, were taken as optimal thresholds. Additionally, the bootstrapping method was used to generate confidence intervals in the selected thresholds, confirming that the chosen values were statistically reliable and reproducible over diverse datasets. This helped to lessen over-segmentation and confirmed a reliable performance. The morphological operation was utilized in segmentation to avoid issues related to the unwanted regions. The steps involved in the modified region growing method are pointed out below, as follows:
  • Initially, the image gradients were computed using the sobel operator. Here, the gradient states the rate of change in pixel values in NCCT images. In the context of region growing, the sobel operator utilizes two convolution kernels to identify the variations in both the vertical and horizontal directions. These two kernels compute the gradient magnitude and direction for every pixel in an NCCT image. The region growing criteria were defined according to the gradient magnitude, and the gradient direction was used to guide the direction of region expansion.
  • Then, the NCCT images were segmented into different grids, S i , based on the orientation threshold, O t h , and intensity threshold, I t h . Here, the sobel operator and histogram analysis were used to identify the O t h and I t h , respectively. The sobel operator was used to identify the edges by evaluating the pixel’s gradient magnitude and orientation. Generally, it was used to discover the image intensity variation. This operator computed the level of intensity variation (magnitude) and direction of orientation. Equations (1) and (2) are used to compute the gradient ( E i j ) by the sobel operator.
E i j = e i j e m a x  
e i j = G x 2 + G y 2
where e max = max ( e i j ) , and the gradient value of components at x and y directions are represented as G x and G y , respectively.
  • The process is followed with respect to grids S i , until the total number of grids is similar to the number of grids in the NCCT images. This is conducted as follows:
    • In the grids S i , compute the histogram value L h of every pixel;
    • Select the frequency histogram value F h of the S i t h grid;
    • Select a pixel based on the frequency histogram value F h and assign a respective pixel as the seed point containing the orientation value O p and intensity value I p ;
    • Then, consider the adjacent pixels containing the orientation O n and intensity I n ;
    • Finally, determine the differences in orientation and intensity of the pixels n and p using Equations (3) and (4). The sample-segmented images are graphically presented in Figure 3, where the red area denotes the segmented portions.
df orientation = O n O p
df intensity = I n I p
The process of segmentation using modified region-growing is important, because it mainly impacts the capacity of feature extraction and classification because the poor segmentation causes noise or irrelevant details which affect classification. A precise segmentation separated the hemorrhagic regions from the images and confirmed that only appropriate areas were transferred to feature extraction. This helped in extracting the significant features that lead to the enhancement of the capacity of classifying the SAH grade. After performing the segmentation, the isolated portions were given as input to the feature extraction where the pre-trained models were utilized to extract appropriate features.

3.3. Feature Extraction

In this section, the feature extraction was performed utilizing three pre-trained models: ResNet50 [38], VGG-16 [39], and GoogleNet [40] to extract high-level and low-level vectors from the segmented regions. Specifically, the inception modules and pooling operations of GoogleNet were used to obtain the wide range of patterns and irregularities that exist in the NCCT images that were helpful in identifying the delicate irregularity. The VGG-16 extracted the fine-grained textures and details using its convolutional filters and deep layers. Further, the complex and abstract features were obtained using the residual connections of ResNet50.
ResNet50: It includes 50 CNN layers (1 average-pool layer, 1 max-pool layer, and 48 convolutional layers) for feature extraction. The residual neural network is an Artificial Neural Network (ANN), which creates networks by stacking several residual blocks.
VGG-16: It has 3 dense layers, 5 max-pool layers, and 13 convolutional layers to extract vectors from the segmented NCCT images. The VGG-16 sums up to 21 layers, but only 16 weighted layers are utilized for learning the parameters.
GoogleNet: It has 22 deep layers, and it works based on the so-called inception module. The GoogleNet model uses inception modules, and it allows the network to choose between several convolutional filter sizes in every block.
By utilizing the feature-averaging technique, the feature vectors from the ResNet50, VGG-16, and GoogleNet are combined. In total, 8728 vectors were obtained from segmented NCCT images, which were processed with the SCSO algorithm for optimizing the extracted vectors.

3.4. Feature Optimization

The feature vectors extracted from the NCCT images were passed to the DWSCSO algorithm for optimization. In that, the WWDF of dynamic factor was incorporated to effectively adapt to the complex operations and enhance the possibility of discovering optimum solution. The wandering strategy included a triangle walk scheme and Lévy flight (LF) was used to improve the robustness of exploration capacity. Currently, the SCSO is an efficient metaheuristic-based optimization algorithm, which works on the concept of swarm intelligence and mimics the sand cat’s hunting behavior [41,42]. The wild sand cat searches or attacks the prey based on the prey’s sound frequency, because every sand cat is impressible to sound frequency. In this algorithm, the initialization matrix was generated based on the size of the extracted vectors.

3.4.1. Exploration Phase (Searching for Prey)

The sand cat’s position is represented as P o s , and it even senses the prey below 2 kHz frequency. In conventional SCSO, a r G linearly decreases from 2 to 0 , however it does not adapt well to the difficult multivariate functions. Hence, the WWDF factor is used for considering the advantage of water wave dynamics, thus it adapts to complex operations and enhances the capacity of discovering the optimum solution. The utilization of water wave’s dynamics supports the population for searching over the extensive area, minimizing the blindness of remaining individuals, improving data exchange and learning among populations, population diversity maintenance, and avoiding local optima issue. Meanwhile, the control factor k is included for handling the magnitude decrement of r G and it is expressed in Equation (5). The parameter R in Equation (6) controls the exploitation and exploration ability of the DWSCSO algorithm [43].
r G = 2 × s × exp t T k × r
R = 2 × r G × r a n d 0,1 r G
where the random integer is denoted as s ; random function is r [ 0,1 ] ; k [ 1 , 3 ] ; maximum number of iterations is denoted as T ; the present iteration number is stated as t ; and the value of S M = 2 .
While searching a prey, every sand cat identifies a new position within its sensitivity range r , and it contributes to the exploitation and exploration algorithms. The parameter r is different for every sand cat that avoids falling into the local minima trap, and parameter r is represented in Equation (7). The parameter r G is used for guiding the parameter r .
r = r G × r a n d 0,1
The sand cat searches the prey position based on the parameter r , the current position P o s c ( t ) , and optimal candidate position P o s b c , and it is represented in Equation (8).
P o s t + 1 = r × P o s b c t r a n d 0,1 × P o s c t
A triangle walk scheme is included for the sand cat to move around as it reaches its prey. The distance L 1 is computed among the sand cat and its prey and computes the range of step size L 2 . Accordingly, it determines the walking direction of the sand cat using the Equations (9)–(13).
L 1 = P o s b c t P o s c t
L 2 = r a n d ( ) L 1
β = 2 × π × r a n d ( )
P = L 1 2 + L 2 2 2 × L 1 × L 2 × cos ( β )
P o s n e w = P o s b c t + r × P
where the location obtained via the walking scheme is denoted as P o s n e w .

3.4.2. Exploitation Phase (Attacking Prey)

The distance P o s r n d between the prey and sand cat is computed utilizing Equation (14) for simulating the process of attacking prey. Let us consider that the sand cat’s range of sensitivity is a circle, and the direction of motion utilizes the Roulette Wheel Selection (RWS) scheme for selecting the random angle α . The random angle is chosen among 0 ° and 360 ° and its value ranges between −1 and 1. Further, the prey is attacked based on Equation (15).
P o s r n d = | r a n d 0,1 × P o s b c t P o s c ( t ) |
P o s t + 1 = P o s b c t r × P o s r n d × c o s ( α )
LF is an enhanced approach that adds randomness to the exploitation phase. LF provides a random wandering method with a step length supporting the Lévy distribution. Sometimes, the LF has higher step length, so for making it consistent with the behavior of the sand cat, it is multiplied by the constant ( C = 0.35 ) . This makes the sand cat walk as near as possible. The walking strategy using LF is expressed in Equation (16).
P o s n e w = P o s b c t + P o s b c t P o s c t + r × P
By regulating the parameters R and r G , the algorithm controls the exploitation and exploration capability. If R is less than or equal to one, the prey is attacked by the sand cat, otherwise, the sand cat searches for the prey. This scenario is mathematically presented in Equation (17), and DWSCSO ends once it reaches the maximum iterations.
P o s t + 1 = r × P o s b c t r a n d 0,1 × P o s c t R > 1 ; e x p l o r a t i o n P o s b t P o s r n d × c o s ( α ) × r R 1 ; e x p l o i t a t i o n
In this algorithm, the accuracy from K-nearest neighbor (KNN) is considered as the fitness function for selecting the important features, and the time complexity is based on the sand cat’s population size and number of iterations. The DWSCSO’s time complexity is presented in Equation (18).
O DWSCSO = O defined   parameters + O location   update + O ( population   initialization )
The parameters considered in the DWSCSO are listed as follows: size of population was 100, RWS was [0, 360], S M was 2, and total iterations were 100. From the extracted 8728 vectors, the DWSCSO algorithm selected 5290 vectors, which were passed to the PRSCNN model for SAH grade classification. The pseudocode and flowchart of the DWSCSO-based feature optimization is depicted in Algorithm 1 and Figure 4.
Algorithm 1 Pseudocode of the DWSCSO algorithm
//Step 1: Initialization
Input: Maximum iterations T, population size N, fitness function F, WWDF, α, r, rG
//Initialize population P of N search agents (feature vectors)
For each search agent i in P do search solution space
  Solution space states the range of probable feature values for every agent.
End
//Step 2: Estimate the Initial Fitness
For every search agent i in P do
  Calculate fitness Fi for agent i using the fitness function F
  Fi = Accuracy(knn_classifier(X_train, y_train))
End
//Step 3: Main Loop—Iterate over Maximum Iterations T
For iteration t = 1 to T do
  //Step 3.1: For every search agent, accomplish exploration or exploitation
  For every search agent i in P do
    //Step 3.1.1: Select a random angle α for direction of movement (0° ≤ α ≤ 360°)
  α = random_angle() //Randomly selected using RWS.
    //Step 3.1.2: Identify if exploration or exploitation is to be executed
     If |rG| > 1 then //Exploration phase
      //Step 3.1.2.1: Exploration—Move to a new position using Triangle Walk Scheme
      new_position_Xi = Xi + WWDFrcos(α) ∗ random_step()
     Else //Exploitation phase
      //Step 3.1.2.2: Exploitation—Move using LF
      new_position_Xi = Xi +levy_flight() ∗ (best_positionXi)
      //LF generates a random step based on Lévy distribution.

     End
    //Step 3.1.3: Verify if new position is valid
     If new_position_Xi is invalid (e.g., NaN, Inf, out-of-bounds) then
      //Reset the agent’s position to a valid random location in the search space
       new_position_Xi = random_valid_position()
     End
    //Step 3.1.4: Evaluate the fitness of the new position
    new_fitness_Xi = evaluate_fitness(new_position_Xi)
    //Fitness function estimates the new feature vector for classification accuracy.
    //Step 3.1.5: Update agent’s position and fitness if the new position is better
     If new_fitness_Fi > current_fitness_Fi then
       Xi = new_position_Xi//Update the agent’s position
       Fi = new_fitness_Fi //Update the agent’s fitness
     Else
       Retain current position Xi and fitness Fi
     End
  End
  //Step 3.2: Convergence Check—Monitor improvement
  If no substantial improvement in fitness after X consecutive iterations then
    //Enhance the step size in LF to escape local optima and encourage exploration
     WWDF = WWDF ∗ 1.5 //enhance an exploration factor to cover more search space
  End
  //Step 3.3: Adjust parameters dynamically
  Adjust r and rG based on the iteration number t
End
//Step 4: Return the best feature vector
Determine an agent with greatest fitness score
Return best_position (best feature vector) and corresponding fitness

3.4.3. Error Handling and Edge Case Considerations

The handling of error and edge case considerations are used for confirming the robustness and common applicability of the optimization algorithm. DWSCSO is developed with the following strategies for addressing the convergence issues, local minima traps, and invalid inputs.

Convergence Issues

The WWDF dynamically adjusts in the iterative process to eliminate inactivity during the optimization, which was used to confirm that the population remains to explore different areas of the solution space. The WWDF factor improves the adaptability of DWSCSO in difficult and higher dimensional search spaces. Moreover, the LF is used for incorporating randomness to the exploitation to avoid premature convergence. Therefore, dynamic alteration avoids the inactivity and confirms the convergence towards the global optimum.

Local Minima Traps

The DWSCSO developed with the triangle walk scheme and LF was used to avoid the local minima risk. Premature convergence was avoided by avoiding the repetitive patterns using the triangle walk scheme. Moreover, the risk of local optima is additionally reduced by continuously altering the angle of approach and attack using RWS.

Invalid Inputs Handling

The DWSCSO verifies the input ranges for confirming that the given inputs are in the boundaries or not. If the values are not set based on requirements, the default values are allocated in the optimization. The fitness function computes the feature vector quality chosen in the optimization process. Moreover, if a population faces an invalid position, the agent position is returned to a randomly chosen valid position in the search space. The fitness estimation function is also developed for handling an edge case by confirming all values are constrained within satisfactory boundaries. The fitness function evaluates the updated position. Next, the fitness function eliminates the errors that interrupt the optimization to confirm that the inputs transferred to the fitness function are valid. If any location returns to NaN or Inf, the population is initialized again utilizing a random position in exploration.

Outlier Detection and Robustness

The fitness function integrates outlier detection by evaluating the fitness values over the population. The DWSCSO considers the probability of either invalid input or local minima trap when the outlier is identified, i.e., fitness of population is lesser or higher than the average of population. In these situations, the exploration is enabled, making the affected population set its location based on random searches in a wide area of solution space. This confirms that the DWSCSO is robust even in the presence of corrupted or noisy data.
Therefore, regarding the incorporation of error-handling mechanisms and edge case considerations, the DWSCSO is developed to make it robust and adaptable in an extensive range of optimization tasks. Subsequently, the selected features from the DWSCSO are given as an input to the PRSCNN to accomplish SAH grade classification.

3.5. SAH Grade Classification

The selected 5290 vectors were given to the PRSCNN for SAH grade classification. The PRSCNN model was designed with distinct layers such as an input layer, convolutional layer, PReLU activation layer, pooling layer, and flattened/fully connected layer [44,45,46]. The incorporated PReLU had smoother gradient flow due to the backpropagation process. This smoother flow of gradient leads to stable and effective training which helps to avoid the vanishing gradient issue. The PReLU has the ability to allow the negative values in the training process, which helps to offer a better depiction of the data. Accordingly, it was used to enhance the generalization of unseen data. A brief description of the layers are discussed below, as follows:
Input layer: The selected 5290 vectors are fed to the input layer, and it is represented as a three-dimensional matrix. The dimension is W × H × D , where W is represented as width, H is denoted as height, and D is indicated as depth, where the depth corresponds to the color channels.
Convolutional layer: The convolutional layer computes the output of the nodes, which are interconnected to the local regions of the matrix. Here, the dot product is computed between the values related to an input: local region and a set of weights (filter).
PReLU activation layer: The PReLU function represented in Equation (19) is used to avoid the issue of dying ReLU and it performs well with the negative inputs as well as allows backpropagation. The incorporation of negative parts as input offers reliable predictions. Moreover, it helps to avoid the vanishing gradient issue.
f x i = x i ,                         i f   x i > 0 α i x i ,               otherwise
where x i denotes only one element from vector x ; input vector is denoted as x ; and learnable parameter is denoted as α i .
Pooling layer: Uses down-sampling technique to reduce the height and width of the convolved features.
Fully connected or flattened layer: The convolved features are fed to the flattened layers here. The class probabilities are calculated and outputted in a three-dimensional array with dimensions of 1 × 1 × K , where K is the number of classes (four grades). Grade I is depressed consciousness level or focal deficit, grade II is mild alteration, grade III is severe headache, and grade IV is mild headache.
The parameters of the PRSCNN model are the initial learning rate, which is 0.001, the loss function is log loss, the drop factor of learning rate is 0.2, the drop period of learning rate is 5, the maximum epochs is 100, the minimum batch size is 500, the activation function is ReLU, the optimizer is Adam, and the momentum rate is 0.9.

4. Results and Discussion

The proposed modified region-growing method and DWSCSO-PRSCNN model are simulated using Python 3 on a system with the specifications of a windows operating system, Intel core i10 12th-generation processor, Santa Clara, CA, USA, and 64 GB random access memory. The modified region-growing method’s effectiveness was analyzed using evaluation measures like the Jaccard Index (JI), Dice Similarity Coefficient (DSC), PA, and MPA. Correspondingly, the classification model’s (DWSCSO-PRSCNN) effectiveness was analyzed using evaluation measures like accuracy, Matthews Correlation Coefficient (MCC), and F1 score on a collected dataset where an 80:20 ratio is considered for training and testing purposes.

4.1. Evaluation Measures

The explanation about the undertaken evaluation measures (JI, DSC, PA, MPA, accuracy, MCC, and f1 score) is detailed in this subsection. The JI is determined as the area of intersection between the ground truth region and segmentation region and divided by the area of union between the ground truth region and segmentation region. Additionally, the DSC efficiently calculates the overlap among the ground truth and segmentation region by performing intersection over union between two sets, where A indicates a segmented region by performing the modified region-growing method and B represents ground truth region. The mathematical formulas of JI and DSC are depicted in Equations (20) and (21).
JI = TP TP + FP + FN = | A B | | A B |
DSC = 2 TP 2 TP + FP + FN = 2 | A B | A + | B |
Similarly, the PA is determined as the ratio of precisely segmented pixels divided by the total pixels. For instance, the PA of K + 1 classes (background and K foreground classes) is mathematically denoted in Equation (22). PA can be treated as a semantic segmentation metric, which represents the percentage of pixels which are precisely classified in an image. In addition, the MPA is determined as the ratio of precisely segmented pixels divided by the average of total pixels, and it is mathematically specified in Equation (23).
PA = i = 0 K p i i i = 0 K j = 0 K p i j
MPA = 1 K + 1 i = 0 K p i i j = 0 K p i j
The evaluation measure: Accuracy estimates how many times a classification model (DWSCSO-PRSCNN) made a correct prediction in the entire collected dataset. F1 measures a classification model’s (DWSCSO-PRSCNN) accuracy by combining the recall and precision. Additionally, the Matthews correlation coefficient (MCC) considers all four values (True Negative (TN), True Positive (TP), False Positive (FP), and False Negative (FN)) in the confusion matrix to estimate the efficacy of the classification model. The MCC ranges between −1 and 1, where −1 represents completely wrong multiclass classification and 1 indicates correct multiclass classification. The expressions to compute accuracy, F1 score, and MCC are represented in Equations (24)–(26).
Accuracy = TP + TN TP + TN + FP + FN × 100
MCC = TP × TN FP × FN ( TP + FP ) ( TP + FN ) ( TN + FP ) ( TN + FN ) × 100
F 1   Score = 2 TP FP + 2 TP + FN × 100

4.2. Quantitative Analysis Related to Segmentation

The simulation results of different segmentation methods for collected and CT datasets are depicted in Table 1 and Table 2. The FCM clustering, Otsu thresholding, K-means clustering, superpixel clustering and region-growing are considered for the evaluation of modified region-growing, because all these approaches are parameter sensitive. Viewing the tables shows that the modified region-growing method achieved precise segmentation results on a collected dataset with a JI of 0.94, DSC of 0.95, PA of 0.93, and MPA of 0.90, which are better than comparative segmentation approaches like Fuzzy C Means (FCM) clustering, Otsu thresholding, K-means clustering, and the region-growing method. The clustering based approaches mainly depend on the predefined clusters and initialized centroids. Accordingly, the clustering based segmentation returns the less defined edges, due to the complexity in the identification of fine details and edges. On the other hand, the modified region-growing approach is effective for the images with indefinite textures and intensities, therefore it adaptively enlarges the region according to the pixel similarity instead of depending on the predefined clusters. Consequently, the developed modified region-growing leads to precise delineation of structures even with the complex images. Therefore, an integration of contextual information from adjacent pixels helps to obtain more effective segmentation than the clustering approaches.
If the original NCCT images have clear edges, the modified region-growing method provides good segmentation results. Additionally, the modified region-growing method precisely separates the regions, which have similar properties, compared with the traditional segmentation methods. The modified region-growing method consumed a minimal processing time of 6.32 s for region segmentation in the collected dataset. The comparative segmentation methods—FCM clustering, Otsu thresholding, K-means clustering, and the region-growing method—consumed processing times of 11.32 s, 10.11 s, 15.44 s and 8.20 s in the collected dataset, respectively.

4.3. Quantitative Analysis Related to Feature Extraction

This section provides an analysis about the combination of all models, i.e., GoogleNet, VGG-16, and ResNet50 used in the feature extraction with the individual model’s comparison to know the effectiveness of the combined model. Additionally, some state-of-the-art approaches such as VGG-19 and SqueezeNet developed using Python 3.7, Keras 2.3 and TensorFlow 1.5 were used for analysis. The combined model was the CNN model, so the aforementioned CNN models were used in the comparison. Table 3 and Table 4 provide the results analysis of feature extraction approaches for collected and CT datasets, respectively. Moreover, the GoogleNet, VGG-16, VGG-19, SqueezeNet, ResNet50, and a combined model (GoogleNet + VGG-16 + ResNet50) utilized the computational time of 8 ms, 15 ms, 16 ms, 13 ms, 10 ms, and 33 ms per image of collected dataset. This analysis demonstrated that the combination of all three models provides a more enhanced performance in SAH grade classification than the individual models. The computational time of the combined model was higher when it was analyzed with individual model. However, the combination of GoogleNet, VGG-16, and ResNet5 helps to obtain the wide range of patterns and irregularities, fine-grained textures and details, and complex and abstract features for enhancing the SAH grade classification. The reason for not selecting a less resource-intensive method like SqueezeNet is that it offers less computational costs, but it does not obtain the fine-grained details and complex abstractions that are required for effective SAH detection.
An important trade-off when considering the combination of GoogleNet, VGG-16, and ResNet50 is the increase in computational complexity that causes a higher memory usage, training time, and processing time in both the training and inference phases. However, the obtained performance enhancement from this combined model in clinical task justifies the trade-off in computational time. The developed DWSCSO in feature selection minimizes the overall feature dimension by around 40% and that helps to minimize the training time and complexity.

4.4. Quantitative Analysis Related to Classification

In this research, the important contribution was to perform feature selection using the DWSCSO. Therefore, the developed DWSCSO was analyzed with different optimization algorithms like the Genetic Algorithm (GA), Butterfly Optimization Algorithm (BOA), Artificial Bee Colony (ABC), Whale Optimization Algorithm (WOA), and SCSO. At first, the optimization algorithm was analyzed with different sizes of population, such as 10, 20, 30 and 40, as shown in Figure 5 for the collected dataset. Based on Figure 5, it was concluded that the optimization algorithm with a population size of 30 provided enhanced classification results. On the other hand, the convergence analysis for the collected dataset was performed as depicted in Figure 6. The developed DWSCSO had improved convergence more than the GA, BOA, ABC, WOA and SCSO. The dynamic factor of WWDF was incorporated for adapting according to the complex operations and enhanced the probability of identifying the optimum solution. On the other hand, the wandering strategy was utilized to enhance the exploration capacity. Therefore, both the WWDF and wandering strategy used in the DWSCSO helped to improve the convergence.
The simulation results of the different optimization algorithms with a PRSCNN model were depicted in Table 5 and Table 6 for the collected and CT datasets, respectively. By investigating Table 5 and Table 6, the combination of the DWSCSO algorithm with the PRSCNN model achieved maximum classification results compared with the comparative optimization algorithms. The feature dimensionality reduction or selection of discriminative vectors reduced the training time of the PRSCNN model to 32.22 s for the collected dataset. The DWSCSO achieved an accuracy of 99.48% for the collected dataset that was higher than the remaining optimization algorithms. The higher accuracy over the classification was obtained due to an effective search of the optimum features by the DWSCSO. The water wave dynamics from WWDF was used to search over an extensive area for optimum features and enhanced the population diversity maintenance while the LF was used to enhance the robustness of the exploration capacity of features.
The computational complexity of the DWSCSO for the collected dataset was analyzed in terms of computational time with GA, BOA, ABC, WOA, and SCSO, as shown in Table 7. This evaluation demonstrated that the DWSCSO had lesser computational time due to its improved convergence obtained by the incorporation of WWDF and wandering strategy in the DWSCSO. Moreover, the statistical test, i.e., the Friedman test, was used for the proposed DWSCSO with the collected dataset, as shown in Table 8. This test denoted that the DWSCSO had better rank than the optimization methods. The p-value represents the significant difference among the evaluated methods.
The PRSCNN is a deep learning classifier, therefore it is evaluated with some state-of-the-art deep learning classifiers such as the decision tree, Graph Convolutional Network (GCN), Artificial Neural Network (ANN), Autoencoder, and Convolutional Neural Network (CNN). The simulation results of different classification models are presented in Table 9 and Table 10 for the collected and CT datasets, respectively. Here, the analysis is performed for actual feature vectors and optimized feature vectors using the DWSCSO algorithm. The parameters considered in the comparative classification models are pointed out below, as follows:
  • Decision tree (criterion is Gini, splitter is best, and maximum depth is 30);
  • GCN (layer is 3, hidden size is 64 and dropout rate is 0.2);
  • ANN (learning number is 13, learning rate is 0.001, and target error is 0.001);
  • Autoencoder (dropout rate is 0.5, epoch is 100, batch size is 128, and learning rate is 0.001);
  • CNN and PRSCNN (layer is 7, kernel size is 3 × 3, filters/channels per layer is 128, pooling size is 2 × 2, learning rate = 0.001, batch size = 32, num epochs = 10, and regularization weight = 0.0001).
The accuracy of PRSCNN over 10 multiple runs for the collected dataset was 99.45%, 99.47%, 99.48%, 99.47%, 99.49%, 99.51%, 99.46%, 99.53%, 99.49%, and 99.46% where the standard deviation was 0.0234. Moreover, the accuracy of PRSCNN with augmentation for the collected dataset was 99.48% while without incorporating the data augmentation, it was 96. 18%, which was lower due to the overfitting issue. By examining tables, the combination of the DWSCSO algorithm with the PRSCNN model obtained higher classification results in the collected dataset with a classification accuracy of 99.48%, MCC of 99.53%, and F1 score of 99.48%, respectively. The obtained classification results were higher than the existing classification models such as the decision tree, GCN, ANN, Autoencoder, and CNN. Next, the analysis of PReLU with different activation functions such as the Rectified Linear Unit (ReLU), Leaky ReLU, and Exponential Linear Unit (ELU) is given in Table 11 and Table 12 for the collected and CT datasets, respectively. This result demonstrates that the PReLU had a better performance than the ReLU, Leaky ReLU, and ELU.
In this article, the PRSCNN model efficiently captures the patterns and spatial relationships in the NCCT images. The PRSCNN model learns complex features by stacking several convolutional and pooling layers resulting in higher classification accuracy. The computational time of ReLU, Leaky ReLU, ELU, and PReLU are 0.1 ms, 0.2 ms, 0.5 ms, and 0.5 ms per image, respectively. However, the PReLU obtains smoother gradient flow by backpropagation which helps to obtain stable training for mitigating the vanishing gradient issue. The PReLU has the capacity of processing negative values which additionally helps to obtain reliable prediction. PReLU has a better performance because it learns the negative values during the training, offering enhanced flexibility and adaptability that helps to discover the subtle dissimilarities in SAH grades.
The Receiver Operating Characteristics (ROC), Area Under the Curve (AUC), and confusion matrix for the collected dataset are shown in Figure 7 and Figure 8, respectively. In ROC curve, the blue dashed line denotes the no-skill line that is used to evaluate the performance of developed classifier, while orange line denotes the ROC curve of developed classifier. The ROC curve was utilized for the classification issued in various threshold settings while the AUC denoted the degree of separability measure. On the other hand, the confusion matrix was utilized to summarize the identification of the classification issue. The ROC is a probability curve which was used to determine the level of differentiating among the classes. The ROC curve of Figure 7 represents that PRSCNN has a higher AUC, hence it provides better classification than the other classifiers. Moreover, the confusion matrix of PRSCNN depicts that it has lesser misclassification when compared with the decision tree, GCN, ANN, Autoencoder, and CNN. The misclassification in PRSCNN occurred because some samples from the collected dataset had lesser variation between the different grade images.
Figure 9 shows the accuracy comparison of the collected dataset for all the classifiers. Based on Figure 9, it is confirmed that the PRSCNN had better classification than the decision tree, GCN, ANN, Autoencoder, and CNN. Moreover, the issue of overfitting in PRSCNN was minimized by incorporating the DWSCSO-based feature selection.
The loss graphs for the different activation functions such as ReLU, Leaky ReLU, ELU, and PReLU are shown in Figure 10. This analysis confirms that the PReLU had a better performance than the ReLU, Leaky ReLU, and ELU.
Additionally, the DWSCSO-PRSCNN model was analyzed using K-fold cross validations. In the context of image classification, cross-fold validations provided a robust estimation of the DWSCSO-PRSCNN model’s performance by assessing its generalization ability across various data subsets. This process decreased the overfitting risk to a particular training–testing split. By inspecting Table 13 and Table 14, the proposed DWSCSO-PRSCNN model achieved high classification outcomes in five-fold cross validation (80%:20% training and testing) than the other cross fold validations: two-fold (50:50% training and testing), four-fold (75:25% training and testing), and eight-fold (87.50%:12.50% training and testing).
Edge cases such as hemorrhages with uneven shapes or those positioned near anatomical boundaries cause extra difficulties in the modified region-growing segmentation approach. This segmentation has the complexity in differentiating among the delicate variations in intensity and cause over-segmentation of missed detections. In the scenario of edge cases, the model sensitivity was observed to reduce by 2.1%, mainly due to the misclassification of adjacent structures as hemorrhages. However, the overall performance still remained higher, and these situations emphasized areas for additional improvement in the segmentation process.

4.5. Quantitative Analysis Related to Noisy Images and Handling of Missing Values

Table 15 shows the analysis of the DWSCSO-PRSCNN with noise simulated and noiseless images for further justifying the effectiveness. Noise in the medical images affected the capacity of the model while obtaining the meaningful features, specifically in NCCT images of the brain. NCCT images are vulnerable to artifacts such as low contrast, motion blur, and scanner noise among the affected regions and healthy tissue. These issues affect the segmented region quality that leads to the affecting of the soloing feature extraction. The modified region-growing method generates the over-segmented or under-segmented regions when the input has higher levels of noise. For instance, noise causes the model to discover false positives by discovering portions that appear identical to the hemorrhagic regions. Accordingly, an increment in the FP leads to a decrease in accuracy from 99.68% to 97.32%.
The performance of the PRSCNN for the analysis of handling missing values simulated for 10% of the data is shown in Table 16. Incomplete or missing data is common in real-world clinical settings where specific images or patient information may be corrupted or unavailable. A missing value in NCCT features affects the feature selection of DWSCSO and leads to misclassification. Specifically, the sensitivity is affected, because the model failed to precisely classify hemorrhagic cases.

4.6. Comparative Analysis

This section provides the comparative analysis for the developed DWSCSO-PRSCNN model with a standard benchmark dataset, i.e., the CT dataset [34], where an 80:20 ratio was considered for training and testing purposes. This comparison of DWSCSO-PRSCNN with the CT dataset was performed with DL-ICH [14] and GoogLeNet + (GLCM and LBP) [19], ResNet-50 + (GLCM and LBP) [19], and AlexNet + (GLCM and LBP) [19], as shown in Table 17. The reason for choosing the aforementioned methods is that all the methods come under the category of deep learning classifiers. It demonstrates that the DWSCSO-PRSCNN outperforms better than the existing approaches. The enhanced searching capacity of DWSCSO in feature selection using the WWDF and wandering strategy helps to select an optimum feature subset for enhancing the classification. Further, the mitigation of the vanishing gradient issue and obtaining the stable training enhances the SAH grade classification.
The p-value for accuracy of DWSCSO-PRSCNN was less than 0.001, representing that there was a statistically significant difference among the accuracy of DWSCSO-PRSCNN and other methods. This demonstrates that variation in the performance of DWSCSO-PRSCNN and other methods is not because of random variation but denotes the actual enhancements in the DWSCSO-PRSCNN performance. The confidence interval was computed for accuracy to additionally compute the ambiguity around the mean values. For a 95% confidence level, the intervals for accuracy of DWSCSO-PRSCNN from the results, i.e., accuracy, were from 99.48% to 99.76%. This narrow confidence interval represents the higher level of confidence in the stated results, denoting the reliability and robustness of DWSCSO-PRSCNN performance. After assessing the significance via ANOVA, the Tukey’s Honest Significant Difference was applied to perform post-hoc analysis for discovering the specific groups which varied from each other. This Tukey test validated that the DWSCSO-PRSCNN significantly improved than the other methods where p-value was 0.001 for all pairwise evaluations. This additionally represented that DWSCSO-PRSCNN was statistically best in detecting and classifying the grade of SAH.

4.7. Discussion

As depicted in the earlier sections, the region segmentation of SAH and classification of SAH are integral parts of this article. A modified region-growing method was proposed in this article for precise segmentation of the regions affected by SAH. The effectiveness of the modified region-growing method is depicted in Table 1, where the proposed segmentation method not only improves the segmentation accuracy but also decreases the processing time of the segmentation. Correspondingly, in the classification phase, the combination of the DWSCSO algorithm with the PRSCNN model improved the performance of SAH grade classification more than the traditional classification models and optimization algorithms. The optimal high-level and low-level vectors (selected by the DWSCSO algorithm) were passed to the PRSCNN model for SAH grade classification. The selection of optimal high-level and low-level vectors decreased the computational time to 32.22 s and even the model complexity to linear. The efficiency of the classification model (DWSCSO-PRSCNN) was specified in Table 3, Table 4, Table 5, Table 6, Table 7, Table 8, Table 9, Table 10, Table 11, Table 12, Table 13, Table 14, Table 15, Table 16 and Table 17. The early detection of SAH assists clinicians in timely treatment and efficient therapeutic intervention, which reduces the mortality rate. Moreover, the developed DWSCSO-PRSCNN also provides better performance with the CT dataset than the DL-ICH [14] and GoogLeNet + (GLCM and LBP) [19], ResNet-50 + (GLCM and LBP) [19], and AlexNet + (GLCM and LBP) [19]. Therefore, it is confirmed that the DWSCSO-PRSCNN has better generalization in both the collected dataset and CT dataset. The developed research is clinically beneficial as it provides an automated framework for a precise estimation of blood leakage in SAH cases, which is significant for prompt and effective treatment. The development of rapid and precise segmentation and classification of SAH minimizes the dependency on manual interventions which are prone to errors and time-consuming processes. Thus, the developed research is useful in enhancing diagnostic efficiency in emergency settings, improving decision making, and possibly enhancing patient progress by confirming rapid treatment, making it appropriate for both present and upcoming clinical applications.
The DWSCSO-PRSCNN is trained and analyzed using two different datasets: the collected dataset and CT dataset. However, the changes in the imaging equipment, patient populations, and scanning protocols create differences in the dataset features, impacting the generalization of the model to unseen data. These changes are being considered in upcoming data collection. For example, the images from different hospitals show the variation between resolution, noise level, and contrast setup that leads to the increase in complexity during feature extraction. Furthermore, the issue creates a class imbalance in the dataset, specifically when handling subtle representation or rare cases of SAH. In real-world scenarios, some SAH grades are underrepresented, which makes the model become biased towards the more frequent classes.

5. Conclusions

This research article focuses on two problems in the SAH analysis: DWSCSO-based feature selection and the PRSCNN-based classification of SAH. A modified region-growing method was introduced for segmenting the affected regions in the collected real-time NCCT images. Optimum features were discovered using DWSCSO with the WWDF and wandering strategy. The WWDF used in DWSCSO helped to search over extensive areas for optimum features and enhanced the population diversity maintenance while the LF enhanced the robustness of exploration capacity. Next, the PReLU obtained smoother gradient flow by backpropagation, which helped to achieve the stable classification in the PRSCNN by avoiding the vanishing gradient issue. Accordingly, reliable predictions were obtained in the PRSCNN by processing the negative values. Further, the training time and overfitting risk of the classification model were reduced by selecting the optimum feature vectors. Compared with traditional segmentation methods (FCM clustering, Otsu thresholding, K-means clustering, superpixel clustering, and region growing), the modified region-growing method achieved a high JI in the collected dataset at 0.94, with a DSC of 0.95, PA of 0.93, and MPA of 0.90. The combination of the DWSCSO algorithm with the PRSCNN model achieved high classification results in the collected dataset, i.e., the F1 score was 99.48%, with a MCC of 99.53% and accuracy of 99.48%, which were better than other combinations. Additionally, the developed DWSCSO-PRSCNN also provided better performance in the CT dataset than the DL-ICH and GoogLeNet + (GLCM and LBP), ResNet-50 + (GLCM and LBP), and AlexNet + (GLCM and LBP), which proves the generalizability of detection. From the ANOVA results, post-hoc analysis, and confidence intervals, the DWSCSO-PRSCNN denoted more statistically significant enhancements than the existing methods. These outcomes offer robust proof that the DWSCSO-PRSCNN has better performance in SAH grade detection. However, the generalizability of the model is limited to the variations in imaging protocols and data from different clinical settings. Moreover, the developed PRSCNN is sensitive to noise, missing data, and class imbalance in real-world scenarios.
As a future extension, automated hyperparameter tuning, pruning inactive features, handling of missing data, and cross-domain adaptation can be developed for addressing the effectiveness of the model when processed with large-scale clinical applications. In addition, inactive feature vectors in the CNN feature extraction can be eliminated to further enhance the performance of SAH detection.

Author Contributions

Conceptualization, J.S. and M.Š.; Methodology, V.R., A.P., G.S. and A.D.; Software, J.S.; Validation, S.L. and A.D.; Formal analysis, T.I. and E.M.; Investigation, T.I., E.M., G.S. and R.N.; Resources, V.P. and I.L.; Data curation, V.P., A.B., V.R., I.L. and R.N.; Writing—original draft, S.L.; Writing—review & editing, A.B. and M.Š.; Supervision, R.A.; Project administration, R.A.; Funding acquisition, J.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the European Regional Development Fund under grant agreement with the Research Council of Lithuania (No. 01.2.2-LMT-K-718-03-0091).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Vilnius and Kaunas Regional Biomedical Research Ethics Committees (Protocol No. 2021/9-1370-847, 2021-09-21 and BE-2-113, 2021-09-09, respectively) during September 2021–August 2024. The study was registered at the Protocol registration results system, ClinicalTrials.gov (ID NCT06006975).

Informed Consent Statement

Informed consent was obtained from all subjects and/or their legal guardians for the use of their anonymized clinical data in the retrospective analysis and publication of this paper.

Data Availability Statement

The datasets used and analyzed during the current study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Stetzuhn, M.; Tigges, T.; Pielmus, A.G.; Spies, C.; Middel, C.; Klum, M.; Zaunseder, S.; Orglmeister, R.; Feldheiser, A. Detection of a Stroke Volume Decrease by Machine-Learning Algorithms Based on Thoracic Bioimpedance in Experimental Hypovolaemia. Sensors 2022, 22, 5066. [Google Scholar] [CrossRef] [PubMed]
  2. Malinova, V.; Tsogkas, I.; Behme, D.; Rohde, V.; Psychogios, M.N.; Mielke, D. Defining Cutoff Values for Early Prediction of Delayed Cerebral Ischemia after Subarachnoid Hemorrhage by CT Perfusion. Neurosurg. Rev. 2020, 43, 581–587. [Google Scholar] [CrossRef]
  3. Kaur, S.; Singh, A. A New Deep Learning Framework for Accurate Intracranial Brain Hemorrhage Detection and Classification Using Real-Time Collected NCCT Images. Appl. Magn. Reson. 2024, 55, 629–661. [Google Scholar] [CrossRef]
  4. Yu, D.; Williams, G.W.; Aguilar, D.; Yamal, J.; Maroufy, V.; Wang, X.; Zhang, C.; Huang, Y.; Gu, Y.; Talebi, Y.; et al. Machine Learning Prediction of the Adverse Outcome for Nontraumatic Subarachnoid Hemorrhage Patients. Ann. Clin. Transl. Neurol. 2020, 7, 2178–2185. [Google Scholar] [CrossRef]
  5. Zeiler, F.A.; Lo, B.W.Y.; Akoth, E.; Silvaggio, J.; Kaufmann, A.M.; Teitelbaum, J.; West, M. Predicting Outcome in Subarachnoid Hemorrhage (SAH) Utilizing the Full Outline of UnResponsiveness (FOUR) Score. Neurocrit. Care 2017, 27, 381–391. [Google Scholar] [CrossRef] [PubMed]
  6. Merkel, H.; Lindner, D.; Gaber, K.; Ziganshyna, S.; Jentzsch, J.; Mucha, S.; Gerhards, T.; Sari, S.; Stock, A.; Vothel, F.; et al. Standardized Classification of Cerebral Vasospasm after Subarachnoid Hemorrhage by Digital Subtraction Angiography. JCM 2022, 11, 2011. [Google Scholar] [CrossRef]
  7. Nassar, H.G.E.; Ghali, A.A.; Bahnasy, W.S.; Elawady, M.M. Vasospasm Following Aneurysmal Subarachnoid Hemorrhage: Prediction, Detection, and Intervention. Egypt. J. Neurol. Psychiatry Neurosurg. 2019, 55, 3. [Google Scholar] [CrossRef]
  8. Takano, K.; Hida, K.; Iwaasa, M.; Inoue, T.; Yoshimitsu, K. Three-dimensional Spin-echo-based Black-blood MRA in the Detection of Vasospasm Following Subarachnoid Hemorrhage. Magn. Reson. Imaging 2019, 49, 800–807. [Google Scholar] [CrossRef]
  9. Kanazawa, T.; Takahashi, S.; Minami, Y.; Jinzaki, M.; Toda, M.; Yoshida, K. Early Prediction of Clinical Outcomes in Patients with Aneurysmal Subarachnoid Hemorrhage Using Computed Tomography Texture Analysis. J. Clin. Neurosci. 2020, 71, 144–149. [Google Scholar] [CrossRef]
  10. Ramos, L.A.; Van Der Steen, W.E.; Sales Barros, R.; Majoie, C.B.L.M.; Van Den Berg, R.; Verbaan, D.; Vandertop, W.P.; Zijlstra, I.J.A.J.; Zwinderman, A.H.; Strijkers, G.J.; et al. Machine Learning Improves Prediction of Delayed Cerebral Ischemia in Patients with Subarachnoid Hemorrhage. J. NeuroIntervent. Surg. 2019, 11, 497–502. [Google Scholar] [CrossRef]
  11. Odenstedt Hergès, H.; Vithal, R.; El-Merhi, A.; Naredi, S.; Staron, M.; Block, L. Machine Learning Analysis of Heart Rate Variability to Detect Delayed Cerebral Ischemia in Subarachnoid Hemorrhage. Acta Neurol. Scand. 2022, 145, 151–159. [Google Scholar] [CrossRef] [PubMed]
  12. Hu, P.; Li, Y.; Liu, Y.; Guo, G.; Gao, X.; Su, Z.; Wang, L.; Deng, G.; Yang, S.; Qi, Y.; et al. Comparison of Conventional Logistic Regression and Machine Learning Methods for Predicting Delayed Cerebral Ischemia After Aneurysmal Subarachnoid Hemorrhage: A Multicentric Observational Cohort Study. Front. Aging Neurosci. 2022, 14, 857521. [Google Scholar] [CrossRef] [PubMed]
  13. Kim, K.H.; Koo, H.-W.; Lee, B.-J.; Sohn, M.-J. Analysis of Risk Factors Correlated with Angiographic Vasospasm in Patients with Aneurysmal Subarachnoid Hemorrhage Using Explainable Predictive Modeling. J. Clin. Neurosci. 2021, 91, 334–342. [Google Scholar] [CrossRef]
  14. Mansour, R.F.; Aljehane, N.O. An Optimal Segmentation with Deep Learning Based Inception Network Model for Intracranial Hemorrhage Diagnosis. Neural. Comput. Applic. 2021, 33, 13831–13843. [Google Scholar] [CrossRef]
  15. Hssayeni, M.D.; Croock, M.S.; Salman, A.D.; Al-khafaji, H.F.; Yahya, Z.A.; Ghoraani, B. Intracranial Hemorrhage Segmentation Using a Deep Convolutional Model. Data 2020, 5, 14. [Google Scholar] [CrossRef]
  16. Nijiati, M.; Tuersun, A.; Zhang, Y.; Yuan, Q.; Gong, P.; Abulizi, A.; Tuoheti, A.; Abulaiti, A.; Zou, X. A Symmetric Prior Knowledge Based Deep Learning Model for Intracerebral Hemorrhage Lesion Segmentation. Front. Physiol. 2022, 13, 977427. [Google Scholar] [CrossRef] [PubMed]
  17. Xu, J.; Zhang, R.; Zhou, Z.; Wu, C.; Gong, Q.; Zhang, H.; Wu, S.; Wu, G.; Deng, Y.; Xia, C.; et al. Deep Network for the Automatic Segmentation and Quantification of Intracranial Hemorrhage on CT. Front. Neurosci. 2021, 14, 541817. [Google Scholar] [CrossRef]
  18. Wang, R.; Zhang, J.; Shan, B.; He, M.; Xu, J. XGBoost Machine Learning Algorithm for Prediction of Outcome in Aneurysmal Subarachnoid Hemorrhage. Neuropsychiatr. Dis. Treat 2022, 18, 659–667. [Google Scholar] [CrossRef]
  19. Mohammed, B.A.; Senan, E.M.; Al-Mekhlafi, Z.G.; Rassem, T.H.; Makbol, N.M.; Alanazi, A.A.; Almurayziq, T.S.; Ghaleb, F.A.; Sallam, A.A. Multi-Method Diagnosis of CT Images for Rapid Detection of Intracranial Hemorrhages Based on Deep and Hybrid Learning. Electronics 2022, 11, 2460. [Google Scholar] [CrossRef]
  20. Rau, C.-S.; Wu, S.-C.; Chien, P.-C.; Kuo, P.-J.; Chen, Y.-C.; Hsieh, H.-Y.; Hsieh, C.-H. Prediction of Mortality in Patients with Isolated Traumatic Subarachnoid Hemorrhage Using a Decision Tree Classifier: A Retrospective Analysis Based on a Trauma Registry System. Int. J. Environ. Res. Public Health 2017, 14, 1420. [Google Scholar] [CrossRef]
  21. Malik, P.; Vidyarthi, A. Stacked deep model-based classification of the multiclass brain hemorrhages in CT scans. Int. J. Imaging Syst. Technol. 2024, 34, e22955. [Google Scholar] [CrossRef]
  22. Lee, H.; Yune, S.; Mansouri, M.; Kim, M.; Tajmir, S.H.; Guerrier, C.E.; Ebert, S.A.; Pomerantz, S.R.; Romero, J.M.; Kamalian, S.; et al. An Explainable Deep-Learning Algorithm for the Detection of Acute Intracranial Haemorrhage from Small Datasets. Nat. Biomed. Eng. 2018, 3, 173–182. [Google Scholar] [CrossRef] [PubMed]
  23. Imran, R.; Hassan, N.; Tariq, R.; Amjad, L.; Wali, A. Intracranial brain Haemorrhage segmentation and classification. IKSP J. Comput. Sci. Eng. 2021, 1, 52–56. [Google Scholar]
  24. Liu, P.; Ning, G.; Shi, L.; Zhou, Q.; Chen, X. Fine-grained Classification of Intracranial Haemorrhage Subtypes in Head CT Scans. IET Comput. Vision 2023, 17, 170–188. [Google Scholar] [CrossRef]
  25. Karkkainen, K.; Fazeli, S.; Sarrafzadeh, M. Unsupervised Acute Intracranial Hemorrhage Segmentation With Mixture Models. In Proceedings of the 2021 IEEE 9th International Conference on Healthcare Informatics (ICHI), Victoria, BC, Canada, 9–12 August 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 120–129. [Google Scholar]
  26. Sage, A.; Badura, P. Intracranial Hemorrhage Detection in Head CT Using Double-Branch Convolutional Neural Network, Support Vector Machine, and Random Forest. Appl. Sci. 2020, 10, 7577. [Google Scholar] [CrossRef]
  27. Barros, R.S.; Van Der Steen, W.E.; Boers, A.M.M.; Zijlstra, I.; Van Den Berg, R.; El Youssoufi, W.; Urwald, A.; Verbaan, D.; Vandertop, P.; Majoie, C.; et al. Automated Segmentation of Subarachnoid Hemorrhages with Convolutional Neural Networks. Inform. Med. Unlocked 2020, 19, 100321. [Google Scholar] [CrossRef]
  28. Li, Y.-H.; Zhang, L.; Hu, Q.-M.; Li, H.-W.; Jia, F.-C.; Wu, J.-H. Automatic Subarachnoid Space Segmentation and Hemorrhage Detection in Clinical Head CT Scans. Int. J. CARS 2012, 7, 507–516. [Google Scholar] [CrossRef]
  29. Shahzad, R.; Pennig, L.; Goertz, L.; Thiele, F.; Kabbasch, C.; Schlamann, M.; Krischek, B.; Maintz, D.; Perkuhn, M.; Borggrefe, J. Fully Automated Detection and Segmentation of Intracranial Aneurysms in Subarachnoid Hemorrhage on CTA Using Deep Learning. Sci. Rep. 2020, 10, 21799. [Google Scholar] [CrossRef]
  30. Sun, H.; Jin, L.; Xie, W. Deformable Attention U-Shaped Network with Progressively Supervised Learning for Subarachnoid Hemorrhage Image Segmentation. In Proceedings of the 2022 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Las Vegas, NV, USA, 6 December 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1734–1739. [Google Scholar]
  31. Nagaraju, S.; Prince Mary, S.; Gayatri, N. Optimal Ensemble Transfer Learning Espoused Detection and Classification of Intracranial Haemorrhage. IETE J. Res. 2024, 1–10. [Google Scholar] [CrossRef]
  32. Korra, S.; Mamidi, R.; Soora, N.R.; Kumar, K.V.; Kumar, N.C.S. Intracranial hemorrhage subtype classification using learned fully connected separable convolutional network. Concurr. Comput. Pract. Exp. 2022, 34, e7218. [Google Scholar] [CrossRef]
  33. SS, V.C. Multimodal deep learning approach for identifying and categorizing intracranial hemorrhage. Multimed. Tools Appl. 2023, 82, 42945–42960. [Google Scholar] [CrossRef]
  34. CT Dataset. Available online: https://www.kaggle.com/abdulkader90/brain-ct-hemorrhage-dataset (accessed on 15 January 2024).
  35. Zeebaree, D.Q.; Haron, H.; Abdulazeez, A.M.; Zebari, D.A. Machine Learning and Region Growing for Breast Cancer Segmentation. In Proceedings of the 2019 International Conference on Advanced Science and Engineering (ICOASE), Zakho-Duhok, Iraq, 2–4 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 88–93. [Google Scholar]
  36. Biratu, E.S.; Schwenker, F.; Debelee, T.G.; Kebede, S.R.; Negera, W.G.; Molla, H.T. Enhanced Region Growing for Brain Tumor MR Image Segmentation. J. Imaging 2021, 7, 22. [Google Scholar] [CrossRef] [PubMed]
  37. Sheela, C.J.J.; Suganthi, G. Morphological Edge Detection and Brain Tumor Segmentation in Magnetic Resonance (MR) Images Based on Region Growing and Performance Evaluation of Modified Fuzzy C-Means (FCM) Algorithm. Multimed. Tools Appl. 2020, 79, 17483–17496. [Google Scholar] [CrossRef]
  38. Mukti, I.Z.; Biswas, D. Transfer Learning Based Plant Diseases Detection Using ResNet50. In Proceedings of the 2019 4th International Conference on Electrical Information and Communication Technology (EICT), Khulna, Bangladesh, 20–22 December 2019; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
  39. Srivastava, S.; Kumar, P.; Chaudhry, V.; Singh, A. Detection of Ovarian Cyst in Ultrasound Images Using Fine-Tuned VGG-16 Deep Learning Network. SN Comput. Sci. 2020, 1, 81. [Google Scholar] [CrossRef]
  40. Yuesheng, F.; Jian, S.; Fuxiang, X.; Yang, B.; Xiang, Z.; Peng, G.; Zhengtao, W.; Shengqiao, X. Circular Fruit and Vegetable Classification Based on Optimized GoogLeNet. IEEE Access 2021, 9, 113599–113611. [Google Scholar] [CrossRef]
  41. Seyyedabbasi, A.; Kiani, F. Sand Cat Swarm Optimization: A Nature-Inspired Algorithm to Solve Global Optimization Problems. Eng. Comput. 2023, 39, 2627–2651. [Google Scholar] [CrossRef]
  42. Kiani, F.; Anka, F.A.; Erenel, F. PSCSO: Enhanced Sand Cat Swarm Optimization Inspired by the Political System to Solve Complex Problems. Adv. Eng. Softw. 2023, 178, 103423. [Google Scholar] [CrossRef]
  43. Li, Y.; Wang, G. Sand Cat Swarm Optimization Based on Stochastic Variation With Elite Collaboration. IEEE Access 2022, 10, 89989–90003. [Google Scholar] [CrossRef]
  44. Yadav, S.S.; Jadhav, S.M. Deep Convolutional Neural Network Based Medical Image Classification for Disease Diagnosis. J. Big Data 2019, 6, 113. [Google Scholar] [CrossRef]
  45. Lei, X.; Pan, H.; Huang, X. A Dilated CNN Model for Image Classification. IEEE Access 2019, 7, 124087–124095. [Google Scholar] [CrossRef]
  46. Tripathi, M. Analysis of Convolutional Neural Network Based Image Classification Techniques. J. Innov. Image Process. 2021, 3, 100–117. [Google Scholar] [CrossRef]
Figure 1. Automated SAH classification using DWSCSO and PRSCNN.
Figure 1. Automated SAH classification using DWSCSO and PRSCNN.
Diagnostics 14 02417 g001
Figure 2. Sample-acquired NCCT images.
Figure 2. Sample-acquired NCCT images.
Diagnostics 14 02417 g002
Figure 3. Sample-segmented images.
Figure 3. Sample-segmented images.
Diagnostics 14 02417 g003
Figure 4. Flowchart of the DWSCSO-based feature optimization.
Figure 4. Flowchart of the DWSCSO-based feature optimization.
Diagnostics 14 02417 g004
Figure 5. Analysis of optimization with different size of population for collected dataset.
Figure 5. Analysis of optimization with different size of population for collected dataset.
Diagnostics 14 02417 g005
Figure 6. Analysis of convergence for collected dataset.
Figure 6. Analysis of convergence for collected dataset.
Diagnostics 14 02417 g006
Figure 7. ROC curve for collected dataset, (a) decision tree, (b) GCN, (c) ANN, (d) Autoencoder, (e) CNN, (f) PRSCNN.
Figure 7. ROC curve for collected dataset, (a) decision tree, (b) GCN, (c) ANN, (d) Autoencoder, (e) CNN, (f) PRSCNN.
Diagnostics 14 02417 g007
Figure 8. Confusion matrix for collected dataset, (a) decision tree, (b) GCN, (c) ANN, (d) Autoencoder, (e) CNN, (f) PRSCNN.
Figure 8. Confusion matrix for collected dataset, (a) decision tree, (b) GCN, (c) ANN, (d) Autoencoder, (e) CNN, (f) PRSCNN.
Diagnostics 14 02417 g008
Figure 9. Accuracy graph for collected dataset, (a) decision tree, (b) GCN, (c) ANN, (d) Autoencoder, (e) CNN, (f) PRSCNN.
Figure 9. Accuracy graph for collected dataset, (a) decision tree, (b) GCN, (c) ANN, (d) Autoencoder, (e) CNN, (f) PRSCNN.
Diagnostics 14 02417 g009
Figure 10. Loss graph for collected dataset, (a) ReLU, (b) Leaky ReLU, (c) ELU, (d) PReLU.
Figure 10. Loss graph for collected dataset, (a) ReLU, (b) Leaky ReLU, (c) ELU, (d) PReLU.
Diagnostics 14 02417 g010aDiagnostics 14 02417 g010b
Table 1. Simulation results of different segmentation methods for collected dataset.
Table 1. Simulation results of different segmentation methods for collected dataset.
Segmentation MethodsJIDSCPAMPA
FCM clustering0.720.780.740.69
Otsu thresholding0.730.790.780.73
K-means clustering0.810.880.840.77
Superpixel clustering0.920.930.890.84
Region-growing0.890.920.920.87
Modified region-growing0.940.950.930.90
Table 2. Simulation results of different segmentation methods for CT dataset.
Table 2. Simulation results of different segmentation methods for CT dataset.
Segmentation MethodsJIDSCPAMPA
FCM clustering0.750.800.770.72
Otsu thresholding0.790.820.830.77
K-means clustering0.830.920.880.82
Superpixel clustering0.940.950.920.88
Region-growing0.910.940.950.92
Modified region-growing0.960.960.950.93
Table 3. Simulation results of different feature extraction methods for collected dataset.
Table 3. Simulation results of different feature extraction methods for collected dataset.
Feature Extraction MethodsAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
GoogleNet95.2794.1294.6793.6496.120.952
VGG-1696.3595.1895.8494.2197.060.961
VGG-1996.8195.6796.3394.7697.530.963
SqueezeNet94.0092.8593.4091.7195.190.940
ResNet5097.4296.2596.8196.3198.220.975
GoogleNet + VGG-16 + ResNet5099.4899.5399.4898.5699.470.996
Table 4. Simulation results of different feature extraction methods for CT dataset.
Table 4. Simulation results of different feature extraction methods for CT dataset.
Feature Extraction MethodsAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
GoogleNet95.6795.6994.9193.5096.850.957
VGG-1696.9296.9696.0294.7597.100.963
VGG-1997.3097.3597.1595.0097.800.965
SqueezeNet95.4294.1894.5092.7096.300.946
ResNet5098.4497.2697.7196.7598.800.980
GoogleNet + VGG-16 + ResNet5099.6299.7399.5198.8099.700.997
Table 5. Simulation results of different optimization algorithms for collected dataset.
Table 5. Simulation results of different optimization algorithms for collected dataset.
Optimization AlgorithmsAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
GA93.4495.3395.4393.5895.450.933
BOA95.5896.4496.4895.0396.630.950
ABC96.9297.2497.9296.0897.150.959
WOA97.4697.4597.9496.7597.990.977
SCSO98.6898.1098.6397.8298.150.981
DWSCSO99.4899.5399.4898.5699.470.996
Table 6. Simulation results of different optimization algorithms for CT dataset.
Table 6. Simulation results of different optimization algorithms for CT dataset.
Optimization AlgorithmsAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
GA95.4096.0195.3595.1295.070.939
BOA96.0996.79968495.2895.100.944
ABC96.7395.9896.3196.0596.110.949
WOA97.0397.8096.8897.0396.870.981
SCSO98.7698.2798.8598.1298.350.986
DWSCSO99.6299.7399.5198.8099.700.997
Table 7. Simulation results of computational time for collected dataset.
Table 7. Simulation results of computational time for collected dataset.
Optimization AlgorithmsComputational Time (S)
GA110.30
BOA101.08
ABC81.66
WOA83.90
SCSO78.02
DWSCSO70.12
Table 8. Friedman test for collected dataset.
Table 8. Friedman test for collected dataset.
Optimization AlgorithmsFriedman Rank
GA2.43
BOA3.98
ABC4.37
WOA5.78
SCSO7.04
DWSCSO7.63
p-value 2.0149 × 10 21
Table 9. Simulation results of different classification models for collected dataset.
Table 9. Simulation results of different classification models for collected dataset.
Actual Feature Vectors
ClassifiersAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
Decision tree88.2487.9488.9887.6788.450.873
GCN89.2690.9090.4089.3290.200.907
ANN92.3693.5092.3492.8993.100.915
Autoencoder94.2793.8893.5493.4892.120.939
CNN94.8194.0193.9994.7794.890.944
PRSCNN95.2295.9094.9094.0995.400.955
Optimized Feature Vectors
ClassifiersAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
Decision tree90.3093.2892.2893.4593.810.902
GCN93.9095.0293.8294.9194.800.930
ANN95.3897.3898.3396.8496.990.966
Autoencoder97.3098.7698.7097.6097.730.979
CNN98.0999.0099.0198.2298.840.982
PRSCNN99.4899.5399.4898.5699.470.996
Table 10. Simulation results of different classification models for CT dataset.
Table 10. Simulation results of different classification models for CT dataset.
Actual Feature Vectors
ClassifiersAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
Decision tree89.2289.1589.0489.3289.140.881
GCN90.9890.7091.6090.6491.220.918
ANN93.6893.8692.1293.7593.150.935
Autoencoder94.6694.5395.0695.1494.570.947
CNN95.7794.4795.1595.8395.500.954
PRSCNN96.5496.1695.7296.4396.820.962
Optimized Feature Vectors
ClassifiersAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
Decision tree91.2791.6291.8791.2691.200.929
GCN92.8193.2292.8593.4092.910.940
ANN94.7895.0295.4795.7194.820.956
Autoencoder97.7297.1797.9196.9397.170.978
CNN98.4298.6598.3497.7097.740.983
PRSCNN99.6299.7399.5198.8099.700.997
Table 11. Simulation results of different activation functions for collected dataset.
Table 11. Simulation results of different activation functions for collected dataset.
Activation FunctionsAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
ReLU92.7491.4792.8892.6092.030.937
Leaky ReLU95.8494.2594.2994.7294.230.941
ELU96.9497.7396.2297.1696.830.970
PReLU99.4899.5399.4898.5699.470.996
Table 12. Simulation results of different activation functions for CT dataset.
Table 12. Simulation results of different activation functions for CT dataset.
Activation FunctionsAccuracy (%)MCC (%)F1 Score (%)Sensitivity (%)Specificity (%)AUC
ReLU93.4593.5493.5693.5193.910.942
Leaky ReLU96.9595.9096.0496.4896.900.964
ELU97.3698.8396.7398.2098.020.987
PReLU99.6299.7399.5198.8099.700.997
Table 13. Different K-fold cross validation results of DWSCSO-PRSCNN for collected dataset.
Table 13. Different K-fold cross validation results of DWSCSO-PRSCNN for collected dataset.
MeasuresK = 2K = 4K = 5K = 8
MCC (%)94.9495.8299.5397.21
F1-score (%)93.6394.0499.4893.69
Accuracy (%)97.2498.1999.4896.17
Sensitivity (%)96.3497.3398.5694.67
Specificity (%)95.1197.6799.4794.62
AUC0.9590.9610.9960.950
Table 14. Different K-fold cross validation results of DWSCSO-PRSCNN for CT dataset.
Table 14. Different K-fold cross validation results of DWSCSO-PRSCNN for CT dataset.
MeasuresK = 2K = 4K = 5K = 8
MCC (%)95.7897.6299.7395.81
F1-score (%)96.1098.3999.5195.29
Accuracy (%)96.6698.5899.6296.28
Sensitivity (%)95.7497.0598.8094.25
Specificity (%)95.5396.3399.7094.72
AUC0.9650.9710.9970.945
Table 15. Simulation results of classifiers for noisy images.
Table 15. Simulation results of classifiers for noisy images.
ClassifiersAccuracy (%)
Noiseless ImagesNoisy Images
Decision tree91.2786.53
GCN92.8186.74
ANN94.7888.21
Autoencoder97.7292.56
CNN98.4293.91
PRSCNN99.6297.32
Table 16. Simulation results of classifiers for missing data.
Table 16. Simulation results of classifiers for missing data.
ClassifiersSensitivity (%)
Without Missing DataWith 10% Missing Data
Decision tree91.2684.56
GCN93.4087.35
ANN95.7191.42
Autoencoder96.9392.58
CNN97.7093.51
PRSCNN98.8095.40
Table 17. Comparison of DWSCSO-PRSCNN.
Table 17. Comparison of DWSCSO-PRSCNN.
Activation FunctionsAccuracy (%)
DL-ICH [14]95.06
GoogLeNet + (GLCM and LBP) [19]98.9
ResNet-50 + (GLCM and LBP) [19]99.1
AlexNet + (GLCM and LBP) [19]99.3
DWSCSO-PRSCNN99.62
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sengupta, J.; Alzbutas, R.; Iešmantas, T.; Petkus, V.; Barkauskienė, A.; Ratkūnas, V.; Lukoševičius, S.; Preikšaitis, A.; Lapinskienė, I.; Šerpytis, M.; et al. Detection of Subarachnoid Hemorrhage Using CNN with Dynamic Factor and Wandering Strategy-Based Feature Selection. Diagnostics 2024, 14, 2417. https://doi.org/10.3390/diagnostics14212417

AMA Style

Sengupta J, Alzbutas R, Iešmantas T, Petkus V, Barkauskienė A, Ratkūnas V, Lukoševičius S, Preikšaitis A, Lapinskienė I, Šerpytis M, et al. Detection of Subarachnoid Hemorrhage Using CNN with Dynamic Factor and Wandering Strategy-Based Feature Selection. Diagnostics. 2024; 14(21):2417. https://doi.org/10.3390/diagnostics14212417

Chicago/Turabian Style

Sengupta, Jewel, Robertas Alzbutas, Tomas Iešmantas, Vytautas Petkus, Alina Barkauskienė, Vytenis Ratkūnas, Saulius Lukoševičius, Aidanas Preikšaitis, Indre Lapinskienė, Mindaugas Šerpytis, and et al. 2024. "Detection of Subarachnoid Hemorrhage Using CNN with Dynamic Factor and Wandering Strategy-Based Feature Selection" Diagnostics 14, no. 21: 2417. https://doi.org/10.3390/diagnostics14212417

APA Style

Sengupta, J., Alzbutas, R., Iešmantas, T., Petkus, V., Barkauskienė, A., Ratkūnas, V., Lukoševičius, S., Preikšaitis, A., Lapinskienė, I., Šerpytis, M., Misiulis, E., Skarbalius, G., Navakas, R., & Džiugys, A. (2024). Detection of Subarachnoid Hemorrhage Using CNN with Dynamic Factor and Wandering Strategy-Based Feature Selection. Diagnostics, 14(21), 2417. https://doi.org/10.3390/diagnostics14212417

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop