Next Article in Journal
Promoting Health Education through Mobile Apps: A Quantitative Analysis of American Hospitals
Previous Article in Journal
Performance Evaluation of a Low-Cost Non-Invasive Ventilator during the COVID-19 Pandemic: A Bench Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

White Blood Cell Classification Using Texture and RGB Features of Oversampled Microscopic Images

by
Furqan Rustam
1,2,
Naila Aslam
3,
Isabel De La Torre Díez
4,*,
Yaser Daanial Khan
3,
Juan Luis Vidal Mazón
5,6,7,
Carmen Lili Rodríguez
5,8 and
Imran Ashraf
9,*
1
School of Computer Science, University College Dublin, D04 V1W8 Dublin, Ireland
2
Department of Software Engineering, University of Management and Technology, Lahore 544700, Pakistan
3
Department of Computer Science, School of Systems and Technology, University of Management and Technology, Lahore 54770, Pakistan
4
Department of Signal Theory and Communications and Telematic Engineering, University of Valladolid, Paseo de Belén 15, 47011 Valladolid, Spain
5
Higher Polytechnic School, Universidad Europea del Atlántico, Isabel Torres 21, 39011 Santander, Spain
6
Department of Projects, Universidad Internacional Iberoamericana (UNIB), Arecibo, PR 00613, USA
7
Department of Project, Universidade Internacional do Cuanza (UNIC), Barrio Kaluanda, Cuito EN 250, Angola
8
Department of Project Management, Universidad Internacional Iberoamericana, Campeche 24560, Mexico
9
Department of Information and Communication Engineering, Yeungnam University, Gyeongsan 38541, Korea
*
Authors to whom correspondence should be addressed.
Healthcare 2022, 10(11), 2230; https://doi.org/10.3390/healthcare10112230
Submission received: 7 September 2022 / Revised: 29 October 2022 / Accepted: 2 November 2022 / Published: 8 November 2022

Abstract

:
White blood cell (WBC) type classification is a task of significant importance for diagnosis using microscopic images of WBC, which develop immunity to fight against infections and foreign substances. WBCs consist of different types, and abnormalities in a type of WBC may potentially represent a disease such as leukemia. Existing studies are limited by low accuracy and overrated performance, often caused by model overfit due to an imbalanced dataset. Additionally, many studies consider a lower number of WBC types, and the accuracy is exaggerated. This study presents a hybrid feature set of selective features and synthetic minority oversampling technique-based resampling to mitigate the influence of the above-mentioned problems. Furthermore, machine learning models are adopted for being less computationally complex, requiring less data for training, and providing robust results. Experiments are performed using both machine- and deep learning models for performance comparison using the original dataset, augmented dataset, and oversampled dataset to analyze the performances of the models. The results suggest that a hybrid feature set of both texture and RGB features from microscopic images, selected using Chi2, produces a high accuracy of 0.97 with random forest. Performance appraisal using k-fold cross-validation and comparison with existing state-of-the-art studies shows that the proposed approach outperforms existing studies regarding the obtained accuracy and computational complexity.

1. Introduction

White blood cells (WBC) are potent infection fighters; they normally grow and divide in an orderly way regarding the needs of the human body to fight infections and other diseases [1]. However, in people with leukemia, the bone marrow produces an excessive amount of abnormal WBCs that do not function properly, and this leads to several abnormalities [2,3]. The counts of WBCs in blood cells can provide early clues of different probable abnormalities concerning the number of cells of different types of WBC. The WBCs can be categorized into five major classes: lymphocytes, eosinophils, neutrophils, basophils, and monocytes. The percentage of each type of WBC in a healthy person varies within a range. For example, lymphocytes are 20% to 40%, eosinophils account for 1% to 6%, and monocytes make 2% to 10%, while neutrophils are 40% to 80% [4]. A lower count of WBC can cause blood cancer and many other diseases, as WBC contributes as an important part of the body’s immune system. The number of WBCs varies when the bone marrow stops making them, or when WBC are destroyed by another entity [4]. Healthy WBCs play an important role in preventing different infections and helping to fight other deadly diseases such as COVID-19 [5].
Estimating the WBC can help with making an early prediction of probable diseases. Blood microscopic images can be used to detect the WBC type and the timely diagnosis of the disease. Traditional approaches for WBC-type detection are time-consuming and low in accuracy, which increases the importance of accurate systems for the fast and accurate analysis of WBC. In this regard, the machine learning approach plays an important role. Several machine learning approaches are introduced recently, which can predict the type of WBC using microscopic images. These include the study [6] using convolutional neural networks (CNN) and an extreme learning machine (ELM) for WBC type detection using microscopic images. CNN with traditional deep learning approaches and handcrafted features for WBC is deployed by the study [7].
This study also follows a machine learning-based approach for WBC type detection using blood microscopic images. We used a WBC image dataset acquired from the IEEE data port, which has imbalanced class distribution and a poor feature set. We proposed a novel approach by combining a data re-sampling technique and a hybrid feature engineering technique to alleviate the influence of such shortcomings and to make the following contributions.
  • An improved feature set is obtained by combining the texture features and RGB features to make a more correlated feature set with target classes to obtain a high accuracy. Later, Chi-squared (Chi2) is used to select an important and equal number of features for the models’ training. The models are evaluated using texture features and RGB features, in comparison to the proposed hybrid features.
  • The imbalanced distribution of different classes of WBC is tackled using data resampling, which helps to reduce the model over-fitting. The synthetic minority oversampling technique (SMOTE) is applied for data resampling in this study for evaluating the influence of data resampling on the performance of machine learning models.
  • Besides using various machine learning models such as decision tree (DT), random forest (RF), k-nearest neighbor (KNN), and support vector machine (SVM), state-of-the-art pre-trained deep learning models are also employed, including ResNet50 and VVG16. In addition, a custom-designed CNN is also used. Experiments are performed using the original dataset, augmented dataset, and oversampled dataset.
  • The performances of all models is analyzed regarding different performance evaluation metrics. Furthermore, k-fold cross-validation and statistical t-tests are carried out. Additionally, the performance comparison with recent state-of-the-art approaches is made to analyze the performance of the proposed approach regarding accuracy and response time.
This paper is further divided into three sections. Section 2 discusses several important studies in the context of WBC-type detection. The proposed feature engineering approach and the processes of WBC-type classification are explained in Section 3. Section 4 presents the results and discussion, while the conclusions are provided in Section 5.

2. Related Work

Research on WBC is one of the most important domains in bioinformatics, while the use of machine learning models for WBC has also been regarded as potentially significant [8,9,10]. The classification of WBC using microscopic images has been investigated by several researchers [6,7,11]. Despite that, several challenges remain unresolved, such as the pure WBC image dataset, the high accuracy machine learning approach, the efficiency and reduction in computational time, etc. In the following, several prominent research works are discussed.
The authors propose a deep learning-based automatic approach for WBC classification in [9]. The study utilizes pre- and post-preprocessing to improve the performance of the CNN model. In preprocessing, data normalization, filled holes, and data augmentation are applied, while in post-preprocessing, the study deployed localization and data segmentation techniques. The study achieved a 95.73% accuracy score using the CNN model on the cytological images dataset. The study [12] used the CNN model and recurrent neural network (RNN), as well as, a combination of both CNN and RNN for WBC classification to resolve the multiple cells overlap problem. The Canonical Correlation Analysis method is used in the study with the BCCD dataset. The achieved accuracy score is 95%. The study [13] proposed a 3D convolutional network, called deep hyper, for WBC microscopic image classification. Spectral and spatial features are used with a deep hyper model to obtain 96% accuracy. Using the deep hyper model with an attention module resulted in a 97% accuracy score.
The study [14] used a hybrid approach for WBC detection, in which the scale-invariant feature transform (SIFT) and CNN model are combined. SIFT is used for feature detection, which is used for CNN training using the LISC and WBCis datasets. The proposed model achieved 95.84% and 97.33% accuracy scores, respectively, for both datasets. Similarly, the study [15] proposed an approach for the classification of WBC using the CNN model. They deployed the proposed approach on the Kaggle WBC images dataset and achieved significant accuracy. The study [16] proposed a multi-level CNN model for the WBC classification for four types of cell classification. At the first level, Faster R-CNN is applied for the detection of the region of interest while at the second level, CNN-based architecture MobileNet is used for cell-type classification.
Besides proposing novel architectures, several studies adopt pre-trained models for WBC classification. The study [6] used a supervised machine learning approach for WBC detection which follows a CNN architecture and ELM model on a microscopic image dataset. Several CNN-based pre-trained models are used, such as AlexNet, GoogleNet, VGG-16, and ResNet for feature extraction while ELM is trained on those features to obtain a 96.03% accuracy. Similarly, [7] used deep learning CNN with handcrafted features to achieve a higher accuracy. The study worked on six types of WBC, including lymphocytes, monocytes, basophils, neutrophils, eosinophils, and abnormal cells. The experimental results are promising. Along the same directions, [17] used a pre-trained CNN architecture for WBC classification. The authors deployed ResNet and Inception variants with fine-tuned parameters to obtain 100% training accuracy with the ResNet50 model for four classes of WBC; however, the computational complexity of this approach is very high, as it uses 3000 epochs for a very deep ResNet50 model. The study achieved an accuracy score of 98.4%. The study [18] used the DenseNet121 model for the classification of the WBC. Data normalization and data augmentation are used with the optimized DenseNet121 model. The proposed model achieved 98.84% accuracy on the KBC dataset.
Table 1 presents a comparative analysis of the cited research works. Predominantly, the above-discussed studies on WBC classification adopt deep learning approaches where the computational costs are higher than for machine learning models. Additionally, several studies use a lower number of WBC classes, and the reported accuracy is high. In addition, a few studies experimented with imbalanced datasets where the models’ over-fitting probability is high. This study resolves the problem of high computational cost by deploying machine learning models, and aims at achieving high accuracy using feature engineering. Using a balanced dataset, the probability of model over-fitting is also reduced.

3. Materials and Methods

This study works on WBC type detection using the image dataset and machine learning approach. In the following sections, the background on WBC, details of the dataset, and a description of feature engineering and models are provided.

3.1. White Blood Cells

WBC are colorless cells of the human blood because they do not have any pigment. They comprise 7000 to 8000 cells in one cubic mm of blood [4]. Their size is much larger than the red blood cells. They can be classified into five different types, based on the shape of the nucleus and the density of the granules in the cytoplasm. WBCs have a single large bilobed nucleus, which distinguishes them from the other blood cells. They are formed in the bone marrow cell and then move to the blood and lymph [4]. Types of WBC and their brief description are provided here.
  • Neutrophils make up about 62% of the total WBC. They can engulf any foreign particles such as viruses and bacteria, and destroy them to neutralize their effects. They are about twice the size of a red blood cell. Their nucleus contains about 2 to 5 lobes [21]. Figure 1 presents the types of WBC.
  • Basophils comprise approximately less than 1% of WBCs. They are about twice the size of a red blood cell, with a bilobed nucleus. These cells release the protein called heparin, which prevents blood clotting. Basophils also release another protein called histamine, which causes inflammation. They release antibodies and antibiotics, which protect the body from the effects of any foreign objects. Figure 1b shows the types of basophils of WBCs Figure 1b.
  • Monocytes forms approximately 3% of the total WBC. They are about two to three times larger than red blood cells. The nucleus is an almost round to lobed shape. It produces macrophages, which can destroy the larger particles via phagocytosis. The life span of macrophages in the blood is about 8 to 10 h, after which they move toward the lymphoid tissue and become a macrophage. The monocyte type of WBC is shown in Figure 1c.
  • Eosinophils comprise about 2% of the WBC. They are about twice the size of a red blood cell, while their nucleus is bilobed. They inactivate the inflammation-producing substance, and attack parasites and worms. Figure 1d shows the sample image of the basophil types of WBC.
  • Lymphocytes make up approximately 32% of the total WBCs (Figure 1b). Their size is nearly equal to the red blood cell. They produce antibodies. The life span of lymphocytes in the blood may be months and even years, depending upon the activity of the cell. Lymphocyte samples are shown in Figure 1e.
Figure 1. Types of WBC. (a) Neutrophils, (b) Basophils, (c) Eosinophils, (d) Monocytes, and (e) Lymphocytes.
Figure 1. Types of WBC. (a) Neutrophils, (b) Basophils, (c) Eosinophils, (d) Monocytes, and (e) Lymphocytes.
Healthcare 10 02230 g001
Figure 2 shows the hierarchy of the WBC type according to the root types, and also illustrates the differences in various types of WBC.

3.2. Flow of Implemented Methodology

In this study, first, a dataset containing the WBC images is acquired from the IEEE data port, which is later used to extract the features from the WBC images. We extract two types of features, including texture features and RGB features. Texture features are extracted using the skimage library [22]. After extracting the features, we deploy the feature selection technique to extract the prominent and most important features using Chi2.
We select 3000 features each from the RGB and texture features. Both types of important features (RGB and texture) are combined to form a hybrid feature set. For resolving the problems related to the imbalanced dataset, we performed data oversampling using SMOTE, which helps to reduce the model over-fitting towards the majority class. Several machine learning and deep learning models are deployed to perform WBC-type classification. To train and test the learning models, we split the dataset into training and testing sets with 0.8 to 0.2 ratios for training and testing, respectively. In the end, we evaluate all models in terms of accuracy, precision, recall, and F1 score. Figure 3 shows the proposed methodology of this study.

3.3. WBC Dataset

The dataset that is used in this study comprises five categories of WBCs, including neutrophil lymphocyte, monocyte, eosinophil, and basophil. The dataset is acquired from the IEEE Dataport [23]. Each type of WBC consists of a different number of images, as shown in Table 2. The total number of images in the dataset is 3539, with 667 raw images, 1464 augmented images, and 1408 cropped and classified images.

3.4. Feature Engineering

For feature engineering, we deployed feature extraction and feature selection techniques to generate an important feature set that can enhance the performance of learning models. We extracted two types of features from the dataset, which are RGB features and texture features, because color and shape are the best parameters for classifying the WBC images.

3.4.1. RGB Features

Color is an important part of the image, and is widely used in image processing. The color information is conveyed through three channels, and these color features are invariant to the rotation of the pixels in an image. The red, blue, and green pixel values represent the content in the image, and can be used for the training of machine learning models [24]. These RGB features are very feasible, as color can potentially be used to discriminate between images of the different target classes.

3.4.2. Texture Features

Texture features can be extracted from grayscaled images, disregarding the color information [25]. Texture features provide information about the intensities of the pixels in an image. They are characterized by the distribution of intensity levels in the neighborhood.
Both RGB and texture features have their own advantages and disadvantages, similar to their significance toward improving the performances of machine learning models. For RGB features, we used the CV2 library, while the texture features are obtained using the skimage library [22]. Table 3 shows the number of features for RGB and texture features.
We combined both features vertically and generated a hybrid feature set. Before combining the features, we performed feature selection and extracted the most important features, which reduced the size of the hybrid feature set to improve the accuracy and efficiency of the models. For feature selection, we used the Chi-squared (Chi2) technique.

3.4.3. Chi2

Chi2 is a feature selection technique that finds the relationship between the feature set and the target classes [26]. It finds the dependence of the target class on each of the data features. It assumes the null hypothesis, indicating that the feature distribution is independent. Chi2 calculates the x 2 scores and then orders them in descending order. The higher the value of x 2 , the more important the feature is. The formula to calculate the x 2 is given as (1)
x 2 = i = 1 k j = 1 m ( O i j E i j ) 2 E i j
where O i j is the observed frequency, E i j is the expected frequency, k is the number of rows and m is the number of columns, while ( i , j ) denotes the cell.
We extract 3000 features each from RGB and texture features and make a hybrid feature set. Table 4 shows the number of features with each approach, and Figure 4 shows the process of forming the hybrid feature set.

3.5. Synthetic Minority Oversampling Technique

SMOTE is an oversampling technique that generates artificial samples to make a balanced dataset that can help to reduce the probability of models overfitting [27]. SMOTE generates samples for the minority class to equal the number of majority class samples. It is mostly used for the numeric data features, while this study uses it to increase the number samples for the image data. Table 5 shows the ratio of samples after the oversampling of data. Figure 5 shows the approach for data normalizing and balancing using the SMOTE approach.

3.6. Data Splitting

Data splitting is performed to train and test the learning models. We split the dataset with 0.8 to 0.2 ratios into cases of both oversampling and the original dataset. This ratio is more optimal as the dataset size is not too large, so we obtain the best results with this ratio. Table 6 shows the number of samples for training and testing with the original dataset and the over-sampled dataset.

3.7. Machine Learning Models

Machine learning models can be used to solve different domain problems such as text analysis [28], computer vision application [29,30], IoT [31,32], and image processing [33,34], etc. In this study, we used machine learning models for WBC image classification. We deployed several machine learning models on hybrid features set to show the significance of our feature engineering approach. Four state-of-the-art machine learning models such as DT, RF, KNN, and SVM have been employed with their best hyper-parameters settings according to the dataset used, as shown in Table 7.
DT is a tree-based machine learning model used for classification and regression tasks [35,36]. DT constructs a tree using the features set, and put important features on the root nodes, while leaf nodes function as decision nodes for the DT. The features’ importance in DT can be calculated using the Gini Index or Information Gain algorithms. These algorithms provide purity and impurity in the feature set. For tree split, we used the Gini criterion [37], which can be calculated as
G = i = 1 T p ( i ) ( 1 p ( i ) )
where T is the total number of target classes and p ( i ) is the probability of picking the data point with class i.
The DT is used with max_depth parameter with value 200, which is the most optimal value for it, because after a 200 value of max_depth, there is no change in model accuracy. The increase in tree depth increases the complexity of the model, so we restrict it to 200-level depth.
RF is also a tree-based model with an ensemble of several DTs. RF can be used for classification and regression tasks. It can perform well on nonlinear and imbalanced datasets [38]. RF combines a number of DTs under majority voting. All DTs in RF make their prediction on the test data, and then the most predicted class by the DTs is regarded as the final prediction by the RF. RF can be defined mathematically as
r f = m o d e { D T 1 , D T 2 , D T 3 , . . . , D T n }
or,
r f = m o d e { i = 1 n D T n }
where D T 1 , D T 2 , D T 3 are the decision trees and n is the number of decision trees.
We used RF with n_estimator parameters with 200 and max_depth parameters with 200 as we used in DT. The n_estimators indicate the number of DTs participating in the prediction procedure of RF.
KNN is a simple machine learning algorithm for classification and regression. KNN, also known as lazy leaner, uses the whole training data. It performs matching on test data with training data, and finds the distance with the nearest target in the training data [39]. Several distance metrics can be used with KNN, while the Euclidean distance is most commonly used:
d ( i , j ) = p = 1 n ( i p j p ) 2
where i and j are two points in Euclidean n-space, and i p j p shows the Euclidean vectors, starting from the origin of the space.
KNN is used with only one parameter, which is n_neighbors with a value of 5, indicating that the model uses five neighbors to predict the class of a given sample.
SVM is a linear model used for classification and regression tasks. SVM draws multiple hyper-planes to classify the data into several classes [40]. Hyper-plane with the best margin from data points will be the selected hyper-plane for the SVM. Hyper-plane can be defined mathematically as
Z . x + b = 0
where Z is a vector normal to hyper-plane and b is an offset. We used SVM with poly kernel and C = 4.0.

3.8. WBC Classification

Algorithm 1 illustrates the WBC classification approach used in this study. The given input image is classified into one of the five classes. First, RGB and texture features are extracted from all images one by one, and then concatenated vertically (axis = 1) to make a hybrid feature set. The feature set goes through the SMOTE technique to make the dataset balanced, which is later used to train machine learning models.
Algorithm 1: WBC classification algorithm
Input: WBC Images
Result: WBC type detection ⟶ (Neutrophil Lymphocyte, Monocyte, Eosinophil, and Basophil)
initialization;
loop (Img in Images)
               RGB R f (Img)
               Txt T f (Img)
                R f . a p p e n d ( R G B )
                T f . a p p e n d ( T x t )
loop end
HF c o n c a t e n a t e ( R f , T f a x i s = 1 )
PFS S M O T E ( H F )
Prediction M L m o d e l ( P S F )
Scores E v a l u a t i o n ( P r e d i c t i o n )

4. Results

This section contains the results of machine learning and deep learning models for WBC-type detection. We deployed DT, RF, KNN, and SVM from machine learning models with both texture and RGB features. We used a Core i7 11th generation machine with NVIDIA GPU to perform the experiments. The system consists of 16 RAM and 1TB SSD. We used a Jupyter notebook and Python language to implement the approach.

4.1. Results Using the Original Dataset

Table 8 contains the results of machine learning models on the original dataset with texture and RGB features. The results of all models are very poor on the original dataset because the target ratio in the original dataset is highly imbalanced. All models show the over-fitting for the majority class data. With both the RGB and texture features, the performance of RF is better as compared to other models in terms of accuracy on the original dataset, as it achieves 0.66 and 0.62 accuracy scores, respectively. The F1 scores of all models are poor because of the highly imbalanced dataset. To resolve this imbalanced dataset problem, we used data augmentation.

4.2. Results Using the Augmented Dataset

For data augmentation, ref. [23] used Keras preprocessing layers, where the images are augmented to generate a larger dataset. The data augmentation influences the learning models positively and marginally improved the accuracy. The marginal improvement is due to the uncorrelated distribution of features during augmentation. Table 9 contains the results of machine learning models on the augmented dataset with both the texture and RGB features. For the augmented dataset, machine learning models with RGB features are somehow better as compared to the texture features, as RF achieves a 0.64 accuracy score with a 0.54 F1 score. With texture features, RF achieves a 0.35 F1 score. The reason behind the accuracy improvement of the learning models with RGB features after augmentation is that the WBC color is not changed after augmentation, while the texture features are changed after flipping or rotating the image during augmentation.

4.3. Performance of Models on Hybrid Features

We try to improve the performance of learning models by making a balanced dataset using the SMOTE technique. The SMOTE technique yields a highly balanced dataset, which increases the number of samples for training and testing for machine learning models and results in significant improvements. Table 10 shows the results of machine learning models using the SMOTE balanced dataset. Using the balanced dataset with SMOTE, the performances of models has been elevated, as the RF achieves the highest accuracy score of 0.92 and the same F1 score using the texture features. SMOTE does not impact on the texture features and the RGB features, as it generates new samples artificially to reduce the models’ over-fitting. RF is also good with RGB features after oversampling with SMOTE, as it achieves a 0.87 accuracy score. DT and KNN do not perform well because KNN shows a poor performance when used with a large feature set.
A novel hybrid feature engineering approach is deployed with the machine learning models to analyze their performances against the individual texture and RGB features. The hybrid features set combines both the texture features and RGB features selected using the Chi2 technique. In addition, SMOTE is applied to perform the oversample for data balancing.
Table 11 shows the results of machine learning models using the hybrid feature engineering technique. The results suggest that this approach significantly improves the performance of all learning models, with RF being the highest accuracy-preserving model, as it obtains an accuracy score of 0.97. The performance of RF is better, owing to its ensemble architecture which combines 300 decision trees on the hybrid feature set to make the final prediction. SVM also shows significant performance with the hybrid feature, as it achieves a 0.89 accuracy score and a 0.89 F1 score. The impact of the hybrid feature set is two-fold, where it provides a highly correlated feature set in the first place and a balanced feature set in the second place, which improves the training process of the machine learning models and enhances their performances. On average, the performances of all the models have been elevated when used with the hybrid feature set.
A performance comparison of all machine learning models used in this study is illustrated in Figure 6, using different types of features such as original texture and RGB features, augmented texture and RGB features, and hybrid features. It shows that the best performance of the models is obtained when trained on oversampled hybrid features, with RF as the best performer. Figure 7 show the confusion matrix for the best performer RF using machine learning models.
Figure 8 shows the comparison of different features, including RGB features, texture features, and hybrid features. It is helpful in understanding the performance of machine learning models when trained on different feature sets. Figure 8a shows the feature space of RGB features, indicating that the distribution of samples of various classes is not well separable, which explains why the performance of machine learning models is not good when using RGB features. The same is true for texture features where samples of different WBC types are mixed, as shown in Figure 8b. However, the feature space of hybrid features on oversampled data, as given in Figure 8c, shows that the sample distribution of WBC types is more separable and distinguishable as compared to both texture and RGB features, which leads to a better performance of the machine learning models.

4.4. Performance Analysis of State-of-the-Art Deep Learning Models

We deployed several deep learning models in comparison with the proposed approach. For this purpose, CNN, VVG16, and ResNet-150 are adopted for WBC-type detection. The architectures of all models are illustrated in Figure 9. Each model consists of a dense layer in the end, with five neurons and a Sigmoid activation function. The models are compiled using the categorical_crossentropy loss function and an Adam optimizer [33]. All models are fitted with 15 epochs, and 15% of the data are used for validation.
Figure 10 shows the training and validation loss and accuracy for the CNN and pre-trained models. All models are fitted with 15 epochs, and the results for each epoch are given. The loss and accuracy per epoch are illustrated with respect to the original dataset, augmented dataset, and oversampled dataset. The figure shows that the training and validation accuracy, and loss of CNN are better, as compared to other models.
The performance results of deep learning models are shown in Table 12. The results are presented for three types of datasets, including the original, augmented, and oversampled datasets. However, the handcrafted features are not used for deep learning models such as texture or RGB features, because deep learning models work well with their own extracted features. On the original dataset, pre-trained models are good as compared to CNN, as VVG16 achieves a 0.74 accuracy score with a 0.30 F1 score. The performance of VVG16 is the same in the case of the augmented dataset. In both cases, the original and augmented dataset models are over-fitted on the majority class and show poor performance for the minority class, which results in a poor F1 score. The oversampling technique resolves this issue by providing balanced samples for training, which results in a better performance of CNN, with similar results for the accuracy and F1 score. CNN achieves an accuracy of 0.73 and an F1 score of 0.71 using the oversampled dataset.

4.5. Validation of the Proposed Approach

We validate the proposed approach using the k-fold cross-validation technique. From this point of view, a 10-fold cross-validation is applied to all of the machine learning models used in this study with the proposed hybrid features on the oversampled dataset. The results given in Table 13 show that the RF obtains the best performance, with a 0.95 accuracy score and a standard deviation (SD) of ±0.03, followed by the SVM, with a 0.90 accuracy score and ±0.02 SD, which validates the superior performance of RF with the proposed hybrid feature set.

4.6. Best Performer Optimization Using Particle Swarm Optimization

We used the swam optimizer for hyperparameter tuning to select the best hyperparameters in comparison with our selected hyperparameters. For this purpose, we used the mealpy 1.0.2 library [42,43]. We imported a particle swarm optimization (PSO) model named BasePSO, and tuned the best performer RF with its two hyperparameters, n_estimators and max_depth. We set upper bound and lower values at 300 and 50, respectively, for both n_estimators and max_depth. We set the epochs value to 2 and the population size to 50 for PSO. The best solution using swam optimizer achieves a 0.94 accuracy as it optimizes the n_estimators value of 273 and max_depth value of 110. Table 14 shows the PSO epoch-wise RF results. It can be observed that RF is good as compared to other models, and the approach is good in terms of accuracy and computational cost.

4.7. Performance Comparison with State-of-the-Art Approaches

The performance comparison of the proposed approach is also carried out with existing state-of-the-art approaches. We performed a fair comparison by evaluating all approaches on the dataset that is used. We implemented the models from the selected studies according to their base paper architecture; however, the dataset of the current study is used. For comparison, only those studies are selected which investigate WBC-type classification using medical image datasets. We compared the results of the current study with [15], which proposed a CNN model for WBC classification. Similarly, the study [12] proposed a hybrid model by combining CNN and RNN to perform WBC classification. We selected these studies for comparison because they share the same topic. These studies used an imbalanced medical image dataset. In addition, two additional studies [34,44] that investigate medical image classification are also considered. For a fair comparison, the models presented in these studies are implemented using the dataset used in this study. Performance comparison is carried out in terms of accuracy and computation time. Table 15 shows the performance comparison of selected studies with the current study. It can be observed that the current study outperforms the existing studies both regarding the obtained accuracy and computation time, which shows the supremacy of this approach for WBC-type classification.

4.8. Statistical t-test

This study also performs a statistical significant t-test to show the significance of the proposed approach [45]. We deployed a t-test on each approach in comparison with the proposed approach and obtained results in form of acceptance or rejection of the null hypothesis.
  • Null Hypothesis ( H 0 ): if both compared results are statistically equal and there is no significant difference in results, it accepts the null hypothesis.
  • Alternative Hypothesis ( H a ): if both compared results are not statistically equal and there is a significant difference in results, it rejects the null hypothesis and accepts the alternative hypothesis.
Table 16 shows the t-test results for all cases. The t-test gives the T score and critical value ( C V ) score as outputs, and if T C V , then it accepts the null hypothesis; else it rejects the null hypothesis. t-test rejects the null hypothesis and accepts the alternative hypothesis in all cases when we compared the proposed approach results with other approaches which show that the proposed approach is statistically significant in comparison with other approaches.

5. Conclusions

White blood cells are an important part of the immune system, and they protect the body against infections and foreign substances. WBCs consist of different types, and abnormalities in a type of WBC may potentially represent a disease advocating the significance of WBC-type classification. Existing studies are limited by poor accuracy, model over-fitting due to an imbalanced dataset, and classification for a lower number of WBC types. From this perspective, this study presents a hybrid feature set of selective features using Chi2- and SMOTE-based oversampling to alleviate the influence of the above-mentioned problems. In addition, machine learning models are adopted to overcome the limitation of the data-intensive training time of deep learning models and robust results. The experimental results indicate that the feature set of both texture and RGB features from microscopic images selected using Chi2 produces a high accuracy of 0.97 with the RF model. Performance appraisals using k-fold cross-validation, T-statistic test, and comparison with existing state-of-the-art studies show the supremacy of the proposed approach, both in terms of the obtained accuracy and computational complexity. In the future, we intend to perform further experiments using derived features from microscopic data.

Author Contributions

Conceptualization, F.R.; Formal analysis, F.R., N.A., I.D.L.T.D., Y.D.K. and C.L.R.; Funding acquisition, I.D.L.T.D.; Investigation, N.A. and C.L.R.; Methodology, F.R. and Y.D.K. Project administration, Y.D.K. and J.L.V.M.; Resources, I.D.L.T.D. and J.L.V.M.; Software, N.A. and J.L.V.M.; Supervision, I.A.; Validation, F.R.; I.A.; Visualization, C.L.R.; Writing—original draft, F.R.; Writing—review and editing, I.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the European University of the Atlantic.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interests.

References

  1. Phosri, S.; Jangpromma, N.; Chang, L.C.; Tan, G.T.; Wongwiwatthananukit, S.; Maijaroen, S.; Anwised, P.; Payoungkiattikun, W.; Klaynongsruang, S. Siamese crocodile white blood cell extract inhibits cell proliferation and promotes autophagy in multiple cancer cell lines. J. Microbiol. Biotechnol. 2018, 28, 1007–1021. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. McKenzie, F.E.; Prudhomme, W.A.; Magill, A.J.; Forney, J.R.; Permpanich, B.; Lucas, C.; Gasser, R.A., Jr.; Wongsrichanalai, C. White blood cell counts and malaria. J. Infect. Dis. 2005, 192, 323–330. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Monteiro, A.C.B.; França, R.P.; Arthur, R.; Iano, Y. AI Approach Based on Deep Learning for Classification of White Blood Cells as a for e-Healthcare Solution. In Intelligent Interactive Multimedia Systems for e-Healthcare Applications; Springer: Berlin/Heidelberg, Germany, 2022; pp. 351–373. [Google Scholar]
  4. White Blood Cell. Available online: https://www.britannica.com/science/white-blood-cell (accessed on 10 April 2022).
  5. Zhu, B.; Feng, X.; Jiang, C.; Mi, S.; Yang, L.; Zhao, Z.; Zhang, Y.; Zhang, L. Correlation between white blood cell count at admission and mortality in COVID-19 patients: A retrospective study. BMC Infect. Dis. 2021, 21, 574. [Google Scholar] [CrossRef] [PubMed]
  6. Özyurt, F. A fused CNN model for WBC detection with MRMR feature selection and extreme learning machine. Soft Comput. 2020, 24, 8163–8172. [Google Scholar] [CrossRef]
  7. Hegde, R.B.; Prasad, K.; Hebbar, H.; Singh, B.M.K. Comparison of traditional image processing and deep learning approaches for classification of white blood cells in peripheral blood smear images. Biocybern. Biomed. Eng. 2019, 39, 382–392. [Google Scholar] [CrossRef]
  8. Bandyopadhyay, S.; Maulik, U.; Roy, D. Gene identification: Classical and computational intelligence approaches. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2007, 38, 55–68. [Google Scholar] [CrossRef]
  9. Min, S.; Lee, B.; Yoon, S. Deep learning in bioinformatics. Briefings Bioinform. 2017, 18, 851–869. [Google Scholar] [CrossRef] [Green Version]
  10. Bergeron, B.P. Bioinformatics Computing; Prentice Hall Professional: Hoboken, NJ, USA, 2003. [Google Scholar]
  11. Khouani, A.; El Habib Daho, M.; Mahmoudi, S.A.; Chikh, M.A.; Benzineb, B. Automated recognition of white blood cells using deep learning. Biomed. Eng. Lett. 2020, 10, 359–367. [Google Scholar] [CrossRef]
  12. Patil, A.; Patil, M.; Birajdar, G. White blood cells image classification using deep learning with canonical correlation analysis. IRBM 2021, 42, 378–389. [Google Scholar] [CrossRef]
  13. Wang, Q.; Wang, J.; Zhou, M.; Li, Q.; Wen, Y.; Chu, J. A 3D attention networks for classification of white blood cells from microscopy hyperspectral images. Opt. Laser Technol. 2021, 139, 106931. [Google Scholar] [CrossRef]
  14. Manthouri, M.; Aghajari, Z.; Safary, S. Computational Intelligence Method for Detection of White Blood Cells Using Hybrid of Convolutional Deep Learning and SIFT. Comput. Math. Methods Med. 2022, 2022, 9934144. [Google Scholar] [CrossRef] [PubMed]
  15. Girdhar, A.; Kapur, H.; Kumar, V. Classification of White blood cell using Convolution Neural Network. Biomed. Signal Process. Control 2022, 71, 103156. [Google Scholar] [CrossRef]
  16. Cheuque, C.; Querales, M.; León, R.; Salas, R.; Torres, R. An Efficient Multi-Level Convolutional Neural Network Approach for White Blood Cells Classification. Diagnostics 2022, 12, 248. [Google Scholar] [CrossRef] [PubMed]
  17. Habibzadeh, M.; Jannesari, M.; Rezaei, Z.; Baharvand, H.; Totonchi, M. Automatic white blood cell classification using pre-trained deep learning models: ResNet and Inception. In Proceedings of the Tenth International Conference on Machine Vision (ICMV 2017), Vienna, Austria, 13–15 November 2017; International Society for Optics and Photonics: Bellingham, DC, USA, 2018; Volume 10696, p. 1069612. [Google Scholar]
  18. Sharma, S.; Gupta, S.; Gupta, D.; Juneja, S.; Gupta, P.; Dhiman, G.; Kautish, S. Deep Learning Model for the Automatic Classification of White Blood Cells. Comput. Intell. Neurosci. 2022, 2022, 7384131. [Google Scholar] [CrossRef] [PubMed]
  19. WBC Image Dataset. Available online: https://github.com/zxaoyou/segmentation_WBC (accessed on 23 October 2022).
  20. BCCD Dataset. Available online: https://www.kaggle.com/datasets/paultimothymooney/blood-cells (accessed on 23 October 2022).
  21. What Are White Blood Cells? Available online: https://www.urmc.rochester.edu/encyclopedia/content.aspx?ContentID=35&ContentTypeID=160 (accessed on 23 October 2022).
  22. Van der Walt, S.; Schönberger, J.L.; Nunez-Iglesias, J.; Boulogne, F.; Warner, J.D.; Yager, N.; Gouillart, E.; Yu, T. Scikit-image: Image processing in Python. PeerJ 2014, 2, e453. [Google Scholar] [CrossRef]
  23. Alipo-on, J.R.; Escobar, F.I.; Novia, J.L.; Atienza, M.M.; Mana-ay, S.; Tan, M.J.; AlDahoul, N.; Yu, E. Dataset for Machine Learning-Based Classification of White Blood Cells of the Juvenile Visayan Warty Pig. 2022. Available online: https://ieee-dataport.org/documents/dataset-machine-learning-based-classification-white-blood-cells-juvenile-visayan-warty-pig (accessed on 15 July 2022).
  24. Álvarez, M.J.; González, E.; Bianconi, F.; Armesto, J.; Fernández, A. Colour and texture features for image retrieval in granite industry. Dyna 2010, 77, 121–130. [Google Scholar]
  25. Gosain, M.; Kaur, S.; Gupta, S. Review of Rotation Invariant Classification Methods for Color Texture Analysis. 2021. Available online: https://www.academia.edu/45598787/IRJET_Review_of_Rotation_Invariant_Classification_Methods_for_Color_Texture_Analysis?from=cover_page (accessed on 15 July 2022).
  26. Liu, H.; Setiono, R. Chi2: Feature selection and discretization of numeric attributes. In Proceedings of the 7th IEEE International Conference on Tools with Artificial Intelligence, Herndon, VA, USA, 5–8 November 1995; pp. 388–391. [Google Scholar]
  27. Rupapara, V.; Rustam, F.; Shahzad, H.F.; Mehmood, A.; Ashraf, I.; Choi, G.S. Impact of SMOTE on imbalanced text features for toxic comments classification using RVVC model. IEEE Access 2021, 9, 78621–78634. [Google Scholar] [CrossRef]
  28. Rustam, F.; Ashraf, I.; Mehmood, A.; Ullah, S.; Choi, G.S. Tweets classification on the base of sentiments for US airline companies. Entropy 2019, 21, 1078. [Google Scholar] [CrossRef] [Green Version]
  29. Mendieta, M.; Neff, C.; Lingerfelt, D.; Beam, C.; George, A.; Rogers, S.; Ravindran, A.; Tabkhi, H. A Novel Application/Infrastructure Co-design Approach for Real-time Edge Video Analytics. In Proceedings of the 2019 SoutheastCon, Raleigh, NC, USA, 10–14 April 2019; pp. 1–7. [Google Scholar]
  30. George, A.; Ravindran, A. Distributed middleware for edge vision systems. In Proceedings of the 2019 IEEE 16th International Conference on Smart Cities: Improving Quality of Life Using ICT & IoT and AI (HONET-ICT), Charlotte, NC, USA, 6–9 October 2019; pp. 193–194. [Google Scholar]
  31. Shafique, R.; Siddiqui, H.U.R.; Rustam, F.; Ullah, S.; Siddique, M.A.; Lee, E.; Ashraf, I.; Dudley, S. A novel approach to railway track faults detection using acoustic analysis. Sensors 2021, 21, 6221. [Google Scholar] [CrossRef]
  32. George, A.; Ravindran, A.; Mendieta, M.; Tabkhi, H. Mez: An adaptive messaging system for latency-sensitive multi-camera machine vision at the iot edge. IEEE Access 2021, 9, 21457–21473. [Google Scholar] [CrossRef]
  33. Rustam, F.; Reshi, A.A.; Aljedaani, W.; Alhossan, A.; Ishaq, A.; Shafi, S.; Lee, E.; Alrabiah, Z.; Alsuwailem, H.; Ahmad, A.; et al. Vector mosquito image classification using novel RIFS feature selection and machine learning models for disease epidemiology. Saudi J. Biol. Sci. 2022, 29, 583–594. [Google Scholar] [CrossRef] [PubMed]
  34. Reshi, A.A.; Rustam, F.; Mehmood, A.; Alhossan, A.; Alrabiah, Z.; Ahmad, A.; Alsuwailem, H.; Choi, G.S. An efficient CNN model for COVID-19 disease detection based on X-ray image classification. Complexity 2021, 2021, 6621607. [Google Scholar] [CrossRef]
  35. Brijain, M.; Patel, R.; Kushik, M.; Rana, K. A survey on decision tree algorithm for classification. Int. J. Eng. Dev. Res. 2014, 1, 5. [Google Scholar]
  36. Patange, A.; Jegadeeshwaran, R.; Bajaj, N.; Khairnar, A.; Gavade, N. Application of Machine Learning for Tool Condition Monitoring in Turning. Sound Vib 2022, 56, 127–145. [Google Scholar] [CrossRef]
  37. Kotsiantis, S.B. Decision trees: A recent overview. Artif. Intell. Rev. 2013, 39, 261–283. [Google Scholar] [CrossRef]
  38. Qi, Y. Random forest for bioinformatics. In Ensemble Machine Learning; Springer: Berlin/Heidelberg, Germany, 2012; pp. 307–323. [Google Scholar]
  39. Sarkar, M.; Leong, T.Y. Application of K-nearest neighbors algorithm on breast cancer diagnosis problem. In Proceedings of the AMIA Symposium, American Medical Informatics Association, Los Angeles, CA, USA, 4–8 November 2000; p. 759. [Google Scholar]
  40. Suthaharan, S. Support vector machine. In Machine Learning Models and Algorithms for Big Data Classification; Springer: Berlin/Heidelberg, Germany, 2016; pp. 207–235. [Google Scholar]
  41. Ji, Q.; Huang, J.; He, W.; Sun, Y. Optimized deep convolutional neural networks for identification of macular diseases from optical coherence tomography images. Algorithms 2019, 12, 51. [Google Scholar] [CrossRef] [Green Version]
  42. Thieu, N.V.; Mirjalili, S. MEALPY: A Framework of The State-of-the-Art Meta-Heuristic Algorithms in Python. Available online: https://pypi.org/project/mealpy/1.0.2/ (accessed on 15 July 2022).
  43. Deo, T.Y.; Patange, A.D.; Pardeshi, S.S.; Jegadeeshwaran, R.; Khairnar, A.N.; Khade, H.S. A white-box SVM framework and its swarm-based optimization for supervision of toothed milling cutter through characterization of spindle vibrations. arXiv 2021, arXiv:2112.08421. [Google Scholar]
  44. Rustam, F.; Siddique, M.A.; Siddiqui, H.U.R.; Ullah, S.; Mehmood, A.; Ashraf, I.; Choi, G.S. Wireless capsule endoscopy bleeding images classification using CNN based model. IEEE Access 2021, 9, 33675–33688. [Google Scholar] [CrossRef]
  45. Omar, B.; Rustam, F.; Mehmood, A.; Choi, G.S. Minimizing the overlapping degree to improve class-imbalanced learning under sparse feature selection: Application to fraud detection. IEEE Access 2021, 9, 28101–28110. [Google Scholar]
Figure 2. Hierarchy and classification of WBCs.
Figure 2. Hierarchy and classification of WBCs.
Healthcare 10 02230 g002
Figure 3. Flow of the proposed methodology.
Figure 3. Flow of the proposed methodology.
Healthcare 10 02230 g003
Figure 4. Process followed to make the hybrid feature set.
Figure 4. Process followed to make the hybrid feature set.
Healthcare 10 02230 g004
Figure 5. Approach followed for data balance using SMOTE.
Figure 5. Approach followed for data balance using SMOTE.
Healthcare 10 02230 g005
Figure 6. Comparison between all machine learning approaches used for classification of WBC types.
Figure 6. Comparison between all machine learning approaches used for classification of WBC types.
Healthcare 10 02230 g006
Figure 7. RF confusion matrix using HB features.
Figure 7. RF confusion matrix using HB features.
Healthcare 10 02230 g007
Figure 8. Comparison of original features space and proposed features space. We visualize features in 3D space and show that the features set makes a good correlation with target classes or not. If there will be less overlapping in different target samples, this means that the feature set is good for achieving significant results. We used PCA to convert our features into 3 dimensions, and then illustrated them using a scatter plot. (a) Original dataset RGB features, (b) Original dataset texture features, (c) Hybrid features space at 0°, and (d) Hybrid features space at 90°.
Figure 8. Comparison of original features space and proposed features space. We visualize features in 3D space and show that the features set makes a good correlation with target classes or not. If there will be less overlapping in different target samples, this means that the feature set is good for achieving significant results. We used PCA to convert our features into 3 dimensions, and then illustrated them using a scatter plot. (a) Original dataset RGB features, (b) Original dataset texture features, (c) Hybrid features space at 0°, and (d) Hybrid features space at 90°.
Healthcare 10 02230 g008
Figure 9. Architecture of the state-of-the-art deep learning models. (a) CNN, (b) ResNet50 [41], and (c) VVG16.
Figure 9. Architecture of the state-of-the-art deep learning models. (a) CNN, (b) ResNet50 [41], and (c) VVG16.
Healthcare 10 02230 g009
Figure 10. Training and testing per epochs accuracy and loss graphs. (a) Accuracy on original dataset, (b) Loss on original dataset, (c) Accuracy on augmented dataset, (d) Loss on augmented dataset, (e) Accuracy on over-sampled dataset, and (f) Loss on over-sampled dataset.
Figure 10. Training and testing per epochs accuracy and loss graphs. (a) Accuracy on original dataset, (b) Loss on original dataset, (c) Accuracy on augmented dataset, (d) Loss on augmented dataset, (e) Accuracy on over-sampled dataset, and (f) Loss on over-sampled dataset.
Healthcare 10 02230 g010
Table 1. Summary of literature review.
Table 1. Summary of literature review.
Ref.YearApproachAim and ResultsDataset
[17]2018Pre-trained models ResNet and InceptionWBC detection, Achieved 99.84% and 99.46% accuracy rates with ResNet V1 152 and ResNet 101WBC microscopic image dataset [19].
[7]2019CNN with handcrafted featuresWBC detection: CNN achieved 99% accuracyRBCs, WBCs images dataset [7]. The augmented dataset consists of 2697 cropped images in the `training set’ and 816 cropped images in the `test set’.
[6]2020CNN-based models features and ELM modelWBC detection, ELM achieved 96.03% accuracy.BCCD dataset, Microscopic WBC dataset consists of 12,494 images, including 3120 eosinophils, 3108 lymphocytes, 3095 monocytes, and 3171 neutrophils [20].
[12]2021Canonical Correlation Analysis (CCA) applied to (CNN + RNN)WBC types classification, Model achieved 95.89% accuracyBCCD dataset [20] and Kaggle dataset
[13]20213D CNN and 3D CNN with attention moduleWBC types classification, Model achieved 97.72% accuracy score.Three independent datasets with 215 patient samples.
[14]2022SIFT features with CNN modelWBC detection, Models achieved LISC = 95.84% and WBCis = 97.33% accuracies.LISC and WBCis datasets [19].
[15]2022CNN modelWBC types classification, Model achieved 98.55% accuracy score.WBC Kaggle dataset [20].
[16]2022Multi-level CNN modelWBC types classification, Model achieved 98.4% accuracy score.Blood Cell Detection (BCD) dataset, Complete Blood Count (CBC) dataset, White Blood Cells (WBC) dataset, Kaggle Blood Cell Images (KBC) dataset, Leukocyte Images for Segmentation and Classification (LISC) dataset
[18]2022DenseNet121 modelWBC types classification, Model achieved 98.84% accuracy score.KBC dataset
Table 2. Sample count in the original dataset.
Table 2. Sample count in the original dataset.
WBC TypeOriginalAugmented
Neutrophil319194
Lymphocyte905-
Monocyte82418
Eosinophil82405
Basophil20447
Table 3. Number of features with individual technique.
Table 3. Number of features with individual technique.
Feature TypeNo. of Features
RGB67,500
Texture22,500
Table 4. Number of features.
Table 4. Number of features.
FeatureTotal No. of FeaturesChi2
RGB67,5003000
Texture22,5003000
HF6000
Table 5. Sample count after over-sampling of data.
Table 5. Sample count after over-sampling of data.
WBC TypeOriginalAugmentedOversampled
Neutrophil319194905
Lymphocyte905-905
Monocyte82418905
Eosinophil82405905
Basophil20447905
Table 6. Target class count in the training and testing sets.
Table 6. Target class count in the training and testing sets.
ClassOversampling Data
Training SetTesting SetTotal
Neutrophil709196905
Lymphocyte728177905
Monocyte719186905
Eosinophil726179905
Basophil738167905
Original Data
Neutrophil24970319
Lymphocyte732173905
Monocyte691382
Eosinophil592382
Basophil17320
Table 7. Hyperparameters setting for machine learning models.
Table 7. Hyperparameters setting for machine learning models.
ModelHyperparametersHyperparameters Tuning
DTmax_depth = 200max_depth = {2 to 500}
RFmax_depth = 200,
n_estimators = 300
max_depth = {2 to 500}, n_estimators = {50 to 500}
KNNn_neighbors = 5n_neighbors = {2 to 10}
SVMKernel = poly, C = 4.0Kernel = {poly, linear, sigmoid}
C = {1.0 to 5.0}
Table 8. Results of machine learning modes on the original dataset.
Table 8. Results of machine learning modes on the original dataset.
FeatureModelAccuracyPrecisionRecallF1
TextureDT0.420.190.190.19
RF0.660.130.200.16
KNN0.430.190.230.19
SVM0.660.130.200.16
RGBDT0.490.220.220.22
RF0.620.360.210.19
KNN0.550.230.210.20
SVM0.560.250.230.23
Table 9. Results of models on the augmented dataset.
Table 9. Results of models on the augmented dataset.
FeatureModelAccuracyPrecisionRecallF1
TextureDT0.260.240.240.24
RF0.440.330.370.35
KNN0.250.060.250.10
SVM0.330.080.250.12
RGBDT0.410.390.380.38
RF0.640.630.570.54
KNN0.400.290.340.31
SVM0.440.410.410.41
Table 10. Machine Learning Results on Over-sampled Dataset.
Table 10. Machine Learning Results on Over-sampled Dataset.
FeatureModelAccuracyPrecisionRecallF1
TextureDT0.740.740.740.74
RF0.920.920.920.92
KNN0.770.650.750.68
SVM0.770.690.760.70
RGBDT0.610.610.620.61
RF0.870.870.870.87
KNN0.700.690.700.69
SVM0.810.810.810.81
Table 11. Results of models using the hybrid features from an oversampled dataset.
Table 11. Results of models using the hybrid features from an oversampled dataset.
FeatureModelAccuracyPrecisionRecallF1
HFDT0.730.730.730.73
RF0.970.970.970.97
KNN0.780.800.780.72
SVM0.890.890.890.89
Table 12. Deep learning models performances with each approach.
Table 12. Deep learning models performances with each approach.
SamplingModelAccuracyPrecisionRecallF1 Score
OriginalCNN0.580.190.220.20
VVG160.740.290.310.30
ResNet-500.680.250.290.27
AugmentionCNN0.580.190.220.20
VVG160.740.290.310.30
ResNet-500.680.250.290.27
OversamplingCNN0.730.740.730.71
VVG160.200.060.200.07
ResNet-500.200.040.200.06
Table 13. Ten-fold cross-validation results of machine learning models using the proposed approach.
Table 13. Ten-fold cross-validation results of machine learning models using the proposed approach.
ModelAccuracySD
DT0.76±0.03
KNN0.79±0.03
RF0.95±0.03
SVM0.90±0.02
Table 14. RF hyperparameters optimization using PSO.
Table 14. RF hyperparameters optimization using PSO.
EpochCurrent BestGlobal BestRuntimeN_estimatorsMax_Depth
Epoch 10.940.94691169253
Epoch 20.940.94655273110
Best Fitness0.94
Table 15. Comparison with existing state-of-the-art approaches.
Table 15. Comparison with existing state-of-the-art approaches.
Ref.YearModelAcc.T (s)
[34]2021CNN46%1244
[44]2021MobileNet+CNN57%601
[12]2021CNN+RNN51%9612
[15]2022CNN72%1109
This study2022RF97%34
Table 16. Statistically significant t-test.
Table 16. Statistically significant t-test.
CaseT CV H 0
HF + SMOTE V
Original
12.1250.000Rejected
HF + SMOTE V
Augmented
25.6980.000Rejected
HF + SMOTE V
Over-sampled
16.1870.000Rejected
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rustam, F.; Aslam, N.; De La Torre Díez, I.; Khan, Y.D.; Mazón, J.L.V.; Rodríguez, C.L.; Ashraf, I. White Blood Cell Classification Using Texture and RGB Features of Oversampled Microscopic Images. Healthcare 2022, 10, 2230. https://doi.org/10.3390/healthcare10112230

AMA Style

Rustam F, Aslam N, De La Torre Díez I, Khan YD, Mazón JLV, Rodríguez CL, Ashraf I. White Blood Cell Classification Using Texture and RGB Features of Oversampled Microscopic Images. Healthcare. 2022; 10(11):2230. https://doi.org/10.3390/healthcare10112230

Chicago/Turabian Style

Rustam, Furqan, Naila Aslam, Isabel De La Torre Díez, Yaser Daanial Khan, Juan Luis Vidal Mazón, Carmen Lili Rodríguez, and Imran Ashraf. 2022. "White Blood Cell Classification Using Texture and RGB Features of Oversampled Microscopic Images" Healthcare 10, no. 11: 2230. https://doi.org/10.3390/healthcare10112230

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop