Next Article in Journal
Effects of the Government Response and Community Mobility on the COVID-19 Pandemic in Southeast Asia
Previous Article in Journal
Efficacy of Nasal High-Flow Oxygen Therapy in Chronic Obstructive Pulmonary Disease Patients in Long-Term Oxygen and Nocturnal Non-Invasive Ventilation during Exercise Training
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Cervical Cancer Diagnosis Using an Integrated System of Principal Component Analysis, Genetic Algorithm, and Multilayer Perceptron

Department of Systems Science and Industrial Engineering, Binghamton University, Binghamton, NY 13902, USA
*
Author to whom correspondence should be addressed.
Healthcare 2022, 10(10), 2002; https://doi.org/10.3390/healthcare10102002
Submission received: 9 September 2022 / Revised: 6 October 2022 / Accepted: 7 October 2022 / Published: 11 October 2022

Abstract

:
Cervical cancer is one of the most dangerous diseases that affect women worldwide. The diagnosis of cervical cancer is challenging, costly, and time-consuming. Existing literature has focused on traditional machine learning techniques and deep learning to identify and predict cervical cancer. This research proposes an integrated system of Genetic Algorithm (GA), Multilayer Perceptron (MLP), and Principal Component Analysis (PCA) that accurately predicts cervical cancer. GA is used to optimize the MLP hyperparameters, and the MLPs act as simulators within the GA to provide the prediction accuracy of the solutions. The proposed method uses PCA to transform the available factors; the transformed features are subsequently used as inputs to the MLP for model training. To contrast with the PCA method, different subsets of the original factors are selected. The performance of the integrated system of PCA–GA–MLP is compared with nine different classification algorithms. The results indicate that the proposed method outperforms the studied classification algorithms. The PCA–GA–MLP model achieves the best accuracy in diagnosing Hinselmann, Biopsy, and Cytology when compared to existing approaches in the literature that were implemented on the same dataset. This study introduces a robust tool that allows medical teams to predict cervical cancer in its early stage.

1. Introduction

Cancer is a leading cause of death across the world. In 2020, around 604,000 women were diagnosed with cervical cancer and 342,000 cervical cancer deaths were recorded [1]. Cervical cancer is one of the most dangerous diseases for women, and approximately 80% of those diagnosed were aged 15 to 45.
Cervical cancer is caused by mutations in genes that regulate cell division and proliferation. Two associated symptoms of early-stage cervical cancer are pelvic discomfort and vaginal bleeding. Because cervical cancer cannot be diagnosed in its early stages, these symptoms are the only early warning signs. If gone unnoticed, cervical cancer can spread to other body regions, such as the lungs and abdomen. Cervical cancer can be identified with diffusion-weighted and Magnetic Resonance Imaging in its later stages (diffusion-weighted imaging). Symptoms, which include tiredness, back discomfort, leg pain, weight loss, and potential bone fractures, often become more severe as the disease progresses [1].
Cervical cancer risk is raised by early pregnancy, contraception usage, numerous pregnancies, cigarette use, and Human Papillomavirus (HPV) [2]. HPV, one of the most critical risk factors for cervical cancer, is a DNA virus that spreads mainly through sexual interaction. HPV is identified in cervical cancer patients 99.7% of the time and consists of more than 100 distinct factors [3].
Cervical cancer is detected using four specific tests: Hinselmann, Schiller, Cytology, and Biopsy (HSCB). Hinselmann is a member of the “colposcopy with acetic acid” group. Schiller, Cytology, and Biopsy, on the other hand, use Lugol iodine [4]. Diagnosing cervical cancer is expensive and time-consuming. Unfortunately, low-income nations face significant challenges in raising cancer awareness and screening. In addition, a lack of resources, such as medical expertise, equipment, and specialist doctors, contributes to the spread of cervical cancer in developing nations. As a result, patient fatality rates are rising [5].
In recent decades, Machine Learning (ML) techniques have been used to identify and predict cervical cancer. Among the most widely used approaches are Neural Network (NN), Ensemble Learning, Support Vector Machine (SVM), K-Nearest Neighbor (KNN), Texture Analysis, and Deep Learning [6,7,8,9,10,11,12,13,14,15,16,17].
One of the most effective techniques for predicting cervical cancer is NN. There are several types of NNs, such as Multilayer Perceptron (MLP), Convolution Neural Network (CNN), Probabilistic Neural Network (PNN), and Recurrent Neural Network (RNN) [18,19]. Many researchers over the last decade have endorsed the use of MLP when predicting cervical cancer because it provides a decent classification accuracy [19,20].
The MLP is a feedforward NN that learns via the Backpropagation method. It has an input layer of neurons that function as receivers, and it contains one or more hidden layers of neurons that compute data and iterate. Finally, an output layer predicts the result, as summarized in Figure 1.
Genetic Algorithm (GA) is a stochastic population searching technique that examines large search areas efficiently. GA has been used to optimize a range of MLP parameters, including momentum, size of the neurons, stopping criteria, solver(s), and activation function(s) [19,21].
This paper is organized as follows: Section 2 discusses related literature on cervical cancer prediction. Section 3 summarizes the contribution of this research. Section 4 describes the research methods, data preprocessing, dataset for the study, and procedures used to combine GA with MLP. Section 5 discusses the results of the proposed method and compares the proposed method with other classification approaches. Finally, Section 6 provides the conclusions and future research directions.

2. Related Literature

This section discusses and criticizes relevant current ML approaches for predicting cervical cancer. Most of the studies that have been conducted to predict cervical cancer used the cervical cancer dataset from the University of California Irvine (UCI) Machine Learning Repository.
Wu and Zhou [4] implemented three different methods to diagnose four target variables of cervical cancer: HSCB. The dataset contained 32 factors that potentially cause cervical cancer. The authors used the Random Oversampling method to balance the dataset of 668 patients. They combined SVM with Recursive Feature Elimination (RFE), combined SVM with Principal Component Analysis (PCA), and employed the traditional SVM, and they compared these method’s characterization of cervical cancer. The classification results show that SVM–RFE and SVM–PCA provide more accurate results when selecting only eight factors instead of using the traditional SVM techniques [4,22].
Abdoh et al. [22] predicted cervical cancer by using RFE and PCA to eliminate several features. The researchers used the Synthetic Minority Oversampling Technique (SMOTE) to balance the cervical cancer dataset from UCI. Random Forest (RF) was used to diagnose cervical cancer and was compared with RF–PCA and RF–RFE. Their results show that integrating SMOTE with RF classifiers enhances classification accuracy by about 4% when compared with the work conducted by Wu and Zhou [4].
Deng et al. [23] applied RF, SVM, and eXtreme Gradient Boosting (XGBoost) to diagnose four target variables of cervical cancer: HSCB. Similar to Abdoh et al. [22], the authors used SMOTE to balance the dataset. However, their research did not use feature elimination techniques. Their results show that RF and XGBoost performed better than SVM in terms of classification accuracy. The results of Deng et al. [23] are similar to those reported by Abdoh et al. [22].
Adem et al. [15] used Deep Learning to predict the same dataset of cervical cancer by employing Softmax classification with a stacked autoencoder. The stacked autoencoder was used as a dimensional reduction tool, while the softmax layer was used to predict HSCB of cervical cancer. The authors validated their approach with six ML methods: RF, Decision Tree (DT), MLP, SVM, Rotation Forest models, and KNN [15]. Their proposed approach for diagnosing the four target variables (HSCB) of cervical cancer achieved better performance than the method of Wu and Zhou (2017), with close to 4% improvement in classification accuracy.
Alsmariy et al. [24] used RF, Logistic Regression (LR), and DT to diagnose HSCB of cervical cancer. Similar to Deng et al. [23] and Abdoh et al. [22], the authors applied SMOTE to balance the dataset. They used PCA as a feature reduction approach. They implemented a voting technique that enables several algorithms to vote to select a winner. Their results perform better than those of Abdoh et al. [22], Deng et al. [23], and Wu and Zhou [4].
Wahid and Al-Mazini [5] adopted a meta-heuristic algorithm, Ant Colony Optimization (ACO) with the Ant-Miner data classification rule, to select the most critical risk factors to diagnose HSCB of cervical cancer. The authors claimed that their method was used for the first time in the literature for the UCI cervical cancer dataset. Their classification results perform better than those of Wu and Zhou [4] but are inferior to the reported results of Alsmariy et al. [24], Abdoh et al. [22], and Deng et al. [23].
Other researchers such as Devi et al. [10] and Fernandes et al. [14] used image processing and Deep Learning to diagnose and predict cervical cancer with a different dataset from the UCI. Devi et al. [10] classified cervical cancer into normal and abnormal cells using NN and Learning Vector Quantification (LVQ). Digital photographs of patients were used as inputs. To diagnose cervical cancer, LVQ was used to obtain the coefficient mean value of the extracted photographs. Their model achieves a 90% classification accuracy [10].
In contrast, deep learning algorithms were used by Fernandes et al. [14] to predict cervical cancer. Their approach is founded on a loss function that permits dimensional reduction to identify the most important classification variables. They concentrate on a specific form of cervical cancer called biopsy. Their algorithm achieves an Area Under the Curve (AUC) of 0.6875 [14].
There were a few studies that combined GA with MLP to diagnose cervical cancer. In one study, GA was used to determine the optimal initial weights and bias of MLP to classify cervical cancer [25]. The study was conducted on a dataset with 401 patients, of which 51.2% had cervical cancer, and included 16 risk factors. The accuracy of their proposed model improved from 94.51% to 96.26% when combining GA with MLP. Similarly, another study adopted GA to optimize the MLP’s initial weights and threshold to identify Nanoparticle (NP) sensors in the early diagnosis of cervical cancer cells. Their study compared the performance of the GA and MLP combination with a standalone MLP. Their results indicated that combining GA with MLP achieved statistically better root mean Square (RMS) and Mean Absolute Error (MAE) than MLP alone [26].
In summary, most of the research used classical machine learning classification algorithms and deep learning approaches to diagnose cervical cancer. The literature on diagnosing HSCB cited above is also summarized in Section 5 where it compared with the proposed hybrid system of PCA–GA–MLP. There were two studies that used GA to optimize MLP’s initial weights, threshold, or bias to diagnose cervical cancer. However, none of the research optimized the parameters of MLP, which include the size of each hidden layer, solvers, and activation functions, to diagnose cervical cancer. Further, a hybrid model of PCA–GA–MLP for the diagnosis of cervical cancer has not been proposed in the literature.

3. Contribution

This study is the first that integrates PCA, GA, and MLP altogether in one framework that accurately predicts cervical cancer using the benchmark dataset from UCI. The proposed method transforms all available features using the PCA method. The transformed features are utilized in model constructions of the MLP, which is within a hybrid system of GA and MLP. GA is used to optimize the MLP parameters, whereas the MLP acts as a simulator within the GA. The hybrid system iteratively evolves the optimal design of MLP that provides the best cervical cancer classification accuracy. The developed framework introduces a robust tool that allows medical teams to predict cervical cancer as a preventive strategy that reduces cervical cancer rates and costs while improving the quality of care for cancer patients.

4. Research Methodology

This research has four main steps, as summarized in Figure 2. Step 1 involves preprocessing and balancing the dataset. Step 2 describes the application of the feature selection process. In this research, four feature selection approaches separately diagnose each target variable/test of cervical cancer: using the transformed features from PCA; using all original features; using the top 18 features based on RF importance; and using the top 10 features based on RF importance. Step 3 explains how GA is used as an optimization tool to determine the optimal parameters of the MLP to predict cervical cancer. This process is applied to the four target variables (HSCB) of cervical cancer separately. Four feature selection approaches are implemented for each target variable, which results in 16 different scenarios (i.e., four scenarios for each cervical cancer variable: PCA–GA–MLP, GA–MLP using all 30 factors, GA–MLP using the top 18 factors, and GA–MLP using the top 10 factors). Step 4 determines the performance measures for each scenario using a 5-fold cross-validation and compares each scenario’s results with nine other classification algorithms. Five-fold cross-validation is used because it allows all available data instances to be used for both model development and model validation [27]. The nine algorithms are as follows: RF, Linear Discriminant Analysis (LDA), SVM, LR, Gaussian Naïve Bayes (NB), KNN, DT, Adaptive Boosting (AdaBoost), and Centroid-Displacement-based KNN (CD-KNN) [28]. Therefore, a total of 160 different experiments are implemented.
Over the past decade, researchers have advocated the use of MLP in cervical cancer prediction due to its respectable classification accuracy [10,18,19,20,25,26,29,30]; therefore, MLP is selected for optimization by GA in this study. In addition, various researchers used SVM, RF, LR DT, KNN, NB, LDA, and AdaBoost to classify cervical cancer [4,9,10,11,13,22,23,30,31,32,33]. Those algorithms are among the most common classification algorithms in modeling medical datasets. Therefore, those algorithms are used in this study to validate the proposed approach, and their results are compared.

4.1. Dataset Description and Data Preprocessing

UCI published the cervical cancer dataset. The dataset contains 858 patients that have 32 cervical cancer factors as summarized in Table 1. It has four cervical cancer target variables: Hinselmann, Schiller, Cytology, and Biopsy. Each cervical cancer target variable is treated as a separate problem in this research. Therefore, four datasets are prepared to be used separately: Hinselmann, Schiller, Cytology, and Biopsy.
Because some patients refused to answer personal questions, some data are missing. As a result, the dataset needed to be preprocessed to account for the null data. Two factors were removed due to having a large amount of missing data (i.e., factor numbers 27, and 28 in Table 1). Further, some samples were removed for the same reason. As a result, there are 668 patients in the final (preprocessed) dataset. Each data instance has 30 distinct risk factors and four target variables (HSCB) of cervical cancer. Normalization is used for some numerical factors to eliminate data redundancy and avoid any undesirable characteristics due to the wide range of values in those factors. The percentage of cervical cancer patients within the dataset is 4.5% for Hinselmann, 9.4% for Schiller, 5.8% for Cytology, and 6.7% for Biopsy. Therefore, the random oversampling technique is used to balance the unbalanced dataset.

4.2. Feature Selection and Principal Component Analysis

Feature selection is the process of minimizing the number of input variables when creating a predictive model. The number of input variables might be reduced to decrease the computational cost of modeling and, in some circumstances, to improve the model’s performance [33,34].
RF is a Bagging algorithm that mixes several different DTs. Tree-based techniques in RFs are naturally rated by how well they improve node purity or, in other words, how well they minimize impurity (Gini impurity) across all trees. The nodes with the greatest reduction in impurity are located at the beginning of the trees, while those with the smallest reduction are found at the conclusion. A subset of the most important features may be created by pruning trees below a certain node [35,36,37]. The proposed framework investigates two feature selection methods, PCA and RF importance, where RF is used to select the best 18 factors and the best 10 factors with the highest relative importance in predicting cervical cancer, as shown in Table 2 and Table 3. The factors listed in Table 2 and Table 3 refer to the factor numbers provided in Table 1.
PCA is a statistical method that uses the eigenvector to determine the orientation of features. PCA’s fundamental idea is to map a j-dimensional feature space into an i-dimensional space, which is generally known as the principal components, where i < j. The covariance matrix is calculated, and the eigenvectors and eigenvalues are computed. Because an eigenvalue shows the most significant relationship between the dataset characteristics, the eigenvector with the greatest eigenvalue is selected as the principal component of the cervical cancer dataset. The eigenvalues are sorted in ascending order to select the most significant principal component(s), while the lowest eigenvalues are discarded. This process reduces large dimensional datasets to smaller dimensional datasets. The variance measures the dispersion of the data in the cervical cancer dataset. Lastly, eigenvectors and eigenvalues for the covariance matrix are computed. Eigenvalues are transformed using varimax orthogonal rotation or oblique rotation [38].
In this research, PCA transforms the 30 original factors into a two-dimensional space, which accounts for most of the variance in the original dataset. The PCA approach reduces the noise in the cervical cancer dataset and also reduces the computational time for model development.

4.3. Combination of GA–MLP

GA is used in this research as an optimization tool to determine the optimal hyperparameters for the MLP that provides the highest classification accuracy in diagnosing each target variable of cervical cancer. The MLP has several hyperparameters that need to be fine-tuned, which include the size of each hidden layer, solvers, and activation functions. The hyperparameters of MLP are encoded as chromosomes in the GA. The population of solutions/chromosomes in the GA represents a population of MLPs. The classification accuracy of each MLP, after network training is completed, is used as the fitness value of that solution.
Initially, the GA has a random population of solutions. In each generation, each solution (MLP) goes through the training process to determine its fitness value (classification accuracy). The evaluated solutions will then go through the typical evolution process of the GA: select two parent solutions, crossover the parent solutions to create two children with a probability of Pc, and mutate the children with a probability of Pm. Once the internal loop is completed and reaches half of the population size (n/2), the replacement process will then cull all the parent solutions, the children will advance to the next generation, and the generation counter will increase by one. This process will continue until the termination criteria are met (gmax). Finally, the best solution out of all generations will be selected, which represents the MLP’s optimized parameters for predicting patients with cervical cancer. Algorithm 1 represents the pseudocode of the hybrid GA–MLP for optimizing the MLP parameters.
Algorithm 1. Hybrid GA–MLP for optimizing MLP parameters
1: Set GA parameters (Pc, Pm, n, gmax)
2: Encode solutions (MLP parameters) using real value encoding
3: Randomly generate n solutions
4: Calculate the fitness value of each solution by the trained MLPs
5: for i = 1, until gmax do
6:   for i = 1, until n/2 do
7:     Select two parents
8:     Crossover to create two children with Pc
9:     Mutate children with Pm
10:   end for
11: Replace parents with children
12: end for
13: Return the best solution

4.4. Main Operators and MLP Hyperparameters

The following parameters are used to design the GA: tournament selection is used to select the parents (k = 4) for breeding; the crossover probability is set at 1 to perform a single-point crossover operation; the mutation probability is set at 0.001; the population size is fixed at 50; and the stopping criterion is 200 generations.
The hyperparameters of MLP are the encoded solution vectors in the GA. Each solution vector considers 4 types of activation functions, 3 types of solvers, and 50 different sizes for the first and the second layers in the MLP. However, some hyperparameters for the MLP are fixed: the learning rate is set at 0.001, the momentum is set at 0.90, and the stopping criterion is set to 200 iterations. Each solution undergoes network training, and the classification accuracy of the trained MLP is used as its fitness value in the GA. The proposed methodology is coded in the Python 3.9 environment.

4.5. Performance Metrics

Accuracy, sensitivity, specificity, and precision are the most common data mining performance metrics [39]. These performance metrics are defined in Equations (1)–(5). The confusion matrix output determines the following metrics: True Negative (TN), True Positive (TP), False Negative (FN), and False Positive (FP) [40]. TP is the number of correct predictions that a patient has cervical cancer, i.e., cervical cancer is correctly classified as cervical cancer. TN is the number of correct predictions that a patient does not have cervical cancer, i.e., non-cervical cancer is correctly classified as non-cervical cancer. FN is the number of incorrect predictions that a patient does not have cervical cancer, i.e., cervical cancer is identified as non-cervical cancer. FP is the number of incorrect predictions that a patient has cervical cancer, i.e., non-cervical cancer is detected as cervical cancer. Table 4 summarizes the confusion matrix for diagnosing cervical cancer.
Accuracy = TP + TN TP + FP + TN + FN   ×   100
Sensitivity = TP TP + FN   ×   100
Specificity = TN TN + FP   ×   100
Precision = TP TP + FP   ×   100
F 1 Score = 2 ×   Sensitivity   ×   Precision     Sensitivity   +   Precision   ×   100
In cervical cancer, the performance metrics are interpreted as follows: accuracy refers to how well the model can accurately categorize TP and TN cervical cancer cases out of all instances. Sensitivity refers to the percentage that the model correctly classifies TP cervical cancer cases out of all patients with cervical cancer. The model’s specificity measures how well it can categorize individuals as not having cervical cancer out of those diagnosed without the disease. Precision refers to the percentage that the model correctly classifies TP cervical cancer cases out of all cases that are classified as cervical cancer. Finally, F1-score is generally considered a better performance measure; it represents a weighted average between sensitivity and precision.

5. Results and Discussion

Experimental results of each target variable of cervical cancer are discussed separately in this section. The proposed approach is compared with nine different classification algorithms. Further, the performance of the PCA–GA–MLP model is compared with the best results in the studies that were conducted on the same cervical cancer dataset from UCI.

5.1. Target Variable: Hinselmann

Table 5 summarizes the performance measures obtained when comparing GA–MLP with other classifiers to diagnose Hinselmann. When adopting PCA as a dimensional reduction tool, PCA–GA–MLP outperforms all nine classification algorithms; it has an accuracy of 98.20%, a sensitivity of 100.00%, a specificity of 96.37%, a precision of 96.54%, and an F1-score of 98.24%. PCA–AdaBoost is the next best method; it has an accuracy of 94.67%, a sensitivity of 100.00%, a specificity of 89.29%, a precision of 90.41%, and an F1-score of 94.96%.
When using all (available) 30 factors to diagnose Hinselmann, GA–MLP has the best accuracy, sensitivity, specificity, precision, and F1-score (97.57%, 100.00%, 95.15%, 95.36%, and 97.62%, respectively), followed by AdaBoost, KNN, and CD-KNN. When incorporating the top 18 factors to diagnose Hinselmann, GA–MLP provides the best results across all performance measures. When selecting the top 10 factors to diagnose Hinselmann, GA–MLP achieves the best accuracy, specificity, precision, and F1-score.
Figure 3 illustrates the performance of the hybrid system of PCA–GA–MLP and the three GA–MLPs that used the top 10 factors, the top 18 factors, and all 30 factors to diagnose Hinselmann. The figure indicates that marginal improvement in model performance is observed when a larger number of factors is used in GA–MLP. Furthermore, the hybrid system of PCA–GA–MLP outperforms all three GA–MLPs across all performance measures. The PCA–GA–MLP offers a simpler input structure in the MLP as a result of the dimension reduction in PCA.
As shown in Table 6, PCA–GA–MLP performs better than the methods reported by Wu and Zhou [4], Adem et al. [15], Alsmariy et al. [24], Abdoh et al. [22], Deng et al. [23], and Wahid and Al-Mazini [5], in terms of accuracy, sensitivity, and F1-score. In terms of specificity, PCA–GA–MLP performs better than the method reported by Wu and Zhou [4].

5.2. Target Variable: Schiller

The performance measures of GA–MLP and other algorithms in diagnosing Schiller are summarized in Table 7. When applying PCA as a feature selection method, PCA–GA–MLP achieves the highest accuracy, sensitivity, specificity, precision, and F1-score of 96.78%, 100.00%, 93.61%, 93.97%, and 96.87%, respectively. PCA–CD-KNN is the second-best method, with an accuracy of 87.52%, a sensitivity of 99.15%, a specificity of 75.88%, a precision of 80.50%, and an F1-score of 88.83%.
When using all 30 factors to diagnose Schiller, GA–MLP has the best accuracy, precision, and F1-score (94.13%, 89.82%, and 94.45%, respectively), followed by CD-KNN and KNN. When using the best 18 factors to diagnose Schiller, GA–MLP provides the best results across all five performance measures. In selecting the best 10 factors to diagnose Schiller, GA–MLP performs better than all other classifiers, in terms of accuracy, sensitivity, precision, and F1-score (93.39%, 99.17%, 88.98%, and 93.79%, respectively).
Figure 4 illustrates the performance of the hybrid system of PCA–GA–MLP and the GA–MLPs that incorporated the top 10 factors, the top 18 factors, and all 30 factors to diagnose Schiller. The hybrid system of PCA–GA–MLP outperforms all three GA–MLPs across all performance measures (with 2–3% improvement in accuracy, 4–6% improvement in specificity, 3–5% improvement in precision, and 2–3% improvement in F1-score). The GA–MLPs that used the top 10 factors, the top 18 factors, and all 30 factors to diagnose Schiller all have similar performance, regardless of the number of factors included in the GA–MLP.
The proposed method, PCA–GA–MLP, is compared with the best approaches in the literature in the diagnosis of Schiller, using the same benchmark dataset from UCI. As shown in Table 8, PCA–GA–MLP performs better than the methods reported by Wu and Zhou [4], Abdoh et al. [22], Deng et al. [23], and Wahid and Al-Mazini [5] in terms of accuracy, sensitivity, and F1-score. The PCA–GA–MLP method achieves the highest sensitivity when compared with all other approaches. However, the method of Alsmariy et al. [24] has slightly better accuracy, specificity, and F1-score than the proposed PCA–GA–MLP method.

5.3. Target Variable: Cytology

Table 9 compares the performance of GA–MLPs with other algorithms in diagnosing Cytology. When implementing PCA as a dimensional reduction technique, PCA–GA–MLP achieves the highest accuracy, specificity, precision, and F1-score of 97.54%, 95.15%, 95.27%, and 97.56%, respectively. The next best method is PCA–RF, which has an accuracy of 91.58%, a sensitivity of 95.30%, a specificity of 88.06%, a precision of 88.73%, and an F1-score of 91.83%.
When using all available (30) factors, only the top 18 factors, or only the top 10 factors to diagnose Cytology, GA–MLP outperforms all other methods across all performance measures, followed by KNN and CD-KNN.
Figure 5 compares the performance of the hybrid system of PCA–GA–MLP and the GA–MLPs that used the top 10 factors, the top 18 factors, and all 30 factors to diagnose Cytology. As shown in Figure 5, the hybrid system of PCA–GA–MLP outperforms all three GA–MLPs across all but one performance indicator. Comparing the performance of GA–MLPs that used the top 10 factors, the top 18 factors, and all 30 factors to diagnose Cytology, some improvement in model performance is observed when a larger number of factors is used in the GA–MLP.
Table 10 compares PCA–GA–MLP with other approaches from the literature in diagnosing Cytology. PCA–GA–MLP achieves better accuracy and F1-score than all reported approaches. In terms of sensitivity, PCA–GA–MLP ranks second, next to the method reported by Wu and Zhou [4]. In terms of specificity, PCA–GA–MLP performs better than the methods reported by Wu and Zhou [4] and Alsmariy et al. [24].

5.4. Target Variable: Biopsy

Table 11 compares the performance measures of GA–MLP with other classifiers to diagnose Biopsy. When adopting PCA as a dimensional reduction tool, PCA–GA–MLP achieves the highest accuracy, specificity, precision, and F1-score of 97.75%, 95.54%, 95.63%, and 97.76%, respectively. PCA–AdaBoost is the second-best method with an accuracy of 90.93%, a sensitivity of 94.24%, a specificity of 87.90%, a precision of 88.47%, and an F1-score of 91.19%.
When using all 30 factors or the top 10 factors to diagnose Biopsy, GA–MLP outperforms all other approaches in all performance measures. In the case of using the top 18 factors to diagnose Biopsy, GA–MLP outperforms all other approaches in terms of accuracy, sensitivity, precision, and F1-score.
Figure 6 compares the performance of the hybrid system of PCA–GA–MLP, and the three GA–MLPs that used the top 10 factors, the top 18 factors, and all 30 factors to diagnose Biopsy. As shown in Figure 6, the hybrid system of PCA–GA–MLP outperforms all three GA–MLPs across all but one performance indicators. All three GA–MLPs have similar performance regardless of the number of factors included in the GA–MLP models.
Table 12 compares the performance of PCA–GA–MLP with other approaches from the literature in diagnosing Biopsy. PCA–GA–MLP achieves better accuracy and F1-score than all reported approaches. In terms of sensitivity, PCA–GA–MLP ranks third, next to the methods reported by Wu and Zhou [4] and Alsmariy et al. [24].

5.5. Four Target Variables

This section compares the classification accuracy of the proposed PCA–GA–MLP method with reported studies from the literature for all four target variables of cervical cancer (HSCB) on the same dataset from UCI. The classification accuracy is the only performance measure that is reported in the six relevant studies from the literature.
As shown in Figure 7, PCA–GA–MLP achieves the highest classification accuracy in diagnosing Hinselmann, Cytology, and Biopsy. In diagnosing Schiller, PCA–GA–MLP ranks third, next to the methods reported by Alsmariy et al. [24] and Adem et al. [15].

6. Conclusions and Future Work

This research proposed an integrated system of PCA, GA, and MLP for diagnosing cervical cancer cases using the benchmark dataset from UCI. There are four target variables of cervical cancer in the dataset: Hinselmann, Schiller, Cytology, and Biopsy. Four feature selection approaches were explored; dimensional reduction was performed using the PCA method, and different subsets of the original factors were selected based on Random Forest Importance.
Experimental results show that the hybrid system of PCA–GA–MLP outperforms all other classification algorithms and the three GA–MLP versions on all four target variables. In comparison with the existing approaches in the literature that were implemented on the same cervical cancer dataset, the PCA–GA–MLP model achieves the highest classification accuracy in Hinselmann, Cytology, and Biopsy (98.20%, 97.54%, and 97.75%, respectively).
Given the growing interest in cervical cancer research using ML algorithms, this research developed a robust predictive tool for cervical cancer. Physicians and healthcare providers can use the proposed model to identify patients with cervical cancer in its early stages. For the identified cases, the medical team can focus their effort on preventive actions and plans to improve women’s care and ultimately reduce cervical cancer rates and the associated costs.
The proposed research has a few limitations. The current approach did not consider RFE as a tool for selecting the best risk factors. Furthermore, the random oversampling technique was used to balance the unbalanced dataset. For future work, feature selection using RFE and advanced data balancing techniques such as SMOTE and cost-sensitive learning can be explored in the proposed method to further improve the overall performance.

Author Contributions

Conceptualization, O.Y.D. and S.S.L.; methodology, O.Y.D. and S.S.L.; software, O.Y.D.; validation, O.Y.D. and S.S.L.; formal analysis, O.Y.D. and S.S.L.; investigation, O.Y.D. and S.S.L.; resources, O.Y.D.; data curation, O.Y.D. and S.S.L.; writing—original draft preparation, O.Y.D.; writing—review and editing, S.S.L.; visualization, O.Y.D.; supervision, S.S.L.; project administration, S.S.L.; funding acquisition, O.Y.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

UCI machine learning repository published the cervical cancer dataset (on UCI Machine Learning Repository: Cervical cancer (Risk Factors) Data Set).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sung, H.; Ferlay, J.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef]
  2. Luhn, P.; Walker, J.; Schiffman, M.; Zuna, R.E.; Dunn, S.T.; Gold, M.A.; Smith, K.; Mathews, C.; Allen, R.A.; Zhang, R.; et al. The Role of Co-Factors in the Progression from Human Papillomavirus Infection to Cervical Cancer. Gynecol. Oncol. 2013, 128, 265–270. [Google Scholar] [CrossRef] [Green Version]
  3. Nour, N.M. Cervical Cancer: A Preventable Death. Rev. Obs. Gynecol. 2009, 2, 240–244. [Google Scholar]
  4. Wu, W.; Zhou, H. Data-Driven Diagnosis of Cervical Cancer with Support Vector Machine-Based Approaches. IEEE Access 2017, 5, 25189–25195. [Google Scholar] [CrossRef]
  5. Wahid, J.; Al-Mazini, H.F.A. Classification of Cervical Cancer Using Ant-Miner for Medical Expertise Knowledge Management. In Proceedings of the Knowledge Management International Conference (KMICe), Miri Sarawak, Malaysia, 25–27 July 2018; pp. 25–27. [Google Scholar]
  6. Ramdhani, Y.; Riana, D. Hierarchical Decision Approach Based on Neural Network and Genetic Algorithm Method for Single Image Classification of Pap Smear. In Proceedings of the 2017 Second International Conference on Informatics and Computing (ICIC), Jayapura-Papua, Indonesia, 2–4 November 2017; pp. 1–6. [Google Scholar]
  7. Athinarayanan, S.; Srinath, M.V.; Kavitha, R. Detection and Classification of Cervical Cancer in Pap Smear Images Using EETCM, EEETCM & CFE Methods Based Texture Features and Various Classification Techniques. Int. J. Sci. Res. Sci. Eng. Technol. 2016, 2, 533–549. [Google Scholar]
  8. Khan, F.S.; Maqbool, F.; Razzaq, S.; Irfan, K.; Zia, T. The Role of Medical Expert Systems in Pakistan. In Proceedings of the World Academy of Science, Engineering and Technology. Int. J. Soc. Behav. Educ. Econ. Bus. Ind. Eng. 2008, 2, 12–14. [Google Scholar]
  9. Lokanayaki, K.; Malathi, A. Exploring on Various Prediction Model in Data Mining Techniques for Disease Diagnosis. Int. J. Comput. Appl. 2013, 77, 26–29. [Google Scholar] [CrossRef]
  10. Devi, M.A.; Ravi, S.; Vaishnavi, J. Classification of Cervical Cancer Using Artificial Neural Networks. Procedia Comput. Sci. 2016, 89, 465–472. [Google Scholar] [CrossRef] [Green Version]
  11. Sharma, M.; Singh, S.K.; Agrawal, P.; Madaan, V. Classification of Clinical Dataset of Cervical Cancer Using KNN. Indian J. Sci. Technol. 2016, 10, 1–5. [Google Scholar] [CrossRef] [Green Version]
  12. Ghoneim, A.; Muhammad, G.; Hossain, M.S. Cervical Cancer Classification Using Convolutional Neural Networks and Extreme Learning Machines. Future Gener. Comput. Syst. 2020, 102, 643–649. [Google Scholar] [CrossRef]
  13. Lu, J.; Song, E.; Ghoneim, A.; Alrashoud, M. Machine Learning for Assisting Cervical Cancer Diagnosis: An Ensemble Approach. Future Generation Computer Systems. Future Gener. Comput. Syst. 2020, 106, 199–205. [Google Scholar] [CrossRef]
  14. Fernandes, K.; Chicco, D.; Cardoso, J.S.; Fernandes, J. Supervised Deep Learning Embeddings for the Prediction of Cervical Cancer Diagnosis. PeerJ. Comput. Sci. 2018, 4, e154. [Google Scholar] [CrossRef] [Green Version]
  15. Adem, K.; Kiliçarslan, S.; Cömert, O. Classification and Diagnosis of Cervical Cancer with Stacked Autoencoder and Softmax Classification. Expert Syst. Appl. 2019, 115, 557–564. [Google Scholar] [CrossRef]
  16. Ma, Y.; Liang, F.; Zhu, M.; Chen, C.; Chen, C.; Lv, X. FT-IR Combined with PSO-CNN Algorithm for Rapid Screening of Cervical Tumors. Photodiagn. Photodyn. Ther. 2022, 39, 103023. [Google Scholar] [CrossRef]
  17. Park, Y.R.; Kim, Y.J.; Ju, W.; Nam, K.; Kim, S.; Kim, K.G. Comparison of Machine and Deep Learning for the Classification of Cervical Cancer Based on Cervicography Images. Sci. Rep. 2021, 11, 16143. [Google Scholar] [CrossRef] [PubMed]
  18. Driss, S.B.; Soua, M.; Kachouri, R.; Akil, M. A Comparison Study between MLP and Convolutional Neural Network Models for Character Recognition. In Proceedings of the Real-Time Image and Video Processing, San Diego, CA, USA, 6–10 August 2017; International Society for Optics and Photonics SPIE: Anaheim, CA, USA, 2017; p. 1022306. [Google Scholar]
  19. Mirjalili, S. Evolutionary Algorithms and Neural Networks; Springer: Cham, Switzerland, 2019; Volume 780. [Google Scholar]
  20. Sokouti, B.; Haghipour, S.; Tabrizi, A.D. A Framework for Diagnosing Cervical Cancer Disease Based on Feedforward MLP Neural Network and ThinPrep Histopathological Cell Image Features. Neural Comput. Appl. 2014, 24, 221–232. [Google Scholar] [CrossRef]
  21. Katoch, S.; Chauhan, S.S.; Kumar, V. A Review on Genetic Algorithm: Past, Present, and Future. Multimed. Tools Appl. 2021, 80, 8091–8126. [Google Scholar] [CrossRef]
  22. Abdoh, S.F.; Abo Rizka, M.; Maghraby, F.A. Cervical Cancer Diagnosis Using Random Forest Classifier with SMOTE and Feature Reduction Techniques. IEEE Access 2018, 6, 59475–59485. [Google Scholar] [CrossRef]
  23. Deng, X.; Luo, Y.; Wang, C. Analysis of Risk Factors for Cervical Cancer Based on Machine Learning Methods. In Proceedings of the 2018 5th IEEE International Conference on Cloud Computing and Intelligence Systems (CCIS), Nanjing, China, 23–25 November 2018; IEEE: New York, NY, USA, 2018; pp. 631–635. [Google Scholar]
  24. Alsmariy, R.; Healy, G.; Abdelhafez, H. Predicting Cervical Cancer Using Machine Learning Methods. Int. J. Adv. Comput. Sci. Appl. 2020, 11, 173–184. [Google Scholar] [CrossRef]
  25. Brawijaya, H.; Widodo, S.; Samudi. Improving the Accuracy of Neural Network Technique with Genetic Algorithm for Cervical Cancer Prediction. In Proceedings of the 2018 6th International Conference on Cyber and IT Service Management, CITSM 2018, Parapat, Indonesia, 7–9 August 2018; IEEE: New York, NY, USA, 2018. [Google Scholar] [CrossRef]
  26. Hu, Y.; Sharma, A.; Dhiman, G.; Shabaz, M. The Identification Nanoparticle Sensor Using Back Propagation Neural Network Optimized by Genetic Algorithm. J. Sens. 2021, 2021, 7548329. [Google Scholar] [CrossRef]
  27. Twomey, J.M.; Smith, A.E.; Member, S. Bias and Variance of Validation Methods for Function Approximation Neural Networks Under Conditions of Sparse Data. IEEE Trans. Syst. Man Cybern. Part C 1998, 28, 417–430. [Google Scholar]
  28. Nguyen, B.P.; Tay, W.L.; Chui, C.K. Robust Biometric Recognition from Palm Depth Images for Gloved Hands. IEEE Trans. Hum. Mach. Syst. 2015, 45, 799–804. [Google Scholar] [CrossRef]
  29. Fekri-Ershad, S.; Ramakrishnan, S. Cervical Cancer Diagnosis Based on Modified Uniform Local Ternary Patterns and Feed Forward Multilayer Network Optimized by Genetic Algorithm. Comput. Biol. Med. 2022, 144, 105392. [Google Scholar] [CrossRef]
  30. Tanimu, J.J.; Hamada, M.; Hassan, M.; Kakudi, H.A.; Abiodun, J.O. A Machine Learning Method for Classification of Cervical Cancer. Electronics 2022, 11, 463. [Google Scholar] [CrossRef]
  31. Sharma, M. Cervical Cancer Prognosis Using Genetic Algorithm and Adaptive Boosting Approach. Health Technol. 2019, 9, 877–886. [Google Scholar] [CrossRef]
  32. Chankong, T.; Theera-Umpon, N.; Auephanwiriyakul, S. Automatic Cervical Cell Segmentation and Classification in Pap Smears. Comput. Methods Programs Biomed. 2014, 113, 539–556. [Google Scholar] [CrossRef]
  33. Kruczkowski, M.; Drabik-Kruczkowska, A.; Marciniak, A.; Tarczewska, M.; Kosowska, M.; Szczerska, M. Predictions of Cervical Cancer Identification by Photonic Method Combined with Machine Learning. Sci. Rep. 2022, 12, 3762. [Google Scholar] [CrossRef]
  34. Remeseiro, B.; Bolon-Canedo, V. A Review of Feature Selection Methods in Medical Applications. Comput. Biol. Med. 2019, 112, 103375. [Google Scholar] [CrossRef]
  35. Kursa, M.B.; Rudnicki, W.R. Feature Selection with the Boruta Package. J. Stat. Softw. 2010, 36, i02. [Google Scholar] [CrossRef] [Green Version]
  36. Hasan, M.A.M.; Nasser, M.; Ahmad, S.; Molla, K.I. Feature Selection for Intrusion Detection Using Random Forest. J. Inf. Secur. 2016, 7, 129–140. [Google Scholar] [CrossRef] [Green Version]
  37. Bommert, A.; Sun, X.; Bischl, B.; Rahnenführer, J.; Lang, M. Benchmark for Filter Methods for Feature Selection in High-Dimensional Classification Data. Comput. Stat. Data Anal. 2020, 143, 106839. [Google Scholar] [CrossRef]
  38. Hasan, B.M.S.; Abdulazeez, A.M. A Review of Principal Component Analysis Algorithm for Dimensionality Reduction. J. Soft. Comput. Data. Min. 2021, 2, 20–30. [Google Scholar]
  39. Ballabio, D.; Grisoni, F.; Todeschini, R. Multivariate Comparison of Classification Performance Measures. Chemom. Intell. Lab. Syst. 2018, 174, 33–44. [Google Scholar] [CrossRef]
  40. Gu, Q.; Li, Z.; Zhihua, C. Evaluation Measures of the Classification Performance of Imbalanced Data Sets. In Proceedings of the International Symposium on Intelligence Computation and Applications, Huangshi, China, 23–25 October 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 461–471. [Google Scholar]
Figure 1. MLP architecture.
Figure 1. MLP architecture.
Healthcare 10 02002 g001
Figure 2. The framework of the proposed approach.
Figure 2. The framework of the proposed approach.
Healthcare 10 02002 g002
Figure 3. Comparison of the proposed method using four feature selection methods to diagnose Hinselmann.
Figure 3. Comparison of the proposed method using four feature selection methods to diagnose Hinselmann.
Healthcare 10 02002 g003
Figure 4. Comparison of the proposed method using four feature selection methods to diagnose Schiller.
Figure 4. Comparison of the proposed method using four feature selection methods to diagnose Schiller.
Healthcare 10 02002 g004
Figure 5. Comparison of the proposed method using four feature selection methods to diagnose Cytology.
Figure 5. Comparison of the proposed method using four feature selection methods to diagnose Cytology.
Healthcare 10 02002 g005
Figure 6. Comparison of the proposed method using four feature selection methods to diagnose Biopsy.
Figure 6. Comparison of the proposed method using four feature selection methods to diagnose Biopsy.
Healthcare 10 02002 g006
Figure 7. Comparison of classification accuracy of PCA–GA–MLP with the best methods in the literature in diagnosing cervical cancer using the benchmark dataset from UCI.
Figure 7. Comparison of classification accuracy of PCA–GA–MLP with the best methods in the literature in diagnosing cervical cancer using the benchmark dataset from UCI.
Healthcare 10 02002 g007
Table 1. Risk factors of cervical cancer.
Table 1. Risk factors of cervical cancer.
Factor NumberFactor NameFactor NumberFactor Name
1Age17STDs: vulvo-perineal condylomatosis
2Number of sexual partners18STDs: syphilis
3First sexual intercourse (age)19STDs: pelvic inflammatory disease
4Number of pregnancies20STDs: genital herpes
5Smokes21STDs: molluscum contagiosum
6Smokes (years)22STDs: AIDS
7Smokes (packs/year)23STDs: HIV
8Hormonal contraceptives24STDs: Hepatitis B
9Hormonal contraceptives (years)25STDs: HPV
10Intrauterine Device (IUD)26STDs: Number of diagnoses
11IUD (years)27STDs: Time since first diagnosis
12Sexually Transmitted Diseases (STDs)28STDs: Time since last diagnosis
13STDs (number)29Dx: Cancer
14STDs: condylomatosis30Dx: CIN
15STDs: cervical condylomatosis31Dx: HPV
16STDs: vaginal condylomatosis32Diagnosis (Dx)
Table 2. Top 18 factors for four target variables based on RF relative importance sequence.
Table 2. Top 18 factors for four target variables based on RF relative importance sequence.
HinselmannSchillerCytologyBiopsy
1111
9939
3393
2424
4242
3111712
67117
78611
23172332
116318
1731326
12122617
85818
2923510
1429145
5161829
26261230
16102921
Table 3. Top 10 factors for four target variables based on RF relative importance sequence.
Table 3. Top 10 factors for four target variables based on RF relative importance sequence.
HinselmannSchillerCytologyBiopsy
1111
9939
3393
2424
4242
3111712
67117
78611
23172332
116318
Table 4. Confusion matrix.
Table 4. Confusion matrix.
Predicted Class
Non-Cervical CancerCervical Cancer
Actual ClassNon-Cervical CancerTNFP
Cervical CancerFNTP
Table 5. Comparison of the proposed method using four feature selection methods with nine other classification algorithms in diagnosing Hinselmann.
Table 5. Comparison of the proposed method using four feature selection methods with nine other classification algorithms in diagnosing Hinselmann.
Feature Selection:
PCA/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
PCA–GA–MLP98.20100.0096.3796.5498.24
PCA–RF92.1698.6785.5787.2292.58
PCA–LDA61.0552.1769.9363.2457.05
PCA–SVM75.2384.8165.5271.0777.28
PCA–LR64.1160.6567.5664.9562.67
PCA–NB60.34100.0020.5855.7571.59
PCA–KNN85.81100.0071.6777.9387.57
PCA–DT83.3995.6971.3176.9885.21
PCA–AdaBoost94.67100.0089.2990.4194.96
PCA–CD-KNN84.4099.5569.2976.5386.49
Feature Selection:
All 30 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP97.57100.0095.1595.3697.62
RF85.2786.0984.4784.5985.09
LDA72.3376.0168.6570.8373.28
SVM75.3987.4263.3670.5878.06
LR72.3375.1469.4371.0772.96
NB51.57100.003.1150.7967.35
KNN85.58100.0071.1477.7387.42
DT71.4095.7147.2464.6977.06
AdaBoost87.0793.5880.4782.6987.74
CD-KNN85.6699.3971.9078.0787.42
Feature Selection:
Top 18 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP97.02100.0094.0894.3697.09
RF85.1188.0482.2383.1785.32
LDA70.9273.2468.4669.9171.48
SVM71.7184.7558.6367.2774.95
LR71.0073.2468.6370.0371.53
NB55.25100.0010.3952.7669.07
KNN85.42100.0070.8277.5087.29
DT71.3295.7147.0964.6277.01
AdaBoost87.0793.5880.4782.6987.74
CD-KNN84.4099.5569.2976.5386.49
Feature Selection:
Top 10 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP96.7999.2194.4094.6596.84
RF83.3995.6871.2477.1585.18
LDA66.9368.1365.5866.5067.28
SVM67.7968.2367.4967.6267.68
LR66.4668.1564.6665.9766.95
NB61.9850.8573.0164.5855.01
KNN84.40100.0068.7976.2486.50
DT70.6197.2544.0463.6276.83
AdaBoost85.7492.1379.1581.5886.43
CD-KNN83.8699.3968.4075.8686.01
Table 6. Comparison of the proposed method with the best results in the literature in diagnosing Hinselmann.
Table 6. Comparison of the proposed method with the best results in the literature in diagnosing Hinselmann.
StudyValidationMethodAccuracySensitivitySpecificityPrecisionF1-Score
Proposed Method5-Fold Cross-ValidationPCA–GA–MLP98.20100.0096.3796.5498.24
[4]5-fold cross-validationSVM
SVM–RFE
SVM–PCA
93.79100.0089.96N/AN/A
[15]70% training and 30% testingDeep Learning (stacked autoencoder and softmax)96.70N/AN/AN/AN/A
[24]10-fold cross-validationPCA, SMOTE with voting on
LR, RF, DT
96.7396.5097.69NA96.85
[22]10-fold cross-validationSMOTE–RF
SMOTE–RF–RFE
SMOTE–RF–PCA
97.6096.6598.54N/AN/A
[23]58% training and 42% testingSMOTE with RF, SVM, and XGBoost97.3995.1499.49N/A96.50
[5]5-fold cross-validationACO and Ant-Miner data classification rule95.45N/AN/AN/AN/A
Table 7. Comparison of the proposed method using four feature selection methods with nine other classification algorithms in diagnosing Schiller.
Table 7. Comparison of the proposed method using four feature selection methods with nine other classification algorithms in diagnosing Schiller.
Feature Selection:
PCA/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
PCA–GA–MLP96.78100.0093.6193.9796.87
PCA–RF86.0381.9990.0589.2285.42
PCA–LDA65.0452.3877.6169.9259.74
PCA–SVM70.5054.1886.7480.1164.25
PCA–LR63.8851.6276.1368.1758.70
PCA–NB51.3295.357.3250.7366.20
PCA–KNN76.6994.0059.3169.9680.08
PCA–DT75.0468.3181.6580.6872.60
PCA–AdaBoost86.5390.2782.8483.9886.98
PCA–CD-KNN87.5299.1575.8880.5088.83
Feature Selection:
All 30 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP94.1399.6788.6889.8294.45
RF71.0754.3787.6881.6565.11
LDA64.7151.0878.2570.1358.98
SVM69.7550.2089.1682.5962.16
LR63.8051.0876.4568.3258.36
NB51.16100.002.3350.5967.18
KNN75.3793.3257.3468.6979.08
DT67.6955.1380.4975.4562.17
AdaBoost72.6468.4276.8274.5871.29
CD-KNN85.2995.8674.7079.1986.72
Feature Selection:
Top 18 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP94.3098.9789.5690.5094.53
RF71.4054.1888.4882.9265.35
LDA64.7951.0878.4470.1959.02
SVM68.6051.3385.7078.3461.79
LR63.4750.9375.9467.7958.06
NB50.4197.033.8450.2266.17
KNN75.3793.3257.3368.7179.08
DT67.6055.1380.3275.2762.10
AdaBoost85.7492.1379.1581.5886.43
CD-KNN85.7096.6874.6979.3287.13
Feature Selection:
Top 10 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP93.3999.1787.5388.9893.79
RF73.0656.2989.8584.6767.49
LDA62.8152.2773.3966.0258.25
SVM65.8745.6886.0876.7856.96
LR62.4052.5872.2165.3358.20
NB61.9045.3278.4467.6854.18
KNN75.2993.9856.5068.4179.14
DT69.0163.7274.5272.0366.89
AdaBoost71.7469.6273.8672.8071.12
CD-KNN86.2896.6875.8380.0987.60
Table 8. Comparison of the proposed method with the best results in the literature in diagnosing Schiller.
Table 8. Comparison of the proposed method with the best results in the literature in diagnosing Schiller.
StudyValidationMethodAccuracySensitivitySpecificityPrecisionF1-Score
Proposed Method5-Fold Cross-ValidationPCA–GA–MLP96.78100.0093.6193.9796.87
[4]5-fold cross-validationSVM
SVM–RFE
SVM–PCA
90.8198.9984.63N/AN/A
[15]70% training and 30% testingDeep Learning (stacked autoencoder and softmax)97.90N/AN/AN/AN/A
[24]10-fold cross-validationPCA, SMOTE with voting on
LR, RF, DT
98.4998.6098.60N/A98.37
[22]10-fold cross-validationSMOTE–RF
SMOTE–RF–RFE
SMOTE–RF–PCA
95.0193.2496.68N/AN/A
[23]58% training and 42% testingSMOTE with RF, SVM, and XGBoost95.5993.9297.25N/A96.00
[5]5-fold cross-validationACO and Ant-Miner data classification rule90.56N/AN/AN/AN/A
Table 9. Comparison of the proposed method using four feature selection methods with nine other classification algorithms in diagnosing Cytology.
Table 9. Comparison of the proposed method using four feature selection methods with nine other classification algorithms in diagnosing Cytology.
Feature Selection:
PCA/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
PCA–GA–MLP97.5496.7895.1595.2797.56
PCA–RF91.5895.3088.0688.7391.83
PCA–LDA62.0447.5676.8967.6955.24
PCA–SVM71.0164.8677.0173.8369.01
PCA–LR61.8851.0073.2165.9356.81
PCA–NB53.5394.9312.1851.9967.10
PCA–KNN83.56100.0067.0875.3385.87
PCA–DT80.5492.7768.4174.4682.55
PCA–AdaBoost89.8396.2583.5085.3690.41
PCA–CD-KNN87.0699.7074.4579.7088.51
Feature Selection:
All 30 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP95.39100.0092.0292.3895.61
RF74.2766.3282.6179.2071.96
LDA63.6256.0771.1066.0660.47
SVM65.0560.6169.0668.3162.72
LR63.3158.3068.2864.7461.22
NB52.90100.005.7451.5267.94
KNN83.24100.0066.4374.8985.61
DT70.6258.5082.8077.2966.51
AdaBoost77.0584.3370.0773.6578.45
CD-KNN86.10100.0072.2978.3387.77
Feature Selection:
Top 18 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP94.92100.0089.9290.7595.12
RF74.9868.6781.5578.7773.25
LDA63.4655.6471.3366.1760.21
SVM66.0964.4067.2066.4565.25
LR63.1557.2769.0864.9960.68
NB53.5397.379.5751.8867.64
KNN83.16100.0066.2574.7985.55
DT70.5458.5082.6377.1766.46
AdaBoost76.7384.6169.1673.1878.31
CD-KNN85.94100.0071.8878.1187.65
Feature Selection:
Top 10 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP94.04100.0088.0089.3693.32
RF80.0682.0378.1278.7880.30
LDA60.6155.3166.6862.6358.14
SVM68.7972.7965.0867.5869.95
LR62.5962.0663.8163.3562.30
NB55.7677.6334.2454.3763.71
KNN82.76100.0065.5774.3485.24
DT65.8552.9879.0173.6060.06
AdaBoost74.8280.9768.8572.1276.13
CD-KNN86.02100.0071.9978.1287.69
Table 10. Comparison of the proposed method with the best results in the literature in diagnosing Cytology.
Table 10. Comparison of the proposed method with the best results in the literature in diagnosing Cytology.
StudyValidationMethodAccuracySensitivitySpecificityPrecisionF1-Score
Proposed Method5-Fold Cross-ValidationPCA–GA–MLP97.5496.7895.1595.2797.56
[4]5-fold cross-validationSVM
SVM–RFE
SVM–PCA
92.75100.0087.92N/AN/A
[15]70% training and 30% testingDeep Learning (stacked autoencoder and softmax)97.50N/AN/AN/AN/A
[24]10-fold cross-validationPCA, SMOTE with voting on
LR, RF, DT
92.8993.1294.59N/A93.35
[22]10-fold cross-validationSMOTE–RF
SMOTE–RF–RFE
SMOTE–RF–PCA
96.9494.8299.01N/AN/A
[23]58% training and 42% testingSMOTE with RF, SVM, and XGBoost96.5694.3298.51N/A97.00
[5]5-fold cross-validationACO and Ant-Miner data classification rule94.64N/AN/AN/AN/A
Table 11. Comparison of the proposed method using four feature selection methods with nine other classification algorithms in diagnosing Biopsy.
Table 11. Comparison of the proposed method using four feature selection methods with nine other classification algorithms in diagnosing Biopsy.
Feature Selection:
PCA/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
PCA–GA–MLP97.7596.7895.5495.6397.76
PCA–RF89.0087.5990.2790.0288.77
PCA–LDA69.4256.4082.7276.4864.69
PCA–SVM74.2360.6088.2083.5670.07
PCA–LR67.8255.7380.2673.8463.20
PCA–NB54.3492.3116.2852.4966.81
PCA–KNN85.07100.0070.1176.8786.89
PCA–DT84.7598.5071.1077.3386.51
PCA–AdaBoost90.9394.2487.9088.4791.19
PCA–CD-KNN88.60100.0077.2281.3189.66
Feature Selection:
All 30 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP96.39100.0092.9093.1996.45
RF76.0061.3190.9787.0471.73
LDA69.6658.4181.2475.6665.60
SVM74.8059.8590.0185.4770.20
LR69.9859.4180.8675.6066.27
NB52.25100.004.4451.1467.56
KNN84.51100.0069.0976.2586.46
DT73.1157.5189.3584.8468.05
AdaBoost80.0978.1682.2481.1979.46
CD-KNN89.49100.0079.2182.5490.35
Feature Selection:
Top 18 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP96.23100.0092.4892.9296.32
RF77.6962.9992.5889.3573.74
LDA69.0257.3681.0775.0064.72
SVM71.1158.9783.7678.1866.95
LR68.4658.0779.1773.4964.62
NB51.6199.144.1750.8567.09
KNN83.87100.0067.7775.5285.99
DT72.8757.5188.8484.1867.85
AdaBoost81.6281.0582.1081.9081.39
CD-KNN88.3698.2578.7081.9289.27
Feature Selection:
Top 10 Factors/Method
Performance Measures
AccuracySensitivitySpecificityPrecisionF1-Score
GA–MLP96.23100.0092.5992.8696.28
RF79.8567.3092.5789.9776.83
LDA68.3857.9878.9573.1364.54
SVM72.3958.1286.7881.3267.63
LR68.2258.1578.4772.7764.49
NB67.5753.6281.7474.4262.16
KNN83.95100.0067.9475.6686.07
DT73.1161.8183.4581.3769.48
AdaBoost79.9380.1879.8279.9479.93
CD-KNN89.97100.0079.9583.1790.78
Table 12. Comparison of the proposed method with the best results in the literature in diagnosing Biopsy.
Table 12. Comparison of the proposed method with the best results in the literature in diagnosing Biopsy.
StudyValidationMethodAccuracySensitivitySpecificityPrecisionF1-Score
Proposed method5-fold cross-validationPCA–GA–MLP97.7596.7895.5495.6397.76
[4]5-fold cross-validationSVM
SVM–RFE
SVM–PCA
94.13100.0090.21N/AN/A
[15]70% training and 30% testingDeep Learning (stacked autoencoder and softmax)96.60N/AN/AN/AN/A
[24]10-fold cross-validationPCA, SMOTE with voting on
LR, RF, DT
97.4497.7998.01N/A97.44
[22]10-fold cross-validationSMOTE–RF
SMOTE–RF–RFE
SMOTE–RF–PCA
96.0694.5597.51N/AN/A
[23]58% training and 42% testingSMOTE with RF, SVM, and XGBoost97.0695.5098.85N/A97.00
[5]5-fold cross-validationACO and Ant-Miner data classification rule94.76N/AN/AN/AN/A
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Dweekat, O.Y.; Lam, S.S. Cervical Cancer Diagnosis Using an Integrated System of Principal Component Analysis, Genetic Algorithm, and Multilayer Perceptron. Healthcare 2022, 10, 2002. https://doi.org/10.3390/healthcare10102002

AMA Style

Dweekat OY, Lam SS. Cervical Cancer Diagnosis Using an Integrated System of Principal Component Analysis, Genetic Algorithm, and Multilayer Perceptron. Healthcare. 2022; 10(10):2002. https://doi.org/10.3390/healthcare10102002

Chicago/Turabian Style

Dweekat, Odai Y., and Sarah S. Lam. 2022. "Cervical Cancer Diagnosis Using an Integrated System of Principal Component Analysis, Genetic Algorithm, and Multilayer Perceptron" Healthcare 10, no. 10: 2002. https://doi.org/10.3390/healthcare10102002

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop