Breast Cancer Diagnosis Using a Novel Parallel Support Vector Machine with Harris Hawks Optimization

: Three contributions are proposed. Firstly, a novel hybrid classiﬁer (HHO-SVM) is introduced, which is a combination between the Harris hawks optimization (HHO) and a support vector machine (SVM) is introduced. Second, the performance of the HHO-SVM is enhanced using the conventional normalization method. The ﬁnal contribution is to improve the efﬁciency of the HHO-SVM by adopting a parallel approach that employs the data distribution. The proposed models are evaluated using the Wisconsin Diagnosis Breast Cancer (WDBC) dataset. The results show that the HHO-SVM achieves a 98.24% accuracy rate with the normalization scaling technique, outperforming other related works. On the other hand, the HHO-SVM achieves a 99.47% accuracy rate with the equilibration scaling technique, which is better than other previous works. Finally, to compare the three effective scaling strategies on four CPU cores, the parallel version of the proposed model provides an acceleration of 3.97.


Introduction
Breast cancer is the most common disease in men and women of all ages, accounting for 11.7 percent of all cancer cases in 2020 [1]. It is the most common cancer in women worldwide, accounting for 24.5 percent of all new cases diagnosed in 2020. Breast cancer must be detected early in order to receive appropriate treatment and to reduce the number of fatalities caused by the disease.
Expert systems and artificial intelligence techniques can aid breast cancer detection professionals in avoiding costly mistakes. These expert systems can review medical data in less time and provide assistance to junior physicians. Breast cancer has been detected with excellent accuracy using a variety of artificial intelligence techniques. Marcano-Cedeo et al. [2] proposed the artificial metaplasticity MLP (AMMLP) method with a 99.26 percent accuracy. An RS-SVM classifier for breast cancer diagnosis was used by Chen et al. [3] and achieved 100% and 96.87% for the highest and average accuracy, respectively. Hui-Ling Chen et al. [4] obtained a 99.3% accuracy using a PSO-SVM. For the breast cancer dataset, Liu and Fu [5] presented the CS-PSO-SVM model, which merged a support vector machine (SVM), particle swarm optimization (PSO), and cuckoo search (CS) and obtained an accuracy of 91.3% versus 90% for both the PSO-SVM and GA-SVM achieve massive passing. The authors of [34] introduced and discussed parallel dynamic programming methods. Reference [35] gives a survey of numerous strategies for parallelizing algorithms. Reference [36] introduces a parallel approach to constraint-solving methods. Polap et al. [37] proposed three strategies for improving traditional procedures that reduced the solution space by using a neighborhood search. The second was to reduce the calculation time by limiting the number of possible solutions. In addition, the two procedures indicated above were combined. Metaheuristic algorithms such as ABC, FPA, BA, PSO, and MFO have been used to optimize SVMs and extreme learning machines, allowing them to readily overcome local minima and overfitting difficulties. The reader can refer to [38,39], which present the advantages and disadvantages of traditional machine learning methods such as SVMs and deep learning methods.
Three achievements are presented in this work. The first is a new hybrid classifier (HHO-SVM) that combines the Harris hawks optimization (HHO) and support vector machine (SVM) techniques. In order to increase the HHO-SVM performance, the second contribution compares three efficient scaling algorithms with the usual normalizing methodology. The final contribution is to improve the efficiency of the HHO-SVM by adopting a parallel approach that employs the data distribution. The proposed models are tested on the Wisconsin Diagnosis Breast Cancer (WDBC) dataset. The results show that the HHO-SVM achieves a 98.24% accuracy rate with the normalization scaling technique outperforming the results in [6][7][8]15,[17][18][19]. On the other hand, the HHO-SVM achieves a 99.47% accuracy rate with the equilibration scaling technique, better than the results in [6][7][8]10,11,15,[17][18][19][20]. Finally, the parallel version of the suggested model achieves a speedup of 3.97 on four CPU cores.
The sections that follow are grouped as such: Section 2 delves into SVM and HHO. Section 3 contains an explanation of the suggested model. Section 4 provides a complete study of three unique scaling methods: the equilibration, arithmetic, and geometric means. Section 5 explains the parallel version of the HHO-SVM. Section 6 has an experimental design that includes data descriptions, experimental setup, performance evaluation measures, and a comparative analysis. The experimental results and discussion are found in Section 7. Finally, Section 8 provides a conclusion as well as future work.

Preliminaries
Support vector machines (SVM) and the Harris hawks optimization (HHO) are introduced and studied in this section.

Support Vector Machine (SVM)
The goal of an SVM is to find an N-dimensional hyperplane that classifies the available data vectors with the least amount of error. An SVM employs convex quadratic programming to avoid local minima [40]. If we assume a binary classification problem and have a training dataset with a class label: (x 1 , y 1 ) . . . (x n , y n ), x i ∈ R d and y i ∈ (−1, +1) where x i is the class label and y i is the input or feature vector, the best hyperplane is as follows: such that w, x, and b indicate the weight, input vector, and bias, respectively. The letters w and b fulfill the following requirements: The goal of the SVM model training is to find the w and b that maximize the margin 1 w 2 . Nonlinearly separable problems are common. To transform the nonlinear problem to a linear one, the input space is converted into a higher-dimensional space.
Kernel functions [41] could be used to extend the data's dimensions and turn the problem into a linear one. The linear and nonlinear SVMs are depicted in Figure 1. Furthermore, kernel functions may be useful in speeding up calculations in high-dimensional space. For example, in the extended feature space, the linear kernel can be used to compute the dot product of two features. The most frequent SVM kernels are RBF and polynomial. They can be expressed as: such that the parameters γ and p are the width of the Gaussian kernel and the polynomial order, respectively. Setting proper model parameters has been demonstrated to increase the accuracy of SVM classification [42]. The adjustment of SVM parameters is a very delicate process. These parameters are C, gamma, and the SVM kernel function which finds the mapping from the nonlinear to linear problem by increasing the dimension.
The goal of the SVM model training is to find the w and b that maximize the margin ‖ ‖ .
Nonlinearly separable problems are common. To transform the nonlinear problem to a linear one, the input space is converted into a higher-dimensional space.
Kernel functions [41] could be used to extend the data's dimensions and turn the problem into a linear one. The linear and nonlinear SVMs are depicted in Figure 1. Furthermore, kernel functions may be useful in speeding up calculations in high-dimensional space. For example, in the extended feature space, the linear kernel can be used to compute the dot product of two features. The most frequent SVM kernels are RBF and polynomial. They can be expressed as: , such that the parameters and are the width of the Gaussian kernel and the polynomial order, respectively. Setting proper model parameters has been demonstrated to increase the accuracy of SVM classification [42]. The adjustment of SVM parameters is a very delicate process. These parameters are C, gamma, and the SVM kernel function which finds the mapping from the nonlinear to linear problem by increasing the dimension.

Harris Hawks Optimization (HHO)
Heidari et al. [43] developed an algorithm called HHO (Harris hawks optimization). It derives from the hunting style and cooperation of Harris's hawks. Some hawks cooperate when attacking their prey from different directions to surprise and disable it. Furthermore, to aid in the selection of different hunting strategies, it is dependent on various sceneries and kinds of prey flying. Exploring a prey, transitioning from exploration to exploitation, and exploitation are the three primary phases of the HHO. In this diagram, all phases of the HHO are depicted ( Figure 2). The following is a diagram of each phase:

Harris Hawks Optimization (HHO)
Heidari et al. [43] developed an algorithm called HHO (Harris hawks optimization). It derives from the hunting style and cooperation of Harris's hawks. Some hawks cooperate when attacking their prey from different directions to surprise and disable it. Furthermore, to aid in the selection of different hunting strategies, it is dependent on various sceneries and kinds of prey flying. Exploring a prey, transitioning from exploration to exploitation, and exploitation are the three primary phases of the HHO. In this diagram, all phases of the HHO are depicted ( Figure 2). The following is a diagram of each phase:

Exploration Phase
This phase is mathematically modeled primarily for waiting, searching, and prey detection. Harris's hawks are the alternative or best at every step. Harris's hawks' position ( + 1) can be formulated according to Equation (6):

Exploration Phase
This phase is mathematically modeled primarily for waiting, searching, and prey detection. Harris's hawks are the alternative or best at every step. Harris's hawks' position X(i + 1) can be formulated according to Equation (6): where i is the current iteration, X rabbit is the rabbit's position, X rand is a randomly chosen hawk at the current population, r j , j = 1, 2, 3, 4, q are random numbers between 0 and 1, and X m is the average position of the hawks, which can be calculated by: where the vector X j denotes the position of each hawk j, and N is the number of hawks.

Transition from Exploration to Exploitation
The HHO alternates between exploration and exploitation depending on the rabbit's escaping energy. Moreover, the rabbit's energy can be calculated using the formula below: where E indicates the rabbit's escaping energy, T denotes the maximum size of the iterations, and E 0 ∈ (−1, 1) presents the initial energy at each step.
The HHO can determine the state of a rabbit based on the direction of E 0 (the HHO enters the exploration phase in order to locate the prey when |E| ≥ 1, otherwise, during the exploitation steps, this strategy seeks to exploit the solutions' proximity).

Exploitation Phase
At this phase, hawks besiege the prey from all directions to hunt it, and this siege is hard or soft according to the remaining prey's energy. During this siege, the prey's escape depends on the chance r (it succeeds in escaping if r < 0.5). Moreover, if |E| ≥ 0.5, the HHO is besieging softly, otherwise, it is besieging hard. According to the phenomenon of prey escape and hawks-hawks' strategies in pursuit, the HHO implements four attack strategies: a soft siege, a hard siege, a soft siege with progressive rapid dives, a hard siege with progressive rapid dives. In particular, the rabbit has enough energy to escape if |E| ≥ 0.5; however, the prey's ability to escape or not depends on both values of |E| and r.
Soft Siege (r ≥ 0.5 and |E| ≥ 0. 5) This procedure can be written as: where ∆X(i) indicates the difference between the rabbit's current location and the rabbit's location vector at the i iteration, J = 2(1 − r 5 ) is the intensity of the rabbit's random jumping during the escape process, and r 5 ∈ (0, 1) is a random number.
Hard Siege (r ≥ 0.5 and |E| < 0. 5) In this strategy, current positions can be updated with the following formula: Soft Siege with Progressive Rapid Dives (|E| ≥ 0.5 and r < 0.5) As for the soft siege, hawks decide their next move with the following equation: The hawks dive according to the following rules based on the LF-based patterns: in which D indicates the dimension of problem, and S 1×D denotes a random vector. The levy flight (LF) can be calculated by Equation (15): where µ and v represent a range of random numbers between 0 and 1. As a result, Equation (16) can be used to describe the final strategy of this phase, which is to update the positions of the hawks: Hard Siege with Progressive Rapid Dives (|E| < 0.5 and r < 0.5) The hawk is always in close proximity to the prey during this step. The following is a model of the behavior: The following formulas can be used to calculate Y and Z: where The main purpose of this study was to employ new scaling approaches to scale breast cancer data, compute the SVM parameter using the HHO algorithm to efficiently classify breast tumors, and use a parallel approach to reduce the proposed model's execution time.

The Proposed HHO-SVM Classification Model
The HHO-SVM system is implemented in two stages. The HHO algorithm determines the SVM parameters automatically for the first phase. The optimized SVM algorithm diagnoses a breast tumor as benign or malignant in the second phase. To obtain the best accurate result, a ten-fold cross-validation (CV) is used. To test the SVM parameters, the HHO-SVM model applies the root-mean-square error (RMSE) as the fitness function. The following formula is used to calculate the RMSE: such that N is the number of entities in the test dataset.
In the HHO-SVM algorithm for breast cancer, the population size is set to N, and each hawk represents X i (i = 1, 2, . . . , N), the maximum number of iterations is set to T, the number of dimensions is set to dim, the upper bound is set to ub, the lower bound is set to lb, and the boundary of positions is set to X rabbit . X rabbit is the position of the rabbit, and all hawks update their positions. After that, random values are used to form the initial population (N*dim). After the data have been loaded, we use one of the scaling strategies to modify it. It uses a k-fold cross-validation and conducts several procedures for each fold to evaluate the model's efficiency. If the number of iterations does not equal T, the model repeats the steps below for each iteration.
To begin, it passes each bird through two specified functions and sets its output to the SVM (C and γ) parameters, then trains the SVM and classifies the test set. Then, it calculates the fitness function (RMSE) from Equation (21), updates X rabbit , according to the smallest fitness value, and update the initial energy E 0 , jump strength J, and the position of the current hawk according to the X rabbit , E 0 , J, E, and r values, where r is a random value and E is the energy. Then, the algorithm checks whether ( |E| ≥ 1); if it is, then it enters the exploration phase and updates the location vector using Equation (6); if ( |E| < 1), then it enters the exploitation phase, which may be a soft siege, hard siege, soft siege with progressive rapid dives or a hard siege with progressive rapid dives.
Therefore, the algorithm checks whether ( |E| ≥ 0.5 and r ≥ 0.5); if true, then it is a soft siege, and the location vector is updated using Equation (10). If ( |E| < 0.5 and r ≥ 0.5), then it is a hard siege, and the location vector is updated using Equation (12). If ( |E| ≥ 0.5 and r < 0.5), then it is a soft siege with progressive rapid dives and the location vector is updated using Equation (16), F(Y) and F(Z) are calculated by passing Y or Z to two particular functions, and the parameters of the SVM (C and γ) are equal to its output. Then, the algorithm trains the SVM and classifies the test set. It computes the RMSE from Equation (21) as the value of F(Y) or F(Z). If ( |E| < 0.5 and r < 0.5), then it is a hard siege with progressive rapid dives. The location vector is updated using Equation (17), F(Y) and F(Z) are calculated by passing Y or Z to two particular functions, and the parameters of the SVM (C and γ) are equal to its output; then, the algorithm trains the SVM and classifies the test set. It computes the RMSE from Equation (21) as the value of F(Y) or F(Z). Then, if the number of iterations does not surpass T, it goes back to step 4 in the process (Algorithm 1). We move on to the next fold and return to step 3 if T is satisfied. If T and the fold number k are equal, we proceed to step 5. Finally, we compute the averages of the RMSE and the accuracy of the k folds and return them. No. o f dimensions k No. o f f olds Output: Average RMSE: Average classi f ication accuracy rates
Apply one of the scaling techniques after loading the data. 3.
for (each fold j) do Divide the data into train and test subsets randomly 4.
while (t < T) do for (each hawk (X i )) do Pass X i to particular functions Set function's output to parameter of SVM (C, γ) Train and test the SVM model Evaluate the fitness X i with EQ (21) Update X rabbit as the position of the rabbit (best position based on the fitness value) end (for) for (every hawk (X i )) do Update E 0 and J (initial energy and jump strength) Update the E by EQ (8) if (|E| ≥ 1) then Exploration phase Update the position vector by EQ (6) if (|E| < 1) then Exploration phase if (r ≥ 0.5 and |E| ≥ 0.5) then Soft siege Update the position vector by EQ (10) else if (r ≥ 0.5 and |E| < 0.5) then Hard siege Update the position vector by EQ (12) else if (r < 0.5 and |E| ≥ 0.5) then Soft siege with PRD Update the position vector by EQ (16) F(Y), F(Z) and F(X i ) calculated by using RMSE else if (r < 0.5 and |E| < 0.5) then Hard siege with PRD Update the position vector by EQ (17) end (for) t=t+1 end (while) t=0 end (for)

5.
Return averages of RMSE and classification accuracy for all folds

Scaling Techniques
Before introducing the scaling techniques, some of the necessary mathematical symbols should be presented. We treat the breast cancer data as a matrix and present some mathematical symbols as shown in Table 1. The final scaled matrix is denoted by RAS, where R = diag (r 1 , . . . , r m ) and S = diag (s 1 , . . . , s n ).

Term Meaning
A a ij m × n matrix (with m entities and n attributes) The scaling factor of row i s j The scaling factor of row j The scaled matrix by row R scaling factor The scaled matrix in its final form.
All of the scaling approaches presented in this section scale the rows first, then the columns. Equations (22) and (23) show the steps for scaling the matrix.
(1) Arithmetic mean: The variance between nonzero entries in the coefficient matrix A is reduced using the arithmetic mean scaling technique. As shown in Equation (24), the rows are scaled by dividing each row by the mean of the absolute value of the nonzero values: Each column (attribute) is scaled by dividing the modulus value of the nonzero items in that column by the mean of the modulus of the nonzero entries in that column as shown in Equation (25): (2) Equilibration scaling technique: This scaling method's cornerstone is the largest value in absolute value. The row scaling is done by dividing every row (instance) of matrix A by the absolute value of the row's largest value. Then, we divide every column of the matrix by the absolute value of the largest value in that column, which is scaled by the row factor. The final scaled matrix A has a range of [−1, 1].
(3) Geometric mean: To begin, Equation (26) depicts the scaling of the rows, in which every row is split by the geometric mean of the nonzero elements in that row.
Second, Equation (27) represents the column scaling where every column is divided by the geometric mean of the modulus of the nonzero elements in that column.
(4) Normalization [−1, 1]: Equation (28) represents the normalization within the range [−1, 1] where a, a , max k , and min k are the initial value, the scaled value, the maximum value, and the minimum value of feature k, respectively.

The Parallel Metaheuristic Algorithm
We implemented a parallel metaheuristic algorithm based on the population, where the population is divided into different parts that are easy to exchange, that evolve separately, and that are then later combined. In this paper, the parallel approach was implemented by dividing the population into several sets on different cores. The number of cores, Nc, was identified. The starting population consisted of n particles randomly initialized. The group size was calculated as follows: The proposed model steps are shown in Algorithm 2. The ceil function was used to obtain an integer number of particles to be distributed on the cores. The basic algorithm steps were executed for all sets in a standalone thread. Nc best particles were chosen as a solution for the optimization problem when these phases were completed. Moreover, these particles were combined to obtain the best particles in general on all cores and update the position according to them.

Experimental Design
This part contains a description of the data, a performance evaluation measure, as well as a comparative study.

Data Description
The proposed model was tested on the Wisconsin diagnostic Breast Cancer (WDBC) dataset, which is available at the University of California, Irvine Machine Learning Repository [44]. There are 569 examples in the dataset, which are separated into two groups (malignant and benign). There are 357 cases of malignant tumors and 212 cases of benign tumors, respectively. Each database record has thirty-two attributes. Table 2 lists the thirty-two qualities. Fractal dimension ("coastline approximation"-1)

Symmetry
In both directions, the length difference between lines perpendicular to the major axis and the cell boundary.

Experimental Setup
MATLAB was used to create the suggested HHO-SVM detection method. Chang and Lin [45] created the SVM method, and their implementation was improved. The computing environment for the experiment is described in Table 3. The k-fold CV was proposed by Salzberg [46], and it was used to ensure that the results were genuine. k = 10 in this study. The following are the HHO-SVM's detailed settings: 1000, 19, 25, and 10 were the values for the iterations, search agents, dimensions, and k-fold, respectively. The [lb, ub] lower and upper bounds were set to [−5, 5].

Performance Metrics
Six metrics, sensitivity, specificity, accuracy, precision, G-mean, and F-score, were used to assess the efficacy of the suggested HHO-SVM model. These metrics are defined as follows according to the confusion matrix: If the dataset has two classes ("M" for malignant and "B" for benign), then the true positives (TP) are the total number of cases with classification result "M" when they are actually "M" in the dataset; the true negatives (TN) are the total number of cases with classification result "B" when they are actually "B" in the dataset; the false positives (FP) are the total number of cases with classification result "M" when they are "B" in the dataset; and the false negatives (FN) are the total number of cases with classification result "B" when they are "M" in the dataset.

Comparative Study
In this study, the efficiency of the presented HHO-SVM algorithm was compared to the SVM algorithm with the grid search technique. Figure 3 shows how the SVM algorithm works with the grid search technique If the dataset has two classes ("M" for malignant and "B" for benign), then the true positives (TP) are the total number of cases with classification result "M" when they are actually "M" in the dataset; the true negatives (TN) are the total number of cases with classification result "B" when they are actually "B" in the dataset; the false positives (FP) are the total number of cases with classification result "M" when they are "B" in the dataset; and the false negatives (FN) are the total number of cases with classification result "B" when they are "M" in the dataset.

Comparative Study
In this study, the efficiency of the presented HHO-SVM algorithm was compared to the SVM algorithm with the grid search technique. Figure 3 shows how the SVM algorithm works with the grid search technique

Empirical Results and Discussion
In this study, the abbreviations S0, S1, S2, S3, and S4 are used to denote no scaling, a normalization in [−1, 1], the arithmetic mean, the geometric mean, and the equilibration scaling techniques, respectively. Experiments on the WBCD dataset were used to assess the efficacy of the proposed HHO-SVM model for breast cancer against the SVM algorithm with a grid search technique. First and foremost, our findings show the value of the grid search methodologies, the usefulness of the HHO-SVM model that was developed sequentially, and the superiority of the most recent scaling strategies over the previous normalizing methodology. Finally, the results show that the parallel version of the proposed model achieves a speedup of 3.97 for four cores.
Tables 4-6 demonstrate a comparison of the SVM classification accuracies using the grid search algorithm with S0, S1, S2, S3, and S4. Tables 4 and 5 show that the average accuracy rates obtained by the SVM using S3 (98.59%) are higher than those produced by the SVM using S1 (96.66%) (98.59%). On the other hand, the S4 technique outperforms all other scaling techniques with an accuracy of 98.95% compared to that obtained by the SVM.

Empirical Results and Discussion
In this study, the abbreviations S0, S1, S2, S3, and S4 are used to denote no scaling, a normalization in [−1, 1], the arithmetic mean, the geometric mean, and the equilibration scaling techniques, respectively. Experiments on the WBCD dataset were used to assess the efficacy of the proposed HHO-SVM model for breast cancer against the SVM algorithm with a grid search technique. First and foremost, our findings show the value of the grid search methodologies, the usefulness of the HHO-SVM model that was developed sequentially, and the superiority of the most recent scaling strategies over the previous normalizing methodology. Finally, the results show that the parallel version of the proposed model achieves a speedup of 3.97 for four cores.
Tables 4-6 demonstrate a comparison of the SVM classification accuracies using the grid search algorithm with S0, S1, S2, S3, and S4. Tables 4 and 5 show that the average accuracy rates obtained by the SVM using S3 (98.59%) are higher than those produced by the SVM using S1 (96.66%) (98.59%). On the other hand, the S4 technique outperforms all other scaling techniques with an accuracy of 98.95% compared to that obtained by the SVM.  Tables 7-11 and Figure 4 show the importance of the data scaling techniques in improving classification accuracy, with the average classification accuracy rate without scaling the data (89.11%) being lower than the average classification accuracy rate when using any other scaling technique, and when comparing the normalization and other scaling techniques, we found that the novel scaling techniques outperformed the normalization in terms of both accuracy rates and CPU time. It is obvious that the HHO-SVM with the arithmetic mean scaling approach (98.25) achieved higher average accuracy rates than the HHO-SVM with normalization and the scaling strategy in the range [−1, 1] (98.24%). With an accuracy of 99.47 percent, the equilibration scaling technique outperforms all the other scaling strategies, including the HHO-SVM. Table 7. Grid-SVM accuracy with S0, S1, S2, S3, and S4.

No
Symbol Accuracy CPU Time      Tables 7-11 and Figure 4 show the importance of the data scaling techniques in im proving classification accuracy, with the average classification accuracy rate withou scaling the data (89.11%) being lower than the average classification accuracy rate whe using any other scaling technique, and when comparing the normalization and othe scaling techniques, we found that the novel scaling techniques outperformed the no malization in terms of both accuracy rates and CPU time. It is obvious that th HHO-SVM with the arithmetic mean scaling approach (98.25) achieved higher averag accuracy rates than the HHO-SVM with normalization and the scaling strategy in th range [−1, 1] (98.24%). With an accuracy of 99.47 percent, the equilibration scaling tech nique outperforms all the other scaling strategies, including the HHO-SVM.    Tables 8-17 show the importance of the data scaling techniques in improving the classification accuracy, with the average classification accuracy rate without scaling the data (89.11 percent) being lower than the average classification accuracy rate when using any other scaling technique, and when comparing the normalization and other scaling techniques, we found that the novel scaling techniques outperformed the normalization in terms of both accuracy rates and CPU time. The HHO-SVM with the arithmetic mean scaling approach (98.25) clearly outperformed the HHO-SVM with the normalization scaling strategy in the range [−1, 1] (98.24%). With an accuracy of 99.47%, S4 outperformed all other scaling procedures.       The results of all scaling strategies obtained by the HHO-SVM in terms of accuracies and CPU times are summarized in Table 18 and Figures 5 and 6. In terms of accuracy and CPU time, the equilibration scaling technique clearly outperformed all other scaling techniques. In terms of precision, however, the equilibration scaling technique was the least accurate. According to CPU time, the normalization scaling in the range [−1, 1] was the greatest.   The accuracy rate of the proposed HHO-SVM model was compared to that of the conventional SVM employing a grid search technique in Table 19. For the scaling procedures S4, S2, S3, and S1, the accuracy rates of the proposed HHO-SVM model were 99.47, 98.25, 98.24, and 98.24, respectively. For the scaling approaches S4 and S1, the accuracy rates of the classic SVM with a grid search algorithm were 98.95 and 96.49, respectively.  The accuracy rate of the proposed HHO-SVM model was compared to that of the conventional SVM employing a grid search technique in Table 19. For the scaling procedures S4, S2, S3, and S1, the accuracy rates of the proposed HHO-SVM model were 99.47, 98.25, 98.24, and 98.24, respectively. For the scaling approaches S4 and S1, the accuracy rates of the classic SVM with a grid search algorithm were 98.95 and 96.49, respectively. The parallel version of the HHO-SVM algorithm was provided to reduce its running time. CPU timings for all scaling strategies produced by the HHO-SVM on different cores are shown in Table 20 and Figure 7.  In addition, Table 21 and Figure 8 show the speedup obtained by the HHO-SVM for all scaling strategies. In addition, Table 21 and Figure 8 show the speedup obtained by the HHO-SVM for all scaling strategies.  In addition, Table 21 and Figure 8 show the speedup obtained by the HHO-SVM for all scaling strategies.  For the scaling techniques S4, S3, S2, and S1, the speedups for four cores were 3.57, 3.74, 3.66, and 3.97, respectively. Table 22 shows that the performance of the presented HHO-SVM model against other related models developed in the literature, demonstrating the usefulness of our method. Table 22 shows that the classification accuracy of our created HHO-SVM diagnostic system is equivalent to or better than that of existing classifiers on the WBCD database.  Table 22 shows that the performance of the presented HHO-SVM model against other related models developed in the literature, demonstrating the usefulness of our method. Table 22 shows that the classification accuracy of our created HHO-SVM diagnostic system is equivalent to or better than that of existing classifiers on the WBCD database. Table 22. A comparison between related works against to our model.

Study
Year Method Accuracy (%) The final contribution was to improve the efficiency of the HHO-SVM by adopting a parallel approach that employed the data distribution. On the Wisconsin Diagnosis Breast Cancer (WDBC) dataset, the proposed models were tested. The results showed that the HHO-SVM achieved a 98.24% accuracy rate with the normalization scaling technique, thus outperforming the results in [6][7][8]11,15,[17][18][19]. On the other hand, the HHO-SVM achieved a 99.47% accuracy rate with the equilibration scaling technique, outperforming the results in [6][7][8]10,11,15,[17][18][19][20]. Finally, on four CPU cores, the parallel HHO-SVM model delivered a speedup of 3.97. The proposed approach will be evaluated in various medical datasets in future research. In addition, we are attempting to incorporate various measuring techniques that will reduce the running time and improve the proposed diagnostic system's efficiency.

Data Availability Statement:
The data used to support the findings of this study are available from the corresponding author upon request.