Next Article in Journal
Usefulness of the CDC/AAP and the EFP/AAP Criteria to Detect Subclinical Atherosclerosis in Subjects with Diabetes and Severe Periodontal Disease
Previous Article in Journal
The Application and Validity of a New Composite Radiographic Index for Patients with Osteonecrosis of the Jaws
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Evaluating the Nuclear Reaction Optimization (NRO) Algorithm for Gene Selection in Cancer Classification

Department of Information Technology, College of Computer and Information Sciences, King Saud University, P.O. Box 51178, Riyadh 11543, Saudi Arabia
*
Author to whom correspondence should be addressed.
Diagnostics 2025, 15(7), 927; https://doi.org/10.3390/diagnostics15070927
Submission received: 24 February 2025 / Revised: 29 March 2025 / Accepted: 31 March 2025 / Published: 3 April 2025

Abstract

Background/Objectives: Cancer classification using microarray datasets presents a significant challenge due to their extremely high dimensionality. This complexity necessitates advanced optimization methods for effective gene selection. Methods: This study introduces and evaluates the Nuclear Reaction Optimization (NRO)—drawing inspiration from nuclear fission and fusion—for identifying informative gene subsets in six benchmark cancer microarray datasets. Employed as a standalone approach without prior dimensionality reduction, NRO was assessed using both Support Vector Machine (SVM) and k-Nearest Neighbors (k-NN). Leave-One-Out Cross-Validation (LOOCV) was used to rigorously evaluate classification accuracy and the relevance of the selected genes. Results: Experimental results show that NRO achieved high classification accuracy, particularly when used with SVM. In select datasets, it outperformed several state-of-the-art optimization algorithms. However, due to the absence of additional dimensionality reduction techniques, the number of selected genes remains relatively high. Comparative analysis with Harris Hawks Optimization (HHO), Artificial Bee Colony (ABC), Particle Swarm Optimization (PSO), and Firefly Algorithm (FFA) shows that while NRO delivers competitive performance, it does not consistently outperform all methods across datasets. Conclusions: The study concludes that NRO is a promising gene selection approach, particularly effective in certain datasets, and suggests that future work should explore hybrid models and feature reduction techniques to further enhance its accuracy and efficiency.

1. Introduction

Cancer remains a significant global health challenge, necessitating advancements in diagnostic and treatment methodologies [1]. The advent of microarray technology has revolutionized molecular oncology, enabling the simultaneous analysis of thousands of genes to identify potential biomarkers for various cancers. This technological leap has brought new opportunities for precision medicine [2]. Still, these high-dimensional datasets—often incorporating thousands of genes but only a limited number of patient samples—can lead to complications, including an elevated risk of overfitting, unwieldy computational demands, and interpretive complexity.
A commonly used strategy to address these issues is feature (gene) selection. The idea is to single out a smaller set of genes that captures the essence of the data, thereby enhancing model accuracy, reducing noise, and keeping the analysis computationally manageable [3]. Throughout this paper, the terms “feature” and “gene” refer to the same concept, as each feature in the dataset represents the expression level of a specific gene. Therefore, “feature selection” and “gene selection” are used interchangeably. A variety of methods have been developed for this task, including a growing list of bio-inspired optimization algorithms. Particle Swarm Optimization (PSO) [4] and Harris Hawks Optimization (HHO) [5] have demonstrated significant capabilities in high-dimensional spaces. These methods have been successfully applied to feature-selection tasks in cancer classification, achieving high performance in identifying optimal gene subsets.
This study is the first to explore the potential of the Nuclear Reaction Optimization (NRO) algorithm [6] for optimizing gene selection in cancer microarray data. Introduced in 2019 as a general optimization algorithm, NRO has been previously applied in gene selection for cancer classification using RNA Sequencing (RNA-Seq) data [7]. However, its effectiveness with microarray data remains unexamined. This research assessed the standalone capability of the NRO algorithm to select gene subsets for cancer classification across six benchmark microarray datasets.
Here, we focused solely on the inherent strengths of the NRO algorithm, foregoing additional dimensionality-reduction techniques. By dimensionality reduction, we refer to preprocessing methods—such as filtering, statistical ranking, or feature extraction—that reduce the number of genes before applying optimization algorithms. To provide a clear baseline for NRO’s performance, we intentionally excluded prior dimensionality reduction. This approach allowed us to evaluate NRO’s capability in handling raw, high-dimensional microarray data without external influence from filtering methods.
The selected gene subsets were subsequently evaluated using robust machine learning classifiers, such as Support Vector Machines (SVMs) and k-Nearest Neighbors (k-NNs), with a primary focus on classification accuracy. The outcomes of this study were intended to position the NRO algorithm as a formidable and reliable tool in bioinformatics feature selection, thus facilitating the development of more effective and interpretable diagnostic models.
The rest of this paper is organized as follows: Section 2 provides a detailed background on classification techniques and optimization algorithms, with a focus on NRO. Section 3 reviews related works in the domain of gene selection for cancer classification. Section 4 describes the materials and methods used in this study, including the datasets, preprocessing techniques, and the implementation of the NRO algorithm. Section 5 presents the results and analysis, comparing NRO’s performance with those of other optimization methods. Finally, Section 6 concludes the paper with key findings, limitations, and directions for future research.

2. Background

2.1. Classification Techniques in Cancer Diagnosis

In the field of bioinformatics, particularly in cancer classification, specific classifiers have proven to be especially effective due to their adaptability and performance with complex, high-dimensional datasets like those generated from microarray gene-expression profiles. Here, we explore two widely used classifiers: Support Vector Machine (SVM) and k-Nearest Neighbors (k-NNs), which have distinct characteristics making them suitable for cancer detection and classification.

2.1.1. Support Vector Machine (SVM)

The SVM is particularly effective for microarray data because of its ability to work well in high-dimensional spaces and its robustness to overfitting with small sample sizes. It identifies an optimal hyperplane to separate data classes, maximizing classification margins, which enhances generalization [8]. The SVM has been extensively used in cancer classification tasks to distinguish between cancer types or between cancerous and non-cancerous samples based on gene-expression profiles. Its capability to manage noisy, complex data makes it a strong benchmark for evaluating gene-selection methods [9]. The SVM is highly effective in binary classification problems and can be adapted for multiclass scenarios using strategies like one-vs.-all (OVA) or one-vs.-one (OVO) [10].

2.1.2. k-Nearest Neighbors (k-NNs)

The k-Nearest Neighbors (k-NNs) is a non-parametric classifier that ranks among the most straightforward and effective methods for medical diagnosis, including cancer classification. This method operates based on a simple principle: it classifies each new case by analyzing the most common class among its nearest neighbors, determined by a chosen distance metric, typically Euclidean [11]. In gene selection, the k-NNs is valuable for assessing whether the selected genes preserve meaningful data structures and class separability. Its sensitivity to local data distributions allows it to highlight the effectiveness of gene subsets in capturing disease-specific patterns in cancer datasets [12,13].

2.2. Nuclear Reaction Optimization (NRO)

The Nuclear Reaction Optimization (NRO) algorithm is a physics-inspired metaheuristic designed to solve optimization problems by mimicking the natural processes of nuclear fission and nuclear fusion [6]. These processes simulate how nuclei split or merge to release energy. A nuclide splitting into two or more smaller nuclides is known as nuclear fission, as shown is Figure 1. The opposite is nuclear fusion. A bigger nuclide is created when two or more nuclides fuse together, as shown in Figure 2. Extremely large amounts of energy could be released by nuclear fission or fusion. In this context, it translates to exploring the solution space for optimal solutions. NRO employs a dynamic balance between exploration (searching for new potential solutions) and exploitation (refining existing solutions).

2.2.1. Nuclear Fission Process

The nuclear fission process in NRO creates new candidate solutions by splitting an existing solution into smaller parts. This helps the algorithm explore different areas of the solution space and avoid becoming stuck in one region. The process uses the following equation:
X i F i G a u s s i a n ( X b e s t , σ 1 ) + ( r a n d n · X b e s t P n e s   ·   N e i ) ,         i f   r a n d     P β , G a u s s i a n ( X i , σ 2 ) + ( r a n d n   ·   X b e s t P n e e   ·   N e i ) ,                   i f   r a n d > P β ,
Here,
  • X i F i :   the new solution generated during fission.
  • X b e s t : the best solution found so far.
  • G a u s s i a n ( X , σ ) : a random value generated around X , with σ defining the spread.
  • σ 1   ,   σ 2 step sizes controlling the exploration range (see Equations (3) and (4) below).
  • r a n d n :   random number introducing variability in the solution.
  • P n e s   ,   P n e e :   mutation factors determining the scale of adjustments for subaltern and essential fission products, respectively.
  • N e i : heated neutron, calculated as N e i = X i X j , where X i   a n d   X j are two random solutions.
  • P β :   the probability governing whether subaltern or essential fission products are produced.
This approach ensures that some new solutions are close to the best one X b e s t (local search) while others are farther away (global search), providing a good mix of exploration and refinement.
Step Size Adjustment in Fission
The effectiveness of the fission process depends on the step size, which determines how far new solutions deviate from existing ones. The step sizes were calculated as follows:
σ 1 = log g g X i X best ,
σ 2 = log g g X r X best .
Here,
  • g : current generation number; the term log g g ensures that step sizes decrease as iterations progress.
  • X i X best :   distance between the current solution and the best-known solution.
  • X r X best : distance between a random solution and the best-known solution.
At first, large step sizes help explore the solution space. Over time, the step sizes shrink, allowing the algorithm to refine the best solutions and reach an optimal result.
Mutation Factors in Fission
Mutation factors, defined as follows:
P ne s = round r a n d + 1 ,
P ne e = round r a n d + 2 ,
control the magnitude of adjustments applied to fission products. These factors ensure variability and are critical for balancing exploration and exploitation:
  • r a n d :   a random number uniformly distributed between 0 and 1.
  • The integer rounding ensures discrete adjustment levels for the mutation process.
These factors are incorporated into the fission equation to refine or diversify the search space, avoiding premature convergence and ensuring effective optimization.

2.2.2. Nuclear Fusion Process

The nuclear fusion process refines solutions by combining promising candidates. It consists of two sub-phases: ionization and fusion.
Ionization Step
In this step, solutions are adjusted based on the differences between randomly selected ones. The formula is as follows:
X i , d I o n X r 1 , d F i + r a n d   ·   ( X r 2 , d F i X i , d F i ) ,     i f   r a n d     0.5 , X r 1 , d F i r a n d   ·   ( X r 2 , d F i X i , d F i )                     i f   r a n d > 0.5 ,  
Here,
  • X r 1 , d F i ,   X r 2 , d F i : components of two randomly selected fission solutions.
  • X i , d F i :   current solution.
  • r a n d : random value for diversity.
This step prevents the algorithm from converging too early by adding diversity. If the two selected solutions are too similar, Levy flight [14] is applied to make bigger changes and avoid stagnation:
  X i , d Ion = X i , d Fi + α Levy β d X i , d Fi X best , d Fi ,
Here,
  • α : a scaling factor controlling the magnitude of jumps.
  • Levy β :   heavy-tailed random step size, introducing both small and large adjustments.
  • : indicates element-wise multiplication.
  • X best , d Fi : best-known solution in the d t h   dimension.
Levy flight is specifically applied when the difference term X r 2 , d F i X i , d F i approaches zero, which could lead to stagnation in the search process. By introducing varying step sizes, this adjustment ensures that the algorithm explores new areas of the solution space, avoiding local optima.
Fusion Step
The fusion step combines the strengths of promising solutions to refine the search. It is expressed as follows:
X i Fu = X i Ion + r a n d X r 1 Ion X best + r a n d X r 2 Ion X best ,
Here,
  • X i Fu : refined solution after fusion.
  • X best :   best-known solution guiding the search.
  • X r 1 Ion ,   X r 2 Ion : ionized solutions selected for comparison.
  • r a n d : random value for diversity.
When solutions become too similar X r 1 Ion = X r 2 Ion , the algorithm may stagnate. To avoid this, Levy flight introduces random jumps, helping the algorithm explore new areas and escape local optima. This is governed by the following equation:
X i Fu = X i Ion + α Levy β X i Ion X best Ion ,
Here, Levy flight ensures that the fused solutions are not constrained to a narrow region of the search space, especially in situations in which the algorithm might otherwise fail to differentiate between nearly identical candidates.

3. Related Works

Gene selection for cancer classification is critical for addressing the challenges of high-dimensional microarray datasets. Numerous bio-inspired optimization algorithms have been proposed to enhance dimensionality reduction and improve classification accuracy, some of which apply dimensionality reduction before optimization [15] while others rely solely on the optimization algorithm without any prior filtering. This section reviews recent advancements in optimization-based gene-selection approaches, focusing on their methodologies and relevance to this study.
AlMazrua and AlShamlan [16] proposed Harris Hawks Optimization (HHO) for gene selection combined with SVM and k-NN classifiers to tackle the dimensionality of microarray datasets. Their approach integrated redundancy analysis and relevance scoring for preprocessing, followed by HHO-based feature selection. Evaluated on six datasets, the method achieved superior classification accuracy and reduced gene subsets compared with traditional algorithms, showcasing the efficiency of HHO in bioinformatics applications.
Alweshah et al. [17] introduced the Monarch Butterfly Optimization (MBO) algorithm as a wrapper-based feature-selection method, utilizing the k-nearest neighbor (k-NN) classifier to enhance classification accuracy and reduce computational complexity. Evaluated across 18 benchmark datasets, the MBO approach demonstrated superior performance compared with other metaheuristic algorithms, achieving an average classification accuracy of 93% while significantly reducing the number of selected features. This study highlights MBO’s effectiveness in balancing global and local search capabilities for feature-selection tasks.
Almugren and Alshamlan [18] proposed FF-SVM, a wrapper-based gene-selection algorithm combining the Firefly Algorithm (FFA) with a Support Vector Machine (SVM) classifier. The method aims to optimize cancer classification by identifying the most informative genes from high-dimensional microarray datasets. Using FFA for feature selection, followed by SVM classification with Leave-One-Out Cross-Validation, the algorithm achieved high classification accuracy with a minimal subset of genes across five benchmark datasets. Comparative experiments demonstrated the FF-SVM’s superior performance over several state-of-the-art methods in terms of accuracy and dimensionality reduction, highlighting its effectiveness in bioinformatics applications.
Nssibi et al. [19] introduced a hybrid optimization approach called iBABC-CGO, which combines an island-based Artificial Bee Colony (iABC) algorithm with Chaos Game Optimization (CGO) for gene selection. The method addresses the challenges of high-dimensional microarray datasets by using a binary representation to identify informative genes while maintaining classification accuracy. The hybrid algorithm leverages CGO principles to improve convergence and avoid local optima during the migration process, and its binary version, iBABC-CGO, ensures efficient exploration and exploitation. Experimental results on 15 biological datasets demonstrated the approach’s superior performance compared with state-of-the-art methods, highlighting its ability to achieve high accuracy with minimal feature subsets.
AlMazrua and AlShamlan [20] proposed a novel feature-selection approach for cancer classification using the Gray Wolf Optimizer (GWO). This bio-inspired optimization algorithm mimics the leadership hierarchy and cooperative hunting behavior of gray wolves to effectively explore and exploit high-dimensional datasets. Their framework utilized the GWO to identify the most significant features, achieving a balance between classification accuracy and dimensionality reduction. Their experimental results demonstrated that the method successfully reduced the number of selected features while maintaining high classification performance, highlighting its potential for bioinformatics and medical diagnostic applications.
Overall, the reviewed studies demonstrate the effectiveness of bio-inspired optimization algorithms in gene selection for cancer classification. Methods such as HHO, MBO, FFA, iBABC-CGO, and GWO have achieved high classification accuracy while selecting minimal and informative gene subsets. This study evaluates the Nuclear Reaction Optimization (NRO) algorithm as a novel contribution to further enhance gene selection and classification performance.

4. Materials and Methods

This study explores the application of the Nuclear Reaction Optimization (NRO) algorithm for gene selection in cancer classification. The methodology consists of dataset preprocessing, optimization using NRO, evaluation through machine learning classifiers, and fitness assessment. The implementation was carried out using Python (version 3.x), with numpy and pandas for preprocessing gene-expression data, scipy.io.arff for handling microarray datasets, and sklearn for feature scaling and classification. The NRO algorithm was implemented by coding its mathematical foundations, including nuclear fission and fusion processes, Lévy flight-based step size adjustments, and mutation mechanisms for optimal gene selection. Figure 3 shows the flowchart summarizing the entire methodology, from data preprocessing and optimization to fitness evaluation and termination.

4.1. Dataset and Preprocessing

Using six well-known binary and multiclass microarray cancer datasets that we obtained from http://www.gems-system.org/ (accessed on 15 February 2025), we assessed the overall effectiveness of the gene-selection techniques. In the discipline of bioinformatics, these datasets are frequently used to compare the effectiveness of gene-selection techniques. SRBCT [21], Lymphoma [22], and Leukemia2 [23] are multiclass microarray datasets, whereas the binary-class microarray datasets are Lung [24], Colon [25], and Leukemia1 [25,26]. We provide a thorough description of these six benchmark microarray gene-expression datasets in Table 1, including information on the number of classes, samples, and genes.
To prepare the datasets for gene selection and classification, we applied a series of preprocessing techniques commonly used in bioinformatics to ensure the datasets were clean, consistent, and suitable for machine learning algorithms. Firstly, any missing values in the datasets were handled by replacing them with the mean value of the respective gene’s expression levels using mean imputation. In our analysis, missing values were found only in the Lymphoma dataset, accounting for 4.91% of its total data, affecting 2796 genes. Given this relatively low percentage, mean imputation was applied to maintain data integrity without significantly impacting the classification performance. To standardize the features and eliminate the effects of varying scales across genes, we applied Z-score normalization, transforming the expression levels of each gene to have a mean of 0 and a standard deviation of 1. Additionally, for multiclass datasets, we employed label encoding to convert categorical class labels into a numerical format, making them compatible with classification algorithms. These preprocessing steps ensured that the datasets were optimally prepared for the subsequent feature-selection and classification phases.

4.2. Apply Nuclear Reaction Optimization (NRO) Algorithm

The Nuclear Reaction Optimization (NRO) algorithm was adapted in this study specifically for gene selection in cancer classification using microarray data. Microarray datasets are characterized by their high dimensionality, with thousands of genes but relatively few samples, posing challenges for classification and optimization algorithms. NRO addresses these challenges by iteratively refining subsets of genes, balancing the need for compact feature sets with high classification accuracy.
Each candidate solution in the NRO algorithm is a binary vector, where 1 indicates that a gene is selected, and 0 indicates exclusion. The algorithm initialized a population of 500 such solutions, each representing a potential subset of genes. We chose a population size of 500 to ensure a diverse set of candidate solutions, which helps in effectively exploring the high-dimensional gene space while maintaining computational efficiency. A smaller population might limit diversity and lead to premature convergence, whereas a significantly larger one would increase computational cost without substantial accuracy gains. Over a maximum of 30 generations, the population evolved through nuclear fission and fusion processes. This number of generations was selected as a balance between exploration and convergence, allowing the algorithm sufficient iterations to optimize gene selection while avoiding excessive computational overhead. In our empirical tests, increasing the generations beyond 30 provided no improvements in classification accuracy.
In the context of microarray data, the nuclear fission phase introduced diversity by splitting solutions into smaller fragments, which correspond to alternative gene subsets. Equation (1) governs this process, generating new solutions by mutating the existing ones. Step sizes, dynamically adjusted using Equations (2) and (3), control the degree of variability, enabling broad exploration in early generations to identify different sets of candidate genes. As the algorithm progresses, these step sizes decrease, focusing on the refinement of the most promising gene subsets. Mutation factors from Equations (4) and (5) further adjust the solutions, ensuring that the algorithm does not converge prematurely and continues to explore the potential of unselected genes.
The nuclear fusion phase refines gene subsets by combining promising solutions. During the ionization step, solutions are adjusted using differences between randomly selected subsets, following Equation (6). This step ensures the algorithm can evaluate combinations of genes that might not have been initially included in a single solution. For microarray data, in which many genes have weak but complementary contributions to classification, this step is crucial. When the differences between the selected solutions are small, Lévy flight from Equation (7) is applied to introduce large jumps, allowing the algorithm to escape local optima and explore new subsets of genes.
In the fusion step, refined subsets are combined using Equation (8), producing solutions that integrate information from the best-performing gene subsets of previous generations. Lévy flight, applied when necessary through Equation (9), ensures that the fusion phase continues to explore new regions of the search space rather than stagnating on similar subsets. This capability was critical for handling the complexity of microarray data, in which interactions among genes can lead to nonlinear relationships affecting the classification accuracy.
The algorithm evaluated each solution using the fitness function described in Section 4.4, which combines LOOCV-based classification accuracy with a penalty for large subsets. The process continued until the maximum number of generations was reached. The final output was an optimized subset of genes that balanced high classification accuracy with minimal dimensionality, ensuring that the selected genes were both computationally efficient and biologically informative for cancer classification. The pseudo code of the NRO algorithm is shown in Algorithm 1.
Algorithm 1: Nuclear Reaction Optimization (NRO) Algorithm for Gene Selection
>
Require: Population size N = 500, Maximum generations T = 30
Ensure: Optimized subset of gene
                           ▷ Initialization
1: Initialize a population of 500 binary solutions x i , where x i [j] = 1 if gene
   j is selected, otherwise x i [j] = 0
2: Set bounds [0, 1] for solutions
3: Initialize global best solution X best
4: Compute initial fitness of X best using LOOCV
5: for g = 1 to T do
                     ▷ Fission Phase: Introduce Diversity
6:     for each solution x i in the population do
7:       Generate new solutions as per Equation (1)
8:       Adjust step sizes dynamically as per Equations (2) and (3)
9:       Apply mutation factors as per Equations (4) and (5)
10:     end for
                     ▷ Fusion Phase: Refine Solutions
11:     for each solution x i in the population do
12:       Adjust solutions through ionization as per Equation (6)
13:       if Ionized solutions are nearly identical then
14:         Apply Lévy flight adjustment as per Equation (7)
15:       end if
16:       Combine solutions through fusion as per Equation (8)
17:       if Fused solutions are nearly identical then
18:         Apply Lévy flight adjustment as per Equation (9)
19:       end if
20:     end for
                     ▷ Fitness Evaluation using LOOCV
21:     for each solution x i in the population do
22:       Compute classification accuracy using LOOCV
23:       Compute fitness: Fitness( x i ) = Accuracy Penalty(size( x i ))
24:       if Fitness ( x i )   >   Fitness ( X best ) then
25:            Update   X best   x i
26:       end if
27:     end for
28: end for
29: Return X best Optimized subset of genes

4.3. Apply Classifiers

To evaluate the performance of the gene subsets selected by the NRO algorithm, two machine learning classifiers were employed: Support Vector Machine (SVM) and k-Nearest Neighbors (k-NNs). These classifiers were chosen for their established effectiveness in handling high-dimensional data, making them particularly suitable for microarray datasets with thousands of genes but relatively few samples.
The SVM was configured with a linear kernel and a penalty parameter (C = 1), which is effective for binary classification tasks in which the data are high-dimensional and sparse. Its ability to maximize the margin between classes ensures robustness in separating cancerous and non-cancerous samples, as well as in distinguishing between different cancer subtypes.
The k-NN classifier was configured with k = 5, using Euclidean distance to measure similarity between samples. Its non-parametric nature allows it to adapt well to the inherent clustering within microarray datasets, particularly when the dimensionality has been reduced through gene selection.
Both classifiers were integrated into the fitness evaluation process, in which they were used to calculate the classification accuracy through Leave-One-Out Cross-Validation (LOOCV). By applying SVM and k-NNs to the gene subsets generated by NRO, the methodology ensures that the selected subsets are rigorously evaluated for their ability to distinguish between cancer classes.

4.4. Fitness Evaluation

The fitness evaluation process measures the quality of each candidate gene subset by balancing classification performance with dimensionality reduction. The fitness of a solution is computed as follows:
Fitness = LOOCV   Accuracy Penalty ,
where Equation (10) outlines the basic computation. The classification accuracy was calculated using LOOCV (Leave-One-Out Cross-Validation), a robust cross-validation technique suited to the small sample sizes typical of microarray datasets. In LOOCV, each sample serves as the test set exactly once, while the remaining samples form the training set. This ensures that the evaluation is both comprehensive and unbiased.
The penalty term was designed to discourage the selection of excessively large gene subsets, promoting dimensionality reduction. It was defined as follows:
Penalty = 0.0001 × N u m b e r   o f   S e l e c t e d   G e n e s
where Equation (11) indicates how the penalty scales with both the size of the gene subset and the progress of the optimization process, ensuring that the algorithm focuses on selecting compact subsets in later generations.
During the evaluation of alternative cross-validation frameworks, both 5-fold and 10-fold methods were tested across all datasets. These approaches resulted in classification accuracies approximately 1–2% lower than those obtained using Leave-One-Out Cross-Validation (LOOCV). Although LOOCV required significantly more computational time—typically 2 to 10 times longer depending on the dataset—it consistently produced higher accuracy. Therefore, LOOCV was selected as the cross-validation method in this study. This reflects a deliberate decision to prioritize accuracy over computational efficiency, especially in the context of small sample sizes in which each data point is critical. Moreover, numerous studies in bioinformatics have demonstrated that LOOCV provides better model reliability and generalization for high-dimensional microarray data [27,28,29,30], further supporting its adoption.

5. Results and Analysis

This section presents the results of applying the Nuclear Reaction Optimization (NRO) algorithm for gene selection in cancer classification across six benchmark microarray datasets. The performance of NRO was evaluated in terms of the number of selected genes and classification accuracy, followed by an analysis of precision, recall, and F1-score to further assess classification robustness. Additionally, the results are compared with state-of-the-art gene-selection algorithms to assess the relative effectiveness of NRO.
The results presented in Table 2 reveal that the NRO algorithm effectively selects informative gene subsets for cancer classification, yielding high classification accuracies across multiple datasets. However, due to the absence of dimensionality-reduction or filtering techniques, the number of selected genes remains relatively high. This was expected, as no feature-elimination techniques were applied before the optimization process.
Notably, the classification accuracy achieved with SVM consistently outperformed that of k-NNs across all the datasets. This aligned with the expectation that SVM, being more robust to high-dimensional data, would be better suited for the gene-expression datasets. The best accuracy for SVM was observed in the SRBCT dataset (99.26%), followed by Lymphoma (98.81%) and Lung (95.36%). Meanwhile, k-NNs yielded its highest accuracy in the Lung dataset (97.78%), but its overall performance was less stable, particularly for high-dimensional datasets like Leukemia1 and Leukemia2, where the classification accuracy dropped significantly for lower-performing runs.
Another key observation is that the number of selected genes varies across datasets, with some requiring a larger subset for optimal classification. For example, while only 74 genes were selected in the SRBCT dataset for SVM, the Colon dataset required 494 genes for its best-performing result. This indicates that NRO was highly adaptive in its selection process, though the lack of dimensionality reduction results in relatively large gene subsets, which could be further refined in future research.
NRO’s performance varied across datasets due to differences in gene count, sample size, and data complexity. It achieved the highest accuracy on SRBCT and Lymphoma, which had moderate gene counts and yielded smaller gene subsets. Lung also performed well despite high dimensionality, likely due to its larger sample size. Leukemia1 and Leukemia2 showed more variability, reflecting the challenge of optimizing in high-dimensional, small-sample datasets. Colon, despite having fewer genes, resulted in the lowest accuracy and the largest subsets, suggesting that feature selection was more difficult, possibly due to dataset-specific complexity.
To complement the evaluation of classification accuracy, a detailed analysis of the 95% confidence intervals (CIs) offers valuable insights into the stability and reliability of the results presented in Table 2. Confidence intervals provide a statistical range within which the true average accuracy is expected to fall, offering a measure of how consistently the NRO algorithm performs across multiple runs. In this study, the CIs were remarkably narrow, with widths ranging from 0.0019 to 0.0061 and an average width of 0.0036, indicating minimal variability and high repeatability in classification outcomes. For instance, the narrowest CI was observed for the Colon dataset using k-NNs ([0.6197, 0.6216]), while the widest CI appeared in Leukemia1 with k-NNs ([0.3731, 0.3792]); even in this case, the interval remained tight and statistically sound. Additionally, all the CIs were logically bounded between 0 and 1, validating the accuracy of the calculations. These results underscore the consistency of NRO’s performance and affirm that the reported accuracies are not outliers but represent stable, repeatable outcomes. Including confidence intervals, therefore, enriches the analysis by quantifying uncertainty and reinforcing the robustness of the findings.
While accuracy provides an overall measure of classification performance, it does not fully capture the model’s ability to minimize false positives and false negatives. To further evaluate the effectiveness of NRO-selected genes, Table 3 presents the precision, recall, and F1-score, which offer a deeper understanding of classification reliability.
Precision measures the proportion of correctly predicted positive cases out of all predicted positives. A higher precision means fewer false positives, which is crucial in medical applications to avoid unnecessary treatments. As shown in Table 3, SVM achieved perfect precision (100%) in the Lung, Lymphoma, and SRBCT datasets, indicating that it classifies all positive cases correctly without false positives. In other datasets, SVM maintained high precision, such as in Leukemia1 (98.60%) and Leukemia2 (95.81%). Meanwhile, k-NNs generally had lower precision, with its best performance in Lymphoma and Lung but lower values in other datasets, such as Colon (75.70%).
Recall, or sensitivity, measures the ability of the classifier to correctly identify all actual positive cases. A high recall reduces the risk of missing cancer cases, which is critical in medical diagnostics. SVM achieved perfect recall (100%) in the Lung, Lymphoma, and SRBCT datasets, ensuring that no positive cases are overlooked. In other datasets, recall remained high, such as in Leukemia1 (98.57%) and Leukemia2 (95.72%). The k-NNs also achieved 100% recall in the Lymphoma dataset, matching SVM, but in other cases, it performed worse than SVM, such as in Leukemia2 (80.70%) and Colon (74.39%).
The F1-score is the harmonic mean of the precision and recall, providing a balanced measure of classification performance. A high F1-score indicates that the classifier effectively minimizes both false positives and false negatives. SVM achieved an F1-score of 100% in the Lung, Lymphoma, and SRBCT datasets, reflecting its strong performance in these cases. In Leukemia1 and Leukemia2, SVM also maintained high F1-scores (98.57% and 95.70%, respectively). k-NNs reached 100% in the Lymphoma dataset but struggled in others, particularly Colon (71.34%) and SRBCT (79.70%), where its lower recall affected the overall performance.
To further evaluate the performance of NRO, Table 4 compared its SVM-based classification accuracy with other gene-selection algorithms, including Harris Hawks Optimization (HHO), Artificial Bee Colony (ABC), Particle Swarm Optimization (PSO), and Firefly Algorithm (FF). This comparison focuses solely on classification accuracy, as the majority of studies on gene-selection algorithms evaluate performance based only on accuracy. Other metrics such as precision, recall, and F1-score are not included, as they are rarely reported in gene-selection research. Accuracy remains the primary benchmark for assessing the effectiveness of selected gene subsets in classification tasks.
The results show that NRO was competitive but did not consistently outperform other methods. It achieved the highest accuracy only in the SRBCT dataset (99.26%), surpassing FF (98.8%) and other algorithms. However, in all other datasets, NRO was outperformed. For example, in the Colon dataset, NRO’s accuracy (82.16%) was the lowest, while FF achieved 98.2%. In Leukemia1, NRO reached 95.53%, which was lower than HHO (97.22%) and FF (100%). In Lung, both HHO and FF achieved 100% accuracy, surpassing NRO’s 95.36%. Similarly, in Leukemia2, NRO’s 92.47% was outperformed by ABC (97.22%) and FF (97.2%). In Lymphoma, NRO achieved 98.81%, slightly below HHO (100%).
Despite not achieving the highest accuracy in all the datasets, NRO consistently delivered a strong and reliable performance across diverse cancer datasets. Its ability to outperform methods like HHO and ABC in datasets such as Leukemia2 and SRBCT, and to achieve near-perfect accuracy in Lymphoma (98.81%) and SRBCT (99.26%), demonstrates its robustness in handling high-dimensional gene-expression data. These results highlight NRO’s effectiveness and adaptability, validating its optimization approach based on nuclear reaction mechanisms. Overall, NRO proved to be a promising and competitive algorithm for gene selection in cancer classification, with the potential for further enhancement through hybrid or refined optimization strategies.

6. Conclusions

This study explored the application of the Nuclear Reaction Optimization (NRO) algorithm for gene selection in cancer classification using six benchmark microarray datasets. The results demonstrated that NRO effectively identified relevant gene subsets, leading to high classification accuracy, particularly when paired with SVM. Compared with state-of-the-art optimization methods, including HHO, ABC, PSO, and FF, NRO showed competitive performance, outperforming some approaches in specific datasets such as SRBCT and Leukemia2. NRO demonstrated strong adaptability in high-dimensional feature spaces. Its ability to refine solutions through nuclear fission and fusion highlights its potential as a powerful bioinformatics tool for cancer classification. However, despite its promising results, certain limitations hinder its full efficiency in gene-selection tasks.
A key limitation of this study is the large number of selected genes. Since no dimensionality-reduction techniques were applied before running NRO, this led to reduced interpretability and higher computational costs. Additionally, the Leave-One-Out Cross-Validation (LOOCV) method, while chosen for its superior accuracy, further increased computational expense due to its iterative training process. Future work should integrate filtering techniques before applying NRO to remove irrelevant genes, improving both classification accuracy and efficiency. Furthermore, hybrid optimization approaches, in which a secondary metaheuristic is used before NRO to refine the initial population, could help accelerate convergence, enhance feature-selection quality, and reduce unnecessary computations. These improvements would make the method more scalable, interpretable, and effective for larger and more complex datasets.
Additionally, this study did not assess the biological relevance of the selected genes. Future research will include pathway enrichment analysis and comparison with known cancer biomarkers to enhance the interpretability of the results. While NRO has previously been applied to RNA-Seq data, this study demonstrates its effectiveness on microarray data. Building on these findings, future studies will evaluate NRO’s scalability and performance on larger and more complex genomic datasets.

Author Contributions

Conceptualization, H.A.; data curation, S.A.; formal analysis, H.A. and S.A.; investigation, S.A.; methodology, H.A.; supervision, H.A.; validation, S.A.; writing—original draft, S.A.; writing—review and editing, S.A. and H.A. All authors have read and agreed to the published version of the manuscript.

Funding

This is funded by Research Supporting Project Number (RSPD2025R904), King Saud University, Riyadh, Saudi Arabia.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors really appreciate the kind support from the King Saud University, Riyadh, Saudi Arabia.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fitzgerald, R.C.; Antoniou, A.C.; Fruk, L.; Rosenfeld, N. The future of early cancer detection. Nat. Med. 2022, 28, 666–677. [Google Scholar] [CrossRef] [PubMed]
  2. Krzyszczyk, P.; Acevedo, A.; Davidoff, E.J.; Timmins, L.M.; Marrero-Berrios, I.; Patel, M.; White, C.; Lowe, C.; Sherba, J.J.; Hartmanshenn, C.; et al. The growing role of precision and personalized medicine for cancer treatment. Technology 2018, 06, 79–100. [Google Scholar] [CrossRef] [PubMed]
  3. Lu, Y.; Han, J. Cancer classification using gene expression data. Inf. Syst. 2003, 28, 243–268. [Google Scholar]
  4. Alsaykhan, L.K.; Maashi, M.S. A hybrid detection model for acute lymphocytic leukemia using support vector machine and particle swarm optimization (SVM-PSO). Sci. Rep. 2024, 14, 23483. [Google Scholar] [CrossRef]
  5. Heidari, A.A.; Mirjalili, S.; Faris, H.; Aljarah, I.; Mafarja, M.; Chen, H. Harris hawks optimization: Algorithm and applications. Future Gener. Comput. Syst. 2019, 97, 849–872. [Google Scholar] [CrossRef]
  6. Wei, Z.; Huang, C.; Wang, X.; Han, T.; Li, Y. Nuclear Reaction Optimization: A Novel and Powerful Physics-Based Algorithm for Global Optimization. IEEE Access 2019, 7, 66084–66109. [Google Scholar] [CrossRef]
  7. Gene Selection via Improved Nuclear Reaction Optimization Algorithm for Cancer Classification in High-Dimensional Data|Journal of Big Data|Full Text. Available online: https://journalofbigdata.springeropen.com/articles/10.1186/s40537-024-00902-z (accessed on 4 October 2024).
  8. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  9. Wang, H.; Huang, G. Application of support vector machine in cancer diagnosis. Med. Oncol. 2011, 28, 613–618. [Google Scholar] [CrossRef]
  10. Falcari, T.; Saotome, O.; Pires, R.; Campo, A.B. Evaluation of multi-class support-vector machines strategies and kernel adjustment levels in hand posture recognition by analyzing sEMG signals acquired from a wearable device. Biomed. Eng. Lett. 2019, 10, 275–284. [Google Scholar] [CrossRef]
  11. Laaksonen, J.; Oja, E. Classification with learning k-nearest neighbors. In Proceedings of the International Conference on Neural Networks (ICNN’96), Washington, DC, USA, 3–6 June 1996; Volume 3, pp. 1480–1483. [Google Scholar] [CrossRef]
  12. Reserach scholar, Pacific University, Udaipur, Rajasthan, India; Moon, K.; Jetawat, A. Predicting Lung Cancer with K-Nearest Neighbors (KNN): A Computational Approach. Indian J. Sci. Technol. 2024, 17, 2199–2206. [Google Scholar] [CrossRef]
  13. Ehsani, R.; Drabløs, F. Robust Distance Measures for kNN Classification of Cancer Data. Cancer Inf. 2020, 19, 1176935120965542. [Google Scholar] [CrossRef]
  14. Zhuoran, Z.; Changqiang, H.; Hanqiao, H.; Shangqin, T.; Kangsheng, D. An optimization method: Hummingbirds optimization algorithm. J. Syst. Eng. Electron. 2018, 29, 386–404. [Google Scholar] [CrossRef]
  15. Panda, P.; Bisoy, S.K.; Panigrahi, A.; Pati, A.; Sahu, B.; Guo, Z.; Liu, H.; Jain, P. BIMSSA: Enhancing cancer prediction with salp swarm optimization and ensemble machine learning approaches. Front. Genet. 2025, 15, 1491602. [Google Scholar] [CrossRef]
  16. AlMazrua, H.; AlShamlan, H. A New Algorithm for Cancer Biomarker Gene Detection Using Harris Hawks Optimization. Sensors 2022, 22, 7273. [Google Scholar] [CrossRef] [PubMed]
  17. Alweshah, M.; Khalaileh, S.A.; Gupta, B.B.; Almomani, A.; Hammouri, A.I.; Al-Betar, M.A. The monarch butterfly optimization algorithm for solving feature selection problems. Neural Comput. Applic 2022, 34, 11267–11281. [Google Scholar] [CrossRef]
  18. Almugren, N.; Alshamlan, H. FF-SVM: New FireFly-based Gene Selection Algorithm for Microarray Cancer Classification. In Proceedings of the 2019 IEEE Conference on Computational Intelligence in Bioinformatics and Computational Biology (CIBCB), Siena, Italy, 9–11 July 2019; pp. 1–6. [Google Scholar] [CrossRef]
  19. Nssibi, M.; Manita, G.; Chhabra, A.; Mirjalili, S.; Korbaa, O. Gene selection for high dimensional biological datasets using hybrid island binary artificial bee colony with chaos game optimization. Artif. Intell. Rev. 2024, 57, 51. [Google Scholar] [CrossRef]
  20. AlMazrua, H.; AlShamlan, H.; AlMazrua, H.; AlShamlan, H. A New Approach for Selecting Features in Cancer Classification Using Grey Wolf Optimizer. In The International Conference on Deep Learning, Big Data and Blockchain (DBB 2022); Awan, I., Younas, M., Bentahar, J., Benbernou, S., Eds.; Springer International Publishing: Cham, Switzerland, 2023; pp. 53–64. [Google Scholar] [CrossRef]
  21. Khan, J.; Wei, J.S.; Ringnér, M.; Saal, L.H.; Ladanyi, M.; Westermann, F.; Berthold, F.; Schwab, M.; Antonescu, C.R.; Peterson, C.; et al. Classification and diagnostic prediction of cancers using gene expression profiling and artificial neural networks. Nat. Med. 2001, 7, 673–679. [Google Scholar] [CrossRef]
  22. Alizadeh, A.A.; Eisen, M.B.; Davis, R.E.; Ma, C.; Lossos, I.S.; Rosenwald, A.; Boldrick, J.C.; Sabet, H.; Tran, T.; Yu, X.; et al. Distinct types of diffuse large B-cell lymphoma identified by gene expression profiling. Nature 2000, 403, 503–511. [Google Scholar] [CrossRef]
  23. Armstrong, S.A.; Staunton, J.E.; Silverman, L.B.; Pieters, R.; den Boer, M.L.; Minden, M.D.; Sallan, S.E.; Lander, E.S.; Golub, T.R.; Korsmeyer, S.J. MLL translocations specify a distinct gene expression profile that distinguishes a unique leukemia. Nat. Genet. 2002, 30, 41–47. [Google Scholar] [CrossRef]
  24. Beer, D.G.; Kardia, S.L.R.; Huang, C.-C.; Giordano, T.J.; Levin, A.M.; Misek, D.E.; Lin, L.; Chen, G.; Gharib, T.G.; Thomas, D.G.; et al. Gene-expression profiles predict survival of patients with lung adenocarcinoma. Nat. Med. 2002, 8, 816–824. [Google Scholar] [CrossRef]
  25. Alon, U.; Barkai, N.; Notterman, D.A.; Gish, K.; Ybarra, S.; Mack, D.; Levine, A.J. Broad patterns of gene expression revealed by clustering analysis of tumor and normal colon tissues probed by oligonucleotide arrays. Proc. Natl. Acad. Sci. USA 1999, 96, 6745–6750. [Google Scholar] [CrossRef] [PubMed]
  26. Golub, T.R.; Slonim, D.K.; Tamayo, P.; Huard, C.; Gaasenbeek, M.; Mesirov, J.P.; Coller, H.; Loh, M.L.; Downing, J.R.; Caligiuri, M.A.; et al. Molecular Classification of Cancer: Class Discovery and Class Prediction by Gene Expression Monitoring. Science 1999, 286, 531–537. [Google Scholar] [CrossRef]
  27. Lumumba, V.; Sang, D.; Mpaine, M.; Makena, N.; Musyimi, D. Kavita Comparative Analysis of Cross-Validation Techniques: LOOCV, K-folds Cross-Validation, and Repeated K-folds Cross-Validation in Machine Learning Models. Am. J. Theor. Appl. Stat. 2024, 13, 127–137. [Google Scholar] [CrossRef]
  28. Alshamlan, H.M.; Badr, G.H.; Alohali, Y.A. Genetic Bee Colony (GBC) algorithm: A new gene selection method for microarray cancer classification. Comput. Biol. Chem. 2015, 56, 49–60. [Google Scholar] [CrossRef] [PubMed]
  29. mRMR-ABC: A Hybrid Gene Selection Algorithm for Cancer Classification Using Microarray Gene Expression Profiling—Alshamlan—2015—BioMed Research International—Wiley Online Library. Available online: https://onlinelibrary.wiley.com/doi/full/10.1155/2015/604910 (accessed on 22 March 2025).
  30. Alanni, R.; Hou, J.; Azzawi, H.; Xiang, Y. A novel gene selection algorithm for cancer classification using microarray datasets. BMC Med. Genom. 2019, 12, 10. [Google Scholar] [CrossRef]
  31. Computer Science Department, King Saud University; Alshamlan, H.M.; Badr, G.H.; Alohali, Y.A. ABC-SVM: Artificial Bee Colony and SVM Method for Microarray Gene Selection and Multi Class Cancer Classification. J. Mach. Learn. Comput. 2016, 6, 184–190. [Google Scholar] [CrossRef]
  32. Shen, Q.; Shi, W.-M.; Kong, W.; Ye, B.-X. A combination of modified particle swarm optimization algorithm and support vector machine for gene selection and tumor classification. Talanta 2007, 71, 1679–1683. [Google Scholar] [CrossRef] [PubMed]
  33. Abdi, M.J.; Giveki, D. Automatic detection of erythemato-squamous diseases using PSO–SVM based on association rules. Eng. Appl. Artif. Intell. 2013, 26, 603–608. [Google Scholar] [CrossRef]
Figure 1. Nuclear fission process, illustrating the inspiration for solution diversification in the NRO algorithm.
Figure 1. Nuclear fission process, illustrating the inspiration for solution diversification in the NRO algorithm.
Diagnostics 15 00927 g001
Figure 2. Nuclear fusion process, illustrating the inspiration for solution refinement in the NRO algorithm.
Figure 2. Nuclear fusion process, illustrating the inspiration for solution refinement in the NRO algorithm.
Diagnostics 15 00927 g002
Figure 3. Flowchart of the methodology.
Figure 3. Flowchart of the methodology.
Diagnostics 15 00927 g003
Table 1. Statistics of microarray cancer datasets.
Table 1. Statistics of microarray cancer datasets.
Microarray DatasetClassesSamplesTotal Genes
Colon [25]2622000
Leukemia1 [26]2727129
Lung [24]2967129
SRBCT [21]4832308
Lymphoma [22]3624026
Leukemia2 [23]3727129
Table 2. Summary of accuracy results using NRO.
Table 2. Summary of accuracy results using NRO.
DatasetTotal GenesClassifierSelected GenesAccuracyCI (95%)
BestAverageWorst
Colon2000SVM49482.16%69.53%60.74%[0.6904, 0.6933]
k-NNs61576.11%62.57%54.23%[0.6197, 0.6216]
Leukemia17129SVM30895.53%55.90%27.37%[0.5341, 0.5368]
k-NNs17178.84%38.96%10.65%[0.3731, 0.3792]
Leukemia27129SVM19892.47%52.28%24.55%[0.5046, 0.5083]
k-NNs41179.22%37.32%9.35%[0.3564, 0.3625]
Lung7129SVM46395.36%56.53%29.11%[0.5482, 0.5515]
k-NNs11797.78%55.33%27.67%[0.5395, 0.5437]
Lymphoma4026SVM11998.81%75.25%59.74%[0.7437, 0.7469]
k-NNs7299.28%75.79%59.74%[0.7439, 0.7486]
SRBCT2308SVM7499.26%85.89%76.96%[0.8535, 0.8557]
k-NNs25980.54%67.42%57.83%[0.6708, 0.6732]
Table 3. Precision, recall, and F1-score Using NRO.
Table 3. Precision, recall, and F1-score Using NRO.
DatasetTotal GenesClassifierPrecisionRecallF1-Score
Colon2000SVM81.25%81.46%81.28%
k-NNs75.70%74.39%71.34%
Leukemia17129SVM98.60%98.57%98.57%
k-NNs84.23%82.22%80.67%
Leukemia27129SVM95.81%95.72%95.70%
k-NNs83.37%80.70%80.10%
Lung7129SVM100%100%100%
k-NNs99.11%99.02%99.04%
Lymphoma4026SVM100%100%100%
k-NNs100%100%100%
SRBCT2308SVM100%100%100%
k-NNs84.05%81.27%79.70%
Table 4. Comparison of gene-selection algorithms’ accuracies with SVM classifier for six microarray datasets.
Table 4. Comparison of gene-selection algorithms’ accuracies with SVM classifier for six microarray datasets.
DatasetNROHHOABC [31]PSO [32]PSO [33]FF [18]
Colon82.16%90.32%95.61%85.48%87.01%98.2%
Leukemia195.53%97.22%93.05%94.44%93.06%100%
Leukemia292.47%84.72%97.22%--97.2%
Lung95.36%100%97.91%--100%
Lymphoma98.81%100%96.96%---
SRBCT99.26%92.77%95.36%--98.8%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alkamli, S.; Alshamlan, H. Evaluating the Nuclear Reaction Optimization (NRO) Algorithm for Gene Selection in Cancer Classification. Diagnostics 2025, 15, 927. https://doi.org/10.3390/diagnostics15070927

AMA Style

Alkamli S, Alshamlan H. Evaluating the Nuclear Reaction Optimization (NRO) Algorithm for Gene Selection in Cancer Classification. Diagnostics. 2025; 15(7):927. https://doi.org/10.3390/diagnostics15070927

Chicago/Turabian Style

Alkamli, Shahad, and Hala Alshamlan. 2025. "Evaluating the Nuclear Reaction Optimization (NRO) Algorithm for Gene Selection in Cancer Classification" Diagnostics 15, no. 7: 927. https://doi.org/10.3390/diagnostics15070927

APA Style

Alkamli, S., & Alshamlan, H. (2025). Evaluating the Nuclear Reaction Optimization (NRO) Algorithm for Gene Selection in Cancer Classification. Diagnostics, 15(7), 927. https://doi.org/10.3390/diagnostics15070927

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop