You are currently viewing a new version of our website. To view the old version click .
Bioengineering
  • Article
  • Open Access

13 November 2025

A Multi-Task Ensemble Strategy for Gene Selection and Cancer Classification

,
,
and
1
School of Cyberspace Security, Hainan University, Haikou 570228, China
2
School of Computing and Information Science, Faculty of Science and Engineering, Anglia Ruskin University, Cambridge CB1 1PT, UK
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
This article belongs to the Section Biosignal Processing

Abstract

Gene expression-based tumor classification aims to distinguish tumor types based on gene expression profiles. This task is difficult due to the high dimensionality of gene expression data and limited sample sizes. Most datasets contain tens of thousands of genes but only a small number of samples. As a result, selecting informative genes is necessary to improve classification performance and model interpretability. Many existing gene selection methods fail to produce stable and consistent results, especially when training data are limited. To address this, we propose a multi-task ensemble strategy that combines repeated sampling with joint feature selection and classification. The method generates multiple training subsets and applies multi-task logistic regression with 2 , 1 group sparsity regularization to select a subset of genes that appears consistently across tasks. This promotes stability and reduces redundancy. The framework supports integration with standard classifiers such as logistic regression and support vector machines. It performs both gene selection and classification in a single process. We evaluate the method on simulated and real gene expression datasets. The results show that it outperforms several baseline methods in classification accuracy and the consistency of selected genes.

1. Introduction

Gene expression-based tumor classification is an important task in cancer research. It involves identifying tumor types using gene expression profiles. These datasets often contain tens of thousands of genes, which leads to very high-dimensional feature spaces. At the same time, the number of samples is usually small. This combination creates major challenges for analysis. To reduce complexity and improve performance, it is important to select a smaller subset of genes that carry the most useful information. Good gene selection can improve classification accuracy and also make the models easier to understand [,,,]. Finding patterns in gene expression data linked to specific cancer types can also help with diagnosis, prognosis, and designing targeted treatments.
Many gene selection methods have been proposed. They are usually grouped into three main types [,,]: filter methods, wrapper methods, and embedded methods. Filter methods select genes before training the model. They do not depend on any specific classifier [,]. These methods use statistical scores to evaluate each gene, so they are fast and work well with high-dimensional data. Common examples are ReliefF and mRMR. ReliefF [] assigns scores to genes based on how well they separate samples from different classes, using distance to nearby samples. mRMR (Minimum Redundancy Maximum Relevance) [] selects genes that are both highly relevant to the class labels and not redundant with each other.
To combine the strengths of different measures, Zhang et al. [] proposed a hybrid method that mixes ReliefF and mRMR. Omuya et al. [] combined principal component analysis with information gain. Gong et al. [] created a filter method that uses both Pearson correlation and mutual information. These hybrid approaches improve performance by selecting features that are both informative and diverse. In other areas, such as neurological disorder detection, hybrid methods have also shown strong results [].
While filter methods are fast and scale well, they require users to set the number of selected genes in advance. This often needs expert knowledge. Also, since these methods run before model training, they do not adjust based on the classifier. This can reduce their performance, because the selected features may not match the model’s needs.
In contrast to filter methods, wrapper methods use a specific learning algorithm to evaluate different gene subsets. These methods run an iterative search to find the subset that gives the highest classification accuracy []. For example, Ghosh et al. [] proposed a recursive memetic algorithm that refines gene subsets in steps to improve classification. GeFeS [] is a genetic algorithm-based approach that uses multiple populations and adaptive weights to perform selection in high-dimensional settings. Other recent work includes a Markov blanket-based wrapper method by Hassan et al. [], and a two-stage greedy search method for sentiment classification developed by Saugbacs and Arif [].
Compared to filter methods, wrapper methods often reach higher classification accuracy because they directly test gene subsets during model training. However, they are slower and more computationally demanding. They are also more likely to overfit because the model is evaluated many times. To reduce these issues, some researchers have combined filter and wrapper methods. For instance, Sun et al. [] combined ReliefF with ant colony optimization to improve tumor classification. Hu et al. [] proposed a method that combines information gain, the Fisher Score, and evolutionary computation to perform efficient feature selection.
Embedded methods take a different approach. These methods include gene selection as part of the model training process. This makes them more efficient and often more generalizable. They usually rely on regularization to force sparsity in the learned weights, which helps interpret the selected genes [,,,,]. For example, adaptive elastic net models select features by balancing relevance and redundancy during training []. Cawley et al. [] used sparse logistic regression with Bayesian regularization to identify genes relevant to cancer, improving both accuracy and interpretability. Xing et al. [] developed a hypergraph-based method that integrates gene selection into model learning using dictionary learning. Liu et al. [] proposed a method that accounts for correlations between genes to improve selection and classification performance.
More recently, deep learning-based methods have been used for feature selection in gene expression analysis. These models can learn complex feature interactions during training and are well suited to high-dimensional data [,,,,,]. Deep learning models have also been used in related fields, such as medical image segmentation, where they are trained on noisy or AI-generated labels []. Other medical imaging studies show that learning texture-based patterns can improve diagnostic results. For example, Li et al. [] used external image references to enhance MRI resolution and improve the visibility of fine structures. Similar methods have also been applied in plant disease detection using UAV images and an improved YOLO model [].
Although many gene selection methods have been developed, most still struggle to produce gene sets that are both stable and biologically meaningful, especially when the data has high dimensionality and limited sample size. In gene expression–based cancer classification, this “high-dimension, low-sample” problem makes it hard to identify reliable biomarkers. Standard methods, including those based on evolutionary algorithms, are still widely used but often show inconsistent results and require careful parameter tuning. Deep learning and graph neural network (GNN) models have shown strong classification performance, but they usually need large datasets and still lack interpretability. Benchmarking practices for these models also remain unsettled.
To address these problems, we propose a multi-task ensemble strategy for gene selection in tumor classification. The method starts by generating multiple training subsets using random sampling. Each subset is then modeled using multi-task logistic regression with 2 , 1 group sparsity regularization. This regularization, combined with ensemble sampling, helps identify genes that appear consistently across subsets, improving selection stability. Similar sampling-based strategies have been used in other domains to improve robustness and reduce bias during training []. The framework is general and can work with different classifiers; in this study, we use logistic regression. Because classification and gene selection are done together, the selected genes are more likely to be relevant across the dataset. This joint learning process improves both interpretability and classification performance compared to standard approaches.
The main contributions of this work can be summarized as follows:
  • We propose a novel ensemble-based framework that systematically enhances gene selection stability by leveraging a sampling approach combined with multi-task learning.
  • The use of 2 , 1 regularization supports the selection of a compact yet informative gene set across multiple sampling subsets, reducing redundancy and improving model interpretability.
  • Experimental results validate the effectiveness of our proposed ensemble approach, demonstrating that it achieves superior performance in terms of classification and gene selection consistency across datasets.
The remainder of this paper is organized as follows: Section 1 presents the proposed multi-task ensemble strategy for gene selection and cancer classification. Section 3 details the optimization procedure and provides the corresponding convergence analysis. Section 4 reports the experimental results and performance evaluation. Finally, Section 5 concludes the paper and outlines potential directions for future work.

2. Methodology

This section describes the proposed multi-task ensemble strategy for gene selection in tumor classification. The method combines data sampling with multi-task logistic regression and incorporates 2 , 1 group sparsity regularization. The goal is to select genes that are both stable across samples and relevant for classification.
Let the gene expression data be X R n × d , where each row is a sample and each column is a gene. Let the label vector be Y { 1 , 1 } n . We first generate m training subsets by randomly sampling 70% of the data without replacement. This results in m separate tasks, with datasets { X 1 , X 2 , , X m } and corresponding labels { Y 1 , Y 2 , , Y m } . Each subset is used to train a task-specific logistic regression model. The objective function is:
L ( X , Y , W , c ) = k = 1 m 1 n k i = 1 n k l o g 1 + e x p Y i k X i k W k + c k
Here, W = [ W 1 , W 2 , , W m ] is the matrix of model weights for the m tasks, with each W k R d . The term c k is the bias for task k. Each n k is the number of samples in subset k, and  X i k and Y i k are the i-th sample and its label in that subset.
In addition, the objective of this formulation is to select a set of relevant genes that are consistent across multiple tasks, while simultaneously maximizing overall classification accuracy. The goal is to identify a stable set of relevant genes that perform well in classifying tumor samples, ensuring that the selected genes are consistent and informative across all tasks. To achieve this, we utilize 2 , 1 group sparsity regularization, which promotes shared gene selection across tasks. This regularization technique ensures that the genes chosen by the model are not only relevant to the classification task but also stable and consistent across different data subsets, enhancing the robustness and interpretability of the gene selection process. The Formulation (1) can be further defined as:
L ( X , Y , W , c ) = k = 1 m 1 n k i = 1 n k l o g 1 + e x p Y i k X i k W k + c k + λ k = 1 m W k 2 , 1
where k = 1 m W k 2 , 1 is 2 , 1 group sparsity regularization and λ is a regularization parameter that controls the degree of sparsity. This 2 , 1 norm encourages the selection of genes that are consistently relevant across tasks, thus promoting shared gene selection and ensuring that the model remains both efficient and interpretable. The overall processing of the proposed method is shown in Figure 1.
Figure 1. Diagram of the proposed multi-task ensemble framework. The pipeline consists of: (1) sampling multiple training subsets from the input gene expression data, (2) performing joint gene selection and classification using multi-task logistic regression with 2 , 1 group sparsity, and (3) aggregating outputs to determine the final model. The colors are used only to distinguish different sampled subsets and task components, and they have no biological meaning.

3. Optimization

In this section, we employ the proximal gradient descent method to optimize the proposed model. Specifically, we approximate the objective function in each iteration using a quadratic expansion around the current estimate. The approximation formulation for the weights W at iteration t + 1 is given by:
W ( t + 1 ) = m i n W , C L W ( t ) + W ( t + 1 ) W ( t ) , L ( W ( t ) ) + γ 2 W ( t + 1 ) W ( t ) 2 2 + λ k = 1 m W k 2 , 1
where γ is the step size, which can be determined by line search, and  W ( t + 1 ) represents the updated estimate for W at iteration t + 1 . To further simplify, Equation (3) can be reformulated as:
W ( t + 1 ) = m i n W 1 2 W A 2 2 + λ γ k = 1 m W k 2 , 1
where A = W ( t ) ( 1 / γ ) L ( W ( t ) ) . Overall, the following formulation allows for a closed-form solution for W ( t + 1 ) using the proximal operator associated with the 2 , 1 norm, which promotes shared sparsity across tasks:
W j = 1 λ A j 2 γ + A j ; for   each j = 1 , 2 , , d
The matrix W can be estimated using the above block-wise thresholding function.
Algorithm 1 shows the data sampling and optimization steps used to solve Equation (2). First, random sampling is used to create multiple subsets from the training data. These subsets define the tasks in the multi-task logistic regression model. Then, the weight matrix W is updated using proximal gradient descent. A line search adjusts the step size γ at each iteration. The algorithm repeats these steps until the objective function stops decreasing.
Algorithm 1 Optimization
1:
Input:  X , Y , λ , m
2:
Initialize:  γ = 1 , η = 0.7 (Samping rate)
3:
Data Sampling:
4:
for  k = 1 : m   do
5:
    Based on η = 0.7 , randomly generate subset index Γ k
6:
     X k = X Γ k
7:
     Y k = Y Γ k
8:
end for
9:
Proximal gradient descent:
10:
repeat
11:
    Compute gradient L ( W ( t ) ) , set A = ( W ( t ) 1 γ L ( W ( t ) ) )
12:
    for  j = 1 , 2 , , d  do
13:
          W j = 1 λ A j 2 γ + A j
14:
    end for
15:
    if  L ( W ) L ( W ( t ) ) <  then
16:
         break and output W ( t + 1 ) = W
17:
    else
18:
          γ = γ · α where α is user-defined
19:
    end if
20:
until convergence
21:
Output:  W ( t + 1 )
For λ -selection, we use 5-fold cross-validation over a logarithmic grid, select λ using mean validation log-loss (equivalently, maximizing validation accuracy), retrain on the full training subset with λ , and evaluate on the held-out test set. The same λ is shared across tasks to align sparsity patterns.
Classification. Our model jointly performs feature selection and logistic regression within a unified framework, enabling direct classification once the optimal parameters ( W , c ) are obtained. For a test sample x i R d , we define the task-specific logits as
z i ( k ) = x i W k + c k , k = 1 , , m .
We then aggregate the predictions by averaging the logits across all m subsampled tasks:
z ¯ i = 1 m k = 1 m z i ( k ) .
The aggregated logit is mapped through the logistic link function to produce the predicted probability:
p ^ i = σ ( z ¯ i ) = 1 1 + exp ( z ¯ i ) .
Finally, the predicted class label is assigned as
y ^ i = 1 , p ^ i > τ , 0 , p ^ i τ ,
where τ = 0.5 by default.
Vectorized form. For a test data matrix X R n × d and an all-ones vector 1 R n , the prediction can be expressed compactly as
p ^ = σ 1 m k = 1 m ( X W k + 1 c k ) ,
where σ ( · ) is applied elementwise.

Computational and Convergence Analysis

The proposed multi-task ensemble method uses random sampling, multi-task logistic regression, and 2 , 1 group sparsity regularization. Optimization is done using proximal gradient descent. The total computational complexity is O ( T · m · n · d ) , where T is the number of iterations, m is the number of sampled subsets, n is the number of samples, and d is the number of genes. The cost grows linearly with all key variables, so the method can scale to high-dimensional gene expression data. The 2 , 1 regularization encourages shared sparsity across tasks, which improves the consistency of selected genes. The objective function is convex, so the algorithm converges to a global minimum. Related approaches using sparse regularization in multi-task settings have been explored for objective reduction in many-objective optimization problems [,]. Compared to filter methods, this approach is more computationally demanding but does not require setting the number of selected genes in advance and achieves better accuracy and stability.
Let L ( W ) = f ( W ) + λ W 2 , 1 where f ( W ) is convex and has a Lipschitz continuous gradient with Lipschitz constant L. Let γ ( 0 , 1 / L ] be the step size. Then the following convergence results hold:
  • The sequence { W ( t ) } generated by the proximal gradient descent algorithm satisfies:
    • The objective value sequence { L ( W ( t ) ) } is non-increasing.
    • L ( W ( t ) ) L * , where L * is the global minimum.
    • Every limit point of { W ( t ) } is a minimizer of L ( W ) .
  • Since f ( W ) is convex with Lipschitz continuous gradient, we have:
    f ( W ( t + 1 ) ) f ( W ( t ) ) + f ( W ( t ) ) , W ( t + 1 ) W ( t ) + L 2 W ( t + 1 ) W ( t ) F 2 .
  • Combining with the update rule using the proximal operator, we obtain that:
    L ( W ( t + 1 ) ) L ( W ( t ) ) ,
It shows monotonic decrease. By the descent property and the convexity of both f and W 2 , 1 , standard results in proximal gradient methods ensure convergence to a global minimum. This convergence analysis guarantees that the proposed optimization procedure reliably finds a globally optimal solution under mild assumptions.

4. Experiment

This section presents the experimental results and analysis of the proposed method, evaluated on several simulated datasets and public microarray datasets characterized by high dimensionality and small sample sizes. First, the simulation and real gene expression datasets and the competing algorithms used in this analysis are briefly described. The experimental results are then presented and discussed from various perspectives.

4.1. Simulation Data

To simulate the high-dimensional nature of gene expression data, we generate high-dimensional datasets with small sample sizes. These datasets contain a large number of irrelevant features and a small set of relevant variables. Following the approach outlined in [], the sample matrix is generated from a multivariate normal distribution. Specifically, we generate six different simulated datasets, each with varying sample and feature sizes. Each dataset includes 20 predefined relevant genes. Detailed information on the generated simulation data can be found in Table 1.
Table 1. Simulation dataset specifications. Each dataset consists of a specified number of samples and gene expression features.

4.2. Real Gene Expression Data

This study evaluates the proposed method and competing approaches using four publicly available gene expression datasets:
  • Leukemia Dataset: The preprocessed Leukemia dataset comprises 3571 genes and 72 samples, including 47 cases of acute lymphoblastic leukemia (ALL) and 25 cases of acute myeloid leukemia (AML).
  • Colon Dataset: The Colon microarray dataset contains 2000 genes, with 22 samples from normal tissues and 40 samples from cancerous tissues.
  • Lung Dataset: This dataset consists of 12,533 genes across 181 tissue samples, including 31 mesothelioma (MPM) samples and 150 adenocarcinoma (ADCA) samples.
  • DLBCL Dataset: The used DLBCL dataset contains 6285 genes and comprises 58 samples of diffuse large B-cell lymphoma (DLBCL) and 19 samples of follicular lymphoma (FL).
Detailed information of the gene expression data can be found in Table 2.
Table 2. Summary of real gene expression datasets used for evaluation.

4.3. Competing Methods

In this study, we compare the performance of the proposed method in a simulation setting with four baseline algorithms: the L 1 -penalized model, the Elastic Net penalized model, mRMR with logistic regression, and ReliefF with logistic regression. Since mRMR and ReliefF are filter methods that select key genes before training the classification model, we set the number of selected genes to range from 10 to 50, with steps of 10 (i.e., [10, 20, 30, 40, 50]) for each dataset.
Both our method and the L 1 -penalized model, as well as the Elastic Net penalized model, involve a parameter λ to control the sparsity of the model for gene selection and learning. Therefore, we use 5-fold cross-validation to automatically select the optimal parameter for the final model.

4.4. Experimental Setup

In our experimental studies, we randomly select 50% and 60% of the entire dataset as training data, with the remaining portion used as test data to evaluate the performance of each method. For each setting, the experiments are repeated 10 times, and the average results are reported.

4.5. Comparison Results

The experimental results presented in Figure 2, Figure 3, Figure 4 and Figure 5 provide a comprehensive comparison of the proposed multi-task ensemble strategy for gene selection against several baseline algorithms on both simulated and real gene expression datasets. The performance of each method was evaluated in terms of prediction accuracy using varying proportions of training data (50% and 60%). Specifically, Figure 2 and Figure 3 illustrate the prediction accuracy of the proposed approach and baseline algorithms on six simulated datasets with 50% and 60% training data, respectively. The results demonstrate that the proposed method consistently outperforms the baseline algorithms, including the L 1 -penalized model, the Elastic Net penalized model, mRMR combined with logistic regression, and ReliefF combined with logistic regression. Similarly, for the real datasets, the proposed multi-task ensemble strategy exhibits superior performance, with particularly significant improvements in classification accuracy observed in the Leukemia and Colon datasets. Overall, models that integrate gene selection and classification learning, such as the L 1 -penalized model, the Elastic Net penalized model, and the proposed multi-task ensemble strategy, generally outperform filter-based approaches like mRMR and ReliefF.
Figure 2. Prediction accuracy of five methods on six simulated datasets with 50% training data, comparing the proposed approach with baseline algorithms.
Figure 3. Prediction accuracy of five methods on six simulated datasets with 60% training data, comparing the proposed approach with baseline algorithms.
Figure 4. Prediction accuracy of five methods on 4 real gene expression datasets (Leukemia, Colon, Lung, DLBCL) with 50% training data, comparing the proposed approach with baseline algorithms.
Figure 5. Prediction accuracy of five methods on 4 real gene expression datasets with 60% training data, comparing the proposed approach with baseline algorithms.
The results further highlight that the integration of multi-task learning and 2 , 1 group sparsity regularization in the proposed strategy not only improves the interpretability of the selected gene subsets but also enhances the overall robustness of the classification models.

4.6. The Influence of Multi-Task Ensemble Learning Strategy

The multi-task ensemble learning strategy is a key component of the proposed method. To further evaluate its effectiveness, we conducted experiments on four datasets. Specifically, we examined the performance of our method under varying numbers of sampling subsets k within the range [ 2 , 4 , 6 , 8 , 10 ] . The average accuracy (ACC) results are presented in Figure 6. From these results, it can be observed that our method achieves the best performance within the range [ 6 , 8 , 10 ] , demonstrating that a sufficient number of sampling subsets enhances the model’s robustness. This further verifies the effectiveness of the proposed multi-task ensemble learning strategy, as it leverages complementary information from multiple subsampled tasks to reduce estimator variance and mitigate overfitting. By aggregating predictions across diverse tasks, the model captures consistent patterns shared among different data partitions, leading to more reliable and generalizable representations. Consequently, the ensemble structure not only stabilizes the learning process but also improves predictive accuracy across heterogeneous gene expression datasets.
Figure 6. Effect of varying the number of sampling subsets ( k = 2 , 4 , 6 , 8 , 10 ) on classification accuracy. Results are averaged over multiple runs on four datasets.

4.7. The Effectiveness of Gene Selection

The primary objective of the proposed method is to identify relevant gene subsets. To evaluate its effectiveness, we report the mean number of selected genes and the mean number of relevant genes identified on the simulated dataset, comparing our method with competing approaches, as summarized in Table 3.
Table 3. Average number of selected genes and, after the slash, the number of those that match the 20 relevant genes embedded in the simulated data. Format: selected/relevant.
The ground truth number of relevant genes is 20. For mRMR and ReliefF, which are filtering-based methods, the number of selected genes must be predefined; in this evaluation, it is set to 100. The results show that, compared with competing methods, our approach effectively identifies most of the relevant genes (close to 20) across all settings while selecting a smaller and more meaningful subset of features. In contrast, mRMR and ReliefF rely solely on pairwise relevance measures between features and labels, which may overlook complex inter-feature dependencies or redundant information. These methods also require a user-specified number of selected genes, making their performance sensitive to this hyperparameter. In comparison, our method integrates feature selection and model learning within a unified optimization framework, enabling the adaptive determination of informative genes without prior knowledge of their number. This joint learning strategy not only reduces manual parameter tuning but also improves interpretability and biological relevance, as the selected genes are directly linked to the model’s predictive objective. Consequently, our method provides a more data-driven and principled approach than traditional filtering techniques such as mRMR and ReliefF.

4.8. Biological Analysis of the Selected Genes

To validate the biological interpretability of our proposed method, we conducted a detailed analysis of the top-ranked genes identified in the experiments. Table 4 summarizes the top five genes selected for the Colon and Leukemia datasets. Notably, all genes identified by our approach have also been reported in previous studies using established gene selection techniques, thereby confirming the biological relevance and robustness of our findings.
Table 4. Top five genes selected by the proposed method on the Colon and Leukemia datasets. All listed genes have been previously reported in the literature and are known to be biologically relevant to the corresponding cancer types.
For the Colon dataset, for instance, the gene H06524 (Gelsolin precursor, plasma) was also reported by [], and T94579 (Human chitotriosidase precursor) was identified by []. Similarly, for the Leukemia dataset, M23197 (CD33 antigen) was also selected by [], further demonstrating the consistency of our results with well-recognized biological evidence.

5. Discussion

The proposed gene-selection strategy is both effective and stable. Our group-sparse multi-task formulation induces shared sparsity across tasks, yielding compact and interpretable panels. In simulation, it correctly recovers most of the 20 ground-truth genes while keeping the selected set comparatively small, and, unlike filter methods, it does not require pre-specifying the subset size.
Despite these strengths, two practical constraints remain—common to many gene-selection frameworks. First, the high-dimensional, low-sample regime can limit generalization even with ensembling. Second, our current evaluation uses only transcriptomic measurements. To address these, in the future, we plans to integrate multi-omic signals (e.g., RNA-seq, copy-number variation, DNA methylation, proteomics) via early/late fusion within the same multi-task objective to better exploit shared structure.

6. Conclusions

This study presents a novel multi-task ensemble strategy for gene selection in gene expression-based tumor classification, aiming to address the challenges posed by high-dimensional data and the demand for both effective and interpretable models. The proposed method integrates gene selection and classification within a unified multi-task logistic regression framework, incorporating 2 , 1 group sparsity regularization to promote consistent and robust gene selection across multiple data subsets. Extensive experiments on both simulated and publicly available gene expression datasets demonstrate that the proposed approach consistently outperforms existing baseline methods in terms of classification accuracy and gene selection stability. By leveraging multi-task learning and structured sparsity, the model enhances both robustness and biological interpretability, offering a promising direction for gene expression analysis and cancer diagnosis.

Author Contributions

Conceptualization, S.L., Z.L., J.Z. and M.-F.L.; methodology, S.L., Z.L. and M.-F.L.; software, S.L. and Z.L.; validation, J.Z. and M.-F.L.; formal analysis, S.L. and Z.L.; investigation, S.L. and Z.L.; resources, J.Z. and M.-F.L.; data curation, S.L.; writing—original draft preparation, S.L. and Z.L.; writing—review and editing, J.Z. and M.-F.L.; visualization, S.L. and Z.L.; supervision, M.-F.L.; project administration, M.-F.L.; funding acquisition, J.Z. and M.-F.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Abdulqader, D.M.; Abdulazeez, A.M.; Zeebaree, D.Q. Machine learning supervised algorithms of gene selection: A review. Mach. Learn. 2020, 62, 233–244. [Google Scholar]
  2. Pashaei, E.; Pashaei, E. Gene selection using hybrid dragonfly black hole algorithm: A case study on RNA-seq COVID-19 data. Anal. Biochem. 2021, 627, 114242. [Google Scholar] [CrossRef]
  3. Perscheid, C.; Grasnick, B.; Uflacker, M. Integrative gene selection on gene expression data: Providing biological context to traditional approaches. J. Integr. Bioinform. 2019, 16, 20180064. [Google Scholar] [CrossRef]
  4. Azadifar, S.; Rostami, M.; Berahmand, K.; Moradi, P.; Oussalah, M. Graph-based relevancy-redundancy gene selection method for cancer diagnosis. Comput. Biol. Med. 2022, 147, 105766. [Google Scholar] [CrossRef]
  5. Alhenawi, E.; Al-Sayyed, R.; Hudaib, A.; Mirjalili, S. Feature selection methods on gene expression microarray data for cancer classification: A systematic review. Comput. Biol. Med. 2022, 140, 105051. [Google Scholar] [CrossRef]
  6. Alomari, O.A.; Makhadmeh, S.N.; Al-Betar, M.A.; Alyasseri, Z.A.A.; Doush, I.A.; Abasi, A.K.; Awadallah, M.A.; Zitar, R.A. Gene selection for microarray data classification based on gray wolf optimizer enhanced with TRIZ-inspired operators. Knowl.-Based Syst. 2021, 223, 107034. [Google Scholar] [CrossRef]
  7. Khurma, R.; Castillo, P.; Sharieh, A.; Aljarah, I. New fitness functions in binary Harris Hawks optimization for gene selection in microarray datasets. In Proceedings of the 12th International Joint Conference on Computer Intelligence, Budapest, Hungary, 2–4 November 2020; pp. 139–146. [Google Scholar]
  8. Sánchez-Maroño, N.; Alonso-Betanzos, A.; Tombilla-Sanromán, M. Filter methods for feature selection: A comparative study. In Proceedings of the International Conference on Intelligent Data Engineering and Automated Learning, Birmingham, UK, 16–19 December 2007; pp. 178–187. [Google Scholar]
  9. Zahoor, J.; Zafar, K. Classification of microarray gene expression data using an infiltration tactics optimization (ITO) algorithm. Genes 2020, 11, 819. [Google Scholar] [CrossRef]
  10. Kononenko, I. Estimating attributes: Analysis and extensions of RELIEF. In Proceedings of the European Conference on Machine Learning, Catania, Italy, 6–8 April 1994; pp. 171–182. [Google Scholar]
  11. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information: Criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef]
  12. Zhang, Y.; Ding, C.; Li, T. Gene selection algorithm by combining reliefF and mRMR. BMC Genom. 2008, 9, S27. [Google Scholar] [CrossRef] [PubMed]
  13. Omuya, E.O.; Okeyo, G.O.; Kimwele, M.W. Feature selection for classification using principal component analysis and information gain. Expert Syst. Appl. 2021, 174, 114765. [Google Scholar] [CrossRef]
  14. Gong, H.; Li, Y.; Zhang, J.; Zhang, B.; Wang, X. A new filter feature selection algorithm for classification task by ensembling pearson correlation coefficient and mutual information. Eng. Appl. Artif. Intell. 2024, 131, 107865. [Google Scholar] [CrossRef]
  15. Ali, L.; Leung, M.-F.; Khan, M.A.; Nour, R.; Imrana, Y.; Vasilakos, A.V. Chiga-Net: A genetically optimized neural network with refined deeply extracted features using χ2 statistical score for trustworthy Parkinson’s disease detection. Neurocomputing 2025, 624, 129450. [Google Scholar] [CrossRef]
  16. El Aboudi, N.; Benhlima, L. Review on wrapper feature selection approaches. In Proceedings of the 2016 International Conference on Engineering & MIS (ICEMIS), Agadir, Morocco, 22–24 September 2016; pp. 1–5. [Google Scholar]
  17. Ghosh, M.; Begum, S.; Sarkar, R.; Chakraborty, D.; Maulik, U. Recursive memetic algorithm for gene selection in microarray data. Expert Syst. Appl. 2019, 116, 172–185. [Google Scholar] [CrossRef]
  18. Sahebi, G.; Movahedi, P.; Ebrahimi, M.; Pahikkala, T.; Plosila, J.; Tenhunen, H. GeFeS: A generalized wrapper feature selection approach for optimizing classification performance. Comput. Biol. Med. 2020, 125, 103974. [Google Scholar] [CrossRef]
  19. Hassan, A.; Paik, J.H.; Khare, S.R.; Hassan, S.A. A wrapper feature selection approach using Markov blankets. Pattern Recognit. 2025, 158, 111069. [Google Scholar] [CrossRef]
  20. Sağbaş, E.A. A novel two-stage wrapper feature selection approach based on greedy search for text sentiment classification. Neurocomputing 2024, 590, 127729. [Google Scholar] [CrossRef]
  21. Sun, L.; Kong, X.; Xu, J.; Zhai, Z.; Xue, R.; Zhang, S. A hybrid gene selection method based on ReliefF and ant colony optimization algorithm for tumor classification. Sci. Rep. 2019, 9, 8978. [Google Scholar] [CrossRef]
  22. Hu, P.; Zhu, J. A filter-wrapper model for high-dimensional feature selection based on evolutionary computation. Appl. Intell. 2025, 55, 581. [Google Scholar] [CrossRef]
  23. Liang, Y.; Liu, C.; Luan, X.Z.; Leung, K.S.; Chan, T.M.; Xu, Z.; Zhang, H. Sparse logistic regression with a L1/2 penalty for gene selection in cancer classification. BMC Bioinform. 2013, 14, 198. [Google Scholar] [CrossRef] [PubMed]
  24. Li, C.; Li, H. Network-constrained regularization and variable selection for analysis of genomic data. Bioinformatics 2008, 24, 1175–1182. [Google Scholar] [CrossRef]
  25. Algamal, Z.Y.; Lee, M.H. Regularized logistic regression with adjusted adaptive elastic net for gene selection in high dimensional cancer classification. Comput. Biol. Med. 2015, 67, 136–145. [Google Scholar] [CrossRef] [PubMed]
  26. Liu, C.; Liang, Y.; Luan, X.Z.; Leung, K.S.; Chan, T.M.; Xu, Z.; Zhang, H. The L1/2 regularization method for variable selection in the Cox model. Appl. Soft Comput. 2014, 14, 498–503. [Google Scholar] [CrossRef]
  27. Saberi-Movahed, F.; Rostami, M.; Berahmand, K.; Karami, S.; Tiwari, P.; Oussalah, M.; Band, S.S. Dual regularized unsupervised feature selection based on matrix factorization and minimum redundancy with application in gene selection. Knowl.-Based Syst. 2022, 256, 109884. [Google Scholar] [CrossRef]
  28. Wang, Y.; Yang, X.-G.; Lu, Y. Informative gene selection for microarray classification via adaptive elastic net with conditional mutual information. Appl. Math. Model. 2019, 71, 286–297. [Google Scholar] [CrossRef]
  29. Cawley, G.C.; Talbot, N.L.C. Gene selection in cancer classification using sparse logistic regression with Bayesian regularization. Bioinformatics 2006, 22, 2348–2355. [Google Scholar] [CrossRef]
  30. Xing, Z.; Zhu, W.; Tang, C.; Wang, M. Gene selection for microarray data classification via adaptive hypergraph embedded dictionary learning. Gene 2019, 706, 188–200. [Google Scholar] [CrossRef]
  31. Liu, C.; Wong, H.S. Structured penalized logistic regression for gene selection in gene expression data analysis. IEEE/ACM Trans. Comput. Biol. Bioinform. 2017, 16, 312–321. [Google Scholar] [CrossRef]
  32. Sharma, A.; Lysenko, A.; Boroevich, K.A.; Vans, E.; Tsunoda, T. DeepFeature: Feature selection in nonimage data using convolutional neural network. Brief. Bioinform. 2021, 22, bbab297. [Google Scholar] [CrossRef]
  33. Zeebaree, D.Q.; Haron, H.; Abdulazeez, A.M.; Zebari, D.A. Trainable model based on new uniform LBP feature to identify the risk of the breast cancer. In Proceedings of the International Conference on Advanced Science and Engineering (ICOASE), Zakho, Iraq, 2–4 April 2019; pp. 106–111. [Google Scholar]
  34. Eluri, N.R.; Kancharla, G.R.; Dara, S.; Dondeti, V. Cancer data classification by quantum-inspired immune clone optimization-based optimal feature selection using gene expression data: Deep learning approach. Data Technol. Appl. 2022, 56, 247–282. [Google Scholar] [CrossRef]
  35. Rukhsar, L.; Bangyal, W.H.; Khan, M.S.A.; Ibrahim, A.A.A.; Nisar, K.; Rawat, D.B. Analyzing RNA-seq gene expression data using deep learning approaches for cancer classification. Appl. Sci. 2022, 12, 1850. [Google Scholar] [CrossRef]
  36. Zhang, S.; Xie, W.; Li, W.; Wang, L.; Feng, C. GAMB-GNN: Graph neural networks learning from gene structure relations and Markov blanket ranking for cancer classification in microarray data. Chemom. Intell. Lab. Syst. 2023, 232, 104713. [Google Scholar] [CrossRef]
  37. Hossain, M.S.; Shorfuzzaman, M. Early Diabetic Retinopathy Cyber-Physical Detection System Using Attention-Guided Deep CNN Fusion. IEEE Trans. Netw. Sci. Eng. 2025, 12, 1898–1910. [Google Scholar] [CrossRef]
  38. Song, Y.; Liu, Y.; Lin, Z.; Zhou, J.; Li, D.; Zhou, T.; Leung, M.-F. Learning from AI-generated annotations for medical image segmentation. IEEE Trans. Consum. Electron. 2025, 71, 1473–1481. [Google Scholar] [CrossRef]
  39. Li, Y.; Hao, W.; Zeng, H.; Wang, L.; Xu, J.; Routray, S.; Jhaveri, R.H.; Gadekallu, T.R. Cross-Scale Texture Supplementation for Reference-based Medical Image Super-Resolution. IEEE J. Biomed. Health Inform. (Early Access) 2021, 1–15. [Google Scholar] [CrossRef]
  40. Wang, L.; Cai, J.; Wang, T.; Zhao, J.; Gadekallu, T.R.; Fang, K. Pine wilt disease detection based on uav remote sensing with an improved yolo model. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 19230–19242. [Google Scholar] [CrossRef]
  41. Zhang, X.; Feng, H.; Hossain, M.S.; Chen, Y.; Wang, H.; Yin, Y. Scaled Background Swap: Video Augmentation for Action Quality Assessment with Background Debiasing. ACM Trans. Multimed. Comput. Commun. Appl. 2025, 21, 1–18. [Google Scholar] [CrossRef]
  42. Li, G.; Wang, Z.; Sun, J.; Zhang, Q. Objective Extraction for Simplifying Many-Objective Solution Sets. IEEE Trans. Emerg. Top. Comput. Intell. 2024, 8, 337–349. [Google Scholar] [CrossRef]
  43. Li, G.; Zhang, Q. Multitask Feature Selection for Objective Reduction. In Proceedings of the International Conference on Evolutionary Multi-Criterion Optimization, Shenzhen, China, 28–31 March 2021; Lecture Notes in Computer Science. Springer: Cham, Switzerland, 2021; Volume 12654, pp. 77–88. [Google Scholar]
  44. Nagra, A.A.; Khan, A.H.; Abubakar, M.; Faheem, M.; Rasool, A.; Masood, K.; Hussain, M. A gene selection algorithm for microarray cancer classification using an improved particle swarm optimization. Sci. Rep. 2024, 14, 19613. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  45. Huang, T.M.; Kecman, V. Gene extraction for cancer diagnosis by support vector machines. Int. Conf. Artif. Neural Netw. 2005, 35, 185–194. [Google Scholar]
  46. Han, F.; Tang, D.; Sun, Y.W.T.; Cheng, Z.; Jiang, J.; Li, Q.W. A hybrid gene selection method based on gene scoring strategy and improved particle swarm optimization. BMC Bioinform. 2019, 20 (Suppl. S8), 289. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.