An Adaptive Feature-Based Quantum Genetic Algorithm for Dimension Reduction with Applications in Outlier Detection †
Abstract
:1. Introduction
2. Related Work
3. Materials and Method
3.1. Genetic Algorithms
- Elitism Selection: this method selects the top k chromosomes based on their fitness scores and passes them directly to the next generation without modification.
- Roulette Wheel Selection: in this approach, each chromosome is assigned a portion of a wheel proportional to its fitness value. Chromosomes with higher fitness occupy larger segments, increasing their probability of being selected as parents.
- Generational Replacement: the entire population is replaced by offspring.
- Elitism: the best individuals from the parent generation are carried over to the next generation to preserve good solutions.
- Steady-state: offspring are evaluated, then merged with the parent generation and sorted. Finally, less fit individuals are removed.
3.2. Quantum Genetic Algorithms
- Initialization: A quantum chromosome population is randomly generated. at generation t, where n is the size of population, and is a Q-bit individual defined as Equation (4).The number of features is represented by m. After that, in the act of observation, n Q-bit individuals are converted into (or collapsed) n binary individuals (). One binary solution; ; is a binary string of the length m, and is formed by selecting each bit using the probability of q-bit following the rules below Equation (5).The fitness function, also known as the evaluation function, is designed to assess the quality of each quantum feature subset. In feature selection problems, whether a lower or higher fitness value indicates a more optimal solution depends on the objective function. If the objective function is a minimization problem (e.g., minimizing error or redundancy as in Equation (6)), a lower fitness value is preferable. Since fitness functions are crucial for evaluating candidate solutions during optimization, various methods have been proposed in the literature. These include evaluation metrics such as accuracy, F1 score, precision, and recall, as well as information-theoretic measures, error or loss functions, and linear combinations of subset cardinality with evaluation or loss metrics. An example of a fitness function is given in Equation (6).Here, acts as a weighting factor for relative to the size of the feature subset. It is designed to identify a feature subset that minimizes both the classification error () and the number of selected features. The parameter ranges between 0 and 1, where represents the selected feature subset, and denotes the total number of features.
- Rotation: In QGA, the update of Q-bit individuals is performed through a quantum rotation. At generation , all individuals adjust their positions by rotating toward the best individual from the previous generation . A rotation gate is utilized to modify a Q-bit individual. For instance, at position is updated to at , as expressed in Equation (7) and illustrated in Figure 5.The rotation angle governs the adjustment of the qubit’s state, influencing its angular approximation toward the |0〉 or |1〉 amplitudes. The superposition state of the qubit, defined by and , dictates the probability of measuring the qubit in either the 0 or 1 state. The rotation angle is dynamically modified based on the fitness function of the quantum individuals. Through iterative updates of these rotation angles, the algorithm explores the quantum state space in search of optimal solutions.
4. The Proposed Method
- (A)
- Initialization: This step remains the same as in the classic QGA, where a population of Q-bit individuals is randomly generated and then converted into binary individuals. We use an SVM classifier to compute the fitness values.
- (B)
- Adaptive Rotation: Typically, a quantum rotation operation involves two key components: the rotation direction and the rotation angle. The details of each component are provided below in this paper.
- Rotation direction: A quantum individual undergoes rotation only when the following conditions are met: the quantum individual is inferior to the global best solution , and the corresponding binary digit of the individual differs from . This mechanism is illustrated in Figure 7. Here, f represents the fitness function, is the binary value of the feature of the individual, and refers to the binary value of the global best solution.
- Rotation angle: we propose a novel adaptive rotation angle that calculates the magnitude of the rotation angle for each Q-bit of each quantum individual in the Equation (8)Here, represents the rotation angle of the Q-bit in the Q-bit individual at generation . is the minimum rotation angle, set to . The population size and the total number of features are denoted by n and m, respectively.The feature frequency vector, , is shown in Figure 8 and is calculated as the sum of all binary individuals. Specifically, represents the frequency of the feature appearing in the binary population at the previous generation . The term refers to the average fitness of the population at the generation.The core concept is that different features have different levels of importance. To evaluate their contributions, we measure the frequency of their occurrence in the population and examine how this frequency influences the population’s fitness. To avoid bias due to variations in population size (n) and the total number of features (m), we normalize the results using these factors. As the population’s fitness improves or worsens, the rotation angle adjusts accordingly, increasing or decreasing in response. Similarly, when the frequency of a feature changes, the rotation angle for that feature adjusts proportionally to the change in fitness.
- (C)
- Adaptive Mutation: We employ the adaptive mutation rate based on the population diversity [20]. The population diversity is measured through the standard deviation of the population’s fitness values, as described in Equation (9).Here, represents the average fitness of the generation, and is the fitness value of individual in the generation.As the standard deviation decreases, the mutation rate is increased to promote greater population diversity. The adaptive mutation rate is given by Equation (10).In this equation, is a fixed mutation rate, and is the highest fitness in the generation. The adaptive mutation rate is influenced by both the fixed mutation rate and the population’s diversity. As the population diversity decreases, will increase proportionally.
5. Experiment Design
5.1. Dataset and Data Preprocessing
- The Taiwan bankruptcy dataset were collected from the Taiwan Economic Journal for the years 1999 to 2009. Company bankruptcy was defined based on the business regulations of the Taiwan Stock Exchange.
- The Musk dataset describes a set of 102 molecules of which 39 are judged by human experts to be musks and the remaining 63 molecules are judged to be non-musks. The 166 features that describe these molecules depend upon the exact shape, or conformation, of the molecule. The label is either 0 if the record is non-musk and 1 if the record is musk.
- The Parkinson’s Disease (PD) dataset is composed of a range of biomedical voice measurements. The main aim of the data is to discriminate healthy people from those with PD, according to “status” column which is set to 0 for healthy and 1 for PD.
- The Backdoor dataset has a hybrid of the real modern normal and the contemporary synthesized attack activities of the network traffic. It contains nine attack categories and normal network traffic. There are 49 features in the original UNSW-NB15 dataset developed by Moustafa and Slay [23]. Han et al. [22] synthetically create the Backdoor dataset based on the UNSW-NB15 dataset by adding more anomalies and irrelevant features up to more than 50% of the total input features. They follow and enrich the approach in [24] to generate realistic synthetic anomalies.
- The Campaign dataset is related with direct marketing campaigns (phone calls) of a Portuguese banking institution. The classification goal is to predict if the client will subscribe a term deposit.
Dataset | # Features | # Samples | # Class | Domain | Outlier Rate |
---|---|---|---|---|---|
Taiwan bankruptcy [25] | 96 | 6819 | 2 | Finance | 3.23% |
Musk [26] | 167 | 3062 | 2 | Chemistry | 3.17% |
Parkinson’s Disease [27] | 756 | 754 | 2 | Healthcare | 25.39% |
Backdoor [22] | 196 | 95,329 | 10 | Network | 2.44% |
Campaign [22,28] | 62 | 41,188 | 2 | Finance | 11.27% |
5.2. Classifier
5.3. Evaluation Metrics
- Classification Accuracy: This metric determines the percentage of accurately classified instances relative to the total number of classifications. It is calculated as:
- Precision (P): This refers to the ratio of correctly identified outlier instances to the total number of instances predicted as outliers:
- Recall (R): This is often referred to as sensitivity, which measures the proportion of actual outlier instances correctly identified as outliers. It is computed by dividing the number of true positive predictions (TP) by the total number of actual outlier instances:
- F1-Score (F-Measure): This is the mean of the precision and recall, defined as:
- G-Mean: To evaluate the performance of a classification algorithm when dealing with imbalanced data, sensitivity and specificity can be combined into a single metric. This metric, called the geometric mean (G-Mean), provides a balanced measure of both factors. It is calculated as follows:
5.4. Comparable Methods
5.4.1. Fixed Value Rotation Angle
5.4.2. Generation-Based Rotation Angle
5.4.3. Individual-Based Rotation Angle
5.5. Parameters
6. Results
6.1. Performance Analysis
6.2. Feature Reduction Analysis
- Taiwan Bankruptcy Dataset: With 96 features, FbQGA (31.52 features) demonstrated the most efficient reduction, followed closely by GbQGA and IbQGA, which also reduced dimensionality effectively. GA, in contrast, selected 41.10 features, leaving more irrelevant features and highlighting its limitations in reducing dimensions for datasets with relatively fewer features.
- Musk Dataset: Among 167 features, FbQGA (22.13 features) again achieved the most substantial reduction, outperforming the other quantum-inspired variants and significantly outpacing GA (61.35 features). This dataset’s moderate dimensionality underscores FbQGA’s adaptability, where dynamic adjustments to feature importance contribute to efficient reduction.
- Parkinson’s Disease Dataset: With 754 features, this dataset posed a significant challenge. FbQGA (236.91 features) emerged as the most effective algorithm, selecting the fewest features and maintaining consistency. Other quantum-inspired algorithms, including GbQGA and IbQGA, performed well but selected slightly more features than FbQGA. GA (367.20 features) struggled, retaining nearly half of the original features.
- Backdoor dataset, FbQGA selects the smallest number of features (13), significantly reducing dimensionality compared to GA (31) and even outperforming other quantum genetic algorithms such as IbQGA (17), GbQGA (19), and FvQGA (24). This suggests that FbQGA is the most efficient in feature selection, minimizing redundancy while maintaining useful information. Given its lower recall in the classification results, this aggressive reduction may contribute to missing some relevant fraud indicators, indicating a trade-off between feature selection and model sensitivity.
- Campaign dataset, FbQGA again selects the fewest features (11), demonstrating its strong feature reduction capability compared to GA (18), FvQGA (16), GbQGA (15), and IbQGA (15). This suggests that FbQGA efficiently filters out irrelevant or redundant features while preserving those most critical for classification. Since FbQGA also achieved the highest accuracy in this dataset, the reduced feature set appears to be well-optimized, helping to enhance model performance without sacrificing predictive power.
- Taiwan Bankruptcy Dataset: GA had the highest processing time (50.93 s), while FvQGA (31.92 s) and FbQGA (33.15 s) were the fastest. The quantum-inspired algorithms demonstrated efficiency, with FbQGA striking a balance between speed and feature selection effectiveness.
- Musk Dataset: GA required the longest processing time (347.33 s), reflecting its inefficiency in handling moderate dimensionality. FvQGA (208.46 s) was the fastest but lacked the feature selection precision of FbQGA (215.46 s), which offered a superior trade-off between processing time and dimensionality reduction.
- Parkinson’s Disease Dataset: For this highly complex dataset, GA (143.88 s) was the slowest, while FbQGA (80.12 s) and GbQGA (80.01 s) were the most efficient. Despite their similar times, FbQGA demonstrated better feature selection, making it the optimal choice for extreme dimensionality.
- Backdoor dataset, the FbQGA algorithm has an average running time of 212 s, making it one of the more efficient methods compared to GA (564 s) and FvQGA (413 s). While it is slightly slower than GbQGA (199 s) and IbQGA (232 s), its performance remains competitive within the quantum-inspired algorithms. The error bars indicate moderate variability, suggesting consistent efficiency in solving this dataset. FbQGA’s significant reduction in runtime compared to GA highlights its advantage in computational efficiency.
- Campaign dataset, FbQGA has an average running time of 93 s, which is nearly identical to GbQGA (92 s) and IbQGA (89 s). Compared to GA (151 s), FbQGA shows a significant improvement, cutting execution time by almost 40%. The error bars suggest minimal variation, reinforcing FbQGA’s reliability in handling this dataset efficiently. This consistency across multiple datasets indicates that FbQGA is a strong candidate for reducing computational costs while maintaining performance.
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
Abbreviations
GA | Genetic Algorithms |
QGA | Quantum Genetic Algorithm |
FbQGA | Adaptive Feature-based Quantum Genetic Algorithm |
FvQGA | Fixed value Quantum Genetic Algorithm |
GbQGA | Generation-based Quantum Genetic Algorithm |
IbQGA | Individual-based Quantum Genetic Algorithm |
References
- L’Heureux, A.; Grolinger, K.; Elyamany, H.F.; Capretz, M.A.M. Machine Learning with Big Data: Challenges and Approaches. IEEE Access 2017, 5, 7776–7797. [Google Scholar] [CrossRef]
- Mandal, A.K.; Chakraborty, B. Quantum computing and quantum-inspired techniques for feature subset selection: A review. Knowl. Inf. Syst. 2024, 67, 2019–2061. [Google Scholar] [CrossRef]
- Pudjihartono, N.; Fadason, T.; Kempa-Liehr, A.W.; O’Sullivan, J.M. A Review of Feature Selection Methods for Machine Learning-Based Disease Risk Prediction. Front. Bioinform. 2022, 2, 927312. [Google Scholar] [CrossRef] [PubMed]
- Rao, H.; Shi, X.; Rodrigue, A.K.; Feng, J.; Xia, Y.; Elhoseny, M.; Yuan, X.; Gu, L. Feature selection based on artificial bee colony and gradient boosting decision tree. Appl. Soft Comput. 2019, 74, 634–642. [Google Scholar] [CrossRef]
- van Zyl, C.; Ye, X.; Naidoo, R. Harnessing eXplainable artificial intelligence for feature selection in time series energy forecasting: A comparative analysis of Grad-CAM and SHAP. Appl. Energy 2024, 353, 122079. [Google Scholar] [CrossRef]
- Li, J.; Tang, J.; Liu, H. Reconstruction-based Unsupervised Feature Selection: An Embedded Approach. In Proceedings of the 26th International Joint Conference on Artificial Intelligence, Melbourne, VIC, Australia, 19–25 August 2017; pp. 2159–2165. [Google Scholar] [CrossRef]
- Afzal, W.; Torkar, R. On the application of genetic programming for software engineering predictive modeling: A systematic review. Expert Syst. Appl. 2011, 38, 11984–11997. [Google Scholar] [CrossRef]
- Brezočnik, L.; Fister, I.; Podgorelec, V. Swarm Intelligence Algorithms for Feature Selection: A Review. Appl. Sci. 2018, 8, 1521. [Google Scholar] [CrossRef]
- Xue, B.; Zhang, M.; Browne, W.N.; Yao, X. A Survey on Evolutionary Computation Approaches to Feature Selection. IEEE Trans. Evol. Comput. 2016, 20, 606–626. [Google Scholar] [CrossRef]
- Holland, J.H. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence; MIT Press: Cambridge, MA, USA, 1992. [Google Scholar]
- Han, K.H.; Kim, J.H. Genetic quantum algorithm and its application to combinatorial optimization problem. In Proceedings of the 2000 Congress on Evolutionary Computation. CEC00 (Cat. No.00TH8512), La Jolla, CA, USA, 16–19 July 2000; Volume 2, pp. 1354–1360. [Google Scholar] [CrossRef]
- Han, K.H.; Kim, J.H. Quantum-inspired evolutionary algorithm for a class of combinatorial optimization. IEEE Trans. Evol. Comput. 2002, 6, 580–593. [Google Scholar] [CrossRef]
- Han, K.H.; Park, K.H.; Lee, C.H.; Kim, J.H. Parallel quantum-inspired genetic algorithm for combinatorial optimization problem. In Proceedings of the 2001 Congress on Evolutionary Computation (IEEE Cat. No.01TH8546), Seoul, Republic of Korea, 27–30 May 2001; Volume 2, pp. 1422–1429. [Google Scholar] [CrossRef]
- Ramos, A.C.; Vellasco, M. Quantum-inspired Evolutionary Algorithm for Feature Selection in Motor Imagery EEG Classification. In Proceedings of the 2018 IEEE Congress on Evolutionary Computation (CEC), Rio de Janeiro, Brazil, 8–13 July 2018; pp. 1–8. [Google Scholar] [CrossRef]
- Ramos, A.C.; Vellasco, M. Chaotic Quantum-inspired Evolutionary Algorithm: Enhancing feature selection in BCI. In Proceedings of the 2020 IEEE Congress on Evolutionary Computation (CEC), Glasgow, UK, 19–24 July 2020; pp. 1–8. [Google Scholar] [CrossRef]
- Tayarani, M.; Esposito, A.; Vinciarelli, A. What an “Ehm” Leaks About You: Mapping Fillers into Personality Traits with Quantum Evolutionary Feature Selection Algorithms. IEEE Trans. Affect. Comput. 2022, 13, 108–121. [Google Scholar] [CrossRef]
- Kaur, B.; Rathi, S.; Agrawal, R. Enhanced depression detection from speech using Quantum Whale Optimization Algorithm for feature selection. Comput. Biol. Med. 2022, 150, 106122. [Google Scholar] [CrossRef] [PubMed]
- Agrawal, R.; Kaur, B.; Sharma, S. Quantum based Whale Optimization Algorithm for wrapper feature selection. Appl. Soft Comput. J. 2020, 89, 106092. [Google Scholar] [CrossRef]
- Dabba, A.; Tari, A.; Meftali, S. Hybridization of Moth flame optimization algorithm and quantum computing for gene selection in microarray data. J. Ambient. Intell. Humaniz. Comput. 2021, 12, 2731–2750. [Google Scholar] [CrossRef]
- Pham, T.H.; Raahemi, B. An Adaptive Quantum-based Genetic Algorithm Feature Selection for Outlier Detection. In Proceedings of the 10th International Conference on Control, Decision and Information Technologies (CoDIT), Vallette, Malta, 1–4 July 2024; pp. 1034–1039, ISSN 2576-3555. [Google Scholar] [CrossRef]
- Kelly, M.; Longjohn, R.; Nottingham, K. The UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu (accessed on 1 December 2024).
- Han, S.; Hu, X.; Huang, H.; Jiang, M.; Zhao, Y. ADBench: Anomaly Detection Benchmark. arXiv 2022, arXiv:2206.09426. [Google Scholar] [CrossRef]
- Moustafa, N.; Slay, J. UNSW-NB15: A comprehensive data set for network intrusion detection systems (UNSW-NB15 network data set). In Proceedings of the 2015 Military Communications and Information Systems Conference (MilCIS), Canberra, ACT, Australia, 10–12 November 2015; pp. 1–6. [Google Scholar] [CrossRef]
- Steinbuss, G.; Böhm, K. Benchmarking Unsupervised Outlier Detection with Realistic Synthetic Data. ACM Trans. Knowl. Discov. Data 2021, 15, 1–20. [Google Scholar] [CrossRef]
- Taiwanese Bankruptcy Prediction. UCI Machine Learning Repository. 2020. Available online: https://doi.org/10.24432/C5004D (accessed on 1 December 2024).
- Chapman, D.; Jain, A. Musk (Version 1). UCI Machine Learning Repository. 1994. Available online: https://doi.org/10.24432/C5ZK5B (accessed on 1 December 2024).
- Sakar, C.; Serbes, G.; Gunduz, A.; Nizam, H.; Sakar, B. Parkinson’s Disease Classification. UCI Machine Learning Repository. 2018. Available online: https://doi.org/10.24432/C5MS4X (accessed on 1 December 2024).
- Pang, G.; Shen, C.; van den Hengel, A. Deep Anomaly Detection with Deviation Networks. arXiv 2019, arXiv:1911.08623. [Google Scholar]
- Zhang, J.; Mani, I. KNN Approach to Unbalanced Data Distributions: A Case Study Involving Information Extraction. In Proceedings of the ICML’2003 Workshop on Learning from Imbalanced Datasets, Washington, DC, USA, 21 August 2003. [Google Scholar]
- Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
- Xiong, H.; Wu, Z.; Fan, H.; Li, G.; Jiang, G. Quantum rotation gate in quantum-inspired evolutionary algorithm: A review, analysis and comparison study. Swarm Evol. Comput. 2018, 42, 43–57. [Google Scholar] [CrossRef]
- Wang, Z.; Yang, B.; Lv, X.; Cui, D. An Improved Quantum Genetic Algorithm. J. Xi’an Univ. Technol. 2012, 28, 145–151. [Google Scholar]
- Wilcoxon, F. Individual Comparisons by Ranking Methods. Biom. Bull. 1945, 1, 80–83. [Google Scholar] [CrossRef]
Algorithm | Acronym | Description |
---|---|---|
Genetic Algorithms | GA | The basic Genetic Algorithm with the steps described in Section 3.1 |
Fixed value Quantum Genetic Algorithm | FvQGA | The Quantum Genetic Algorithm proposed by Han and Kim [11,12] that employs a fixed value for the quantum rotation angle. |
Generation-based Quantum Genetic Algorithm | GbQGA | The variant of Quantum Genetic Algorithm which has the quantum rotation angle decreases proportionately to the generation. |
Individual-based Quantum Genetic Algorithm | IbQGA | The variant of Quantum Genetic Algorithm where each individual develops its quantum rotation angle based on its distance to the global best solution. |
Adaptive Feature-based Quantum Genetic Algorithm | FbQGA | Our proposed method, details in Section 4, that formulates the quantum rotation angle for each feature based on its level of contribution. |
Dataset | Subset | # Samples | Outlier Rate |
---|---|---|---|
Taiwan Bankruptcy | Train | 356 | 50% |
Validation | 614 | 3.25% | |
Test | 682 | 3.22% | |
Musk | Train | 156 | 50% |
Validation | 276 | 3.26% | |
Test | 307 | 3.25% | |
Parkinson’s Disease | Train | 312 | 50% |
Validation | 68 | 25% | |
Test | 76 | 25% | |
Backdoor | Train | 1674 | 50% |
Validation | 851 | 2.46% | |
Test | 858 | 2.44% | |
Campaign | Train | 732 | 50% |
Validation | 408 | 11.27% | |
Test | 412 | 11.16% |
Parameter | All Variants of QGA | GA |
---|---|---|
Number of generations | 100 | 100 |
Number of independent runs | 30 | 30 |
Population size | 10 | 10 |
0.99 | 0.99 | |
1 − F1 | 1 − F1 | |
Binary conversion threshold | Random as in Equation (5) | 0.5 |
(fixed mutation rate) | 0.01 | 0.01 |
Crossover rate | - | 0.8 |
Crossover strategy | - | Random-point crossover |
Selection method | - | Roulette wheel |
Next generation | - | Steady-state |
Dataset | Evaluation Metric | GA | GA std | FbQGA | FbQGA std | SSD |
---|---|---|---|---|---|---|
Taiwan Bankruptcy | F1 | 0.5242 | 4.26% | 0.5886 | 1.17% | yes |
Precision | 0.511 | 1.79% | 0.5663 | 1.23% | yes | |
Recall | 0.5381 | 3.25% | 0.6127 | 3.06% | yes | |
Accuracy | 0.8489 | 4.98% | 0.8961 | 4.50% | yes | |
Avg. No. of Selected Features | 41.10 | 11.28% | 31.52 | 23.12% | yes | |
Running Time (s) | 50.92 | 7.82% | 33.15 | 15.90% | yes | |
Musk | F1 | 0.6096 | 11.51% | 0.6965 | 5.48% | yes |
Precision | 0.5795 | 5.48% | 0.654 | 4.15% | yes | |
Recall | 0.6431 | 7.21% | 0.7031 | 6.37% | yes | |
Accuracy | 0.6053 | 13.37% | 0.7142 | 6.37% | yes | |
Avg. No. of Selected Features | 61.35 | 14.60% | 22.13 | 24.97% | yes | |
Running Time (s) | 347.33 | 3.16% | 215.46 | 15.83% | yes | |
Parkinson’s Disease | F1 | 0.6553 | 2.66% | 0.7031 | 2.74% | yes |
Precision | 0.6481 | 2.56% | 0.6872 | 2.34% | yes | |
Recall | 0.6627 | 2.79% | 0.7731 | 2.72% | yes | |
Accuracy | 0.7269 | 1.89% | 0.745 | 2.67% | yes | |
Avg. No. of Selected Features | 367.20 | 2.96% | 236.91 | 25.24% | yes | |
Running Time (s) | 143.87 | 2.53% | 80.11 | 16.05% | yes | |
Backdoor | F1 | 0.7918 | 3.13% | 0.8422 | 1.94% | yes |
Precision | 0.9105 | 4.33% | 0.9205 | 5.10% | no | |
Recall | 0.7005 | 4.48% | 0.7762 | 2.08% | yes | |
Accuracy | 0.9462 | 1.62% | 0.9547 | 2.20% | no | |
Avg. No. of Selected Features | 31.00 | 8.71% | 12.50 | 3.78% | yes | |
Running Time (s) | 563.83 | 15.61% | 212.35 | 23.96% | yes | |
Campaign | F1 | 0.6568 | 2.27% | 0.6934 | 1.96% | yes |
Precision | 0.6418 | 2.26% | 0.6826 | 1.63% | yes | |
Recall | 0.6861 | 2.84% | 0.7046 | 3.18% | no | |
Accuracy | 0.8408 | 1.83% | 0.8733 | 1.20% | yes | |
Avg. No. of Selected Features | 18.00 | 4.71% | 10.70 | 4.32% | yes | |
Running Time (s) | 150.94 | 6.96% | 92.83 | 16.06% | yes |
Dataset | Evaluation Metric | FvQGA | FvQGA std | FbQGA | FbQGA std | SSD |
---|---|---|---|---|---|---|
Taiwan Bankruptcy | F1 | 0.5598 | 2.17% | 0.5886 | 1.17% | yes |
Precision | 0.539 | 1.23% | 0.5663 | 1.32% | yes | |
Recall | 0.5823 | 3.06% | 0.6127 | 3.06% | yes | |
Accuracy | 0.8761 | 3.20% | 0.8961 | 4.50% | no | |
Avg. No. of Selected Features | 33.20 | 27.07% | 31.52 | 23.12% | no | |
Running Time (s) | 31.9178 | 16.41% | 33.1522 | 15.90% | yes | |
Musk | F1 | 0.6392 | 7.48% | 0.6965 | 5.48% | yes |
Precision | 0.6011 | 4.10% | 0.654 | 4.15% | yes | |
Recall | 0.6824 | 6.13% | 0.745 | 6.13% | yes | |
Accuracy | 0.6604 | 8.37% | 0.712 | 6.37% | no | |
Avg. No. of Selected Features | 28.9 | 18.11% | 22.13 | 24.97% | yes | |
Running Time (s) | 208.4605 | 2.95% | 215.4651 | 15.83% | no | |
Parkinson’s Disease | F1 | 0.6756 | 2.95% | 0.6822 | 2.72% | yes |
Precision | 0.6672 | 2.61% | 0.6872 | 2.34% | yes | |
Recall | 0.685 | 2.89% | 0.7031 | 2.61% | yes | |
Accuracy | 0.723 | 2.44% | 0.7731 | 2.74% | yes | |
Avg. No. of Selected Features | 255.25 | 18.18% | 236.91 | 25.24% | yes | |
Running Time (s) | 77.1235 | 12.18% | 80.1122 | 16.05% | yes | |
Backdoor | F1 | 0.8523 | 4.55% | 0.8422 | 1.94% | no |
Precision | 0.9305 | 4.02% | 0.9205 | 5.10% | no | |
Recall | 0.7862 | 4.10% | 0.7762 | 2.08% | no | |
Accuracy | 0.9547 | 2.57% | 0.9423 | 2.20% | no | |
Avg. No. of Selected Features | 24.10 | 5.05% | 12.50 | 3.78% | yes | |
Running Time (s) | 412.68 | 23.21% | 212.35 | 23.96% | yes | |
Campaign | F1 | 0.6611 | 3.08% | 0.6934 | 1.96% | yes |
Precision | 0.6588 | 3.98% | 0.6826 | 1.63% | no | |
Recall | 0.6715 | 2.43% | 0.7046 | 3.18% | yes | |
Accuracy | 0.8340 | 2.90% | 0.8733 | 1.20% | no | |
Avg. No. of Selected Features | 15.90 | 3.73% | 10.70 | 4.32% | yes | |
Running Time (s) | 87.88 | 15.56% | 92.83 | 16.06% | no |
Dataset | Evaluation Metric | GbQGA | GbQGA std | FbQGA | FbQGA std | SSD |
---|---|---|---|---|---|---|
Taiwan Bankruptcy | F1 | 0.5684 | 2.13% | 0.5886 | 1.17% | yes |
Precision | 0.549 | 1.33% | 0.5663 | 1.32% | yes | |
Recall | 0.5893 | 4.90% | 0.6127 | 3.06% | yes | |
Accuracy | 0.8811 | 3.15% | 0.8961 | 4.50% | no | |
Avg. No. of Selected Features | 31.1 | 24.07% | 31.52 | 23.12% | no | |
Running Time (s) | 32.9178 | 15.98% | 33.1522 | 15.90% | no | |
Musk | F1 | 0.6756 | 4.48% | 0.6965 | 5.48% | yes |
Precision | 0.6331 | 4.10% | 0.654 | 4.15% | yes | |
Recall | 0.7242 | 6.13% | 0.745 | 6.13% | no | |
Accuracy | 0.6912 | 7.19% | 0.712 | 6.37% | no | |
Avg. No. of Selected Features | 24.3 | 25.70% | 22.13 | 24.97% | yes | |
Running Time (s) | 213.4605 | 14.59% | 215.4651 | 15.83% | no | |
Parkinson’s Disease | F1 | 0.6759 | 2.70% | 0.6822 | 2.72% | yes |
Precision | 0.6673 | 2.83% | 0.6872 | 2.34% | yes | |
Recall | 0.6855 | 2.89% | 0.7031 | 2.61% | yes | |
Accuracy | 0.7441 | 2.44% | 0.7731 | 2.74% | yes | |
Avg. No. of Selected Features | 244.67 | 26.43% | 236.91 | 25.24% | yes | |
Running Time (s) | 80.1235 | 12.12% | 80.1122 | 16.05% | no | |
Backdoor | F1 | 0.8392 | 3.25% | 0.8422 | 1.94% | no |
Precision | 0.9405 | 1.43% | 0.9205 | 5.10% | no | |
Recall | 0.7576 | 3.22% | 0.7762 | 2.08% | no | |
Accuracy | 0.9647 | 1.57% | 0.9547 | 2.20% | no | |
Avg. No. of Selected Features | 18.80 | 5.17% | 12.50 | 3.78% | yes | |
Running Time (s) | 198.99 | 23.23% | 212.35 | 23.96% | no | |
Campaign | F1 | 0.6613 | 2.49% | 0.6934 | 1.96% | no |
Precision | 0.6553 | 2.50% | 0.6826 | 1.63% | no | |
Recall | 0.6726 | 3.23% | 0.7046 | 3.18% | no | |
Accuracy | 0.8573 | 1.61% | 0.8733 | 1.20% | no | |
Avg. No. of Selected Features | 15.40 | 2.01% | 10.70 | 4.32% | yes | |
Running Time (s) | 88.50 | 15.32% | 92.83 | 16.06% | no |
Dataset | Evaluation Metric | IbQGA | IbQGA std | FbQGA | FbQGA std | SSD |
---|---|---|---|---|---|---|
Taiwan Bankruptcy | F1 | 0.5643 | 2.17% | 0.5886 | 1.17% | yes |
Precision | 0.55 | 1.33% | 0.5663 | 1.32% | yes | |
Recall | 0.5793 | 4.90% | 0.6127 | 3.06% | yes | |
Accuracy | 0.872 | 3.15% | 0.8961 | 4.50% | no | |
Avg. No. of Selected Features | 32.05 | 25.70% | 31.52 | 23.12% | no | |
Running Time (s) | 35.7801 | 13.92% | 33.1522 | 15.90% | no | |
Musk | F1 | 0.6762 | 4.48% | 0.6965 | 5.48% | yes |
Precision | 0.642 | 5.10% | 0.654 | 4.15% | no | |
Recall | 0.7245 | 6.13% | 0.745 | 6.13% | no | |
Accuracy | 0.6912 | 7.37% | 0.712 | 6.37% | no | |
Avg. No. of Selected Features | 30.12 | 21.13% | 22.13 | 24.97% | yes | |
Running Time (s) | 226.6581 | 17.72% | 215.4651 | 15.83% | no | |
Parkinson’s Disease | F1 | 0.6781 | 3.41% | 0.6822 | 2.72% | yes |
Precision | 0.6713 | 2.89% | 0.6872 | 2.34% | yes | |
Recall | 0.6831 | 2.89% | 0.7031 | 2.61% | yes | |
Accuracy | 0.745 | 2.67% | 0.7731 | 2.74% | yes | |
Avg. No. of Selected Features | 243.26 | 29.45% | 236.91 | 25.24% | yes | |
Running Time (s) | 78.5213 | 16.12% | 80.1122 | 16.05% | no | |
Backdoor | F1 | 0.7826 | 4.49% | 0.8422 | 1.94% | yes |
Precision | 0.8711 | 2.10% | 0.9205 | 5.10% | yes | |
Recall | 0.7105 | 4.50% | 0.7762 | 2.08% | yes | |
Accuracy | 0.9247 | 2.20% | 0.9547 | 2.20% | yes | |
Avg. No. of Selected Features | 17.30 | 2.49% | 12.50 | 3.78% | yes | |
Running Time (s) | 231.89 | 27.18% | 212.35 | 23.96% | no | |
Campaign | F1 | 0.6914 | 1.78% | 0.6934 | 1.96% | no |
Precision | 0.6880 | 1.77% | 0.6826 | 1.63% | no | |
Recall | 0.6948 | 2.14% | 0.7046 | 3.18% | no | |
Accuracy | 0.8709 | 1.29% | 0.8733 | 1.20% | no | |
Avg. No. of Selected Features | 15.40 | 3.13% | 10.70 | 4.32% | yes | |
Running Time (s) | 92.48 | 14.24% | 92.83 | 16.06% | no |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Pham, T.H.; Raahemi, B. An Adaptive Feature-Based Quantum Genetic Algorithm for Dimension Reduction with Applications in Outlier Detection. Algorithms 2025, 18, 154. https://doi.org/10.3390/a18030154
Pham TH, Raahemi B. An Adaptive Feature-Based Quantum Genetic Algorithm for Dimension Reduction with Applications in Outlier Detection. Algorithms. 2025; 18(3):154. https://doi.org/10.3390/a18030154
Chicago/Turabian StylePham, Tin H., and Bijan Raahemi. 2025. "An Adaptive Feature-Based Quantum Genetic Algorithm for Dimension Reduction with Applications in Outlier Detection" Algorithms 18, no. 3: 154. https://doi.org/10.3390/a18030154
APA StylePham, T. H., & Raahemi, B. (2025). An Adaptive Feature-Based Quantum Genetic Algorithm for Dimension Reduction with Applications in Outlier Detection. Algorithms, 18(3), 154. https://doi.org/10.3390/a18030154