Adaptive Two-Index Fusion Attribute-Weighted Naive Bayes
Abstract
:1. Introduction
- We propose a general framework, ATFNB, which fuses any two indexes from two categories of data description. Our framework can derive all existing filter-attribute-weighted NB models by selecting difference indexes.
- We introduce a regulatory factor , which can adaptively adjust the optimal ratio of two indexes, and shows which index is more important on various datasets.
- A quick range query method is proposed to obtain the optimal value of the regulatory factor . Compared to the traditional method, step-length searching, our method obviously speeds up the optimization.
- The existing two-index NB methods’ performance are significantly improved by introducing the regulatory factor .
2. Related Work
3. ATFNB
3.1. The General Framework of ATFNB
3.2. Index Selection
3.3. Range Query Method for the Regulatory Factor
Algorithm 1: RQRF |
Input: class–attribute (), attribute–attribute (), dataset D For each instance in D: For each class c in C: Calculate and in Equation (13). According and , get . End If instance label is : Find the value that satisfies Equation (14); it is recorded as , otherwise = ∅. End End For each Find the subinterval that conforms to Equation (15). End Output: |
3.4. The Implementation of ATFNB
Algorithm 2: ATFNB Framework |
Input: Training set D, test set X (1) For each attribute in D Calculate (attribute–attribute) index Calculate (class–attribute) index . (2) According to RQRF, the value of the regulatory factor is solved. (3) According to Equation (12), the weight matrix is obtained. (4) According to Equation (4), the class label of each instance in X is predicted. Output: The class label of instances in X |
4. Experiments and Results
4.1. Experimental Data
4.2. Experimental Setting
4.3. The Effectiveness of the Regulatory Factor
4.4. Experimental Results on UCI Datasets
4.5. Experimental Results on Flavia Dataset
5. Discussion
5.1. The Influence of Instance and Attribute Number
- (1)
- On 78.26% of datasets with less than 500 instances, ATFNB could achieve the highest classification accuracy. On datasets with a number of instances greater than or equal to 500, 56.25% could achieve the maximum classification accuracy. By comparison, ATFNB is more advantageous on datasets with fewer instances.
- (2)
- For datasets with a number of attributes less than 15, the percentage of datasets with the highest classification accuracy for ATFNB (67.74%) was also higher than that with a number of attributes greater than or equal to 15 (63.16%).
- (3)
- When the number of instances was less than 500 and the number of attributes was greater than 15, the percentage of datasets with the highest classification accuracy for ATFNB (87.5%) was significantly higher than that of the other three types of datasets (73.33%, 62.50%, 45.45%).
5.2. The Distribution of the Regulatory Factor
- (1)
- If the number of instances was less than 500, the upper bound value of in 52.17% of the datasets was less than 0.5. If the number of instances was more than 500, the upper bound value of in 40.74% of the datasets was less than 0.5.
- (2)
- From the perspective of the number of attributes, regardless of the number of attributes, the upper bound value of was less than 0.5 in most datasets.
- (3)
- (3) Considering the number of instance and attributes simultaneously, the upper bound value of in 62.50% of the datasets with a number of instances less than 500 and number of attributes greater than 15 was less than 0.5. On the dataset with a number of instances greater than 500 and number of attributes greater than 15, the upper bound value of in 45.46% of the datasets was less than 0.5.
5.3. The Impact of Different Index Combinations
5.4. Computation Time
6. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Zhang, H.; Jiang, L.; Yu, L. Class-specific attribute value weighting for Naive Bayes. Inf. Sci. 2020, 508, 260–274. [Google Scholar] [CrossRef]
- Koivisto, M.; Sood, K. Exact Bayesian structure discovery in Bayesian networks. J. Mach. Learn. Res. 2004, 5, 549–573. [Google Scholar]
- Friedman, N.; Koller, D. Being Bayesian about network structure. A Bayesian approach to structure discovery in Bayesian networks. Mach. Learn. 2003, 50, 95–125. [Google Scholar] [CrossRef]
- Friedman, N.; Geiger, D.; Goldszmidt, M. Bayesian network classifiers. Mach. Learn. 1997, 29, 131–163. [Google Scholar] [CrossRef]
- Jiang, L.; Zhang, H.; Cai, Z. A novel bayes model: Hidden naive bayes. IEEE Trans. Knowl. Data Eng. 2008, 21, 1361–1371. [Google Scholar] [CrossRef]
- Jiang, L.; Wang, S.; Li, C.; Zhang, L. Structure extended multinomial naive Bayes. Inf. Sci. 2016, 329, 346–356. [Google Scholar] [CrossRef]
- Diab, D.M.; El Hindi, K.M. Using differential evolution for fine tuning naïve Bayesian classifiers and its application for text classification. Appl. Soft Comput. 2017, 54, 183–199. [Google Scholar] [CrossRef]
- El Hindi, K. Fine tuning the Naïve Bayesian learning algorithm. AI Commun. 2014, 27, 133–141. [Google Scholar] [CrossRef]
- Ryu, D.; Jang, J.I.; Baik, J. A hybrid instance selection using nearest-neighbor for cross-project defect prediction. J. Comput. Sci. Technol. 2015, 30, 969–980. [Google Scholar] [CrossRef]
- Xie, Z.; Hsu, W.; Liu, Z.; Lee, M.L. Snnb: A selective neighborhood based naive Bayes for lazy learning. In Proceedings of the Pacific-Asia Conference on Knowledge Discovery and Data Mining, Taipei, Taiwan, 6–8 May 2002; Springer: Berlin/Heidelberg, Germany, 2002; pp. 104–114. [Google Scholar]
- Zhang, H.; Jiang, L.; Yu, L. Attribute and instance weighted naive Bayes. Pattern Recognit. 2021, 111, 107674. [Google Scholar] [CrossRef]
- Elkan, C. Boosting and naive bayesian learning. In Proceedings of the International Conference on Knowledge Discovery and Data Mining, Newport Beach, CA, USA, 14–17 August 1997. [Google Scholar]
- Jiang, L.; Cai, Z.; Wang, D. Improving naive Bayes for classification. Int. J. Comput. Appl. 2010, 32, 328–332. [Google Scholar] [CrossRef]
- Chen, J.; Huang, H.; Tian, S.; Qu, Y. Feature selection for text classification with Naïve Bayes. Expert Syst. Appl. 2009, 36, 5432–5435. [Google Scholar] [CrossRef]
- Choubey, D.K.; Paul, S.; Kumar, S.; Kumar, S. Classification of Pima indian diabetes dataset using naive bayes with genetic algorithm as an attribute selection. In Proceedings of the Communication and Computing Systems: Proceedings of the International Conference on Communication and Computing System (ICCCS 2016), Nanjing, China, 19–31 July 2017; pp. 451–455.
- Hall, M.A.; Holmes, G. Benchmarking attribute selection techniques for discrete class data mining. IEEE Trans. Knowl. Data Eng. 2003, 15, 1437–1447. [Google Scholar] [CrossRef]
- Jiang, L.; Cai, Z.; Zhang, H.; Wang, D. Not so greedy: Randomly selected naive Bayes. Expert Syst. Appl. 2012, 39, 11022–11028. [Google Scholar] [CrossRef]
- Lee, C.; Lee, G.G. Information gain and divergence-based feature selection for machine learning-based text categorization. Inf. Process. Manag. 2006, 42, 155–165. [Google Scholar] [CrossRef]
- Deng, X.; Li, Y.; Weng, J.; Zhang, J. Feature selection for text classification: A review. Multimed. Tools Appl. 2019, 78, 3797–3816. [Google Scholar] [CrossRef]
- Lee, C.H. An information-theoretic filter approach for value weighted classification learning in naive Bayes. Data Knowl. Eng. 2018, 113, 116–128. [Google Scholar] [CrossRef]
- Yu, L.; Jiang, L.; Wang, D.; Zhang, L. Toward naive Bayes with attribute value weighting. Neural Comput. Appl. 2019, 31, 5699–5713. [Google Scholar] [CrossRef]
- Jiang, L.; Zhang, L.; Yu, L.; Wang, D. Class-specific attribute weighted naive Bayes. Pattern Recognit. 2019, 88, 321–330. [Google Scholar] [CrossRef]
- Wu, J.; Pan, S.; Zhu, X.; Cai, Z.; Zhang, P.; Zhang, C. Self-adaptive attribute weighting for Naive Bayes classification. Expert Syst. Appl. 2015, 42, 1487–1502. [Google Scholar] [CrossRef]
- Hall, M. A decision tree-based attribute weighting filter for naive Bayes. In Proceedings of the International Conference on Innovative Techniques and Applications of Artificial Intelligence, Cambridge, UK, 11–13 December 2006; Springer: Berlin/Heidelberg, Germany, 2006; pp. 59–70. [Google Scholar]
- Taheri, S.; Yearwood, J.; Mammadov, M.; Seifollahi, S. Attribute weighted Naive Bayes classifier using a local optimization. Neural Comput. Appl. 2014, 24, 995–1002. [Google Scholar] [CrossRef] [Green Version]
- Jiang, L.; Li, C.; Wang, S.; Zhang, L. Deep feature weighting for naive Bayes and its application to text classification. Eng. Appl. Artif. Intell. 2016, 52, 26–39. [Google Scholar] [CrossRef]
- Zhang, L.; Jiang, L.; Li, C.; Kong, G. Two feature weighting approaches for naive Bayes text classifiers. Knowl.-Based Syst. 2016, 100, 137–144. [Google Scholar] [CrossRef]
- Wu, J.; Cai, Z. Attribute weighting via differential evolution algorithm for attribute weighted naive bayes (wnb). J. Comput. Inf. Syst. 2011, 7, 1672–1679. [Google Scholar]
- Jiang, L.; Zhang, L.; Li, C.; Wu, J. A correlation-based feature weighting filter for Naive Bayes. IEEE Trans. Knowl. Data Eng. 2018, 31, 201–213. [Google Scholar] [CrossRef]
- Ferreira, J.; Denison, D.; Hand, D. Weighted Naive Bayes Modelling for Data Mining; Technical Report; Imperial College: London, UK, 2001. [Google Scholar]
- Zhang, H.; Sheng, S. Learning weighted naive Bayes with accurate ranking. In Proceedings of the Fourth IEEE International Conference on Data Mining (ICDM’04), Online. 1 November 2004; pp. 567–570. [Google Scholar]
- Quinlan, J.R. C4.5: Programs for Machine Learning; Elsevier: San Mateo, CA, USA, 2014. [Google Scholar]
- Lee, C.H.; Gutierrez, F.; Dou, D. Calculating feature weights in naive bayes with kullback-leibler measure. In Proceedings of the 2011 IEEE 11th International Conference on Data Mining, Vancouver, BC, Canada, 11–14 December 2011; pp. 1146–1151. [Google Scholar]
- Kullback, S.; Leibler, R.A. On information and sufficiency. Ann. Math. Stat. 1951, 22, 79–86. [Google Scholar] [CrossRef]
- Hall, M.A. Correlation-Based Feature Selection of Discrete and Numeric Class Machine Learning; University of Waikato, Department of Computer Science (Working paper 00/08); University of Waikato, Department of Computer Science: Hamilton, NZ, USA, 2000. [Google Scholar]
- Asuncion, A.; Newman, D. UCI Machine Learning Repository; Irvine University of California: Irvine, CA, USA, 2007. [Google Scholar]
- Kerber, R. Chimerge: Discretization of numeric attributes. In Proceedings of the Tenth National Conference on Artificial Intelligence, San Jose, CA, USA, 12–16 July 1992; pp. 123–128. [Google Scholar]
- Sachar, S.; Kumar, A. Survey of feature extraction and classification techniques to identify plant through leaves. Expert Syst. Appl. 2021, 167, 114181. [Google Scholar] [CrossRef]
- Langley, P.; Iba, W.; Thompson, K. An analysis of Bayesian classifiers. In Proceedings of the AAAI, San Jose, CA, USA, 12–16 July 1992; Volume 90, pp. 223–228. [Google Scholar]
- Nadeau, C.; Bengio, Y. Inference for the generalization error. Mach. Learn. 2003, 52, 239–281. [Google Scholar] [CrossRef] [Green Version]
Class | …… | |||||
---|---|---|---|---|---|---|
…… | ||||||
…… | ||||||
…… | ||||||
…… | ||||||
…… |
Dataset | Instance Number | Attribute Number | Class Number |
---|---|---|---|
abalone | 4177 | 8 | 3 |
acute | 120 | 6 | 2 |
aggregation | 788 | 2 | 7 |
balance-scale | 625 | 4 | 3 |
bank | 4521 | 16 | 2 |
banknote | 1372 | 4 | 2 |
blood | 748 | 4 | 2 |
breast-cancer | 286 | 9 | 2 |
breast-tissue | 106 | 9 | 6 |
bupa | 345 | 6 | 2 |
car | 1728 | 6 | 4 |
chart_Input | 600 | 60 | 6 |
climate-simulation | 540 | 18 | 2 |
congressional-voting | 435 | 16 | 2 |
connectionist | 208 | 60 | 2 |
dermatology | 366 | 34 | 6 |
diabetes | 768 | 8 | 2 |
ecoli | 336 | 7 | 8 |
energy-y1 | 768 | 8 | 3 |
fertility | 100 | 9 | 2 |
glass | 214 | 9 | 6 |
haberman-survival | 306 | 3 | 2 |
iris | 150 | 4 | 3 |
jain | 373 | 2 | 2 |
knowledge | 172 | 5 | 4 |
libras | 360 | 90 | 15 |
low-res-spect | 531 | 100 | 9 |
lymphography | 148 | 18 | 4 |
magic | 19,020 | 10 | 2 |
mammographic | 961 | 5 | 2 |
promoters | 106 | 57 | 2 |
splice | 3190 | 60 | 3 |
nursery | 12,960 | 8 | 5 |
page-blocks | 5473 | 10 | 5 |
pima | 768 | 8 | 2 |
planning | 182 | 12 | 2 |
post-operative | 90 | 8 | 3 |
robotnavigation | 5456 | 24 | 4 |
seeds | 210 | 7 | 3 |
sonar | 208 | 60 | 2 |
soybean | 683 | 35 | 18 |
spect | 265 | 22 | 2 |
synthetic-control | 600 | 60 | 6 |
tic-tac-toe | 958 | 9 | 2 |
titanic | 2201 | 3 | 2 |
twonorm | 7400 | 20 | 2 |
wall-following | 5456 | 24 | 4 |
waveform | 5000 | 21 | 3 |
wilt | 4839 | 5 | 2 |
wine | 178 | 13 | 3 |
Dataset | Instance Number | Attribute Number | Class Number |
---|---|---|---|
869 | 14 | 15 | |
888 | 14 | 15 | |
865 | 14 | 15 | |
887 | 14 | 15 | |
884 | 14 | 15 | |
919 | 14 | 15 | |
892 | 14 | 15 | |
881 | 14 | 15 | |
864 | 14 | 15 | |
879 | 14 | 15 | |
895 | 14 | 15 | |
888 | 14 | 15 | |
927 | 14 | 15 | |
924 | 14 | 15 | |
904 | 14 | 15 |
Dataset | The Interval of Regulatory Factor | Time (s) | |||
---|---|---|---|---|---|
SLS | RQRF | SLS | RQRF | Speed | |
bupa | [0.17, 0.59] | [0.1687, 0.5937] | 7.4908 | 0.0119 | ×629 |
abalone | [0.70, 0.75] | [0.6988, 0.7521] | 16.826 | 0.1068 | ×157 |
breast-cancer | [0.23, 0.31] | [0.2257, 0.3129] | 6.8023 | 0.0389 | ×174 |
knowledge | [0.27, 0.29] | [0.2688, 0.2954] | 6.1298 | 0.0229 | ×267 |
Dataset | NB | WNB | CFW | ATFNB | CFW- |
---|---|---|---|---|---|
abalone | 0.5886 | 0.5871 * | 0.5890 * | 0.5908 | 0.5926 |
acute | 0.9958 | 0.9521 * | 0.9948 | 0.9635 | 0.9813 |
aggregation | 0.9890 | 0.9882 | 0.9761 * | 0.9875 | 0.9824 |
balance-scale | 0.8592 * | 0.8728 | 0.8312 * | 0.8984 | 0.8581 |
bank | 0.8765 | 0.8831 | 0.8901 | 0.8822 | 0.9076 |
banknote | 0.8636 | 0.8468 | 0.8491 | 0.8498 | 0.8338 |
blood | 0.7597 * | 0.7733 | 0.7720 * | 0.7847 | 0.7990 |
breast-cancer | 0.7214 * | 0.7059 * | 0.7331 | 0.7472 | 0.7422 |
breast-tissue | 0.5727 | 0.5955 | 0.5818 * | 0.6091 | 0.6158 |
bupa | 0.6232 | 0.5942 * | 0.6174 * | 0.6333 | 0.6299 |
car | 0.8523 | 0.6965 * | 0.7671 * | 0.8014 | 0.8101 |
chart_Input | 0.9533 | 0.9367 | 0.9558 | 0.9455 | 0.9488 |
climate-simulation | 0.9137 | 0.9178 | 0.9174 | 0.9181 | 0.9209 |
congressional-voting | 0.6149 * | 0.6345 * | 0.6253 * | 0.6506 | 0.6614 |
connectionist | 0.7238 * | 0.7429 * | 0.7214 * | 0.7667 | 0.7560 |
dermatology | 0.9797 | 0.9644 | 0.9757 | 0.9649 | 0.9665 |
diabetes | 0.7377 | 0.6584 * | 0.7403 | 0.7422 | 0.7611 |
Ecoli | 0.8135 | 0.7706 * | 0.7588 * | 0.8245 | 0.7981 |
energy-y1 | 0.8874 | 0.8225 * | 0.8701 | 0.8463 | 0.8813 |
fertility | 0.8400 * | 0.8500 | 0.8350 * | 0.8650 | 0.8669 |
glass | 0.7023 | 0.6837 * | 0.6930 | 0.7193 | 0.7233 |
haberman-survival | 0.7532 * | 0.7468 * | 0.7403 * | 0.7710 | 0.7791 |
Iris | 0.9133 | 0.9100 * | 0.9167 | 0.9367 | 0.9099 |
Jain | 0.9464 | 0.9368 | 0.9379 | 0.9397 | 0.9399 |
knowledge | 0.7371* | 0.7743 * | 0.7629 * | 0.8057 | 0.7989 |
libras | 0.5903 | 0.5917 | 0.5847 | 0.5965 | 0.6122 |
low-res-spect | 0.8037 * | 0.8018 * | 0.8131* | 0.8318 | 0.8411 |
lymphography | 0.8122 * | 0.7889 * | 0.8233 | 0.8334 | 0.8399 |
magic | 0.7300 | 0.6885 * | 0.7411 | 0.7674 | 0.7782 |
mammographic | 0.8290 * | 0.8394 | 0.8446 | 0.8549 | 0.8679 |
promoters | 0.9091 * | 0.9045 * | 0.9242 * | 0.9545 | 0.9302 |
splice | 0.9475 | 0.9376 * | 0.9580 | 0.9414 | 0.9677 |
nursery | 0.9043 | 0.8089 * | 0.8812 | 0.8961 | 0.9002 |
page-blocks | 0.9300 * | 0.9404 | 0.9545 * | 0.9684 | 0.9690 |
pima | 0.7338 * | 0.6688 * | 0.7330 * | 0.7599 | 0.7613 |
planning | 0.6000 * | 0.7189 | 0.6919 * | 0.7378 | 0.7500 |
post-operative | 0.7222 * | 0.8519 * | 0.7593 * | 0.9074 | 0.8489 |
robotnavigation | 0.8760 * | 0.9159 | 0.9095 | 0.9179 | 0.9199 |
seeds | 0.8747 | 0.8622 | 0.8762 | 0.8655 | 0.8881 |
sonar | 0.7625 | 0.7429 * | 0.7571 * | 0.7734 | 0.7662 |
soybean | 0.9036 | 0.8730 | 0.9117 | 0.8781 | 0.9049 |
spect | 0.6566 * | 0.6604 * | 0.6792 * | 0.7151 | 0.7288 |
synthetic-control | 0.9677 | 0.9458 | 0.9698 | 0.9567 | 0.9675 |
tic-tac-toe | 0.7141 | 0.6589 * | 0.7109 | 0.7005 | 0.7201 |
titanic | 0.7782 | 0.6680 * | 0.7751 | 0.7822 | 0.7991 |
twonorm | 0.9384 | 0.9364 | 0.9388 | 0.9489 | 0.9346 |
wall-following | 0.8032 | 0.7964 | 0.8137 | 0.7976 | 0.8199 |
waveform | 0.8080 * | 0.7960 * | 0.8172 * | 0.8355 | 0.8317 |
wilt | 0.9472 | 0.9374 * | 0.9475 | 0.9523 | 0.9538 |
wine | 0.9694 | 0.9625 | 0.9750 | 0.9697 | 0.9622 |
Average | 0.8146 | 0.8028 | 0.8169 | 0.8317 | 0.8345 |
G/W/L | 9/15/35 | 0/2/48 | 8/12/38 | 33/ / |
Algorithm | ATFNB | WNB | CFW | NB |
---|---|---|---|---|
ATFNB | — | 2 (0) | 12 (4) | 15 (5) |
WNB | 48 (28) | — | 35 (16) | 34 (19) |
CFW | 38 (22) | 15 (5) | — | 20 (8) |
NB | 35 (19) | 16 (9) | 30 (11) | — |
Algorithm | ATFNB | WNB | CFW | NB |
---|---|---|---|---|
ATFNB | — | 1268 | 1007.5 | 961.5 |
WNB | 7 | — | 308.5 | 391.5 |
CFW | 267.5 | 966.5 | — | 771 |
NB | 313.5 | 883.5 | 504 | — |
Algorithm | ATFNB | WNB | CFW | NB |
---|---|---|---|---|
ATFNB | — | ∘ | ∘ | ∘ |
WNB | • | — | ∘ | ∘ |
CFW | • | • | — | |
NB | • | • | — |
Group | NB | WNB | CFW | ATFNB | CFW- |
---|---|---|---|---|---|
G_1 | 0.8253 * | 0.8506 * | 0.8552 * | 0.8805 | 0.9011 |
G_2 | 0.8337 | 0.8629 | 0.8742 | 0.8444 | 0.8668 |
G_3 | 0.7874 * | 0.8484 | 0.8312 | 0.8786 | 0.8771 |
G_4 | 0.8562 * | 0.8854 * | 0.8899 | 0.8987 | 0.9022 |
G_5 | 0.8016 * | 0.8129 | 0.8050 * | 0.8174 | 0.8177 |
G_6 | 0.8822 | 0.8729 * | 0.8903 | 0.8843 | 0.8801 |
G_7 | 0.9134 * | 0.9137 * | 0.9322 | 0.9233 | 0.9400 |
G_8 | 0.9011 | 0.8812 * | 0.8927 | 0.8904 | 0.9022 |
G_9 | 0.9122 | 0.9100 | 0.9033 * | 0.9422 | 0.8891 |
G_10 | 0.8135 * | 0.8213 * | 0.8200 * | 0.8422 | 0.8399 |
G_11 | 0.8572 | 0.8734 | 0.8534 | 0.8799 | 0.8912 |
G_12 | 0.8356 | 0.8132 * | 0.8224 | 0.8233 | 0.8335 |
G_13 | 0.7724 * | 0.7787 * | 0.7732 * | 0.7987 | 0.7887 |
G_14 | 0.8342 * | 0.8344 | 0.8322 * | 0.8458 | 0.8422 |
G_15 | 0.9169 * | 0.9224 * | 0.9243 * | 0.9321 | 0.9095 |
Average | 0.8495 | 0.8588 | 0.8600 | 0.8721 | 0.8720 |
G/W/L | 2/2/13 | 0/1/14 | 3/4/11 | 10 / / |
Algorithm | ATFNB | WNB | CFW | NB |
---|---|---|---|---|
ATFNB | — | 1 (0) | 4 (1) | 2 (0) |
WNB | 14 (9) | — | 8 (3) | 4 (1) |
CFW | 11 (7) | 7 (4) | — | 5 (1) |
NB | 13 (9) | 11 (6) | 10 (5) | — |
Algorithm | ATFNB | WNB | CFW | NB |
---|---|---|---|---|
ATFNB | — | 110 | 96 | 110.5 |
WNB | 10 | — | 52 | 89 |
CFW | 24 | 68 | — | 87 |
NB | 9.5 | 31 | 33 | — |
Algorithm | ATFNB | WNB | CFW | NB |
---|---|---|---|---|
ATFNB | — | ∘ | ∘ | ∘ |
WNB | • | — | ||
CFW | • | — | ||
NB | • | — |
Data Characteristics | Number | ATFNB (%) | Competitors (%) | |
---|---|---|---|---|
Instance number | <500 | 23 | 78.26 | 21.74 |
≥500 | 27 | 56.25 | 43.75 | |
Attribute number | <15 | 31 | 67.74 | 32.26 |
≥15 | 19 | 63.16 | 36.84 | |
Instance and attribute | <500 and <15 | 15 | 73.33 | 26.67 |
<500 and ≥15 | 8 | 87.50 | 12.50 | |
≥500 and <15 | 16 | 62.50 | 37.50 | |
≥500 and ≥15 | 11 | 45.45 | 54.55 |
Dataset | Interval () | Mark | Dataset | Interval () | Mark |
---|---|---|---|---|---|
abalone | [0.7122, 0.8311] | ◯ | libras | [0.5377, 0.8832] | ◯ |
acute | [0.4418, 0.9433] | △ | low-res-spect | [0.6552, 0.7211] | ◯ |
aggregation | [0.5529, 0.8832] | △ | lymphography | [0.3344, 0.4834] | ⬜ |
balance-scale | [0.3233, 0.4537] | ⬜ | magic | [0.6733, 0.8122] | ◯ |
bank | [0.4198, 0.7691] | △ | mammographic | [0.1229, 0.3879] | ⬜ |
banknote | [0.3144, 0.3914] | ⬜ | promoters | [0.4876, 0.8867] | △ |
blood | [0.2243, 0.5532] | △ | splice | [0.0512, 0.1321] | ⬜ |
breast-cancer | [0.2311, 0.3521] | ⬜ | nursery | [0.5211, 0.5908] | ◯ |
breast-tissue | [0.3566, 0.4513] | ⬜ | page-blocks | [0.6322, 0.7109] | ◯ |
bupa | [0.1533, 0.6588] | △ | pima | [0.1566, 0.3118] | ⬜ |
car | [0.3211, 0.3987] | ⬜ | planning | [0.1829, 0.4721] | ⬜ |
chart_Input | [0.4592, 0.8311] | △ | post-operative | [0.0187, 0.2100] | ⬜ |
climate-simulation | [0.2301, 0.3255] | ⬜ | robotnavigation | [0.7122, 0.7830] | ◯ |
congressional-voting | [0.2199, 0.3472] | ⬜ | seeds | [0.4288, 0.8543] | △ |
connectionist | [0.0912, 0.1388] | ⬜ | sonar | [0.0521, 0.1487] | ⬜ |
dermatology | [0.3365, 0.8987] | △ | soybean | [0.2759, 0.3108] | ⬜ |
diabetes | [0.2355, 0.5243] | △ | spect | [0.1802, 0.2499] | ⬜ |
Ecoli | [0.7360, 0.9211] | ◯ | synthetic-control | [0.2480, 0.4033] | ⬜ |
energy-y1 | [0.3211, 0.9219] | △ | tic-tac-toe | [0.1213, 0.1870] | ⬜ |
fertility | [0.0511, 0.4390] | ⬜ | titanic | [0.4697, 0.6122] | △ |
glass | [0.1229, 0.1833] | ⬜ | twonorm | [0.5833, 0.6291] | ◯ |
haberman-survival | [0.2166, 0.6345] | △ | wall-following | [0.4128, 0.4736] | ⬜ |
iris | [0.3522, 0.8799] | △ | waveform | [0.7398, 0.7933] | ◯ |
jain | [0.3409, 0.8577] | △ | wilt | [0.6103, 0.6899] | ◯ |
knowledge | [0.3012, 0.4522] | ⬜ | wine | [0.3881, 0.9220] | △ |
Data Characteristics | Number | ◯ (%) | △ (%) | □ (%) | |
---|---|---|---|---|---|
Instance number | <500 | 23 | 8.70 | 39.13 | 52.17 |
≥500 | 27 | 33.33 | 25.93 | 40.74 | |
Attribute number | <15 | 31 | 19.35 | 38.71 | 41.94 |
≥15 | 19 | 26.32 | 21.05 | 52.63 | |
Instance and attribute | <500 and <15 | 15 | 6.66 | 46.67 | 46.67 |
<500 and ≥15 | 8 | 12.50 | 25.00 | 62.50 | |
≥500 and <15 | 16 | 31.25 | 31.25 | 37.50 | |
≥500 and ≥15 | 11 | 36.36 | 18.18 | 45.46 |
Algorithm | Stage 1 (s) | Stage 2 (s) | Stage 3 (s) | Total (s) |
---|---|---|---|---|
NB | 0.5233 | 0 | 0.0793 | 0.6026 |
WNB | 0.6811 | 0 | 0.0803 | 0.7614 |
CFW | 0.8794 | 0 | 0.0647 | 0.9441 |
CFW- | 0.8794 | 0.1092 | 0.0721 | 1.0607 |
ATFNB | 0.8413 | 0.1125 | 0.0823 | 1.0361 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhou, X.; Wu, D.; You, Z.; Wu, D.; Ye, N.; Zhang, L. Adaptive Two-Index Fusion Attribute-Weighted Naive Bayes. Electronics 2022, 11, 3126. https://doi.org/10.3390/electronics11193126
Zhou X, Wu D, You Z, Wu D, Ye N, Zhang L. Adaptive Two-Index Fusion Attribute-Weighted Naive Bayes. Electronics. 2022; 11(19):3126. https://doi.org/10.3390/electronics11193126
Chicago/Turabian StyleZhou, Xiaoliang, Donghua Wu, Zitong You, Dongyang Wu, Ning Ye, and Li Zhang. 2022. "Adaptive Two-Index Fusion Attribute-Weighted Naive Bayes" Electronics 11, no. 19: 3126. https://doi.org/10.3390/electronics11193126
APA StyleZhou, X., Wu, D., You, Z., Wu, D., Ye, N., & Zhang, L. (2022). Adaptive Two-Index Fusion Attribute-Weighted Naive Bayes. Electronics, 11(19), 3126. https://doi.org/10.3390/electronics11193126