Splitting Choice and Computational Complexity Analysis of Decision Trees
Abstract
:1. Introduction
2. Splitting Bias
2.1. Bias Due to Missing Values
2.2. Bias Related to More Values or Categories
: X is independent of Y; | : X is dependent on Y |
3. Influence of Noise Variables on CART Computational Complexity
- All the independent variables can be divided into effective variables and noise variables. The criterion is whether they are used in the tree growing process or not. As the most effective variables will be chosen for splitting firstly. Those variables not chosen have less effect than those chosen. A tree building process includes both a growing process and pruning process (or stopping criteria). This time, the tree is assumed to choose the stopping criteria, so that we only need to concentrate on the growing process. Noise variables refer to variables that are not used in the tree growing process.
- All variables are categorial variables for convenience of calculation.
- For every split, no matter how many categories the independent variable has, there are always two child nodes after the parent node since CART is a binary tree. All nodes are assumed to stop splitting at the same time which means the depth is the same for every branch on the same level.
- When one independent variable is chosen as a split, it will not be chosen again.
3.1. Computational Complexity without Noise Variables
3.2. Computational Complexity with Noise Variables
3.3. Computational Complexity Increase
4. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Breiman, L.; Friedman, J.; Stone, C.J.; Olshen, R.A. Classification and Regression Trees; Taylor & Francis Group: Abingdon, UK, 1984. [Google Scholar]
- Hothorn, T.; Hornik, K.; Zeileis, A. Unbiased recursive partitioning: A conditional inference framework. J. Comput. Graph. Stat. 2006, 15, 651–674. [Google Scholar] [CrossRef] [Green Version]
- Hothorn, T.; Hornik, K.; Van De Wiel, M.A.; Zeileis, A. A Lego system for conditional inference. Am. Stat. 2006, 60, 257–263. [Google Scholar] [CrossRef] [Green Version]
- Funahashi, K.I. On the approximate realization of continuous mappings by neural networks. Neural Netw. 1989, 2, 183–192. [Google Scholar] [CrossRef]
- Specht, D.F. Probabilistic neural networks. Neural Netw. 1990, 3, 109–118. [Google Scholar] [CrossRef]
- Suykens, J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
- Cauwenberghs, G.; Poggio, T. Incremental and decremental support vector machine learning. In The Proceedings of the 2000 Neural Information Processing Systems (NIPS) Conference; MIT Press: Cambridge, MA, USA, 2001; pp. 409–415. [Google Scholar]
- Walker, S.H.; Duncan, D.B. Estimation of the probability of an event as a function of several independent variables. Biometrika 1967, 54, 167–179. [Google Scholar] [CrossRef] [PubMed]
- Cox, D.R. The regression analysis of binary sequences. J. R. Stat. Soc. Ser. 1958, 20, 215–242. [Google Scholar] [CrossRef]
- Goodman, K.E.; Lessler, J.; Cosgrove, S.E.; Harris, A.D.; Lautenbach, E.; Han, J.H.; Milstone, A.M.; Massey, C.J.; Tamma, P.D. A clinical decision tree to predict whether a bacteremic patient is infected with an extended-spectrum β-Lactamase–producing organism. Clin. Infect. Dis. 2016, 63, 896–903. [Google Scholar] [CrossRef] [PubMed]
- Abdar, M.; Kalhori, S.R.N.; Sutikno, T.; Subroto, I.M.I.; Arji, G. Comparing Performance of Data Mining Algorithms in Prediction Heart Diseases. Int. J. Electr. Comput. Eng. 2015, 5, 1569–1576. [Google Scholar] [CrossRef]
- Sun, P.; Cárdenas, D.A.; Harrill, R. Chinese customers evaluation of travel website quality: A decision-tree analysis. J. Hosp. Mark. Manag. 2016, 25, 476–497. [Google Scholar] [CrossRef]
- Zhang, Y.; Wang, S.; Phillips, P.; Ji, G. Binary PSO with mutation operator for feature selection using decision tree applied to spam detection. Knowl.-Based Syst. 2014, 64, 22–31. [Google Scholar] [CrossRef]
- Loh, W.Y. Classification and regression trees. Wiley Interdiscip. Rev. Data Min. Knowl. Discov. 2011, 1, 14–23. [Google Scholar] [CrossRef]
- Gini, C. Variabilità e mutabilità; 1912. Available online: https://ui.adsabs.harvard.edu/abs/1912vamu.book.....G/abstract (accessed on 23 September 2021).
- Iversen, G.R.; Norpoth, H. Analysis of Variance; SAGE: New York, NY, USA, 1987. [Google Scholar]
- Cohen, J. The analysis of variance. Stat. Power Anal. Behav. Sci. 1988, 2, 273–403. [Google Scholar]
- Zhao, X.; Barber, S.; Taylor, C.C.; Milan, Z. Classification tree methods for panel data using wavelet-transformed time series. Comput. Stat. Data Anal. 2018, 127, 204–216. [Google Scholar] [CrossRef] [Green Version]
- Zhao, X.; Barber, S.; Taylor, C.C.; Milan, Z. Interval forecasts based on regression trees for streaming data. Adv. Data Anal. Classif. 2019. [Google Scholar] [CrossRef]
- Strobl, C.; Boulesteix, A.L.; Augustin, T. Unbiased split selection for classification trees based on the Gini index. Comput. Stat. Data Anal. 2007, 52, 483–501. [Google Scholar] [CrossRef] [Green Version]
- Kass, G.V. An exploratory technique for investigating large quantities of categorical data. Appl. Stat. 1980, 29, 119–127. [Google Scholar] [CrossRef]
- Therneau, T.M.; Atkinson, E.J.; Foundation, M. An Introduction to Recursive Partitioning Using the RPART Routines. 1997. Available online: https://stat.ethz.ch/R-manual/R-patched/library/rpart/doc/longintro.pdf (accessed on 24 September 2021).
- Therneau, T.; Atkinson, B.; Ripley, B. rpart: Recursive Partitioning and Regression Trees. R Package Version 2014, 4, 1–9. [Google Scholar]
- Quinlan, J.R. Induction of decision trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef] [Green Version]
- Quinlan, J.R. C4.5: Programs for Machine Learning; Elsevier: Amsterdam, The Netherlands, 2014. [Google Scholar]
- Rokach, L.; Maimon, O. Data Mining with Decision Trees: Theory and Applications; World Scientific: Singapore, 2008. [Google Scholar]
- Bonferroni, C. Teoria statistica delle classi e calcolo delle probabilita. Pubbl. Del Ist. Super. Sci. Econ. Commericiali Firenze 1936, 8, 3–62. (In Italian) [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zhao, X.; Nie, X. Splitting Choice and Computational Complexity Analysis of Decision Trees. Entropy 2021, 23, 1241. https://doi.org/10.3390/e23101241
Zhao X, Nie X. Splitting Choice and Computational Complexity Analysis of Decision Trees. Entropy. 2021; 23(10):1241. https://doi.org/10.3390/e23101241
Chicago/Turabian StyleZhao, Xin, and Xiaokai Nie. 2021. "Splitting Choice and Computational Complexity Analysis of Decision Trees" Entropy 23, no. 10: 1241. https://doi.org/10.3390/e23101241
APA StyleZhao, X., & Nie, X. (2021). Splitting Choice and Computational Complexity Analysis of Decision Trees. Entropy, 23(10), 1241. https://doi.org/10.3390/e23101241