Constructing Explainable Classifiers from the Start—Enabling Human-in-the Loop Machine Learning
Abstract
:1. Introduction
“It is obvious that the interactive approach to knowledge acquisition cannot keep pace with the burgeoning demand for expert systems; Feigenbaum terms this the ‘bottleneck problem’. This perception has stimulated the investigation of machine learning as a means of explicating knowledge”.[9]
“A definite loss of any communication abilities is contrary to the spirit of AI. AI systems are open to their user who must understand them”.[11]
2. Dataset Visualisations for Involving Experts in Classifier Construction
- The visualisation displays only two attributes at a time; this is critically restrictive.
- The space to display region bars is minuscule, impeding users’ observation of differences to decide which two attributes to display.
- Despite the immense literature on techniques for splitting a node to grow and construct a decision tree, the system does not provide any split-suggestion to the user.
- Unless users depart from the attribute visualisation window (losing context of the current splitting task), the tree under construction is not visible.
- Visualisation techniques (such as colour, or size) are not used. So, the user cannot inspect any properties of a node or an edge nor any relationship between a node and the dataset under analysis.
3. Iterative Construction of DTCs Supported by Visualisations in Parallel Coordinates
3.1. Using Parallel Coordinates
- We colour the corresponding leaf T to illustrate the purity of T (this directly correlates with the classification accuracy of the rule terminating at T). For instance, the left leaf in the tree of Figure 1a indicates it containing an almost even split of two classes. However, the right leaf is practically pure. The depth of the leaf T inversely correlates with the applicability and generality of that rule terminating at T. Understandability and interpretability also inversely correlate with leaf depth.
- The system allows the user to select whether to display values of predictability power of attributes, such as the information gain.
3.2. The Splits the User Shall Apply
3.3. Information That Supports Interaction
3.4. Visualising the Tree
3.5. Visualising Rules
3.5.1. Condensing the Decision Path
3.5.2. Visualising Negated Splits
4. Materials and Methods
4.1. Recruiting Participants
- Which of the following best describes the sector you primarily work in? Information Technology, Science, Technology, Engineering and Mathematics;
- What is your first language? English;
- Which of these is the highest level of education you have completed? Undergraduate degree (BA/BSc/other);
- Do you have computer programming skills? Yes/No.
4.2. Survey Design
4.2.1. Experiment 1—DTC Node Colouring
- A reasonably accurate (>90% accuracy) DTC can be learnt for each dataset with small trees sizes that remain interpretable to a human.
- All attributes have humanly understandable names and semantic meaning.
- Their prediction of the accuracy of each DTC.
- The time taken to predict the accuracy of each DTC.
- The leaf node selected by the participant as most impure for each DTC.
- The time taken to select the most impure leaf node in each DTC.
4.2.2. Experiment 2—Rule Visualisation
4.2.3. Experiment 3—Human-in-the-Loop Video Survey
- The first pair of videos examines the ability of a user to find an effective split for a node in a DTC. In the UserClassifier video, survey participants are shown how the dataset is visualised using a two-dimensional scatter-plot as well as how a user can select attributes using the small bar visualisation of each attribute on the right of the screen. The video then demonstrates how a user can construct a split for a node by selecting a region on the two-dimensional scatter-plot. The PC-based video shows how a dataset is visualised using PC and how a user can create a split by selecting a range on an axis. This video also shows how a user can rearrange axes and remove axes that are not of interest. In addition, participants are shown how a user can ask the PC-based system to reorder axes so that interesting axes appear on the far left. This rearrangement is achieved by ordering axes based on the best gain ratio of each attribute, as discussed in Section 3.3. At the end of both videos, participants are asked, ‘Which system do you believe provides a better method of finding splits to build a decision tree classifier?’
- The next pair of videos examines a user’s ability to navigate and understand a DTC in each HITL-ML system. The UserClassifier video shows how a user can switch to a separate tab from the training set visualisation to observe the current state of the DTC being constructed. Participants are shown how internal and leaf nodes are displayed as grey circles and rectangles, respectively. This video also shows how splits for each internal node can be observed by selecting the node and switching tabs to see the selected region in the dataset visualisation tab. The PC-based video demonstrates how the system splits one window into two sections. The video shows how the left section visualises the DTC using the coloured nodes. This video also demonstrates how a user can interact with this DTC and select a node, at which point any split for that node is shown on the far left axis in PC. After viewing both videos, the survey system asks participants, ‘Which system allows you to better navigate and understand the current state of a decision tree classifier as it is being constructed?’
- The third set of videos looks at a user’s ability to estimate the accuracy of a DTC using each HITL-ML system. The UserClassifier video shows participants how a user can look at the counts of instances shown in each leaf node to determine the majority class and how often this leaf misclassified training instances. This video also demonstrates how a user can assess the quality of a split for a node by visualising it with the training set in the second tab of the system. The PC-based video shows participants how a user can use the colouring of the nodes in the DTC to determine the accuracy of each leaf node. Participants are also shown how a user can assess the quality of a split using the visualisation of the split in PC. After both videos, participants are asked, ‘Which system allows you to more easily determine how often a tree will predict the correct class of an instance?’
- The fourth pair of videos examines the ability of a user to locate nodes in a DTC that requires further refinement. In the UserClassifier video, participants are shown how a user can use the visualisation of the DTC and the numbers within each node to find leaf nodes with a large number of instances from multiple classes. The PC-based video shows participants how a user can use the colouring of nodes to determine which nodes require further refinement. Following these videos, participants are asked, ‘Which system allows you to more easily find nodes in a decision tree classifier that need additional splits?’
- The final pair of videos look at the algorithmic assistance features offered by each system. In the UserClassifier video, participants are shown how a user can use an automated algorithm to complete a subtree. This video points out to participants that the user has no way of visualising the generated subtree or determining any of its characteristics. The PC-based video shows how a user can ask the system to suggest a test for a node in the tree and how this test can be visualised for the user. This video also shows how the system can complete a subtree for the user, which can be visualised and edited as deemed appropriate by the user. Following these videos, participants are asked, ‘Which system would provide better assistance to you when constructing a decision tree?’
4.3. Methods
4.3.1. Introduction to Decision Trees
- The structure of a DTC.
- How to evaluate univariate splits on internal nodes in a DTC.
- How to traverse a DTC.
- How a DTC can be used to classify an instance.
- How the class label assigned to a leaf node is determined via the majority class of instances reaching that leaf.
- Examples of a DTC incorrectly classifying instances.
- How a user might use the presented visualisation of a DTC to estimate its accuracy.
4.3.2. Introduction to Parallel Coordinates
- How each instance is represented in parallel coordinates.
- The use of colour in each poly-line to indicate the class of an instance.
- The ability for a user to toggle a numerical scale for all axes on and off.
- Coloured buttons at the top of the PC visualisation which allow a user to show/hide individual classes.
4.3.3. Decision Tree Classifiers with Parallel Coordinates
- How selecting a node in the DTC filters the instances shown in PC.
- How univariate splits can be represented in PC as a range on an axis.
- How the poly-lines for all instances in this visualisation originate from an origin point before passing through each axis in PC.
- How all splits in the path to a leaf node are shown as a series of ranges on PC axes.
- How axes are ordered to match the order of attributes appearing in the path to the leaf node.
- How the poly-line for each instance is terminated when it does not match the range on a PC axis.
- How each poly-line is dimmed after it passes through all required ranges to reach a leaf node.
4.3.4. Evaluating Decision Tree Classifiers
4.3.5. Human-in-the-Loop-Learning of Decision Tree Classifiers
5. Results
5.1. Experiment 1
5.2. Experiment 2
5.3. Experiment 3
5.4. Validity Threats
6. Discussion
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
Abbreviations
CART | Classification and Regression Trees |
DTC | Decision Tree Classifier |
XAI | Explainable Artificial Intelligence |
HITL-ML | Human-In-The-Loop Machine Learning |
IML | Interactive Machine Learning |
ML | Machine Learning |
NC | Nested Cavities Algorithm |
PC | Parallel Coordinates |
References
- Dzindolet, M.T.; Peterson, S.A.; Pomranky, R.A.; Pierce, L.G.; Beck, H.P. The role of trust in automation reliance. Int. J. Hum.-Comput. Stud. 2003, 58, 697–718. [Google Scholar] [CrossRef]
- Darlington, K. Aspects of Intelligent Systems Explanation. Univers. J. Control. Autom. 2013, 1, 40–51. [Google Scholar] [CrossRef]
- Dominguez-Jimenez, C. PROSE: An Architecture for Explanation in Expert Systems. In Proceedings of the Third COGNITIVA Symposium on at the Crossroads of Artificial Intelligence, Cognitive Science, and Neuroscience, COGNITIVA 90, Madrid, Spain, 20–23 November 1990; North-Holland Publishing Co.: Amsterdam, The Netherlands, 1991; pp. 305–312. [Google Scholar]
- Swartout, W.R.; Moore, J.D. Explanation in Second Generation Expert Systems. In Second Generation Expert Systems; David, J.M., Krivine, J.P., Simmons, R., Eds.; Springer: Berlin/Heidelberg, Germany, 1993; pp. 543–585. [Google Scholar]
- Ye, R.L.; Johnson, P.E. The Impact of Explanation Facilities on User Acceptance of Expert Systems Advice. MIS Q. 1995, 19, 157–172. [Google Scholar] [CrossRef]
- Wang, N.; Pynadath, D.V.; Hill, S.G. Trust calibration within a human-robot team: Comparing automatically generated explanations. In Proceedings of the 2016 11th ACM/IEEE International Conference on Human-Robot Interaction (HRI), Christchurch, New Zealand, 7–10 March 2016; pp. 109–116. [Google Scholar] [CrossRef]
- Machlev, R.; Perl, M.; Belikov, J.; Levy, K.Y.; Levron, Y. Measuring Explainability and Trustworthiness of Power Quality Disturbances Classifiers Using XAI—Explainable Artificial Intelligence. IEEE Trans. Ind. Inform. 2022, 18, 5127–5137. [Google Scholar] [CrossRef]
- Weiss, S.M.; Kulikowski, C.A. Computer Systems That Learn: Classification and Prediction Methods from Statistics, Neural Nets, Machine Learning, and Expert Systems; Morgan Kaufmann Publishers Inc.: San Francisco, CA, USA, 1991. [Google Scholar]
- Quinlan, J.R. Induction of Decision Trees. Mach. Learn. 1986, 1, 81–106. [Google Scholar] [CrossRef]
- Henery, R.J. Chapter 2: Classification. In Machine Learning, Neural and Statistical Classification; Michie, D., Spiegelhalter, D.J., Taylor, C.C., Eds.; Ellis Horwood: New York, NY, USA, 1995; pp. 6–16. [Google Scholar]
- Kodratoff, Y. Chapter 8: Machine Learning. In Knowledge Engineering Volume I Fundamentals; Adeli, H., Ed.; McGraw-Hill, Inc.: New York, NY, USA, 1990; pp. 226–255. [Google Scholar]
- Ankerst, M.; Elsen, C.; Ester, M.; Kriegel, H.P. Visual Classification: An Interactive Approach to Decision Tree Construction. In Proceedings of the Fifth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD’99, San Diego, CA, USA, 15–18 August 1999; ACM: New York, NY, USA, 1999; pp. 392–396. [Google Scholar] [CrossRef]
- Estivill-Castro, V. Collaborative Knowledge Acquisition with a Genetic Algorithm. In Proceedings of the 9th International Conference on Tools with Artificial Intelligence, ICTAI’97, Newport Beach, CA, USA, 3–8 November 1997; IEEE Computer Society: Newport Beach, CA, USA, 1997; pp. 270–277. [Google Scholar] [CrossRef]
- Ware, M.; Frank, E.; Holmes, G.; Hall, M.; Witten, I.H. Interactive machine learning: Letting users build classifiers. Int. J. Hum.-Comput. Stud. 2001, 55, 281–292. [Google Scholar] [CrossRef]
- Webb, G.I. Integrating machine learning with knowledge acquisition through direct interaction with domain experts. Knowl.-Based Syst. 1996, 9, 253–266. [Google Scholar] [CrossRef] [Green Version]
- Sacha, D.; Kraus, M.; Keim, D.A.; Chen, M. VIS4ML: An Ontology for Visual Analytics Assisted Machine Learning. IEEE Trans. Vis. Comput. Graph. 2019, 25, 385–395. [Google Scholar] [CrossRef]
- Amershi, S.; Cakmak, M.; Knox, W.B.; Kulesza, T. Power to the People: The Role of Humans in Interactive Machine Learning. AI Mag. 2014, 35, 105–120. [Google Scholar] [CrossRef]
- Mosqueira-Rey, E.; Hernández-Pereira, E.; Alonso-Ríos, D.; Bobes-Bascarán, J.; Fernández-Leal, A. Human-in-the-loop machine learning: A state of the art. Artif. Intell. Rev. 2022, 35, 1–19. [Google Scholar] [CrossRef]
- Fails, J.A.; Olsen, D.R. Interactive Machine Learning. In Proceedings of the 8th International Conference on Intelligent User Interfaces, IUI’03, Miami, FL, USA, 12–15 January 2003; Association for Computing Machinery: New York, NY, USA, 2003; pp. 39–45. [Google Scholar] [CrossRef]
- Krak, I.; Barmak, O.; Manziuk, E.; Kudin, H. Approach to Piecewise-Linear Classification in a Multi-dimensional Space of Features Based on Plane Visualization. In Advances in Intelligent Systems and Computing, Proceedings of the International Scientific Conference “Intellectual Systems of Decision Making and Problem of Computational Intelligence”, Kherson, Ukraine, 25–29 May 2020; Lytvynenko, V., Babichev, S., Wójcik, W., Vynokurova, O., Vyshemyrskaya, S., Radetskaya, S., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 35–47. [Google Scholar] [CrossRef]
- Freitas, A.A. Comprehensible classification models: A position paper. SIGKDD Explor. 2013, 15, 1–10. [Google Scholar] [CrossRef]
- Guidotti, R.; Monreale, A.; Ruggieri, S.; Turini, F.; Giannotti, F.; Pedreschi, D. A survey of methods for explaining black box models. ACM Comput. Surv. (CSUR) 2018, 51, 1–42. [Google Scholar] [CrossRef]
- Lavrač, N. Selected techniques for data mining in medicine. Artif. Intell. Med. 1999, 16, 3–23. [Google Scholar] [CrossRef]
- Mues, C.; Huysmans, J.; Vanthienen, J.; Baesens, B. Comprehensible Credit-Scoring Knowledge Visualization Using Decision Tables and Diagrams. In Enterprise Information Systems VI; Springer: Dordrecht, The Netherlands, 2006; pp. 109–115. [Google Scholar] [CrossRef]
- Verbeke, W.; Martens, D.; Mues, C.; Baesens, B. Building comprehensible customer churn prediction models with advanced rule induction techniques. Expert Syst. Appl. 2011, 38, 2354–2364. [Google Scholar] [CrossRef]
- Freitas, A.A.; Wieser, D.; Apweiler, R. On the Importance of Comprehensible Classification Models for Protein Function Prediction. IEEE/ACM Trans. Comput. Biol. Bioinform. 2010, 7, 172–182. [Google Scholar] [CrossRef]
- Sahu, M.; Dash, R. A Survey on Deep Learning: Convolution Neural Network (CNN). In Smart Innovation, Systems and Technologies, Proceedings of the Intelligent and Cloud Computing, Smart Innovation, Systems and Technologies, Bhubaneswar, India, 22–23 October 2021; Mishra, D., Buyya, R., Mohapatra, P., Patnaik, S., Eds.; Springer: Singapore, 2021; Volume 153. [Google Scholar] [CrossRef]
- Cervantes, J.; Garcia-Lamont, F.; Rodríguez-Mazahua, L.; Lopez, A. A comprehensive survey on support vector machine classification: Applications, challenges and trends. Neurocomputing 2020, 408, 189–215. [Google Scholar] [CrossRef]
- Dong, X.; Yu, Z.; Cao, W.; Shi, Y.; Ma, Q. A survey on ensemble learning. Front. Comput. Sci. 2020, 14, 241–258. [Google Scholar] [CrossRef]
- Moore, A.; Murdock, V.; Cai, Y.; Jones, K. Transparent Tree Ensembles. In Proceedings of the 41st International ACM SIGIR Conference on Research & Development in Information Retrieval, SIGIR’18, Ann Arbor, MI, USA, 8–12 July 2018; Association for Computing Machinery: New York, NY, USA, 2018; pp. 1241–1244. [Google Scholar] [CrossRef]
- Liao, Q.V.; Singh, M.; Zhang, Y.; Bellamy, R.K. Introduction to Explainable AI. In Proceedings of the Extended Abstracts of the 2020 CHI Conference on Human Factors in Computing Systems, CHI EA’20, Honolulu, HI, USA, 25–30 April 2020; Association for Computing Machinery: New York, NY, USA, 2020; pp. 1–4. [Google Scholar] [CrossRef]
- Samek, W.; Müller, K.R. Towards Explainable Artificial Intelligence. In Explainable AI: Interpreting, Explaining and Visualizing Deep Learning; Springer International Publishing: Cham, Switzerland, 2019; pp. 5–22. [Google Scholar] [CrossRef]
- Blanco-Justicia, A.; Domingo-Ferrer, J. Machine Learning Explainability through Comprehensible Decision Trees. In Lecture Notes in Computer Science, Proceedings of the Machine Learning and Knowledge Extraction, Canterbury, UK, 26–29 August 2019; Holzinger, A., Kieseberg, P., Tjoa, A.M., Weippl, E., Eds.; Springer International Publishing: Cham, Switzerland, 2019; pp. 15–26. [Google Scholar] [CrossRef]
- Bourguin, G.; Lewandowski, A.; Bouneffa, M.; Ahmad, A. Towards Ontologically Explainable Classifiers. In Lecture Notes in Computer Science, Proceedings of the Artificial Neural Networks and Machine Learning—ICANN 2021, Bratislava, Slovakia, 14–17 September 2021; Farkaš, I., Masulli, P., Otte, S., Wermter, S., Eds.; Springer International Publishing: Cham, Switzerland, 2021; pp. 472–484. [Google Scholar]
- Samek, W.; Binder, A.; Montavon, G.; Lapuschkin, S.; Müller, K. Evaluating the Visualization of What a Deep Neural Network Has Learned. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2660–2673. [Google Scholar] [CrossRef]
- Lakkaraju, H.; Kamar, E.; Caruana, R.; Leskovec, J. Faithful and Customizable Explanations of Black Box Models. In Proceedings of the 2019 AAAI/ACM Conference on AI, Ethics, and Society, AIES’19, Honolulu, HI, USA, 27–28 January 2019; Association for Computing Machinery: New York, NY, USA, 2019; pp. 131–138. [Google Scholar] [CrossRef]
- Ali, H.; Khan, M.S.; Al-Fuqaha, A.; Qadir, J. Tamp-X: Attacking Explainable Natural Language Classifiers through Tampered Activations. Comput. Secur. 2022, 120, 102791. [Google Scholar] [CrossRef]
- Rudin, C. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nat. Mach. Intell. 2019, 1, 206–215. [Google Scholar] [CrossRef]
- Kingsford, C.; Salzberg, S.L. What are decision trees? Nat. Biotechnol. 2008, 26, 1011–1013. [Google Scholar] [CrossRef] [PubMed]
- Wu, X.; Kumar, V.; Quinlan, J.R.; Ghosh, J.; Yang, Q.; Motoda, H.; McLachlan, G.; Ng, A.; Liu, B.; Yu, P.; et al. Top 10 algorithms in data mining. Knowl. Inf. Syst. 2008, 14, 1–37. [Google Scholar] [CrossRef]
- Quinlan, J. C4.5: Programs for Machine Learning; Morgan Kaufmann Publishers: San Mateo, CA, USA, 1993. [Google Scholar]
- Hunt, E.; Martin, J.; Stone, P. Experiments in Induction; Academic Press: New York, NY, USA, 1966. [Google Scholar]
- Tan, P.N.; Steinbach, M.; Kumar, V. Introduction to Data Mining; Addison-Wesley Publishing Co.: Reading, MA, USA, 2006. [Google Scholar]
- Breiman, L.; Friedman, J.; Olshen, R.; Stone, C. Classification and Regression Trees; Wadsworth and Brooks: Monterrey, CA, USA, 1984. [Google Scholar]
- Estivill-Castro, V.; Gilmore, E.; Hexel, R. Human-In-The-Loop Construction of Decision Tree Classifiers with Parallel Coordinates. In Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, SMC, Toronto, ON, Canada, 11–14 October 2020; pp. 3852–3859. [Google Scholar] [CrossRef]
- Inselberg, A. Parallel Coordinates: Visual Multidimensional Geometry and Its Applications; Springer: New York, NY, USA, 2009. [Google Scholar]
- Dudley, J.J.; Kristensson, P.O. A Review of User Interface Design for Interactive Machine Learning. ACM Trans. Interact. Intell. Syst. 2018, 8, 1–37. [Google Scholar] [CrossRef]
- Ramos, R.; Meek, C.; Simard, P.; Suh, J.; Ghorashi, S. Interactive machine teaching: A human-centered approach to building machine-learned models. Hum.-Comput. Interact. 2020, 35, 413–451. [Google Scholar] [CrossRef]
- Estivill-Castro, V.; Gilmore, E.; Hexel, R. Constructing Interpretable Decision Trees Using Parallel Coordinates. In Lecture Notes in Computer Science, Proceedings of the Artificial Intelligence and Soft Computing—19th International Conference, ICAISC, Part II, Zakopane, Poland, 12–14 October 2020; Rutkowski, L., Scherer, R., Korytkowski, M., Pedrycz, W., Tadeusiewicz, R., Zurada, J.M., Eds.; Springer: Berlin/Heidelberg, Germany, 2020; Volume 12416, pp. 152–164. [Google Scholar] [CrossRef]
- Han, J. Data Mining Tasks and Methods: Rule Discovery: Characteristic Rules. In Handbook of Data Mining and Knowledge Discovery; Oxford University Press, Inc.: New York, NY, USA, 2002; pp. 339–344. [Google Scholar]
- Estivill-Castro, V.; Gilmore, E.; Hexel, R. Interpretable decisions trees via human-in-the-loop-learning. In Proceedings of the 20th Australasian Data Mining Conference (AusDM’22), Sydney, Australia, 12–16 December 2022. [Google Scholar]
- Nauta, M.; Trienes, J.; Pathak, S.; Nguyen, E.; Peters, M.; Schmitt, Y.; Schlötterer, J.; van Keulen, M.; Seifert, C. From Anecdotal Evidence to Quantitative Evaluation Methods: A Systematic Review on Evaluating Explainable AI. arXiv 2022, arXiv:2201.08164. [Google Scholar] [CrossRef]
- Ankerst, M.; Ester, M.; Kriegel, H.P. Towards an Effective Cooperation of the User and the Computer for Classification. In Proceedings of the Sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD’00, Boston, MA, USA, 20–23 August 2000; ACM: New York, NY, USA, 2000; pp. 179–188. [Google Scholar] [CrossRef] [Green Version]
- Inselberg, A.; Avidan, T. Classification and visualization for high-dimensional data. In Proceedings of the Sixth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Boston, MA, USA, 20–23 August 2000; ACM: Boston, MA, USA, 2000; pp. 370–374. [Google Scholar] [CrossRef]
- Lai, P.L.; Liang, Y.J.; Inselberg, A. Geometric Divide and Conquer Classification for High-dimensional Data. In Proceedings of the DATA 2012—International Conference on Data Technologies and Applications, Rome, Italy, 25–27 July 2012; SciTePress: Setúbal, Portugal, 2012; pp. 79–82. [Google Scholar] [CrossRef]
- Inselberg, A. III.14 Parallel Coordinates: Visualization, Exploration and Classification of High-Dimensional Data. In Handbook of Data Visualization; Springer: Berlin/Heidelberg, Germany, 2008. [Google Scholar] [CrossRef]
- Gilmore, E.; Estivill-Castro, V.; Hexel, R. More Interpretable Decision Trees. In Lecture Notes in Computer Science, Proceedings of the Hybrid Artificial Intelligent Systems—16th International Conference, HAIS 2021, Bilbao, Spain, 22–24 September 2021; Sanjurjo-González, H., Pastor-López, I., García Bringas, P., Quintián, H., Corchado, E., Eds.; Springer: Berlin/Heidelberg, Germany, 2021; Volume 12886, pp. 280–292. [Google Scholar] [CrossRef]
- Hunt, E. Concept Learning—An Information Processing Problem, 2nd ed.; John Wiley: New York, NY, USA, 1962. [Google Scholar]
- Cohen, P.R.; Feigenbaum, E.A. The Handbook of Artificial Intelligence, Volume III; HeurisTech Press: Stanford, CA, USA, 1982. [Google Scholar]
- Teoh, S.T.; Ma, K. StarClass: Interactive Visual Classification using Star Coordinates. In Proceedings of the Third SIAM International Conference on Data Mining, SIAM, San Francisco, CA, USA, 1–3 May 2003; Volume 112, pp. 178–185. [Google Scholar] [CrossRef]
- Teoh, S.T.; Ma, K.L. PaintingClass: Interactive Construction, Visualization and Exploration of Decision Trees. In Proceedings of the Ninth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, KDD’03, Washington, DC, USA, 24–27 August 2003; Association for Computing Machinery: New York, NY, USA, 2003; pp. 667–672. [Google Scholar] [CrossRef]
- Choo, J.; Lee, H.; Kihm, J.; Park, H. iVisClassifier: An interactive visual analytics system for classification based on supervised dimension reduction. In Proceedings of the 2010 IEEE Symposium on Visual Analytics Science and Technology, Salt Lake City, UT, USA, 25–26 October 2010; pp. 27–34. [Google Scholar] [CrossRef]
- Krak, I.; Barmak, O.; Manziuk, E. Using visual analytics to develop human and machine-centric models: A review of approaches and proposed information technology. Comput. Intell. 2022, 38, 921–946. [Google Scholar] [CrossRef]
- Tam, G.K.L.; Kothari, V.; Chen, M. An Analysis of Machine- and Human-Analytics in Classification. IEEE Trans. Vis. Comput. Graph. 2017, 23, 71–80. [Google Scholar] [CrossRef]
- Chaudhuri, S.; Dayal, U. An Overview of Data Warehousing and OLAP Technology. SIGMOD Rec. 1997, 26, 65–74. [Google Scholar] [CrossRef] [Green Version]
- Hall, M.; Frank, E.; Holmes, G.; Pfahringer, B.; Reutemann, P.; Witten, I.H. The WEKA data mining software: An update. SIGKDD Explor. 2009, 11, 10–18. [Google Scholar] [CrossRef]
- Cantú-Paz, E.; Kamath, C. Inducing oblique decision trees with evolutionary algorithms. IEEE Trans. Evol. Comput. 2003, 7, 54–68. [Google Scholar] [CrossRef]
- Heath, D.; Kasif, S.; Salzberg, S. Induction of Oblique Decision Trees. In Proceedings of the 13th International Joint Conference on Artificial Intelligence, Chambery, France, 28 August–3 September 1993; Morgan Kaufmann: Chambéry, France, 1993; pp. 1002–1007. [Google Scholar]
- Murthy, S.K.; Kasif, S.; Salzberg, S. A System for Induction of Oblique Decision Trees. J. Artif. Int. Res. 1994, 2, 1–32. [Google Scholar] [CrossRef]
- Brodley, C.E.; Utgoff, P.E. Multivariate Decision Trees. Mach. Learn. 1995, 19, 45–77. [Google Scholar] [CrossRef]
- Koziol, M.; Wozniak, M. Multivariate Decision Trees vs. Univariate Ones. In Computer Recognition Systems 3; Kurzynski, M., Wozniak, M., Eds.; Advances in Intelligent and Soft Computing; Springer: Berlin/Heidelberg, Germany, 2009; Volume 57, pp. 275–284. [Google Scholar] [CrossRef]
- Hurley, P.J. A Concise Introduction to Logic, 12th ed.; Cengage: Boston, MA, USA, 2015. [Google Scholar]
- Palan, S.; Schitter, C. Prolific.ac—A subject pool for online experiments. J. Behav. Exp. Financ. 2018, 17, 22–27. [Google Scholar] [CrossRef]
- Dua, D.; Graff, C. UCI Machine Learning Repository; University of California, School of Information and Computer Science: Irvine, CA, USA, 2017; Available online: http://archive.ics.uci.edu/ml (accessed on 3 April 2020).
- Shapiro, S.S.; Wilk, M.B. An analysis of variance test for normality (complete samples). Biometrika 1965, LII, 591–611. [Google Scholar] [CrossRef]
- Hui, E.G.M. Learn R for Applied Statistics: With Data Visualizations, Regressions, and Statistics; Springer: Berkeley, CA, USA, 2019. [Google Scholar]
- Nguyen, T.D.; Ho, T.; Shimodaira, H. Interactive Visualization in Mining Large Decision Trees. In Lecture Notes in Computer Science, Proceedings of the Knowledge Discovery and Data Mining, Current Issues and New Applications, 4th Pacific-Asia Conference PADKK 2000, Kyoto, Japan, 18–20 April 2000; Springer: Kyoto, Japan, 2000; Volume 1805, pp. 345–348. [Google Scholar] [CrossRef]
- Jiang, L.; Liu, S.; Chen, C. Recent research advances on interactive machine learning. J. Vis. 2019, 22, 401–417. [Google Scholar] [CrossRef] [Green Version]
Tree No. | Dataset | Accuracy |
---|---|---|
1 | Wine | 69 |
2 | Wine | 99 |
3 | Wine | 90 |
4 | Cryotherapy | 72 |
5 | Cryotherapy | 92 |
6 | Seeds | 72 |
7 | Seeds | 95 |
8 | Seeds | 82 |
Question No. | Question |
---|---|
Q1 | Which system do you believe provides a better method of finding splits to build a decision tree classifier? |
Q2 | Which system allows you to better navigate and understand the current state of a decision tree classifier as it is being constructed? |
Q3 | Which system allows you to more easily determine how often a tree will predict the class correct class of an instance? |
Q4 | Which system allows you to more easily find nodes in a decision tree classifier that need additional splits? |
Q5 | Which system would provide better assistance to you when constructing a decision tree? |
Q6 | Based on the videos you have seen, which system would you prefer to use to build a decision tree classifier? |
Group | Mean Accuracy Differences | Mean Accuracy Time (s) | Mean | Mean Leaf Choice Time (s) |
---|---|---|---|---|
A | 17.8% | 80.2 | 0.259 | 17.6 |
B | 19.0% | 49.3 | 0.282 | 17.1 |
Group | Mean Accuracy | Mean Time per Leaf (s) |
---|---|---|
A | 77.5% | 24.0 |
B | 86.7% | 6.7 |
Question | Q1 | Q2 | Q3 | Q4 | Q5 | Q6 |
---|---|---|---|---|---|---|
Strongly PC-based system | 38 | 42 | 39 | 44 | 58 | 57 |
Somewhat PC-based system | 32 | 27 | 32 | 33 | 25 | 26 |
Strongly UserClassifier | 7 | 7 | 6 | 4 | 5 | 7 |
Somewhat UserClassifier | 23 | 21 | 21 | 13 | 8 | 7 |
No differences | 4 | 7 | 6 | 10 | 8 | 7 |
Subtotal PC-based system | 70 | 69 | 71 | 77 | 83 | 83 |
Subtotal UserClassifier | 30 | 28 | 27 | 17 | 13 | 14 |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Estivill-Castro, V.; Gilmore, E.; Hexel, R. Constructing Explainable Classifiers from the Start—Enabling Human-in-the Loop Machine Learning. Information 2022, 13, 464. https://doi.org/10.3390/info13100464
Estivill-Castro V, Gilmore E, Hexel R. Constructing Explainable Classifiers from the Start—Enabling Human-in-the Loop Machine Learning. Information. 2022; 13(10):464. https://doi.org/10.3390/info13100464
Chicago/Turabian StyleEstivill-Castro, Vladimir, Eugene Gilmore, and René Hexel. 2022. "Constructing Explainable Classifiers from the Start—Enabling Human-in-the Loop Machine Learning" Information 13, no. 10: 464. https://doi.org/10.3390/info13100464
APA StyleEstivill-Castro, V., Gilmore, E., & Hexel, R. (2022). Constructing Explainable Classifiers from the Start—Enabling Human-in-the Loop Machine Learning. Information, 13(10), 464. https://doi.org/10.3390/info13100464