Next Article in Journal
Effect of Pole Head Profile on Torque Performance of a Disc-Type Multipole Magnetorheological Brake
Next Article in Special Issue
Factors Associated with Medial Elbow Torque Measured Using a Wearable Sensor in Junior High School Baseball Pitchers
Previous Article in Journal
The Machine-Learning-Based Prediction of the Punching Shear Capacity of Reinforced Concrete Flat Slabs: An Advanced M5P Model Tree Approach
Previous Article in Special Issue
Material Design in Implantable Biosensors toward Future Personalized Diagnostics and Treatments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Agile Machine Learning Model Development Using Data Canyons in Medicine: A Step towards Explainable Artificial Intelligence and Flexible Expert-Based Model Improvement

1
Faculty of Electrical Engineering and Computer Science, University of Maribor, 2000 Maribor, Slovenia
2
Community Healthcare Center Dr. Adolf Drolc Maribor, 2000 Maribor, Slovenia
3
Alma Mater Europaea—ECM, 2000 Maribor, Slovenia
4
Science and Research Center Koper, 6000 Koper, Slovenia
5
Faculty of Natural Sciences and Mathematics, University of Maribor, 2000 Maribor, Slovenia
6
Faculty of Health and Social Sciences Slovenj Gradec, 2380 Slovenj Gradec, Slovenia
7
University Clinical Centre Maribor, 2000 Maribor, Slovenia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(14), 8329; https://doi.org/10.3390/app13148329
Submission received: 3 July 2023 / Revised: 13 July 2023 / Accepted: 17 July 2023 / Published: 19 July 2023
(This article belongs to the Special Issue Intelligent Diagnosis and Decision Support in Medical Applications)

Abstract

:

Featured Application

In the field of medicine, daily high-stake decision-making scenarios arise. To enhance these decision-making processes, the utilization of modern machine learning techniques and their infusion with expert knowledge in a seamless manner while adhering to agile principles enables swift action and facilitates an understanding of the obtained knowledge, fostering trust in the decision-making process.

Abstract

Over the past few decades, machine learning has emerged as a valuable tool in the field of medicine, driven by the accumulation of vast amounts of medical data and the imperative to harness this data for the betterment of humanity. However, many of the prevailing machine learning algorithms in use today are characterized as black-box models, lacking transparency in their decision-making processes and are often devoid of clear visualization capabilities. The transparency of these machine learning models impedes medical experts from effectively leveraging them due to the high-stakes nature of their decisions. Consequently, the need for explainable artificial intelligence (XAI) that aims to address the demand for transparency in the decision-making mechanisms of black-box algorithms has arisen. Alternatively, employing white-box algorithms can empower medical experts by allowing them to contribute their knowledge to the decision-making process and obtain a clear and transparent output. This approach offers an opportunity to personalize machine learning models through an agile process. A novel white-box machine learning algorithm known as Data canyons was employed as a transparent and robust foundation for the proposed solution. By providing medical experts with a web framework where their expertise is transferred to a machine learning model and enabling the utilization of this process in an agile manner, a symbiotic relationship is fostered between the domains of medical expertise and machine learning. The flexibility to manipulate the output machine learning model and visually validate it, even without expertise in machine learning, establishes a crucial link between these two expert domains.

1. Introduction

Machine learning has become a staple in several fields and industries, most prominently in the field of medicine. Applications in medicine include but are not limited to Anti-cancer Drug Discovery [1], screening and diagnosis of diabetes [2], HIV clinical research and care [3], medical image analysis [4], automated seizure detection [5], and COVID-19 pandemic mitigation [6,7]. Discovering patterns in large datasets and the consequential information gain that allows experts in multiple fields of medicine to apply those findings to save lives is an indispensable quality, especially in this day and age where big data is becoming the standard data quantity [8].
While machine learning provides a clear solution to an enormous problem, it does so at a cost. The transparency and the ability to explain the decisions behind all black-box models is a challenge that is being tackled and intensely studied in over recent years under the umbrella term explainable as artificial intelligence (XAI) [9]. The aim of XAI is to add an explanation and a layer of transparency to models that are inherently black-box and therefore do not enable the user to infer the logic behind the decisions made by the model. On the other hand, XAI also promotes the usage and development of white-box models [10] that, by default, enable the user to grasp the idea behind the decision. The white-box machine learning approach focuses on developing transparent and interpretable models, allowing humans to understand and validate their decision-making process. By examining internal variables, coefficients, or rules, researchers gain insights into how and why predictions are made. This approach facilitates model debugging, error analysis, and bias identification, enhancing performance and fairness. It is particularly useful in domains like medicine and finance, where interpretability is crucial for regulatory compliance and ethical considerations. The white-box approach promotes transparency, accountability, and trustworthiness in machine learning systems. White-box models differ by the type of visualization and the possibility of interaction and allow professionals in the domain of application to visualize and, if possible, interact with the resulting model.
In software development, various methods promote efficient, robust, and agile development. Among those, the most prevalent principle is the agile development process [11] which can be applied to almost any area of interest where there is a need to be able to quickly, efficiently, robustly, and continuously develop or manage a solution. Because the main aim of the agile development process is software development, it is worth noting that all machine learning algorithms are software and that a system consisting of several software components is still software. Agility, speed, and adaptation are the most prevalent needs in medical systems when agile as a process is concerned [12]. Introducing this popular software development principle to the area of white-box machine learning algorithms and, with this, allowing medical experts to ad hoc improve on the final model presents a unique possibility for expert knowledge infusion into the end models.
The development of an agile machine learning expert-driven framework is highly dependent on the white-box algorithm and its representation of the model, the need for XAI in the field of application, and the expertise of the end user concerning the train data. Data canyons were the applied machine learning white-box model because they provide a graphical explanation layer that can be easily understood and interpreted. Additionally, Data canyons allow easy manipulation of attributes and their inclusion or restriction while offering a clear instance-based visualization of instance alignment. Medicine presents a field of application of machine learning algorithms where there is a great need for XAI because of the high-risk decision-making nature of the domain. Providing that the end user of the agile framework is an expert in the domain of application, the framework presents a great potential for creating new expert-infused machine learning models. All three presented parts are equally important for the development of an agile machine learning framework. Without an intuitive and transparent white-box model, the expert cannot easily understand and manipulate the model. If the field of application has no inherent need for transparency and the decision-making process has low risks associated with it, there is no need for white-box machine learning algorithms or agile expert-infused models. Lastly, if the user is not a domain expert, his knowledge and experiences cannot be infused into the final model.
The field of XAI is gradually gaining traction, as evidenced by the increasing number of publications in recent years. A literature search conducted on Scopus using the term “explainable artificial intelligence” revealed that the most relevant subject area for its application is medicine, preceded by computer science, engineering, and mathematics, which primarily focus on method development. Within the domain of medicine, a significant majority (over 90%) of the publications have been published within the last 3 years, indicating a current and pressing interest in this area. Notably, one highly cited article has expressed concerns regarding the limitations of existing XAI approaches [13], while other relevant articles have primarily focused on showcasing XAI applications in medicine. However, the inclusion of the term “agile” in the search string did not yield any results, suggesting a lack of research on agile frameworks for XAI in the context of medicine. This literature survey highlights the currency and relevance of XAI in medicine and indicates the need for further advancements and novel approaches in this domain, including the exploration of agile methodologies.
The aim of this paper is to present the process of continuous improvement of a novel white-box machine learning algorithm, namely Data canyons, using agile medical expert interaction. Data canyons allow for a unique presentation of the created model that enables experts to interpret the model as such and manipulate it based on their own understanding and expert knowledge in order to better understand the dynamics of the attributes interaction and to infuse their knowledge that lies outside of the boundaries of the used data. This approach combines XAI’s objectives and agile development principles to form a symbiotic human–computer machine learning agent. In this test scenario, a labeled dataset using patient bloodwork obtained during infarction and reinfarction was used to create a model to classify the possibility of reinfarction.

2. Materials and Methods

The presented framework encompasses two main parts: the white-box machine learning model and the interactive visualization and manipulation user interface, as can be seen in Figure 1. The domain expert provides the data to the machine learning algorithm. In the first instance, the algorithm runs with the base parameters and with all parameters. The algorithm separately creates a canyon for each class using the provided data and parameters. The stored metadata and the model are forwarded to the visualization and manipulation user interface. The expert can then visually inspect the canyons through the interactive 3D plots for all canyons. The canyons can be easily rotated and enlarged. Individual instances can be viewed in the canyons to test how they relate to each canyon. The expert can evaluate the results through the visual inspection, testing of instances, and standard metrics and can then adjust the inclusion of attributes and the parameters of the machine learning algorithm and start a new phase.

2.1. The Data

The data for presenting the approach of expert-driven agile machine learning model creation were gathered from University Medical Centre Maribor, which is a Slovenian-based non-profit public health care institute performing health care services on the secondary and tertiary level. The data were anonymized, and the ethical comity of University Medical Centre Maribor authorized the use of the data. The data in question consist of detailed blood work of patients, including the diagnosis. To examine a meaningful case, we focused on a subset of this data, where the aim was to distinguish between people with infarction and people with reinfarction. The final dataset consists of 2274 entries, of which 308 are examples of reinfarctions. Each entry consists of 8 attributes: age, gender, S-Lp(a), S-cholesterol, S-HDL-cholesterol, S-LDL-cholesterol, S-triglycerides, and whether or not it is classified as reinfarction.

2.2. Data Canyons

As the aim of this article is to present the symbiotic cooperation between computer science, medical expertise, and an agile approach to manipulation and improvement of the output machine learning model, we will only briefly introduce Data canyons to understand the underlying logic and the visualization concept. Data canyons are a novel machine learning approach with a human-interpretable output model. Because it is inherently interpretable, it aligns well with the concepts of XAI. As the name suggests, the concepts of Data canyons are derived from the natural phenomenon of river canyons. The idea of transferring nature’s concepts into machine learning is not new; there are several concepts and algorithms of machine learning and AI that owe their existence to nature [14,15].
The base elements of the natural phenomenon, namely a river canyon, which were utilized in the machine learning algorithm were the river stream, the land mass, and time. The test data represent the river stream and are, throughout the algorithm, referred to as the data stream. The land mass is represented via a 3D surface plot where we represent the terrain of the canyon. What remains is the element of time which, in the case of Data canyons, is represented via the iterative processing of each data instance as a separate data stream in the given time frame. Data canyons are created for each class separately. For n classes, the output gives us n Data canyons.
All Data canyons consist of sections that represent the sequential attributes from the train data and can be thought of as cross sections of the canyon that are sequentially equidistant. The terrain of the Data canyon sections is created for each attribute separately, measuring the occurrences of values between the maximum and minimum values of the current attribute. The bridging of the distance between adjoining attributes is created using the same principle as the attribute sections, except that the value is dictated by the connection between the two values of the adjoining attributes. There are two main variables that can be set to fine-tune the performance of the canyons, the length between adjoining attributes, and the width of the canyon. The length between attributes increases or decreases the emphasis on the importance of the connection between two parameters. Longer distance puts the emphasis on the connection between two parameters. A shorter distance puts the emphasis on the values of the sections. The width parameter controls the amount of detail the canyon should capture. Wider canyons focus on more detail, while narrow canyons focus on the higher abstraction of parameters. The difference in the changes of the output canyons, if the main attributes of the canyons are focused on length and width or are neutral, can be observed in Figure 2, Figure 3 and Figure 4.
The algorithm gives a visual output of the model in the form of a 3D surface plot where the depth is presented using gradient color changes to mimic the natural appearance of canyons. The depth of each Data canyon presents the occurrence of attribute values in the given area of the canyon. The width and depth are normalized to minimize the influence of unbalanced datasets on the final output. This visualization technique presents the end user with the ability to visually compare any given instance correlation to the output canyons in the form of a color-coded scatter plot. The scatter plot is similar to the canyon coded with a gradient except that there is an interchange from green to red and vice versa, where green presents a perfect fit to the canyon and red a complete mismatch. An example can be seen in Figure 5.
Data canyons provide an objective method to adjust parameters according to the expert’s knowledge, the visualization of output models, the correlation between the canyon and any given instance, and regular metrics [16] for algorithms of machine learning.

2.3. Visualization Framework and Explainability

The base Data canyon algorithm is written in Python [17], where the visualization is performed using Plotly [18] and Dash [19]. Plotly is a Python graphing library that enables data scientists to create complex interactive graphs. Dash, on the other hand, is a framework for the rapid development of data apps in several programming languages that allows for a full-stack web app development with interactive data visualization. The usage of Plotly and Dash enables data scientists to quickly and dynamically represent data and create an additional layer of interaction for the end user. The utilization of those technologies enabled the creation of the interactive platform used to visualize and manipulate the model of Data canyons. The interface can be hosted as a local web page or as a regular web page that can be accessed via the web address, depending on the needs of the end user. The main areas on the web page focus on the interactive presentation of the model and the correlation between the given test instance and the bed of the Data canyons, as shown in Figure 6. The metadata of the test instance is also presented to provide the experts with a comprehensive picture, as seen in Figure 7. Additionally, some metadata of the canyons can be seen if the expert deems it important. The metadata contains the maximum and minimum values of each Data canyon which is presented in Figure 8. The last portion of the interface is the interaction part, where the end user can select which parameters of the train set he will incorporate to create the Data canyon, as shown in Figure 9. Accepting the selected attributes instantly presents the end user with the new model, including the first test instance.

2.4. Medical Expert-Driven Agile Model Development and Improvement

Experts in the field where machine learning algorithms are applied are rarely also experts in the field of computer science and machine learning. Therefore, it is important that the interaction with machine learning algorithms and their visualization is presented in a transparent and understandable manner. Base principles of user experience (UX) [20] need to be considered and, if feasible, applied in order to achieve the most interactive and coherent interface. Experts need to be able to relate to the main concepts and elements of the user interface (UI) without being overwhelmed and confused. The concepts of UX go hand in hand with agile principles of development since agile development inherently demands swift transitions and agility, which can be only achieved if the development environment, in our case the visualization and interaction platform, allows for such development. The base algorithm is like any other machine learning algorithm. It is a complex algorithm and can, in most cases, be only used as an input–output program, adding an abstract layer for interaction, at least for white-box models. The input–output model transforms into an interactive model, where the UX-centered interface allows for an agile exploration and visualization platform. There are other platforms that allow for similar testing and manipulation [21]. However, what makes our solution special and different is the unique visualization layer of Data canyons that adds a simple graphical interpretation model and the possibility to combine it with on-the-fly interaction through the expert.

2.5. The Interviews

Several methods for evaluating XAI methods exist [22], and a recent effort has been made to create a model usability evaluation framework for explainable artificial intelligence (MUsE) [23] that featured a set of questions that were used in interviews to assess the usability of evaluation frameworks. Since white-box machine learning algorithms and their visual representation are subject to the same scrutiny as XAI-based interpretation tools such as Local Interpretable Model-Agnostic Explanations (LIME) [24] and SHapley Additive exPlanations (SHAP) [25], the approach of MUsE was therefore consequently applicable to the approach presented in this paper. The only exception were questions that were not applicable to the presented solution and were therefore left out. The interviews were focused on medical experts with extensive knowledge of the medical matter in question. In total, the interview was held with five medical experts. Each interview took 15 min, with a 5 min individual introduction before the interview to familiarize the participants with the logic behind the visualization of Data canyons and the individual parts of the web page. After the initial 15 min introduction, all participants were confident in their ability to use and explore the presented system.

3. Results

The interview was preceded by the previously mentioned introduction, after which the participants were asked the following questions. The questions were then evaluated and are shown in Table 1. The rating column shows the ratings of individual participants based on the question regarding the interpretability of the results of the provided predictions. The interpretability of the Data canyons output was rated with an average of 8.8, while the average rating for LIME in the study that introduced MUsE [23] was 7.08.

3.1. The Interview

What do you see in this graph?
Almost all participants initially knew what they were looking at and what they were looking for. Those who were initially struggling to understand the whole picture quickly made sense of it after they found the relevant attributes and then understood the relation between the instance and the appropriate Data canyons.
Which feature influences the prediction and how?
All participants understood the relationship between the color-coded instance and the influence of the attributes in the appropriate Data canyon since Data canyons are 3D figures that medical experts are used to interpreting.
Do you know why the model made this prediction?
All participants understood that the relation between the colors of the instance in one canyon in relation to the colors of the same instance in another canyon was the key component in the decision-making process.
How well can you interpret the results of the prediction of the graph on an increasing scale from 1–10?
Most participants had no trouble interpreting the results, while some had to glance at the provided metadata to interpret the results confidently. The average score was 8.8, which is a consequence of all the parts of the agile framework.
Is there anything that stands out as strange or unusual?
There were no noteworthy remarks made when confronted with this question.
The overall evaluation of the presented system was very positive. Participants had a clear understanding of what they were trying to achieve and a visually adequate representation that they could interpret. What stood out was the positive attitude towards the color-coded nature of the Data canyon and of the instance.

3.2. How Are Data Canyons and the Integrated System in Terms of Achieving Model Interpretability?

Here the focus is effectiveness as to three main factors: completion, accuracy, and negative consequences.
(a)
How complete is the explanation on a local level?
The explanation is relatable and can be concisely summarized using the provided metadata explored.
(b)
How complete is the explanation on a global level?
Since the system was integrated with a white-box machine learning algorithm and not an interpretation tool of a black-box algorithm, the explanation on a global level does not suffer.
(c)
Could accurate results be misinterpreted?
Looking at the whole output that is provided by the presented solution, a misinterpretation by a medical expert is highly unlikely since the metadata combined with the visualization of the white-box algorithm presents a complete picture. However, there is always a chance of misinterpretation, and therefore, in medicine, a conclusion is never reached based only on the output of a machine learning algorithm.
(d)
What negative consequences arise from a misinterpretation?
In the field of medicine, the output of a support system is only one piece of the puzzle. It is the medical expert’s duty to look at the whole picture and form his own opinion. Therefore, a misinterpretation should not have fatal consequences.

3.3. What Resources Are Consumed in Order to Achieve Interpretability?

Here the focus was on resource efficiency, which was derived from task time, time efficiency, cost-effectiveness, productive time ratio, unnecessary actions, and fatigue.
(a)
How much time does it take to use the presented system?
The concept behind the visualization is relatively trivial and intuitive. Therefore, interpretation and adaptation to the usage take very little time.
(b)
What other costs are involved?
From an application standpoint, it would have to be integrated into existing national medical systems, and for wide use, workshops for medical professionals would need to be held.
(c)
Does this process cause fatigue?
The picture says more than a thousand words hold true in this case of Data canyons as we can grasp the main decision at a glance, which takes very little effort.

3.4. How Satisfying Is the Application of the Presented System?

This last part is highly dependent on the user of the system since it focuses on his satisfaction which can be determined on a per-user basis.
(a)
Do we have a positive or negative attitude towards the tool?
Since the solution is a standalone system that is served through a web page, it is easily accessible and easy to use. However, for the same reasons, changes to the underlying algorithm or any adjustment would be out of the skill base of the end user and would have to be performed by someone with a high degree of knowledge in computer science.
(b)
What emotions arise from using it?
The tool arouses favorable emotions since it increases confidence and allows for dynamic manipulation from the end-user site, which implies the importance of expert knowledge in the creation of quality decision-making algorithms.
(c)
How satisfying is the final result?
The overall experience of working with this tool is positive, and the possibility of interaction visually and the agile approach make for a new and exciting expert-driven framework.

4. Discussion

Given the lack of XAI solutions and because of the fact that both the area of XAI and the MuSE framework are relatively new, our results can only be compared to the results of the LIME algorithm [23]. The responses from the participants in the interview are mostly very positive and are an indicator that using agile principles and an inclusive approach where we allow experts to infuse their knowledge into the final output model while presenting the model using a white-box machine learning visualization is the right path to try and tackle some challenges of XAI. Comparing the results of this paper with the results presented in the study where LIME was tested for its XAI effectiveness using MuSE paints a very positive picture since the results are better in this study. The reasons for that might be the presentation layer and their white-box algorithm’s nature. Additionally, Data canyons create a unique perspective on data transparency; being a multidimensional visualization tool allows for more transparency than 2D-based explanation tools like Shapley and SLIM. Each layer of transparency that we can add to increase the confidence of end users in the decision made by a machine learning model should be encouraged, and a combination of traditional white-model algorithms with tools like Shapley and SLIM should be encouraged. The participants in the interview gave some suggestions about where they would apply the presented framework and had an enthusiastic attitude towards tools that enforce XAI. One potential area of additional research has presented itself with the potential of Data canyons to be used as an XAI presentation framework for black-box models. Providing tools that enrich and diversify the area of XAI helps to improve the feasibility of the application of machine learning algorithms since there is a great need for transparent machine learning solutions in medicine [26,27,28,29,30].

5. Conclusions

The use of Data canyons in agile XAI driven system, at least in medicine, presents a unique opportunity to form a symbiotic framework that, while boosting confidence in the solution, allows for the expert to be an integral part in developing a complex decision system based on both the medical data and the experience and knowledge of the medical expert. The need to understand or interpret the decisions behind machine learning algorithms in medicine is considerable, having alternative solutions that expand the range of the medical expert’s arsenal in the struggle to extend the frontiers of knowledge. The agile approach to actively include medical experts in the development process of machine learning models and with that allowing for expert knowledge infusion is one way to combine XAI and expert validation. The current findings present an opportunity to combine other white-box machine learning algorithms into an agile XAI-oriented framework that allows for expert knowledge infusion. Additionally, there are several XAI-based explanation systems for black-box models that could either be used with white-box models as an additional layer of XAI or as a mechanism to introduce black-box models into an agile XAI focused interactive expert-driven system for development of machine learning algorithms.

Author Contributions

Conceptualization, B.Ž. and P.K.; methodology, B.Ž.; software, B.Ž.; validation, J.Z., H.B.V., T.Z., and P.K.; formal analysis, P.K., D.Š. and J.Z.; investigation, B.Ž., J.Z., H.B.V., and P.K.; writing—original draft preparation, B.Ž.; writing—review and editing, P.K., T.Z. and H.B.V.; visualization, B.Ž.; supervision, J.Z., D.Š., and P.K.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy and ethical reasons.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, L.; Song, Y.; Wang, H.; Zhang, X.; Wang, M.; He, J.; Li, S.; Zhang, L.; Li, K.; Cao, L. Advances of Artificial Intelligence in Anti-Cancer Drug Design: A Review of the Past Decade. Pharmaceuticals 2023, 16, 253. [Google Scholar] [CrossRef] [PubMed]
  2. Uddin, M.J.; Ahamad, M.M.; Hoque, M.N.; Walid, M.A.A.; Aktar, S.; Alotaibi, N.; Alyami, S.A.; Kabir, M.A.; Moni, M.A. A Comparison of Machine Learning Techniques for the Detection of Type-2 Diabetes Mellitus: Experiences from Bangladesh. Information 2023, 14, 376. [Google Scholar] [CrossRef]
  3. Bisaso, K.R.; Anguzu, G.T.; Karungi, S.A.; Kiragga, A.; Castelnuovo, B. A Survey of Machine Learning Applications in HIV Clinical Research and Care. Comput. Biol. Med. 2017, 91, 366–371. [Google Scholar] [CrossRef]
  4. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [Green Version]
  5. Abbasi, B.; Goldenholz, D.M. Machine Learning Applications in Epilepsy. Epilepsia 2019, 60, 2037–2047. [Google Scholar] [CrossRef]
  6. Bhattacharya, S.; Reddy Maddikunta, P.K.; Pham, Q.-V.; Gadekallu, T.R.; Krishnan, S.S.R.; Chowdhary, C.L.; Alazab, M.; Jalil Piran, M. Deep Learning and Medical Image Processing for Coronavirus (COVID-19) Pandemic: A Survey. Sustain. Cities Soc. 2021, 65, 102589. [Google Scholar] [CrossRef]
  7. Kushwaha, S.; Bahl, S.; Bagha, A.K.; Parmar, K.S.; Javaid, M.; Haleem, A.; Singh, R.P. Significant Applications of Machine Learning for COVID-19 Pandemic. J. Ind. Integr. Manag. 2020, 5, 453–479. [Google Scholar] [CrossRef]
  8. Santosh, K.C.; Ghosh, S. COVID-19 Imaging Tools: How Big Data Is Big? J. Med. Syst. 2021, 45, 71. [Google Scholar] [CrossRef]
  9. Barredo Arrieta, A.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef] [Green Version]
  10. Clement, T.; Kemmerzell, N.; Abdelaal, M.; Amberg, M. XAIR: A Systematic Metareview of Explainable AI (XAI) Aligned to the Software Development Process. Mach. Learn. Knowl. Extr. 2023, 5, 78–108. [Google Scholar] [CrossRef]
  11. Abrahamsson, P.; Salo, O.; Ronkainen, J.; Warsta, J. Agile Software Development Methods: Review and Analysis. arXiv 2002, arXiv:1709.08439. [Google Scholar]
  12. Kokol, P.; Blažun Vošner, H.; Kokol, M.; Završnik, J. Role of Agile in Digital Public Health Transformation. Front. Public Health 2022, 10, 899874. [Google Scholar] [CrossRef] [PubMed]
  13. Ghassemi, M.; Oakden-Rayner, L.; Beam, A.L. The False Hope of Current Approaches to Explainable Artificial Intelligence in Health Care. Lancet Digit. Health 2021, 3, e745–e750. [Google Scholar] [CrossRef]
  14. Fister, I., Jr.; Yang, X.-S.; Fister, I.; Brest, J.; Fister, D. A Brief Review of Nature-Inspired Algorithms for Optimization. arXiv 2013, arXiv:1307.4186v1. [Google Scholar]
  15. Zang, H.; Zhang, S.; Hapeshi, K. A Review of Nature-Inspired Algorithms. J. Bionic. Eng. 2010, 7, S232–S237. [Google Scholar] [CrossRef]
  16. Hand, D.J. Measuring Classifier Performance: A Coherent Alternative to the Area under the ROC Curve. Mach. Learn. 2009, 77, 103–123. [Google Scholar] [CrossRef] [Green Version]
  17. Sanner, M.F. Python: A Programming Language for Software Integration and Development. J. Mol. Graph. Model. 1999, 17, 57–61. [Google Scholar]
  18. Plotly Python Graphing Library. Available online: https://plotly.com/python/ (accessed on 16 March 2023).
  19. Introduction, Dash for Python Documentation. Plotly. Available online: https://dash.plotly.com/introduction (accessed on 16 March 2023).
  20. Yablonski, J. Laws of UX: Using Psychology to Design Better Products & Services; O’Reilly Media: Sebastopol, CA, USA, 2020. [Google Scholar]
  21. Attwal, K.P.S.; Dhiman, A.S. Exploring Data Mining Tool-Weka and Using Weka to Build and Evaluate Predictive Models. Adv. Appl. Math. Sci. 2020, 19, 451–469. [Google Scholar]
  22. Moreno-Sánchez, P. Methods and Metrics for Evaluating Explainable Artificial Inteligence in Healthcare Domain. Bachelor’s Thesis, Tampere University, Tampere, Finland, 2023. [Google Scholar]
  23. Dieber, J.; Kirrane, S. A Novel Model Usability Evaluation Framework (MUsE) for Explainable Artificial Intelligence. Inf. Fusion 2022, 81, 143–153. [Google Scholar] [CrossRef]
  24. Dieber, J.; Kirrane, S. Why Model Why? Assessing the Strengths and Limitations of LIME. arXiv 2020, arXiv:2012.00093. [Google Scholar]
  25. Fryer, D.; Strümke, I.; Nguyen, H. Shapley Values for Feature Selection: The Good, the Bad, and the Axioms. IEEE Access 2021, 9, 144352–144360. [Google Scholar] [CrossRef]
  26. Antoniadi, A.M.; Du, Y.; Guendouz, Y.; Wei, L.; Mazo, C.; Becker, B.A.; Mooney, C. Current Challenges and Future Opportunities for XAI in Machine Learning-Based Clinical Decision Support Systems: A Systematic Review. Appl. Sci. 2021, 11, 5088. [Google Scholar] [CrossRef]
  27. Van der Velden, B.H.; Kuijf, H.J.; Gilhuijs, K.G.; Viergever, M.A. Explainable Artificial Intelligence (XAI) in Deep Learning-Based Medical Image Analysis. Med. Image Anal. 2022, 79, 102470. [Google Scholar] [CrossRef] [PubMed]
  28. Tjoa, E.; Guan, C. A Survey on Explainable Artificial Intelligence (Xai): Toward Medical Xai. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 4793–4813. [Google Scholar] [CrossRef] [PubMed]
  29. de Vries, B.M.; Zwezerijnen, G.J.C.; Burchell, G.L.; van Velden, F.H.P.; Menke-van der Houven van Oordt, C.W.; Boellaard, R. Explainable Artificial Intelligence (XAI) in Radiology and Nuclear Medicine: A Literature Review. Front. Med. 2023, 10, 1180773. [Google Scholar] [CrossRef]
  30. Borys, K.; Schmitt, Y.A.; Nauta, M.; Seifert, C.; Krämer, N.; Friedrich, C.M.; Nensa, F. Explainable AI in Medical Imaging: An Overview for Clinical Practitioners—Saliency-Based XAI Approaches. Eur. J. Radiol. 2023, 162, 110787. [Google Scholar] [CrossRef]
Figure 1. Agile framework diagram.
Figure 1. Agile framework diagram.
Applsci 13 08329 g001
Figure 2. Example of canyons where the length is prioritized.
Figure 2. Example of canyons where the length is prioritized.
Applsci 13 08329 g002
Figure 3. Example of canyons where the width is prioritized.
Figure 3. Example of canyons where the width is prioritized.
Applsci 13 08329 g003
Figure 4. Example of canyons where the length and width are neutral.
Figure 4. Example of canyons where the length and width are neutral.
Applsci 13 08329 g004
Figure 5. Visualization of an instance on a Data canyon.
Figure 5. Visualization of an instance on a Data canyon.
Applsci 13 08329 g005
Figure 6. Visualization of the output model and the instance in the framework.
Figure 6. Visualization of the output model and the instance in the framework.
Applsci 13 08329 g006
Figure 7. Visualization of the data of the test instance.
Figure 7. Visualization of the data of the test instance.
Applsci 13 08329 g007
Figure 8. Visualization of each canyons metadata.
Figure 8. Visualization of each canyons metadata.
Applsci 13 08329 g008
Figure 9. Attribute selection.
Figure 9. Attribute selection.
Applsci 13 08329 g009
Table 1. Summary of the participants’ understanding of the presented output (ratings on a scale from 1–10).
Table 1. Summary of the participants’ understanding of the presented output (ratings on a scale from 1–10).
IDPrior KnowledgeUnderstood OutputUnderstood PredictionRating
1YesYesYes10
2YesYesYes10
3YesYesYes7
4NoYesNo9
5YesYesYes8
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Žlahtič, B.; Završnik, J.; Blažun Vošner, H.; Kokol, P.; Šuran, D.; Završnik, T. Agile Machine Learning Model Development Using Data Canyons in Medicine: A Step towards Explainable Artificial Intelligence and Flexible Expert-Based Model Improvement. Appl. Sci. 2023, 13, 8329. https://doi.org/10.3390/app13148329

AMA Style

Žlahtič B, Završnik J, Blažun Vošner H, Kokol P, Šuran D, Završnik T. Agile Machine Learning Model Development Using Data Canyons in Medicine: A Step towards Explainable Artificial Intelligence and Flexible Expert-Based Model Improvement. Applied Sciences. 2023; 13(14):8329. https://doi.org/10.3390/app13148329

Chicago/Turabian Style

Žlahtič, Bojan, Jernej Završnik, Helena Blažun Vošner, Peter Kokol, David Šuran, and Tadej Završnik. 2023. "Agile Machine Learning Model Development Using Data Canyons in Medicine: A Step towards Explainable Artificial Intelligence and Flexible Expert-Based Model Improvement" Applied Sciences 13, no. 14: 8329. https://doi.org/10.3390/app13148329

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop