Protocol for Evaluating Explainability in Actuarial Models
Abstract
:1. Introduction
2. Context and Related Work
2.1. Definition: Explainable Artificial Intelligence (XAI)
- Data and feature visualization is an essential technique for helping people understand how input features relate to the model’s outputs [5]. Charts and diagrams can intuitively show these relationships and facilitate comprehension.
- Feature importance techniques assess the relative importance of input features in the model’s predictions, helping to identify which variables influence the model’s decisions the most, for instance in fair risk prediction application [18].
- Employing AI models that are inherently more interpretable, such as decision trees or linear regressions, instead of more complex black box models like deep neural networks [19], is the most standardized practice in the industry.
- Some approaches allow AI models to generate rules or explanations describing how they made a particular decision. This can be especially useful in critical applications where justification is needed [20].
- Natural language: generating natural language explanations that describe the models decisions in terms understandable by humans [21].
2.2. XAI Techniques
2.3. Contribution of XAI to the Actuarial Context
- Interpretable models: More interpretable algorithms, such as decision trees or linear regressions. These models are more accessible to explain and understand than black-box models [8] like neural networks. They are commonly used in highly regulated contexts, such as technical provisions.
- Localization of explanations: Provides specific explanations for each prediction. For instance, why was an insurance application denied [32]? What factors contributed to that decision? What surcharges could be applied to offer the requested coverage?
- Model auditing [33]: The periodic audits of AI models to detect any unexpected behavior or bias. This could be by governance models, internal control processes, external audits, or regulatory bodies. XAI techniques reduce the need for exact replicability and allow evaluation of the robustness of the established relationships, facilitating reviews.
- Transparency in decision-making: Clearly, communication about how AI models is used in underwriting, claims processing, and risk evaluation. This builds trust among decision-makers and regulators, (AI Act [34]). In this context, processes are highly technical and must, understandably, be communicated to senior management and the board, identifying the most sensitive variables to facilitate understanding and impact evaluation.
- Education and training: Employees and insurance agents training on how AI models work and how to interpret their results [35]. Internal understanding is essential for successful implementation, especially for those models’ affecting commercialization and not used by developers.
- Sensitivity tests evaluate how predictions change when certain features are modified, helping better understand the relationships between variables and the models decisions. XAI techniques help identify the most influential variables, aiding in anticipating effects due to variations or potential impacts from environmental or management changes. Other relevant use is for feature selection [36] and model debugging.
3. Selection of XAI Techniques
3.1. Usability
3.1.1. Model Consistency
3.1.2. Facilitating Decision-Making
3.1.3. Replicability in Production
3.1.4. Fairness and Bias
3.2. Interpretability
3.2.1. Simplicity of Results
3.2.2. Coherence
- A priori list: a top list of variables based on their relevance in the study context provided by domain experts.
- A posteriori list: the average result from different feature importance determination techniques. At least three techniques should be applied to derive a mean ranking from the results.
3.3. Confidence
3.3.1. Faithfulness
3.3.2. Stability and Robustness:
4. Proposed XAI Selection Protocol for Actuarial Problems Discussion
4.1. Actuarial Application
4.1.1. Understanding the Model
4.1.2. Results Explanation
- Relevance and precision of explanations: Evaluate the relevance and precision of explanations about the specific actuarial problem. Explanations must be relevant and valuable for actuarial decision-making. This aspect can be evaluated with FI, in any of its versions.
- Clarity and understandability of explanations: Assess the clarity and understandability of the explanations provided by the XAI model. The explanations should be accessible to actuaries and other end users without requiring deep technical knowledge. For example, if a pricing model combines diverse sources of information, the explanations should be evaluated according to the number of highlighted features; too many features could make the explanation difficult for insurance agents to follow. Coherence means whether the most relevant features remain similar for clients with similar profiles.
4.1.3. Model Inspection
- Transparency in the modeling process: Evaluate the transparency of the modeling process provided by the XAI model. Actuaries must understand how the model was trained, what data were used, and how the explanations were generated. This characteristic will be a fundamental part of model governance documentation, and a similar process should allow for similar explainability results, even if the core model does not allow for exact replicability.
- Regulatory and ethical compliance [43]: Assess whether the XAI model complies with the actuarial field’s relevant regulatory and ethical requirements. They can include data privacy, fairness, and non-discrimination, among other factors. The results from feature importance and fairness tests will facilitate this type of assessment of the models.
4.2. Proposed Protocol
4.2.1. Understanding the Problem
4.2.2. Define the Explainability Goal
- Part of the model development cycle–technical validation:
- Model optimization: Identify interactions and their impact on model output, enabling the evaluation of feature relevance, prioritization of variables, optimization, or identifying potential deficiencies or areas of improvement in the models.
- Detection of biases or technical discrimination: Identify possible biases generated by AI models, ensuring balanced inputs or hyperparameters to guarantee equitable models.
- Model selection: When multiple models are available, prediction accuracy may not be the only factor considered when implementing or deploying the final model. Bias analysis, regulatory compliance, or ease of implementation can become additional factors. In this regard, XAI techniques can be used to evaluate metrics beyond prediction accuracy.
- Understanding established relationships in the base model: Identify interactions and their impact on model output, which allows for evaluating the effects of varying relevant characteristics, facilitating sensitivity analysis while reducing total processing costs. It also allows for the consistent evaluation of the model’s development context and economic rationale.
- Evaluate financial models: Analyze the utilized financial models. This includes understanding their architecture, the data they use, and the relevant features for financial decision-making.
- Decision-making:
- Explanation to third parties: Understand how the relationships established by the model work, which features or variables influence the predictions most, and the rationale behind these relationships. It will aid in communicating with non-technical stakeholders and serve as a decision-making tool.
- Contrasting explainability results: XAI techniques rely on training and test data, are sensitive to inputs, and are influenced by the nature of the selected model. Therefore, it is advisable to compare results with one or more models.
- Expert review and contrast analysis: Compare the results with expert judgment to ensure coherence and validate the findings.
- Audit Process:
- Internal control: In developing the internal review processes carried out by key functions in the second and third lines of defense, replicating complex or deep learning models is not always feasible. Implementing XAI techniques facilitates the evaluation of the models proposed outputs without requiring the exact recalculation of results, which may not be viable in some models.
- Regulatory compliance: Facilitate the evaluation of compliance with regulatory requirements, such as ensuring no gender-based discrimination in insurance pricing or establishing consistency between variables and the context of the problem, allowing for bias detection.
4.2.3. XAI Model Selection
4.2.4. Evaluation of Results and Selection of Techniques
4.2.5. Implementation of XAI Techniques in Governance Model
4.2.6. Re-Evaluation of Needs
4.3. Assessment Framework
5. Case Study
5.1. Dataset
- Delinquency and inquiries (Group 1)
- Credit history and activity (Group 2)
- Yield and financial risk (Group 3)
5.2. Machine Learning and Explainability Results
5.3. Protocol Application
5.4. Protocol Result: Understanding Established Relationships
- Objective: To compare two different credit granting prediction models (random forest, gradient boosting) using XAI techniques to select the best model based on relevance and accuracy of the explanations.
- Attribute evaluation: Attributes of usability, interpretability, and confidence in the explanations generated by each XAI technique will be evaluated. The results are presented in Table 10 for gradient boosting model.
- The results for random forest are presented in Table 11.
5.5. Protocol Result: Audit Process
- Objective: To review the modeling process and its results, evaluating transparency, regulatory compliance, and internal audit.
- Techniques: Same as one target.
6. Discussion and Conclusions
6.1. Evaluation of the Framework
6.2. Practical Implications
6.3. Limitations and Challenges
7. Future Work
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Richman, R. AI in Actuarial Science—A Review of Recent Advances; Cambridge University Press: Cambridge, UK, 2020. [Google Scholar]
- Richman, R. AI in actuarial science—A review of recent advances—Part 2. Ann. Actuar. Sci. 2021, 15, 230–258. [Google Scholar] [CrossRef]
- Barredo, A.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
- Lorentzen, C.; Mayer, M. Peeking into the Black Box: An Actuarial Case Study for Interpretable Machine Learning; Swiss Association of Actuaries SAV: Estocolmo, Sweden, 2020. [Google Scholar]
- Greewell, B. pdp: An R Package for Constructing. R J. 2017, 9, 421–436. [Google Scholar] [CrossRef]
- Matthews, S.; Hartman, B. mSHAP: SHAP Values for Two-Part Models. Risks 2022, 10, 3. [Google Scholar] [CrossRef]
- Greenwell, B.; Boehmke, B.; McCarthy, A. A Simple and Effective Model-Based Variable Importance Measure. arXiv 2018, arXiv:1805.04755. [Google Scholar]
- Owens, E.; Sheehan, B.; Mullins, M.; Cunneen, M.; Ressel, J.; Castignani, G. Explainable Artificial Intelligence (XAI) in Insurance. Risks 2022, 10, 230. [Google Scholar] [CrossRef]
- Yang, G.; Ye, Q.; Xia, J. Unbox the black-box for the medical explainable AI via multi-modal and multi-centre data fusion: A mini-review, two showcases and beyond. Inf. Fusion 2022, 77, 29–52. [Google Scholar] [CrossRef] [PubMed]
- Zhang, Z.; Al Hamadi, H.; Damiani, E.; Yeun, C.Y.; Taher, F. Explainable Artificial Intelligence Applications in Cyber Security: State-of-the-Art in Research. IEEE Access 2022, 10, 93104–93139. [Google Scholar] [CrossRef]
- Gerlings, J.; Constantiou, I. Machine Learning in Transaction Monitoring: The Prospect of xAI. In Proceedings of the 56th International Conference on System Sciences, Maui, HI, USA, Virtual Event, 4–7 January 2022. [Google Scholar]
- Černevičienė, J.; Kabašinskas, A. Explainable artificial intelligence (XAI) in finance: A systematic literature review. Artif. Intell. Rev. 2024, 57, 216. [Google Scholar] [CrossRef]
- Golpayegani, D.; Pandit, H.; Lewis, D. Comparison and Analysis of 3 Key AI Documents: EU’s Proposed AI Act, Assessment List for Trustworthy AI (ALTAI), and ISO/IEC 42001 AI Management System. In Irish Conference on Artificial Intelligence and Cognitive Science; Springer: Cham, Switzerland, 2023; pp. 189–200. [Google Scholar]
- Salih, A.; Raisi-Estabragh, Z.; Galazzo, I.B.; Radeva, P.; Petersen, S.E.; Lekadir, K.; Menegaz, G. A Perspective on Explainable Artificial Intelligence Methods: SHAP and LIME. arXiv 2024, arXiv:2305.02012v3. [Google Scholar] [CrossRef]
- Fauzan, M.A.; Murfi, H. The accuracy of XGBoost for insurance claim prediction. Int. J. Adv. Soft Comput. Its Appl. 2018, 10, 159–171. [Google Scholar]
- Lin, W.; Wu, Z.; Lin, L.; Wen, A.; Li, J. An ensemble random forest algorithm for insurance big data analysis. IEEE Access 2017, 5, 16568–16575. [Google Scholar] [CrossRef]
- Bahalul, H.; Najmul, I.; Patrick, M. Explainable Artificial Intelligence (XAI) from a user perspective: A synthesis of prior literature and problematizing avenues for future research. Technol. Forecast. Soc. Change 2023, 186, 122120. [Google Scholar]
- Ning, Y.; Li, S.; Ng, Y.Y.; Chia, M.Y.C.; Gan, H.N.; Tiah, L.; Mao, D.R.; Ng, W.M.; Leong, B.S.-H.; Doctor, N.; et al. Variable importance analysis with interpretable machine learning for fair risk prediction. PLoS Digit. Health 2024, 3, e0000542. [Google Scholar] [CrossRef] [PubMed]
- Lipton, Z.C. The Mythos of Model Interpretability: In machine learning, the concept of interpretability is both important and slippery. Queue 2018, 16, 31–57. [Google Scholar] [CrossRef]
- Ghorbani, A.; Berenbaum, D.; Ivgi, M.; Dafna, Y.; Zou, J. Beyond Importance Scores: Interpreting Tabular ML by Visualizing Feature Semantics. arXiv 2021, arXiv:2111.05898. [Google Scholar] [CrossRef]
- Danilevsky, M.; Qian, K.; Aharonov, R.; Katsis, Y.; Kawas, B.; Sen, P. A Survey of the State of Explainable AI for Natural Language Processing. In Proceedings of the 10th International Joint Conference on Natural Language Processing, Suzhou, China, 4–7 December 2020. [Google Scholar]
- Wachter, S.; Mittelstadt, B.; Russell, C. Counterfactual Explanations without Opening the Black Box: Automated Decisions and the GDPR. Harv. J. Law Technol. 2018, 31, 842–889. [Google Scholar] [CrossRef]
- Ben-Haim, Y.; Tom-Tov, E. A Streaming Parallel Decision Tree Algorithm. J. Mach. Learn. Res. 2010, 11, 849–872. [Google Scholar]
- Goldstein, A.; Kapelner, A.; Bleich, J.; Pitkin, E. Peeking Inside the Black Box: Visualizing Statistical Learning with Plots of Individual Conditional Expectation. J. Comput. Graph. Stat. 2015, 24, 44–65. [Google Scholar] [CrossRef]
- Apley, D.W.; Zhu, J. Visualizing the effects of predictor variables in black box supervised learning models. J. R. Stat. Soc. Stat. Methodol. Ser. B 2020, 82, 1059–1086. [Google Scholar] [CrossRef]
- Friedman, J.H.; Popescu, B.E. Predictive Learning via Rule Ensembles. Ann. Appl. Stat. 2008, 2, 916–954. [Google Scholar] [CrossRef]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. “Why should I trust you?” Explaining the predictions of any classifier. arXiv 2016, arXiv:1602.04938. [Google Scholar]
- Andrews, R.; Diederich, J.; Tickle, A.B. Survey and critique of techniques for extracting rules from trained artificial neural networks. Knowl.-Based Syst. 1995, 8, 373–389. [Google Scholar] [CrossRef]
- Saltelli, A.; Ratto, M.; Andres, T.; Campolongo, F.; Cariboni, J.; Gatelli, D.; Saisana, M.; Tarantola, S. Global Sensitivity Analysis. The Primer; John Wiley & Sons, Ltd.: Hoboken, NJ, USA, 2007. [Google Scholar]
- Lundberg, S.; Lee, S.-I. A Unified Approach to Interpreting Model Predictions. arXiv 2017, arXiv:1705.07874. [Google Scholar]
- Molnar, C. Interpretable Machine Learning: A Guide for Making Black Box Models Explainable. Available online: https://christophm.github.io/interpretable-ml-book/ (accessed on 31 January 2025).
- Delcaillau, D.; Ly, A.; Papp, A.; Vermet, F. Model Transparency and Interpretability: Survey and Application to the Insurance Industry. Eur. Actuar. J. 2022, 12, 443–484. [Google Scholar] [CrossRef]
- Zhang, C.A.; Cho, S.; Vasarhelyi, M. Explainable Artificial Intelligence (XAI) in auditing. Int. J. Account. Inf. Syst. 2022, 46, 100572. [Google Scholar] [CrossRef]
- European Union. European Union Official Journal. Available online: https://digital-strategy.ec.europa.eu/es/policies/regulatory-framework-ai (accessed on 31 January 2025).
- Koster, O.; Kosman, R.; Visser, J. A Checklist for Explainable AI in the Insurance Domain. arXiv 2021, arXiv:2107.14039. [Google Scholar]
- Zacharias, J.; von Zahn, M.; Chen, J.; Hinz, O. Designing a feature selection method based on explainable artificial intelligence. Electron. Mark. 2022, 32, 2159–2184. [Google Scholar] [CrossRef]
- Kuo, K.; Lupton, D. Towards Explainability of Machine Learning Models in Insurance Pricing. arXiv 2020, arXiv:2003.10674. [Google Scholar]
- Holzinger, A.; Carrington, A.; Müller, H. Measuring the Quality of Explanations: The System Causability Scale (SCS). KI—Kunstliche Intell. 2020, 34, 193–198. [Google Scholar] [CrossRef]
- Alvarez-Melis, D.; Jaakkola, T.S. On the robustness of interpretability methods. arXiv 2018, arXiv:1806.08049. [Google Scholar]
- Lozano, C.; Romero, F.; Serrano, J.; Olivas, J.A. A comparison between explainable machine learning methods for classification and regression problems in the actuarial context. Mathematics 2023, 11, 3088. [Google Scholar] [CrossRef]
- Pagano, T.; Loureiro, R.; Lisboa, F.; Peixoto, R.; Guimaraes, G.; Cruz, G.; Araujo, M.; Santos, L.; Cruz, M.; Oliveira, E.; et al. Bias and Unfairness in Machine Learning Models: A Systematic review on datasets, tools, fairness metrics, and identification and mitigation methods. Big Data Cogn. Comput. 2023, 7, 31. [Google Scholar] [CrossRef]
- Ning, Y.; Li, S.; Ng, Y.Y.; Chia, M.Y.C.; Gan, H.N.; Tiah, L.; Ouyang, L.; Wu, J.; Jiang, X.; Almeida, D.; et al. Training language models to follow instructions with human feedback. arXiv 2022, arXiv:2203.02155. [Google Scholar]
- Martin, M.; Christopher, P.H.; Martin, C. Creating ethics guidelines for artificial intelligence and big data analytics customers: The case of the consumer European insurance market. Patterns 2021, 2, 100362. [Google Scholar]
- Tredger, E.R.W.; Lo, J.T.H.; Haria, S.; Lau, H.H.K.; Bonello, N.; Hlavka, B.; Scullion, C. Bias, guess and expert judgement in actuarial work. Br. Actuar. J. 2016, 21, 545–578. [Google Scholar] [CrossRef]
- Tsang, H.; Lee, W.; Tsui, E. AHP-Driven Knowledge Leakage Risk Assessment Model: A Construct-Apply-Control Cycle Approach. Int. J. Knowl. Syst. Sci. (IJKSS) 2016, 7, 1–18. [Google Scholar] [CrossRef]
- Fair Isaac Corporation. FICO Community. Available online: https://community.fico.com/s/explainable-machine-learning-challenge (accessed on 31 January 2024).
- Moscato, V.; Picariello, A.; Sperlí, G. A benchmark of machine learning approaches for credit score prediction. Expert Syst. Appl. 2021, 165, 113986. [Google Scholar] [CrossRef]
- Morio, J.; Balesdent, M. Estimation of Rare Event Probabilities in Complex Aerospace and Other Systems; Elsevier: Amsterdam, The Netherlands, 2015. [Google Scholar]
- AAE Artificial Intelligence and Data Science Working Group, Explainable Artificial Inteligence for C-Level Executives in Insurance. Actuarial Association of Europe. Discussion Paper. 2024. Available online: https://actuary.eu/wp-content/uploads/2024/08/AAE-Discussion-Paper-on-XAI-DEF.pdf (accessed on 7 April 2025).
- Lundberg, S.M.; Erion, G.; Chen, H.; DeGrave, A.; Prutkin, J.M.; Nair, B.; Katz, R.; Himmelfarb, J.; Bansal, N.; Lee, S.-I. Explainable AI for Trees: From Local Explanations to Global Understanding. arXiv 2019, arXiv:1905.04610. [Google Scholar] [CrossRef]
Techniques | Interpretability | Origin | Scope |
---|---|---|---|
Counterfactual explanations [22] | Post hoc | Agnostic | Local |
Decision trees [23] | Intrinsic | Specific | Global |
Feature importance [7] | Post hoc | Agnostic | Global |
Individual conditional expectation (ICE) [24] | Post hoc | Agnostic | Local |
Accumulated local effects (ALE) [25] | Post hoc | Agnostic | Global |
Interaction measure (H-statistic) [26] | Post hoc | Agnostic | Local |
LIME [27] | Post hoc | Agnostic | Local |
Partial dependence plot [5] | Post hoc | Agnostic | Global |
Rule extraction [28] | Post hoc | Agnostic | Global |
Sensitive analysis [29] | Post hoc | Agnostic | Local and global |
SHAP (Shapley explanations) [30] | Post hoc | Agnostic | Local and global |
Surrogate models [31] | Post hoc | Agnostic | Local and global |
Criteria | Key Points |
---|---|
Model Type | Different XAI methods work better with certain models (e.g., deep learning vs. linear/decision trees); agnostic techniques are model-independent. |
Available Methods | A variety exists (rule-based, post-processing, visualization); selection depends on computational resources and specific explainability goals. |
Multiple Techniques | Combining techniques can yield complementary insights and enhance interpretability. |
Evaluation | Assess based on the quality of information, ease of interpretation, computational cost, and robustness/sensitivity to input variations. |
Aspect | Key Point/Example |
---|---|
Interpretation of Key Features | Explain critical variables (e.g., income and age in credit models). |
Consistency and Stability | Ensure explanations are robust and consistent (e.g., high C(E) values). |
Feedback | Incorporate actuary feedback for improvements (e.g., surcharges based on age/activity). |
Ease of Use | Seamlessly integrate into workflows and assess computational efficiency (using TO formula). |
Goal | Technique |
---|---|
Technical Validation: | |
Model optimization | Global–local |
Detection of biases or technical discrimination | Global |
Model Selection | Feature importance Partial dependence plot Rule extraction Surrogate models |
Understanding relationships in the base model | Local |
Decision-Making: | |
Explanation to third parties | Shape PDP |
Contrasting explainability results | Global |
Expert review and contrast analysis | Shape PDP |
Audit Process: | |
Internal control and/or regulatory compliance | Global and local: Shapley, LIME, PDP |
XAI Attribute | Usability | Interpretability | Confidence | |||||
---|---|---|---|---|---|---|---|---|
Explainability Goal | Consistency | Decision-Making | Replicability | Fairness | Coherence | Simplicity | Faithfulness | Stability and Robustness |
Consistency and stability of explanations | 18% | 5% | 7% | 3% | 14% | 10% | 21% | 22% |
Feedback and continuous improvement | 22% | 7% | 21% | 3% | 10% | 5% | 14% | 18% |
Ease of use and practical applicability | 5% | 21% | 22% | 3% | 7% | 18% | 10% | 14% |
Relevance and accuracy of explanations. | 10% | 7% | 5% | 3% | 21% | 14% | 22% | 18% |
Clarity and comprehensibility of explanations. | 21% | 7% | 5% | 3% | 10% | 22% | 14% | 18% |
Transparency of the modeling process | 14% | 5% | 3% | 18% | 10% | 7% | 22% | 21% |
Regulatory compliance | 10% | 3% | 7% | 22% | 14% | 5% | 21% | 18% |
Models | AUC | Accuracy | Precision | Recall | F1 |
---|---|---|---|---|---|
Decision tree (DT) | 0.71 | 0.71 | 0.71 | 0.67 | 0.69 |
KNN | 0.70 | 0.70 | 0.67 | 0.73 | 0.70 |
Artificial neural networks (ANN) | 0.71 | 0.71 | 0.73 | 0.70 | 0.71 |
Random forest (RF) | 0.79 | 0.71 | 0.72 | 0.65 | 0.69 |
Gradient boosting (GB) | 0.80 | 0.73 | 0.73 | 0.69 | 0.70 |
XAI Attribute | Usability | Interpretability | Confidence | |||||
---|---|---|---|---|---|---|---|---|
Explainability Goal | Consistency | Decision-Making | Replicability | Fairness | Coherence | Simplicity | Faithfulness | Stability and Robustness |
LIME | 0.8298 | 0.500 | 0.739 | 1.000 | 0.667 | 0.333 | 0.500 | 0.908 |
SHAP | 0.9418 | 0.600 | 0.245 | 1.000 | 0.500 | 0.500 | 0.495 | 0.705 |
PERM | 0.6905 | 0.200 | 0.727 | 1.000 | 0.500 | 1.000 | 0.500 | 0.962 |
MS | 0.6379 | 0.200 | 0.879 | 1.000 | 0.500 | 0.250 | 0.490 | 0.416 |
PDV | 0.8653 | 0.500 | 0.696 | 1.000 | 0.667 | 0.667 | 0.485 | 0.912 |
XAI Attribute | Usability | Interpretability | Confidence | |||||
---|---|---|---|---|---|---|---|---|
Explainability Goal | Consistency | Decision-Making | Replicability | Fairness | Coherence | Simplicity | Faithfulness | Stability and Robustness |
LIME | 0.8298 | 0.500 | 0.706 | 1.000 | 0.900 | 0.500 | 0.492 | 0.945 |
SHAP | 0.9418 | 0.600 | 0.211 | 1.000 | 0.857 | 0.714 | 0.482 | 0.635 |
PERM | 0.6905 | 0.200 | 0.689 | 1.000 | 0.818 | 0.455 | 0.497 | 0.970 |
MS | 0.6379 | 0.200 | 0.732 | 1.000 | 0.600 | 0.333 | 0.492 | 0.573 |
PDV | 0.8653 | 0.500 | 0.691 | 1.000 | 0.800 | 1.000 | 0.482 | 0.962 |
Features | Evaluation of Prediction | Models Audit | |
---|---|---|---|
Usability | Consistency | 24.75% | 14.00% |
Make decision | 18.00% | 4.00% | |
Replicability | 22.00% | 10.00% | |
Interpretability | Coherency | 7.00% | 18.00% |
Simplicity | 4.00% | 7.00% | |
Confidence | Fidelity | 14.00% | 24.75% |
Stability and robustness | 10.00% | 22.00% |
LIME | SHAP | PERM | MS | PDV | |
---|---|---|---|---|---|
Usability | 0.458 | 0.395 | 0.367 | 0.387 | 0.457 |
Interpretability | 0.060 | 0.055 | 0.075 | 0.045 | 0.073 |
Confidence | 0.161 | 0.140 | 0.166 | 0.110 | 0.159 |
Results | 0.679 | 0.590 | 0.608 | 0.543 | 0.690 |
LIME | SHAP | PERM | MS | PDV | |
---|---|---|---|---|---|
Usability | 0.451 | 0.388 | 0.359 | 0.355 | 0.456 |
Interpretability | 0.083 | 0.089 | 0.075 | 0.055 | 0.096 |
Confidence | 0.163 | 0.131 | 0.166 | 0.126 | 0.164 |
Results | 0.697 | 0.607 | 0.600 | 0.536 | 0.716 |
LIME | SHAP | PERM | MS | PDV | |
---|---|---|---|---|---|
Usability | 0.210 | 0.180 | 0.177 | 0.185 | 0.211 |
Interpretability | 0.143 | 0.125 | 0.160 | 0.108 | 0.167 |
Confidence | 0.324 | 0.278 | 0.335 | 0.213 | 0.321 |
Results | 0.677 | 0.583 | 0.673 | 0.506 | 0.698 |
LIME | SHAP | PERM | MS | PDV | |
---|---|---|---|---|---|
Usability | 0.207 | 0.177 | 0.174 | 0.171 | 0.210 |
Interpretability | 0.197 | 0.204 | 0.179 | 0.131 | 0.214 |
Confidence | 0.330 | 0.259 | 0.336 | 0.248 | 0.331 |
Results | 0.733 | 0.640 | 0.689 | 0.550 | 0.755 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Lozano-Murcia, C.; Romero, F.P.; Gonzalez-Ramos, M.C. Protocol for Evaluating Explainability in Actuarial Models. Electronics 2025, 14, 1561. https://doi.org/10.3390/electronics14081561
Lozano-Murcia C, Romero FP, Gonzalez-Ramos MC. Protocol for Evaluating Explainability in Actuarial Models. Electronics. 2025; 14(8):1561. https://doi.org/10.3390/electronics14081561
Chicago/Turabian StyleLozano-Murcia, Catalina, Francisco P. Romero, and Mᵃ Concepción Gonzalez-Ramos. 2025. "Protocol for Evaluating Explainability in Actuarial Models" Electronics 14, no. 8: 1561. https://doi.org/10.3390/electronics14081561
APA StyleLozano-Murcia, C., Romero, F. P., & Gonzalez-Ramos, M. C. (2025). Protocol for Evaluating Explainability in Actuarial Models. Electronics, 14(8), 1561. https://doi.org/10.3390/electronics14081561