Correlation Metrics for Safe Artificial Intelligence
Abstract
1. Introduction
2. Extending the SAFE AI Metrics
3. Application
Benchmarking
4. Discussion
5. Conclusions
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Achakzai, Muhammad Atif Khan, and Peng Juan. Using machine learning meta-classifiers to detect financial frauds. Finance Research Letters 48: 102915. [CrossRef]
- Ainslie, Russell, John McCall, Sid Shakya, and Gilbert Owusu. 2017. Predicting service levels using neural networks. In International Conference on Innovative Techniques and Applications of Artificial Intelligence. Berlin/Heidelberg: Springer, pp. 411–416. [Google Scholar] [CrossRef]
- Aldasoro, Iñaki, Leonardo Gambacorta, Paolo Giudici, and Thomas Leach. 2022. The drivers of cyber risk. Journal of Financial Stability 60: 100989. [Google Scholar] [CrossRef]
- Babaei, Golnoosh, Paolo Giudici, and Emanuela Raffinetti. 2025. A rank graduation box for safe AI. Expert Systems with Applications 259: 125239. [Google Scholar] [CrossRef]
- Binns, Reuben. 2018. Fairness in machine learning: Lessons from political philosophy. Paper presented at 2018 Conference on Fairness, Accountability and Transparency, PMLR, New York, NY, USA, 23–24 February; pp. 149–159. [Google Scholar]
- Bracke, Philippe, Anupam Datta, Carsten Jung, and Shayak Sen. 2019. Machine Learning Explainability in Finance: An Application to Default Risk Analysis. Working Paper 816. London: Bank of England, August. [Google Scholar] [CrossRef]
- Bücker, Michael, Gero Szepannek, Alicja Gosiewska, and Przemyslaw Biecek. 2022. Transparency, auditability, and explainability of machine learning models in credit scoring. Journal of the Operational Research Society 73: 70–90. [Google Scholar] [CrossRef]
- Bussmann, Niklas, Paolo Giudici, Dimitri Marinelli, and Jochen Papenbrock. 2020. Explainable machine learning in credit risk management. Computational Economics 57: 2013–16. [Google Scholar] [CrossRef]
- Chen, Zhiyuan, Le Dinh Van Khoa, Ee Na Teoh, Amril Nazir, Ettikan Kandasamy Karuppiah, and Kim Sim Lam. 2018. Machine learning techniques for anti-money laundering (aml) solutions in suspicious transaction detection: A review. Knowledge and Information Systems 57: 245–85. [Google Scholar] [CrossRef]
- Diebold, Francis X., and Roberto S. Mariano. 1995. Comparing predictive accuracy. Journal of Business & Economic Statistics 13: 253–63. [Google Scholar] [CrossRef]
- European Commission. 2022. Artificial Intelligence Act. Available online: https://artificialintelligenceact.eu (accessed on 18 August 2025).
- Floridi, Luciano, Josh Cowls, Monica Beltrametti, Raja Chatila, Patrice Chazerand, Virginia Dignum, Christoph Luetge, Robert Madelin, Ugo Pagallo, Francesca Rossi, and et al. 2018. Ai4people—An ethical framework for a good ai society: Opportunities, risks, principles, and recommendations. Minds and Machines 28: 689–707. [Google Scholar] [CrossRef]
- Frost, J., L. Gambacorta, Y. Huang, and P. Zbindnen. 2019. Bigtech and the changing structure of financial intermediation. Economic Policy 34: 761–99. [Google Scholar] [CrossRef]
- Ganesh, A. Deiva., and P. Kalpana. 2022. Future of artificial intelligence and its influence on supply chain risk management—A systematic review. Computers and Industrial Engineering 169: 108206. [Google Scholar] [CrossRef]
- Giudici, Paolo, and Emanuela Raffinetti. 2021. Explainable ai in cyber risk management. Quality and Reliability Engineering International 38: 1318–26. [Google Scholar] [CrossRef]
- Giudici, Paolo, and Emanuela Raffinetti. 2025. RGA: A unified measure of predictive accuracy. Advances in Data Analysis and Classification 19: 67–93. [Google Scholar] [CrossRef]
- Giudici, Paolo, and Gloria Polinesi. 2021. Crypto price discovery through correlation networks. Annals of Operations Research 299: 443–57. [Google Scholar] [CrossRef]
- Giudici, Paolo, and Iman Abu-Hashish. 2019. What determines bitcoin exchange prices? A network var approach. Finance Research Letters 28: 309–18. [Google Scholar] [CrossRef]
- Goodfellow, Ian J., Jonathon Shlens, and Christian Szegedy. 2015. Explaining and harnessing adversarial examples. Paper presented at International Conference on Learning Representations (ICLR), San Diego, CA, USA, May 7–9. [Google Scholar]
- Hardt, Moritz, Eric Price, and Nati Srebro. 2016. Equality of opportunity in supervised learning. Advances in Neural Information Processing Systems 29: 3315–23. [Google Scholar]
- International Organization for Standardization and International Electrotechnical Commission. 2023. Information Technology—Artificial Intelligence—Guidance on Risk Management. ISO/IEC 23894:2023. London: British Standards Institution.
- Jobin, Anna, Marcello Ienca, and Effy Vayena. 2019. The global landscape of ai ethics guidelines. Nature Machine Intelligence 1: 389–99. [Google Scholar] [CrossRef]
- Kleinberg, Jon, Sendhil Mullainathan, and Manish Raghavan. 2016. Inherent trade-offs in the fair determination of risk scores. arXiv arXiv:1609.05807. [Google Scholar] [CrossRef]
- Kuiper, O., M. Van den Berg, J. Van der Burgt, and S. Leijnen. 2021. Exploring explainable ai in the financial sector: Perspectives of banks and supervisory authorities. In Artificial Intelligence and Machine Learning. Berlin/Heidelberg: Springer. [Google Scholar] [CrossRef]
- Liu, Wanan, Hong Fan, and Meng Xia. 2022. Credit scoring based on tree-enhanced gradient boosting decision trees. Expert Systems with Applications 189: 116034. [Google Scholar] [CrossRef]
- Manheim, David, and Scott Garrabrant. 2018. Multiparty dynamics and failure modes for machine learning systems. arXiv arXiv:1810.10862. [Google Scholar]
- Melançon, Gabrielle Gauthier, Philippe Grangier, Eric Prescott-Gagnon, Emmanuel Sabourin, and Louis-Martin Rousseau. 2021. A machine learning-based system for predicting service-level failures in supply chains. Informs Journal on Applied Analytics 51: 200–12. [Google Scholar] [CrossRef]
- Moscato, Vincenzo, Antonio Picariello, and Giancarlo Sperlì. 2021. A benchmark of machine learning approaches for credit score prediction. Expert Systems with Applications 165: 113986. [Google Scholar] [CrossRef]
- Naim, Arshi. 2022. Role of artificial intelligence in business risk management. American Journal of Business Management, Economics and Banking 1: 55–66. [Google Scholar]
- Organisation for Economic Cooperation and Development. 2022. Framework for the Classification of AI Systems. Available online: https://oecd.ai/en/ai-principles (accessed on 18 August 2025).
- Pathade, Chetan. 2025. Red teaming the mind of the machine: A systematic evaluation of prompt injection and jailbreak vulnerabilities in llms. arXiv arXiv:2505.04806. [Google Scholar] [CrossRef]
- Rufibach, Kaspar. 2010. Use of brier score to assess binary predictions. Journal of Clinical Epidemiology 63: 938–39. [Google Scholar] [CrossRef]
- Sachan, Swati, Jian-Bo Yang, Dong-Ling Xu, David Eraso Benavides, and Yang Li. 2020. An explainable ai decision-support-system to automate loan underwriting. Expert Systems with Applications 144: 113100. [Google Scholar] [CrossRef]
- Selbst, Andrew D, Danah Boyd, Sorelle A Friedler, Suresh Venkatasubramanian, and Janet Vertesi. 2019. Fairness and abstraction in sociotechnical systems. Paper presented at Conference on Fairness, Accountability, and Transparency, Rio de Janeiro, Brazil, June 3–6; pp. 59–68. [Google Scholar]
- Steinhardt, Jacob, Pang Wei Koh, and Percy Liang. 2017. Certified defenses for data poisoning attacks. Advances in Neural Information Processing Systems 30: 3517–29. [Google Scholar]
- Sundra, B. M., K. Sathiyamurthi, and G. Subramanian. 2023. Critical evaluation of applying machine learning approaches for better handling bank risk management in the post-modern era. Scandinavian Journal of Information Systems 35: 1228–31. [Google Scholar]
- Szegedy, Christian, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian Goodfellow, and Rob Fergus. 2014. Intriguing properties of neural networks. arXiv arXiv:1312.6199. [Google Scholar] [CrossRef]
- United States National Institute of Standards and Technologies. 2023. Ai Risk Management Framework. Available online: https://www.nist.gov/itl/ai-risk-management-framework (accessed on 18 August 2025).
| Control Category | ISO/IEC 23894:2023 Requirement Actions | S.A.F.E. Actions | Who Measures What | When in the AI Lifecycle | Who Escalates/Rolls Back |
|---|---|---|---|---|---|
| Context Establishment | Define scope, objectives, boundaries, and stakeholders; identify intended use and potential misuse. | Define AI applications and, for each of them, response and explanatory variables. | Business owners, risk managers—define scope and variables. | Conception/ project initiation. | Governance board or project steering committee. |
| Risk Identification | Identify risks from data, models, processes, deployment environment, and misuse scenarios across lifecycle. | For AI application identify which S.A.F.E. metric applies. | Data scientists, domain experts—map risks and metrics. | Data collection and model design. | Risk manager with compliance team. |
| Risk Analysis | Assess likelihood, severity, and uncertainty of risks; consider emergent and systemic risks. | Calculate the identified metrics, equate likelihood with their complement to one, and severity with the quantity and quality of impacted stakeholders. | Data scientists, quantitative risk analysts—compute metrics. | Model development and validation. | Chief Risk Officer (CRO) or equivalent oversight function. |
| Risk Evaluation | Compare risks to acceptance criteria; prioritize risks for treatment. | Calculate thresholds, using appropriate statistical tests, such as Diebold and Mariano (1995). | Validation team—apply thresholds, statistical testing. | Pre-deployment validation. | Independent risk committee or model validation unit. |
| Risk Treatment | Apply measures such as data quality and governance checks, bias and fairness analysis, robustness testing, explainability, human oversight, security safeguards, fallback and incident response mechanisms. | Interpret deviation from threshold as potential lack of fairness, robustness and security, explainability and human oversight or fallback accuracy. | ML engineers, auditors—monitor controls and implement safeguards. | Deployment and operational readiness. | AI ethics board, compliance lead, or IT security head. |
| Monitoring | Continuously monitor AI performance and risks; implement feedback loops and update risk assessments. | Update model, metrics and comparison with threshold as new training data arrives. | Operations/ monitoring team—track metrics over time. | Post-deployment, continuous operation. | Incident response manager, system owner. |
| Communication | Engage stakeholders in risk decisions; ensure transparency and reporting of risk management outcomes. | Report S.A.F.E. metrics values, in comparison with thresholds, to stakeholders. | Risk managers, reporting officers—communicate metrics and risks. | Throughout lifecycle, periodic reviews. | Senior management, external regulators if required. |
| Variable | Definition |
|---|---|
| salary | a numeric variable, used as response variable: current salary in US dollars |
| age | a numeric variable: age in years |
| edu | a numeric variable: educational level in years |
| startsal | a numeric variable: beginning salary in US dollars |
| jobtime | a numeric variable: months since hire |
| prevexp | a numeric variable: previous work experience in months |
| minority | a factor variable: minority classification with levels min, indicating minority, and no_min, no minority |
| gender | a factor variable: gender type with levels f, indicating female, and m, indicating male |
| jobcat | a factor variable: type of job with levels Clerical, Custodial, and Manager |
| Variable | |
|---|---|
| edu | 0.4523 |
| gender | 0.1405 |
| prevexp | 0.0392 |
| jobtime | 0.0274 |
| minority | 0.0207 |
| age | 0.0001 |
| Variable | |
|---|---|
| edu | 0.0910 |
| gender | 0.0256 |
| jobtime | 0.0094 |
| prevexp | 0.0069 |
| minority | 0.0063 |
| age | 0.0000 |
| Variable | |
|---|---|
| age | 0.3329 |
| jobtime | 0.1386 |
| prevexp | 0.0009 |
| gender | 0.0005 |
| edu | 0.0001 |
| minority | 0.0000 |
| Variable | |
|---|---|
| age | 0.9992 |
| jobtime | 0.3684 |
| prevexp | 0.0091 |
| gender | 0.0016 |
| edu | 0.0008 |
| minority | 0.000 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Babaei, G.; Giudici, P. Correlation Metrics for Safe Artificial Intelligence. Risks 2025, 13, 178. https://doi.org/10.3390/risks13090178
Babaei G, Giudici P. Correlation Metrics for Safe Artificial Intelligence. Risks. 2025; 13(9):178. https://doi.org/10.3390/risks13090178
Chicago/Turabian StyleBabaei, Golnoosh, and Paolo Giudici. 2025. "Correlation Metrics for Safe Artificial Intelligence" Risks 13, no. 9: 178. https://doi.org/10.3390/risks13090178
APA StyleBabaei, G., & Giudici, P. (2025). Correlation Metrics for Safe Artificial Intelligence. Risks, 13(9), 178. https://doi.org/10.3390/risks13090178

