Performance, Fairness, and Explainability in AI-Based Credit Scoring: A Systematic Literature Review
Abstract
1. Introduction
2. Methodology
2.1. Research Questions (RQs) Formulation
- RQ1: To what extent can credit scoring frameworks achieve compelling performance while balancing explainability and fairness trade-offs? The extent to which AI models operate opaquely in determining creditworthiness not only hinders their adoption but also poses threats to their fairness and trustworthiness (Ribeiro-Flucht et al., 2024). In response, numerous efforts within the credit scoring landscape have prioritized transparency and explainability to develop complementary or embedded methods that clarify the rationale and reasoning behind AI-generated outcomes. Model explainability and interpretability are foundational in the deployment of AI and are considered no less important than the performance itself. Apart from describing the motive, both can fundamentally serve the ability to explain causes leading to biased or uncertain decisions (Alufaisan et al., 2021). High-stakes applications depend critically on the ability to reason about and justify model decisions, whereby the lack of transparent decision-making poses a significant drawback and may result in mistrust and non-compliance with local regulations (Wang et al., 2020). Considering the criticality of the three pillars, this research question examines the extent to which these pillars can be elevated jointly or if there are potential trade-offs.
- RQ2: How do historical repayment data, class imbalance, and protected attributes contribute to biased predictions, and what mitigation strategies are most effective? A common issue across datasets used to train AI models is class imbalance, which hinders AI models from producing accurate results (Chen et al., 2024). This problem is pervasive across credit scoring datasets where the number of defaulters is significantly less than that of non-defaulters. Such an imbalance adversely affects accuracy, suggesting the need for more adaptive techniques that treat all classes equitably. In addition, the presence of protected attributes across different credit scoring datasets potentially amplifies the historical discrimination against minority groups, leaving structural traces in training data (Hurlin et al., 2024; Talaat et al., 2024). For instance, certain ethnic groups have historically been granted credit less frequently, thereby appearing more frequently in the “bad” class, not due to actual risk but because they were denied favorable products or guidance. Addressing this research question reveals the relationships among class imbalance, as well as their potential effect on protected attributes, and provides means to understand the mitigation strategies that eliminate biased decisions.
- RQ3: How do regulatory frameworks and human-in-the-loop (HITL) approaches influence the interpretation of fairness across different contexts, and how can they be incorporated into ethically aligned AI models? Given that AI models are prone to biased decisions and lack transparency in how their results are determined, the intervention of regulatory bodies underscores the importance of consciously adopting AI in domains involving monetary decisions and fundamental human rights. While precision and accuracy were the ultimate goals sought in the past, the loss of transparency has proven far more costly in critical and regulated domains, making explainability a necessity rather than an option (Chen et al., 2024). Ensuring this balance enables the responsible and accountable deployment of AI in credit scoring domains, supported by adequate human oversight to maintain compliance with local regulations, and ensures that results remain comprehensible to human decision-makers (Peng et al., 2023).
2.2. Search Strategy
2.3. Selection Criteria
2.3.1. Inclusion Criteria
2.3.2. Exclusion Criteria
2.4. Screening Process
2.5. Data Extraction
2.6. Quality Assessment
3. Results
3.1. Study Selection
3.2. Characteristics of Selected Studies
3.3. Topic Coverage
4. Discussion
4.1. Compatibilities and Trade-Offs (RQ1)
4.2. Fairness Strategies in Deployment Pipelines (RQ2)
4.3. Regulatory, Ethical, and Governance Foundations for Fair AI Credit Scoring (RQ3)
5. Gaps and Future Directions
6. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Appendix A
| ID | Author(s) | Title | Exclusion Reason | PRISMA Bucket |
|---|---|---|---|---|
| 1 | Zacharias et al. (2022) | Designing a feature selection method based on explainable artificial intelligence | Excluded due to overlapping scope, as explainability is applied mainly as post-hoc SHAP-based feature attribution within a performance-driven pipeline that is already well represented among the included studies. | Overlapping scope |
| 2 | Corrales-Barquero et al. (2021) | A Review of Gender Bias Mitigation in Credit Scoring Models | Excluded due to overlapping scope, as a survey-style review summarizing bias mitigation strategies, which is already covered by more recent or synthesis-relevant included sources, contributing limited additional evidence to the review objectives. | Overlapping scope |
| 3 | de Castro Vieira et al. (2025) | Towards Fair AI: Mitigating Bias in Credit Decisions—A Systematic Literature Review | Excluded due to overlapping scope, as the study is primarily a survey summarizing fairness and bias mitigation without providing distinct empirical evidence or integrated analysis across performance, fairness, and explainability beyond included works. | Overlapping scope |
| 4 | Vuković et al. (2025) | AI integration in financial services: a systematic review of trends and regulatory challenges | Excluded as out of scope, as it addresses AI in financial services broadly rather than AI-based credit scoring and does not provide credit scoring-specific evidence aligned with the review eligibility criteria. | Did not meet eligibility |
| 5 | Cornacchia et al. (2023) | A General Architecture for a Trustworthy Creditworthiness- Assessment Platform in the Financial Domain | Excluded due to overlapping scope, as the study emphasizes a performance-oriented architecture with explainability treated primarily as a post-hoc component, overlapping with included works addressing similar post-hoc explainability configurations. | Overlapping scope |
| 6 | Mou et al. (2024) | Cost-aware Credit-scoring Framework Based on Resampling and Feature Selection | Excluded as performance-oriented only, focusing on class imbalance handling, resampling, and feature selection for cost-aware optimization without substantive treatment of fairness or explainability as primary review pillars. | Did not meet eligibility |
| 7 | Cao et al. (2021) | Ensemble methods for credit scoring of Chinese peer-to-peer loans | Excluded as performance-oriented only, as it evaluates ensemble learning primarily for predictive performance without explicit fairness, explainability, or regulatory/ HITL considerations aligned with the review scope. | Did not meet eligibility |
| 8 | Wu et al. (2025) | A ‘divide and conquer’ reject inference approach leveraging graph-based semi-supervised learning | Excluded as reject-inference/performance-oriented, where reject inference is used to address sample selection bias and improve predictive performance without protected-attribute fairness analysis or explainability objectives aligned with the review synthesis goals. | Did not meet eligibility |
| ID | Author(s) | Title | Exclusion Reason | PRISMA Bucket |
|---|---|---|---|---|
| 9 | Liao et al. (2022) | Combating Sampling Bias: A Self-Training Method in Credit Risk Models | Excluded as performance/sample-bias oriented, focusing on accepted-only sampling bias using self-training without explicitly addressing fairness across protected attributes or explainability as core objectives. | Did not meet eligibility |
| 10 | C-Rella et al. (2025) | Cost-sensitive reinforcement learning for credit risk | Excluded as performance-oriented only, proposing cost-sensitive reinforcement learning to optimize credit risk decisioning without operationalizing fairness, explainability, or regulatory/HITL requirements central to the review eligibility criteria. | Did not meet eligibility |
| 11 | Koulu (2019) | Human control over automation: EU policy and AI ethics | Excluded as not eligible, as it is primarily a legal/policy discussion of automation and AI ethics without credit scoring-specific empirical methods or operational evidence supporting synthesis across performance, fairness, and explainability. | Did not meet eligibility |
| 12 | Z. Li et al. (2020) | Inferring the outcomes of rejected loans: an application of semisupervised clustering | Excluded as reject-inference/performance-oriented, focusing on outcome inference for rejected applicants to enhance prediction performance, with fairness and explainability not treated as central, operationalized objectives. | Did not meet eligibility |
| 13 | Tiukhova et al. (2025) | Boosting Credit Risk Data Quality Using Machine Learning and eXplainable AI Techniques | Excluded due to overlapping scope, as XAI is applied mainly for data/model diagnostics, and explainability is treated as post-hoc analysis, closely overlapping with included SHAP-post-hoc explainability studies. | Overlapping scope |
| 14 | W. Li et al. (2022) | A data-driven explainable case-based reasoning approach for financial risk detection | Excluded as out of scope, as it targets financial risk detection rather than credit scoring/creditworthiness assessment, and it does not align with the review’s domain-specific eligibility criteria. | Did not meet eligibility |
| 15 | Chacko and Aravindhar (2025) | Enhancing Fairness and Accuracy in Credit Score Analysis: A Novel Framework Utilizing Kernel PCA | Excluded due to insufficient operationalization of fairness, as fairness is referenced but not clearly defined using explicit metrics or evaluation protocols that support structured synthesis under the review eligibility criteria. | Did not meet eligibility |
Appendix B
| ID | Author(s) | Title | Summary | Related RQs |
|---|---|---|---|---|
| 1 | Kozodoi et al. (2022) | Fairness in Credit Scoring: Assessment, Implementation and Profit Implications | Examines trade-offs between fairness and profitability in credit scoring. Integrates fairness metrics into ML pipelines and evaluates pre-, in-, and post-processing methods (reweighing, prejudice remover, adversarial debiasing, reject option) across seven datasets. Concludes fairness can improve without major performance loss, supporting regulatory compliance and ethical lending. | RQ1, RQ2 |
| 2 | Dessain et al. (2023) | Cost of Explainability in AI: An Example with Credit Scoring Models | Explores the explainability–performance trade-off in credit scoring. Compares black-box and interpretable models (XGBoost, NN, LR, GAMs) under ECB compliance. Introduces isotonic smoothing to align expert judgement with regulatory master-scale grading. Finds GAM-style models achieve near-black-box accuracy while preserving inherent interpretability and meeting regulatory standards. | RQ1, RQ3 |
| 3 | Moldovan (2023) | Algorithmic Decision Making Methods for Fair Credit Scoring | Assesses algorithmic bias and compares 12 mitigation strategies across five fairness metrics using German and Romanian datasets. Highlights that no single fairness measure satisfies fairness, performance, and profitability simultaneously. Shows incompatibilities among metrics (independence, separation, sufficiency) and stresses regulatory ambiguity and the need for balanced, multi-method approaches. | RQ1, RQ2, RQ3 |
| 4 | Zehlike et al. (2025) | Beyond Incompatibility: Trade-offs Between Mutually Exclusive Fairness Criteria in Machine Learning and Law | Introduces the Fair Interpolation Method (FAIM), a post-processing algorithm using optimal transport to interpolate between calibration, balance for positives, and balance for negatives. Motivated by the EU AI Act, it addresses fairness incompatibility and legal ambiguity, emphasizing regulator involvement and human oversight for trade-off management across jurisdictions. | RQ1, RQ2, RQ3 |
| 5 | Das et al. (2023) | Algorithmic Fairness | Situates fairness within ECOA, FCRA, and supervisory rules; distinguishes individual vs. group fairness. Argues bias can enter pre-, in-, or post-training and catalogues dataset biases and metrics (e.g., DI, EO, equalized odds, predictive parity). Advocates a systemic approach combining data quality, interpretability, and regulatory alignment. | RQ2, RQ3 |
| 6 | Brzezinski et al. (2024) | Properties of Fairness Measures in the Context of Varying Class Imbalance and Protected Group Ratios | Analyzes effects of class imbalance and protected-group ratios on fairness metrics using the UCI Adult dataset. Finds Statistical Parity Difference and Disparate Impact are highly sensitive to imbalance, while Equal Opportunity and Average Odds Difference are more stable. Recommends contextual evaluation combining fairness and performance indicators. | RQ2 |
| ID | Author(s) | Title | Summary | Related RQs |
|---|---|---|---|---|
| 7 | Langenbucher and Corcoran (2022) | Responsible AI Credit Scoring–A Lesson from Upstart.com | Examines regulatory and ethical challenges in AI-driven credit scoring under GDPR, ECOA, FCRA, and GLBA. Highlights privacy risks, proxy discrimination, and fairness–accuracy trade-offs. Recommends transparency, fairness audits, human-in-the-loop oversight, and regulator collaboration to ensure compliant and explainable lending decisions. | RQ1, RQ3 |
| 8 | Langenbucher (2020) | Responsible A.I.-based Credit Scoring–A Legal Framework | Outlines a legal framework for responsible AI credit scoring based on transparency, fairness, and accountability. Warns opaque models can conflict with GDPR’s “right to explain.” Recommends embedding interpretability, validating fairness throughout model phases, enforcing human oversight, and assigning accountability roles for lawful deployment. | RQ3 |
| 9 | Valdrighi et al. (2025) | Best Practices for Responsible Machine Learning in Credit Scoring | Addresses bias, transparency, and reject inference in AI-based credit scoring across German, Taiwanese, and Home Credit datasets. Discusses bias origins, fairness metrics (group and individual), and mitigation across pre-, in-, and post-processing. Highlights transparency tools (LIME, SHAP, PD, ICE) and emphasizes inclusive, responsible deployment. | RQ1, RQ2 |
| 10 | Kuiper et al. (2021) | Exploring Explainable AI in the Financial Sector: Perspectives of Banks and Supervisory Authorities | Reports qualitative findings from Dutch banks and regulators on integrating XAI into credit scoring. Defines explainability as transparency into model reasoning, data, and design. Notes reliance on interpretable traditional models, human safeguards, and phased deployment. Positions explainability as essential for ethical, accountable, regulatory-aligned AI. | RQ1, RQ3 |
| 11 | Balashankar and Lees (2022) | The Need for Transparent Demographic Group Trade-Offs in Credit Risk and Income Classification | Highlights fairness as jurisdiction-dependent. Uses Pareto principles to balance group- and overall-level accuracies with human-in-the-loop oversight. Shows intersecting protected attributes reduce sample sizes and degrade accuracy, motivating data improvements. Advocates transparent trade-off visualization to align fairness, performance, and social objectives. | RQ1, RQ2, RQ3 |
| 12 | Amekoe et al. (2024) | Exploring Accuracy and Interpretability Trade-off in Tabular Learning with Novel Attention-based Models | Quantifies accuracy–interpretability trade-offs in tabular learning using 45 datasets. Finds less than 4% accuracy loss between black-box and inherently interpretable models. Proposes a TabSRA attention-based ensemble inspired by GAMs, offering feature-level interpretability and stable performance; argues for inherent interpretability in high-stakes domains. | RQ1, RQ3 |
| ID | Author(s) | Title | Summary | Related RQs |
|---|---|---|---|---|
| 13 | S. Liu and Vicente (2022) | Accuracy and Fairness Trade-offs in Machine Learning: A Stochastic Multi-objective Approach | Formulates fairness–accuracy trade-offs as a stochastic multi-objective optimization problem. Proposes the Stochastic Multi-Gradient (SMG) algorithm using Disparate Impact and Equal Opportunity as constraints. Demonstrates Pareto frontiers on the UCI Adult dataset, showing tension between fairness and accuracy driven by proxy and protected attributes. | RQ1, RQ2 |
| 14 | Martinez et al. (2020) | Minimax Pareto Fairness: A Multi-objective Perspective | Models group fairness as multi-objective optimization where each sensitive group defines a fairness objective. Proposes Minimum-Maximum Pareto Fairness (MMPF) using neural networks with post-hoc corrections to reduce risk disparity. Evaluated on German Credit and Adult Income datasets with accuracy, Brier score, and cross-entropy metrics. | RQ1, RQ2 |
| 15 | Badar and Fisichella (2024) | Fair-CMNB: Advancing Fairness-Aware Stream Learning with Naïve Bayes and Multi-Objective Optimization | Proposes Fair-CMNB, a fairness- and imbalance-aware Mixed Naïve Bayes model for streaming credit data using multi-objective optimization. Introduces dynamic instance weighting to prioritize minority updates and control discrimination. Reports improved accuracy and fairness over baselines, emphasizing scalability and practicality for high-stakes credit scoring. | RQ1, RQ2 |
| 16 | J. Liu et al. (2024) | Credit Risk Prediction Based on Causal Machine Learning: Bayesian Network Learning, Default Inference, and Interpretation | Applies causal ML with Bayesian networks to model cause–effect relations and enable transparent decision analysis via DAGs. Uses SMOTE and L1 regularization for imbalance handling and feature selection. Supports interpretable, regulation-oriented what-if analysis across six real datasets. | RQ1, RQ2, RQ3 |
| 17 | Hickey et al. (2020) | Fairness by Explicability and Adversarial SHAP Learning | Proposes an adversarial SHAP framework linking fairness and explainability by penalizing predictions correlated with protected attributes via SHAP-based regularization. Uses surrogate auditing to mirror oversight. Demonstrates improved fairness and interpretability on Adult Income and proprietary credit datasets while maintaining strong predictive performance. | RQ1, RQ2, RQ3 |
| 18 | Lainez and Gardner (2023) | Algorithmic Credit Scoring in Vietnam: A Legal Proposal for Maximizing Benefits and Minimizing Risks | Analyzes regulatory gaps and ethical risks in Vietnam’s adoption of algorithmic credit scoring. Highlights discrimination, bias, opacity, and privacy concerns under FCRA, ECOA, FCA, and the EU AI Act. Advocates stronger legal oversight and interpretability standards to restore trust and fairness. | RQ3 |
| ID | Author(s) | Title | Summary | Related RQs |
|---|---|---|---|---|
| 19 | Nwafor et al. (2024) | Enhancing Transparency and Fairness in Automated Credit Decisions: An Explainable Novel Hybrid Machine Learning Approach | Proposes a hybrid CNN–XGBoost stacking model to improve accuracy and interpretability in credit scoring. Uses SHAP global explanations to examine feature effects on the Lending Club dataset. Reports high predictive performance while enhancing transparency and trust in automated lending decisions. | RQ1, RQ2 |
| 20 | S. Han et al. (2024) | NOTE: Non-parametric Oversampling Technique for Explainable Credit Scoring | Introduces NOTE, a non-parametric oversampling approach combining stacked autoencoders and conditional Wasserstein GANs to address severe class imbalance. Evaluated on Home Equity and Give Me Some Credit datasets; reports improved accuracy over ADSGAN and DeepSMOTE. Uses SHAP for global interpretability, linking imbalance correction with explainable credit scoring. | RQ1, RQ2 |
| 21 | Kozodoi et al. (2025) | Fighting Sampling Bias: A Framework for Training and Evaluating Credit Scoring Models | Proposes BASL, a bias-aware self-learning framework addressing sampling bias from missing rejected applicants. Uses a semi-supervised Bayesian approach to iteratively label unlabeled data while filtering outliers. Applied to Monedo’s real-world dataset; recovers up to 36% performance loss due to sampling bias and outperforms reweighting and Heckman-style methods. | RQ1, RQ2 |
| 22 | Shi et al. (2025) | Credit Scoring Prediction Using Deep Learning Models in the Financial Sector | Proposes a hybrid LSTM-based framework capturing temporal borrower behavior for credit scoring. Uses SMOTE, normalization, and one-hot encoding for imbalance handling. Introduces a hybrid loss with interpretability regularization enforcing feature sparsity, aiming to retain transparency without relying on post-hoc explainers. | RQ1, RQ2 |
| 23 | Bueff et al. (2022) | Machine Learning Interpretability for a Stress Scenario Generation in Credit Scoring Based on Counterfactuals | Compares SHAP with counterfactual methods for interpretability in credit scoring. Uses Genetic Algorithms to generate counterfactuals identifying minimal feature changes needed to alter outcomes. Links interpretability to robustness under stress scenarios and highlights how counterfactuals expose sensitive decision boundaries and bias-prone features. | RQ1, RQ2 |
| 24 | Kumar et al. (2022) | Equalizing Credit Opportunity in Algorithms: Aligning Algorithmic Fairness Research with U.S. Fair Lending Regulation | Aligns algorithmic fairness research with US anti-discrimination laws (ECOA, FCRA, HMDA). Discusses disparate impact/treatment as legal fairness criteria and the role of proxy attributes in bias. Advocates causal and counterfactual analysis and regulatory oversight for equitable, transparent AI-driven lending practices. | RQ2, RQ3 |
| ID | Author(s) | Title | Summary | Related RQs |
|---|---|---|---|---|
| 25 | Chai et al. (2025) | Farmers’ Credit Risk Evaluation with an Explainable Hybrid Ensemble Approach: A Closer Look in Microfinance | Proposes a hybrid ADASYN–LCE model combining adaptive resampling and local cascading ensembles for microfinance credit scoring. Uses SHAP and LIME for interpretability and fairness validation. Reports improved robustness and visibility for underserved populations through balanced learning and explainable ensemble modeling. | RQ1, RQ2 |
| 26 | Hlongwane et al. (2024) | A Novel Framework for Enhancing Transparency in Credit Scoring: Leveraging Shapley Values for Interpretable Credit Scorecards | Integrates SHAP explanations into credit scoring pipelines using XGBoost and Random Forest models. Improves transparency, regulatory alignment, and consumer trust by visualizing feature attributions. Evaluated on Taiwanese and Home Credit datasets, demonstrating interpretable performance aligned with explainability expectations in lending. | RQ1 |
| 27 | Zhang et al. (2025) | An Interpretable Credit Risk Assessment Model with Boundary Sample Identification | Proposes IAIBS, combining logistic regression and deep learning to handle ambiguous boundary samples. Uses ARPD to classify noise/anomalies and applies SHAP for interpretability. Reports improved AUC while enhancing transparency via boundary-aware pre-processing. | RQ1 |
| 28 | Hjelkrem and Lange (2023) | Explaining Deep Learning Models for Credit Scoring with SHAP: A Case Study Using Open Banking Data | Compares 1D-CNN and transfer-learning BERT models using open banking transactions for credit scoring. Uses SHAP to interpret deep models and support justification under regulatory mandates. Finds that 1D-CNN outperforms BERT in AUC and Brier score, emphasizing explainable deep learning for compliant credit assessment. | RQ1 |
| 29 | Bulut and Arslan (2025) | A Hybrid Approach to Credit Risk Assessment Using Bill Payment Habits Data and Explainable Artificial Intelligence | Addresses multi-class credit risk prediction using hybrid ensembles (LR, RF, SVM, NB, MLP) with SMOTE and ADASYN. Uses Mutual Information to capture proxy interactions and applies SHAP/LIME for interpretability. Reports strong performance with tree-based models and highlights explainable, balanced risk assessment in alternative-data settings. | RQ1, RQ2 |
| 30 | Ali Shahee and Patel (2025) | An Explainable ADASYN-Based Focal Loss Approach for Credit Assessment | Proposes an ANN integrating ADASYN resampling and Focal Loss to mitigate imbalance. Tested on the German dataset; reports improved accuracy and AUC over baselines. Uses SHAP and LIME for feature attribution, aiming to combine predictive performance with interpretability in credit assessment. | RQ1, RQ2 |
| ID | Author(s) | Title | Summary | Related RQs |
|---|---|---|---|---|
| 31 | Dastile et al. (2022) | Model-Agnostic Counterfactual Explanations in Credit Scoring | Introduces a GA-based counterfactual explanation framework for black-box credit models. Searches neighbouring instances that flip predictions with minimal feature changes, exposing decision boundaries and potential bias sources. Validated on German and HMEQ datasets, supporting model-agnostic interpretability for transparency and auditing. | RQ1 |
| 32 | Atif (2025) | VAE-INN: Variational Autoencoder with Integrated Neural Network Classifier for Imbalanced Credit Scoring | Proposes VAE-INN, a variational autoencoder guided by weighted loss to counter class imbalance in latent space. Assigns higher weights to minority classes to reduce Type II errors. Tested on the Taiwanese credit dataset and reports improved balanced accuracy and reliability over SMOTE/ADASYN-based baselines. | RQ1, RQ2 |
| 33 | Hartomo et al. (2025) | A Novel Weighted Loss TabTransformer Integrating Explainable AI for Imbalanced Credit Risk Datasets | Combines TabTransformer with weighted loss to address class imbalance while preserving interpretability. Applies SHAP for global feature attribution. Evaluated on German and BISAID datasets, reporting accuracy/AUC improvements and demonstrating explainable, balanced performance for tabular credit scoring. | RQ1, RQ2 |
| 34 | W. Han et al. (2023) | A Multi-layer Multi-view Stacking Model for Credit Risk Assessment | Introduces MLMVS, a stacking ensemble (LR, MLP, RF, KNN) with multi-view partitioning (personal, behavioral, history features). Uses LIME for instance-level interpretability. Reports gains in accuracy, precision, and specificity over baselines, supporting interpretable ensemble learning for default prediction. | RQ1 |
| 35 | Ridzuan et al. (2024) | AI in the Financial Sector: The Line Between Innovation, Regulation and Ethical Responsibility | Discusses AI governance in finance, emphasizing regulation, ethical responsibility, and human oversight. Identifies key challenges (privacy, fairness, accountability) and positions explainability as central for human decision-making. Advocates governance approaches aligned with societal values to foster trust in regulated financial AI. | RQ3 |
| 36 | Perry et al. (2023) | Algorithms for All: Can AI in the Mortgage Market Expand Access to Homeownership? | Examines whether AI can expand mortgage access while managing bias and equity concerns. Warns historical data may perpetuate discrimination. Recommends aligning AI outcomes with legal and ethical frameworks, ensuring demographic fairness, transparency, and human oversight to prevent disparate impact. | RQ2, RQ3 |
| ID | Author(s) | Title | Summary | Related RQs |
|---|---|---|---|---|
| 37 | Repetto (2025) | Multicriteria Interpretability Driven Deep Learning | Proposes a deep learning framework that injects interpretability constraints into training via multi-objective optimization. Uses soft constraints and weighted-sum scalarization to balance criteria. Demonstrates on the Polish bankruptcy dataset and visualizes effects using ALE plots, indicating interpretability-aware training can support generalization in high-stakes tasks. | RQ1 |
| 38 | Sulastri et al. (2025) | Sensitivity Analysis: Improving Inclusive Credit Scoring Algorithm Through Feature Weight and Penalty-Based Approach | Proposes feature-weight adjustment, penalty-based modeling, and a hybrid method to enhance inclusion in credit scoring. Evaluates inclusivity and performance across extensive hyperparameter combinations using XGBoost, CatBoost, RF, and DT. Improves inclusion by reweighting sensitive features but notes risks of dataset-specific overfitting and limited generalizability. | RQ1, RQ2 |
| 39 | L. H. Li et al. (2025) | Explainable AI-based LightGBM Prediction Model to Predict Default Borrower in Social Lending Platform | Implements LightGBM with LIME and SHAP for global/local interpretability in credit scoring. Uses sampling and RFE to address imbalance and dimensionality on the Lending Club dataset. Reports strong predictive performance and provides a reference pipeline for integrating explainability with ensemble models in social lending. | RQ1 |
| 40 | Dastile and Celik (2024) | Counterfactual Explanations with Multiple Properties in Credit Scoring | Proposes a counterfactual explanation method optimizing validity and sparsity to improve interpretability. Uses GA and PSO to find minimal feature changes that alter predictions. Highlights challenges such as drift sensitivity and missing data, and positions counterfactuals as a transparent alternative to feature-importance explanations for auditing. | RQ1, RQ3 |
| 41 | Aruleba and Sun (2024) | Effective Credit Risk Prediction Using Ensemble Classifiers With Model Explanation | Presents an ensemble framework (RF, AdaBoost, XGBoost, LightGBM) with SMOTE-ENN for imbalance correction and SHAP for interpretation. Evaluated on German and Australian datasets; reports improved recall/specificity and argues balanced, explainable ensemble learning improves generalization and auditability. | RQ1, RQ2 |
| 42 | Patron et al. (2020) | An Interpretable Automated Machine Learning Credit Risk Model | Proposes an AutoML framework integrating LIME for local interpretability of complex models. Reports near–deep learning performance while maintaining transparency through local perturbation-based explanations, supporting expert validation and contestability in credit risk decisions. | RQ1 |
| 43 | C. Li et al. (2024) | The Effect of AI-enabled Credit Scoring on Financial Inclusion: Evidence from an Underserved Population of Over One Million | Analyzes AI’s impact on financial inclusion using data from over one million underserved borrowers. Introduces weak signals (features weakly tied to financial status) to study inclusion and bias trade-offs. Warns protected attributes may amplify discrimination and argues for balanced adoption to improve access while managing equity risks. | RQ2, RQ3 |
Appendix C
| Author | Category | Method | Mechanism | Strengths | Limitations |
|---|---|---|---|---|---|
| Kozodoi et al. (2022) | Pre-processing | Reweighing | Adjusts training sample weights so disadvantaged groups receive higher influence during training, targeting independence (parity). | Model-agnostic; simple to apply before training; can reduce discrimination at low implementation cost. | Smaller fairness gains than the strongest post- processing option in their comparison; may require repeated data-pipeline adjustments. |
| Kozodoi et al. (2022) | Pre-processing | Disparate Impact Remover (DIR) | Transforms feature values to reduce distribution differences across protected groups, reducing dependence on protected attributes. | Improves fairness without changing model architecture; model-agnostic. | Worse profit–fairness trade-off than the best in-processing option (PRR) in their reported results. |
| Kozodoi et al. (2022) | In-processing | Prejudice Remover Regularizer (PRR) | Adds a regularization term to the training objective that penalizes unfairness using a prejudice index, with a tunable penalty weight. | Tunable trade-off; achieves better profit– fairness trade-off than DIR in their evaluation. | Invasive; modifies the training objective/ scorecards and increases implementation burden. |
| Kozodoi et al. (2022) | In-processing | Adversarial Debiasing | Trains a predictor while an auxiliary adversary tries to infer the protected attribute; penalizes the predictor when the adversary succeeds. | Tunable fairness–profit balance via meta-parameters. | Requires retraining and pipeline changes; more invasive than post-processing. |
| Kozodoi et al. (2022) | In-processing | Meta-fair Classification | Optimizes a classifier under fairness constraints (e.g., independence/separation) with trade-off meta-parameters controlling accuracy vs. fairness. | Explicit control over fairness–accuracy trade-offs. | Model/training specific; requires retraining and integration effort. |
| Kozodoi et al. (2022) | Post-processing | Reject Option Classification (ROC) | Relabels decisions in an uncertainty region in favor of the disadvantaged group to improve group parity. | Strong fairness gains; largely preserves the existing scoring pipeline. | Can reduce profitability compared to in-processing approaches; acts only on the decision boundary. |
| Author | Category | Method | Mechanism | Strengths | Limitations |
|---|---|---|---|---|---|
| Kozodoi et al. (2022) | Post-processing | Equalized Odds Post-processing | Uses group-specific thresholds to equalize error rates across groups (separation/equalized odds). | Post-hoc; model-agnostic; can reduce discrimination at low cost up to a point on the Pareto frontier. | Strict fairness may require large profit/utility sacrifices across datasets. |
| Kozodoi et al. (2022) | Post-processing | Group-wise Platt Scaling | Calibrates predicted probabilities per group to satisfy sufficiency (risk meaning consistent across groups). | Post-hoc; preserves pipeline; supports calibration for sufficiency- oriented compliance. | Inherits post-processing trade-offs; does not address upstream bias; cannot satisfy all criteria simultaneously. |
| Moldovan (2023) | In-processing | GerryFair | Learner–auditor adversarial approach minimizing unfairness via iterative constraint enforcement (targets individual fairness). | Explicitly targets individual-level unfairness, not only group parity. | Hard to tune; may overfit on small credit datasets. |
| Moldovan (2023) | In-processing | Grid Search Reduction | Reformulates learning as a cost-sensitive reduction and searches constraint weights to obtain fairness–accuracy trade-offs. | Allows explicit exploration and selection of trade-off points. | Accuracy may degrade sharply under strict constraints. |
| Zehlike et al. (2025) | Post-processing | FAIM | Optimal-transport interpolation between incompatible fairness criteria (calibration, balance for positives, balance for negatives) using weighted constraints. | Provides a tunable mechanism to navigate incompatibility between fairness criteria. | Weight selection embeds normative/legal judgment; may not match a single regulatory interpretation. |
| Valdrighi et al. (2025) | In-processing | Demographic Parity/Equal Opportunity Classifier | Modifies LR training to minimize loss subject to a correlation constraint between predictions and sensitive attributes (tunable constant). | Simple and interpretable; tunable trade-off. | Requires sensitive attributes during training; remains a trade-off rather than a guarantee. |
| Author | Category | Method | Mechanism | Strengths | Limitations |
|---|---|---|---|---|---|
| Valdrighi et al. (2025) | In-processing | FairGBM (constrained gradient boosting) | Alters boosting to jointly minimize prediction loss and a differentiable proxy of a fairness metric (e.g., DP/EO) during training. | Strong tabular performance with embedded fairness control using constrained learning. | Model-class specific; depends on proxy design and differentiability of fairness objectives. |
| Valdrighi et al. (2025) | Post-processing | Threshold Optimizer | Builds separate ROC curves per group and selects thresholds that minimize loss within the feasible fair region, yielding group-specific thresholds. | Consistently reaches fairness targets with minimum accuracy loss in their comparisons. | Requires sensitive attributes at prediction time, which may be legally/practically constrained. |
| Valdrighi et al. (2025) | Post-processing (general) | Position on post-processing (general) | Alters outputs of black-box models to satisfy fairness constraints (e.g., group thresholds) without changing model training. | Versatile and suitable for black-boxes and fixed scorecards. | Less tunable than in-processing; may yield weaker improvements relative to pre-/in-processing. |
| S. Liu and Vicente (2022) | In-processing | Stochastic Multi-Gradient (SMG) bi-objective optimization | Frames fairness as stochastic multi-objective optimization and aggregates gradients of prediction loss and fairness penalty along the Pareto frontier (e.g., DI/EO constraints). | Produces smooth Pareto frontiers and stable convergence across fairness constraints. | Not reported (explicit method-level limitations not stated beyond general trade-offs). |
| Badar and Fisichella (2024) | Hybrid | Fair-CMNB | Stream-learning Mixed Naïve Bayes with multi-objective optimization; dynamic instance weighting for imbalance and discrimination control (targets Statistical Parity; uses causal fairness via ATE/FACE). | Low discrimination (SP near 0) while improving accuracy relative to baselines; supports streaming settings. | Does not guarantee global fairness across settings; gains are dataset-dependent. |
| S. Han et al. (2024) | Pre-processing (oversampling) | NOTE | Non-parametric stacked autoencoders to learn latent structure, then conditional Wasserstein GAN oversampling for mixed categorical/ numerical features. | Stronger performance than classic oversampling (e.g., SMOTE) in their comparisons. | Not reported. |
| Author | Category | Method | Mechanism | Strengths | Limitations |
|---|---|---|---|---|---|
| Kozodoi et al. (2025) | Pre-processing (bias-aware rejection inference) | BASL: Bias-Aware Self-Learning | Semi-supervised approach that iteratively infers labels for rejected/unlabeled instances to reduce sampling bias and improve training representativeness. | Outperforms parceling, reweighting, and Heckman-style correction; recovers a substantial share of predictive loss attributed to sampling bias in their case study. | Not reported. |
| Chai et al. (2025) | Pre-processing | ADASYN-LCE | Combines ADASYN oversampling with a Local Cascading Ensemble (bagging/boosting/local cascading) to improve robustness for underserved populations under imbalance. | Improves generalization under imbalance; LCE balances bias–variance across subsets of the data. | May struggle to address bias and variance simultaneously under some conditions (as noted by the authors). |
| Hartomo et al. (2025) | In-processing | Weighted TabTransformer | Integrates a weighted cross-entropy objective into TabTransformer to give larger gradients to minority classes. | Supports joint improvements in performance under imbalance and fairness-related objectives in their framing. | Prone to overfitting if weighting is mis-specified. |
References
- Adegoke, T., Ofodile, O., Ochuba, N., & Akinrinol, O. (2024). Evaluating the fairness of credit scoring models: A literature review on mortgage accessibility for under-reserved populations. GSC Advanced Research and Reviews, 18(3), 189–199. [Google Scholar] [CrossRef]
- Ali Shahee, S., & Patel, R. (2025). An explainable ADASYN-based focal loss approach for credit assessment. Journal of Forecasting, 44, 1513–1530. [Google Scholar] [CrossRef]
- Alufaisan, Y., Marusich, L. R., Bakdash, J. Z., Zhou, Y., & Kantarcioglu, M. (2021). Does explainable artificial intelligence improve human decision-making? In Proceedings of the AAAI conference on artificial intelligence (Vol. 35, pp. 6618–6626). AAAI Press. [Google Scholar] [CrossRef]
- Alves, G., Bernier, F., Couceiro, M., Makhlouf, K., Palamidessi, C., & Zhioua, S. (2023). Survey on fairness notions and related tensions. EURO Journal on Decision Processes, 11, 100033. [Google Scholar] [CrossRef]
- Amekoe, K. M., Azzag, H., Dagdia, Z. C., Lebbah, M., & Jaffre, G. (2024). Exploring accuracy and interpretability trade-off in tabular learning with novel attention-based models. Neural Computing and Applications, 36(30), 18583–18611. [Google Scholar] [CrossRef]
- Aruleba, I., & Sun, Y. (2024). Effective credit risk prediction using ensemble classifiers with model explanation. IEEE Access, 12, 115015–115025. [Google Scholar] [CrossRef]
- Atif, D. (2025). VAE-INN: Variational autoencoder with integrated neural network classifier for imbalanced credit scoring, utilizing weighted loss for improved accuracy. Computational Economics. [Google Scholar] [CrossRef]
- Badar, M., & Fisichella, M. (2024). Fair-CMNB: Advancing fairness-aware stream learning with naïve bayes and multi-objective optimization. Big Data and Cognitive Computing, 8(2), 16. [Google Scholar] [CrossRef]
- Balashankar, A., & Lees, A. (2022). The need for transparent demographic group trade-offs in credit risk and income classification. In Proceedings of the international conference on information (pp. 344–354). Springer International Publishing. [Google Scholar]
- Bartlett, R., Morse, A., Stanton, R., & Wallace, N. (2022). Consumer-lending discrimination in the FinTech era. Journal of Financial Economics, 143(1), 30–56. [Google Scholar] [CrossRef]
- Basel Committee on Banking Supervision. (2013). Principles for effective risk data aggregation and risk reporting (Technical report no. 8, Basel committee publication no. 239). Bank for International Settlements. Available online: https://www.bis.org/publ/bcbs239.htm (accessed on 30 September 2025).
- Berg, T., Burg, V., Gombović, A., & Puri, M. (2020). On the rise of Fintechs: Credit scoring using digital footprints. The Review of Financial Studies, 33(7), 2845–2897. [Google Scholar] [CrossRef]
- Brzezinski, D., Stachowiak, J., Stefanowski, J., Szczech, I., Susmaga, R., Aksenyuk, S., & Yasinskyi, O. (2024). Properties of fairness measures in the context of varying class imbalance and protected group ratios. ACM Transactions on Knowledge Discovery from Data, 18(7), 1–18. [Google Scholar] [CrossRef]
- Bueff, A. C., Cytryński, M., Calabrese, R., Jones, M., Roberts, J., Moore, J., & Brown, I. (2022). Machine learning interpretability for a stress scenario generation in credit scoring based on counterfactuals. Expert Systems with Applications, 202, 117271. [Google Scholar] [CrossRef]
- Bulut, C., & Arslan, E. (2025). A hybrid approach to credit risk assessment using bill payment habits data and explainable artificial intelligence. Applied Sciences, 15(10), 5723. [Google Scholar] [CrossRef]
- Bücker, M., Szepannek, G., Gosiewska, A., & Biecek, P. (2022). Transparency, auditability, and explainability of machine learning models in credit scoring. Journal of the Operational Research Society, 73(1), 70–90. [Google Scholar] [CrossRef]
- Cao, W., He, Y., Wang, W., Zhu, W., & Demazeau, Y. (2021). Ensemble methods for credit scoring of chinese peer-to-peer loans. Journal of Credit Risk, 17, 79–115. [Google Scholar] [CrossRef]
- Caton, S., & Haas, C. (2024). Fairness in machine learning: A survey. ACM Computing Surveys, 56(7), 1–38. [Google Scholar] [CrossRef]
- Chacko, A., & Aravindhar, D. J. (2025, February 21–22). Enhancing fairness and accuracy in credit score analysis: A novel framework utilizing kernel PCA. 2025 International Conference on Information Technology, Innovation and Intelligent Systems (ICITIIT), Kottayam, India. [Google Scholar] [CrossRef]
- Chai, N., Abedin, M. Z., Yang, L., & Shi, B. (2025). Farmers’ credit risk evaluation with an explainable hybrid ensemble approach: A closer look in microfinance. Pacific-Basin Finance Journal, 89, 102612. [Google Scholar] [CrossRef]
- Chen, Y., Calabrese, R., & Martin-Barragán, B. (2024). Interpretable machine learning for imbalanced credit scoring datasets. European Journal of Operational Research, 312(1), 357–372. [Google Scholar] [CrossRef]
- Cornacchia, G., Anelli, V. W., Narducci, F., Ragone, A., & Di Sciascio, E. (2023). A general architecture for a trustworthy creditworthiness-assessment platform in the financial domain. AETiC, 7, 56–64. [Google Scholar] [CrossRef]
- Corrales-Barquero, R., Marín-Raventós, G., & Barrantes, E. G. (2021, October 27–28). A review of gender bias mitigation in credit scoring models. 2021 International Conference on Electrical, Electronics and Related Data Science (EE-RDS), Johannesburg, South Africa. [Google Scholar] [CrossRef]
- C-Rella, J., Martínez-Rego, D., & Vilar Fernández, J. M. (2025). Cost-sensitive reinforcement learning for credit risk. Expert Systems with Applications, 272, 126708. [Google Scholar] [CrossRef]
- Das, S., Stanton, R., & Wallace, N. (2023). Algorithmic fairness. Annual Review of Financial Economics, 15(1), 565–593. [Google Scholar] [CrossRef]
- Dastile, X., & Celik, T. (2024). Counterfactual explanations with multiple properties in credit scoring. IEEE Access, 12, 110713–110728. [Google Scholar] [CrossRef]
- Dastile, X., Celik, T., & Vandierendonck, H. (2022). Model-agnostic counterfactual explanations in credit scoring. IEEE Access, 10, 69543–69554. [Google Scholar] [CrossRef]
- de Castro Vieira, J. R., Barboza, F. L. D. M., Cajueiro, D. O., & Kimura, H. (2025). Towards fair AI: Mitigating bias in credit decisions—A systematic literature review. Journal of Risk and Financial Management, 18, 228. [Google Scholar] [CrossRef]
- Dessain, J., Bentaleb, N., & Viñas, F. (2023). Cost of explainability in AI: An example with credit scoring models. In Proceedings of the world conference on explainable artificial intelligence (pp. 498–516). Springer Nature. [Google Scholar] [CrossRef]
- European Central Bank. (2024). Supervisory guide on risk data aggregation and risk reporting (Technical report, supervisory guide). European Central Bank, Banking Supervision. Available online: https://www.bankingsupervision.europa.eu/ecb/pub/pdf/ssm.supervisory_guides240503_riskreporting.en.pdf (accessed on 30 September 2025).
- Goethals, S., Martens, D., & Calders, T. (2024). Precof: Counterfactual explanations for fairness. Machine Learning, 113(5), 3111–3142. [Google Scholar] [CrossRef]
- Griffith, M. A. (2023). AI lending and the ECOA: Avoiding accidental discrimination. North Carolina Banking Institute, 27, 349–381. Available online: https://scholarship.law.unc.edu/ncbi/vol27/iss1/16 (accessed on 30 September 2025).
- Han, S., Jung, H., Yoo, P. D., Provetti, A., & Calì, A. (2024). NOTE: Non-parametric oversampling technique for explainable credit scoring. Scientific Reports, 14(1), 26070. [Google Scholar] [CrossRef]
- Han, W., Gu, X., & Jian, L. (2023). A multi-layer multi-view stacking model for credit risk assessment. Intelligent Data Analysis, 27(5), 1457–1475. [Google Scholar] [CrossRef]
- Hartomo, K. D., Arthur, C., & Nataliani, Y. (2025). A novel weighted loss tabtransformer integrating explainable AI for imbalanced credit risk datasets. IEEE Access, 13, 31045–31056. [Google Scholar] [CrossRef]
- Hickey, J. M., Di Stefano, P. G., & Vasileiou, V. (2020). Fairness by explicability and adversarial SHAP learning. In Joint European conference on machine learning and knowledge discovery in databases (ECML PKDD) (pp. 174–190). Springer International Publishing. [Google Scholar]
- Hjelkrem, L. O., & Lange, P. E. D. (2023). Explaining deep learning models for credit scoring with textual transaction data. Journal of Risk and Financial Management, 16(4), 221. [Google Scholar] [CrossRef]
- Hlongwane, R., Ramabao, K., & Mongwe, W. (2024). A novel framework for enhancing transparency in credit scoring: Leveraging Shapley values for interpretable credit scorecards. PLoS ONE, 19(8), e0308718. [Google Scholar] [CrossRef] [PubMed]
- Hurlin, C., Pérignon, C., & Saurin, S. (2024). The fairness of credit scoring models. Management Science, 70(11), 1234–1256. [Google Scholar] [CrossRef]
- Jiang, Y., Fang, X., & Wang, Z. (2024). Disparity and discrimination in consumer credit markets: Evidence from online peer-to-peer lending. Pacific-Basin Finance Journal, 83, 102237. [Google Scholar] [CrossRef]
- Kanaparthi, V. (2023, April 26–28). Credit risk prediction using ensemble machine learning algorithms. 2023 International Conference on Inventive Computation Technologies (ICICT) (pp. 41–47), Lalitpur, Nepal. [Google Scholar] [CrossRef]
- Keele, S. (2007). Guidelines for performing systematic literature reviews in software engineering (EBSE technical report, version 2.3). Available online: https://www.elsevier.com/__data/promis_misc/525444systematicreviewsguide.pdf (accessed on 17 October 2025).
- Koulu, R. (2020). Human control over automation: EU policy and AI ethics. European Journal of Legal Studies, 12, 9–46. [Google Scholar] [CrossRef]
- Kozodoi, N., Jacob, J., & Lessmann, S. (2022). Fairness in credit scoring: Assessment, implementation and profit implications. European Journal of Operational Research, 297(3), 1083–1094. [Google Scholar] [CrossRef]
- Kozodoi, N., Lessmann, S., Alamgir, M., Moreira-Matias, L., & Papakonstantinou, K. (2025). Fighting sampling bias: A framework for training and evaluating credit scoring models. European Journal of Operational Research, 324(2), 616–628. [Google Scholar] [CrossRef]
- Kuiper, O., van den Berg, M., van der Burgt, J., & Leijnen, S. (2021). Exploring explainable AI in the financial sector: Perspectives of banks and supervisory authorities. In Proceedings of the benelux conference on artificial intelligence (pp. 105–119). Springer International Publishing. [Google Scholar]
- Kumar, I. E., Hines, K. E., & Dickerson, J. P. (2022). Equalizing credit opportunity in algorithms: Aligning algorithmic fairness research with US fair lending regulation. In Proceedings of the 2022 AAAI/ACM conference on AI, ethics, and society (pp. 357–368). Association for Computing Machinery. [Google Scholar] [CrossRef]
- Lainez, N., & Gardner, J. (2023). Algorithmic credit scoring in Vietnam: A legal proposal for maximizing benefits and minimizing risks. Asian Journal of Law and Society, 10(3), 401–432. [Google Scholar] [CrossRef]
- Langenbucher, K. (2020). Responsible AI-based credit scoring—A legal framework. European Business Law Review, 31(4), 527–572. [Google Scholar] [CrossRef]
- Langenbucher, K., & Corcoran, P. (2022). Responsible AI credit scoring—A lesson from upstart.com. In Digital finance in Europe: Law, regulation, and governance. De Gruyter. [Google Scholar]
- Li, C., Wang, H., Jiang, S., & Gu, B. (2024). The effect of AI-enabled credit scoring on financial inclusion: Evidence from an underserved population of over one million. MIS Quarterly, 48(4), 1803–1834. [Google Scholar] [CrossRef]
- Li, L. H., Sharma, A. K., & Cheng, S. T. (2025). Explainable AI based LightGBM prediction model to predict default borrower in social lending platform. Intelligent Systems with Applications, 26, 200514. [Google Scholar] [CrossRef]
- Li, W., Paraschiv, F., & Sermpinis, G. S. (2022). A data-driven explainable case-based reasoning approach for financial risk detection. Quantitative Finance, 22, 2257–2274. [Google Scholar] [CrossRef]
- Li, Y., Wang, X., Djehiche, B., & Hu, X. (2020). Credit scoring by incorporating dynamic networked information. European Journal of Operational Research, 286(3), 1103–1112. [Google Scholar] [CrossRef]
- Li, Z., Hu, X., Li, K., Zhou, F., & Shen, F. (2020). Inferring the outcomes of rejected loans: An application of semisupervised clustering. Journal of the Royal Statistical Society: Series A, 183, 631–654. [Google Scholar] [CrossRef]
- Liao, J., Wang, W., Xue, J., Lei, A., Han, X., & Lu, K. (2022). Combating sampling bias: A self-training method in credit risk models. Proceedings of the AAAI Conference on Artificial Intelligence, 36, 12566–12572. [Google Scholar] [CrossRef]
- Liu, J., Zhang, X., & Xiong, H. (2024). Credit risk prediction based on causal machine learning: Bayesian network learning, default inference, and interpretation. Journal of Forecasting, 43(5), 1625–1660. [Google Scholar] [CrossRef]
- Liu, S., & Vicente, L. N. (2022). Accuracy and fairness trade-offs in machine learning: A stochastic multi-objective approach. Computational Management Science, 19(3), 513–537. [Google Scholar] [CrossRef]
- Martinez, N., Bertran, M., & Sapiro, G. (2020, July 13–18). Minimax pareto fairness: A multi-objective perspective. International Conference on Machine Learning (ICML) (pp. 6755–6764), Virtual. [Google Scholar]
- Mestiri, S., & Hiboun, S. M. (2024). Credit scoring using machine learning and deep learning-based models. Data Science in Finance and Economics, 4(2), 236–248. [Google Scholar] [CrossRef]
- Moher, D., Liberati, A., Tetzlaff, J., Altman, D. G., & Group, P. (2010). Preferred reporting items for systematic reviews and meta-analyses: The PRISMA statement. International Journal of Surgery, 8(5), 336–341. [Google Scholar] [CrossRef]
- Moldovan, D. (2023). Algorithmic decision making methods for fair credit scoring. IEEE Access, 11, 59729–59743. [Google Scholar] [CrossRef]
- Mou, Y., Pu, Z., Feng, D., Luo, Y., Lai, Y., Huang, J., Tian, Y., & Xiao, F. (2025). Cost-aware credit-scoring framework based on resampling and feature selection. Computational Economics, 66, 3007–3032. [Google Scholar] [CrossRef]
- Muñoz-Cancino, R., Bravo, C., Ríos, S. A., & Graña, M. (2023). On the dynamics of credit history and social interaction features, and their impact on creditworthiness assessment performance. Expert Systems with Applications, 218, 119599. [Google Scholar] [CrossRef]
- Nwafor, C. N., Nwafor, O., & Brahma, S. (2024). Enhancing transparency and fairness in automated credit decisions: An explainable novel hybrid machine learning approach. Scientific Reports, 14(1), 25174. [Google Scholar] [CrossRef]
- Page, M. J., McKenzie, J. E., Bossuyt, P. M., Boutron, I., Hoffmann, T. C., Mulrow, C. D., Shamseer, L., Tetzlaff, J. M., Akl, E. A., Brennan, S. E., Chou, R., Glanville, J., Grimshaw, J. M., Hróbjartsson, A., Lalu, M. M., Li, T., Loder, E. W., Mayo-Wilson, E., McDonald, S., & Moher, D. (2021). Updating guidance for reporting systematic reviews: Development of the PRISMA 2020 statement. Journal of Clinical Epidemiology, 134, 103–112. [Google Scholar] [CrossRef]
- Patron, G., Leon, D., Lopez, E., & Hernandez, G. (2020). An interpretable automated machine learning credit risk model. In Workshop on engineering applications (pp. 16–23). Springer International Publishing. [Google Scholar]
- Peng, Z., Mo, W., Duan, C., Li, Q., & Zhou, B. (2023). Learning from active human involvement through proxy value propagation. In Proceedings of the 37th conference on neural information processing systems (NeurIPS 2023). Curran Associates, Inc. Available online: https://metadriverse.github.io/pvp (accessed on 17 October 2025).
- Perry, V. G., Martin, K., & Schnare, A. (2023). Algorithms for all: Can AI in the mortgage market expand access to homeownership? AI, 4(4), 888–903. [Google Scholar] [CrossRef]
- Ratul, Q. E. A., Serra, E., & Cuzzocrea, A. (2021, December 15–18). Evaluating attribution methods in machine learning interpretability. 2021 IEEE International Conference on Big Data (Big Data) (pp. 5239–5245), Orlando, FL, USA. [Google Scholar]
- Repetto, M. (2025). Multicriteria interpretability driven deep learning. Annals of Operations Research, 346(2), 1621–1635. [Google Scholar] [CrossRef]
- Ribeiro-Flucht, L., Chen, X., & Meurers, D. (2024). Explainable AI in language learning: Linking empirical evidence and theoretical concepts in proficiency and readability modeling of portuguese. In Proceedings of the 19th workshop on innovative use of NLP for building educational applications (BEA 2024) (pp. 199–209). Association for Computational Linguistics. Available online: https://aclanthology.org/2024.bea-1.17 (accessed on 4 October 2025).
- Ridzuan, N. N., Masri, M., Anshari, M., Fitriyani, N. L., & Syafrudin, M. (2024). AI in the financial sector: The line between innovation, regulation and ethical responsibility. Information, 15(8), 432. [Google Scholar] [CrossRef]
- Roa, L., Correa-Bahnsen, A., Suarez, G., Cortés-Tejada, F., Luque, M. A., & Bravo, C. (2021). Super-app behavioral patterns in credit risk models: Financial, statistical and regulatory implications. Expert Systems with Applications, 169, 114486. [Google Scholar] [CrossRef]
- Shi, X., Tang, D., & Yu, Y. (2025). Credit scoring prediction using deep learning models in the financial sector. IEEE Access, 13, 130731–130746. [Google Scholar] [CrossRef]
- Sulastri, R., Ding, A. Y., & Janssen, M. (2025, June 18–20). Sensitivity analysis: Improving inclusive credit scoring algorithm through feature weight and penalty-based approach. Proceedings of the 2025 Eleventh International Conference on Edemocracy & Egovernment (ICEDEG) (pp. 54–61), Bern, Switzerland. [Google Scholar] [CrossRef]
- Talaat, F. M., Aljadani, A., Badawy, M., & Elhosseini, M. (2024). Toward interpretable credit scoring: Integrating explainable artificial intelligence with deep learning for credit card default prediction. Neural Computing and Applications, 36(9), 4847–4865. [Google Scholar] [CrossRef]
- Tiukhova, E., Salcuni, A., Oguz, C., Niglio, M., Storti, G., Forte, F., Baesens, B. M., & Snoeck, M. (2025). Boosting credit risk data quality using machine learning and eXplainable AI techniques. In Machine learning and principles and practice of knowledge discovery in databases. Springer. [Google Scholar] [CrossRef]
- Valdrighi, G., Ribeiro, A. M., Pereira, J. S. B., Guardieiro, V., Hendricks, A., Miranda Filho, D., & Medeiros Raimundo, M. (2025). Best practices for responsible machine learning in credit scoring. Neural Computing and Applications, 37, 20781–20821. [Google Scholar] [CrossRef]
- Vuković, D. B., Dekpo-Adza, S., & Matović, S. (2025). AI integration in financial services: A systematic review of trends and regulatory challenges. Humanities and Social Sciences Communications, 12, 562. [Google Scholar] [CrossRef]
- Wang, W., Lesner, C., Ran, A., Rukonic, M., Xue, J., & Shiu, E. (2020). Using small business banking data for explainable credit risk scoring. In Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 34, pp. 13396–13401). AAAI Press. [Google Scholar] [CrossRef]
- Wu, Z., Dong, Y., Li, Y., & Liu, Y. (2025). A ‘Divide and Conquer’ reject inference approach leveraging graph-based semi-supervised learning. Annals of Operations Research. [Google Scholar] [CrossRef]
- Xie, W., He, J., Huang, F., & Ren, J. (2025). Operational risk assessment of commercial banks’ supply chain finance. Systems, 13(2), 76. [Google Scholar] [CrossRef]
- Zacharias, J., von Zahn, M., Chen, J., & Hinz, O. (2022). Designing a feature selection method based on explainable artificial intelligence. Electronic Markets, 32, 2159–2184. [Google Scholar] [CrossRef]
- Zehlike, M., Loosley, A., Jonsson, H., Wiedemann, E., & Hacker, P. (2025). Beyond incompatibility: Trade-offs between mutually exclusive fairness criteria in machine learning and law. Artificial Intelligence, 340, 104280. [Google Scholar] [CrossRef]
- Zhang, R., Li, I., & Ding, Z. (2025). An Interpretable credit risk assessment model with boundary sample identification. PeerJ Computer Science, 11, e2988. [Google Scholar] [CrossRef] [PubMed]




| Component | Definition |
|---|---|
| Population | Defines the domain and subject of deployed AI models highlighted in past studies. It refers to credit scoring predictions that are based on evaluating default risks using historical financial information that is exclusive to application risk assessments. |
| Intervention | Refers to the technique or method employed to tackle single or joint problems given in credit scoring, including performance, fairness and explainability. It corresponds to the deployment of AI to solve the credit scoring prediction problem and can incorporate multiple pillars. |
| Comparison | Specifies what the intervention is evaluated against. In this context, studies must make use of existing baselines to benchmark their methods or highlight the trade-offs across the pillars. This element is crucial, as it analyzes pillar interactions. |
| Outcome | Captures what was measured and reported, illustrating the outcomes of interest, such as performance indicators, fairness metrics, explainability with human comprehension, and quantitative assessment of the trade-off analysis between pillars. |
| Context | Defines the application environment and publication constraints. It determines whether the study is exclusive to the credit scoring domain, published in a peer-reviewed journal or conference, or demonstrates recency in terms of reporting fairness and explainability integrated into modeling. In addition, it pinpoints where trade-offs arise or other gaps exist, given the three pillars. |
| TITLE-ABS-KEY (“credit scoring” OR “credit risk”) |
| AND TITLE-ABS-KEY (“machine learning” OR “deep learning” OR “artificial intelligence” OR “reinforcement learning” OR “deep reinforcement learning”) |
| AND TITLE-ABS-KEY (“explainable AI” OR “interpretability” OR “model transparency” OR “XAI” OR “fairness” OR “bias” OR “discrimination” OR “protected attribute*”) |
| AND PUBYEAR ≥ 2020 AND PUBYEAR ≤ 2025 |
| Dimension | Submetric | Description | Scoring Criteria (0–3) |
|---|---|---|---|
| Relevance | Pillar Alignment (R1) | Does the study explicitly address at least one of the six conceptual pillars (Explainable AI, Fairness, Imbalance, Protected Attributes, Regulation, Human Intervention)? | none, peripheral, central and focus |
| Relevance | RQ Fit (R2) | Does the paper contribute evidence toward one or more of the three research questions (RQs)? | none, weak, moderate and strong |
| Rigor | Methodological Soundness (R3) | Are the models or methods clearly described, validated, and reproducible? | poor, basic, robust, and state of the art |
| Rigor | Evaluation Depth (R4) | Does the study use real-world datasets, multiple metrics (e.g., AUC, fairness measures), or comparative baselines? | minimal, partial, strong and comprehensive |
| Reach | Cross-Context Awareness (R5) | Does the study consider regulatory compliance (e.g., ECOA, GDPR) or cross-national fairness transferability? | none, partial, clear attempt, and deep analysis |
| Reach | Integration of Dimensions (R6) | Does it combine multiple pillars (e.g., fairness + explainability or imbalance + regulation)? | siloed, minor combination, partial integration and holistic framework |
| Quality | Transparency (Q1) and Reproducibility | Are code, data, or supplementary reproducibility resources available? | not available, vague, partial and open and reproducible |
| Quality | Practical Relevance (Q2) | Does the study provide actionable insights for deployment (e.g., industry adoption, human oversight, legal compliance)? | theoretical, limited, moderate and strong |
| Database/Publisher Family | Count | % |
|---|---|---|
| IEEE Xplore | 6 | 13.95% |
| SpringerLink | 5 | 11.63% |
| Elsevier (ScienceDirect) | 3 | 6.98% |
| Open Access (Public) | 3 | 6.98% |
| ACM Digital Library | 1 | 2.33% |
| MDPI | 1 | 2.33% |
| Other/Misc. | 24 | 55.81% |
| Dimension | Intersect | Count |
|---|---|---|
| Explainability | Imbalance | 10 |
| Fairness | 7 | |
| Protected Attributes | 6 | |
| Regulation | 6 | |
| Human Intervention | 6 | |
| Fairness | Protected Attributes | 21 |
| Regulation | 11 | |
| Human Intervention | 11 | |
| Imbalance | 6 | |
| Protected Attributes | Regulation | 11 |
| Human Intervention | 11 | |
| Imbalance | 4 | |
| Regulation | Human Intervention | 10 |
| Paper | Dataset | Models Compared | Metric | Interp. | Black-Box | |
|---|---|---|---|---|---|---|
| Nwafor et al. (2024) | LendingClub | LR vs. XGB | AUC | 0.95 | 0.99 | +0.04 |
| H-score | 0.95 | 0.95 | 0.00 | |||
| Precision (w) | 0.93 | 0.95 | +0.02 | |||
| Recall (w) | 0.92 | 0.94 | +0.02 | |||
| F1-score (w) | 0.92 | 0.94 | +0.02 | |||
| S. Han et al. (2024) | HE & GMSC (best-case) | LR vs. RF/GB | AUC | 0.9750 | 0.9891 | +0.0141 |
| Chai et al. (2025) | Farmers (best-case) | DT vs. LCE | AUC | 0.622 | 0.784 | +0.162 |
| Hlongwane et al. (2024) | Taiwan | LR vs. RF (best-case) | AUC | 0.74891 | 0.75929 | +0.01038 |
| Home Credit | LR vs. XGB (best-case) | AUC | 0.69644 | 0.69766 | +0.00122 | |
| Zhang et al. (2025) | PCL | DT vs. IAIBS | AUC | 86.81 | 89.17 | +2.36 |
| Accuracy | 76.95 | 79.32 | +2.37 | |||
| F1 | 56.79 | 59.39 | +2.60 | |||
| FICO | LR vs. IAIBS | AUC | 77.48 | 79.86 | +2.38 | |
| Accuracy | 71.52 | 74.55 | +3.03 | |||
| F1 | 73.93 | 76.51 | +2.58 | |||
| CCF | DT vs. IAIBS | AUC | 96.04 | 97.48 | +1.44 | |
| LR vs. IAIBS | Accuracy | 97.45 | 97.56 | +0.11 | ||
| F1 | 86.71 | 88.69 | +1.98 | |||
| VL | LR vs. IAIBS | AUC | 61.91 | 66.03 | +4.12 | |
| Accuracy | 59.32 | 62.70 | +3.38 | |||
| F1 | 60.18 | 63.31 | +3.13 | |||
| Ali Shahee and Patel (2025) | Proprietary | DT vs. ANN (ADASYN+FL) | Accuracy | 0.720 | 0.783 | +0.063 |
| F1-score | 0.644 | 0.747 | +0.103 | |||
| AUC | 0.737 | 0.812 | +0.075 | |||
| G-mean | 0.602 | 0.747 | +0.145 | |||
| L. H. Li et al. (2025) | LendingClub (2007–2020) | LR vs. LightGBM | AUC | 0.91 | 0.94 | +0.03 |
| LR vs. CatBoost | AUC | 0.91 | 0.94 | +0.03 | ||
| LR vs. RF | AUC | 0.91 | 0.93 | +0.02 | ||
| LR vs. MLP | AUC | 0.91 | 0.91 | +0.00 | ||
| SVM vs. LightGBM | AUC | 0.88 | 0.94 | +0.06 | ||
| SVM vs. CatBoost | AUC | 0.88 | 0.94 | +0.06 | ||
| SVM vs. RF | AUC | 0.88 | 0.93 | +0.05 | ||
| SVM vs. MLP | AUC | 0.88 | 0.91 | +0.03 | ||
| NB vs. LightGBM | AUC | 0.89 | 0.94 | +0.05 | ||
| NB vs. CatBoost | AUC | 0.89 | 0.94 | +0.05 | ||
| NB vs. RF | AUC | 0.89 | 0.93 | +0.04 | ||
| NB vs. MLP | AUC | 0.89 | 0.91 | +0.02 | ||
| LR vs. LightGBM | Accuracy | 0.83 | 0.87 | +0.04 | ||
| LR vs. CatBoost | Accuracy | 0.83 | 0.86 | +0.03 | ||
| LR vs. RF | Accuracy | 0.83 | 0.86 | +0.03 | ||
| LR vs. MLP | Accuracy | 0.83 | 0.84 | +0.01 | ||
| SVM vs. LightGBM | Accuracy | 0.83 | 0.87 | +0.04 | ||
| SVM vs. CatBoost | Accuracy | 0.83 | 0.86 | +0.03 | ||
| SVM vs. RF | Accuracy | 0.83 | 0.86 | +0.03 | ||
| SVM vs. MLP | Accuracy | 0.83 | 0.84 | +0.01 | ||
| NB vs. LightGBM | Accuracy | 0.81 | 0.87 | +0.06 | ||
| NB vs. CatBoost | Accuracy | 0.81 | 0.86 | +0.05 | ||
| NB vs. RF | Accuracy | 0.81 | 0.86 | +0.05 | ||
| NB vs. MLP | Accuracy | 0.81 | 0.84 | +0.03 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2026 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license.
Share and Cite
Bahlool, R.; Hewahi, N.; Elmedany, W. Performance, Fairness, and Explainability in AI-Based Credit Scoring: A Systematic Literature Review. J. Risk Financial Manag. 2026, 19, 104. https://doi.org/10.3390/jrfm19020104
Bahlool R, Hewahi N, Elmedany W. Performance, Fairness, and Explainability in AI-Based Credit Scoring: A Systematic Literature Review. Journal of Risk and Financial Management. 2026; 19(2):104. https://doi.org/10.3390/jrfm19020104
Chicago/Turabian StyleBahlool, Rashed, Nabil Hewahi, and Wael Elmedany. 2026. "Performance, Fairness, and Explainability in AI-Based Credit Scoring: A Systematic Literature Review" Journal of Risk and Financial Management 19, no. 2: 104. https://doi.org/10.3390/jrfm19020104
APA StyleBahlool, R., Hewahi, N., & Elmedany, W. (2026). Performance, Fairness, and Explainability in AI-Based Credit Scoring: A Systematic Literature Review. Journal of Risk and Financial Management, 19(2), 104. https://doi.org/10.3390/jrfm19020104

