Next Article in Journal
Influence of Category and Sex in Game Structure Variables in Young Elite Tennis
Previous Article in Journal
A 13.44-Bit Low-Power SAR ADC for Brain–Computer Interface Applications
Previous Article in Special Issue
Machine Selection for Inventory Tracking with a Continuous Intuitionistic Fuzzy Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Cross-Validation of Predictive Models Used in Credit Default Risk

by
Jose Vicente Alonso
1,* and
Lorenzo Escot
2
1
Department of Applied Mathematics, National University of Distance Education (UNED), 28040 Madrid, Spain
2
Research Institute for Statistics and Data Science, Complutense University of Madrid (UCM), 28040 Madrid, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(10), 5495; https://doi.org/10.3390/app15105495
Submission received: 11 April 2025 / Revised: 6 May 2025 / Accepted: 12 May 2025 / Published: 14 May 2025
(This article belongs to the Special Issue Soft Computing Methods and Applications for Decision Making)

Abstract

:

Featured Application

Predictive model selection when rare events and data scarcity are involved (credit default, fraud detection, or any other event probability model).

Abstract

Model validation is a challenging Machine Learning task, usually more difficult for consumer credit default models because of the availability of small datasets, the modeling of low-frequency events (imbalanced data), and the bias in the explanatory variables induced by the train/test sets split of the validation techniques (covariate shift). While many methodologies have been developed, cross-validation is perhaps the most widely accepted, often being part of the model development process by optimizing the hyperparameters of predictive algorithms. This experimental research focuses on evaluating existing robust cross-validation variants to address the issues of validating credit default models. In addition, some improvements to those methods are proposed and compared with a wide range of validation techniques, including fuzzy methods. To reach solid and practical conclusions, this work limits its scope to logistic regression, as it is the best-practice modeling technique in real-world applications of this context. It is shown that robust cross-validation algorithms lead to more stable estimates, as expected due to the more homogeneous partitions, which have a positive impact on the selection of credit default models. In addition, the enhancements proposed to existing robust techniques lead to improved results when there are data restrictions.

1. Introduction

Predictive models are widely used in different industry sectors as a basis for decision modeling. They are applied in multiple use cases according to the available data and the modeled event, and examples include fraud detection, targeting marketing campaigns, credit risk assessment, disease diagnosis, predictive asset maintenance, etc.
In all cases, it is necessary to choose the best model in terms of the potential predictive performance in new population samples where decisions should be taken in the future. There is a well-known trade-off between the accuracy of the model in past observations and the model performance in new, unknown samples when the model complexity increases; that is, when a larger number of predictive variables are used to train the model. Then, it is key in any industry sector to identify the optimal model complexity according to the data patterns before using the predictive model for decision making.
Many methods have been investigated and proposed to select the combination of explanatory variables of the final model that maximizes predictive performance or minimizes model error. More simplistic methods like hold-out test samples make a deterministic estimation of error bias, while more advanced methods like cross-validation (CV) and bootstrapping (BTS) are stochastic in nature, considering both bias and variance of the model error [1,2,3,4,5,6,7,8]. The less computational requirements, better practicality, and good results of cross-validation techniques have made CV the most popular stochastic validation methodology for model error estimation, yielding a broad range of CV variants and benchmarking studies over the last decades [6,9,10,11].
The main motivation of this research is that, in the particular case of credit default risk models, there is a simultaneous concurrence of several validation difficulties: the predicted events have low frequency and high severity, while the available data sample is usually small and always imbalanced because the number of defaulters is much smaller than non-defaulters due to the nature of the credit events, leading to a covariate-shift when cross-validation techniques are used. This results in less stable model performance estimations that increase the risk of not selecting the best model. There are studies that address these issues separately [12,13,14], and others propose solutions that are not feasible for all credit default explanatory variables [15], while there are very few robust methods that face all the issues simultaneously [9,16,17].
All the above facts support the motivation basis of this work, in a context where a failure in the prediction of an event could be critical, as happens in credit lending, where the economic impact of a bad loan is many times the profit of any performing loan for the financial institution. In fact, it is even worse for the credit applicant, and it could mean the bankruptcy of any person. In addition, the referenced studies in validation methodologies do not tackle this context specifically, and most of the work conducted specifically for credit lending usually proposes and assesses different modeling techniques without focusing on the validation aspect of the problem. Then, we can pose the questions: Are some existing methods more appropriate than others to validate credit default models? In that case, is it possible to improve those methods even further?
This paper conducts experimental research to test the hypothesis that reducing the volatility of the results of the validation process using more robust cross-validation methods, the estimations should outperform any other validation methodology when tackling all the above issues simultaneously for model selection in the consumer credit default risk context. In addition, some enhancements are proposed and tested to better leverage categorical explanatory variables, leading to the best results in data-scarce situations. In addition, a fuzzy approach to cross-validation has also been proposed and added to the benchmarking.
In order to make this research reproducible and applicable, real-life public credit default datasets from Europe, Asia, Oceania, and America were used together with the most commonly accepted methodologies for the probability of credit default, the logistic regression, and the area under the receiver-operating characteristic curve (ROC). Thus, it has been decided to assess a wide range of validation methodologies and their variants while keeping the models, data, and performance metrics bounded to the consumer credit default context. Presumably, this approach will lead to more conclusive results than testing fewer validation algorithms with more types of models and datasets.
In order to test the performance of each validation methodology, a set of models of very different predefined complexities was used. Then, the validation assessment will cover as many potential situations as possible, ranging from models with a poor predictive performance to the most accurate ones.
All model trainings and validations were made using a random sample of the population of each entire dataset. As the validation results can be compared against the real model performance in the unseen population, both aspects of model validation (selection and assessment) could be tested. Nevertheless, this research has been limited to model selection purposes only, as this is a more realistic application of the validation procedures with limited data availability.
Due to the number of validation algorithms used in the comparison, from now on referred to as “validators”, there are validators with similar performance or efficiency, leading to close results at the end of a set of experiments made on a particular dataset. To make sure that there are statistically significant differences between the performance of the validators before ranking them, non-parametric paired tests have been carried out for comparison among all of them.
Among the main conclusions, it is clear that robust cross-validation algorithms need more data processing to calculate the nearest neighbors of the observations, which allows the homogeneity of the train and test datasets to be increased, therefore reducing the variability of the estimations. The main contribution of this research is that, despite the higher computer resource requirements, these validation techniques outperform the rest of the validation methods for the selection of consumer credit default models in data-scarce situations. As a second contribution, the proposed methodological improvements to better leverage the categorical covariates result in even better model selection capabilities. This is potentially applicable in other statistical data modeling contexts where rare events with a large diversity of explanatory variables and data limitations are present.

2. Credit Default Modeling Use Case

Many times, research papers on model validation experiment with several types of models on a variety of datasets try to draw general conclusions rather than address a particular use case or complex situation. For example, there are publications that compare models such as support vector machines, decision trees, nearest neighbors, linear discriminant analysis, etc., on datasets of different contexts (clinical or medical data, environmental, agricultural, financial products or many others) that do not share common specific problems [9,17].
Given the great diversity of existing datasets and predictive models, it is impossible to include all possible cases in a single study, and any attempt to reach universal conclusions could not be feasible, as stated by the “no free lunch” theorem for any optimization problem. In addition, logistic regressions are by far the most common practice for credit default modeling in real-life applications in financial institutions. As a consequence, only logistic regression models are used in this study to fairly compare the validation techniques in the consumer credit default context. Nevertheless, it has to be clear that the validators used in this research can be used for any algorithm able to make predictive discriminations for binary response variables in other contexts, like fraud detection or anti-money laundering, where less interpretable Machine Learning models can be applied.

2.1. Logistic Regression

Credit default models must be explainable and understandable to make sure that they properly reflect the underlying risks of the lending activities, and they can be used together with different business assumptions to make decisions in other potential scenarios. More importantly, the models must be explainable to banking regulators as part of the legal requirements originated by the Basel Accords set by the Basel Committee on Banking Supervision. Therefore, the models should be flexible and easy to understand, like linear regressions, while taking into consideration that predictions must belong to the (0, 1) interval, as happens with the probability of a credit default. Logistic regressions are very convenient for this task as they link a linear regression model with predictions in the (0, 1) domain via a logit function that is analytically easy to handle. These are the reasons why logistic regression is the only predictive modeling technique underlying the credit scorecards used for rating credit applications for more than 50 years.
Logistic regressions can predict with some degree of accuracy the probability of an event in a specific time horizon. When used for credit default modeling, financial institutions can assess which credit applications could be worthy of lending, considering their probability of default (PD).
Logistic regression is a particular case of generalized linear models commonly used when the outcome variable is dichotomous; that is, the response is a binary random variable that represents the occurrence of an event or otherwise non-event.
In the context of consumer credit default modeling, a data sample consists of a certain number of credit customers (or credit lines), each of which can be assigned to a specific group of credit behavior that is defined by the values taken by certain customer and/or product attributes.
Let N be the number of possible credit behavior groups, then each group is characterized by the probability of default of any of its members or PDi where i = 1, 2, … N. Define ni as the number of observations in group i, and define also the random variable Yi as the number of defaults in group i considering all default events as independent among them. Then, each random variable Yi follows a binomial distribution:
P r o b Y i = y i = n i y i P D i y i 1 P D i n i y i
The log-likelihood function of the joint probability distribution of random variables Y1, Y2, … YN is:
l P D 1 ,   ,   P D N ;   y 1 ,   , y N = i = 1 N y i log   P D i + n i y i log   1 P D i + l o g n i y i
Building a probability of default model means representing the dependency between the response or dependent variable PDi of each homogeneous credit default group and some of the customer attributes given by a vector of independent variables X = (X1, X2,… Xp).
Generally, the logit function is used as the link between the probability of default PD and a linear combination of the covariates {Xi}i=1,…p:
L o g   P D 1 P D = β 0 + β 1 × X 1 + β 2 × X 2 +   β p × X p
The left-hand side of this expression or logit function represents the logarithm of odds, and the p coefficients β in the right-hand side are estimated by maximum likelihood, thus maximizing the log-likelihood function.
In order to include categorical explanatory variables in the above formula, it is very common to transform each categorical variable into a set of dummy or binary variables. When a categorical variable takes K different values, a set of K binary variables can be created, setting the value 1 for one specific value and 0 otherwise. This is usually called variable binarization.
Another possibility is to use the weight of evidence (woe) of the categorical variables as the covariates to include in the model formula. To calculate the woe of a given variable X that takes K different values X = {x1, x2, ….xK}, the following formula is used to substitute each of the values:
W O E ( X = x i ) = l n ( q i q e i ) / N E q e i / E   i = 1 , 2 , K
where qi is the number of observations with X = xi; qei is the number of defaulted observations with X = xi; NE is the total number of non-defaulted observations in the sample; and E is the total number of defaulted observations or events.
In the case where there are values xi with only a few observations, it would be convenient to aggregate some of the variable values to obtain a new variable X’ with a smaller number of different categories, but more observations in all or some of them. In addition, if K is large, then model interpretation and implementation become quite complex in practice. Therefore, the original variables are usually aggregated to obtain covariates with no more than 10 or 20 different values in real applications. This practice is usually called variable binning.

2.2. Area Under ROC Curve

Predictive model performance metrics, such as accuracy or misclassification error, have been commonly used in many research papers. However, these are not the most convenient metrics to assess the performance of credit default models, and the area under the receiver operating characteristic curve (area under ROC curve or AUR) should be used instead [18,19,20].
The AUR is suitable for imbalanced data [9] because it considers the trade-off between the benefit of right predictions and the cost of errors. Normally, a binary classifier that increases the number of true positives (captured events) will also increase the number of false positives, and this relationship should be taken into consideration in use cases where the cost of a false negative is very different from the cost of a false positive. This is the reason why AUR has been adopted as the best practice in the field when modeling rare events with high severity, as happens in credit defaults.
To calculate the AUR for a particular PD model over a sample dataset, all the observations in the sample must be ranked first according to the predicted PD, from highest to lowest probability of default. Then, a two-way table of the observed versus predicted defaulting events must be created for each different PD value that is considered as a cut-off value to differentiate defaults from non-defaults. For any specific cut-off PD value, correctly predicted defaults are those defaulted observations with a predicted PD greater than or equal to the cut-off PD, while incorrectly predicted defaults are the non-defaulted observations with a predicted PD also above or equal to the cut-off value.
For each PD value, the corresponding ROC curve point (XROC, YROC) is obtained from its two-way table as the pair “sensitivity” vs. “false alarm rate” in the Y-axis and X-axis, respectively. Sensitivity at a specific cut-off PD is calculated as the correctly predicted defaults divided by all the observed defaults of the sample. The false alarm rate is calculated at a particular cut-off PD value as the number of observations incorrectly predicted as defaults divided by all the observed non-defaulters of the sample.
Then, if there are M different PD values predicted by the model PD1 > PD2 > … > PDM, the AUR can be simply calculated as the area under the discrete ROC curve by:
A U R = i = 1 M 1 Y R O C P D i + 1 Y R O C P D i   2 X R O C P D i + 1 X R O C P D i
There are several ways to approximate the area under the curve; it is not of great importance which is chosen as long as the formula used is always the same when comparing model performances. In any case, the greater the AUR, the better the model performance. AUR equals 1 in a perfect model with no predicted errors, while the AUR of a pure random model is 0.5, where the response variable is independent from the covariates.

3. Model Validation Methodologies

Model validation has several applications according to the purpose of the overall analysis [10,21]. On one hand, the validation of models to select the best among them is called “model selection”. On the other hand, model validation with the aim of estimating what the performance might be on the underlying population is often called “model assessment”.
To carry out an evaluation of models with certain reliability, some requirements must be fulfilled by the data samples available for both the training and testing to be performed during the validation process. Mainly, the data must be highly representative of the underlying population, as otherwise there will always be an indeterminate bias in the estimated expected model performance in the rest of the population.
In real use cases, the lack of representativeness of the data will usually not allow a trustworthy assessment of the predictive models. So, the uncertainty of the evaluation should also be considered [7,13,14], although it is difficult to evaluate [3]. For this reason, model validation is usually more reliable for model selection purposes.
At the same time, the selection of models can be executed within different scopes that have their own particularities. One may be the comparison of previously trained models to pick which one should be used in a production environment with future data samples. Another framework is the model training process itself [10,21] since many iterations that compare eligible intermediate models are performed before coming up with the final model. In this case, resource consumption during the validation process becomes critical, and this workload could make the use of certain sophisticated methodologies that are computationally intensive unfeasible.
In this research, the validation scope is model selection among a predefined set of models with different combinations of explanatory variables, ranging from a few variables to all variables available in the data sample. The more covariates used by the model, the more complex it is.
There is a trade-off between model complexity and model robustness: too many variables will achieve more precision on the training data but less precision when making predictions on new data [22]. This potential loss of model performance is due to over-fitting to the training data in such a way that they do not correctly reflect the true patterns present in the rest of the population.
Therefore, it is critical to consider this circumstance in the validation process and avoid overestimating the performance of the model by finding a balance between model accuracy and model complexity. This will allow us to select the model with the best expected performance in new samples taken from the underlying population. The different approaches to solving this issue will be described next.

3.1. In-Sample Validation

Under the in-sample approach, the validation procedure is carried out with the same data sample used to train the model. There are many goodness-of-fit tests, like the Chi-squared, Kolmogorov–Smirnov, Anderson–Darling, and Cramér–Von Mises, but maximizing any of these metrics will usually lead to more complex models that overfit to the training data, increasing the generalization error in future samples [4]. This is especially harmful when the sample size is small and not fully representative of the underlying population, making it necessary to account for this potential overfitting when selecting a model to make predictions in new datasets.
In this paper, by “in-sample” validation, we refer to those methods where it is the validation metric that directly penalizes model complexity against model performance. One such metric is the Akaike Information Criterion (AIC) that uses the deviance as a measure of the model error. In this metric, the expected prediction error of a probability of default model PD(.) trained with N observations {xi,yi}i=1,…N is estimated by:
A I C = 1 N i = 1 N L y i ,   P D x i + 2 β N
where β is the PD model complexity measured as the number of parameters to fit by maximum likelihood, P D ( x i ) is the default probability predicted for observation x i and y i is the response variable: y i = 1 for defaults and y i = 0 for non-defaults. The deviation error for an observation ( x i , y i ) is calculated as:
L ( y i ,   P D x i ) = 2   l n ( 1 P D x i ) ,   y i = 0 2   l n ( P D x i ) ,    y i = 1
Another in-sample validation metric is the Bayesian Information Criterion (BIC), which penalizes complexity more than AIC does, leading to models with a smaller number of parameters. BIC is calculated as:
B I C = i = 1 N L ( y i ,   P D x i ) + β   l n N
Between AIC and BIC, it is not clear which metric is better for model selection. In general, BIC outperforms AIC when N because AIC usually selects more complex models. On the other hand, for less representative samples, AIC tends to perform better because BIC selects too simplistic models.
In practical modeling applications, AIC and BIC are more commonly used to evaluate the importance of each predictor individually, ranking both of them the independent variables in the same order. For model validation, it is more common to use the approaches described in the following subsections.

3.2. Hold-Out Test Sample

A widely adopted methodology is to separate part of the available data into an independent test set or hold-out test sample (HO sample) that will be used to evaluate the performance of the model after training it with the rest of the available data. This technique can be used with any model performance measure and reduces the bias of the estimation made by the training error [22].
However, when there are few data available, both the bias of the results and their variance (which can become greater than the bias) increase notably, and the use of stochastic methodologies such as bootstrapping (BTS) or repeated cross-validation (CV) becomes necessary [1,12].

3.3. Bootstrapping

This is a general-purpose technique that can be used to estimate a probability distribution for any statistic that can be calculated on a data sample [1,2,4,21,23].
In this technique, an arbitrary number R of subsamples of N observations { Ω 1 , Ω 2 , Ω R } are generated randomly with replacement from the original available dataset Ω of size N. Then, a PD model is trained in each of the R subsamples, and the model performance is calculated in the rest of the available data not included in each corresponding subsample { Ω 1 * , Ω 2 * , Ω R * } .
In this way, if L(Y,PD(X)) is the function chosen as the PD model prediction error for any data point (X,Y), the bootstrap error is calculated as:
E R R b o o t = 1 R j = 1 R E R R Ω j * = 1 R j = 1 R 1 N j * ( X , Y ) Ω j L Y , P D j ( X )
where N j * is the number of observations in the dataset Ω j * that only includes the observations (X,Y) not in the subsample Ω j used to train the model P D j ( X ) .
If the model error is taken as 1-AUR—that is, the model performance is measured via the AUR—then it can be written:
A U R b o o t = 1 R j = 1 R A U R P D j ; Ω j *
where A U R P D j ; Ω j * is the AUR calculated in the Ω j * test set using the PDj model trained in subsample Ωj.
There is an enhanced bootstrap variant called “0.632 bootstrap” aimed at reducing the bias in the bootstrap estimation due to the fact that, on average, around one-third of the observations are not used to train each PDj model. This could lead to poor model performance in situations of limited sample sizes. This variant is a weighted average of the performance AURtrain of the PD model trained in the entire sample Ω and the bootstrap estimation:
A U R 0.632 = 0.368   A U R t r a i n + 0.632   A U R b o o t
An additional refinement is the variant known as the “0.632+ bootstrap” that modifies the weights of the above average to decrease the impact of the training performance when the sample size is small and there is potential overfitting to the training data that could underestimate the real model error [23]. It is calculated as:
A U R 0.632 + = ( 1 ω )   A U R t r a i n + ω   A U R b o o t
where ω = 0.632/(1 − 0.368 ρ) and ρ is the ratio of overfitting defined in terms of the “non-information performance” φ that is the model performance on a population where the response variable Y is independent from the covariates:
ρ = ( A U R b o o t A U R t r a i n ) / ( φ A U R t r a i n )

3.4. Cross-Validation

Usually, the hold-out test sample technique is used to validate a model when data in excess are available, taking out a large enough independent test set separated from the training data to avoid performance overestimation. For those situations in which there is data scarcity, a smaller validation set should be used to allow for enough training data. However, this can add significant bias to the validation results due to the lower representativeness of the test set. The cross-validation method (CV) was specifically designed to tackle this situation.
The basis of cross-validation is to use all the N observations in the sample Ω to validate the classification model, calculating its performance on all the available data. Then, the k-fold cross-validation method performs a random partition of the data into k subsets of equal size { Ω 1 , Ω 2 , Ω k } as shown in Figure 1, and uses each subset Ωj of N/k elements (test set) to validate a model that has been previously trained with the remaining N(1 − 1/k) observations of Ω\Ωj (train set).
Thus, all the data are used once each time, either to train the model or to validate it. Finally, the average of the performances obtained in the k test sets is used as an estimation of the model’s performance:
A U R C V ( k ) = 1 k j = 1 k A U R P D j ; Ω j
where A U R P D j ; Ω j is the AUR calculated in the Ω j dataset using the PDj model trained in Ω\Ωj.
The k-fold cross-validation technique can be applied iteratively by repeating the process r times. That is, once an estimate of the performance has been obtained by cross-validation, a different partition of the available data can be made to estimate again the performance of the predictive model in the same way. Finally, all the results given by the r repetitions are averaged to obtain a more stable final estimate of the predictive model performance:
A U R C V ( k , r ) = 1 r   i = 1 r A U R C V r ( k )
where A U R C V r ( k ) is A U R C V ( k ) calculated for a particular k-fold partition corresponding to the repetition r.
Given the flexibility to choose the number of partition segments, the number of repetitions, and the method used for partitioning, a significant direct impact of this parameter setting is expected on the CV results [24].
According to several studies, some of the most interesting results obtained may be the following:
  • The number of partitions most commonly accepted as the best choice is k = 10 or even k = 5 when compared to other typical values such as k = 2 or k = N [2,25], although this should not be taken as a rigorous rule of universal application [10].
  • The iterative repetition of the CV estimates converges as the number of repetitions increases [9] and significantly improves the validation results by reducing their variance [5,8,10,26,27].
  • The variance of the CV estimator decreases as the size of the data sample increases [8,26].
  • Usually, there is not enough information about the population and the bias of the data samples to calibrate the validation result to the true model performance. For this reason, cross-validation could overestimate or underestimate the actual performance of the predictive model, making it a more appropriate technique for model selection than for model assessment [7,13,14].
The versatility of the CV methodology allows for continuous research on how to leverage this technique to improve the predictions of increasingly sophisticated artificial intelligence algorithms that are applicable in many specialized fields, including credit default risk.
As will be explained in Section 4, many CV variants have been developed from the original methodology described above. Possibly, stratified cross-validation (SCV) is the most popular and commonly used variant among all of them due to its simplicity and potential benefits. SCV is a k-fold cross-validation where the data are partitioned with a random sampling stratified on the target variable [2]. This methodological improvement makes the test sets (and the training sets) more homogeneous among themselves, reducing the volatility of the model performance estimates.
The design of the SCV technique avoids biased distributions of the target variable in the train and test sets, and it could therefore be useful in cases of imbalanced data. Although this technique does not address the problem of covariate shift, it will provide a good basis for comparison against the cross-validation enhancement proposed in Section 5.

3.5. Other Approaches

Some authors have highlighted the limitations of the above stochastic methodologies in situations where there are few data available and suggest giving the results in the form of confidence intervals that conservatively take into account the uncertainty of the estimates [13].
Other researchers prefer to increase the size or the heterogeneity of the samples used [14], but this is not possible in real situations where data availability is limited. Then, the only way to make a better model selection is to optimize the validation technique regarding the issues present in each particular use case, as discussed in the next section.

4. The Partitioning Problem and Cross-Validation Variants

Over the years, various studies have identified cross-validation as generally more convenient than bootstrapping because of both the stability of the results and the computational cost [2,5,6,25]. Numerous studies have been carried out to analyze the variance of cross-validation results using different methods [3,8,26,28].
Given that, this work focuses on experimenting with different cross-validation variants to assess whether there are particular techniques that could present extra benefits for validating credit default models regarding the way they were designed.
Cross-validation methodologies usually differ among them in some parameters that define how the original available data are split into validation and training sets, e.g., how many splits must be performed and how these splits have to be chosen. Consequently, different CV methods do not always select the same set of explanatory variables or even the same number of them for the best predictive model [2]. So it is very difficult or even impossible to select the best CV method and parameterization for any possible dataset because the available dataset size, structure, underlying patterns, and response variable will affect the performance of the chosen CV method.
The researcher must make subjective decisions on the set of parameters to use when applying CV methods to their particular case study. This is the reason why much of the research carried out in the past focused on analyzing the impact of the different parameter settings and CV variants on the results. Ideally, the splitting of the available data should be performed to maximize the representativeness of the test and training sets, avoiding the introduction of additional artificial bias between them.
The main parameters of the k-fold cross-validation are the number of repetitions to perform, the number of segments of the partitions, and the specific partitioning technique. Using different values for the number of test sets and the number of repetitions, several known variants of the CV method can be obtained, such as 5-fold cross-validation repeated twice (5 × 2 CV) or 10-fold cross-validation without repetition (10-fold CV). Other procedural aspects that can be modified are the aggregation level of the results and the chaining of cross-validations.
Regardless of the first two above-mentioned main parameters, the partition technique applied to split the observations into k subsets or segments can be a problem when the data distribution is highly skewed (either in the response variable or in the explanatory variables). In these cases, data partitioning can lead to a shift in the data distribution of the training set compared to the data distribution of the test set. This issue is already known as dataset shift.
This dataset shift may cause additional volatility in the CV results, thus the convergence of the validation algorithm will require more repetitions and, for example, the true predictive model performance may be underestimated if there is a large bias between the training set used to calibrate the model and the test set used to evaluate its performance.
One of the main sources of dataset shift is the availability of a dataset not large enough to have equally representative train and test sets. Another is the existence of anomalous observations that could be present in the test set but not in the training set (or vice versa). Additionally, imbalanced data will favor different proportions of events between the training and test sets.
These three issues are often related since having only a few available data points makes it more likely that any partitioning sampling will result in significant differences between the train and test distributions of any variable. However, many of the studies carried out to date usually analyze each problem separately, while in the context of consumer credit default models, all the issues are simultaneously present: few data available, low frequency events, and very diverse debtors’ attributes due to differences in sociodemographic characteristics, financial information, and delinquency history.
In the following section, several techniques developed during the last decades are discussed. The goal is to identify beforehand if there are CV variants more convenient to be used in a benchmarking study of credit default model validation.

4.1. Monte-Carlo Cross-Validation

Monte-Carlo cross-validation methodology (MC-CV) is not a true cross-validation technique in which all observations are finally used to validate the predictive model. This method basically consists of multiple executions of hold-out test sample validations using a different test sample each time that is drawn randomly without replacement from the available data [11,15].
Some variants could be derived from this basic idea, like the stratified Monte-Carlo cross-validation (SMC-CV), where the hold-out test samples are stratified in the response variable. Going further, a couple of fuzzy variants are proposed here for experimentation by simply considering the size of the hold-out test sample as a random variable distributed uniformly in a predefined interval. Then, a fuzzy Monte-Carlo cross-validation (FMC-CV) and a fuzzy stratified Monte-Carlo cross-validation (FSMC-CV) will be additionally implemented and tested in the present analysis.
The disadvantage of this methodology is that not all the data will be finally used to train and test the model, and not all the training and testing observations will be used to the same extent. The number of times that any data are used for training or testing is left to chance, so it is expected that the results obtained will have some induced volatility. Then, in order to include this technique in the comparison, a larger number of repetitions must be performed compared to true cross-validation methods.

4.2. Leave-One-Out Cross-Validation

One of the most studied techniques in the field of cross-validation is the leave-one-out cross-validation (LOOCV). Fundamentally, this is a k-fold cross-validation where k is equal to the number of observations (k = N). That is, the test sets always contain a single data item [11,12,15,24]. Using LOOCV, there is no need to choose either the value of k (since it is determined by the number of observations of the data sample) or the value of r, since only one repetition is possible.
To combine LOOCV with the AUR metric, it is necessary to perform the cross-validation algorithm in a slightly different way: first, the model predictions of all the individual test observations are computed, and then only one single AUR figure is calculated using all the predicted values. In this case, there is no way to average the performance measure over the test sets.
An important disadvantage is the large computational workload required since N models will have to be trained using N − 1 data. This computational complexity is not compensated for by an improvement in the estimations since unstable results are expected due to the low representativeness of the test sets and because this variance cannot be reduced by repetition [12]. Additionally, the unfeasibility of stratified sampling usually yields a high estimation bias when applied to small data samples. For these reasons, 10-fold cross-validation usually outperforms LOOCV, but this method will also be included as another validator in this study for verification.

4.3. Distribution Balanced Stratified Cross-Validation

An interesting method to cope with the partitioning problem is the Distribution Balanced Stratified Cross-Validation (DB-SCV). This method tries to reduce the deviation in both the target variable and the covariates by making each segment of the partition as intrinsically heterogeneous as possible, and as homogeneous as possible with respect to the rest of the k − 1 segments. That is, it distributes the classes of the response variable as uniformly as possible and, for each of these response classes, it keeps the distribution of the covariates as similar as possible between the test and the train set of each k-fold split.
There are two versions of this validator published by their authors. In the first version [16], from now on referred to as DB-SCV1, for each class, all distances are calculated from an artificial data point taken as a constant reference that is built with the minimum value of each continuous variable and one particular value of each of the categorical variables.
In the second version of this validator [9], referred to as DB-SCV2 from now on, for each class, all the distances are calculated from a reference observation that is selected at random from the data sample at the beginning of the process. The rest of the validation procedure is the same.

4.4. Distribution Optimally Balanced Stratified Cross-Validation

The Distribution Optimally Balanced Stratified Cross-Validation (DOB-SCV) is an enhanced alternative to the DB-SCV. In this case, there is not only one single reference observation for each class. Instead, a new observation of the class is selected at random each time that k observations are going to be evenly distributed through the k segments of the partition. The k − 1 nearest observations to the reference one in terms of the model covariates are chosen each time [9,17].
Since there are only minor differences between DOB-SCV, DB-SCV1, and DB-SCV2, it will be necessary to evaluate with experiments whether there are statistically significant advantages to any of them when selecting the best models in the current experimental context where, according to our knowledge, they have not been compared yet.

4.5. Representative Splitting Cross-Validation

Representative Splitting Cross-Validation (RS-SCV) is another CV variant whose goal is to carry out the cross-validation partitioning in a way that makes both the train and test sets as representative as possible of the available sample data. Thus, sample data should be distributed as evenly as possible in both sets [15].
It has therefore the same aim as the DB-SCV and DOB-SCV techniques, but achieved with an opposite implementation approach: instead of using a clustering algorithm that identifies the nearest observations to distribute them in different segments of each partition, in RS-SCV, the furthest observations are identified to include them in the same segment.
The RS-SCV method uses the DUPLEX algorithm to divide the available dataset into two subsets of equal size, as uniformly as possible between them, identifying the data pairs that are furthest away from each other by means of the Euclidean distance and placing them alternately in the first and the second subset.
There are two main disadvantages to this technique. One is that the DUPLEX algorithm cannot use categorical variables, so this type of variables is not leveraged when distributing the observations uniformly.
The second issue is that applying the DUPLEX algorithm successively yields k-folds that are powers of two: 2, 4, 8, 16, 32…. This makes it more difficult to compare this validation method against the commonly used k-fold values of 5 and 10. For these situations, the solution adopted has been to apply the RS-SCV for k = 4 and 8, respectively, and for other k values using the larger power of two smaller than the original k value as the k-fold parameter. After this, the observations are redistributed evenly among the original number of 5 or 10 folds, keeping the same proportion of observations belonging to the four or eight segments previously made by the DUPLEX algorithm.
This additional random behavior of redistributing the observations when k is not a power of two allows to make several repetitions with different results for averaging. For case k = 2 or any other power of 2, there will be no repetitions as the partitioning technique is deterministic.

4.6. Stochastic Cross-Validation

This technique consists of an ordinary cross-validation with repetition, where a random value for the k parameter is chosen for each repetition [11]. The random distributions originally proposed to select a k value each time are the uniform and the normal distributions.
The result of the estimations could be seen as an average over a mixture of cross-validations with different k-folds. This could be interpreted as a fuzzy approach to cross-validation. The technique was designed to avoid the selection of a particular value for the k parameter because it is usually impossible to know beforehand which one will perform better.

4.7. Fuzzy Cross-Validation

To go deeper into the fuzzy aspect of the stochastic cross-validation previously described, a new fuzzy cross-validation method is proposed in this study to be compared with the rest of the validation methodologies. The aim is to use a truly fuzzy value for the k parameter in each repetition instead of using a particular k value in each repetition.
The basic idea is to make irregular partitions where the size of each segment is fuzzy, meaning that the data sample will be divided into subsets of different numbers of observations. Therefore, there is no proper k-fold parameter, although the number of resulting segments of the irregular partition could be seen as a corresponding fuzzy k parameter.
To choose the size of a particular segment of the partition, a normal distribution or uniform distribution can be used to select a hypothetical k value among a set of possible values that will be used to decide the number of observations of the segment. Then, the observations to be assigned to this segment will be taken randomly from the remaining observations of the data sample not yet belonging to any segment.
The process of selecting the size of the next segment and assigning its observations will continue until the last choice of the segment size is equal to or larger than the remaining observations not yet assigned to any particular segment.
Then, no k parameter needs to be predefined as the number of resulting segments of the partition in each repetition is a consequence of this random process of selecting the size of each segment sequentially.
A simple pseudo-code of this fuzzy cross-validation is given just to state this validation procedure more clearly (Algorithm 1).
Algorithm 1: Fuzzy cross-validation (Source: own elaboration)
For iter = 1 to Total number of repetitions
  Current fold number = 0
  Number of observations not assigned to any fold = Total number of observations
  While Number of observations not assigned to any fold > 0
    Random selection of a k value from a Normal or Uniform distribution
    Test set size = floor(Total number of observations/k)
    If Number of observations not assigned to any fold < Test set size then
      Assign the remaining observations to Current fold number
    Else
      Current fold number = Current fold number + 1
      Select randomly a test set from the observations without fold assigned
      Assign the Current fold number to the observations of the test set
    End if
  End while
  Total number of folds = Current fold number
  For i = 1 to Total number of folds
    Train a model in the set of observations not belonging to fold number i
    MP(i) = compute the model performance with the observations of fold number i
  End for i
  ModelPerform(iter) = average of the performances MP(i) on the Total number of folds.
End for iter
Final model performance estimation by averaging all the estimations ModelPerform(iter) on the Total number of repetitions.

4.8. Maximally Shifted Stratified Cross-Validation

The aim of Maximally Shifted SCV (MS-SCV) is the opposite of making homogeneous partitions, as it tries to maximize the covariate shift, generating the most different possible test and training sets [9]. This method can be seen as the opposite of DB-SCV2, because using the same reference observation per class, MS-SCV assigns the same segment of the partition to the k nearest neighbors instead of distributing them evenly across all k folds. The observation assignment to the same segment continues until the fold is full of observations of that class, and the process continues with the next fold to be filled.
The interest in implementing and testing this method is to verify that it performs very badly for model selection in a credit default context, according to the arguments given in Section 5. MS-SCV could serve as an additional verification of the convenience of homogenous partitioning if its performance as a validator in the current context is finally confirmed as very poor.

4.9. Other Cross-Validation Variants

The flexibility of cross-validation methodology has allowed the introduction of even more variants than the ones discussed above; however, they were finally discarded as potentially useful validators in this study for different reasons. Some of them will be briefly mentioned here.
Leave-p-out cross-validation (LpO-CV) is an exhaustive method where all possible test sets of p observations are used as hold-out test samples. Given the huge number of possible test sets of p elements taken from a population of N elements, where N has a minimum order of magnitude of hundreds or thousands, this variant has been discarded as it is computationally infeasible in real-life applications. Only the special case of p = 1 is considered, as it is the LOOCV technique.
Relevant to mention is Importance Weighted Cross-Validation (IWCV), which was especially designed to address the problem of mitigating the a priori existence of a covariate shift rather than trying to avoid such an induced shift during the data partition [29]. IWCV measures the importance of the covariate deviation and considers it when calculating the model performance to decrease the bias of the estimations as much as possible. For this reason, this technique does not try to reduce the existing deviation in the covariates but rather to correct the deviation afterwards in the calculations. Therefore, this is a complementary technique that could be combined with other CV variants and is, therefore, not a benchmark for direct comparison against any other.
Nested cross-validation is a methodology that chains an additional cross-validation when building the k models within the training datasets. When a regular k-fold cross-validation is carried out to perform a model validation, each of the k times a model is trained in one of the corresponding train sets, another cross-validation with a different k parameter (usually lower) is conducted to make that model training. Then, this additional k-fold cross-validation has the objective of building a better model by means of selecting the most convenient model complexity to avoid overfitting [27]. In the present research, model complexity is already predefined in a set of given models to be compared; therefore, nested cross-validation cannot be used in this experimental design.

4.10. Summary of Validation Methods

To provide a general picture of the main validation techniques previously explained, Figure 2 shows the more relevant features of those methodologies in a hierarchical classification.

5. Robust Cross-Validation Methods

As stated in the introductory section, the objective of this work is to identify potentially better validation methodologies to be used for credit default data and models. As discussed previously, cross-validation is the most accepted validation method nowadays, so the research methodology is to identify potentially robust cross-validation methods, those with more stable expected estimations, to compare them with the rest of the validation methodologies for model selection purposes. Going further in the research, another goal is to identify new variants of the robust methods by improving the strategy used to reduce the volatility of the estimations.
Among the different methodological variants of cross-validation presented in the previous section, only a few can be considered robust methods because they were designed to make more homogeneous partitioning, potentially resulting in more stable and less volatile estimations. This could be beneficial when validating credit default models, where data scarcity, imbalanced data, and covariate shift are generally present, because these three issues increase the volatility of the results.
Then, robust cross-validation methodologies of interest to be tested and analyzed in this research are DB-SCV and DOB-SCV. Nevertheless, they have some weaknesses that are going to be addressed here to potentially improve their performance for model selection.
The main problem of these methods for validating credit default models is that they rely on Euclidean distances to distribute the sample observations more uniformly. This is a problem when dealing with categorical variables because a Euclidean distance cannot be calculated, and consumer credit models usually have several categorical explanatory variables.
Euclidean distance in several dimensions can only be calculated properly for quantitative attributes that are comparable in their magnitudes. Both DB-SCV and DOB-SCV deal with this issue by using a “binarized” distance for the categorical variables between two observations. This distance is 0 when both observations have the same category and 1 otherwise, considering all possible categories as equidistant from each other. If there are pcateg categorical covariates, the binarized distance between two observations X and Z is calculated similarly to a Euclidean distance as:
D c a t e g X , Z = i = 1 p c a t e g δ i ( X , Z )   :   δ i X , Z = 1   i f     X i Z i ;   δ i X , Z = 0   i f     X i = Z i
To calculate the Euclidean distance between observations using continuous covariates, the values of the covariates must be normalized first to make them comparable. This normalization is made by considering the observed range of values in the sample. Then, the Euclidean distance between two observations X and Z when they have pcont continuous covariates is calculated as:
D c o n t X ,   Z = i = 1 p c o n t X i min   ( X i ) max X i min   ( X i ) Z i min   ( Z i ) max Z i min   ( Z i ) 2
In case the two observations X and Z have both continuous and categorical variables, the distance between them is calculated following the expression:
D X , Z = D c o n t X , Z 2 + D c a t e g ( X , Z ) 2
In order to enhance the performance of these validators, it is proposed here to use the weight of evidence of the categorical variables to calculate a Euclidean distance for Dcateg. This could lead to better assessing the similarity or the distance between observations when categorical variables are involved, making them more fairly comparable in terms of the same quantitative magnitude.
This does not mean that the woe variables must be used as the model covariates. The covariates in the model could be the woe variables, the binarized dummy variables, or any other transformed variable. However, in order to enhance the process of finding the nearest neighbors, the woe of each categorical covariate will be used to calculate the distance Dcateg. Then, the same Euler’s formula used for Dcont will be used for Dcateg just by means of directly considering the woe of the categorical covariates as if they were continuous variables. This requires the normalization of the woe variables between 0 and 1 using their maximum and minimum values.
This enhancement can be applied to the original DOB-SCV validation methodology, using distances calculated with the weight of evidence for all the categorical covariates in the model being validated. This validator will be referred to as DOB-WOE for the rest of this document, while the same enhancement for DB-SCV will be labelled as DB-WOE2, as it was made for the more recent second version of the DB-SCV algorithm.
In summary, the aim of these robust k-fold cross-validation methods and their proposed enhancements is to make the training and test sets created during the partitioning as similar as possible. Estimating the performance of the model on more homogeneous partitions will yield more robust estimates with less volatility, and, therefore, quicker convergence to a final result when repetitions are made.
It is important to note that these robust methods are going to be tested for model selection, not model assessment, because limiting the possible partitions to a smaller subspace of more homogeneous partitions could introduce additional bias in the validation estimates, probably overestimating the model performance. This is in agreement with the “no free lunch” theorem. An enhancement or optimization made in an algorithm for a specific goal could worsen the performance of the same algorithm for a different goal. In this regard, model selection can be seen as an optimization task that aims to find the best model in a search space of alternative models with different complexities.

6. Experimental Framework

The experiments held in this study try to replicate a real-life validation procedure for model selection in the sense that several predefined models validated in a data sample from a larger population are ranked by each validation algorithm and the resulting rankings are verified using the true performance of the models in the unseen population not used to validate the models. This experiment is conducted multiple times using different data samples taken at random from several populations to perform statistical assessments of the results.
In each experiment, there will be validators giving a model ranking more similar to the true model performance ranking. Then, the validators could also be ranked in each experiment based on their discrimination capability to differentiate between better and worse models. The metric chosen to measure this ability to distinguish good models from bad ones was the Kendall Tau, as will be explained later on.

6.1. Credit Default Datasets

The set of experiments was made with several consumer credit datasets from developed countries of distinct geographical areas to test the validation methodologies with potentially different credit default patterns. These datasets have been used in previous research and are publicly available to make the conclusions comparable with any other future or past studies.
Table 1 shows the datasets used, their size, and bad credit rates, as well as the source repository where they can be found. The event rates of these credit datasets go from 14% to 44%. These high rates are not normally found in credit products; for example, typical rates for credit cards usually vary between 1% and 7% [30], while rates higher than 10% were observed only during the “credit-crunch” of 2007–2008 for some products. However, it is a common practice in predictive modeling of credit defaults to make an oversampling of delinquency events that allows reaching ratios of bad/good credits from 1:1 to 1:5, which facilitates the modeling procedure. So, event rates after oversampling generally vary between 17% and 50%. In this sense, the range of event rates of the credit data used in the experiments aligns very well with those over-sampling conditions given in practice.
We have to make sure that the three issues faced in the validation of consumer credit default models are also present in the current experiments. The imbalanced data issue is already taken into consideration due to the low default rate of most of the datasets, usually making it more difficult to both build and validate predictive models [9,17,35].
To include data scarcity in the experiments, samples of 500 records have been used for the German and Lending Club datasets. The Australian dataset has only 690 records in total, so a sample size of 230 observations was used, representing one-third of the available data. For the Taiwan dataset, 500 observations are probably not representative enough to make modeling exercises, so samples of 10% of the population were used to have sufficient data to build predictive models, but still small enough to be considered as a potential data shortage situation.
The explanatory variables included in the datasets mainly concern socio-demographic, financial, and professional information of the applicant, as well as characteristics related to the credit application (purpose, amount, duration…).
Due to the large number of categorical levels of some variables of the Lending Club 2009 file, the original categorical explanatory variables of this dataset were aggregated in 10 bins when there were more than 10 levels, merging the bins with more similar weight of evidence according to Section 2.1 and using the resulting woe of the merged bin as the new variable value.
Finally, the covariate shift issue will be a result of the data scarcity because the test/train data splits will significantly decrease the representativeness of these datasets. This will lead to partitions with biased covariates if almost all the observations of the experimental sample are needed to fully capture the data patterns of the underlying population.
This means that the hold-out test samples and the partitioning of cross-validation techniques could induce significant deviations between the distributions of the test and training sets, which can cause undesired, unstable results. This situation should be better managed by robust partitioning techniques, which is the hypothesis to be tested in this work.

6.2. Model Set

In consumer credit approval, logistic regressions are the market standard for probability of default models. This makes it possible to reach practical, applicable conclusions without the need to test other types of models, such as neural networks, random forest, SVM, or others not used in this business context.
A set of 63 models was created systematically from different subsets of explanatory variables to diversify the model complexity and the validation results. This allowed us to test the validators for model selection in a wide range of model performances, typically from AURs less than 0.6 to more than 0.8. In addition, as the number of models created was quite large, most of the time there were models with similar performance results, which made the model selection more difficult, resulting in different model performance rankings for similar validation algorithms.
To automate the model set generation, first, an ordered list of all variables was created for each credit dataset. The list order was determined by decreasing variable importance measured with the Akaike Information Criterion AIC. First, the list was divided into three equal parts (top, middle, bottom) or sub-lists; then, four subsets were created from each one of the three sub-lists. These subsets contained all variables, none of them, the variables in even positions of the sub-list, and the variables in odd positions. Taking all possible combinations formed by one subset from each sub-list, a total of 43 − 1 = 63 variable combinations were made, excluding the empty set.
In this way, the models trained in each of the experiments covered a wide range of predictive performance and model complexity, going from a few of the worst explanatory variables to all the variables available in each credit datasets.
For each experiment, each combination of covariates was used on a logistic regression trained with the whole available sample of each credit dataset using the “glm” function of the “Stats” package in R version 4.4.1. Details and syntax of this function can be found on the official R documentation website [36]. In fact, since the underlying population used for sampling was known, it was possible to find out the true performance of each of the models trained by applying them to the remaining observations not included in each sample.
Table A1 in Appendix A shows the expected true AUR performance of each logistic regression considered, calculated by averaging over 150 experiments of each credit dataset, and their corresponding standard deviation. These details give an overview of the large variety of model performances tested in this study, where the averaged true AUR goes from 0.50 to 0.93, allowing us to challenge the model selection capabilities of the validation algorithms in many potential scenarios, thus providing strong support to the conclusions.
The explanatory variables corresponding to each model are shown in Table A1 as a combination of the subsets that correspond to the top, middle, and bottom sub-lists of the variable list ordered by increasing AIC. The subsets of each third of the list have been labelled as follows: “A” = all variables of the corresponding sub-list; “N” = none of the variables; “O” = variables in odd positions of the sub-list; and “E” = variables in even positions.

6.3. Validation Algorithms

A total of 65 validation algorithms have been tested and compared with each other. These validators can be classified into groups according to the methodology used. The acronyms of the validators belonging to each validation methodology are explained to be able to identify their corresponding results, given in Section 7 and Appendix A.

6.3.1. In-Sample Validation

This group is identified as “InSample” and includes the AIC and BIC validation techniques identified as “InSample/AIC” and “InSample/BIC”, respectively.

6.3.2. Hold-Out Test Sample Validation

This is the “HO” validation group, where the acronyms used are MCCV (Monte-Carlo Cross-Validation), SMCCV (stratified MCCV), FMCCV (fuzzy MCCV), and FSMCCV (fuzzy stratified MCCV). In addition to the group prefix, two suffixes were added to these acronyms, indicating the percentage of the data used for testing and the number of repetitions made. For example, “HO:SMCCV/0.2/1000” stands for Stratified Monte-Carlo Cross-Validation with 20% test data and 1000 simulations. For fuzzy methods, the suffix relative to the percentage of test data is 0 because there is no specific test set size, as a random size was taken for each repetition from a uniform distribution between 5% and 50%.

6.3.3. Bootstrap Validation

The bootstrapping validators (BTS) implemented are identified as “boot” (original bootstrap), “sboot” (stratified bootstrap), “b632” (0.632 bootstrap), “sb632” (stratified 0.632 bootstrap), “b632+” (0.632+ bootstrap), and “sb632+” (stratified 0.632+ bootstrap). There is no choice of the test size in these techniques, so only the last suffix with the number of simulations has meaning. Then, “BTS:b632+/0/1000” refers to the 0.632+ bootstrap using 1000 simulations.

6.3.4. Cross-Validation

This is the largest validation group (CV) this study focused on, mainly due to the usual superior performance of this methodology compared to the previous ones, as will be confirmed by the results of the next section, and is aligned with previous studies in more generic contexts.
There are 48 different variants implemented, divided into several subgroups according to the methodology used in the partitioning. In all cases, 50 repetitions [2,27] were conducted to make sure that a stable enough estimate was reached by all the cross-validation techniques.
The original k-fold cross-validation and stratified cross-validation techniques are identified by “CV:kfold/k/r” and “CV:SCV/k/r”, respectively, where the “k” suffix indicates the value of the k-fold parameter and “r” is the number of repetitions made (r = 50). Exceptions are the LOOCV and the RSCV with k = 2 “CV:RSCV/2”, because no repetitions are possible.
Other CV methods have been named in the same way. The representative splitting cross-validation is identified as “CV:RSCV/k/r”; distribution optimally balanced stratified cross-validation as “CV:DOB-SCV/k/r”; the proposed enhancement to this method as “CV:DOB-WOE/k/r”; the distribution balanced cross-validation with algorithm versions 1 and 2 as “CV:DB-SCV1/k/r” and “CV:DB-SCV2/k/r”; and the corresponding enhancement to the latter as “CV:DB-WOE2/k/r”.
Similarly, the acronym chosen for maximally shifted cross-validation was “CV:MS-SCV/k/r”. This partitioning methodology is more heterogeneous by definition and should lead to more volatile results in contrast with the robust methods that are going to be challenged. The MS-SCV has been included in this study to check if its performance is the opposite of the most homogenous partitioning methods in the same situation.
Stochastic cross-validation using a discrete uniform distribution over the integers between 2 and 20 to choose the k-fold value for each repetition has been labelled as “CV:RndCV/0/r”.
Finally, the proposed fuzzy cross-validation methods have been labelled as “CV:FUkfold/0/r” for the uniform distribution version where the integer k-fold parameter was chosen randomly from 2 to 20, while the same method using a normal distribution N(10, 5) capped to minimum and maximum values of 2 and 20, respectively, has been identified as “CV:FNkfold/0/r”.
Except for the stochastic and fuzzy methods, where the k parameter is not specified deterministically, the values tested for the k-fold parameter were k = 2, 5, 10, 15, and 20. The unusual higher values of k = 15 and 20 have been added to the more common values k = 2, 5, and 10 of other researches [2,6,12,25,26] to check if the robust partitioning algorithms decrease the variability of smaller test sets enough to leverage the larger size of the training sets used for calibrating the models.

6.4. Validator Performance Assessment and Comparison

To statistically determine whether a particular validator can be considered better than others at discriminating between good and bad models, the validation algorithms will be considered as if they were classification algorithms whose task is to order a set of predictive models in a ranking according to their estimated performance.
Then, it is necessary to assess how good the ranking provided by a particular validator is compared to the true model ranking that can be built based on the model performance in the population not used to train the models.
To achieve this, the Kendall Tao ranking distance can be used, which counts the number of discrepancies between two ordered lists or rankings. This distance metric can be defined as the number of discordant pairs between two ordered lists L1 and L2, that is, the number of pairs in which the order given by L1 to the two elements of each pair is different from the order given by L2 to the same elements.
Let R1(i) be the rank of element i in list L1 and R2(i) the rank of element i in list L2; then the Kendall Tau distance KT(L1, L2) between the two lists can be calculated using the following expression:
K T L 1 , L 2 = | { i , j : i < j , R 1 i < R 1 j     R 2 i > R 2 j     [ R 1 i < R 1 j     R 2 i > R 2 ( j ) ] } |
The distance is zero in cases where the two lists have the same order, while the maximum distance is achieved when one list has the reverse ordering of the other. This maximum value is m(m − 1)/2, where m is the number of elements of the lists. Then, a normalized Kendall Tau distance in the interval [0, 1] can be obtained by dividing KT by m(m − 1)/2.
The normalized Kendall Tau distance has been used as a measure of the error that each validator makes when ranking the models by their estimated AUR in comparison with the ranking based on the true AUR of the models. From now on, this metric will be referred to as the “selection error” of the validation algorithm and will be used to assess the validator’s performance for model selection purposes.
This approach allows the assessment of each validator’s performance using a set of models with a wide range of different performances, since in practical applications, there will be situations where the models to be selected will perform well, and others where the AUR of the models will be lower.
In this study, a validator will be considered worse than others for a specific credit default dataset if its selection error is statistically larger when measured on a sufficient number of samples taken randomly from the dataset.
One may think about comparing the selection error of each validator averaged over many samples, but there is no easy way to assess the variance of the selection error to make a proper comparison. Then, the selection error of all the validators needs to be compared in each of the experimental samples one by one, which allows to perform a paired test of the validator performances. Therefore, validation results will differ due to the particular validation method used, avoiding any variability coming from the sample randomness.
To this end, the non-parametric paired Wilcoxon one-sided rank test was used to compare the selection errors with a confidence level of 95% using the R function wilcox.test() over 150 experiments performed in each of the default credit datasets considered. In addition, an overall paired comparison using all the 600 samples was performed, applying a single Wilcoxon test to all the samples of all datasets.

6.5. Overall Experimental Research Methodology

The experimental elements previously described through this Section 6 were performed in subsequent iterations according to the following list of steps that summarizes the overall research methodology:
1. For each credit dataset, the individual importance of each variable is determined using the AIC and generating a variable ranking.
2. Using the variable importance ranking, 63 different combinations of variables were generated in a systematic way.
3. A total of 150 samples of limited size were taken without replacement from each dataset to carry out different experiments in each dataset population.
4. For each experiment, a logistic regression model was trained for each variable combination (model complexity) with all available sample data, and the true model performance was calculated using the remaining population not used to train the model. Therefore, a true model performance ranking could be established in each experiment.
5. Each model was validated by 65 different validation algorithms (validators), and a model ranking was generated for each validator based on its model performance estimations.
6. The model selection error of each validator in each sample was calculated as the normalized Kendall Tau of the comparison of each validator’s model ranking and the true model ranking of the corresponding experiment.
7. A validator performance ranking for model selection was determined for each credit dataset based on the selection error averaged in all the experiments, although non-parametric paired tests were used to determine when the differences in the selection error between validators were statistically significant.
8. Finally, an overall ranking was generated in the same way as step 7, but using all the experiments of all the datasets together.

7. Results and Discussion

The following main results are shown comparing the performance of the validation methods in all the experiments carried out, and other aspects of the validation process. The focus of the analysis is on model selection, and no evaluation of the proper AUR estimation has been made, giving priority to the correct discrimination between better and worse predictive models.

7.1. Model Selection

To make a one-dimensional ranking for the performance of the validators in a given dataset, first, each validator has been ordered in a list according to its average selection error on the 150 experiments from lower to higher values. Then, a minimum rank value has been calculated for each validator, rejecting the null hypothesis that its selection error is equal or lower than the selection error of some previous validator using a one-sided Wilcoxon paired test: if the Wilcoxon test is positive compared to the previous validator in the list, the rank is increased by one. Otherwise, the one-sided Wilcoxon test is evaluated against the following previous validator in the list until a positive paired test is found, supporting the alternative hypothesis that the selection error of the validator is greater at a 95% confidence level. This is considered a minimum rank value for the validator. Then, the rank of the validator is determined as the maximum between the rank of the previous validator in the ordered list and the minimum rank found for it. The resulting ranking in each dataset is shown in Table A2 of Appendix A.
To make a validator performance ranking based on all the credit datasets, an overall rank has been calculated in the same way, performing a single Wilcoxon test in all the experiments of all the datasets, as shown in Table A2 of Appendix A. The validators have been ordered according to this overall rank from best to worst performance for model selection.
In this way, it can be seen in Table A2 that only 20 different ranks were found in a list with 65 different validation techniques. This means that many times the Kendall Tau distance is not different enough to consider one validator better than the next one. Something similar happens with the results of the individual credit datasets shown in the same table.
Table A2 shows that there are more draws in the ranking of the validators in the individual datasets, different from Taiwan, because the samples and populations are smaller, probably with less complex default patterns, which prevents them from easily discriminating among the performances of the validators. This is also the reason why the selection error is larger in these datasets.
Despite some differences found in the performance of the validation methodologies throughout the credit datasets, it was found that the robust cross-validation techniques are usually in the best positions in the rankings. In particular, the enhancement made to the DOB-SCV using the weight of evidence to calculate the distances for categorical variables presents better results than the original method for the same value of the k-fold parameter. In fact, looking at the overall ranking, almost all variants of the robust cross-validation algorithms are in the top positions for any value of the k-fold parameter, making this type of validator very convenient in situations where there is no clue about what k value to choose.
It is remarkable that the robust methods present better performance than the rest of the validation techniques, even for high values of the k parameter, supporting the initial hypothesis that homogeneous partitioning makes more stable test sets, allowing the use of more data for training without harming the results too much.
Of course, it can be seen from Table A2 that there is still an impact of the k parameter, and for this credit default use case, k = 10 and 5 were found to be the more suitable values, aligned with the findings of other research in different contexts.
Roughly speaking, following the more robust cross-validation variants, there is a large group of validators in positions eight and nine of the overall rank, mainly with traditional and fuzzy cross-validation techniques. There are also some robust methods in this group with k = 2, as the negative impact of this parameter is too large to be mitigated by the optimal partitioning. As there is data scarcity in the samples, taking only half of the data for training does not allow for the discovery of enough patterns, decreasing the discriminatory power of any type of cross-validation technique.
Following in position 10, the validation algorithms based on bootstrapping are all close together with the same overall rank, meaning that the methodological variants among them are not quite relevant for model selection purposes in the credit default use case.
The part of the list with the least performing validators with an overall rank greater than 10 is composed of the LOOCV and distinct types of validators due to the less convenient test sizes of 50% and 5% corresponding to the largest and smallest values tried. The disadvantage of taking only 5% of the data for testing (k = 20) is again due to data scarcity, but the problem now is the instability and heterogeneity of smaller test sets that will lead to higher variance of the AUR estimations. More data available for training is better for discovering more data patterns, but only without shortening the test set size too much.
In this respect, fuzzy methods do not use a particular k value, achieving intermediate results between the best and worst choices of the k parameter.
As expected, among the validation approaches that use independent test samples, maximally shifted CV methods are the least performing methods due to the large covariate shift induced by their partitioning technique, as anticipated in Section 4.
Finally, the in-sample methodologies are closing the list, meaning that some kind of independent test set should be used for model selection, at least when there are data restrictions in the available sample.
To better compare the different cross-validation methodologies, the results can be grouped in subsets of the same test set size to eliminate the effect of this factor in the comparison, except for the techniques that use a set of mixed parameter values. The selection of the hold-out test sample size or the k-fold parameter in cross-validation is subject to expert judgement in each practical application and will lead to better or worse validation performance in each particular context, depending on the representativeness of the available sample and the difficulty of modeling the underlying data patterns.
In order to fairly rank the cross-validation methods grouped by the same value of the k-fold parameter, two rankings of k = 5 and 10 have been made again with separate Wilcoxon non-parametric paired tests instead of directly using the ranks provided in the full list in Table A2. The results are shown in Table 2 and Table 3, together with the rest of the CV methods that do not have the need to establish a predefined value for the k parameter.
It can be seen from Table 2 and Table 3 that the distribution optimally balanced stratified cross-validation is really enhanced by the use of weight of evidence for calculating the distance between observations, being the most robust methodology for model selection in the current context.
However, for distribution balance cross-validation techniques, using the weight of evidence has no improvement, as it can also be verified for the rest of the k values of Table A2. In particular, the first version of this technique tends to perform better than the second version for increasing values of k. This is probably due to the larger randomness of the second version when distributing the observations in the partitions, giving slightly larger variances in the results that are more relevant when the test sets are smaller and more unstable.
In general, the robust CV variants are leading the classification, followed by the traditional and fuzzy methods. The Monte-Carlo hold-out variants have an irregular performance, but the stratification seems to be a disadvantage in these cases.
In particular, LOOCV performs as well as the best validation algorithms for the Australian dataset, probably due to the fact that this dataset has fewer observations and data diversity, leading to less unstable test sets than usual. Then, the more homogeneous and complete training sets that capture all the data patterns are more than enough to compensate for the test set instability in this specific credit dataset.
Closing the lists is MS-SCV, probably because maximally shifting the train/test sets should lead to higher volatility in the estimations, and therefore be more prone to mistakes when deciding between models.

7.2. Sample Size Impact

It is expected that the advantages of using robust validation methods will tend to disappear when increasing the size of the available dataset because the greater representativeness of the training and test sets will increase the homogeneity of the partitions for all kinds of partitioning methods.
To test this hypothesis, the same experiments were carried out with double and triple-sized samples for the Lending Club 2009 dataset, which is the one with larger selection errors. As these samples are more representative of the underlying population patterns, the selection errors are reduced significantly for all validation methods. The full validator comparison for sample sizes of 500, 1000, and 1500 observations is given in Table A3 of Appendix A.
To compare more clearly the effect of sample size for the cross-validation methods, Table 4 and Table 5 show separately the results for two groups corresponding to k = 5 and k = 10, respectively. The methods without a specific k-fold parameter value have also been added to the comparisons.
Table 4 and Table 5 show that there is no significant enhancement using the weight of evidence of the categorical covariates for the distribution optimally balanced stratified cross-validation methods when the sample size is large enough. Nevertheless, the robust validators DOB-SCV and DB-SCV are still at the top of the performance ranking, with some advantage for the DOB variant.
In Table 4, where k = 5, other non-robust methods scale up in the performance list, diluting some of the benefits of the homogeneous partitioning against the random resampling or fuzzy methods. In the case of k = 10, this is not yet clearly seen because the test sets are smaller and therefore more unstable than in the k = 5 case, still giving room for the benefits of homogeneous partitioning.

7.3. Best Model Selection

Apart from analyzing the capability of the validators to rank and discriminate among a wide range of credit default models with different predictive performances in terms of AUR, it is also interesting to check if the same validators have the ability to select the best ones. It could happen that a validator that differentiates well among intermediate models is not able to choose the best one or the top ones.
To check this, the number of times the model selected as the best one was in fact the best model in the unseen population has been counted for each validation technique, as shown in Table 6. In case of any draw in the number of successes, the number of second and third-best models correctly selected was used to break the tie and rank the validators.
Again, it is confirmed that the top of the rank is led by robust cross-validation methods, although the representative splitting cross-validation performs surprisingly much better than expected according to the overall selection error shown in Table A2 of Appendix A.
It is also surprising how much the stratified bootstrapping has improved in this rank compared with the results shown in Table A2. Then, to make a more robust assessment about which validators perform better when selecting good predictors, Table 7 shows the total number of times the model selected as the best one by each validator was in fact among the top five performances in the unseen population for all experiments in all datasets. To break any tie in the results, the validator with more second or third-best options in the true top five was preferred.
As shown in Table 7, the robust cross-validation methods are still leading the ranking as the most reliable in general for model selection purposes when the validated models are good predictors. In addition, Monte Carlo cross-validation seems to be a trustworthy alternative when using 20% of the data for testing.
Something important to highlight is that distribution optimally balanced stratified cross-validation, enhanced with the weight of evidence for categorical variables, is usually among the best validators, even for different values of the k-fold parameter. This is an important advantage because the selection of the k parameter is usually subject to expert judgement, and it is not clear when to choose one or another value. The DOB-WOE methodology mitigates the risk of not selecting an optimal value for k because the more homogeneous partitions and more stable results make this technique less sensitive to the train/test split size. This benefit could be potentially useful in other applications, different from the use case of credit default model selection.

7.4. Result Stability

Robust cross-validation methods achieve better results than other techniques in exchange for more computational requirements. This is due to the need to calculate the nearest neighbors of the observations being partitioned, which extends the validation execution time.
Then, more stable results do not necessarily mean any computational saving, but it is possible to measure the stability of the cross-validation estimations in terms of the number of repetitions needed before reaching a final estimate within a certain tolerance level.
Just to confirm this expected behavior, Figure 3 represents the average number of repetitions, after which all changes in the estimated AUR are below 0.001 for each validation technique when validating the model with the best averaged true AUR in each dataset. The validators in Figure 3 on the horizontal axis are in the same order as in Table A2 to visualize that there is a clear trend and therefore a relationship between the validator performance and the stability of its estimations: more stable results correlate with lower selection errors. The main exceptions to this rule are the two fuzzy methods (FNkfold and FUkfold), which need more repetitions than expected by the trend in all the datasets due to their fuzzy nature and therefore more volatile results.
Figure 3 shows that the distribution optimally balanced stratified cross-validation gives the more stable estimations of the AUR on average. This is coherent with the performance shown by this type of validation method, which is usually the best choice for model selection, particularly the enhanced algorithm that uses the weight of evidence for the distance between categorical covariates (see Table A2 in Appendix A).
The maximally shifted cross-validation methods need more repetitions than any other cross-validation method tested because their more heterogeneous partitions give place to a larger volatility of the AUR estimations.
The representative splitting cross-validation with k = 2 is not represented because the DUPLEX algorithm always divides the sample into two halves in the same way, and no repetition can be performed in this case.
Another way to check the existing relationship between result stability and model selection capability for any validator is to compare the number of repetitions needed to stabilize the results and the selection error in each dataset. It can be seen that the Taiwan dataset has the lowest selection error in Table 2 and Table 3 for any validator and needs the lowest number of repetitions to stabilize the results in Figure 3. On the other hand, the Lending Club dataset needs more repetitions and has greater selection errors than the rest of the datasets, while the German and Australian datasets behave similarly in both aspects.

7.5. Effect of the Number of Repetitions

If the number of repetitions leads to more stable results, it can be questioned if the same benefits of robust partitioning can be achieved by means of increasing the number of repetitions of more traditional cross-validation methods. In particular, as stratified 10-fold cross-validation is the best non-robust methodology (rank 6 of Table A2), this technique performed with a massive number of repetitions should be the most challenging validator to compare against the best robust cross-validation variant, also for k = 10.
In order to discover which of the two ways of reaching more stable results could be more advantageous for model selection, 120 experiments were used, comparing all the data sets, the original 10-fold cross-validation, its stratified version, and the distribution optimally balanced, enhanced with the weights of evidence for categorical covariates. The hypothesis testing was performed by a Wilcoxon one-sided paired test at a 95% confidence level for the selection error measured by the Kendall Tau, exactly in the same way as in the previous comparisons of Section 7.1 for the whole set of validators.
To make a feasible although intensive traditional cross-validation, a total of 1000 repetitions were performed for the 10-fold CV and SCV, while the most robust validator was repeated only 10 and 50 times. The results are shown in Table 8.
According to the overall ranking of Table 8, better model selection is achieved with the best robust methodology (DOB-WOE) compared to the best non-robust method (SCV), even making 100 times fewer repetitions. In addition, it is also shown that if we make a large number of repetitions, there is no material difference between the original 10-fold cross-validation and its stratified version.

7.6. Limitations of the Analysis

The overall rank of Table A2 in Appendix A could be considered the main contribution of this study because it is the comparison of all validation methods using all the paired tests in all the considered datasets. From the overall rank, it can be stated that there is statistical evidence of the better performance of robust validation methods because the 7 best-ranking positions correspond only to the more robust techniques, and any other robust validator with a rank greater than 8 has less convenient k parameter values different from 5 and 10.
Apart from this qualitative conclusion based on extensive quantitative results, one of the limitations faced in the analysis is that making a pure quantitative assessment of the benefits of a robust validation is not easy or feasible because those benefits depend on many circumstances. For example, the selection error measures, via the normalized Kendall Tau, how many positions each validator has altered the correct order of model performances, but how much the selection of one model instead of another one impacts some final generalization error will depend on the particular models considered, the available data, and the severity of the impact of wrong predictions in each practical application.
To provide a simple visualization of the clear division in model selection performance between robust and non-robust validators when there are data limitations, Figure 4 represents the selection error averaged per each rank for the validators in the first 12 ranks of Table A2. The last validators in the ranking that corresponds to the RSCV with k = 2 (which is deterministic), the maximally shifting cross-validations, and the in-sample methods are not represented in the graph to keep the visualization in a better scale, as the selection error is up to five times larger than the smallest one.
From Figure 4 or Table A2, it can be seen that the variation in the model selection error among the existing validation techniques in ranks 8 to 12, such as stratified CV, bootstrapping, Monte-Carlo CV, leave-one-out CV and even the fuzzy methods tested, is similar in magnitude to the difference found between validators in rank 8 and validators in rank 1 or 2. This means that the improvement in model selection using the best robust validators could be as large as the improvements made when choosing between two different methods from the existing validation methodologies and variants. Then, the benefits of robust model selection are potentially relevant in practice, and at least as relevant as the best improvements found among the methods used as potential validation alternatives.
In fact, previous studies making comparisons between bootstrapping, leave-one-out, cross-validation with different k-fold parameters, and other techniques generally reach qualitative conclusions [5,7,11,12,15,17,24,27] without being able to quantify the potential benefits. Even more, in this type of work, there are usually various methods identified as more convenient, but without concluding which one is the best. Therefore, recommendations are generally given to test several of them in each particular situation, highlighting the importance of making the correct choice of the model selection method, overall, for small sample sizes.
On the contrary, in our research scope, it seems that the gap between the performance of more robust validators and the rest of them makes it clear that a robust validation is better, although difficult to assess quantitatively due to the dependency on the data availability, the models compared for selection, and the severity of the prediction errors.
As the main conclusion, it is important to keep in mind that the more stable the stochastic validation method used is, the more reliable the selection process is expected to be. In this regard, robust cross-validation methods will help in selecting potentially better models of credit default in data-scarce situations.

8. Conclusions

This experimental research tackles the main issues of credit default model selection, where the problems of small data sample, imbalanced datasets, and covariate shift are present simultaneously.
Distribution-balanced and, to a greater extent, optimally balanced stratified cross-validation methods have been identified as robust stochastic methodologies for model selection in the credit default use case when compared against many other validation variants based on hold-out test samples, resampling techniques, and other cross-validation methodologies, including new proposed fuzzy variants.
In particular, the enhancements proposed to those robust CV methods have shown statistically significant improvements in the results, identifying DOB-WOE with k = 5 or 10 as the best robust validators for selecting consumer credit default models in situations of data scarcity when the sample size is large enough to build capable models, but it is quite sensible to set aside test data for validation purposes.
The key aspect to achieve better results in model selection in these situations with imbalanced data and scarce samples is the homogeneity of the train/test splits used to average the cross-validation results, reducing the variability of the AUR estimates in the k-fold partition, regardless of the bias.
The downside of the robust cross-validation methods is the computation required because it is necessary to calculate the nearest neighbors of the observations when distributing them in the partitions. Nevertheless, this could be performed just once for all the observations in the data sample, allowing for quickly performing many repetitions of the validation algorithm if desired. In fact, significantly increasing the number of repetitions instead of the partitioning homogeneity does not seem to be a better strategy, so it is probably better to dedicate some computational resources to distribute the observations more evenly using robust validation methods than focusing only on increasing the number of repetitions of traditional cross-validation.
As expected, when the size of the available data increases, the benefits of robust validation methods decrease because larger data samples directly lead to more homogeneous train/test splits for any splitting technique.
The above conclusions are supported not only by the overall results but also by each individual credit dataset used. Therefore, robust cross-validation methods could be worth leveraging to challenge current production models in lending institutions. Winnings could be assessed through backtesting the new credit default models selected by robust validation and the current ones, using recent historical data to quantify the differences.
In addition, potential future research on these robust validation algorithms could be carried out in other contexts with similar validation problems where the modeling benchmark involves different data patterns and types of predictive algorithms. For example, a use case particularly interesting for the financial industry and sometimes linked to credit defaults is the detection of credit card fraud, which involves more imbalanced datasets and less data availability than the credit default use case. Thus, it is usually more difficult to predict and manage fraudsters than defaulters.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/app15105495/s1.

Author Contributions

Conceptualization, Methodology, and Formal Analysis, J.V.A. and L.E.; Investigation, Resources, Software, Writing—Review and Editing, J.V.A.; Review and Supervision, L.E. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data supporting reported results can be found as Supplementary Materials, and links to the original public datasets analyzed are listed in the reference section.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A

This section shows detailed results of the models and validation techniques, averaged over all the samples used in each credit dataset.
Table A1. True model performance in the unseen population by averaging the model AUR over the 150 experiments in each dataset: “Model” = model number, “var.comb.” = variable combination, “avg.” = AUR average, “sd.” = AUR standard deviation.
Table A1. True model performance in the unseen population by averaging the model AUR over the 150 experiments in each dataset: “Model” = model number, “var.comb.” = variable combination, “avg.” = AUR average, “sd.” = AUR standard deviation.
Modelvar.comb.TaiwanGermanAustralianLending Club
avg.sd.avg.sd.avg.sd.avg.sd.
1A|N|N0.71640.00280.77480.01810.92860.00740.59650.0166
2O|N|N0.71550.00180.75210.01740.91040.00860.61320.0214
3E|N|N0.67160.00270.68700.01840.83310.01020.60140.0180
4N|A|N0.60580.00700.60670.01850.73430.01540.59080.0163
5A|A|N0.71880.00340.77760.01730.92660.00850.59160.0169
6O|A|N0.71880.00320.76110.01720.91680.00930.59960.0181
7E|A|N0.68150.00430.70010.01670.84840.01230.59250.0172
8N|O|N0.60620.00620.57940.02080.70880.01750.54390.0139
9A|O|N0.71740.00360.77790.01810.92650.00850.59380.0165
10O|O|N0.71750.00300.76140.01800.91350.01050.60450.0170
11E|O|N0.67900.00430.69370.01850.84510.01150.59630.0171
12N|E|N0.58590.00780.56900.02020.64980.01480.52670.0137
13A|E|N0.71750.00310.77440.01730.92930.00770.59270.0173
14O|E|N0.71700.00260.75020.01670.91880.00860.60300.0193
15E|E|N0.67560.00350.69290.01650.84330.01140.59470.0176
16N|N|A0.52170.01080.51270.01880.59750.01950.56810.0223
17A|N|A0.71590.00310.76900.01830.92430.00890.62300.0168
18O|N|A0.71270.00240.74410.01810.90910.01010.63490.0145
19E|N|A0.67270.00410.67810.01960.83090.01150.62380.0176
20N|A|A0.62450.00930.58860.01960.73190.01650.62150.0151
21A|A|A0.71860.00340.77240.01810.92260.00960.62100.0164
22O|A|A0.71740.00310.75320.01840.91340.01020.63030.0146
23E|A|A0.68200.00470.68940.01800.84070.01230.62060.0170
24N|O|A0.62050.00960.55610.02430.70140.02080.59110.0163
25A|O|A0.71770.00340.77390.01870.92170.00960.62240.0164
26O|O|A0.71640.00310.75430.01890.90850.01070.63270.0143
27E|O|A0.67970.00470.68790.01970.83650.01230.62300.0170
28N|E|A0.60000.01060.54580.02090.69010.01740.58720.0166
29A|E|A0.71700.00310.76770.01760.92500.00880.62320.0168
30O|E|A0.71500.00260.74190.01770.91500.00960.63360.0144
31E|E|A0.67710.00420.67940.01750.83860.01200.62380.0171
32N|N|O0.53070.01460.51090.01850.58550.01730.56480.0241
33A|N|O0.71480.00300.77200.01760.92670.00810.62500.0156
34O|N|O0.71300.00240.74890.01780.91010.00930.63940.0132
35E|N|O0.66930.00310.68200.01780.83620.00990.62730.0168
36N|A|O0.62610.00940.59580.01900.73040.01560.62540.0161
37A|A|O0.71860.00340.77600.01720.92470.00890.62250.0164
38O|A|O0.71750.00310.75820.01750.91470.00950.63290.0149
39E|A|O0.68200.00470.69460.01660.84480.01210.62300.0172
40N|O|O0.62290.01010.56460.02140.69830.02120.59220.0184
41A|O|O0.71720.00360.77640.01790.92380.00910.62420.0156
42O|O|O0.71590.00320.75870.01820.90900.01030.63600.0140
43E|O|O0.67950.00470.69000.01810.84080.01180.62600.0165
44N|E|O0.60140.01020.55270.02070.68820.01600.58750.0172
45A|E|O0.71650.00300.77180.01690.92740.00800.62480.0163
46O|E|O0.71530.00260.74730.01700.91660.00870.63680.0146
47E|E|O0.67430.00360.68620.01600.84390.01080.62640.0167
48N|N|E0.51130.00960.49890.02740.53530.02180.51790.0132
49A|N|E0.71490.00300.77100.01870.92790.00860.59730.0179
50O|N|E0.71360.00240.74620.01780.91400.00940.60900.0203
51E|N|E0.66840.00380.68360.02040.83030.01150.59850.0183
52N|A|E0.62200.00870.59700.01950.73780.01580.58810.0154
53A|A|E0.71800.00340.77330.01830.92570.00930.59350.0172
54O|A|E0.71790.00310.75530.01830.91700.01020.59980.0158
55E|A|E0.68000.00450.69450.01860.84480.01230.59280.0173
56N|O|E0.61450.00870.56270.02580.71330.01800.54490.0176
57A|O|E0.71690.00340.77470.01920.92540.00940.59500.0175
58O|O|E0.71660.00310.75630.01880.91370.01060.60310.0160
59E|O|E0.67670.00460.69160.02060.84140.01190.59550.0177
60N|E|E0.58740.01030.55610.02210.66140.01550.53750.0132
61A|E|E0.71620.00310.76960.01790.92830.00850.59480.0176
62O|E|E0.71550.00270.74430.01730.91910.00950.60280.0163
63E|E|E0.67360.00380.68550.01820.83850.01180.59510.0176
Max.0.720.0150.780.0270.930.0220.640.024
Min.0.510.0020.500.0160.540.0070.520.013
Table A2. Selection error and ranking of each validation method given by the normalized Kendall Tau, averaged over 150 random samples taken from each credit dataset. An overall ranking is calculated using all the experiments made in all credit datasets as the final ranking.
Table A2. Selection error and ranking of each validation method given by the normalized Kendall Tau, averaged over 150 random samples taken from each credit dataset. An overall ranking is calculated using all the experiments made in all credit datasets as the final ranking.
ValidatorAll DatasetsTaiwanGermanAustralianLending Club
ErrorRankErrorRankErrorRankErrorRankErrorRank
CV:DOB-WOE/10/500.166110.100620.112020.118310.33371
CV:DOB-WOE/5/500.166410.099210.111110.119220.33622
CV:DOB-WOE/15/500.166720.101830.112220.118810.33421
CV:DOB-WOE/20/500.167530.102340.112630.119120.33602
CV:DOB-SCV/5/500.167730.100420.113030.118910.33882
CV:DOB-SCV/10/500.169030.102230.112930.118410.34243
CV:DOB-SCV/15/500.169840.103050.112830.118610.34463
CV:DB-SCV1/15/500.170650.107080.113030.119520.34273
CV:DB-SCV1/5/500.171460.108390.113130.121040.34303
CV:DB-SCV1/10/500.171660.107380.112730.120030.34624
CV:DOB-SCV/20/500.171660.103760.113030.118710.35095
CV:SCV/10/500.172160.108490.113130.120330.34664
CV:DB-WOE2/10/500.172160.106870.111210.123260.34734
CV:DB-SCV2/5/500.172270.1090100.112630.122350.34483
HO:MCCV/0.1/10000.172480.1115120.114240.121850.34222
CV:DB-WOE2/5/500.172680.108190.112530.123150.34684
CV:RSCV/10/500.172680.108490.113130.120940.34814
CV:DOB-WOE/2/500.172780.100220.114650.127390.34874
CV:kfold/10/500.172980.1089100.113230.120430.34904
CV:DB-SCV2/10/500.172980.107580.111620.122450.35004
CV:DB-WOE2/15/500.172980.106370.110910.123460.35115
CV:DB-SCV1/20/500.172980.107080.112730.120430.35155
CV:SCV/5/500.173080.1100110.113430.120640.34814
CV:kfold/5/500.173180.1100110.113330.120740.34844
HO:FMCCV/0/10000.173280.1111120.114140.120740.34684
CV:RSCV/5/500.173280.1099100.113330.121550.34814
CV:FNkfold/0/500.173280.108790.113030.120740.35034
CV:SCV/15/500.173380.108490.113330.120030.35155
HO:MCCV/0.2/10000.173580.1116120.114450.120840.34744
CV:RndCV/0/500.173590.1093100.113740.121240.34974
CV:kfold/15/500.173790.108590.112930.120530.35295
HO:FSMCCV/0/10000.173990.1117120.114140.121650.34834
HO:SMCCV/0.2/10000.174090.1122130.114240.122050.34784
CV:FUkfold/0/500.174090.1095100.113940.120430.35215
CV:DOB-SCV/2/500.174290.100520.116860.126790.35255
CV:RSCV/15/500.174590.108790.112930.123460.35325
HO:SMCCV/0.1/10000.175090.1138140.115250.121650.34944
CV:DB-SCV2/15/500.1750100.106870.111010.123050.35916
CV:kfold/20/500.1752100.1089100.113530.120530.35786
CV:SCV/20/500.1754100.108790.113940.120840.35816
BTS:sboot/0/10000.1759100.1138140.115860.124570.34944
BTS:boot/0/10000.1761100.1140140.115860.124870.34974
BTS:sb632/0/10000.1762100.1146150.115760.124770.34984
BTS:b632/0/10000.1762100.1142140.115660.124870.35014
BTS:sb632+/0/10000.1762100.1138140.115960.124570.35054
BTS:b632+/0/10000.1763100.1137140.115960.125080.35075
CV:LOOCV0.1763110.106170.120480.118110.36056
CV:DB-WOE2/20/500.1763110.106570.110410.124360.36406
CV:DB-SCV2/20/500.1785110.107280.110810.123460.37247
CV:DB-WOE2/2/500.1786120.1155150.118770.1303100.34984
CV:RSCV/20/500.1790120.1089100.114650.119520.37297
CV:DB-SCV1/2/500.1795120.1155150.118070.1311100.35346
CV:DB-SCV2/2/500.1798120.1170160.117570.1308100.35376
CV:SCV/2/500.1803120.1172160.118070.1315110.35456
CV:kfold/2/500.1806120.1173160.119780.1320110.35346
HO:MCCV/0.5/10000.1806120.1175160.118370.1313110.35526
HO:SMCCV/0.5/10000.1807120.1182170.118780.1309100.35526
CV:RSCV/2/500.1984130.1243180.1372100.1447130.38728
CV:MS-SCV/20/500.2161140.2059190.115560.121240.42199
CV:MS-SCV/15/500.2180150.2122200.117570.119420.42309
CV:MS-SCV/10/500.2219160.2236210.123790.122950.41739
CV:MS-SCV/5/500.2470170.2529220.1483110.1324120.454410
CV:MS-SCV/2/500.3885180.4735230.3557120.1867140.538211
InSample/BIC/0/00.7203190.7797240.8012130.8324150.467810
InSample/AIC/0/00.8314200.8769250.8893140.8744160.685012
Table A3. Impact of the sample size on the selection error and ranking of each validation method for the Lending Club 2009 credit dataset. Selection error is given by the normalized Kendall Tau averaged over 150 random samples of sizes N = 500, 1000, and 1500.
Table A3. Impact of the sample size on the selection error and ranking of each validation method for the Lending Club 2009 credit dataset. Selection error is given by the normalized Kendall Tau averaged over 150 random samples of sizes N = 500, 1000, and 1500.
ValidatorLending Club—N=500Lending Club—N=1000Lending Club—N=1500
ErrorRankErrorRankErrorRank
CV:DOB-WOE/10/500.333710.251420.22333
CV:DOB-WOE/15/500.334210.254630.22604
CV:DOB-WOE/20/500.336020.256450.22785
CV:DOB-WOE/5/500.336220.249110.22232
CV:DOB-SCV/5/500.338820.246210.21911
HO:MCCV/0.1/10000.342220.258760.23016
CV:DOB-SCV/10/500.342430.250620.22132
CV:DB-SCV1/15/500.342730.256340.22845
CV:DB-SCV1/5/500.343030.253530.22855
CV:DOB-SCV/15/500.344630.255240.22483
CV:DB-SCV2/5/500.344830.255740.22986
CV:DB-SCV1/10/500.346240.254430.22855
CV:SCV/10/500.346640.256450.23006
CV:DB-WOE2/5/500.346840.255740.23097
HO:FMCCV/0/10000.346840.255740.23016
CV:DB-WOE2/10/500.347340.256950.22986
HO:MCCV/0.2/10000.347440.255040.23117
HO:SMCCV/0.2/10000.347840.256550.23147
CV:RSCV/10/500.348140.257050.22956
CV:RSCV/5/500.348140.255540.23097
CV:SCV/5/500.348140.255840.23097
HO:FSMCCV/0/10000.348340.256350.23097
CV:kfold/5/500.348440.255040.23046
CV:DOB-WOE/2/500.348740.261370.22805
CV:kfold/10/500.349040.256950.23097
HO:SMCCV/0.1/10000.349440.260170.23086
BTS:sboot/0/10000.349440.259470.23458
CV:RndCV/0/500.349740.257450.23137
BTS:boot/0/10000.349740.259270.23428
BTS:sb632/0/10000.349840.258260.23448
CV:DB-WOE2/2/500.349840.265290.23709
CV:DB-SCV2/10/500.350040.259270.22926
BTS:b632/0/10000.350140.258560.23478
CV:FNkfold/0/500.350340.257050.23016
BTS:sb632+/0/10000.350540.258460.23428
BTS:b632+/0/10000.350750.259070.23428
CV:DOB-SCV/20/500.350950.258960.22724
CV:DB-WOE2/15/500.351150.263080.23137
CV:DB-SCV1/20/500.351550.260070.23056
CV:SCV/15/500.351550.258660.23036
CV:FUkfold/0/500.352150.258250.23026
CV:DOB-SCV/2/500.352550.260670.22363
CV:kfold/15/500.352950.258460.23137
CV:RSCV/15/500.353250.259470.23107
CV:DB-SCV1/2/500.353460.261880.23619
CV:kfold/2/500.353460.2673100.240110
CV:DB-SCV2/2/500.353760.263780.23859
CV:SCV/2/500.354560.266490.238810
HO:MCCV/0.5/10000.355260.264990.23819
HO:SMCCV/0.5/10000.355260.264990.238810
CV:kfold/20/500.357860.261570.23257
CV:SCV/20/500.358160.263980.23247
CV:DB-SCV2/15/500.359160.262180.23046
CV:LOOCV0.360560.2713100.249611
CV:DB-WOE2/20/500.364060.265690.23247
CV:DB-SCV2/20/500.372470.266390.23277
CV:RSCV/20/500.372970.263380.23187
CV:RSCV/2/500.387280.2996110.258712
CV:MS-SCV/10/500.417390.3498120.312713
CV:MS-SCV/20/500.421990.3487120.318914
CV:MS-SCV/15/500.423090.3463120.312613
CV:MS-SCV/5/500.4544100.3822130.325514
InSample/BIC/0/00.4678100.5002140.559315
CV:MS-SCV/2/500.5382110.5094150.492414
InSample/AIC/0/00.6850120.7426160.740216

References

  1. Efron, B.; Tibshirani, R. An Introduction to the Bootstrap; Chapman and Hall/CRC Press: New York, NY, USA, 1993. [Google Scholar] [CrossRef]
  2. Kohavi, R. A study of cross-validation and bootstrap for accuracy estimation and model selection. Int. Jt. Conf. Artif. Intell. 1995, 14, 1137–1145. [Google Scholar]
  3. Bengio, Y.; Grandvalet, Y. No Unbiased Estimator of the Variance of K-Fold Cross-Validation. J. Mach. Learn. Res. 2004, 5, 1089–1105. [Google Scholar]
  4. Hastie, T.; Tibshirani, R.; Friedman, J. The Elements of Statistical Learning, 2nd ed.; Springer: New York, NY, USA, 2009. [Google Scholar] [CrossRef]
  5. Kim, J. Estimating classification error rate: Repeated cross-validation, repeated hold-out and bootstrap. Comput. Stat. Data Anal. 2009, 53, 3735–3745. [Google Scholar] [CrossRef]
  6. Arlot, S.; Celisse, A. A survey of cross-validation procedures for model selection. Stat. Surv. 2010, 4, 40–79. [Google Scholar] [CrossRef]
  7. Rodríguez, J.D.; Pérez, A.; Lozano, J.A. A general framework for the statistical analysis of the sources of variance for classification error estimators. Pattern Recognit. 2013, 46, 855–864. [Google Scholar] [CrossRef]
  8. Jiang, G.; Wang, W. Error estimation based on variance analysis of k-fold cross-validation. Pattern Recognit. 2017, 69, 94–106. [Google Scholar] [CrossRef]
  9. Moreno-Torres, J.G.; Sáez, J.A.; Herrera, F. Study on the Impact of Partition-Induced Dataset Shift on k-fold Cross-Validation. IEEE Trans. Neural Netw. Learn. Syst. 2012, 23, 1304–1312. [Google Scholar] [CrossRef]
  10. Zhang, Y.; Yang, Y. Cross-validation for selecting a model selection procedure. J. Econom. 2015, 187, 95–112. [Google Scholar] [CrossRef]
  11. Xu, L.; Fu, H.; Goodarzi, M.; Cai, C.; Yin, Q.; Wu, Y.; Tang, B.; She, Y. Stochastic cross validation. Chemom. Intell. Lab. Syst. 2018, 175, 74–81. [Google Scholar] [CrossRef]
  12. Beleites, C.; Baumgartner, R.; Bowman, C.; Somorjai, R.; Steiner, G.; Salzer, R.; Sowa, M.G. Variance reduction in estimating classification error using sparse datasets. Chemom. Intell. Lab. Syst. 2005, 79, 91–100. [Google Scholar] [CrossRef]
  13. Isaksson, A.; Wallman, M.; Göransson, H.; Gustafsson, M.G. Cross-validation and bootstrapping are unreliable in small sample classification. Pattern Recognit. Lett. 2008, 29, 1960–1965. [Google Scholar] [CrossRef]
  14. Varoquaux, G. Cross-validation failure: Small sample sizes lead to large error bars. NeuroImage 2018, 180, 68–77. [Google Scholar] [CrossRef]
  15. Xu, L.; Hu, O.; Guo, Y.; Zhang, M.; Lu, D.; Cai, C.; Xie, S.; Goodarzi, M.; Fu, H.; She, Y. Representative splitting cross validation. Chemom. Intell. Lab. Syst. 2018, 183, 29–35. [Google Scholar] [CrossRef]
  16. Zeng, X.; Martinez, T.R. Distribution-balanced stratified cross-validation for accuracy estimation. J. Exp. Theor. Artif. Intell. 2000, 12, 1–12. [Google Scholar] [CrossRef]
  17. López, V.; Fernández, A.; Herrera, F. On the importance of the validation technique for classification with imbalanced datasets: Addressing covariate shift when data is skewed. Inf. Sci. 2014, 257, 1–13. [Google Scholar] [CrossRef]
  18. Bradley, A.P. The use of the area under the ROC curve in the evaluation of machine learning algorithms. Pattern Recognit. 1997, 30, 1145–1159. [Google Scholar] [CrossRef]
  19. Huang, J.; Ling, C.X. Using AUC and accuracy in evaluating learning algorithms. IEEE Trans. Knowl. Data Eng. 2005, 17, 299–310. [Google Scholar] [CrossRef]
  20. Christodoulakis, G.; Satchell, S. The Analytics of Risk Model Validation, 1st ed.; Elsevier: Amsterdam, The Netherlands; Academic Press: Boston, MA, USA, 2008. [Google Scholar]
  21. Stone, M. Cross-Validatory Choice and Assessment of Statistical Predictions. J. R. Stat. Soc. Ser. B 1974, 36, 111–133. [Google Scholar] [CrossRef]
  22. Larson, S.C. The shrinkage of the coefficient of multiple correlation. J. Educ. Psychol. 1931, 22, 45–55. [Google Scholar] [CrossRef]
  23. Efron, B.; Tibshirani, R. Improvements on Cross-Validation: The 0.632+ Bootstrap Method. J. Am. Stat. Assoc. 1997, 92, 548–560. [Google Scholar] [CrossRef]
  24. Wong, T. Performance evaluation of classification algorithms by k-fold and leave-one-out cross validation. Pattern Recognit. 2015, 48, 2839–2846. [Google Scholar] [CrossRef]
  25. Marcot, B.G.; Hanea, A.M. What is an optimal value of k in k-fold cross-validation in discrete Bayesian network analysis? Comput. Stat. 2021, 36, 2009–2031. [Google Scholar] [CrossRef]
  26. Rodríguez, J.D.; Pérez, A.; Lozano, J.A. Sensitivity Analysis of k-Fold Cross Validation in Prediction Error Estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32, 569–575. [Google Scholar] [CrossRef]
  27. Krstajic, D.; Buturovic, L.J.; Leahy, D.E.; Thomas, S. Cross-validation pitfalls when selecting and assessing regression and classification models. J. Cheminform. 2014, 6, 1–15. [Google Scholar] [CrossRef]
  28. Markatou, M.; Tian, H.; Biswas, S.; Hripcsak, G. Analysis of Variance of Cross-Validation Estimators of the Generalization Error. J. Mach. Learn. Res. 2005, 6, 1127–1168. [Google Scholar]
  29. Sugiyama, M.; Krauledat, M.; Müller, K. Covariate Shift Adaptation by Importance Weighted Cross Validation. J. Mach. Learn. Res. 2007, 8, 985–1005. [Google Scholar]
  30. Delinquency Rate on Credit Card Loans, All Commercial Banks (DRCCLACBS)|FRED|St. Louis Fed. Available online: https://fred.stlouisfed.org/series/DRCCLACBS (accessed on 5 April 2025).
  31. Default of Credit Card Clients—UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/dataset/350/default+of+credit+card+clients (accessed on 5 April 2025).
  32. Statlog (German Credit Data)—UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/dataset/144/statlog+german+credit+data (accessed on 5 April 2025).
  33. Statlog (Australian Credit Approval)—UCI Machine Learning Repository. Available online: https://archive.ics.uci.edu/dataset/143/statlog+australian+credit+approval (accessed on 5 April 2025).
  34. All Lending Club Loan Data. Available online: https://www.kaggle.com/datasets/wordsforthewise/lending-club (accessed on 5 April 2025).
  35. Sun, Y.; Wong, A.K.C.; Kamel, M.S. Classification of imbalanced data: A review. Int. J. Pattern Recognit. Artif. Intell. 2009, 23, 687–719. [Google Scholar] [CrossRef]
  36. glm Function—RDocumentation. Available online: https://www.rdocumentation.org/packages/stats/versions/3.6.2/topics/glm (accessed on 5 April 2025).
Figure 1. Dataset partition into k equally sized subsets { Ω 1 , Ω 2 , Ω k } .
Figure 1. Dataset partition into k equally sized subsets { Ω 1 , Ω 2 , Ω k } .
Applsci 15 05495 g001
Figure 2. Classification of validation methods based on fundamental characteristics.
Figure 2. Classification of validation methods based on fundamental characteristics.
Applsci 15 05495 g002
Figure 3. Number of repetitions needed to stabilize the average AUR under a 0.001 threshold per each validator when validating the best model in each dataset and averaging the experiments. Error bars corresponding to the standard deviation are shown only for Taiwan dataset for clarity of the graph.
Figure 3. Number of repetitions needed to stabilize the average AUR under a 0.001 threshold per each validator when validating the best model in each dataset and averaging the experiments. Error bars corresponding to the standard deviation are shown only for Taiwan dataset for clarity of the graph.
Applsci 15 05495 g003
Figure 4. Horizontal axis represents the value of the overall rank of Table A2 (only till rank 12), while vertical axis is the average selection error of the validators in each rank value according to Table A2.
Figure 4. Horizontal axis represents the value of the overall rank of Table A2 (only till rank 12), while vertical axis is the average selection error of the validators in each rank value according to Table A2.
Applsci 15 05495 g004
Table 1. Public credit datasets used: name of the dataset, number of observations, number of variables, global default rate, and reference.
Table 1. Public credit datasets used: name of the dataset, number of observations, number of variables, global default rate, and reference.
Dataset NameRecordsFeaturesEvent RateReference
Taiwan30,0002322%[31]
German10002029%[32]
Australian6901444%[33]
Lending Club 200952811314%[34]
Table 2. Average selection error and ranking of CV methods using k = 5 or non-predefined k value for each credit dataset. An overall ranking is calculated using all the experiments made in all credit datasets as the final ranking.
Table 2. Average selection error and ranking of CV methods using k = 5 or non-predefined k value for each credit dataset. An overall ranking is calculated using all the experiments made in all credit datasets as the final ranking.
ValidatorAll DatasetsTaiwanGermanAustralianLending Club
ErrorRankErrorRankErrorRankErrorRankErrorRank
CV:DOB-WOE/5/500.166410.099210.111110.119210.33621
CV:DOB-SCV/5/500.167720.100420.113020.118910.33882
CV:DB-SCV1/5/500.171430.108340.113120.121020.34303
CV:DB-SCV2/5/500.172230.109050.112620.122330.34483
CV:DB-WOE2/5/500.172640.108140.112510.123140.34683
CV:SCV/5/500.173040.110060.113420.120620.34814
CV:kfold/5/500.173140.110060.113320.120720.34844
HO:FMCCV/0/10000.173240.111170.114130.120720.34684
CV:RSCV/5/500.173240.109950.113320.121530.34814
CV:FNkfold/0/500.173250.108740.113020.120720.35034
HO:MCCV/0.2/10000.173560.111670.114440.120820.34744
CV:RndCV/0/500.173570.109350.113730.121220.34974
HO:FSMCCV/0/10000.173980.111770.114130.121630.34834
HO:SMCCV/0.2/10000.174080.112280.114230.122030.34784
CV:FUkfold/0/500.174090.109550.113930.120420.35215
CV:LOOCV0.176390.106130.120450.118110.36055
CV:MS-SCV/5/500.247090.252990.148360.132450.45446
Table 3. Average selection error and ranking of CV methods using k = 10 or non-predefined k value for each credit dataset. An overall ranking is calculated using all the experiments made in all credit datasets as the final ranking.
Table 3. Average selection error and ranking of CV methods using k = 10 or non-predefined k value for each credit dataset. An overall ranking is calculated using all the experiments made in all credit datasets as the final ranking.
ValidatorAll DatasetsTaiwanGermanAustralianLending Club
ErrorRankErrorRankErrorRankErrorRankErrorRank
CV:DOB-WOE/10/500.166110.100610.112010.118310.33371
CV:DOB-SCV/10/500.169020.102220.112920.118410.34242
CV:DB-SCV1/10/500.171630.107340.112720.120020.34622
CV:SCV/10/500.172140.108450.113120.120320.34662
CV:DB-WOE2/10/500.172140.106830.111210.123240.34732
HO:MCCV/0.1/10000.172440.111570.114240.121840.34222
CV:RSCV/10/500.172640.108450.113120.120930.34813
CV:kfold/10/500.172940.108960.113230.120420.34903
CV:DB-SCV2/10/500.172940.107540.111610.122440.35003
HO:FMCCV/0/10000.173250.111170.114140.120730.34682
CV:FNkfold/0/500.173250.108750.113020.120730.35033
CV:RndCV/0/500.173550.109360.113730.121230.34973
HO:FSMCCV/0/10000.173960.111770.114140.121640.34833
CV:FUkfold/0/500.174060.109560.113930.120420.35213
HO:SMCCV/0.1/10000.175060.113880.115250.121640.34943
CV:LOOCV0.176360.106130.120460.118110.36054
CV:MS-SCV/10/500.221960.223690.123770.122940.41735
Table 4. Impact of the sample size on the selection error and ranking of cross-validation methods with k = 5 or non-predefined k-fold parameter for the Lending Club 2009 credit dataset. Selection error is given by the normalized Kendall Tau averaged over 150 random samples of sizes N = 500, 1000, and 1500.
Table 4. Impact of the sample size on the selection error and ranking of cross-validation methods with k = 5 or non-predefined k-fold parameter for the Lending Club 2009 credit dataset. Selection error is given by the normalized Kendall Tau averaged over 150 random samples of sizes N = 500, 1000, and 1500.
ValidatorLending Club—N = 500Lending Club—N = 1000Lending Club—N = 1500
ErrorRankErrorRankErrorRank
CV:DOB-WOE/5/500.336210.249110.22232
CV:DOB-SCV/5/500.338820.246210.21911
CV:DB-SCV1/5/500.343030.253520.22853
CV:DB-SCV2/5/500.344830.255730.22984
CV:DB-WOE2/5/500.346830.255730.23095
HO:FMCCV/0/10000.346840.255730.23014
HO:MCCV/0.2/10000.347440.255030.23115
HO:SMCCV/0.2/10000.347840.256540.23145
CV:RSCV/5/500.348140.255530.23095
CV:SCV/5/500.348140.255830.23095
HO:FSMCCV/0/10000.348340.256340.23095
CV:kfold/5/500.348440.255030.23044
CV:RndCV/0/500.349740.257440.23135
CV:FNkfold/0/500.350340.257040.23014
CV:FUkfold/0/500.352150.258240.23024
CV:LOOCV0.360550.271350.24965
CV:MS-SCV/5/500.454460.382260.32556
Table 5. Impact of the sample size on the selection error and ranking of cross-validation methods with k = 10 or non-predefined k-fold parameter for the Lending Club 2009 credit dataset. Selection error is given by the normalized Kendall Tau averaged over 150 random samples of sizes N = 500, 1000, and 1500.
Table 5. Impact of the sample size on the selection error and ranking of cross-validation methods with k = 10 or non-predefined k-fold parameter for the Lending Club 2009 credit dataset. Selection error is given by the normalized Kendall Tau averaged over 150 random samples of sizes N = 500, 1000, and 1500.
ValidatorLending Club—N = 500Lending Club—N = 1000Lending Club—N = 1500
ErrorRankErrorRankErrorRank
CV:DOB-WOE/10/500.333710.251410.22332
HO:MCCV/0.1/10000.342220.258740.23014
CV:DOB-SCV/10/500.342420.250610.22131
CV:DB-SCV1/10/500.346220.254420.22853
CV:SCV/10/500.346620.256430.23004
HO:FMCCV/0/10000.346820.255720.23014
CV:DB-WOE2/10/500.347320.256930.22984
CV:RSCV/10/500.348130.257030.22954
HO:FSMCCV/0/10000.348330.256330.23094
CV:kfold/10/500.349030.256930.23095
HO:SMCCV/0.1/10000.349430.260150.23084
CV:RndCV/0/500.349730.257430.23135
CV:DB-SCV2/10/500.350030.259240.22923
CV:FNkfold/0/500.350330.257030.23014
CV:FUkfold/0/500.352130.258230.23024
CV:LOOCV0.360540.271360.24966
CV:MS-SCV/10/500.417350.349870.31277
Table 6. Number of times the best model was correctly selected by each validator, summing up the results of all experiments in all credit datasets used.
Table 6. Number of times the best model was correctly selected by each validator, summing up the results of all experiments in all credit datasets used.
ValidatorNum.BestValidatorNum.BestValidatorNum.Best
CV:RSCV/296CV:SCV/5/5077CV:DB-SCV2/10/5071
CV:DOB-WOE/10/5091CV:DOB-WOE/20/5077CV:DB-SCV2/15/5071
CV:DOB-SCV/10/5081CV:DB-SCV1/2/5077CV:SCV/10/5071
CV:DOB-SCV/2/5081CV:DB-SCV2/20/5077CV:kfold/15/5070
CV:DOB-WOE/15/5080HO:MCCV/0.5/100075CV:RSCV/10/5070
BTS:sboot/0/100080BTS:b632+/0/100075CV:SCV/2/5069
CV:DB-WOE2/10/5080CV:DOB-SCV/15/5075CV:kfold/20/5069
CV:DOB-WOE/5/5080CV:SCV/15/5075BTS:boot/0/100069
CV:DB-WOE2/2/5080CV:DB-SCV1/15/5075CV:RSCV/20/5068
CV:RSCV/15/5080CV:DB-WOE2/5/5075BTS:b632/0/100068
HO:MCCV/0.2/100079CV:DOB-SCV/5/5074CV:SCV/20/5068
HO:SMCCV/0.1/100079CV:kfold/10/5073HO:MCCV/0.1/100066
CV:DB-WOE2/20/5079CV:kfold/2/5073CV:MS-SCV/15/5064
BTS:sb632/0/100079BTS:sb632+/0/100073CV:MS-SCV/10/5062
HO:SMCCV/0.5/100078CV:DB-SCV2/5/5072CV:MS-SCV/5/5061
CV:RSCV/5/5078HO:SMCCV/0.2/100072CV:LOOCV59
CV:DOB-SCV/20/5078CV:DB-SCV2/2/5072CV:MS-SCV/20/5056
CV:DB-SCV1/5/5078CV:kfold/5/5072CV:MS-SCV/2/5048
CV:DB-WOE2/15/5077CV:DB-SCV1/20/5072InSample/AIC/0/04
CV:DB-SCV1/10/5077CV:DOB-WOE/2/5072InSample/BIC/0/02
Table 7. Number of times the best model selected by each validator was among the true top five models in terms of AUR calculated in the unseen population. The counting sums up the results of all experiments in all credit datasets used.
Table 7. Number of times the best model selected by each validator was among the true top five models in terms of AUR calculated in the unseen population. The counting sums up the results of all experiments in all credit datasets used.
ValidatorRank 1–5ValidatorRank 1–5ValidatorRank 1–5
CV:DOB-WOE/10/50449CV:kfold/15/50433BTS:sb632/0/1000423
CV:DOB-WOE/5/50449BTS:b632/0/1000433BTS:boot/0/1000423
HO:MCCV/0.2/1000445CV:DOB-SCV/15/50432CV:DB-SCV1/2/50422
CV:DB-WOE2/5/50445HO:SMCCV/0.1/1000431CV:DOB-SCV/2/50421
CV:DOB-SCV/10/50444CV:SCV/15/50431CV:DOB-WOE/2/50420
CV:DOB-WOE/15/50443CV:RSCV/15/50430CV:RSCV/20/50420
CV:DB-SCV1/5/50442CV:DB-WOE2/20/50430HO:MCCV/0.5/1000416
CV:RSCV/5/50441CV:DB-SCV2/20/50430CV:DB-SCV2/2/50415
CV:DOB-SCV/20/50439CV:kfold/10/50430CV:SCV/2/50412
CV:DB-WOE2/15/50438CV:DB-SCV2/5/50430CV:kfold/2/50409
CV:DB-WOE2/10/50437HO:SMCCV/0.2/1000430CV:RSCV/2391
CV:kfold/5/50437CV:DB-WOE2/2/50429CV:LOOCV391
CV:DOB-WOE/20/50436BTS:sb632+/0/1000428HO:MCCV/0.1/1000357
BTS:sboot/0/1000435CV:DB-SCV1/20/50428CV:MS-SCV/15/50351
CV:DB-SCV1/10/50435CV:kfold/20/50428CV:MS-SCV/10/50345
CV:SCV/5/50435HO:SMCCV/0.5/1000427CV:MS-SCV/20/50337
CV:DB-SCV2/15/50435BTS:b632+/0/1000426CV:MS-SCV/5/50327
CV:DOB-SCV/5/50434CV:SCV/20/50426CV:MS-SCV/2/50180
CV:DB-SCV2/10/50434CV:SCV/10/50425InSample/BIC/0/065
CV:DB-SCV1/15/50433CV:RSCV/10/50425InSample/AIC/0/07
Table 8. Selection error and ranking of selected validation methods given by the normalized Kendall Tau, averaged over 120 random samples taken from each credit dataset. An overall ranking is calculated using all the experiments made in all credit datasets as the final ranking.
Table 8. Selection error and ranking of selected validation methods given by the normalized Kendall Tau, averaged over 120 random samples taken from each credit dataset. An overall ranking is calculated using all the experiments made in all credit datasets as the final ranking.
ValidatorAll DatasetsTaiwanGermanAustralianLending Club
ErrorRankErrorRankErrorRankErrorRankErrorRank
CV:DOB-WOE/10/500.167610.099910.114410.121110.33481
CV:DOB-WOE/10/100.168820.100620.114710.121810.33791
CV:kfold/10/10000.173130.107630.115410.122120.34732
CV:SCV/10/10000.173230.107630.115310.122010.34782
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Alonso, J.V.; Escot, L. Robust Cross-Validation of Predictive Models Used in Credit Default Risk. Appl. Sci. 2025, 15, 5495. https://doi.org/10.3390/app15105495

AMA Style

Alonso JV, Escot L. Robust Cross-Validation of Predictive Models Used in Credit Default Risk. Applied Sciences. 2025; 15(10):5495. https://doi.org/10.3390/app15105495

Chicago/Turabian Style

Alonso, Jose Vicente, and Lorenzo Escot. 2025. "Robust Cross-Validation of Predictive Models Used in Credit Default Risk" Applied Sciences 15, no. 10: 5495. https://doi.org/10.3390/app15105495

APA Style

Alonso, J. V., & Escot, L. (2025). Robust Cross-Validation of Predictive Models Used in Credit Default Risk. Applied Sciences, 15(10), 5495. https://doi.org/10.3390/app15105495

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop