Figure 1.
CUES component visualization for breast cancer (Logistic Regression). Each panel summarizes complementary aspects of model performance: (a) normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components ( and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 1.
CUES component visualization for breast cancer (Logistic Regression). Each panel summarizes complementary aspects of model performance: (a) normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components ( and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 2.
CUES component visualization for diabetes (Bagging). Each panel summarizes complementary aspects of model performance: (a) Normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components () and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 2.
CUES component visualization for diabetes (Bagging). Each panel summarizes complementary aspects of model performance: (a) Normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components () and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 3.
CUES component visualization for diabetes (KNN). Each panel summarizes complementary aspects of model performance: (a) normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components ) and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 3.
CUES component visualization for diabetes (KNN). Each panel summarizes complementary aspects of model performance: (a) normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components ) and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 4.
Macro-OVR CUES visualization for dermatology (logistic model). Each panel summarizes complementary aspects of model performance: (a) macro-averaged normalized net benefit curve (utility, ); (b) macro-OVR reliability diagram (calibration, ); (c) mean CUES component scores (); (d,e) bootstrap-based macro-utility stability analysis; (f) stability band ( SD) around the Macro-OVR CUES curve; (g) per-class CUES curves illustrating class-wise trade-offs; (h) bar chart of per-class utility () values; (i) probability distributions across classes. This figure summarizes how multi-class models balance calibration, utility, equity, and robustness under the Macro-OVR framework.
Figure 4.
Macro-OVR CUES visualization for dermatology (logistic model). Each panel summarizes complementary aspects of model performance: (a) macro-averaged normalized net benefit curve (utility, ); (b) macro-OVR reliability diagram (calibration, ); (c) mean CUES component scores (); (d,e) bootstrap-based macro-utility stability analysis; (f) stability band ( SD) around the Macro-OVR CUES curve; (g) per-class CUES curves illustrating class-wise trade-offs; (h) bar chart of per-class utility () values; (i) probability distributions across classes. This figure summarizes how multi-class models balance calibration, utility, equity, and robustness under the Macro-OVR framework.
Figure 5.
Macro-OVR CUES visualization for heart diseases (logistic model). Each panel summarizes complementary aspects of model performance: (a) macro-averaged normalized net benefit curve (utility, ); (b) macro-OVR reliability diagram (calibration, ); (c) mean CUES component scores (); (d,e) bootstrap-based macro-utility stability analysis; (f) stability band ( SD) around the Macro-OVR CUES curve; (g) per-class CUES curves illustrating class-wise trade-offs; (h) bar chart of per-class utility () values; (i) probability distributions across classes. This figure summarizes how multi-class models balance calibration, utility, equity, and robustness under the Macro-OVR framework.
Figure 5.
Macro-OVR CUES visualization for heart diseases (logistic model). Each panel summarizes complementary aspects of model performance: (a) macro-averaged normalized net benefit curve (utility, ); (b) macro-OVR reliability diagram (calibration, ); (c) mean CUES component scores (); (d,e) bootstrap-based macro-utility stability analysis; (f) stability band ( SD) around the Macro-OVR CUES curve; (g) per-class CUES curves illustrating class-wise trade-offs; (h) bar chart of per-class utility () values; (i) probability distributions across classes. This figure summarizes how multi-class models balance calibration, utility, equity, and robustness under the Macro-OVR framework.
Figure 6.
Macro-OVR CUES visualization for heart diseases (KNN). Each panel summarizes complementary aspects of model performance: (a) macro-averaged normalized net benefit curve (utility, ); (b) Macro-OVR reliability diagram (calibration, ); (c) mean CUES component scores (); (d,e) bootstrap-based macro-utility stability analysis; (f) stability band ( SD) around the Macro-OVR CUES curve; (g) per-class CUES curves illustrating class-wise trade-offs; (h) bar chart of per-class utility () values; (i) probability distributions across classes. This figure summarizes how multi-class models balance calibration, utility, equity, and robustness under the Macro-OVR framework.
Figure 6.
Macro-OVR CUES visualization for heart diseases (KNN). Each panel summarizes complementary aspects of model performance: (a) macro-averaged normalized net benefit curve (utility, ); (b) Macro-OVR reliability diagram (calibration, ); (c) mean CUES component scores (); (d,e) bootstrap-based macro-utility stability analysis; (f) stability band ( SD) around the Macro-OVR CUES curve; (g) per-class CUES curves illustrating class-wise trade-offs; (h) bar chart of per-class utility () values; (i) probability distributions across classes. This figure summarizes how multi-class models balance calibration, utility, equity, and robustness under the Macro-OVR framework.
Figure 7.
Stacked CUES component scores for multiple classifiers trained on the breast cancer dataset. The figure demonstrates how overall CUES values reflect critical dimensions of clinical reliability beyond traditional metrics, highlighting differences in model stability and equity.
Figure 7.
Stacked CUES component scores for multiple classifiers trained on the breast cancer dataset. The figure demonstrates how overall CUES values reflect critical dimensions of clinical reliability beyond traditional metrics, highlighting differences in model stability and equity.
Figure 8.
Stacked CUES component scores for classifiers on the heart disease binary dataset. The components illustrate the importance of calibration and equity, showing that moderate traditional performance does not guarantee clinical suitability when important reliability factors are lacking.
Figure 8.
Stacked CUES component scores for classifiers on the heart disease binary dataset. The components illustrate the importance of calibration and equity, showing that moderate traditional performance does not guarantee clinical suitability when important reliability factors are lacking.
Figure 9.
Stacked CUES component scores for classifiers on the dermatology dataset. Here, interpretable and straightforward models, such as Logistic Regression, achieve high composite CUES scores, demonstrating the benefits of transparency and stability for trustworthy clinical predictions.
Figure 9.
Stacked CUES component scores for classifiers on the dermatology dataset. Here, interpretable and straightforward models, such as Logistic Regression, achieve high composite CUES scores, demonstrating the benefits of transparency and stability for trustworthy clinical predictions.
Figure 10.
Stacked CUES component scores for classifiers on the heart disease multi-class dataset. The figure shows how subgroup variability and equity deficiencies lower CUES values, prompting a refined model selection to achieve more equitable clinical impact.
Figure 10.
Stacked CUES component scores for classifiers on the heart disease multi-class dataset. The figure shows how subgroup variability and equity deficiencies lower CUES values, prompting a refined model selection to achieve more equitable clinical impact.
Figure 11.
Stacked CUES component scores for classifiers on the diabetes dataset. The visualization reveals cases in which stability or equity gaps reduce the overall CUES score, reinforcing CUES’s role in identifying reliable and fair models for clinical deployment.
Figure 11.
Stacked CUES component scores for classifiers on the diabetes dataset. The visualization reveals cases in which stability or equity gaps reduce the overall CUES score, reinforcing CUES’s role in identifying reliable and fair models for clinical deployment.
Figure 12.
Overall, CUES features importance analysis for the binary datasets. The bar chart (A) displays the mean absolute CUES values, summarizing each feature’s overall contribution to model performance for the heart disease dataset using the KNN classifier. The bar chart (B) displays the mean absolute CUES values, summarizing each feature’s overall contribution to model performance for the diabetic dataset using a logistic classifier.
Figure 12.
Overall, CUES features importance analysis for the binary datasets. The bar chart (A) displays the mean absolute CUES values, summarizing each feature’s overall contribution to model performance for the heart disease dataset using the KNN classifier. The bar chart (B) displays the mean absolute CUES values, summarizing each feature’s overall contribution to model performance for the diabetic dataset using a logistic classifier.
Figure 13.
Overall, CUES features importance analysis for the multi-class datasets. The bar chart (A) displays the mean absolute CUES values, summarizing each feature’s overall contribution to model performance in the dermatology dataset using a logistic classifier. The bar chart (B) displays the mean absolute CUES values, summarizing each feature’s overall contribution to model performance for the heart disease dataset using a bagging classifier.
Figure 13.
Overall, CUES features importance analysis for the multi-class datasets. The bar chart (A) displays the mean absolute CUES values, summarizing each feature’s overall contribution to model performance in the dermatology dataset using a logistic classifier. The bar chart (B) displays the mean absolute CUES values, summarizing each feature’s overall contribution to model performance for the heart disease dataset using a bagging classifier.
Figure 14.
CUES component visualization for breast cancer (Bagging) using only the top 50% features selected using the CUES Score. Each panel summarizes complementary aspects of model performance: (a) normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components () and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 14.
CUES component visualization for breast cancer (Bagging) using only the top 50% features selected using the CUES Score. Each panel summarizes complementary aspects of model performance: (a) normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components () and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 15.
CUES component visualization for dermatology (Bagging) using only the top 50% features selected using the CUES Score. Each panel summarizes complementary aspects of model performance: (a) normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components () and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 15.
CUES component visualization for dermatology (Bagging) using only the top 50% features selected using the CUES Score. Each panel summarizes complementary aspects of model performance: (a) normalized net benefit (utility curve, ) across decision thresholds; (b) reliability diagram showing calibration performance (); (c) bar chart of CUES components () and overall CUES score; (d,e) bootstrap-based utility stability analysis showing temporal variation and distribution of ; (f) threshold-wise stability band ( SD) for the CUES curve; (g,h) equity analysis comparing subgroup CUES curves and utility balance (); (i) distribution of predicted probabilities for positive and negative classes. Together, these plots provide a holistic view of model utility, calibration, equity, and stability in imbalanced binary data.
Figure 16.
Accuracy vs. CUES (cross-validation scores): shows differences in optimal hyperparameters when tuned for accuracy versus CUES. CUES-optimized models often achieved slightly lower accuracy but higher overall CUES, reflecting improved calibration and stability.
Figure 16.
Accuracy vs. CUES (cross-validation scores): shows differences in optimal hyperparameters when tuned for accuracy versus CUES. CUES-optimized models often achieved slightly lower accuracy but higher overall CUES, reflecting improved calibration and stability.
Figure 17.
Components and metrics comparison: displays bar plots of accuracy, sensitivity, specificity, and individual CUES components (C, U, E, S). This allows inspection of trade-offs between classical metrics and explainable utility metrics.
Figure 17.
Components and metrics comparison: displays bar plots of accuracy, sensitivity, specificity, and individual CUES components (C, U, E, S). This allows inspection of trade-offs between classical metrics and explainable utility metrics.
Figure 18.
Heatmap: a comprehensive heatmap of all metrics and CUES components highlights each classifier’s performance profile. Darker colors indicate higher values, allowing rapid comparison across classifiers and optimization strategies.
Figure 18.
Heatmap: a comprehensive heatmap of all metrics and CUES components highlights each classifier’s performance profile. Darker colors indicate higher values, allowing rapid comparison across classifiers and optimization strategies.
Figure 19.
Comprehensive anthropometric equity analysis across classifiers. The figure includes overall CUES performance, decomposition into utility, calibration, equity, and stability, a heatmap of feature-wise equity, the relationship between equity and CUES, utility disparities, and a summary of feature-wise equity across all models.
Figure 19.
Comprehensive anthropometric equity analysis across classifiers. The figure includes overall CUES performance, decomposition into utility, calibration, equity, and stability, a heatmap of feature-wise equity, the relationship between equity and CUES, utility disparities, and a summary of feature-wise equity across all models.
Table 1.
Computational complexity and scalability of CUES components.
Table 1.
Computational complexity and scalability of CUES components.
| Component | Primary Operation | Time Complexity | Parallelizable |
|---|
| Utility (U) | Decision Curve Analysis across decision thresholds | O(N × T) | Yes |
| Calibration (C) | Reliability diagram/calibration error computation | O(N) | Yes |
| Equity (E) | Subgroup-wise utility estimation | O(N × G) | Yes |
| Stability (S) | Bootstrap resampling (B iterations) | O(B × N) | Yes |
| CUES Aggregation | Geometric mean of normalized components | O(1) | Yes |
Table 2.
Binary classification: multiple classifiers CUES performance.
Table 2.
Binary classification: multiple classifiers CUES performance.
| Dataset | Classifier | CUES Score | C | U | E | S | AUC CUES |
|---|
| Breast Cancer | Logistic | 0.903 | 0.914 | 0.900 | 0.862 | 0.941 | 0.900 |
| Breast Cancer | SVMrbf | 0.909 | 0.909 | 0.894 | 0.894 | 0.942 | 0.894 |
| Breast Cancer | Bagging | 0.858 | 0.858 | 0.834 | 0.813 | 0.936 | 0.834 |
| Breast Cancer | KNN | 0.866 | 0.870 | 0.846 | 0.824 | 0.932 | 0.846 |
| Diabetes | Logistic | 0.473 | 0.302 | 0.246 | 0.933 | 0.732 | 0.246 |
| Diabetes | SVMrbf | 0.443 | 0.276 | 0.220 | 0.897 | 0.718 | 0.220 |
| Diabetes | Bagging | 0.446 | 0.278 | 0.224 | 0.913 | 0.715 | 0.224 |
| Diabetes | KNN | 0.383 | 0.204 | 0.169 | 0.944 | 0.705 | 0.169 |
| Heart Disease | Logistic | 0.614 | 0.525 | 0.474 | 0.801 | 0.746 | 0.474 |
| Heart Disease | SVMrbf | 0.602 | 0.487 | 0.429 | 0.859 | 0.756 | 0.429 |
| Heart Disease | Bagging | 0.587 | 0.460 | 0.413 | 0.849 | 0.760 | 0.413 |
| Heart Disease | KNN | 0.589 | 0.473 | 0.416 | 0.841 | 0.746 | 0.416 |
Table 3.
Binary classification: multiple classifiers performance comparison between CUES and traditional metrics.
Table 3.
Binary classification: multiple classifiers performance comparison between CUES and traditional metrics.
| Dataset | Classifier | CUES Score | AUC CUES | AUROC | AUPRC | Accuracy | Sensitivity | Specificity |
|---|
| Breast Cancer | Logistic | 0.903 | 0.900 | 0.995 | 0.997 | 0.976 | 0.990 | 0.953 |
| Breast Cancer | SVMrbf | 0.909 | 0.894 | 0.995 | 0.997 | 0.969 | 0.975 | 0.960 |
| Breast Cancer | Bagging | 0.858 | 0.834 | 0.988 | 0.989 | 0.961 | 0.975 | 0.939 |
| Breast Cancer | KNN | 0.866 | 0.846 | 0.987 | 0.986 | 0.965 | 0.989 | 0.925 |
| Diabetes | Logistic | 0.473 | 0.246 | 0.830 | 0.719 | 0.768 | 0.565 | 0.877 |
| Diabetes | SVMrbf | 0.443 | 0.220 | 0.823 | 0.711 | 0.762 | 0.575 | 0.862 |
| Diabetes | Bagging | 0.446 | 0.224 | 0.819 | 0.698 | 0.756 | 0.616 | 0.831 |
| Diabetes | KNN | 0.383 | 0.169 | 0.780 | 0.624 | 0.727 | 0.543 | 0.826 |
| Heart Disease | Logistic | 0.614 | 0.474 | 0.911 | 0.907 | 0.845 | 0.786 | 0.894 |
| Heart Disease | SVMrbf | 0.602 | 0.429 | 0.896 | 0.886 | 0.835 | 0.800 | 0.864 |
| Heart Disease | Bagging | 0.587 | 0.413 | 0.890 | 0.884 | 0.818 | 0.771 | 0.858 |
| Heart Disease | KNN | 0.589 | 0.416 | 0.889 | 0.850 | 0.832 | 0.808 | 0.852 |
Table 4.
Multi-class classification: multiple classifiers CUES performance.
Table 4.
Multi-class classification: multiple classifiers CUES performance.
| Dataset | Classifier | CUES Score | C | U | E | S | AUC CUES |
|---|
| Heart Disease | Logistic | 0.421 | 0.262 | 0.229 | 0.849 | 0.629 | 0.229 |
| Heart Disease | SVMrbf | 0.400 | 0.235 | 0.194 | 0.871 | 0.653 | 0.194 |
| Heart Disease | Bagging | 0.390 | 0.231 | 0.205 | 0.824 | 0.608 | 0.205 |
| Heart Disease | KNN | 0.414 | 0.248 | 0.227 | 0.848 | 0.636 | 0.227 |
| Dermatology | Logistic | 0.877 | 0.948 | 0.926 | 0.747 | 0.910 | 0.926 |
| Dermatology | SVMrbf | 0.875 | 0.948 | 0.931 | 0.739 | 0.907 | 0.931 |
| Dermatology | Bagging | 0.848 | 0.913 | 0.888 | 0.752 | 0.854 | 0.888 |
| Dermatology | KNN | 0.852 | 0.920 | 0.890 | 0.732 | 0.885 | 0.890 |
Table 5.
Multi-class classification: multiple classifiers performance comparison between CUES and traditional metrics.
Table 5.
Multi-class classification: multiple classifiers performance comparison between CUES and traditional metrics.
| Dataset | Classifier | CUES Score | AUC CUES | AUROC | AUPRC | Accuracy | Sensitivity | Specificity |
|---|
| Heart Disease | Logistic | 0.421 | 0.229 | 0.818 | 0.589 | 0.682 | 0.564 | 0.820 |
| Heart Disease | SVMrbf | 0.400 | 0.194 | 0.812 | 0.585 | 0.682 | 0.541 | 0.819 |
| Heart Disease | Bagging | 0.390 | 0.205 | 0.801 | 0.588 | 0.652 | 0.544 | 0.805 |
| Heart Disease | KNN | 0.414 | 0.227 | 0.807 | 0.572 | 0.654 | 0.504 | 0.804 |
| Dermatology | Logistic | 0.877 | 0.926 | 0.998 | 0.992 | 0.978 | 0.975 | 0.996 |
| Dermatology | SVMrbf | 0.875 | 0.931 | 0.999 | 0.993 | 0.975 | 0.973 | 0.995 |
| Dermatology | Bagging | 0.848 | 0.888 | 0.993 | 0.974 | 0.954 | 0.949 | 0.990 |
| Dermatology | KNN | 0.852 | 0.890 | 0.996 | 0.978 | 0.963 | 0.965 | 0.993 |
Table 6.
Correlations between CUES and performance metrics.
Table 6.
Correlations between CUES and performance metrics.
| Scope | Spearman CUES AUROC | Spearman CUES AUPRC |
|---|
| Binary Single | 1.000 | 1.000 |
| Multi Many | 0.952 | 0.881 |
| Binary Many | 0.979 | 0.979 |
Table 7.
Multiple classifiers performance comparison between CUES and traditional metrics using two selected features subsets.
Table 7.
Multiple classifiers performance comparison between CUES and traditional metrics using two selected features subsets.
| Dataset | Classifier | Feature Subset | CUES Score | AUC CUES | AUROC | AUPRC | Accuracy | Sensitivity | Specificity |
|---|
| Binary |
| Breast Cancer | Logistic | Top 50% (15) | 0.950 | 0.908 | 0.997 | 0.998 | 0.975 | 0.970 | 0.948 |
| | | Top 75% (22) | 0.962 | 0.929 | 0.997 | 0.998 | 0.989 | 0.987 | 0.976 |
| | SVMrbf | Top 50% (15) | 0.957 | 0.929 | 0.996 | 0.997 | 0.984 | 0.981 | 0.967 |
| | | Top 75% (22) | 0.964 | 0.944 | 0.997 | 0.998 | 0.988 | 0.984 | 0.972 |
| | Bagging | Top 50% (15) | 0.970 | 0.940 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
| | | Top 75% (22) | 0.979 | 0.957 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
| | KNN | Top 50% (15) | 0.907 | 0.852 | 0.996 | 0.997 | 0.968 | 0.961 | 0.934 |
| | | Top 75% (22) | 0.926 | 0.879 | 0.997 | 0.998 | 0.975 | 0.970 | 0.948 |
| Diabetes | Logistic | Top 50% (4) | 0.357 | 0.128 | 0.754 | 0.594 | 0.710 | 0.641 | 0.868 |
| | | Top 75% (6) | 0.503 | 0.254 | 0.836 | 0.713 | 0.780 | 0.731 | 0.892 |
| | SVMrbf | Top 50% (4) | 0.534 | 0.297 | 0.859 | 0.773 | 0.801 | 0.754 | 0.908 |
| | | Top 75% (6) | 0.570 | 0.337 | 0.882 | 0.807 | 0.815 | 0.770 | 0.920 |
| | Bagging | Top 50% (4) | 0.905 | 0.807 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
| | | Top 75% (6) | 0.908 | 0.807 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
| | KNN | Top 50% (4) | 0.598 | 0.378 | 0.892 | 0.821 | 0.818 | 0.777 | 0.912 |
| | | Top 75% (6) | 0.595 | 0.362 | 0.886 | 0.800 | 0.797 | 0.764 | 0.872 |
| Heart Disease | Logistic | Top 50% (12) | 0.677 | 0.493 | 0.923 | 0.916 | 0.838 | 0.833 | 0.897 |
| | | Top 75% (18) | 0.714 | 0.551 | 0.939 | 0.932 | 0.868 | 0.863 | 0.915 |
| | SVMrbf | Top 50% (12) | 0.750 | 0.587 | 0.951 | 0.942 | 0.881 | 0.878 | 0.915 |
| | | Top 75% (18) | 0.795 | 0.649 | 0.962 | 0.958 | 0.908 | 0.906 | 0.927 |
| | Bagging | Top 50% (12) | 0.905 | 0.804 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
| | | Top 75% (18) | 0.911 | 0.827 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
| | KNN | Top 50% (12) | 0.643 | 0.422 | 0.890 | 0.885 | 0.792 | 0.790 | 0.818 |
| | | Top 75% (18) | 0.709 | 0.515 | 0.927 | 0.923 | 0.855 | 0.853 | 0.873 |
| Multi-class |
| Dermatology | Logistic | Top 50% (17) | 0.750 | 0.902 | 0.999 | 0.993 | 0.975 | 0.973 | 0.995 |
| | | Top 75% (25) | 0.760 | 0.940 | 0.999 | 0.996 | 0.981 | 0.979 | 0.996 |
| | SVMrbf | Top 50% (17) | 0.826 | 0.930 | 0.999 | 0.994 | 0.973 | 0.969 | 0.995 |
| | | Top 75% (25) | 0.829 | 0.942 | 0.999 | 0.995 | 0.975 | 0.972 | 0.995 |
| | Bagging | Top 50% (17) | 0.966 | 0.963 | 1.000 | 1.000 | 0.997 | 0.997 | 0.999 |
| | | Top 75% (25) | 0.970 | 0.968 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
| | KNN | Top 50% (17) | 0.778 | 0.826 | 0.995 | 0.978 | 0.932 | 0.929 | 0.986 |
| | | Top 75% (25) | 0.814 | 0.873 | 0.997 | 0.988 | 0.948 | 0.948 | 0.990 |
| Heart Disease | Logistic | Top 50% (6) | 0.388 | 0.159 | 0.785 | 0.591 | 0.607 | 0.470 | 0.760 |
| | | Top 75% (9) | 0.481 | 0.244 | 0.849 | 0.652 | 0.700 | 0.594 | 0.825 |
| | SVMrbf | Top 50% (6) | 0.435 | 0.188 | 0.856 | 0.695 | 0.719 | 0.593 | 0.830 |
| | | Top 75% (9) | 0.566 | 0.326 | 0.912 | 0.824 | 0.789 | 0.690 | 0.872 |
| | Bagging | Top 50% (6) | 0.826 | 0.698 | 0.981 | 0.976 | 0.941 | 0.941 | 0.966 |
| | | Top 75% (9) | 0.889 | 0.776 | 1.000 | 1.000 | 1.000 | 1.000 | 1.000 |
| | KNN | Top 50% (6) | 0.506 | 0.285 | 0.865 | 0.699 | 0.710 | 0.586 | 0.828 |
| | | Top 75% (9) | 0.523 | 0.296 | 0.873 | 0.722 | 0.706 | 0.595 | 0.825 |
Table 8.
Values for hyperparameter optimization grid search.
Table 8.
Values for hyperparameter optimization grid search.
| Classifier | Parameters |
|---|
| SVM | C = 10, γ = ‘scale’ |
| Bagging | Estimators = 20 |
| KNN | Neighbors = 3 |