Next Article in Journal
A Concept for MRI-Based Cholesteatoma Detection in Cochlear Implant Recipients
Previous Article in Journal
Reuniting and Endolymphatic Duct Macrophages: Localization and Possible Roles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Machine Learning Versus Simple Clinical Models for Cochlear Implant Outcome Prediction

1
Division of Otolaryngology, Head and Neck Surgery, University of Oldenburg, 26129 Oldenburg, Germany
2
Human Genetics, Faculty of Medicine and Health Science, University of Oldenburg, 26129 Oldenburg, Germany
3
AI4Health Division, University of Oldenburg, 26129 Oldenburg, Germany
4
Research Center Neurosensory Science, University of Oldenburg, 26129 Oldenburg, Germany
5
Cluster of Excellence “Hearing4All”, University of Oldenburg, 26129 Oldenburg, Germany
*
Author to whom correspondence should be addressed.
Audiol. Res. 2025, 15(6), 161; https://doi.org/10.3390/audiolres15060161
Submission received: 14 October 2025 / Revised: 14 November 2025 / Accepted: 19 November 2025 / Published: 21 November 2025
(This article belongs to the Section Hearing)

Abstract

Background/Objectives: Cochlear implantation is the most widely used treatment option for patients with severe to profound hearing loss. Despite being a relatively standardized surgical procedure, cochlear implant (CI) outcomes vary considerably among patients. Several studies have attempted to develop predictive models for CI outcomes but achieving accurate and generalizable predictions remains challenging. The present study aimed to evaluate whether simple and complex statistical and machine learning models could outperform the Null model based on various pre-CI implantation variables. Methods: We conducted a retrospective analysis of 236 ears with postlingual profound sensorineural hearing loss (SNHL) and measurable residual hearing (WRSmax > 0%) at the time of implantation. The median postoperative word recognition score with CI (WRS65(CI)) was 75% [Q1: 55%, Q3: 80%]. The dataset was divided using a 70:15:15 split into training (n = 165), validation (n = 35) and test (n = 36) cohorts. We evaluated multiple modeling approaches: different Generalized Linear Model (GLM) approaches, Elastic Net, XGBoost, Random Forest, ensemble methods, and a Null model baseline. Results: All models demonstrated similar predictive performance, with root mean squared errors ranging from 26.28 percentage points (pp) to 30.74 and mean absolute errors ranging from 20.62 pp to 23.75 pp. Coefficients of determination (R2) ranged from −0.468 to −0.073. Bland–Altman analyses revealed wide limits of agreement and consistent negative bias, while Passing–Bablok regression indicated calibration errors. Nonetheless, all models incorporating predictors significantly outperformed the Null model. Conclusions: Increasing model complexity yielded only marginal improvements in predictive accuracy compared with simpler statistical models. Pre-implantation clinical variables showed limited evidence of predictive validity for CI outcomes, although further research is needed.

1. Introduction

Hearing loss is one of the most common sensory impairments worldwide. For those with severe to profound sensorineural hearing loss (SNHL), cochlear implantation is the established standard of care in clinical practice today. A cochlear implant (CI) is an electronic hearing device that converts acoustic signals into patterned electrical stimulation of auditory nerve fibers, thereby bypassing the dysfunctional cochlear hair cells [1,2,3]. Despite incremental technological advancements in CI design, considerable variability in CI outcomes remains. Accurate prediction of postoperative CI performance would substantially improve clinical decision-making and enhance preoperative patient counselling regarding expected outcomes. Several studies have identified potential predictors of CI outcomes, such as age at hearing loss onset, duration of hearing loss, age at implantation, underlying etiology, CI experience, hearing aid use duration, CI manufacturer, percentage of active electrodes [4,5,6,7,8]. These predictors were used in various statistical models: regression models [4,5,6,7,9], random forests algorithms [10] and deep learning neural networks [11]. Comparative analyses between linear regression and machine learning models for CI outcome prediction showed modest advantages for ensemble-based approaches [12]. Reported model performance varied widely, explaining between 10% and 60% of outcome variance [4,6,8,13].
A previously reported general linear regression model (GLM) for the word recognition score with CI at 65 dB SPL (WRS65(CI)) suggested three predictors in a cohort of Cochlear™ recipients: Maximum word recognition score (WRSmax), age at implantation and word recognition score with hearing aids at 65 dB SPL (WRS65(HA)). The GLM showed a median mean absolute error (MAE) of 13 percentage points (pp) for predictions of WRS65(CI) in a subgroup with 4FPTA ≤ 80 dB HL [9] and a median MAE of 11.5 pp for WRS65(CI) in a subgroup with preoperative WRSmax > 0% [14]. These variables have been consistently identified across multiple studies as putative predictors of postoperative cochlear implant outcomes [4,8,15,16,17].
The aim of this study was to evaluate whether advanced machine learning algorithms could enhance the prediction of CI word recognition outcomes beyond traditional statistical models.

2. Methods

This study was approved by the Medical Ethics Committee of the University of Oldenburg (application #2024-021).

2.1. Patient Cohort

This retrospective analysis included CI recipients treated at our center (ENT university clinic Oldenburg at Evangelisches Krankenhaus Oldenburg, Germany) between 2007 and 2023, receiving CIs from Cochlear (Cochlear Deutschland GmbH & Co. KG, Hannover, Germany), Med-EL (MED-EL Elektromedizinische Geräte Gesellschaft m.b.H., Innsbruck, Austria) or Advanced Bionics (Advanced Bionics LLC, Valencia, CA, USA). The study period was defined because earlier records were either incomplete or did not meet the inclusion criteria. The endpoint of 2023 was chosen to ensure that only fully completed follow-up years were included, as 12-month outcome measurements for CI recipients implanted in 2024 were not yet available at the time of data extraction in 2025.
Inclusion criteria were postlingual sensorineural hearing loss with onset at ≥6 years of age, preoperative residual hearing (WRSmax > 0%), and ≥12 months of CI rehabilitation.
Exclusion criteria comprised patients with etiologies documented to negatively impact cochlear implant outcomes (e.g., neuroinflammatory disorders, neurodegenerative conditions, and vestibular schwannoma) [4], individuals with intellectual disabilities that could compromise speech recognition testing reliability, and non-native German speakers whose linguistic background might confound word recognition assessments.
From an initial databank of 667 patients with 828 implanted ears, application of the inclusion and exclusion criteria yielded a final analytical cohort of 205 patients and 236 implanted CIs (174 (~85% unilaterally implanted), 31 (~15% bilaterally implanted).
An additional analysis was conducted on a subcohort consisting exclusively of Cochlear™ recipients, as previous studies have indicated that the predictive accuracy of word recognition outcomes tends to be slightly higher in a more homogenous subcohort [18].

2.2. Predictor Variables

Based on previous studies [4,5,6,7,8,16] and clinical relevance, seven predictor variables were selected:
-
Age at onset of hearing loss (years);
-
Duration of hearing loss (months);
-
Age at implantation (years);
-
Preoperative four-frequency pure tone average (4FPTA: 0.5 kHz, 1 kHz, 2 kHz, 4 kHz) of the implanted ear;
-
Preoperative word recognition score at 65 dB with hearing aids (WRS65(HA)) of the implanted ear;
-
Preoperative maximum word recognition score (WRSmax) of the implanted ear;
-
CI experience (defined as the time span between the implantation date and the WRS65(CI), with a minimum rehabilitation period of 12 months).

2.3. Data Preparation

A 70:15:15 split strategy (70% training, 15% validation, 15% test) was implemented. Hyperparameter optimization and cross-validation procedures were performed on the training set, with the validation set used for model comparison and the test set reserved for final evaluation.
Missing data were imputed using multiple imputation by chain equations (MICE) with m = 5 using the predictive mean matching (PMM) method. To prevent data leakage, imputations were first trained on training data. Subsequently, the trained MICE model was used to impute missing values in the validation and test sets via its transform functionality. This approach preserves the multivariate structure of the data and avoids the biases introduced by simpler methods like mean imputation.

2.4. Models

Null model
A null model is a baseline model that assumes the outcome variable can be predicted without using any predictors. It serves as a reference point for evaluating whether more complex models, which include predictors, offer better predictive performance. In our analyses, the null model predicts the median of the training data outcome (WRS65(CI)) for all observations, allowing us to assess whether incorporating additional variables yields statistically significant gains in model accuracy.
Generalized Linear Model (GLM)
We employed four different GLM-based modeling approaches.
GLM_Full employed linear regression with Gaussian distribution, incorporating all eight predictor variables.
GLM_Backward used the same linear regression framework but applied AIC-based backward stepwise elimination starting from the full model. At each step, the removal of each predictor was tested individually and the one with the greatest AIC improvement was dropped. This process was repeated iteratively until no further improvement in AIC could be achieved. This resulted in the equation: WRS65(CI) = 72.6377 − 0.1960 × age at implantation + 0.1962 × WRSmax.
The GLM_External model implemented the published logistic regression model from Hoppe et al. [9], using the equation WRS65(CI) = 100/(1 + exp(−(β0 + β1 × WRSmax + β2 × Age + β3 × WRS65(HA)))) with fixed coefficients (β0 = 0.84, β1 = 0.012, β2 = −0.0094, β3 = 0.0059). This external model applies a logistic transformation to model the nonlinear relationship between predictors and WRS65(CI). In the original model, WRS65(CI) scores were converted into 20 binary (yes/no) responses [9].
Since the original model was validated in a cohort with residual hearing (WRSmax > 0%), we applied the same inclusion criterion throughout this study to ensure comparability.
GLM_External_Backward follows the same formulation as the GLM_External, but with re-estimated regression coefficients (β) based on the present study population (β0 = −2.257 × 101, β1 = −9.056 × 10−17, β2 = 3.574 × 10−16, β3 = 4.073 × 10−16).
Generalized Additive Model (GAM)
GAM is more flexible than GLM for modelling complex, nonlinear data. Unlike GLMs with fixed coefficients, GAM uses smoothing splines to model relationships between response variables and predictors [19]. In our analysis, we applied smoothing splines to each predictor. The model is fitted with a gaussian family on the training data.
Elastic Net
Elastic net is a regularized linear regression method that combines L1 (Lasso) and L2 (Ridge) penalties to improve model performance. This technique balances feature selection (L1) and coefficient shrinkage (L2) to prevent overfitting and enhance generalizability, making it particularly useful when predictors are highly correlated [20]. We performed grid search with 10-fold cross-validation for hyperparameter optimization (α: 11-point grid from 0 to 1; λ: automatically selected). The final model was then trained on the full training set using the optimal α-λ combination.
Random Forest
Random Forest is an ensemble method that uses multiple decision trees to enhance model robustness and generalizability. It relies on bagging, where predictions are aggregated from decision trees trained on bootstrap samples [21,22]. This reduces overfitting by lowering tree variance. This approach can capture complex, nonlinear relationships and interactions among predictor variables, making it particularly suitable for predicting heterogeneous outcomes in CI recipients. In our analysis, we performed a grid search to tune hyperparameters, including the number of trees (between 100 and 1500), number of predictors randomly selected at each split (2 to 6), the minimum number of samples in nodes (1, 3, 5, 10), and the maximum of nodes per tree (between 0 and 200).
eXtreme Gradient Boosting (XGBoost)
XGBoost is an advanced ensemble learning technique based on the principles of boosting. Boosting combines multiple weak predictive models, typically decision trees, into a single strong model. Each model in the sequence corrects residual errors from the previous ones, improving prediction accuracy. Gradient boosting adjusts model parameters via gradient descent to minimize residual errors [23,24]. XGBoost has shown the potential to accurately predict clinical outcomes due to its iterative refinement, which enhances model precision, and its inherent robustness in managing missing or imbalanced data. This methodological strength renders it particularly well-suited for the analysis of complex, multifactorial outcomes in CI recipients. In our analysis, an enhanced grid search was performed to tune hyperparameters, including the learning rate (0.01, 0.05, 0.1, 0.2, 0.3), maximum tree depth (3 to 10), row subsampling (0.7 to 1.0), column subsampling per tree (0.7 to 1.0), and minimum child weight (1, 3, or 5). Models were trained up to 2000 boosting rounds with early stopping after 100 rounds without improvement on a held-out validation set.
Ensemble
Ensemble methods have demonstrated particular effectiveness in clinical prediction modeling, as they can enhance generalizability across diverse patient populations and clinical settings—an important consideration when modeling outcomes in cochlear implant recipients.
We implemented two ensemble methods combining predictions from the top three models based on validation set performance: Null model, Elastic Net and XGBoost. The simple ensemble calculated the arithmetic mean of the three model outputs to improve prediction stability and reduce variance. The weighted ensemble assigned each model’s contribution proportional to its inverse validation RMSE, favoring better-performing models in the final prediction.

2.5. Model Evaluation and Data Analysis

Data were collected in Microsoft Excel. All calculation, modelling, performance evaluation and figure generation were done using RStudio ([25], version 2024.12.1+563) with assistance from Claude ([26], version 4.0 (2024)). The outputs were reviewed and validated by the authors. Baseline characteristics of the cohort were analyzed using statistical approaches. Shapiro–Wilk normality testing was performed for each continuous variable. Variables with normal distribution were summarized using the mean and standard deviation (SD), whereas non-normally distributed data were described using the median and the percentiles at 25% to 75% (median ± Q1, Q3). Model performance was assessed using root mean squared error (RMSE), mean absolute error (MAE), and the coefficient of determination (R2) (calculated as 1—SS_residual/SS_total), with the latter representing traditional R2 for linear models and pseudo-R2 for nonlinear models. Models were primarily ranked by RMSE due to its direct clinical interpretability in word recognition scores (WRS). R2 and MAE are reported as supplementary performance indicators. Predicted scores were aligned with the corresponding observed values, and performance metrics were computed accordingly.
To assess the agreement between predicted and actual outcomes, Bland–Altman analysis was performed. Bias and limits of agreement (LoA) were calculated for each model to quantify systematic and individual-level error.
Passing-Bablok regression was employed to assess calibration performance. This non-parametric method provides estimates of proportional bias (slope) and constant bias (intercept), allowing for a comprehensive evaluation of systematic deviations from perfect predictions.
To determine whether model performances differed from the Null model, bootstrap resampling was used. Repeated iterations were performed to generate confidence intervals for RMSE, MAE and R2 differences.

3. Results

3.1. Patient Characteristics

A total of 236 implanted ears of adult patients with postlingual hearing loss and residual preoperative speech perception (WRSmax > 0%) were included. After applying a 70:15:15 split, the training cohort comprised 165 implanted ears. Table 1 summarizes the distribution of the cohort’s key clinical variables.

3.2. Model Performance Analysis

We assessed predictive performance across a range of models, including GLM_Full, GLM_Backward, GLM_External_Backward, GLM_External, GAM, Elastic Net, Random Forest, XGBoost, and two ensemble approaches (Weighted and Top3 models). Performances were benchmarked against the Null model.
All models demonstrated comparable RMSE and MAE, ranging between 26.28 pp and 30.74 percentage points (pp) and 20.62 pp and 23.75 pp, respectively (Table 2, Figure 1). R2 were consistently low and negative (−0.468 to −0.073), indicating poor model fit. Among the predictive models, Random Forest yielded the most favorable performance, with the lowest error metrics and highest R2 value. The different GLM models demonstrated marginally poorer accuracy, while the Null model performed worst, exhibiting the highest error metrics and lowest R2 (Table 2, Figure 1).
Across models, predicted outcome values spanned a broad range (Figure 1), reflecting the models’ ability to capture nearly the full spectrum of outcome variability.
Bland–Altman-analysis revealed consistently wide limits of agreement (LoA) for all models, ranging from ~96 to 105 pp, indicating substantial individual-level prediction error and limited agreement with actual outcomes. In addition, all models exhibited negative bias (range: −9.01 pp to −17.36 pp), suggesting a systematic tendency to underpredict outcomes (Table 2). Visual inspection of Bland–Altman plots further supported the presence of systematic bias. All models exhibited a diagonal pattern of residuals, rather than random scatter around zero, indicating systematic deviation from perfect prediction. The inclusion of predictors introduced additional dispersion along this diagonal trend, consistent with a modest but insufficient increase in explained variance (Figure 2).
These findings were corroborated by Passing-Bablok (Pb) regression, which identified notable calibration errors across models. The slopes were markedly less than 1 (range: 0.019–0.177), indicating a proportional underprediction, particularly at higher outcome values. Concurrently, the intercepts were consistently large and positive (range: 58.66 to 69.48 pp), reflecting a substantial systematic offset (Table 2).
Although overall predictive performance was limited, all models incorporating predictors significantly outperformed the Null model in terms of RMSE and R2. In terms of MAE, all models except GAM, Hoppe_External and XGBoost demonstrated statistically significant improvement (Table 3).
The improved performance metrics in all models over the Null model suggest that the inclusion of predictors contributes to enhanced predictive accuracy of the models, even if overall explanatory power remains limited.
Subgroup analyses were conducted in a subset of Cochlear™ recipients (n = 16 in the test cohort) to evaluate model performance in a more clinically homogenous population, which was previously shown to have a slightly better, although not significant, prediction accuracy in the Hoppe_External model [18]. Within this group, predictive accuracy was marginally improved across all models compared to the entire cohort (Table S1, Figure S1). Notably, Hoppe_External achieved the best performance, while XGBoost performed worst, exhibiting the highest error metrics and lowest R2 (Table S1, Figure S1). Similar to the entire cohort, predicted values remained confined to a narrow range, and correlations between actual and predicted CI outcomes were weak (Figure S1).
The Bland–Altman analysis showed wide LoAs (ranging from ~77 to ~90 pp) and consistently negative biases (−3.91 to −8.75 pp), again indicating a persistent tendency toward underprediction (Table S1). Residuals followed a diagonal distribution in Bland–Altman plots, and the inclusion of predictors contributed to increased dispersion along this trend (Figure S2).
Passing–Bablok regression results corroborated these patterns, with slopes substantially below 1 (range: 0.025–0.168), indicating persistent proportional underestimation at higher observed values. Intercepts remained large and positive (range: 59.01 to 70.88 pp), further highlighting systematic prediction error across all models (Table S1).
The improvement of the performance metrics over the Null model was not significant (Table S2), which might be due to the small sample size.

3.3. Feature Importance for CI-Outcome Predictions

Given the limited predictive accuracy of all models, we examined feature importance by assessing correlations between predictor variables and WRS65(CI) using Pearson’s correlation coefficient. Across the entire cohort, all predictors exhibited weak to moderate associations with CI outcome, with preoperative 4FPTA demonstrating the strongest correlation. Notably, duration of hearing loss (HL) and WRSmax showed the weakest association with CI outcomes (Figure 3).
In the Cochlear™ recipient subgroup, age at implantation emerged as the strongest predictor, with a moderate correlation approaching an r ≈ 0.5. Conversely, WRS65(HA), 4FPTA and duration of hearing loss exhibited weak correlations with the CI outcome (Figure S3).
These findings suggest that the limited predictive performance of current models may stem from insufficiently informative predictors.

4. Discussion

In this study, we systematically compared the predictive performance of traditional regression- and more complex machine learning-based models against the Null model in predicting word recognition outcomes among CI recipients. Across all approaches, predictive performance was limited. The RMSE and MAE values were similar across models, with RMSE ranging from 26.28 pp to 30.74 pp and MAE from 20.62 pp to 23.75 pp, respectively. However, all predictive models significantly outperformed the Null model. These results are consistent with, albeit slightly higher than, error ranges reported in previous studies. Shafieibavani et al. [12] compared patient cohorts of three different centers (Vanderbilt University Medical Center (VUMC), Ear Science Institute Australia (ESIA), and Medizinische Hochschule Hannover (MHH)) by using an XGB-RF model and showed MAE ranges from 17.9 to 21.8 pp. Similarly, a recent study compared XGBoost with a traditional regression model and found comparable MAEs ranging between 17% and 22% [27], consistent with the findings from Shafieibavani et al. [12]. In addition, when outcomes were classified into quintile percentiles, XGBoost showed improved predictive performance over the traditional regression [27]. Crowson et al. [11] analyzed outcomes either as individual data points or categorized those into three groups: high, mid and low performers. They reported an RMSE of 25.3 for XGBoost and 0.57 for neural networks when only numerical variables were included. Notably, inclusion of text-based variables in the neural network model reduced classification accuracy from 95.4% to 73.3% and increased RMSE to approximately 25 pp [11], indicating decreased predictive performance with the addition of textual data. Another recent study employed a decision tree regression model and reported a MAE of 17–18 pp, with a relatively high standard deviation of ~12.5 to 14.5 pp [28], reflecting substantial variability in the model’s predictions.
Hoppe et al. [14] reported a lower MAE of 11.5 pp in a more homogeneous cohort of Cochlear™ recipients with WRSmax > 0%. Although we applied identical inclusion criteria in our subgroup analysis, predictive accuracy did not reach the same level. Several factors may explain the discrepancy. First, our cohort had a broader implantation timeframe (2007–2023) with heterogeneous cochlear implant generations. Second, our Cochlear™ subgroup had a smaller sample size. Third, institutional differences such as surgical techniques, audiological protocols, and patient selection criteria may have contributed to the performance gap between studies.
The overall model fit in our models was negligible, with R2 values ranging from −0.468 to −0.073. In all models, predictions exhibited a narrow range and deviated systematically from measured outcomes, as confirmed by Bland–Altman analyses and Passing–Bablok regression. While R2 is commonly interpreted as explained variance, this interpretation does not apply in non-linear models, where R2 functions more generally as an indicator of goodness of fit. Nevertheless, the broader literature reports highly variable explained variances ranging between 10% and 66% [4,5,8,10,11,12,29]. In these studies, higher R2 values in smaller cohorts may reflect statistical overfitting rather than superior predictive capability. While overfitting is detected through validation/test set performance, small test sets (such as our n = 16 subgroup) face the additional limitation of reduced representativeness. Smaller cohorts may not adequately capture population diversity, limiting the generalizability of performance estimates even when overfitting is not present. This should be considered when interpreting our subgroup analysis results. However, a previous study also suggested that sample size alone was unlikely to improve model performance [12]. Across the literature, a wide range of predictor variables have been employed, including age at implantation, duration of hearing loss, preoperative tone audiometry and speech recognition, duration of CI experience, the area ratio of ventral cochlear nerve to facial nerve at the cerebellopontine angle, and etiology [4,5,8,10,12,29,30]. Our findings indicate that the variables included in our models were insufficient to predict post-implantation outcomes for our cohort.
Feature importance analysis corroborated this interpretation: Classical predictors—including age at onset of hearing loss (years), duration of hearing loss (months), age at implantation (years), preoperative 4FPTA of the implanted ear, WRS65(HA) of the implanted ear, preoperative WRSmax of the implanted ear, CI experience (months)—exhibited only weak to moderate correlations with WRS65(CI). Our results are in contrast to previous studies that identified WRSmax, WRS65(HA), and age at implantation as relevant predictors for CI outcome [4,8,15,16,17].
Notably, increasing model complexity—via ensemble methods, Random Forest, gradient boosting, or GAM—did not improve predictive accuracy compared with simpler regression approaches (GLMs with or without variable selection). This aligns with a prior study suggesting that prediction accuracy is constrained by the quality of the input variables, rather than by model choice [12].
This study has several important limitations that should be considered. The cohort exhibited substantial heterogeneity, encompassing a wide implantation timeframe (2007–2023), multiple CI manufacturers, and clinical methodologies that evolved over time. Such temporal and technical variability may have introduced noise into the analyses. As reported by Blamey et al. [4,30], the relative importance of predictive variables can change over time. Thus, future investigation stratifying by implantation period may yield additional insights. However, previous findings indicate that temporally defined subgroups may not differ substantially from randomly distributed subgroups [28]. Secondly, both ears of bilaterally implanted patients were analyzed as independent observations. Differences in the configuration of hearing loss—unilateral, bilateral, or asymmetric—could have affected model performance, as individuals often develop compensatory listening strategies that emphasize the better-hearing ear in daily communication. Thirdly, the models were limited to routinely collected data, primarily demographic and subjective audiometric variables. To enhance predictive accuracy, future work should integrate additional predictors such as imaging-derived metrics, genetic factors, cognitive measures, or more objective audiological assessments. Simply refining statistical or machine learning techniques without improving input variable quality is unlikely to yield substantial gains in predictive accuracy.
Finally, the statistical power was constrained by small test set sizes, potentially affecting the robustness of performance estimates. All models exhibited a systematic underprediction of outcomes. A nested cross-validation for evaluating the performance might lead to more reliable results.
Subgroup analyses were conducted Further research, including external validation across diverse cohorts, will be needed. For this, public multi-center datasets with standardized test sets would be beneficial [31].

5. Conclusions

This study evaluated the predictive performance of linear and non-linear models for CI outcomes. Although the included variables have previously been identified as significant contributors, their assessment within more advanced modeling frameworks did not yield substantial improvements in predictive accuracy. Across all models, the error metrics were high and the explained variance remained minimal, underscoring the limited predictive value of currently available variables. Importantly, increasing model complexity did not enhance performance, suggesting that model architecture alone cannot compensate for weakly informative input features. These findings emphasize the need for the identification and integration of additional or alternative predictors—potentially including neuroimaging, genetic, cognitive, or more refined audiological measures—to improve the precision of CI outcome prediction. Moreover, rigorous external validation across diverse cohorts will be essential to strengthen both the robustness and generalizability of future predictive models.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/audiolres15060161/s1, Figure S1: Actual versus predicted CI-outcome in Cochlear™ recipients across different models using independent train/test splits; Figure S2: Bland–Altman analyses across different model performances for the subcohort of Cochlear™ recipients; Figure S3: Feature importance analysis based on absolute correlation with CI outcome in Cochlear™ recipients; Table S1: Summary statistics of model performances in Cochlear™ recipients; Table S2: Improvement in statistical metrics over the Null model in Cochlear™ recipients.

Author Contributions

Conceptualization, R.B.; Methodology, R.B. and N.S.; Validation, R.O. and R.B.; Formal Analysis, R.B.; Investigation, R.O.; Resources, A.R.; Writing—Original Draft Preparation, R.O.; Writing—Review and Editing, R.B.; Supervision, R.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Forschungspool (PB Rare 2019-003), School of Medicine and Health Sciences at the University Oldenburg.

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki and approved by the Ethics Committee of the University of Oldenburg (protocol code: 2024-021, date of approval: 16 February 2024).

Informed Consent Statement

Patient consent was waived due to the retrospective analysis of anonymized data, as approved by the ethics committee of the University of Oldenburg.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy restrictions.

Acknowledgments

We have used Claude Sonnet 4.0 (Anthropic, San Francisco, CA, USA, 2024) to assist in generating R scripts for data processing, modeling, performance evaluation and figure generation. We have reviewed and edited the output and take full responsibility for the content of this publication.

Conflicts of Interest

R.B. and A.R. received travel compensation from Cochlear and Med-EL. All other authors declare no conflicts of interest.

References

  1. AWMF. S2k-Leitlinie Cochlea-Implantat Versorgung; AWMF-Register-Nr, 017/071; AWMF: Bradford, UK, 2020. [Google Scholar]
  2. Kinney, S.E. Current Status of an Implantable Cochlear Pros thesis. Artif. Organs 1979, 3, 379–382. [Google Scholar] [CrossRef] [PubMed]
  3. WHO. World Report on Hearing; WHO: Geneva, Switzerland, 2021.
  4. Blamey, P.; Artieres, F.; Baskent, D.; Bergeron, F.; Beynon, A.; Burke, E.; Dillier, N.; Dowell, R.; Fraysse, B.; Gallego, S.; et al. Factors affecting auditory performance of postlinguistically deaf adults using cochlear implants: An update with 2251 patients. Audiol. Neurootol. 2013, 18, 36–47. [Google Scholar] [CrossRef] [PubMed]
  5. Goudey, B.; Plant, K.; Kiral, I.; Jimeno-Yepes, A.; Swan, A.; Gambhir, M.; Buchner, A.; Kludt, E.; Eikelboom, R.H.; Sucher, C.; et al. A MultiCenter Analysis of Factors Associated with Hearing Outcome for 2,735 Adults with Cochlear Implants. Trends Hear. 2021, 25, 23312165211037525. [Google Scholar] [CrossRef] [PubMed]
  6. Lazard, D.S.; Vincent, C.; Venail, F.; Van de Heyning, P.; Truy, E.; Sterkers, O.; Skarzynski, P.H.; Skarzynski, H.; Schauwers, K.; O’Leary, S.; et al. Pre-, per- and postoperative factors affecting performance of postlinguistically deaf adults using cochlear implants: A new conceptual model over time. PLoS ONE 2012, 7, e48739. [Google Scholar] [CrossRef]
  7. Plant, K.; McDermott, H.; van Hoesel, R.; Dawson, P.; Cowan, R. Factors Predicting Postoperative Unilateral and Bilateral Speech Recognition in Adult Cochlear Implant Recipients with Acoustic Hearing. Ear Hear. 2016, 37, 153–163. [Google Scholar] [CrossRef]
  8. Roditi, R.E.; Poissant, S.F.; Bero, E.M.; Lee, D.J. A Predictive Model of Cochlear Implant Performance in Postlingually Deafened Adults. Otol. Neurotol. 2009, 30, 449–454. [Google Scholar] [CrossRef]
  9. Hoppe, U.; Hocke, T.; Hast, A.; Iro, H. Cochlear Implantation in Candidates With Moderate-to-Severe Hearing Loss and Poor Speech Perception. Laryngoscope 2021, 131, E940–E945. [Google Scholar] [CrossRef]
  10. Kim, H.; Kang, W.S.; Park, H.J.; Lee, J.Y.; Park, J.W.; Kim, Y.; Seo, J.W.; Kwak, M.Y.; Kang, B.C.; Yang, C.J.; et al. Cochlear Implantation in Postlingually Deaf Adults is Time-sensitive Towards Positive Outcome: Prediction using Advanced Machine Learning Techniques. Sci. Rep. 2018, 8, 18004. [Google Scholar] [CrossRef]
  11. Crowson, M.G.; Dixon, P.; Mahmood, R.; Lee, J.W.; Shipp, D.; Le, T.; Lin, V.; Chen, J.; Chan, T.C.Y. Predicting Postoperative Cochlear Implant Performance Using Supervised Machine Learning. Otol Neurotol. 2020, 41, e1013–e1023. [Google Scholar] [CrossRef]
  12. Shafieibavani, E.; Goudey, B.; Kiral, I.; Zhong, P.; Jimeno-Yepes, A.; Swan, A.; Gambhir, M.; Buechner, A.; Kludt, E.; Eikelboom, R.H.; et al. Predictive models for cochlear implant outcomes: Performance, generalizability, and the impact of cohort size. Trends Hear 2021, 25, 23312165211066174. [Google Scholar] [CrossRef]
  13. Dziemba, O.C.; Merz, S.; Hocke, T. Evaluative audiometry after cochlear implant provision. HNO 2024, 72, 56–62. [Google Scholar] [CrossRef]
  14. Hoppe, U.; Hast, A.; Hornung, J.; Hocke, T. Evolving a Model for Cochlear Implant Outcome. J. Clin. Med. 2023, 12 (Suppl. 1), 6215. [Google Scholar] [CrossRef]
  15. Heutink, F.; Verbist, B.M.; van der Woude, W.-J.; Meulman, T.J.; Briaire, J.J.; Frijns, J.H.M.; Vart, P.; Mylanus, E.A.M.; Huinck, W.J. Factors Influencing Speech Perception in Adults With a Cochlear Implant. Ear Hear. 2021, 42, 949–960. [Google Scholar] [CrossRef] [PubMed]
  16. Hoppe, U.; Hocke, T.; Hast, A.; Iro, H. Maximum monosyllabic score as a predictor for cochlear implant outcome. HNO 2019, 67, 199–206. [Google Scholar] [CrossRef] [PubMed]
  17. Thangavelu, K.; Nitzge, M.; Weiss, R.M.; Mueller-Mazzotta, J.; Stuck, B.A.; Reimann, K. Role of cochlear reserve in adults with cochlear implants following post-lingual hearing loss. Eur. Arch. OtoRhinoLaryngol. 2023, 280, 1063–1071. [Google Scholar] [CrossRef] [PubMed]
  18. Ollermann, R.; Böscke, R.; Neidhardt, J.; Radeloff, A. External Validation and Extension of a Cochlear Implant Performance Prediction Model: Analysis of the Oldenburg Cohort. Audiol. Res. 2025, 15, 69. [Google Scholar] [CrossRef]
  19. Hastie, T.; Tibshirani, R. Generalized Additive Models. Stat. Sci. 1986, 1, 297–318. [Google Scholar] [CrossRef]
  20. Zou, H.; Hastie, T. Regularization and variable selection via the elastic net. J. R. Statist. Soc. B 2005, 67, 301–320. [Google Scholar] [CrossRef]
  21. Breiman, L. Bagging Predictors. Mach. Learn. 1996, 24, 123–140. [Google Scholar] [CrossRef]
  22. Breiman, L. Random Forests. Mach. Learn. 2001, 45, 5–32. [Google Scholar] [CrossRef]
  23. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  24. Friedman, J. Greedy Function Approximation: A Gradient Boosting Machine. Ann. Stat. 2001, 29, 1189–1232. [Google Scholar] [CrossRef]
  25. Posit Team. RStudio: Integrated Development Environment for R; Posit Software, PBC: Boston, MA, USA, 2025. [Google Scholar]
  26. Anthropic. Claude Sonnet 4.0 (Software). San Francisco, CA, USA. 2024. Available online: https://www.anthropic.com/claude/sonnet?utm_source=chatgpt.com (accessed on 20 October 2025).
  27. Shew, M.A.; Pavelchek, C.; Michelson, A.; Ortmann, A.; Lefler, S.; Walia, A.; Durakovic, N.; Phillips, A.; Rejepova, A.; Herzog, J.A.; et al. Machine Learning Feasibility in Cochlear Implant Speech Perception Outcomes-Moving Beyond Single Biomarkers for Cochlear Implant Performance Prediction. Ear Hear. 2025, 46, 1266–1281. [Google Scholar] [CrossRef] [PubMed]
  28. Demyanchuk, A.; Kludt, E.; Lenarz, T.; Buchner, A. A Machine Learning Model to Predict Postoperative Speech Recognition Outcomes in Cochlear Implant Recipients: Development, Validation, and Comparison with Expert Clinical Judgment. J. Clin. Med. 2025, 14, 3625. [Google Scholar] [CrossRef] [PubMed]
  29. Han, J.J.; Suh, M.W.; Park, M.K.; Koo, J.W.; Lee, J.H.; Oh, S.H. A Predictive Model for Cochlear Implant Outcome in Children with Cochlear Nerve Deficiency. Sci. Rep. 2019, 9, 1154. [Google Scholar] [CrossRef]
  30. Blamey, P.; Arndt, P.; Bergeron, F.; Bredberg, G.; Brimacombe, J.; Facer, G.; Larky, J.; Lindstrom, B.; Nedzelski, J.; Peterson, A.; et al. Factors affecting auditory performance of postlinguistically deaf adults using cochlear implants. Audiol. Neurootol. 1996, 1, 293–306. [Google Scholar] [CrossRef] [PubMed]
  31. Mahmood, F. A benchmarking crisis in biomedical machine learning. Nat. Med. 2025, 31, 1060. [Google Scholar] [CrossRef]
Figure 1. Actual versus predicted CI-outcome across different models using independent train/test splits. The black dashed line represents perfect prediction, and the red line shows the linear fit with confidence bands (grey). In all plots, the x-axis shows the predicted WRS65(CI) [%] and the y-axis shows the actual WRS65(CI) [%]. Blue dots represent the implanted ears in the test set cohort. RMSE = root mean squared error; GLM = generalized linear regression model; GAM = generalized additive model; XGBoost = eXtreme Gradient Boosting; n = 36.
Figure 1. Actual versus predicted CI-outcome across different models using independent train/test splits. The black dashed line represents perfect prediction, and the red line shows the linear fit with confidence bands (grey). In all plots, the x-axis shows the predicted WRS65(CI) [%] and the y-axis shows the actual WRS65(CI) [%]. Blue dots represent the implanted ears in the test set cohort. RMSE = root mean squared error; GLM = generalized linear regression model; GAM = generalized additive model; XGBoost = eXtreme Gradient Boosting; n = 36.
Audiolres 15 00161 g001
Figure 2. Bland–Altman analyses across different model performances. Mean difference and limits of agreement (LoA) were calculated for each model. Blue dots represent the implanted ears in the test set cohort. Dashed red lines represent the upper and lower LoA, while the red solid line shows the mean difference between actual and predicted values, indicating the average of bias of the model. The black dotted line represents the line of no difference (zero bias), indicating a perfect agreement between actual and predicted scores. n = 36.
Figure 2. Bland–Altman analyses across different model performances. Mean difference and limits of agreement (LoA) were calculated for each model. Blue dots represent the implanted ears in the test set cohort. Dashed red lines represent the upper and lower LoA, while the red solid line shows the mean difference between actual and predicted values, indicating the average of bias of the model. The black dotted line represents the line of no difference (zero bias), indicating a perfect agreement between actual and predicted scores. n = 36.
Audiolres 15 00161 g002
Figure 3. Feature importance analysis based on absolute correlation with CI outcome. The features WRSmax, CI experience, age at implantation, WRS65(HA), age at onset of hearing loss (HL), duration of HL and the four-frequency pure tone average (4FPTA) were correlated to the WRS65(CI) to assess the importance of these variables. n = 36.
Figure 3. Feature importance analysis based on absolute correlation with CI outcome. The features WRSmax, CI experience, age at implantation, WRS65(HA), age at onset of hearing loss (HL), duration of HL and the four-frequency pure tone average (4FPTA) were correlated to the WRS65(CI) to assess the importance of these variables. n = 36.
Audiolres 15 00161 g003
Table 1. Summary of patient characteristics.
Table 1. Summary of patient characteristics.
VariableNMedian [Q1, Q3]Range
Duration of hearing loss (years)22217 [9–28]1–80
Age at implantation (years)23663 [54–73]18–88
CI experience (months)22312 [12–24.25]10–140
Preoperative 4FPTA of the implanted ear (dB HL)23588.75 [80–96.25]63.75–112.5
WRSmax (%)23630 [15–50]5–100
WRS65(HA) (%)23615 [0–30]0–60
WRS65(CI) (%)23675 [55–80]5–100
Q1 = 25th percentile, Q3 = 75th percentile.
Table 2. Summary statistics of model performances.
Table 2. Summary statistics of model performances.
ModelRMSEMAER2SDBiasUpper LoALower LoAPb SlopePb InterceptCorrelation
Random_Forest26.2820.62−0.07325.04−9.0140.07−58.080.09461.250.248
GAM26.6621.20−0.10424.52−11.2336.82−59.280.17758.660.327
GLM_Full27.1021.26−0.14124.99−11.2837.71−60.260.07064.870.240
ElasticNet27.2021.23−0.15025.40−10.6339.15−60.410.02766.730.161
GLM_External_
Backward
27.5121.19−0.17625.45−11.2638.63−61.150.04466.380.167
GLM_Backward27.5121.24−0.17625.46−11.2638.64−61.160.04666.270.168
Hoppe_External28.0921.50−0.22625.77−11.9738.55−62.490.06765.730.170
XGBoost28.2322.22−0.23826.77−10.0042.47−62.470.03665.540.082
Ensemble_Top328.2321.97−0.23825.59−12.6637.50−62.830.02069.170.115
Ensemble_Weighted28.3322.05−0.24725.58−12.9137.23−63.040.01969.480.116
Null30.7423.75−0.46825.73−17.3633.07−67.80NaNNaNNA
RMSE = root mean squared error, MAE = mean absolute error, SD = Standard deviation; LoA = limits of agreement, Pb = Passing–Bablok; NaN = Not a Number; NA = Not available; Gray = baseline model, Green = best performing model; n = 36.
Table 3. Improvement in statistical metrics over the Null model.
Table 3. Improvement in statistical metrics over the Null model.
ModelMean RMSE Improvement
[Lower CI, Upper CI]
p-ValueMean MAE Improvement
[Lower CI, Upper CI]
p-ValueMean R2 Improvement
[Lower CI, Upper CI]
p-Value
Random_Forest4.446 [1.395, 7.577]0.0013.117 [−0.151, 6.416]0.0310.406 [0.122, 0.711]0.001
GAM4.063 [0.84, 7.199]0.012.554 [−0.892, 6.17]0.0680.373 [0.084, 0.705]0.01
GLM_Full3.617 [1.012, 5.927]0.0032.454 [−0.39, 5.206]0.0390.339 [0.091, 0.624]0.003
ElasticNet3.51 [1.491, 5.342]0.0012.486 [0.154, 4.831]0.020.331 [0.124, 0.583]0.001
GLM_External_
Backward
3.221 [0.955, 5.35]0.0042.517 [0.022, 5.09]0.0230.305 [0.085, 0.571]0.004
GLM_Backward3.216 [0.872, 5.385]0.0042.468 [−0.045, 5.074]0.0320.304 [0.081, 0.575]0.004
Hoppe_External2.658 [−0.219, 5.369]0.0352.217 [−0.714, 5.078]0.060.253 [−0.018, 0.55]0.035
XGBoost2.504 [−0.394, 5.315]0.0471.498 [−1.954, 4.855]0.1940.236 [−0.041, 0.51]0.047
Ensemble_Top32.502 [0.941, 3.846]01.761 [−0.289, 3.705]0.0430.239 [0.082, 0.42]0
Ensemble_Weighted2.404 [0.94, 3.669]01.682 [−0.257, 3.532]0.0410.23 [0.082, 0.403]0
CI = Confidence interval.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ollermann, R.; Strodthoff, N.; Radeloff, A.; Böscke, R. Machine Learning Versus Simple Clinical Models for Cochlear Implant Outcome Prediction. Audiol. Res. 2025, 15, 161. https://doi.org/10.3390/audiolres15060161

AMA Style

Ollermann R, Strodthoff N, Radeloff A, Böscke R. Machine Learning Versus Simple Clinical Models for Cochlear Implant Outcome Prediction. Audiology Research. 2025; 15(6):161. https://doi.org/10.3390/audiolres15060161

Chicago/Turabian Style

Ollermann, Rieke, Nils Strodthoff, Andreas Radeloff, and Robert Böscke. 2025. "Machine Learning Versus Simple Clinical Models for Cochlear Implant Outcome Prediction" Audiology Research 15, no. 6: 161. https://doi.org/10.3390/audiolres15060161

APA Style

Ollermann, R., Strodthoff, N., Radeloff, A., & Böscke, R. (2025). Machine Learning Versus Simple Clinical Models for Cochlear Implant Outcome Prediction. Audiology Research, 15(6), 161. https://doi.org/10.3390/audiolres15060161

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop