Next Article in Journal
Research on a Joint Extraction Method of Track Circuit Entities and Relations Integrating Global Pointer and Tensor Learning
Previous Article in Journal
Electroencephalography-Based Brain-Computer Interfaces in Rehabilitation: A Bibliometric Analysis (2013–2023)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Artificial Intelligence in Audiology: A Scoping Review of Current Applications and Future Directions

1
Maxillofacial Surgery Unit, Department of Medical Biotechnology, S. Maria alle Scotte University Hospital of Siena, 53100 Siena, Italy
2
Phoniatris and Audiology Unit, Department of Neuroscience DNS, University of Padova, 33100 Treviso, Italy
3
Audiology Program, Otorhinolaryngology Unit, Department of Medical and Surgical Sciences for Children and Adults, University of Modena and Reggio Emilia, 41124 Modena, Italy
4
Audiology Program, Department of Maternal, Child and Adult Medical and Surgical Sciences, University of Modena and Reggio Emilia, 41124 Modena, Italy
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(22), 7126; https://doi.org/10.3390/s24227126
Submission received: 14 September 2024 / Revised: 31 October 2024 / Accepted: 4 November 2024 / Published: 6 November 2024
(This article belongs to the Section Biomedical Sensors)

Abstract

:
The integration of artificial intelligence (AI) into medical disciplines is rapidly transforming healthcare delivery, with audiology being no exception. By synthesizing the existing literature, this review seeks to inform clinicians, researchers, and policymakers about the potential and challenges of integrating AI into audiological practice. The PubMed, Cochrane, and Google Scholar databases were searched for articles published in English from 1990 to 2024 with the following query: “(audiology) AND (“artificial intelligence” OR “machine learning” OR “deep learning”)”. The PRISMA extension for scoping reviews (PRISMA-ScR) was followed. The database research yielded 1359 results, and the selection process led to the inclusion of 104 manuscripts. The integration of AI in audiology has evolved significantly over the succeeding decades, with 87.5% of manuscripts published in the last 4 years. Most types of AI were consistently used for specific purposes, such as logistic regression and other statistical machine learning tools (e.g., support vector machine, multilayer perceptron, random forest, deep belief network, decision tree, k-nearest neighbor, or LASSO) for automated audiometry and clinical predictions; convolutional neural networks for radiological image analysis; and large language models for automatic generation of diagnostic reports. Despite the advances in AI technologies, different ethical and professional challenges are still present, underscoring the need for larger, more diverse data collection and bioethics studies in the field of audiology.

1. Introduction

The integration of artificial intelligence (AI) into medical disciplines is rapidly transforming healthcare delivery, with audiology being no exception. As audiology traditionally relied on manual diagnostic methods and subjective assessments [1], advances in AI have opened new avenues for enhancing diagnostic accuracy, treatment efficiency, and patient outcomes. AI refers to technologies which enable machines to perform tasks typically requiring human intelligence. The history of AI dates back to the mid-20th century, beginning with the development of strict rule-based algorithm systems in the 1950s and 1960s [2]. The field experienced a significant resurgence in the 2000s due to increased computational power and the advent of big data [3]. Key milestones included the development of machine learning (ML) algorithms (a subset of artificial intelligence which allows systems to learn from data and make predictions or decisions without explicit programming), the rise of deep learning (DL) in the 2010s, which is any type of ML based on artificial neural networks with multiple layers, and more recently the emergence of generative AI and large language models (LLMs) [4]. In fact, AI includes a broad range of technologies, each with distinct capabilities and applications. ML is a fundamental subset of AI which focuses on developing algorithms enabling computers to learn from and make predictions based on data [5]. Unlike traditional programming, which relies on explicit instructions, ML systems enhance their performance through iterative exposure to data [6]. As more data are processed, these systems refine their algorithms to improve their accuracy and efficiency in tasks ranging from pattern recognition to predictive analytics. Within the realm of ML, DL stands out as a powerful technique utilizing artificial neural networks with multiple layers; hence the term “deep” [2]. These networks can model complex patterns in large datasets, making DL particularly successful in image and speech recognition tasks [7]. The ability to automatically learn hierarchical feature representations allows DL to excel in identifying intricate patterns and correlations which might be missed by simpler algorithms. Generative AI represents another innovative branch of AI, involving algorithms designed to create new contents such as text, images, or music by learning from existing data [8]. Techniques such as generative adversarial networks (GANs) and variational autoencoders (VAEs) are central to this field [9]. These methods have shown remarkable potential in applications like medical image synthesis, where they can generate high-quality images for training purposes, and in drug discovery, where they can create novel molecular structures. LLMs like GPT-4 (OpenAI, San Francisco, CA, USA) epitomize the advanced capabilities of AI in understanding and generating human-like text [10]. Trained on extensive corpora of text data, LLMs are adept at tasks such as language translation, summarization, and conversational interaction. They employ DL techniques to process and generate text which is coherent and contextually relevant, making them invaluable tools in fields requiring sophisticated language understanding and generation. Although these models showed promising applications across various medical specialties [11,12], the evaluation of healthcare information provided by generative AI platforms remains a topic of debate and requires careful consideration [13]. In audiology, AI applications range from enhancing the accuracy of hearing assessment to automating hearing aid and cochlear implant (CI) fitting, as well as optimization of auditory rehabilitation processes [14,15,16,17]. The deployment of AI in these areas promises to (1) improve the capabilities of audiologists, (2) reduce the burden of routine tasks, and (3) deliver personalized patient care. Despite the promise of AI, its integration into audiology also raises several challenges. These include data privacy concerns, the need for large datasets for training algorithms, potential biases in AI models, and the requirement of rigorous validation before clinical implementation [18]. Furthermore, many audiologists lack awareness of the potential applications of AI in their field [1]. Addressing these challenges is crucial to realize the full potential of AI in audiology.
This narrative review aims to provide a comprehensive overview of the current state of AI applications in audiology. We will explore the technological advancements and discuss the benefits and limitations of AI in audiology. By synthesizing the existing literature, the review seeks to inform clinicians, researchers, and policymakers about the potential and challenges of integrating AI into audiological practice.

2. Materials and Methods

A comprehensive literature search was conducted to identify relevant studies and reviews on the application of AI in audiology. The PubMed, Cochrane, and Google Scholar databases were searched on 5 July 2024 with the following query: “(audiology) AND (“artificial intelligence” OR “machine learning” OR “deep learning”)”. The search was limited to articles published in English from 1990 to 2024 to capture the most recent advances in the field. The PRISMA extension for scoping reviews (PRISMA-ScR) checklist was considered to ensure quality of the review [19]. Articles were selected if they primarily focused on the application of AI in audiology. This investigation included studies which provided empirical data on the effectiveness, accuracy, or utility of AI technologies in the context of audiology practices from diagnosis to treatment. Additionally, reviews and meta-analyses which discussed the integration of AI within the audiology field were critically analyzed, given their comprehensive insights and evaluation of multiple studies. Studies which lacked sufficient data or clear methodology were excluded, as these did not provide reliable or verifiable results.
Data extraction was performed independently by three reviewers (V.C., L.F. and A.F.) to ensure accuracy and comprehensiveness. The information extracted included the study design, AI technology used, audiology application, measured outcomes, and key findings. Discrepancies between reviewers were resolved through discussion and consensus. The extracted data were then analyzed to identify the study design, population, intervention, comparison, and outcome. Studies were categorized based on the type of AI technology (ML, DL, or generative AI), clinical field (hearing and balance), and AI application (e.g., diagnosis and treatment). Qualitative result synthesis was performed.

3. Results

The database research yielded 156 results from PubMed, 1200 from Google Scholar, and 3 from Cochrane. The selection process led to the inclusion of 104 manuscripts [1,16,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54,55,56,57,58,59,60,61,62,63,64,65,66,67,68,69,70,71,72,73,74,75,76,77,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92,93,94,95,96,97,98,99,100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116,117,118,119,120,121], as shown in Figure 1.
The integration of AI in audiology has evolved significantly over the succeeding decades. An early pioneer study by Bonadonna [20] laid the initial groundwork for AI applications in medical audiology, marking the inception of this innovative research area. During the last decade of the second millennium and the first decade of the third millennium (1990–2009), there were only three investigations [20,21,22], representing approximately 2.9% of the total studies reviewed. The second decade of the third millennium (2010–2019) saw the establishment of more sophisticated AI models, improved computational power, and greater data availability. Accordingly, we retrieved a substantial increase, with 10 publications (9.6% of our sample) focused on developing ML algorithms and integrating AI technologies into audiology’s diagnostic and therapeutic tools. The growing trend continued into the third decade (2020–2024), which showed a prolific output of 91 studies from 2020 to 2024, representing approximately 87.5% of the total publications. This period has been characterized by the application of DL, the rise of large language models, and a focus on enhancing diagnostic accuracy, treatment efficiency, and patient outcomes. The three most active countries in this field have been the United States, Germany, and China. The United States contributed to 26 of the reviewed studies, including advancements in hearing tests, diagnostic tools, and therapeutic applications [24,25,29,35,36,37,38,43,45,46,55,60,70,81,82,85,86,96,97,101,104,115,120]. Germany followed with 14 contributions, focusing in particular on ML models and clinical decision support systems [30,33,40,42,51,53,62,63,64,65,73,93,95]. Contributing to 12 studies, China has made significant strides in diagnostic applications using DL and predictive modeling [50,59,78,83,87,105,110,116,117,118,119], as depicted in Figure 2.
Table 1 illustrates the most frequently retrieved journals and the published studies’ designs. Ear and Hearing was the leading journal with 13 publications [24,29,32,35,36,37,50,55,57,69,86,96,97] followed by International Journal of Audiology with 6 [26,30,32,33,42,72], both showcasing significant contributions to the field through various investigations of hearing tests, diagnostics, and therapeutic tools. With four publications, Frontiers in Digital Health reflected the growing multidisciplinary interest in integrating digital health technologies with AI in audiology [43,51,73,115]. Observational studies were the most prevalent study design (68%), followed by development and validation studies (16.3%) and reviews (8.7%).

Application Fields and Technical Approaches

The application of AI in audiology spans various fields, reflecting the technology’s versatility and potential impact. The most prevalent application was in therapeutic and prognostic tools, accounting for 34.6% of the studies reviewed. These studies focused on using AI to enhance treatment plans and predict patient outcomes, as exemplified by investigations such as those by Jin et al. (2023) and Doborjeh et al. (2023) [84,86]. The hearing test field was close behind, comprising 33.7% of the research, including significant contributions from groups such as Barbour et al. (2019) and Heisey et al. (2020) [29,37], who explored AI-driven methods to improve the accuracy and efficiency of hearing assessments. The miscellaneous category included 21.1% of the studies, covering a range of unique applications and exploratory audiological research areas. The fields of temporal bone radiology and vestibular diagnosis each accounted for 5.8% of the studies. Research in temporal bone radiology, such as the investigation by Masino et al. [25], focused on applying AI to analyze complex oto-radiology imaging data. Meanwhile, studies in vestibular diagnosis, such as those by Kong et al. [87], applied AI to improve diagnosis and treatment of vestibular disorders. Specific features and results of different included categories of investigations will be covered in the following paragraphs.
ML encompasses a diverse array of algorithms, each suited to specific tasks. Choosing the right algorithm is crucial in data analysis and predictive modeling. In the subsequent sections, we explore various representative algorithms, delving into their mechanisms and practical applications in audiology with the aim to enhance understanding of the intricate relationship between function choice and data analysis.
Gaussian process regression (GPR) and Gaussian process classification (GPC) are probabilistic modeling techniques employed for supervised learning and grounded in Gaussian processes (GPs), which are a powerful framework for defining distributions over functions. A GP is essentially a collection of random variables, any finite number of which have a joint Gaussian distribution. This allows GPs to model the uncertainty and variability in the relationship between input and output data in a flexible, non-parametric way [29]. GPR is used for regression tasks, predicting continuous outputs given input data [115]. It models the relationship between inputs and outputs by placing a GP prior over possible functions, leading to predictions which include both a mean estimate and a measure of uncertainty [24]. Instead, GPC is applied to classification tasks, where the outputs are discrete classes. In GPC, the GP models a latent function—an unobserved or hidden function which describes the relationship between the input data and some underlying variable—which is then mapped to class probabilities via a link function, such as a sigmoid function, an unobserved or hidden function which describes the relationship between the input data and some underlying variable [29]. Both approaches benefit from the inherent ability of GPs to quantify uncertainty, making them valuable for tasks where understanding the confidence in predictions is important. Accordingly, they have recently been employed to develop ML techniques for automated hearing threshold tests [24,29].
Unlike standard regression, which minimizes the difference between the predicted and actual values, least absolute shrinkage and selection operator (LASSO) regression adds a penalty term which constrains the coefficient size and promotes sparsity by setting some coefficients to zero [5]. This process selects the most important variables, reduces overfitting, and improves model accuracy and generalizability, making LASSO a valuable tool for creating robust and interpretable predictive models [47]. This technique was encountered in the research line by the Carl von Ossietzky Universität Oldenburg (Oldenburg, Germany) [30,33,40,73]. Specifically, Saak et al. [40] used ML models to predict a cluster of previously defined common audiological functional parameters (CAFPAs) [30]. The CAFPAs—describing functional aspects of the human auditory system (including the hearing threshold, suprathreshold deficits, binaural hearing, neural processing, cognitive components, and socioeconomic status)—were empirically instantiated by an expert survey conducted on a large dataset of audiological measures [33]. Saak et al. [73] had ML models trained on the same dataset, aiming to improve the objectivity and precision of diagnostic decisions by automating the estimation of CAFPAs from patient data. LASSO regression was utilized to identify the most relevant audiological measures influencing specific CAFPAs [73].
The k-nearest neighbors (k-NN) algorithm is intuitive and straightforward, making it suitable for both classification and regression tasks. It classifies new data points based on their similarity to the k-nearest neighbors [122]. In classification, the most common class among the k neighbors determines the new data’s class, while in regression, predictions are based on the average values of these neighbors [50]. Distance metrics such as the Euclidean or Manhattan distance define the term “nearest” [122]. In medical applications, normalization is crucial to ensure variables contribute proportionately to distance calculations, enhancing accuracy and reliability [50]. Choosing the right k value is vital; too small k value can make the model sensitive to noise, while too large k value can oversimplify the model [43]. Although k-NN is easy to implement, it can be sensitive to outliers and computationally intensive for large datasets [50]. The k-NN and artificial neural network (ANN) models were used by Szaleniec et al. [122] to predict hearing improvement after tympanoplasty surgery in patients with chronic suppurative otitis media. The study involved 150 patients, characterized by variables such as age, gender, preoperative audiometric results, ear pathology, and surgical procedure details. The k-NN model, optimized using a 10 fold cross-validation method with the City block distance metric, achieved a prediction accuracy of 75.8% for the validation and test sets. Additionally, the best-performing ANN model demonstrated superior performance, achieving 98% accuracy in the training set, 89% accuracy in the validation set, and 84% accuracy in the test set [122].
A support vector machine (SVM) is used to define the boundaries between data points, treating them as p-dimensional vectors [22]. The goal is to construct a (p-1)-dimensional hyperplane which maximizes the margin from the nearest data vectors, effectively separating different classes [48]. SVMs can handle both classification and regression tasks, making them versatile [25]. SVMs aim to find the global optimal solution, capturing complex relationships within the data. Therefore, they are widely used in medical research. Recently, Rodrigo et al. [45] used the SVM model to predict the outcomes of Internet-based cognitive behavioral therapy (ICBT) for tinnitus. The SVM model, including both linear and radial basis kernel variations, was used to classify treatment success based on a 13-points reduction in the Tinnitus Functional Index (TFI). A study by Rodrigo et al. [45] involved secondary analysis of data from 228 individuals who had completed ICBT with 33 predictor variables, including demographic, tinnitus, hearing-related, and clinical factors. The SVM models were compared with ANN models, and their performance was evaluated using the mean predictive accuracy and area under the receiver’s operating characteristic curve (AUC). Although the ANN model showed the highest predictive accuracy (mean AUC of 0.73), the SVM models also demonstrated adequate discriminative power, with the linear kernel SVM achieving an AUC of 0.72 and the radial basis kernel SVM reaching an AUC of 0.70.
Decision tree (DT) algorithms create hierarchical models with tree-like structures, where nodes represent decision points and branches represent outcomes [28]. This structure is interpretable and systematic, akin to a flowchart [21]. DTs evaluate input features, making decisions at each node to predict the target output. They are versatile, being applicable to both classification and regression tasks [25]. DTs were proposed early to systematically and interpretably classify oto-neurological diseases based on patient symptoms and history, achieving a high accuracy (90% on a database of 564 patients) but facing limitations with insufficient input parameters for certain conditions (e.g., sudden sensorineural hearing loss) [21]. DTs often serve as the basis for ensemble learning methods like random forest (RF) and extreme Gradient Boosting (XGBoost), as discussed below [28].
RF and random forest regression (RFR) combine multiple decision tree algorithms arranged in parallel, enhancing model robustness through ensemble learning. In clinical applications, their use of random inputs and features for each tree results in low correlation between individual trees, effectively capturing diverse patterns in the data [64]. RF and RFR do not require variable normalization, thus accommodating the variability in clinical data [43]. They are valuable for exploring feature importance, and their ability to handle diverse data types, resist overfitting, and provide interpretable insights makes RF and RFR a preferred choice in clinical research and predictive modeling [25]. Hoppe et al. (2022) applied the RFR model to analyze the relationship between age-related hearing loss and speech recognition decline on a large clinical dataset of 19,801 ears. The RFR model provided more specific information about the time course and amount of degradation in speech recognition, yielding smaller mean absolute errors (MAEs) compared with the generalized linear model (GLM). However, the RFR model indicated some degree of overfitting, as evidenced by the differences between the training and test group MAEs. The above-mentioned study found that speech scores varied based on the specific type of hearing loss and the decade of life. Speech recognition deteriorated by up to 25 percentage points over the entire lifespan for constant pure-tone thresholds, with the most significant decline being 10 percentage points per decade [64].
XGBoost constructs an ensemble learning model by adding DTs sequentially. Each new tree corrects the mistakes of its predecessors, creating a highly adaptive model [81]. XGBoost effectively captures complex relationships within data, making it suitable for intricate patterns. However, it is sensitive to variable selection, affecting performance significantly [114]. Despite this, when variables are chosen carefully, XGBoost outperforms other ensemble learning algorithms in predictive accuracy [102]. The investigation by Balan et al. [81] is an appropriate example. The authors applied various ML models to analyze how hearing thresholds could predict speech-in-noise recognition among individuals with normal audiograms. Utilizing archival data of hearing thresholds (0.25–16 kHz) and speech recognition thresholds (SRTs) from 764 participants, XGBoost outperformed the other models—the ANN, deep neural network (DNN) and RF—with an MAE of 1.62 dB. The ANN and RF models showed comparable performances (MAE = 1.68 and 1.67 dB, respectively), while the DNN exhibited poorer performance (MAE = 1.94 dB) [81]. Moreover, Balan et al. [81] highlighted the significant contributions of age and high-frequency thresholds (16 kHz and 12.5 kHz) to the SRT, underscoring the relevance of extended high frequencies in predicting speech-in-noise recognition in individuals with normal hearing.
Inspired by the anatomical model of the human brain, an ANN—more often referred to as just a neural network (NN)—consists of interconnected nodes (the fundamental computational units (i.e., neurons)) which process information through weighted connections (i.e., synapses) [2]. Nodes are categorized into input, hidden, and output layers. Input nodes handle real-world data, hidden nodes process and transform inputs, and output nodes represent predictions [32]. NNs are powerful tools for predictive modeling and ML due to their ability to capture relationships within data [31]. DL, a subset of NNs, incorporates multiple hidden layers, enabling it to handle complex, nonlinear data effectively. DL is particularly adept at processing image data, where convolutional neural networks (CNNs, which are specific DL algorithms able to analyze an input image, assign importance to objects within the image, and differentiate them from one another) are commonly used. CNNs recognize intricate patterns in images, combining features into connected layers for comprehensive analysis. DL’s ability to handle complex datasets has led to its adoption in fields requiring sophisticated analysis techniques [36]. McKearney et al. [32] utilized a deep CNN to classify paired auditory brainstem response (ABR) waveforms into three categories: “clear response”, “inconclusive”, and “response absent”. This innovative approach involved training the CNN on 190 paired ABR waveforms using stratified 10 fold cross-validation and evaluating it on a separate test set of 42 paired waveforms. The CNN architecture included convolutional, pooling, and fully connected layers optimized to detect complex patterns in the ABR data. The network achieved a high classification accuracy of 92.9%, with sensitivity and specificity values of 92.9% and 96.4%, respectively. The study highlighted the potential of DL models to assist clinicians in interpreting ABR waveforms, ultimately improving the consistency and accuracy of hearing threshold estimations [32].
Generative AI refers to algorithms which create new outputs, such as text, images, or music, by learning from existing data, with applications ranging from creative content generation to medical application [108]. Generative AI, particularly large language models (LLMs), represents a significant advancement in AI [1]. LLMs have been trained on vast amounts of text data, enabling them to generate human-like text based on given prompts [11]. These models, such as OpenAI’s GPT-4, have had a disruptive impact and dissemination in the last few years and have shown remarkable capabilities in understanding and generating natural language, making them useful for a wide range of applications [4]. In audiology, the reported use of LLMs encompassed from passing qualification exams to supporting clinical practice and patient education [108,117]. Wang et al. [117] evaluated ChatGPT-4’s performance in the Taiwan Audiologist Qualification Examination. ChatGPT-4 was tasked with answering multiple-choice questions across six subjects, achieving an overall accuracy of 75%, surpassing the passing criterion of 60%. Quite recently, Jedrzejczak et al. [108] longitudinally assessed ChatGPT version 3.5’s effectiveness in providing information on and support for tinnitus. ChatGPT was presented with a set of questions related to tinnitus, and the responses were assessed for accuracy by experts. The study found that ChatGPT’s responses were generally satisfactory and improved over time, but they also identified some limitations, such as the lack of accurate references and occasional misleading information.

4. Discussion

The integration of AI in audiology is transforming various aspects of auditory and vestibular health management. In order to systematically explore the advances in these fields, we provided five dedicated subsections of the discussion, focusing on hearing tests, vestibular diagnostics, temporal bone radiology, and therapeutic and prognostic applications.

4.1. Hearing Tests

AI is significantly impacting auditory examination practices, particularly in screening, diagnostic, and surgical methods. As traditional gold standard tests for hearing impairment (HI) diagnosis, such as pure-tone hearing tests, are not feasible for large-scale community implementation, AI has been proposed for the development of HI screening text [118]. As a consequence, ML-based HI screening tests could be suitable for the general population and useful for primary care settings [118]. Risk questionnaires have been effectively created using multiple ML algorithms in order to analyze different variables, including patients’ clinical symptoms and hematological test results [118].
To streamline patients’ selection for audiometric evaluation, a separate ML model was also developed to predict the speech-frequency pure-tone average (PTA) based on patients’ demographics, clinical factors, and subjective hearing status. This supervised ML model employed a tree-based architecture to identify individuals needing audiometric testing, thereby optimizing the screening process [85]. For quantitative prediction of hearing thresholds, stimulus-frequency otoacoustic emissions (SFOAEs) elicited by swept tones was also proposed. In this regard, it has been demonstrated that SFOAEs were effective in predicting hearing thresholds and further enhanced through the application of ML models [50]. A stacked ensemble detection method also showed greater performance compared with traditional ABR detection techniques using ML algorithms [69]. It has also been shown that dynamically masked audiograms achieved accurate true threshold estimates and reduced test times compared with current clinical masking procedures [37]. With regard to diagnostic applications, ML employment has primarily focused on the analysis of tympanic membrane images. Advanced DL methods have enabled automatic diagnosis of otitis media based on wideband tympanometry measurements, enhancing the diagnostic capacity [75]. In this context, different studies suggested that AI was able to detect conditions such as tympanic perforations and otitis media and had a significant accuracy score, ranging from 93.67% to 97.9% [59,78].
In terms of treatments, AI has been applied in order to support surgeon indications. In particular, ML was used to assess the effectiveness of Vibrant Soundbridge (VSB) surgery for conductive or mixed hearing loss [47]. Data suggest that ML enhanced the predictive accuracy regarding speech discrimination scores after VSB surgery [47]. AI was also used to predict adult CI appropriate indication using demographic data and standard behavioral audiometry [101]. ML techniques were also used in order to predict the preservation of residual acoustic hearing in patients receiving CIs, enhancing shared clinical decision making and patient outcomes [120]. AI has been used to enhance autonomous CI fitting. In this regard, psychoacoustic self-testing could improve both the fitting process and follow-up care for patients [39]. Moreover, an AI-assisted CI mapping model showed improvements in audiological outcomes, demonstrating comparability or superiority to manual approaches regarding hearing performance and patient comfort [96,97].

4.2. Vestibular Diagnostics

AI has also been applied in vestibular disorder assessment. In this regard, AI was used to recognize nystagmus patterns using DL in order to aid physicians in classifying such conditions [87]. Moreover, various ML algorithms were demonstrated to be effective; they achieved an accuracy rate of 94.53% in distinguishing recurrent vertigo types [116]. In order to screen patients earlier, questionnaire-based ML models were also created, predicting common vestibular disorders. Nonetheless, preliminary data suggested the need for objective evidence alongside patient-reported history to improve diagnostic accuracy [99]. Moreover, AI was used to evaluate vestibular organ functioning in the presence of acute or chronic vestibular disorders. In particular, utricular function was assessed in patients with Meniere’s disease, suggesting that AI might be suitable to support diagnosis and target therapies [82]. In terms of treatment, predictive analysis for vestibular schwannoma management showed an accuracy value of approximately 80% using a simple decision tree model [52]. Conversely, Heman-Ackah et al. [106] found 90.5% accuracy in predicting facial nerve injury prognoses after surgery for vestibular schwannoma [106]. Figure 3 reports the current state of the art regarding integration of AI in audiology and vestibology.

4.3. Temporal Bone Radiology

4.3.1. AI Models in Temporal Bone Radiology: General Principles

Given the intrinsic complexity of temporal bone anatomy in both physiological and pathological conditions, a comprehensive interpretation of the radiological images of this anatomical district often requires specific training and experience. As a result, the application of AI paradigms to temporal bone radiology represents a rapidly expanding field [123]. Generally speaking, the AI systems applied to the radiology setting start from a wide range of input data (including digital radiological images, clinical data, and demographics) and produce as an output a categorization of the tested cases based on different machine analysis approaches (including supervised learning, unsupervised learning, semi-supervised learning, and reinforcement learning) [124,125]. Supervised learning refers to the use of labeled data (consisting of input features, such as risk factors for specific diseases, clinical data, and demographics paired with corresponding outputs, such as clinical manifestations or diagnosis) in training an ML model [125]. Unsupervised learning involves the use of unlabeled data (with no specific feedback or target variable) to identify patterns of variable distributions, thus allowing tasks such as clustering, dimensionality reduction, and anomaly detection [125,126]. Semi-supervised learning combines elements of both supervised and unsupervised learning (leveraging the smaller amount of labeled data along with a large amount of unlabeled data) to improve the learning process in situations where obtaining labeled data is limited, expensive, or time-consuming [127,128]. Reinforcement learning refers to a trial-and-error training process involving positive or negative feedback in which constant interaction with the environment urges the AI system to choose the appropriate actions to achieve rewarding results [129]. Reinforcement learning approaches include Q-Q-learning, deep Q-Q-networks (DQNs), and policy gradient methods [129].
The applications of AI in temporal bone radiology are mainly based on either supervised or unsupervised learning, with the latter gaining increased interest more recently [123]. The main applications of AI in this field are automated segmentation of temporal bone anatomical structures and assisted diagnosis of middle and inner ear diseases.

4.3.2. Automated Temporal Bone Image Segmentation

Image segmentation refers to the process of partitioning a digital image into multiple segments by identifying and labeling discrete sets of pixels or voxels, representing objects of interest which can be individually analyzed [130]. In temporal bone imaging, the possibility to individually identify anatomical structures, obtaining a quantitative evaluation of their morphological and volumetric parameters, may be valuable support in identifying pathological conditions and anatomical variants [131]. In segmenting the 3D bony structures of the skull, manual and semi-automated tools have long been employed to extract the 3D volume and quantitative information [132]. More recently, the spread of AI-powered technologies has prompted the development of automated segmentation tools specifically for temporal bone anatomic structures [133,134]. The application in medical images of convolutional neural networks specific to image classification and analysis, such as U-Net, allowed the development of automated tools to identify and segment inner and middle ear structures in either RM or CT images with a limited amount of training data [133,134,135]. By analyzing 944 MRI images as a training dataset and 99 for validation, Vaidyanathan et al. [133] aimed to apply U-Net to implement an automatic segmentation system specifically for inner ear structures, including the cochlea and labyrinthic structures, with high performance (true positive rate = 91.5%; false discovery rate = 14.8%; false negative rate = 8.49%). Similarly, Wu et al. [134] and Heutink et al. [135] developed U-Net-powered automated segmentation tools based on CT images specifically for identification of the semicircular canals and cochlea, reporting high accuracy levels (Dice coefficients ≥0.90). Aside from the possibility of in-house development of specific AI-assisted tools, commercially available advanced image analysis software such as Materialise Mimics version 20.0 (Leuven, Belgium: Materialise NV) may also offer the possibility of implementing automated segmentation [131]. When training the Materialise Mimics software with 60 annotated CT scans, Ke et al. [131] reported reliable automatic segmentation of temporal bone structures, obtaining correct identification of facial nerves, ossicles, inner ear (including the cochlea, vestibule, and semicircular canals), internal auditory canal, internal carotid artery, jugular bulb, and external auditory canal in most of the validation images.

4.3.3. AI and Radiological Imaging of Middle Ear Diseases

Moving from technical to clinical aspects, the application of AI models to temporal bone radiology has shown promising results in the field of middle ear disease diagnosis. In this setting, differentiating chronic suppurative otitis media from cholesteatoma is a challenging task, often requiring multi-modal imaging strategies (including both CT and MRI with specific diffusion-weighted sequences) [136,137]. Although still limited to experimental settings, the attempts to employ AI-based image classification models to reduce the need for multiple exams have shown promising results. In 2020, Wang et al. [128] proposed a deep learning framework specifically for the diagnosis of chronic otitis media and cholesteatoma based on temporal bone computed CT scans. A dataset of 975 labeled CT scans was used for training, while the deep learning framework contained two distinct networks: one to extract regions of interest from two-dimensional CT slices and the other to classify images into diagnostic groups based on the extracted regions scans. Such a system showed higher overall accuracy and recall rates in identifying chronic suppurative otitis media and cholesteatoma compared with clinical experts [128]. Using a convolutional neural network architecture trained to perform two consecutive classification tasks on CT images, similar results were found by Chen et al. [138] when discriminating normal versus pathological cases and identifying cholesteatoma versus chronic suppurative otitis media. Such a model showed a substantial gain in identifying cholesteatoma compared with the diagnostic performances of resident fellows and attending otologists [138]. Aside from the automatic image analysis setting, AI has been raising interest as a tool to facilitate the extraction and classification of anatomic and diagnostic information contained in radiology reports. In this intriguing field, Masino et al. [25] proposed an ML system based on natural language processing libraries trained with 726 radiology reports (mainly from temporal bone CTs). In this investigation, they achieved an overall good accuracy in classifying reports by both free-text keywords and ICD-9 terms [25].

4.3.4. AI and Radiological Imaging of Inner Ear

In temporal bone radiology, the need to reliably detect extremely thin anatomical structures on CT and MRI images has raised the issue of optimizing the spatial resolution limit, possibly with no increased image noise. During the last decade, iterative CT reconstruction algorithms allowed the reduction of noise [139]. As a further evolution, the introduction of DL tools has allowed the implementation of more effective image reconstruction algorithms, aiming to obtain substantial noise reduction and, at the same time, improve the spatial resolution [140]. Fujita et al. [141] proposed a deep learning reconstruction algorithm specifically for high-resolution CT images of the temporal bone. They reported a significant reduction in image noise obtained with the deep learning-based reconstruction algorithm, compared with the iterative CT reconstruction approaches. In particular, the depiction of the otic capsule, auditory ossicles, and tympanic membrane was significantly improved in images reconstructed with the deep learning algorithm [141]. Aside from the technical aspects, the possibility to automatedly identify and morphologically classify anatomical structures on temporal bone CT or MRI images, provided by the development of AI-based technologies, has prompted the application of such approaches to the field of inner ear radiology. Ogawa et al. [142] described an unsupervised deep learning system based on a 3D variational autoencoder for detecting and localizing inner ear abnormalities in CT images. They trained their system with unlabeled CT images of a normal subject and used both malformation and normal cases for the test. The unsupervised deep learning system showed good diagnostic performance in identifying cases of malformation, with a specificity of 92% and a sensitivity of around 99% [142].

4.4. Therapeutic and Prognostic Tools

During the last decade, accumulating evidence has shown the possible application of AI technologies in supporting clinical decision making by providing clinical experts with probabilities for medical findings or diagnoses which are based on large amounts of patient data [143]. In the field of audiology, the aim of AI-based decision-making support systems is to characterize patients’ clinical profiles, provide presumptive diagnoses, and suggest appropriate treatment or rehabilitation to compensate for functional impairment [62,63]. Such concepts may be applied in clinical practice as the employment of specific AI tools for counseling or self-counseling, therapeutic guidance, and prognostic assessment.

4.4.1. Expert Systems for Counseling and Peer Support in Chronic Audiological Diseases

One of the first applications of AI in clinical management of chronic audiological diseases was the development of expert systems based on inference engines for peer support purposes [23]. In the management of chronic diseases, many forms of peer support have been explored (including person-to-person support, telephone calls, and Internet-based support), with the goal of giving and receiving help founded on the key principles of mutual agreement of what is helpful [144]. AI allows the development of computer-based solutions for self-assessment of symptom severity which can be directly used by patients. Rasku et al. [23] developed an inference engine trained via the collection of necessary and supportive clinical data to profile the severity of Ménière’s disease, as well as the quality-of-life impact in terms of hearing loss, tinnitus, and vertigo. Similarly, Pyykkö et al. [26] reported the impact of an Internet-based, AI-driven peer support program for Ménière’s disease on 740 patients, with 78% of recent onset cases as well as 55% of those with chronic disease rating the program as useful or very useful.

4.4.2. AI-Assisted Therapy and Hearing Rehabilitation

In the audiological field, the application of AI to support therapeutic decisions may be useful, especially in cases in which defining reliable predictive factors of treatment response is intrinsically difficult, such as in patients with tinnitus [45,63]. Regarding this specific setting, in 2021, Schlee et al. [53] proposed some criteria to develop a clinical decision support system specific to tinnitus treatment based on medical, epidemiological, audiological, electrophysiological, genetic, and clinical subtyping data [53]. According to the authors, such a specific clinical decision support system should be able to suggest an optimal treatment strategy for each individual patient, based on the specific input data and in consideration of the previous training dataset [53]. In line with this, Doborjeh et al. [84] proposed the use of a convolutional neural network approach based on electroencephalographic data to identify patients who could most likely benefit from sound therapy. In such an investigation, the prediction accuracy of the convolutional neural network system was 99.07% for the non-responder group and 98.86% for responders [84]. Also, in the field of non-pharmacological approaches to tinnitus, Yin et al. [119] implemented a machine learning model based on the knowledge graph method to identify the likelihood of response of patients with tinnitus to traditional Chinese medicine, based on clinical features and aspects derived from traditional Chinese semiology. According to the authors, such an AI model achieved high prediction performances (99.4% accuracy, 98.5% sensitivity, 99.6% specificity, and 98.7% precision, with an area under the receiver operating characteristic curve of 99%) across 253 test patients [119]. Aside from identification of the predictive factors for treatment purposes in chronic conditions, AI approaches have also been proposed to optimize hearing rehabilitation. In this sense, one of the first applications was AI-based cochlear implant fitting. In 2018, the clinical trial “Programming Cochlear Implant with Artificial Intelligence” [27], based in Belgium, aimed to compare the results in terms of hearing thresholds and speech discrimination of cochlear implants fitted via the AI-powered software FOX® (2G version) with those fitted manually. The first explorative study published from this trial [145] preliminarily showed that patients initially fitted with a traditional approach may experience an improvement in their auditory results when submitted to AI-assisted fitting. In a subsequent study by the same group, a significant improvement in the pure-tone audiometric threshold at 6000 Hz, phonemic discrimination scores, and soft intensity to normal-intensity speech audiometric scores was found after AI-based cochlear implant fitting [96]. The same authors [97] found over a population of 24 patients who received their first cochlear implant, who were randomly assigned to either the manual or FOX®-assisted fitting arms, less variability and significantly better speech intelligibility in the experimental group. However, in some cases, the participants reported preferring the manual map because it felt more comfortable, even if the FOX map gave better measured outcomes [97].

4.4.3. AI-Based Prognostic Models

The possibility of managing large data amounts, while also identifying in a hypostasis-free statistical setting the patterns of variable distributions, has led to many AI-based analysis approaches being increasingly employed to identify prognosticators in many medical fields [11,12]. Audiology is no exception among the clinical fields potentially involved in this use of AI. One of the possible applications of AI technology in prognostic assessment regards estimation of the functional hearing correlates of certain clinical conditions. Gathman et al. [85] proposed a machine learning model to predict PTA based on anamnestic data from the patients [85]. Such a system was trained with demographic, medical, and subjectively assessed hearing data labeled with hearing thresholds, and it was able to predict the PTA with a mean absolute error of 5.29 dB [85]. In 2022, Zeng et al. [78] developed a deep learning model based on otoscopic images to estimate the related conductive hearing loss degree. The model was trained on 2232 otoscopic images and validated on 558 images, showing promising performance in predicting conductive hearing loss, with an area under the ROC curve of 0.74 and an accuracy of 81% [78]. In the prediction setting, AI-based models have also been proposed to estimate the functional prognoses of specific diseases. For example, in sudden sensorineural hearing loss, several machine learning models have been proposed to estimate the residual functional outcome, with most of them based on logistic regression, although especially in recent years, other AI-based statistical approaches have been employed, including support vector machines, multilayer perceptron, random forest, deep belief networks, decision tree, k-nearest neighbor, and least absolute shrinkage and selection operator (LASSO) [16]. AI-based predictive models have also been proposed as valuable tools in the prevention and early diagnosis setting. Regarding the prevention of occupational hearing loss, Madahana et al. [67] developed and described a machine learning model based on logistic regression, support vector machine, decision tree and random forest classifier algorithms to classify in real time the levels of noise recorded by sensors worn by mine workers and generate warning reports. Such systems showed good performance in terms of detecting noise exposure, analyzing signals, and generating recommendations to the workers [67]. Finally, many predictive AI-based models have been proposed to estimate the functional yield of cochlear implants [54]. In 2024, Carlson et al. [101] proposed a machine learning model to predict speech discrimination scores after receiving a cochlear implant based on preoperative audiometric and basic demographical data. Such a model achieved an accuracy of 87% in predicting final speech discrimination performances, with a sensitivity of 90% and precision of 80% [101]. Moreover, AI has also been employed to predict the association between brain cross-modal plasticity patterns in deaf patients and the functional outcomes of cochlear implants. In 2021, Kyong et al. [48] tried to predict via a machine learning approach the outcome of cochlear implants based on patients’ brain cross-modal plasticity, measured using event-related responses. In their pilot study, based on the machine learning classification outputs, the authors found that tactile and visual central processing patterns best classified the cochlear implant outcomes, with an accuracy of 98.83 ± 2.57% and 93.50 ± 4.89%, respectively [48].

4.5. AI-Driven Augmented Sensors in Audiology

Audiological prostheses, including CIs and HAs, have progressively evolved over the past five decades, transitioning from analog to digital signal processing. The recent integration of AI has further propelled advancements in these technologies, notably improving speech recognition, noise reduction, sound source localization, and the fitting process. These innovations have contributed to significantly better hearing outcomes for users [41]. For HA, AI-driven advancements, such as ML with Bayesian and Gaussian optimization, have allowed personalized real-time fitting parameter adjustments, which improve sound quality in different environments. This is crucial as HAs traditionally rely on average models which do not fully account for individual user needs, despite adequate audiological fitting. An example is Widex’s SSL (WS Audiology, Nymoellevej, Denmark), an AI system integrated into an HA which adapts performance through active patient feedback, improving user satisfaction by optimizing gain and adjusting sound parameters to match its preferences [41]. In a survey on SSL satisfaction, among 118 experienced users, 53 participants had used the functionality. Of those, over 70% reported improvements in at least one listening environment, and 80% would recommend the feature [41]. Speech clarity in noisy environments is one of the primary concerns for HA users. The study by Ting et al. [56] showed how deep learning architectures and data augmentation can improve noise reduction in HAs, specifically through optimizing classification of different ambient noise, leading to better speech clarity in noisy environments [56]. Enhancing the fitting process for HAs and CIs can significantly improve audiologists’ effectiveness, expanding their ability to manage a larger patient population and more effectively addressing the healthcare challenges associated with hearing loss. As discussed in Section 4.4.2, the FOX AI-based fitting systems have proven to be equivalent or more effective than manual fittings while reducing the number of visits required in the first year post operation [96,97]. AI has potential to transform CIs and HAs from traditional prosthetic devices into intelligent augmented sensors. A notable example is the incorporation of inertial sensors into HAs for fall detection and step tracking, as demonstrated by Rahme et al.’s study on the Starkey Livio (Starkey Hearing Technologies, Eden Prairie, MN, USA) [44]. These devices can track steps accurately in real-world and treadmill conditions and detect falls effectively during daily activities, establishing HAs’ role as multi-functional health monitors [44]. Similarly, the “EarGait” system integrated AI and inertial sensors into HAs, allowing for continuous monitoring of gait parameters such as stride time and cadence, with the advantage of bilaterality over traditional gait sensor systems integrated into mobile phones and watches. The system provides valuable insights into the user’s mobility, detecting abnormalities which may indicate fall risk or mobility impairments [95]. The use of CIs as augmented sensors includes their ability to capture valuable electrophysiological data from the cochlear nerve and brain. Real-time analysis of these data can be used to monitor and predict cochlear nerve function and auditory outcomes. In a study by Skidmore et al., ML models were developed to predict the functional status of the cochlear nerve in CI users based on electrically evoked compound-action potential parameters [55]. By analyzing these signals, the model can stratify patients based on their neural function, cochlear nerve dimension, and speech recognition results, offering a way to predict CI outcomes and tailor individual treatment plans. This integration of AI has the potential to improve our understanding of how cochlear nerve health influences auditory processing and long-term CI success. To assist the scientific community in interpreting complex electrophysiological signals, Schuerch et al. [94] provided their intracochlear electrocochleography (ECochG) dataset, which consisted of 4924 signals recorded from 46 ears with CIs during and after surgery. These measures provide insight into cochlear health and neural responses. According to the authors, AI algorithms—particularly deep learning networks—will allow an accurate interpretation of these electrophysiological traces [94]. Thus, AI is transforming CIs and HAs into sensors which not only restore hearing, but also provide real-time data on several different biological parameters. This dual role of CIs as both prosthetic devices and diagnostic tools opens new avenues for hearing function understanding, personalized auditory rehabilitation, and continuous monitoring of auditory health.

5. Conclusions

At this moment, a fully comprehensive picture of the state of the art of AI applications in audiology and otology may be hard to achieve, due to the ongoing dramatic evolution of this technology. However, the current review tried to depict an updated summary of the most promising applications of various AI models in this medical discipline, while recognizing the inherent limitations related to the broad inclusion criteria. Judging from this review’s results, a large variety of AI models, based on different algorithms and computational approaches, have been described in the field of audiology. As a result, there are no universally accepted approaches. However, within such a variety, some types of AI models appeared to be consistently used for specific purposes, such as convolutional neural networks for radiological image analysis, large language models for automatic generation of diagnostic reports, and logistic regression or other statistical machine learning tools (e.g., support vector machine, multilayer perceptron, random forest, deep belief network, decision tree, k-nearest neighbor, or LASSO) for clinical predictions. In the field of audiology, despite the promising advances in AI technologies, different challenges, such as the availability of limited datasets and biases inherent in single studies, are still present, underscoring the need for larger and more diverse data collection to enhance predictive capabilities of AI models. Moreover, the ethical and professional concerns which have been raised regarding the application of AI tools in the medical setting should be addressed by specific bioethics studies in the field of audiology. Currently, professional healthcare supervision is essential in every instance where AI is integrated into audiology for patient care.

Author Contributions

Conceptualization, A.F., L.F. and G.M.; methodology, A.F., L.F. and G.M.; software, L.F.; validation, all authors; formal analysis, L.F.; investigation, A.F., L.F., V.C. and G.M.; data curation, A.F., L.F. and V.C.; writing—original draft preparation, A.F., L.F. and V.C.; writing—review and editing, A.F., L.F., V.C. and G.M.; visualization, all authors; supervision, C.d.F., E.G. and G.M.; project administration, G.M.; funding acquisition, G.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Acknowledgments

The authors thank Alison Garside for correcting the English version of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Oremule, B.; Saunders, G.H.; Kulk, K.; d’Elia, A.; Bruce, I.A. Understanding, experience, and attitudes towards artificial intelligence technologies for clinical decision support in hearing health: A mixed-methods survey of healthcare professionals in the UK. J. Laryngol. Otol. 2024, 138, 928–935. [Google Scholar] [CrossRef] [PubMed]
  2. Muthukrishnan, N.; Maleki, F.; Ovens, K.; Reinhold, C.; Forghani, B.; Forghani, R. Brief history of artificial intelligence. Neuroimaging Clin. N. Am. 2020, 30, 393–399. [Google Scholar] [CrossRef] [PubMed]
  3. Garcia, E.V.; Taylor, A.; Halkar, R.; Folks, R.; Krishnan, M.; Cooke, C.D.; Dubovsky, E. RENEX: An expert system for the interpretation of 99mTc-MAG3 scans to detect renal obstruction. J. Nucl. Med. 2006, 47, 320–329. [Google Scholar] [PubMed]
  4. Frosolini, A.; Franz, L.; Benedetti, S.; Vaira, L.A.; de Filippis, C.; Gennaro, P.; Marioni, G.; Gabriele, G. Assessing the accuracy of ChatGPT references in head and neck and ENT disciplines. Eur. Arch. Otorhinolaryngol. 2023, 280, 5129–5133. [Google Scholar] [CrossRef]
  5. Koyama, H. Machine learning application in otology. Auris Nasus Larynx 2024, 51, 666–673. [Google Scholar] [CrossRef]
  6. Choi, R.Y.; Coyner, A.S.; Kalpathy-Cramer, J.; Chiang, M.F.; Campbell, J.P. Introduction to machine learning, neural networks, and deep learning. Transl. Vis. Sci. Technol. 2020, 9, 14. [Google Scholar]
  7. Liu, X.; Faes, L.; Kale, A.U.; Wagner, S.K.; Fu, D.J.; Bruynseels, A.; Mahendiran, T.; Moraes, G.; Shamdas, M.; Kern, C.; et al. A comparison of deep learning performance against health-care professionals in detecting diseases from medical imaging: A systematic review and meta-analysis. Lancet Digit. Health 2019, 1, e271–e297. [Google Scholar] [CrossRef]
  8. Mohamed, A.A.; Lucke-Wold, B. Text-to-video generative artificial intelligence: Sora in neurosurgery. Neurosurg. Rev. 2024, 47, 272. [Google Scholar] [CrossRef]
  9. Paladugu, P.S.; Ong, J.; Nelson, N.; Kamran, S.A.; Waisberg, E.; Zaman, N.; Kumar, R.; Dias, R.D.; Lee, A.G.; Tavakkoli, A. Generative adversarial networks in medicine: Important considerations for this emerging innovation in artificial intelligence. Ann. Biomed. Eng. 2023, 51, 2130–2142. [Google Scholar] [CrossRef]
  10. OpenAI. ChatGPT (Version GPT-4) [Large Language Model]. OpenAI. 2023. Available online: https://www.openai.com/ (accessed on 9 September 2024).
  11. Frosolini, A.; Catarzi, L.; Benedetti, S.; Latini, L.; Chisci, G.; Franz, L.; Gennaro, P.; Gabriele, G. The role of Large Language Models (LLMs) in providing triage for maxillofacial trauma cases: A preliminary study. Diagnostics 2024, 14, 839. [Google Scholar] [CrossRef]
  12. Milne-Ives, M.; de Cock, C.; Lim, E.; Shehadeh, M.H.; de Pennington, N.; Mole, G.; Normando, E.; Meinert, E. The effectiveness of artificial intelligence conversational agents in health care: Systematic review. J. Med. Internet Res. 2020, 22, e20346. [Google Scholar] [CrossRef] [PubMed]
  13. Vaira, L.A.; Lechien, J.R.; Abbate, V.; Allevi, F.; Audino, G.; Beltramini, G.A.; Bergonzani, M.; Boscolo-Rizzo, P.; Califano, G.; Cammaroto, G.; et al. Validation of the Quality Analysis of Medical Artificial Intelligence (QAMAI) tool: A new tool to assess the quality of health information provided by AI platforms. Eur. Arch. Otorhinolaryngol. 2024, 281, 6123–6131. [Google Scholar] [CrossRef] [PubMed]
  14. Alter, I.L.; Chan, K.; Lechien, J.; Rameau, A. An introduction to machine learning and generative artificial intelligence for otolaryngologists-head and neck surgeons: A narrative review. Eur. Arch. Otorhinolaryngol. 2024, 281, 2723–2731. [Google Scholar] [CrossRef] [PubMed]
  15. Rapoport, N.; Pavelchek, C.; Michelson, A.P.; Shew, M.A. Artificial Intelligence in Otology and Neurotology. Otolaryngol. Clin. N. Am. 2024, 57, 791–802. [Google Scholar] [CrossRef] [PubMed]
  16. Aghakhani, A.; Yousefi, M.; Yekaninejad, M.S. Machine learning models for predicting sudden sensorineural hearing loss outcome: A systematic review. Ann. Otol. Rhinol. Laryngol. 2024, 133, 268–276. [Google Scholar] [CrossRef]
  17. Ma, T.; Wu, Q.; Jiang, L.; Zeng, X.; Wang, Y.; Yuan, Y.; Wang, B.; Zhang, T. Artificial intelligence and machine (deep) learning in otorhinolaryngology: A bibliometric analysis based on VOSviewer and CiteSpace. Ear Nose Throat J. 2023, 01455613231185074. [Google Scholar] [CrossRef]
  18. Bonnefon, J.F.; Rahwan, I.; Shariff, A. The moral psychology of artificial intelligence. Annu. Rev. Psychol. 2024, 75, 653–675. [Google Scholar] [CrossRef]
  19. Tricco, A.C.; Lillie, E.; Zarin, W.; O’Brien, K.K.; Colquhoun, H.; Levac, D.; Moher, D.; Peters, M.D.; Horsley, T.; Weeks, L.; et al. PRISMA extension for scoping reviews (PRISMA-ScR): Checklist and explanation. Ann. Intern. Med. 2018, 169, 467–473. [Google Scholar] [CrossRef]
  20. Bonadonna, F. HyperShell: An expert system shell in a hypermedia environment—Application in medical audiology. Med. Inform. 1990, 15, 105–114. [Google Scholar] [CrossRef]
  21. Juhola, M.; Viikki, K.; Laurikkala, J.; Auramo, Y.; Kentala, E.; Pyykkö, I. Application of artificial intelligence in audiology. Scand. Audiol. Suppl. 2001, 52, 97–99. [Google Scholar] [CrossRef]
  22. McCullagh, P.; Wang, H.; Zheng, H.; Lightbody, G.; McAllister, G. A comparison of supervised classification methods for auditory brainstem response determination. Stud. Health Technol. Inform. 2007, 129, 1289–1293. [Google Scholar] [PubMed]
  23. Rasku, J.; Pyykkö, I.; Levo, H.; Kentala, E.; Manchaiah, V. Disease profiling for computerized peer support of Ménière’s disease. JMIR Rehabil. Assist. Technol. 2015, 2, e9. [Google Scholar] [CrossRef] [PubMed]
  24. Song, X.D.; Wallace, B.M.; Gardner, J.R.; Ledbetter, N.M.; Weinberger, K.Q.; Barbour, D.L. Fast, continuous audiogram estimation using machine learning. Ear Hear. 2015, 36, e326–e335. [Google Scholar] [CrossRef] [PubMed]
  25. Masino, A.J.; Grundmeier, R.W.; Pennington, J.W.; Germiller, J.A.; Crenshaw, E.B., 3rd. Temporal bone radiology report classification using open source machine learning and natural language processing libraries. BMC Med. Inform. Decis. Mak. 2016, 16, 65. [Google Scholar] [CrossRef]
  26. Pyykkő, I.; Manchaiah, V.; Levo, H.; Kentala, E.; Juhola, M. Internet-based peer support for Ménière’s disease: A summary of web-based data collection, impact evaluation, and user evaluation. Int. J. Audiol. 2017, 56, 453–463. [Google Scholar] [CrossRef]
  27. Programming Cochlear Implant with Artificial Intelligence. ClinicalTrials.gov Identifier NCT03700268. Available online: https://clinicaltrials.gov/study/NCT03700268 (accessed on 4 September 2024).
  28. Sanchez Lopez, R.; Bianchi, F.; Fereczkowski, M.; Santurette, S.; Dau, T. Data-driven approach for auditory profiling and characterization of individual hearing loss. Trends Hear. 2018, 22, 2331216518807400. [Google Scholar] [CrossRef]
  29. Barbour, D.L.; Howard, R.T.; Song, X.D.; Metzger, N.; Sukesan, K.A.; DiLorenzo, J.C.; Snyder, B.R.D.; Chen, J.Y.; Degen, E.A.; Buchbinder, J.M.; et al. Online machine learning audiometry. Ear Hear. 2019, 40, 918–926. [Google Scholar] [CrossRef]
  30. Buhl, M.; Warzybok, A.; Schädler, M.R.; Lenarz, T.; Majdani, O.; Kollmeier, B. Common Audiological Functional Parameters (CAFPAs): Statistical and compact representation of rehabilitative audiological classification based on expert knowledge. Int. J. Audiol. 2019, 58, 231–245. [Google Scholar] [CrossRef]
  31. Lee, J.Y.; Choi, S.-H.; Chung, J.W. Automated classification of the tympanic membrane using a convolutional neural network. Appl. Sci. 2019, 9, 1827. [Google Scholar] [CrossRef]
  32. McKearney, R.M.; MacKinnon, R.C. Objective auditory brainstem response classification using machine learning. Int. J. Audiol. 2019, 58, 224–230. [Google Scholar] [CrossRef]
  33. Buhl, M.; Warzybok, A.; Schädler, M.R.; Majdani, O.; Kollmeier, B. Common Audiological Functional Parameters (CAFPAs) for single patient cases: Deriving statistical models from an expert-labelled data set. Int. J. Audiol. 2020, 59, 534–547. [Google Scholar] [CrossRef] [PubMed]
  34. Charih, F.; Bromwich, M.; Mark, A.E. Data-driven audiogram classification for mobile audiometry. Sci. Rep. 2020, 10, 3962. [Google Scholar] [CrossRef] [PubMed]
  35. Crowson, M.G.; Franck, K.H.; Rosella, L.C.; Chan, T.C.Y. Predicting depression from hearing loss using machine learning. Ear Hear. 2021, 42, 982–989. [Google Scholar] [CrossRef] [PubMed]
  36. Crowson, M.G.; Lee, J.W.; Hamour, A.; Mahmood, R.; Babier, A.; Lin, V.; Tucci, D.L.; Chan, T.C.Y. AutoAudio: Deep learning for automatic audiogram interpretation. J. Med. Syst. 2020, 44, 163. [Google Scholar] [CrossRef] [PubMed]
  37. Heisey, K.L.; Walker, A.M.; Xie, K.; Abrams, J.M.; Barbour, D.L. Dynamically masked audiograms with machine learning audiometry. Ear Hear. 2020, 41, 1692–1702. [Google Scholar] [CrossRef]
  38. Losorelli, S.; Kaneshiro, B.; Musacchia, G.A.; Blevins, N.H.; Fitzgerald, M.B. Factors influencing classification of frequency following responses to speech and music stimuli. Hear. Res. 2020, 398, 108101. [Google Scholar] [CrossRef]
  39. Meeuws, M.; Pascoal, D.; Janssens de Varebeke, S.; De Ceulaer, G.; Govaerts, P.J. Cochlear implant telemedicine: Remote fitting based on psychoacoustic self-tests and artificial intelligence. Cochlear Implants Int. 2020, 21, 260–268. [Google Scholar] [CrossRef]
  40. Saak, S.K.; Hildebrandt, A.; Kollmeier, B.; Buhl, M. Predicting Common Audiological Functional Parameters (CAFPAs) as interpretable intermediate representation in a clinical decision-support system for audiology. Front. Digit. Health 2020, 2, 596433. [Google Scholar] [CrossRef]
  41. Balling, L.W.; Mølgaard, L.L.; Townend, O.; Nielsen, J.B.B. The collaboration between hearing aid users and artificial intelligence to optimize sound. Semin. Hear. 2021, 42, 282–294. [Google Scholar] [CrossRef]
  42. Buhl, M.; Warzybok, A.; Schädler, M.R.; Kollmeier, B. Sensitivity and specificity of automatic audiological classification using expert-labelled audiological data and Common Audiological Functional Parameters. Int. J. Audiol. 2021, 60, 16–26. [Google Scholar] [CrossRef]
  43. Ellis, G.M.; Souza, P.E. Using machine learning and the National Health and Nutrition Examination Survey to classify individuals with hearing loss. Front. Digit. Health 2021, 3, 723533. [Google Scholar] [CrossRef] [PubMed]
  44. Rahme, M.; Folkeard, P.; Scollie, S. Evaluating the accuracy of step tracking and fall detection in the Starkey Livio artificial intelligence hearing aids: A pilot study. Am. J. Audiol. 2021, 30, 182–189. [Google Scholar] [CrossRef] [PubMed]
  45. Rodrigo, H.; Beukes, E.W.; Andersson, G.; Manchaiah, V. Exploratory data mining techniques (decision tree models) for examining the impact of internet-based cognitive behavioral therapy for tinnitus: Machine learning approach. J. Med. Internet Res. 2021, 23, e28999. [Google Scholar] [CrossRef] [PubMed]
  46. Hart, B.N.; Jeng, F.C. A Demonstration of machine learning in detecting frequency following responses in American neonates. Percept. Mot. Skills 2021, 128, 48–58. [Google Scholar] [CrossRef]
  47. Koyama, H.; Mori, A.; Nagatomi, D.; Fujita, T.; Saito, K.; Osaki, Y.; Yamasoba, T.; Doi, K. Machine learning technique reveals prognostic factors of vibrant soundbridge for conductive or mixed hearing loss patients. Otol. Neurotol. 2021, 42, e1286–e1292. [Google Scholar] [CrossRef]
  48. Kyong, J.S.; Suh, M.W.; Han, J.J.; Park, M.K.; Noh, T.S.; Oh, S.H.; Lee, J.H. Cross-modal cortical activity in the brain can predict cochlear implantation outcome in adults: A machine learning study. J. Int. Adv. Otol. 2021, 17, 380–386. [Google Scholar] [CrossRef]
  49. Li, L.P.; Han, J.Y.; Zheng, W.Z.; Huang, R.J.; Lai, Y.H. Improved environment-aware-based noise reduction system for cochlear implant users based on a knowledge transfer approach: Development and usability study. J. Med. Internet Res. 2021, 23, e25460. [Google Scholar] [CrossRef]
  50. Liu, Y.; Xu, R.; Gong, Q. Maximising the ability of stimulus-frequency otoacoustic emissions to predict hearing status and thresholds using machine-learning models. Int. J. Audiol. 2021, 60, 263–273. [Google Scholar] [CrossRef]
  51. Luengen, M.; Garrelfs, C.; Adiloǧlu, K.; Krueger, M.; Cauchi, B.; Markert, U.; Typlt, M.; Kinkel, M.; Schultz, C. Connected hearing devices and audiologists: The user-centered development of digital service innovations. Front. Digit. Health 2021, 3, 739370. [Google Scholar] [CrossRef]
  52. Profant, O.; Bureš, Z.; Balogová, Z.; Betka, J.; Fík, Z.; Chovanec, M.; Voráček, J. Decision making on vestibular schwannoma treatment: Predictions based on machine-learning analysis. Sci. Rep. 2021, 11, 18376. [Google Scholar] [CrossRef]
  53. Schlee, W.; Langguth, B.; Pryss, R.; Allgaier, J.; Mulansky, L.; Vogel, C.; Spiliopoulou, M.; Schleicher, M.; Unnikrishnan, V.; Puga, C.; et al. Using big data to develop a clinical decision support system for tinnitus treatment. Curr. Top. Behav. Neurosci. 2021, 51, 175–189. [Google Scholar] [CrossRef] [PubMed]
  54. Shafieibavani, E.; Goudey, B.; Kiral, I.; Zhong, P.; Jimeno-Yepes, A.; Swan, A.; Gambhir, M.; Buechner, A.; Kludt, E.; Eikelboom, R.H.; et al. Predictive models for cochlear implant outcomes: Performance, generalizability, and the impact of cohort size. Trends Hear. 2021, 25, 23312165211066174. [Google Scholar] [CrossRef] [PubMed]
  55. Skidmore, J.; Xu, L.; Chao, X.; Riggs, W.J.; Pellittieri, A.; Vaughan, C.; Ning, X.; Wang, R.; Luo, J.; He, S. Prediction of the functional status of the cochlear nerve in individual cochlear implant users using machine learning and electrophysiological measures. Ear Hear. 2021, 42, 180–192. [Google Scholar] [CrossRef] [PubMed]
  56. Ting, P.J.; Ruan, S.J.; Li, L.P. Environmental noise classification with inception-dense blocks for hearing aids. Sensors 2021, 21, 5406. [Google Scholar] [CrossRef] [PubMed]
  57. Wasmann, J.A.; Lanting, C.P.; Huinck, W.J.; Mylanus, E.A.M.; van der Laak, J.W.M.; Govaerts, P.J.; Swanepoel, W.; Moore, D.R.; Barbour, D.L. Computational audiology: New approaches to advance hearing health care in the digital age. Ear Hear. 2021, 42, 1499–1507. [Google Scholar] [CrossRef]
  58. Wimalarathna, H.; Ankmnal-Veeranna, S.; Allan, C.; Agrawal, S.K.; Samarabandu, J.; Ladak, H.M.; Allen, P. Comparison of machine learning models to classify Auditory Brainstem Responses recorded from children with Auditory Processing Disorder. Comput. Methods Programs Biomed. 2021, 200, 105942. [Google Scholar] [CrossRef]
  59. Wu, Z.; Lin, Z.; Li, L.; Pan, H.; Chen, G.; Fu, Y.; Qiu, Q. Deep learning for classification of pediatric otitis media. Laryngoscope 2021, 131, E2344–E2351. [Google Scholar] [CrossRef]
  60. Anderson, S.R.; Jocewicz, R.; Kan, A.; Zhu, J.; Tzeng, S.; Litovsky, R.Y. Sound source localization patterns and bilateral cochlear implants: Age at onset of deafness effects. PLoS ONE 2022, 17, e0263516. [Google Scholar] [CrossRef]
  61. Anh, D.T.; Takakura, H.; Asai, M.; Ueda, N.; Shojaku, H. Application of machine learning in the diagnosis of vestibular disease. Sci. Rep. 2022, 12, 20805. [Google Scholar] [CrossRef]
  62. Buhl, M. Interpretable Clinical Decision Support System for Audiology Based on Predicted Common Audiological Functional Parameters (CAFPAs). Diagnostics 2022, 12, 463. [Google Scholar] [CrossRef]
  63. Buhl, M.; Akin, G.; Saak, S.; Eysholdt, U.; Radeloff, A.; Kollmeier, B.; Hildebrandt, A. Expert validation of prediction models for a clinical decision-support system in audiology. Front. Neurol. 2022, 13, 960012. [Google Scholar] [CrossRef] [PubMed]
  64. Hoppe, U.; Hocke, T.; Iro, H. Age-Related Decline of Speech Perception. Front. Aging Neurosci. 2022, 14, 891202. [Google Scholar] [CrossRef] [PubMed]
  65. Hülsmeier, D.; Kollmeier, B. How much individualization is required to predict the individual effect of suprathreshold processing deficits? Assessing Plomp’s distortion component with psychoacoustic detection thresholds and FADE. Hear. Res. 2022, 426, 108609. [Google Scholar] [CrossRef] [PubMed]
  66. Lenatti, M.; Moreno-Sánchez, P.A.; Polo, E.M.; Mollura, M.; Barbieri, R.; Paglialonga, A. Evaluation of machine learning algorithms and explainability techniques to detect hearing Loss from a speech-in-noise screening test. Am. J. Audiol. 2022, 31, 961–979. [Google Scholar] [CrossRef] [PubMed]
  67. Madahana, M.C.I.; Nyandoro, O.T.C.; Khoza-Shangase, K.; Moroe, N.F. From manual to fully-automated systems and the impact of artificial intelligence: Effects of changes in mining practices on occupational noise-induced hearing loss in Africa. In Occupational Noise-Induced Hearing Loss: An African Perspective; Khoza-Shangase, K., Ed.; AOSIS: Cape Town, South Africa, 2022. [Google Scholar]
  68. Madahana, M.C.; Khoza-Shangase, K.; Moroe, N.; Nyandoro, O.; Ekoru, J. Application of machine learning approaches to analyse student success for contact learning and emergency remote teaching and learning during the COVID-19 era in speech-language pathology and audiology. S. Afr. J. Commun. Disord. 2022, 69, e1–e13. [Google Scholar] [CrossRef]
  69. McKearney, R.M.; Bell, S.L.; Chesnaye, M.A.; Simpson, D.M. Auditory brainstem response detection using machine learning: A comparison with statistical detection methods. Ear Hear. 2022, 43, 949–960. [Google Scholar] [CrossRef]
  70. McMurray, B.; Sarrett, M.E.; Chiu, S.; Black, A.K.; Wang, A.; Canale, R.; Aslin, R.N. Decoding the temporal dynamics of spoken word and nonword processing from EEG. NeuroImage 2022, 260, 119457. [Google Scholar] [CrossRef]
  71. Ntlhakana, L.; Khoza-Shangase, K. Machine learning models: Predictive tools for occupational noise-induced hearing loss in the South African mining industry. In Preventive Audiology: An African Perspective; Khoza-Shangase, K., Ed.; AOSIS: Cape Town, South Africa, 2022. [Google Scholar]
  72. Pitathawatchai, P.; Chaichulee, S.; Kirtsreesakul, V. Robust machine learning method for imputing missing values in audiograms collected in children. Int. J. Audiol. 2022, 61, 66–77. [Google Scholar] [CrossRef]
  73. Saak, S.; Huelsmeier, D.; Kollmeier, B.; Buhl, M. A flexible data-driven audiological patient stratification method for deriving auditory profiles. Front. Neurol. 2022, 13, 959582. [Google Scholar] [CrossRef]
  74. Sandström, J.; Myburgh, H.; Laurent, C.; Swanepoel, W.; Lundberg, T. A Machine learning approach to screen for otitis media using digital otoscope images labelled by an expert panel. Diagnostics 2022, 12, 1318. [Google Scholar] [CrossRef]
  75. Sundgaard, J.V.; Bray, P.; Laugesen, S.; Harte, J.; Kamide, Y.; Tanaka, C.; Christensen, A.N.; Paulsen, R.R. A Deep learning approach for detecting otitis media from wideband tympanometry measurements. IEEE J. Biomed. Health Inform. 2022, 26, 2974–2982. [Google Scholar] [CrossRef] [PubMed]
  76. Wasmann, J.W.; Pragt, L.; Eikelboom, R.; Swanepoel, W. Digital approaches to automated and machine learning assessments of hearing: Scoping review. J. Med. Internet Res. 2022, 24, e32581. [Google Scholar] [CrossRef] [PubMed]
  77. Wimalarathna, H.; Ankmnal-Veeranna, S.; Allan, C.; Agrawal, S.K.; Samarabandu, J.; Ladak, H.M.; Allen, P. Machine learning approaches used to analyze auditory evoked responses from the human auditory brainstem: A systematic review. Comput. Methods Programs Biomed. 2022, 226, 107118. [Google Scholar] [CrossRef] [PubMed]
  78. Zeng, J.; Kang, W.; Chen, S.; Lin, Y.; Deng, W.; Wang, Y.; Chen, G.; Ma, K.; Zhao, F.; Zheng, Y.; et al. A deep learning approach to predict conductive hearing loss in patients with otitis media with effusion using otoscopic images. JAMA Otolaryngol. Head Neck Surg. 2022, 148, 612–620. [Google Scholar] [CrossRef] [PubMed]
  79. Abousetta, A.; El Kholy, W.; Hegazy, M.; Kolkaila, E.; Emara, A.; Serag, S.; Ismail, O. A scoring system for cochlear implant candidate selection using artificial intelligence. Hear. Balance Commun. 2023, 21, 114–121. [Google Scholar] [CrossRef]
  80. Amiri, S.; Abdolali, F.; Neshastehriz, A.; Nikoofar, A.; Farahani, S.; Firoozabadi, L.A.; Askarabad, Z.A.; Cheraghi, S. A machine learning approach for prediction of auditory brain stem response in patients after head-and-neck radiation therapy. J. Cancer Res. Ther. 2023, 19, 1219–1225. [Google Scholar] [CrossRef]
  81. Balan, J.R.; Rodrigo, H.; Saxena, U.; Mishra, S.K. Explainable machine learning reveals the relationship between hearing thresholds and speech-in-noise recognition in listeners with normal audiograms. J. Acoust. Soc. Am. 2023, 154, 2278–2288. [Google Scholar] [CrossRef]
  82. Bragg, P.G.; Norton, B.M.; Petrak, M.R.; Weiss, A.D.; Kandl, L.M.; Corrigan, M.L.; Bahner, C.L.; Matsuoka, A.J. Application of supervised machine learning algorithms for the evaluation of utricular function on patients with Meniere’s disease: Utilizing subjective visual vertical and ocular-vestibular-evoked myogenic potentials. Acta Otolaryngol. 2023, 143, 262–273. [Google Scholar] [CrossRef]
  83. Cao, Z.; Chen, F.; Grais, E.M.; Yue, F.; Cai, Y.; Swanepoel, W.; Zhao, F. Machine learning in diagnosing middle ear disorders using tympanic membrane images: A meta-analysis. Laryngoscope 2023, 133, 732–741. [Google Scholar] [CrossRef]
  84. Doborjeh, M.; Liu, X.; Doborjeh, Z.; Shen, Y.; Searchfield, G.; Sanders, P.; Wang, G.Y.; Sumich, A.; Yan, W.Q. Prediction of tinnitus treatment outcomes based on EEG sensors and TFI score using deep learning. Sensors 2023, 23, 902. [Google Scholar] [CrossRef]
  85. Gathman, T.J.; Choi, J.S.; Vasdev, R.M.S.; Schoephoerster, J.A.; Adams, M.E. Machine learning prediction of objective hearing loss with demographics, clinical factors, and subjective hearing status. Otolaryngol. Head Neck Surg. 2023, 169, 504–513. [Google Scholar] [CrossRef] [PubMed]
  86. Jin, F.Q.; Huang, O.; Kleindienst Robler, S.; Morton, S.; Platt, A.; Egger, J.R.; Emmett, S.D.; Palmeri, M.L. A hybrid deep learning approach to identify preventable childhood hearing loss. Ear Hear. 2023, 44, 1262–1270. [Google Scholar] [CrossRef] [PubMed]
  87. Kong, S.; Huang, Z.; Deng, W.; Zhan, Y.; Lv, J.; Cui, Y. Nystagmus patterns classification framework based on deep learning and optical flow. Comput. Biol. Med. 2023, 153, 106473. [Google Scholar] [CrossRef]
  88. Mahey, P.; Toussi, N.; Purnomu, G.; Herdman, A.T. Generative Adversarial Network (GAN) for simulating electroencephalography. Brain Topogr. 2023, 36, 661–670. [Google Scholar] [CrossRef]
  89. Mohsen, S.; Sadeghijam, M.; Talebian, S.; Pourbakht, A. Use of some relevant parameters for primary prediction of brain activity in idiopathic tinnitus based on a machine learning application. Audiol. Neurootol. 2023, 28, 446–457. [Google Scholar] [CrossRef]
  90. Petsiou, D.P.; Martinos, A.; Spinos, D. Applications of artificial intelligence in temporal bone imaging: Advances and future challenges. Cureus 2023, 15, e44591. [Google Scholar] [CrossRef]
  91. Ramzi, G.; McLoughlin, I.; Palaniappan, R. Did you hear that? Detecting aditory events with EEGNet. In Proceedings of the 2023 45th Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), Sydney, Australia, 24–27 July 2023; pp. 1–4. [Google Scholar] [CrossRef]
  92. Saeed, H.S.; Fergie, M.; Mey, K.; West, N.; Caye-Thomasen, P.; Nash, R.; Saeed, S.R.; Stivaros, S.M.; Black, G.; Bruce, I.A. Enlarged vestibular aqueduct and associated inner ear malformations: Hearing loss prognostic factors and data modeling from an international cohort. J. Int. Adv. Otol. 2023, 19, 454–460. [Google Scholar] [CrossRef]
  93. Schroeer, A.; Andersen, M.R.; Rank, M.L.; Hannemann, R.; Petersen, E.B.; Rønne, F.M.; Strauss, D.J.; Corona-Strauss, F.I. Assessment of vestigial auriculomotor activity to acoustic stimuli using electrodes in and around the ear. Trends Hear. 2023, 27, 23312165231200158. [Google Scholar] [CrossRef]
  94. Schuerch, K.; Wimmer, W.; Dalbert, A.; Rummel, C.; Caversaccio, M.; Mantokoudis, G.; Gawliczek, T.; Weder, S. An intracochlear electrocochleography dataset—From raw data to objective analysis using deep learning. Sci. Data 2023, 10, 157. [Google Scholar] [CrossRef]
  95. Seifer, A.K.; Dorschky, E.; Küderle, A.; Moradi, H.; Hannemann, R.; Eskofier, B.M. EarGait: Estimation of temporal gait parameters from hearing aid integrated inertial sensors. Sensors 2023, 23, 6565. [Google Scholar] [CrossRef]
  96. Wathour, J.; Govaerts, P.J.; Lacroix, E.; Naïma, D. Effect of a CI programming fitting tool with artificial intelligence in experienced cochlear implant patients. Otol. Neurotol. 2023, 44, 209–215. [Google Scholar] [CrossRef] [PubMed]
  97. Wathour, J.; Govaerts, P.J.; Derue, L.; Vanderbemden, S.; Huaux, H.; Lacroix, E.; Deggouj, N. Prospective comparison between manual and computer-assisted (FOX) cochlear implant fitting in newly implanted patients. Ear Hear. 2023, 44, 494–505. [Google Scholar] [CrossRef] [PubMed]
  98. Yang, T.H.; Chen, Y.F.; Cheng, Y.F.; Huang, J.N.; Wu, C.S.; Chu, Y.C. Optimizing age-related hearing risk predictions: An advanced machine learning integration with HHIE-S. BioData Min. 2023, 16, 35. [Google Scholar] [CrossRef]
  99. Yu, F.; Wu, P.; Deng, H.; Wu, J.; Sun, S.; Yu, H.; Yang, J.; Luo, X.; He, J.; Ma, X.; et al. A questionnaire-based ensemble learning model to predict the diagnosis of vertigo: Model development and validation study. J. Med. Internet Res. 2022, 24, e34126. [Google Scholar] [CrossRef] [PubMed]
  100. Ahmed, M.A.O.; Satar, Y.A.; Darwish, E.M.; Zanaty, E.A. Synergistic integration of multi-view brain networks and advanced machine learning techniques for auditory disorders diagnostics. Brain Inform. 2024, 11, 3. [Google Scholar] [CrossRef] [PubMed]
  101. Carlson, M.L.; Carducci, V.; Deep, N.L.; DeJong, M.D.; Poling, G.L.; Brufau, S.R. AI model for predicting adult cochlear implant candidacy using routine behavioral audiometry. Am. J. Otolaryngol. 2024, 45, 104337. [Google Scholar] [CrossRef]
  102. Chen, P.Y.; Yang, T.W.; Tseng, Y.S.; Tsai, C.Y.; Yeh, C.S.; Lee, Y.H.; Lin, P.H.; Lin, T.C.; Wu, Y.J.; Yang, T.H.; et al. Machine learning-based longitudinal prediction for GJB2-related sensorineural hearing loss. Comput. Biol. Med. 2024, 176, 108597. [Google Scholar] [CrossRef]
  103. Chua, K.W.; Yeo, H.K.H.; Tan, C.K.L.; Martinez, J.C.; Goh, Z.H.; Dritsas, S.; Simpson, R.E. A novel ear impression-taking method using structured light imaging and machine learning: A pilot proof of concept study with patients’ feedback on prototype. J. Clin. Med. 2024, 13, 1214. [Google Scholar] [CrossRef]
  104. Deroche, M.L.D.; Wolfe, J.; Neumann, S.; Manning, J.; Hanna, L.; Towler, W.; Wilson, C.; Bien, A.G.; Miller, S.; Schafer, E.; et al. Cross-Modal Plasticity in Children with cochlear implant: Converging evidence from EEG and functional near-infrared spectroscopy. Brain Commun. 2024, 6, fcae175. [Google Scholar] [CrossRef]
  105. Dou, Z.; Li, Y.; Deng, D.; Zhang, Y.; Pang, A.; Fang, C.; Bai, X.; Bing, D. Pure tone audiogram classification using deep learning techniques. Clin. Otolaryngol. 2024, 49, 595–603. [Google Scholar] [CrossRef]
  106. Heman-Ackah, S.M.; Blue, R.; Quimby, A.E.; Abdallah, H.; Sweeney, E.M.; Chauhan, D.; Hwa, T.; Brant, J.; Ruckenstein, M.J.; Bigelow, D.C.; et al. A multi-institutional machine learning algorithm for prognosticating facial nerve injury following microsurgical resection of vestibular schwannoma. Sci. Rep. 2024, 14, 12963. [Google Scholar] [CrossRef] [PubMed]
  107. Jedrzejczak, W.W.; Kochanek, K. Comparison of the audiological knowledge of three chatbots: ChatGPT, Bing Chat, and Bard. Audiol. Neurootol. 2024, 13, 1–7. [Google Scholar] [CrossRef] [PubMed]
  108. Jedrzejczak, W.W.; Skarzynski, P.H.; Raj-Koziak, D.; Sanfins, M.D.; Hatzopoulos, S.; Kochanek, K. ChatGPT for tinnitus information and support: Response accuracy and retest after three and six months. Brain Sci. 2024, 14, 465. [Google Scholar] [CrossRef] [PubMed]
  109. Kohler, I.; Perrotta, M.V.; Ferreira, T.; Eagleman, D.M. Cross-modal sensory boosting to improve high-frequency hearing loss: Device development and validation. JMIRx Med. 2024, 5, e49969. [Google Scholar] [CrossRef] [PubMed]
  110. Liang, S.; Xu, J.; Liu, H.; Liang, R.; Guo, Z.; Lu, M.; Liu, S.; Gao, J.; Ye, Z.; Yi, H. Automatic recognition of auditory brainstem response waveforms using a deep learning-based framework. Otolaryngol. Head Neck Surg. 2024, 171, 1165–1171. [Google Scholar] [CrossRef]
  111. Madahana, M.C.I.; Ekoru, J.E.D.; Sebothoma, B.; Khoza-Shangase, K. Development of an artificial intelligence based occupational noise induced hearing loss early warning system for mine workers. Front. Neurosci. 2024, 18, 1321357. [Google Scholar] [CrossRef]
  112. Sadegh-Zadeh, S.A.; Soleimani Mamalo, A.; Kavianpour, K.; Atashbar, H.; Heidari, E.; Hajizadeh, R.; Roshani, A.S.; Habibzadeh, S.; Saadat, S.; Behmanesh, M.; et al. Artificial intelligence approaches for tinnitus diagnosis: Leveraging high-frequency audiometry data for enhanced clinical predictions. Front. Artif. Intell. 2024, 7, 1381455. [Google Scholar] [CrossRef]
  113. Simon, A.; Bech, S.; Loquet, G.; Østergaard, J. Cortical linear encoding and decoding of sounds: Similarities and differences between naturalistic speech and music listening. Eur. J. Neurosci. 2024, 59, 2059–2074. [Google Scholar] [CrossRef]
  114. Soylemez, E.; Avci, I.; Yildirim, E.; Karaboya, E.; Yilmaz, N.; Ertugrul, S.; Tokgoz-Yilmaz, S. Predicting noise-induced hearing loss with machine learning: The influence of tinnitus as a predictive factor. J. Laryngol. Otol. Published Online 2024, 1–18. [Google Scholar] [CrossRef]
  115. Twinomurinzi, H.; Myburgh, H.; Barbour, D.L. Active transfer learning for audiogram estimation. Front. Digit. Health 2024, 6, 1267799. [Google Scholar] [CrossRef]
  116. Wang, C.; Young, A.S.; Raj, C.; Bradshaw, A.P.; Nham, B.; Rosengren, S.M.; Calic, Z.; Burke, D.; Halmagyi, G.M.; Bharathy, G.K.; et al. Machine learning models help differentiate between causes of recurrent spontaneous vertigo. J. Neurol. 2024, 271, 3426–3438. [Google Scholar] [CrossRef] [PubMed]
  117. Wang, S.; Mo, C.; Chen, Y.; Dai, X.; Wang, H.; Shen, X. Exploring the performance of ChatGPT-4 in the Taiwan audiologist qualification examination: Preliminary observational study highlighting the potential of AI chatbots in hearing care. JMIR Med. Educ. 2024, 10, e55595. [Google Scholar] [CrossRef] [PubMed]
  118. Wang, Y.; Yao, X.; Wang, D.; Ye, C.; Xu, L. A machine learning screening model for identifying the risk of high-frequency hearing impairment in a general population. BMC Public Health 2024, 24, 1160. [Google Scholar] [CrossRef] [PubMed]
  119. Yin, Z.; Kuang, Z.; Zhang, H.; Guo, Y.; Li, T.; Wu, Z.; Wang, L. Explainable AI method for tinnitus diagnosis via neighbor-augmented knowledge graph and traditional Chinese medicine: Development and validation study. JMIR Med. Inform. 2024, 12, e57678. [Google Scholar] [CrossRef] [PubMed]
  120. Zeitler, D.M.; Buchlak, Q.D.; Ramasundara, S.; Farrokhi, F.; Esmaili, N. Predicting acoustic hearing preservation following cochlear implant surgery using machine learning. Laryngoscope 2024, 134, 926–936. [Google Scholar] [CrossRef]
  121. Liu, Y.; Gong, Q. Deep learning models for predicting hearing thresholds based on swept-tone stimulus-frequency otoacoustic emissions. Ear Hear. 2024, 45, 465–475. [Google Scholar] [CrossRef]
  122. Szaleniec, J.; Wiatr, M.; Szaleniec, M.; Składzień, J.; Tomik, J.; Oleś, K.; Tadeusiewicz, R. Artificial neural network modelling of the results of tympanoplasty in chronic suppurative otitis media patients. Comput. Biol. Med. 2013, 43, 16–22. [Google Scholar] [CrossRef]
  123. Fang, Y.; Wu, J.; Wang, Q.; Qiu, S.; Bozoki, A.; Liu, M. Source-free collaborative domain adaptation via multi-perspective feature enrichment for functional MRI analysis. Pattern Recognit. 2025, 157, 110912. [Google Scholar] [CrossRef]
  124. Deo, R.C. Machine learning in medicine. Circulation 2015, 132, 1920–1930. [Google Scholar] [CrossRef]
  125. Handelman, G.S.; Kok, H.K.; Chandra, R.V.; Razavi, A.H.; Lee, M.J.; Asadi, H. eDoctor: Machine learning and the future of medicine. J. Intern. Med. 2018, 284, 603–619. [Google Scholar] [CrossRef]
  126. An, Q.; Rahman, S.; Zhou, J.; Kang, J.J. A comprehensive review on machine learning in healthcare industry: Classification, restrictions, opportunities and challenges. Sensors 2023, 23, 4178. [Google Scholar] [CrossRef] [PubMed]
  127. Jiao, R.; Zhang, Y.; Ding, L.; Cai, R.; Zhang, J. Learning with limited annotations: A survey on deep semi-supervised learning for medical image segmentation. arXiv 2022, arXiv:2207.10085. [Google Scholar] [CrossRef] [PubMed]
  128. Wang, D.; Zhang, Y.; Zhang, K.; Wang, L. FocalMix: Semi-supervised learning for 3D medical image detection. arXiv 2020, arXiv:2007.02701. [Google Scholar]
  129. Liu, Y.; Logan, B.; Liu, N.; Xu, Z.; Tang, J.; Wang, Y. Deep reinforcement learning for dynamic treatment regimes on medical registry data. Healthc. Inform. 2017, 2017, 380–385. [Google Scholar] [PubMed]
  130. Zhachow, S.; Zilske, M.; Hege, H.C. 3D Reconstruction of Individual Anatomy from Medical Image Data: Segmentation and Geometry Processing. In Proceedings of the 25. ANSYS Conference & CADFEM Users’ Meeting, Dresden, Germany, 21–23 November 2007. [Google Scholar]
  131. Ke, J.; Lv, Y.; Ma, F.; Du, Y.; Xiong, S.; Wang, J.; Wang, J. Deep learning-based approach for the automatic segmentation of adult and pediatric temporal bone computed tomography images. Quant. Imaging Med. Surg. 2023, 13, 1577–1591. [Google Scholar] [CrossRef]
  132. Franz, L.; Isola, M.; Bagatto, D.; Tuniz, F.; Robiony, M. A Novel approach to skull-base and orbital osteotomies through virtual planning and navigation. Laryngoscope 2019, 129, 823–831. [Google Scholar] [CrossRef]
  133. European Society of Radiology (ESR). Current practical experience with artificial intelligence in clinical radiology: A survey of the European Society of Radiology. Insights Imaging 2022, 13, 107. [Google Scholar] [CrossRef]
  134. Wu, H.; Liu, J.; Chen, G.; Liu, J.; Chen, G.; Liu, W.; Hao, R.; Liu, L.; Ni, G.; Liu, Y.; et al. Automatic semicircular canal segmentation of CT volumes using improved 3D UNet with attention mechanism. Comput. Intell. Neurosci. 2021, 2021, 9654059. [Google Scholar] [CrossRef]
  135. Heutink, F.; Koch, V.; Verbist, B.; van der Woude, W.J.; Mylanus, E.; Huinck, W.; Sechopoulos, I.; Caballo, M. Multi-scale deep learning framework for cochlea localization, segmentation and analysis on clinical ultra-high-resolution CT images. Comput. Methods Programs Biomed. 2020, 191, 105387. [Google Scholar] [CrossRef]
  136. Lingam, R.K.; Bassett, P. A Meta-analysis on the diagnostic performance of non-echoplanar diffusion-weighted imaging in detecting middle ear cholesteatoma: 10 years on. Otol. Neurotol. 2017, 38, 521–528. [Google Scholar] [CrossRef]
  137. Boucher, F.; Liao, E.; Srinivasan, A. Diffusion-weighted imaging of the head and neck (including temporal bone). Magn. Reson. Imaging Clin. N. Am. 2021, 29, 205–232. [Google Scholar] [CrossRef] [PubMed]
  138. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 785–794. [Google Scholar] [CrossRef]
  139. Yasaka, K.; Furuta, T.; Kubo, T.; Maeda, E.; Katsura, M.; Sato, J.; Ohtomo, K. Full and hybrid iterative reconstruction to reduce artifacts in abdominal CT for patients scanned without arm elevation. Acta Radiol. 2017, 58, 1085–1093. [Google Scholar] [CrossRef] [PubMed]
  140. Higaki, T.; Nakamura, Y.; Tatsugami, F.; Nakaura, T.; Awai, K. Improvement of image quality at CT and MRI using deep learning. Jpn. J. Radiol. 2019, 37, 73–80. [Google Scholar] [CrossRef]
  141. Fujita, N.; Yasaka, K.; Hatano, S.; Sakamoto, N.; Kurokawa, R.; Abe, O. Deep learning reconstruction for high-resolution computed tomography images of the temporal bone: Comparison with hybrid iterative reconstruction. Neuroradiology 2024, 66, 1105–1112. [Google Scholar] [CrossRef]
  142. Ogawa, M.; Kisohara, M.; Yamamoto, T.; Shibata, S.; Ojio, Y.; Mochizuki, K.; Tatsuta, A.; Iwasaki, S.; Shibamoto, Y. Utility of unsupervised deep learning using a 3D variational autoencoder in detecting inner ear abnormalities on CT images. Comput. Biol. Med. 2022, 147, 105683. [Google Scholar] [CrossRef]
  143. Belle, V.; Papantonis, I. Principles and practice of explainable machine learning. Front. Big Data 2021, 4, 688969. [Google Scholar] [CrossRef]
  144. Asadi-Lari, M.; Tamburini, M.; Gray, D. Patients’ needs, satisfaction, and health related quality of life: Towards a comprehensive model. Health Qual. Life Outcomes 2004, 2, 32. [Google Scholar] [CrossRef]
  145. Wathour, J.; Govaerts, P.J.; Deggouj, N. From manual to artificial intelligence fitting: Two cochlear implant case studies. Cochlear Implants Int. 2020, 21, 299–305. [Google Scholar] [CrossRef]
Figure 1. PRISMA diagram of the present review from database search to inclusion.
Figure 1. PRISMA diagram of the present review from database search to inclusion.
Sensors 24 07126 g001
Figure 2. World map depicting top three contributing countries and timeline showing trends of contributions over the last decades. Abbreviation: manuscripts (MM).
Figure 2. World map depicting top three contributing countries and timeline showing trends of contributions over the last decades. Abbreviation: manuscripts (MM).
Sensors 24 07126 g002
Figure 3. Integration of AI in audiology and vestibology.
Figure 3. Integration of AI in audiology and vestibology.
Sensors 24 07126 g003
Table 1. Summary of top journals and study designs in AI and audiology research considering manuscripts included in the present review.
Table 1. Summary of top journals and study designs in AI and audiology research considering manuscripts included in the present review.
JournalsNumber of Publications (%)Study DesignNumber of Publications (%)
Ear and Hearing13 (12.5%)Observational71 (68.3%)
International Journal of Audiology6 (5.8%)Development and Validation17 (16.3%)
Frontiers in Digital Health4 (3.8%)Reviews9 (8.7%)
Others81 (88.7%)Clinical Trials3 (2.9%)
Surveys2 (1.9%)
Case Reports and Study Protocols2 (1.9%)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Frosolini, A.; Franz, L.; Caragli, V.; Genovese, E.; de Filippis, C.; Marioni, G. Artificial Intelligence in Audiology: A Scoping Review of Current Applications and Future Directions. Sensors 2024, 24, 7126. https://doi.org/10.3390/s24227126

AMA Style

Frosolini A, Franz L, Caragli V, Genovese E, de Filippis C, Marioni G. Artificial Intelligence in Audiology: A Scoping Review of Current Applications and Future Directions. Sensors. 2024; 24(22):7126. https://doi.org/10.3390/s24227126

Chicago/Turabian Style

Frosolini, Andrea, Leonardo Franz, Valeria Caragli, Elisabetta Genovese, Cosimo de Filippis, and Gino Marioni. 2024. "Artificial Intelligence in Audiology: A Scoping Review of Current Applications and Future Directions" Sensors 24, no. 22: 7126. https://doi.org/10.3390/s24227126

APA Style

Frosolini, A., Franz, L., Caragli, V., Genovese, E., de Filippis, C., & Marioni, G. (2024). Artificial Intelligence in Audiology: A Scoping Review of Current Applications and Future Directions. Sensors, 24(22), 7126. https://doi.org/10.3390/s24227126

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop