Next Article in Journal
Hyperpolarized Carbon-13 MRI in Breast Cancer
Previous Article in Journal
Optic Nerve Sheath Meningiomas: Solving Diagnostic Challenges with 68Ga-DOTATOC PET/CT
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Diagnosis, Treatment, and Management of Otitis Media with Artificial Intelligence

Department of Otorhinolaryngology Head and Neck Surgery, The Peaking Union Medical College Hospital, No. 1, Shuaifuyuan, Dongcheng District, Beijing 100010, China
*
Authors to whom correspondence should be addressed.
Diagnostics 2023, 13(13), 2309; https://doi.org/10.3390/diagnostics13132309
Submission received: 2 May 2023 / Revised: 4 June 2023 / Accepted: 14 June 2023 / Published: 7 July 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
A common infectious disease, otitis media (OM) has a low rate of early diagnosis, which significantly increases the difficulty of treating the disease and the likelihood of serious complications developing including hearing loss, speech impairment, and even intracranial infection. Several areas of healthcare have shown great promise in the application of artificial intelligence (AI) systems, such as the accurate detection of diseases, the automated interpretation of images, and the prediction of patient outcomes. Several articles have reported some machine learning (ML) algorithms such as ResNet, InceptionV3 and Unet, were applied to the diagnosis of OM successfully. The use of these techniques in the OM is still in its infancy, but their potential is enormous. We present in this review important concepts related to ML and AI, describe how these technologies are currently being applied to diagnosing, treating, and managing OM, and discuss the challenges associated with developing AI-assisted OM technologies in the future.

1. Introduction

Otitis media (OM) is one of the most common pediatric diseases, affecting almost two-thirds of preschoolers [1]. Unfortunately, doctors often fail to correctly diagnose OM [2,3], partly due to the young age of patients, poor awareness of patients and their families, and the insidious nature of the symptoms of the disease. The absence of an early diagnosis of OM can lead to hearing loss, speech impairment, and even serious extracranial and intracranial complications [4,5]. Due to many misdiagnosed cases of OM, antibiotics are overused and unnecessary operations of OM are performed, which also places a burden on global public health [6,7]. While at the same time, the advent of big data and artificial intelligence (AI), as well as the wide use of electronic health records, present an opportunity to revolutionise how patients with OM are treated. Based on a large amount of data collected over time about OM patients, AI can be used to identify latent diseases, narrow down disease types, predict disease risks, personalise treatment options, assist in surgery, and optimise disease management (Figure 1) [8]. This article discusses AI and machine learning (ML) as contributions to OM management, the need for AI algorithms tailored to OM, and the implications of integrating AI and OM care for patients, physicians, surgeons, and the entire multidisciplinary team of care providers.

2. Artificial Intelligence & Machine Learning

2.1. What

In a broad sense, AI is the ability of machines to mimic the functions of human intelligence. As a form of AI, ML uses statistical methods to enable machines to learn tasks without the need for explicit programming [9]. Often, learning involves mathematical model fitting. Mathematically, relationships between features (or variables) are analysed through many data points, resulting in an optimal fitting model. Deep learning (DL) extends ML, while mathematical models are more complex and detailed.

2.2. How

ML relies on data and algorithms, and a large amount of data and advanced algorithms, if trained and adjusted appropriately, can often produce ideal results (Table 1 and Figure 2). Clinical data regarding OM patients can be collected more easily with the widespread use of electronic health record systems. Data included medical history records, physical examination records, laboratory results, otoscopy images, audiology room images, imaging data, drug use records, surgical records, surgical videos, and pathological records. A large amount of data is being collected in the healthcare field from large-scale genetic research and consumer electronic devices (such as wearables and smartphones) that provide physiological and behavioral information.

2.3. Why

By contrast with traditional statistical methods based on hypothesis, ML algorithms can be fitted with large amounts of data to determine the model that will produce the best results. In addition to linear relationships between features and results, ML algorithms can capture more complex, nonlinear interactions [10,11]. ML can also handle large data sets and integrate insights from different data types to provide precision medicine solutions.

3. Materials and Methods

The present review was conducted in accordance with the PRISMA guidelines. The overall workflow is shown in Figure 3 and described below. Ethics approval and patient consent were not required for this review.

3.1. Literature Search

A search of online databases (including PubMed, MEDLINE, EMBASE, Scopus and ResearchGate databases) for articles, abstracts or conference proceedings published between 2011 and 2022 that used Al-based approaches to diagnose, treat or manage OM patients was conducted. Searches were limited to those involving human subjects and those published in the English Language.
Medical subject headings (MeSH) terms and non-MeSH terms related to Al approaches included: ‘artificial intelligence’, ‘machine learning’, ‘deep learning’, ‘convolutional neural networks’, ‘data mining’, ‘computer-assisted diagnosis’, ‘computer-assisted surgery’, ‘natural language process’ and ‘computer vision’. MeSH terms and key-words related to otitis media included: ‘ear’, ‘eardrum’, ‘tympanic membrane’, ‘ear disease’, ‘otitis’, ‘otitis media’, ‘acute otitis media’, ‘chronic otitis media’, ‘chronic suppurative otitis media’, ‘cholesteatoma’, ‘OM’, ‘AOM’, ‘OME’, ‘COM’, ‘CSOM’.

3.2. Selection Criteria

Titles and abstracts were reviewed for eligibility by two independent investigators (Ding, X. and Huang, Y.). Discrepancies between the two investigators were resolved by two senior authors (Feng, G. and Gao, Z.), board certified otolaryngologists. Reference lists of available full-text articles were also manually screened for further studies eligible for inclusion in this review. Studies describing the development of an autonomous algorithm to diagnose, treat or manage OM patients using the Al approaches described above were included. Articles that were excluded consisted of those that published in the non-English Language, did not use Al-based approaches, explore other diseases instead of OM or were review articles or editorials. Study inclusion/exclusion is summarised in a PRISMA flow diagram (Figure 3).

3.3. Data Extraction

Two investigators (Ding, X. and Huang, Y.) independently extracted data from included studies for analysis. The following characteristics were extracted and listed from included studies: primary author, year, study objective, Al technique to achieve study objective, number and datasets, deep learning models and performance (Supplementary Tables S1–S6).

4. Diagnosis

Below, we will proceed to describe in detail recent progress in the diagnosis of otitis media based on AI. This section was divided into Computer Vision (Section 4.1) and Natural Language Processing (Section 4.2) according to the CNN methods, and the Computer Vision part was further classified into Otoscopy (Section 4.1.1), Radiology & Pathology (Section 4.1.2) and Tympanometry (Section 4.1.3) according to the data sources.

4.1. Computer Vision

4.1.1. Otoscopy

A branch of AI known as computer vision (CV) uses deep learning algorithms to analyse images and videos automatically. Recently, CV has shown great promise in OM-related medical images, particularly otoscopy images. Three main types of ear endoscopes are commonly used in clinical practice: the endoscopic otoscope, the electronic otoscope, and the video pneumatic otoscope (VPO). In recent years, special otoscopes have been developed, such as multicolor imaging otoscopes, shortwave infrared otoscopes, and optical coherence tomography (OCT) in vivo. After screening relevant articles on the analysis of OM endoscopic images based on AI, a total of 43 relevant articles were identified (Supplementary Table S1) [12,13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29,30,31,32,33,34,35,36,37,38,39,40,41,42,43,44,45,46,47,48,49,50,51,52,53,54]. In most studies, the ML system achieved a high level of diagnosis and was as accurate as the clinician or even better. A new ML algorithm has been developed by Matthew G. Crowson et al., using 639 otoscopic images containing middle ear effusions and infections. On a test set consisted of 22 images, their algorithm detected abnormalities in the tympanic membrane with an accuracy of 95.5%, significantly higher than the 65% average accuracy of 39 pediatric and ENT physicians [15]. Moreover, Hayoung Byun et al., further constructed a convolutional neural network (CNN) to identify conductive deafness based on 1130 VPO images of patients with OM. With an accuracy of 94.1% (AUC 0.972), their CNN algorithm could distinguish between normal tympanic membranes, conductive hearing loss and sensorineural hearing loss, which is higher than the accuracy rate of 79.0% (AUC 0.773) of experienced ENT physicians [20]
At the same time, many research teams are also focusing on developing special ear endoscopy technology based on AI to diagnose OM. Using the single green channel mode, for example, Michelle Viscaino et al., demonstrated the optimal overall diagnostic performance (accuracy 92%, sensitivity 85%, specificity 95% and precision 86%) by changing the otoscope colour based on multicolour imaging [21]. Rustin G. Kashani et al., developed an algorithm that could accurately detect effusion when OM was present, using 1179 shortwave infrared otoscopic images [29]. Optical coherence tomography (OCT) is a non-invasive test that evaluates the tympanic membrane (TM) and the adjacent middle ear space (OM). Using OCT images, otoscopic images and patient reports, Guillermo L. Monroy et al., trained a neural network capable of automatically identifying the TM and middle ear effusion in OM patients with a diagnostic accuracy of more than 90% [55]. They further increased the OCT images of the chinchilla, the gold-standard pre-clinical animal model used to study human OM in subsequent studies to increase the sample size and robustness, further improving the accuracy of their model (95% accuracy) [56].

4.1.2. Radiology & Pathology

An optimal diagnostic imaging method for chronic otitis media (COM) is high-resolution computed tomography (CT) of the temporal bone since it can provide detailed information about the middle ear structure and shows great sensitivity in detecting the extent and complications of middle ear lesions. Due to the rapid development of deep learning technology, computer vision processing and recognition techniques have been applied to interpret CT images of OM automatically (Supplementary Table S2) [57,58,59,60,61]. Using this method can greatly reduce labor costs and human errors due to repetition, fatigue, or knowledge differences, as well as improve the accuracy of the diagnosis. Yanmei Wang et al., developed an algorithm for diagnosing COM by analysing 1147 CT images containing cholesteatoma and chronic suppurative otitis media (CSOM). Their model reached an AUC of 0.92, with sensitivity and specificity exceeding the average for clinical specialists (sensitivity 83.3% vs. 81.1%, specificity 91.4% vs. 88.8%) [57]. Using the ROI method of Mask R-CNN, Zhenchang Wang et al., conducted the image extraction network. Their model further improved the accuracy of the pathological recognition algorithm (AUC 0.96) compared to Yanmei Wang et al. [58]. Bo Duan et al., used GoogleNet to develop a neural network capable of distinguishing primary ciliated dyskinesia (PCD)-associated OM from otitis media effusion (OME). Their model exhibited a diagnostic accuracy of greater than 90% [61].
Even though CT has many advantages in diagnosing OM, plain radiography or magnetic resonance imaging (MRI) are still used clinically in certain cases. Considering the sensitivity to radiation, plain radiographs were selected for auxiliary diagnosis of OM in children under two years of age. A neural network was trained to diagnose mastoiditis based on 9988 images of mastoid X-rays obtained from 4994 patients by Kyong Joon Lee et al. Their model demonstrated a high diagnostic accuracy (AUC 0.968 ± 0.003) [62]. In addition to being radiation-free, MRI can distinguish cholesteatoma of the middle ear from cholesterol granuloma, exhibiting high signals on both T1 and T2 sequences. In order to diagnose diseases more quickly and accurately, the combination of pathological sections and AI is considered the gold standard. Although AI-assisted MRI and pathological diagnosis have been extensively studied in cardiovascular, tumor and other diseases, research on OM remains limited.
Radiomics is a rapidly growing field of research involving the extraction of quantitative indicators in medical images, known as radiomic features. Radiomic features capture tissue and lesion characteristics, such as heterogeneity and shape, that can be used alone or in combination with demographic, histological, genomic or proteomic data to solve clinical problems. The approach relates the omics with radiology forming a bridge between two areas of research that may prove useful for clinical treatment planning leading to better outcomes. Although radiomics has shown great effects in the diagnosis, treatment and prognosis of various diseases, especially tumor, there are few relevant studies in the field of OM [63,64].

4.1.3. Tympanometry

Tympanometry measures the sound quality of the ear canal by using an acoustic probe (usually 226 Hz or 1 kHz) and a microphone (for measuring sound). In addition, it can provide us with quantitative information regarding the TM and tympanum of the middle ear, as well as determine whether or not there is any fluid present(Supplementary Table S2) [41,65,66,67]. A decision fusion method based on majority voting was adopted by Hamidullah Binol et al., to analyse 73 records with both tympanometry and digital otoscopy data, and their model classification accuracy reached 84.9% [41]. The wideband tympanometry (WBT) method is an extension of standard tympanometry that simultaneously determines the full range of middle ear characteristics at the audiometrically most important frequencies, thereby achieving a more accurate and efficient assessment than a standard 226 Hz or 1 kHz tympanogram. The work by Josefine Vilsøll Sundgaard et al., applied deep learning to WBT, using 1014 measurements for training, and the algorithm reached an average accuracy of 92.6% for the detection of OMs. Nevertheless, the model does not distinguish well between specific types of OM, such as acute otitis media (AOM) and OME [65].

4.2. Natural Language Processing

A large amount of textual data in electronic medical records and laboratory tests allow natural language processing (NLP) to automatically identify characteristic descriptive words, thereby detecting cases of OM (Supplementary Table S3) [68,69,70,71]. Kuruvilla A et al., applied rule-based natural language processing algorithms to the descriptive words of patients’ tympanic membranes to detect patients with OM automatically. First, the authors identified 8 characteristic words related to the tympanic membrane of OM (such as tympanic membrane bulging, bubbles presence, central concavity, translucency, etc.) based on the analysis of 783 tympanic membrane image examination reports of children from otoscopic experts. Following that, they used the random forest classification classifier algorithm to establish the appropriate syntax and finally developed the NLP model for the diagnosis and classification of OM. The algorithm they developed achieved an accuracy level of 89.9%, which is significantly higher than clinicians without special training [68]. H Binol et al., also used natural language processing and the bag-of-words model (BOW) to analyse otoscopic reports of OM patients to determine five characteristic vocabulary words that could be used to diagnose OM. Their proposed model provided an overall F1-score of 90.2% based on neighborhood component analysis (NCA) [69]. In addition to constructing the NLP model using note texts and structured data from children under five, Herigon JC et al., also limited the training object to otoscopic report findings. Their model can be used to identify patients with AOM more accurately to assess antibiotic prescribing practices better [70].

5. Treatment

5.1. Internal Medicine

Currently, the primary treatment for OM is the use of antibacterial drugs. However, paediatricians and ENT doctors are constantly concerned about the long-term safety of drugs and the emergence of antibiotic resistance. The use of ML models can allow us to predict individual susceptibility to antibiotics and resistance to them, improve the effectiveness of antibiotics, and even design new antibacterial drugs. In general, polymyxins are used as a last-resort treatment for Gram-negative infections, and Klebsiella pneumoniae is one of the most common Gram-negative bacteria resistant to polymyxins. According to Nenad Macesic et al., ML was applied to whole-genome sequencing data from >600 Klebsiella pneumoniae clonal group 258 (CG258) genomes, and this approach gave an accurate prediction of polymyxin sensitivity and resistance (AUC 0.894) [72]. A study by Rachael A. Mansbach et al., combined ML with drug development, analysed the representation and definition of compounds using an ML algorithm, determined the characteristic submolecular fragment vocabulary, and fitted the prediction model according to the vocabulary. A set of fragments identified by their algorithm can allow drugs to penetrate the outer membrane of Pseudomonas aeruginosa bacteria, in addition to nine compounds that are regarded as good outer membrane penetrators, five of which have been successfully validated experimentally [73].
Increasing resistance to traditional antibiotics has led many researchers to focus on developing new antibacterial drugs. AI can greatly reduce drug development’s time and capital costs when used as an intelligent auxiliary tool. Francisco R Fields et al., used ML and minimum-region determination methods based on biophysics to automate the analysis of antimicrobial peptides, a type of antibiofilm drug, for identifying antimicrobial peptides suitable for synthesis and therapeutic use. In their model, 28,895 20-mer candidate peptides were evaluated for their biophysical properties. Significant antibacterial activity was observed against Escherichia coli and Pseudomonas aeruginosa for the peptides with the highest overall score, indicating that the method is useful for discovering and synthesising novel candidate bactericin [74]. Deepmind developed AlphaFold in 2021. Using this computer algorithm, it is possible to predict the three-dimensional structure of a protein based solely on its amino acid sequence and to demonstrate its accuracy through further experiments [75]. AlphaFold2 was updated again later by the Deepmind team. In a recent study, David Baker et al., devised RoseTTAFold, a “three-track” neural network capable of analysing the three-dimensional structure of a sequence of proteins within 20 min [76]. The prediction of drug action sites and target protein structures is critical to designing new drugs, predicting drug-receptor interactions, and improving understanding of drug efficacy, adverse reactions, and multidrug interactions.
Nanomaterials loaded with drugs can more accurately reach the lesion area and achieve better therapeutic effect. Fluorescent nanodiamonds (FNDs), possessing extraordinary optical and electronic properties, were used for the ultrasensitive disease and virus detection and the drug delivery. Combined with AI, fluorescence-based biosensors can achieve more accurate and real-time detection and monitor under inevitable conditions or contamination [77]. This may provide new ideas for the rapid diagnosis and targeted drug therapy of OM.

5.2. Surgery

As the middle ear has a small surgical space and delicate components, it is important to identify significant landmarks to perform the procedure accurately and safely. Several teams have been working in OM-related surgery to conduct corresponding research and make certain advances (Supplementary Tables S4 and S5) [37,78,79,80,81,82,83,84,85,86,87,88,89,90,91,92]. For example, Toru Miwa et al., developed an AI system to detect cholesteatoma lesions during otoscopic surgery. It should be noted that, even though the detection accuracy of their algorithm was very limited, it was superior to non-otologists in identifying the cholesteatoma matrix (42.3% vs. 38.5%) [80]. Using a neural network, Andy S. Ding et al., were able to segment anatomical structures in the temporal bone with a submillimeter level of accuracy. As a result, this algorithm achieved a high detection accuracy (six of sixteen significant landmarks, Dice 0.681 ± 0.157) for auditory ossicles, including malleus and incus, important nerves, including cochlear nerve and facial nerve, and important blood vessels, including internal carotid artery and sigmoid sinus [81]. This technology can enhance existing image guidance technologies by informing surgeons of potential contact with critical anatomy in real-time. Using the videos of patients undergoing OM surgery under the microscope, our team developed a neural network capable of identifying the chorda tympani (CT) and the short process of the incus (SPI). The recognition accuracy of our model is fairly high. In addition to locating the facial nerve and other important anatomical structures, subsequent registration technology helps to lay the foundation for surgical navigation.

6. Risk Prediction & Postoperative Care

Although many epidemiological studies have been conducted on OM as a common respiratory complication in children, few studies have been able to analyse big data so that symptom severity, longitudinal patterns and variance in OM patients related to age and seasonality can be accurately described. In order to plan primary health care services and individualised treatments effectively, these factors are important (Supplementary Table S6) [93,94,95]. NLP was used by Anthony Dowell et al., to assess primary care incidence and service utilisation of childhood respiratory illnesses in 77,582 children. Additionally, they demonstrate that the same approach to AI-assisted analysis of big data in healthcare can be applied to other health conditions [93].
COM is typically treated with tympanoplasty. This procedure is intended to improve hearing and repair perforated eardrums. The study by Joanna Szaleniec et al., developed a model that could predict whether a patient’s hearing improved after tympanoplasty utilising feedforward artificial neural networks (FNN) and 21 independent variables (such as age, gender, ear pathology, surgical procedure description, etc.). Their algorithm was able to make 84% of correct predictions [94]. Hajime Koyama et al., developed a model to predict the air-bone gap (ABG) after tympanoplasty using random forest. They found that the correct prediction percentages of their algorithm for predicting postoperative ABG ≤ 15 or >15 dB were 81.5%. They also determined the factors, such as age and ABG, influencing the outcome of tympanoplasty [95]. ENT specialists have also become interested in the study of intelligent follow-up regarding the presence and patency of tympanostomy tubes following their placement. Xin Wang et al., used a support vector machine (SVM) to construct a neural network capable of detecting the presence of a tympanostomy tube in an otoscopic image. Notably, their model achieved an accuracy of 90%, which can be very useful for regularly scheduled follow-ups to check the status of the tympanostomy tubes [79].

7. Challenges and Future Considerations

7.1. Data & Algorithm

It is generally recognised that deep learning requires large data sets to improve its robustness. However, most of the studies are still single-center and small-sample studies. In particular, very few studies have used OM sets in NLP. In addition, many studies intentionally exclude atypical data when establishing their data sets. For example, many otoscopic images that are difficult to identify due to bleeding, blurring, or defocus are often excluded from studies related to otoscopy. While this improves the accuracy of its reporting in the paper, it significantly reduces its robustness. In terms of algorithms, we still lack high-precision and lightweight algorithms, especially for research requiring real-time tracking and recognition, such as surgery, where recognition speed plays an important role in the practical application of AI.

7.2. Application

Various measures are often used to diagnose and treat certain diseases in real clinical practice. An accurate diagnosis of OM requires the combination of the patient’s medical history, otoscopic images, audiogram, CT scan, and other inspections for a comprehensive analysis. Despite this, most researchers tend to focus on a specific method for using AI to assist in diagnosing and treating OM patients. A comprehensive AI-assisted diagnosis, treatment, and disease management system must be developed in the future that combines multiple measures to enhance the potential of AI-assisted medicine. Additionally, it should be noted that most current AI-assisted OM studies are based on retrospective data, but the real world is filled with uncertainties. Future relevant studies should demonstrate the feasibility of AI in practice by applying it to real-world clinical scenarios. Lastly, when AI is applied to solve relevant clinical problems, the “black box” nature of the deep learning network model makes it difficult for users to understand its decision-making mechanism, which not only impedes the optimisation of model structures and safety enhancement but also greatly increases the cost of training and parameter adjustment. While the Grad-CAM can now be used to explain the regions of interest and the decision principles in some network models, the current Explainable Artificial Intelligence (XAI) methods still have significant limitations [96]. We should continue to develop appropriate methods for interpreting deep learning models vigorously in the future so that clinicians may better understand and use relevant AI tools and provide a scientific basis for optimising the structure of deep learning models.

7.3. Privacy & Regulation

Patient privacy is an important consideration in developing AI and other information technologies. However, sharing information is an inevitable trend in establishing multi-center and big datasets. Consequently, ensuring the privacy and confidentiality of patient information is an important issue that needs to be addressed urgently. Known also as distributed ledger technology, blockchain technology is characterised by its decentralisation, openness, and transparency, which allows everyone to participate in database records [97]. In 2016, Google proposed a new AI technology called federated learning. During the big data exchange, it can carry out efficient ML among multiple participants or computing nodes to ensure information security and protect the privacy of terminal information and personal information [98]. Decentralised platforms provide important solutions to protect the privacy of patient data in the future.
It is also urgent to standardise deep learning research. In deep learning research, there tends to be a process-based approach, with rules for data collection, the training of models, and the evaluation of the results. Unfortunately, many studies have different criteria for including and excluding original data and related indicators for evaluating their effects. The selection of relevant training models and the direction of network reconstruction remains a subjective judgment. A major reason for this is that there are no standardised guidelines for the research on medical AI. We can build more secure and effective AI systems and realise realistic transformations and applications of these promising technologies only by establishing standardised deep learning process guidelines and innovation.

8. Conclusions

The articles we reviewed preliminarily demonstrated the potential of AI methods in diagnosis, treatment, and prognosis of OM. Medical history information, endoscopic otoscopy, radiology, audiology, laboratory data, genetic data, and surgical videos have been used to create AI systems to effectively diagnose patients, aid treatment, and predict treatment response and disease activity. As a result, AI provides an advantage to the quality of care that OM patients currently receive, resulting in improved quality of life and improved OM-related health outcomes. However, most of these studies have been conducted in the laboratory. Therefore, further validation of these results in more realistic Settings and on a larger scale is needed to assess their effectiveness in the general OM population.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/diagnostics13132309/s1, Table S1: ML-related articles for the diagnosis of OM with otoscopy; Table S2: ML-related articles for the diagnosis of OM based on CT, X-ray, OCT and tympanogram; Table S3: Articles for diagnosis of OM using NLP; Table S4: Articles related to the application of ML in OM surgery; Table S5: Articles related to the application of ML in the registration of OM surgery; Table S6: Articles related to the application of ML in the health management of OM.

Author Contributions

X.D.: designing research direction, organizing materials, writing papers; Y.H. contribution: helping to design research direction, helping organizing materials Y.Z. and X.T.: helping to design research direction, helping to write papers; G.F. and Z.G. contribution: review and revise papers, guidance of research direction and article writing. All authors have read and agreed to the published version of the manuscript.

Funding

This study is supported by ‘National key research and development program’ (Grant number: 2019YFB311801) and ‘the Fundamental Research Funds for the Central Universities’ (3332021010).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AbbreviationsExtensions
ENTEar-nose-throat
OMOtitis media
OMEOtitis media effusion
AOMAcute otitis media
COMChronic otitis media
CSOMChronic suppurative otitis media
PCDPrimary ciliated dyskinesia
TMTympanic membrane
ABGAir-bone gap
FNDsFluorescent nanodiamonds
VPOVideo pneumatic otoscope
OCTOptical coherence tomography
WBTWideband tympanometry
AIArtificial intelligence
CNNConvolutional neural networks
MLMachine learning
DLDeep learning
CVComputer vision
NLPNatural language processing
XAIExplainable artificial intelligence
Grad-CAMGradient-weighted Class Activation Mapping
BOWBag-of-words model
NCANeighborhood component analysis
FNNFeedforward artificial neural networks

References

  1. Auinger, P.; Lanphear, B.P.; Kalkwarf, H.J.; Mansour, M.E. Trends in Otitis Media Among Children in the United States. Pediatrics 2003, 112, 514–520. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Blomgren, K.; Pitkäranta, A. Is it possible to diagnose acute otitis media accurately in primary health care? Fam. Pract. 2003, 20, 524–527. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Pichichero, M.E.; Poole, M.D. Assessing Diagnostic Accuracy and Tympanocentesis Skills in the Management of Otitis Media. Arch. Pediatr. Adolesc. Med. 2001, 155, 1137–1142. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Brennan-Jones, C.G.; Whitehouse, A.J.; Calder, S.D.; Costa, C.D.; Eikelboom, R.H.; Swanepoel, D.W.; Jamieson, S.E. Does Otitis Media Affect Later Language Ability? A Prospective Birth Cohort Study. J. Speech Lang Hear Res. 2020, 63, 2441–2452. [Google Scholar] [CrossRef]
  5. Abada, R.L.; Mansouri, I.; Maamri, M.; Kadiri, F. Complications of chronic otitis media. Ann. Otolaryngol. Chir. Cervicofac. 2009, 126, 1–5. [Google Scholar] [CrossRef]
  6. Dagan, R. Clinical significance of resistant organisms in otitis media. Pediatr. Infect. Dis. J. 2000, 19, 378–382. [Google Scholar] [CrossRef]
  7. Bosch, J.H.W.-V.D.; Stolk, E.A.; Francois, M.; Gasparini, R.; Brosa, M. The health care burden and societal impact of acute otitis media in seven European countries: Results of an Internet survey. Vaccine 2010, 28 (Suppl. S6), G39–G52. [Google Scholar]
  8. Qureshi, S.A.; Raza, S.E.A.; Hussain, L.; Malibari, A.A.; Nour, M.K.; Rehman, A.U.; Al-Wesabi, F.N.; Hilal, A.M. Intelligent Ultra-Light Deep Learning Model for Multi-Class Brain Tumor Detection. Appl. Sci. 2022, 12, 3715. [Google Scholar] [CrossRef]
  9. Handelman, G.S.; Kok, H.K.; Chandra, R.V.; Razavi, A.H.; Lee, M.J.; Asadi, H. eDoctor: Machine learning and the future of medicine. J. Intern. Med. 2018, 284, 603–619. [Google Scholar] [CrossRef]
  10. Rajkomar, A.; Dean, J.; Kohane, I. Machine learning in medicine. New Engl. J. Med. 2019, 380, 1347–1358. [Google Scholar] [CrossRef]
  11. Deo, R.C. Machine Learning in Medicine. Circulation 2015, 132, 1920–1930. [Google Scholar] [CrossRef] [Green Version]
  12. Sandström, J.; Myburgh, H.; Laurent, C.; Swanepoel, D.W.; Lundberg, T. A Machine Learning Approach to Screen for Otitis Media Using Digital Otoscope Images Labelled by an Expert Panel. Diagnostics 2022, 12, 1318. [Google Scholar] [CrossRef]
  13. Chen, Y.-C.; Chu, Y.-C.; Huang, C.-Y.; Lee, Y.-T.; Lee, W.-Y.; Hsu, C.-Y.; Yang, A.C.; Liao, W.-H.; Cheng, Y.-F. Smartphone-based artificial intelligence using a transfer learning algorithm for the detection and diagnosis of middle ear diseases: A retrospective deep learning study. Eclinicalmedicine 2022, 51, 101543. [Google Scholar] [CrossRef]
  14. Mao, C.; Li, A.; Hu, J.; Wang, P.; Peng, D.; Wang, J.; Sun, Y. Efficient and accurate diagnosis of otomycosis using an ensemble deep-learning model. Front. Mol. Biosci. 2022, 9, 951432. [Google Scholar] [CrossRef]
  15. Crowson, M.G.; Bates, D.W.; Suresh, K.; Cohen, M.S.; Hartnick, C.J. “Human vs Machine” Validation of a Deep Learning Algorithm for Pediatric Middle Ear Infection Diagnosis. Otolaryngol. Neck Surg. 2022, 1945998221119156. [Google Scholar] [CrossRef]
  16. Mohammed, K.K.; Hassanien, A.E.; Afify, H.M. Classification of Ear Imagery Database using Bayesian Optimization based on CNN-LSTM Architecture. J. Digit. Imaging 2022, 35, 947–961. [Google Scholar] [CrossRef]
  17. Choi, Y.; Chae, J.; Park, K.; Hur, J.; Kweon, J.; Ahn, J.H. Automated multi-class classification for prediction of tympanic membrane changes with deep learning models. PLoS ONE 2022, 17, e0275846. [Google Scholar] [CrossRef]
  18. Habib, A.-R.; Crossland, G.; Patel, H.; Wong, E.; Kong, K.; Gunasekera, H.; Richards, B.; Caffery, L.; Perry, C.; Sacks, R.; et al. An Artificial Intelligence Computer-vision Algorithm to Triage Otoscopic Images From Australian Aboriginal and Torres Strait Islander Children. Otol. Neurotol. 2022, 43, 481–488. [Google Scholar] [CrossRef]
  19. Zeng, J.; Kang, W.; Chen, S.; Lin, Y.; Deng, W.; Wang, Y.; Chen, G.; Ma, K.; Zhao, F.; Zheng, Y.; et al. A Deep Learning Approach to Predict Conductive Hearing Loss in Patients with Otitis Media with Effusion Using Otoscopic Images. JAMA Otolaryngol. Neck Surg. 2022, 148, 612–620. [Google Scholar] [CrossRef]
  20. Byun, H.; Park, C.J.; Oh, S.J.; Chung, M.J.; Cho, B.H.; Cho, Y.-S. Automatic Prediction of Conductive Hearing Loss Using Video Pneumatic Otoscopy and Deep Learning Algorithm. Ear Hear. 2022, 43, 1563–1573. [Google Scholar] [CrossRef]
  21. Viscaino, M.; Talamilla, M.; Maass, J.C.; Henríquez, P.; Délano, P.H.; Cheein, C.A.; Cheein, F.A. Color Dependence Analysis in a CNN-Based Computer-Aided Diagnosis System for Middle and External Ear Diseases. Diagnostics 2022, 12, 917. [Google Scholar] [CrossRef] [PubMed]
  22. Alhudhaif, A.; Cömert, Z.; Polat, K. Otitis media detection using tympanic membrane images with a novel multi-class machine learning algorithm. PeerJ Comput. Sci. 2021, 7, e405. [Google Scholar] [CrossRef] [PubMed]
  23. Cai, Y.; Yu, J.-G.; Chen, Y.; Liu, C.; Xiao, L.; Grais, E.M.; Zhao, F.; Lan, L.; Zeng, S.; Zeng, J.; et al. Investigating the use of a two-stage attention-aware convolutional neural network for the automated diagnosis of otitis media from tympanic membrane images: A prediction model development and validation study. BMJ Open 2021, 11, e041139. [Google Scholar] [CrossRef] [PubMed]
  24. Byun, H.; Yu, S.; Oh, J.; Bae, J.; Yoon, M.S.; Lee, S.H.; Chung, J.H.; Kim, T.H. An Assistive Role of a Machine Learning Network in Diagnosis of Middle Ear Diseases. J. Clin. Med. 2021, 10, 3198. [Google Scholar] [CrossRef]
  25. Crowson, M.G.; Hartnick, C.J.; Diercks, G.R.; Gallagher, T.Q.; Fracchia, M.S.; Setlur, J.; Cohen, M.S. Machine Learning for Accurate Intraoperative Pediatric Middle Ear Effusion Diagnosis. Pediatrics 2021, 147, e2020034546. [Google Scholar] [CrossRef]
  26. Pham, V.-T.; Tran, T.-T.; Wang, P.-C.; Chen, P.-Y.; Lo, M.-T. EAR-UNet: A deep learning-based approach for segmentation of tympanic membranes from otoscopic images. Artif. Intell. Med. 2021, 115, 102065. [Google Scholar] [CrossRef]
  27. Sundgaard, J.V.; Harte, J.; Bray, P.; Laugesen, S.; Kamide, Y.; Tanaka, C.; Paulsen, R.R.; Christensen, A.N. Deep metric learning for otitis media classification. Med. Image Anal. 2021, 71, 102034. [Google Scholar] [CrossRef]
  28. Zeng, X.; Jiang, Z.; Luo, W.; Li, H.; Li, H.; Li, G.; Shi, J.; Wu, K.; Liu, T.; Lin, X.; et al. Efficient and accurate identification of ear diseases using an ensemble deep learning model. Sci. Rep. 2021, 11, 10839. [Google Scholar] [CrossRef]
  29. Kashani, R.G.; Młyńczak, M.C.; Zarabanda, D.; Solis-Pazmino, P.; Huland, D.M.; Ahmad, I.N.; Singh, S.P.; Valdez, T.A. Shortwave infrared otoscopy for diagnosis of middle ear effusions: A machine-learning-based approach. Sci. Rep. 2021, 11, 12509. [Google Scholar] [CrossRef]
  30. Tsutsumi, K.; Goshtasbi, K.; Risbud, A.; Khosravi, P.; Pang, J.C.; Lin, H.W.; Djalilian, H.R.; Abouzari, M. A Web-Based Deep Learning Model for Automated Diagnosis of Otoscopic Images. Otol. Neurotol. 2021, 42, e1382–e1388. [Google Scholar] [CrossRef]
  31. Başaran, E.; Cömert, Z.; Çelik, Y. Convolutional neural network approach for automatic tympanic membrane detection and classification. Biomed. Signal Process. Control. 2020, 56, 101734. [Google Scholar] [CrossRef]
  32. Cha, D.; Pae, C.; A Lee, S.; Na, G.; Hur, Y.K.; Lee, H.Y.; Cho, A.R.; Cho, Y.J.; Gil Han, S.; Kim, S.H.; et al. Differential Biases and Variabilities of Deep Learning–Based Artificial Intelligence and Human Experts in Clinical Diagnosis: Retrospective Cohort and Survey Study. JMIR Public Heal. Surveill. 2021, 9, e33049. [Google Scholar] [CrossRef]
  33. Wang, W.; Tamhane, A.; Santos, C.; Rzasa, J.R.; Clark, J.H.; Canares, T.L.; Unberath, M. Pediatric Otoscopy Video Screening With Shift Contrastive Anomaly Detection. Front. Digit. Health 2022, 3, 810427. [Google Scholar] [CrossRef]
  34. Binol, H.; Niazi, M.K.K.; Elmaraghy, C.; Moberly, A.C.; Gurcan, M.N. OtoXNet-automated identification of eardrum diseases from otoscope videos: A deep learning study for video-representing images. Neural Comput. Appl. 2022, 14, 12197–12210. [Google Scholar] [CrossRef]
  35. Wu, Z.; Lin, Z.; Li, L.; Pan, H.; Chen, G.; Fu, Y.; Qiu, Q. Deep Learning for Classification of Pediatric Otitis Media. Laryngoscope 2021, 131, E2344–E2351. [Google Scholar] [CrossRef]
  36. Khan, M.A.; Kwon, S.; Choo, J.; Hong, S.J.; Kang, S.H.; Park, I.-H.; Kim, S.K. Automatic detection of tympanic membrane and middle ear infection from oto-endoscopic images via convolutional neural networks. Neural Netw. 2020, 126, 384–394. [Google Scholar] [CrossRef]
  37. Camalan, S.; Niazi, M.K.K.; Moberly, A.C.; Teknos, T.; Essig, G.; Elmaraghy, C.; Taj-Schaal, N.; Gurcan, M.N. OtoMatch: Content-based eardrum image retrieval using deep learning. PLoS ONE 2020, 15, e0232776. [Google Scholar] [CrossRef]
  38. Viscaino, M.; Maass, J.C.; Delano, P.H.; Torrente, M.; Stott, C.; Cheein, F.A. Computer-aided diagnosis of external and middle ear conditions: A machine learning approach. PLoS ONE 2020, 15, e0229226. [Google Scholar] [CrossRef]
  39. Pham, V.-T.; Tran, T.-T.; Wang, P.-C.; Lo, M.-T. Tympanic membrane segmentation in otoscopic images based on fully convolutional network with active contour loss. Signal Image Video Process. 2021, 15, 519–527. [Google Scholar] [CrossRef]
  40. Habib, A.-R.; Wong, E.; Sacks, R.; Singh, N. Artificial intelligence to detect tympanic membrane perforations. J. Laryngol. Otol. 2020, 134, 311–315. [Google Scholar] [CrossRef]
  41. Binol, H.; Moberly, A.C.; Niazi, M.K.K.; Essig, G.; Shah, J.; Elmaraghy, C.; Teknos, T.; Taj-Schaal, N.; Yu, L.; Gurcan, M.N. Decision Fusion on Image Analysis and Tympanometry to Detect Eardrum Abnormalities; SPIE Medical Imaging: Houston, TX, USA, 2020. [Google Scholar]
  42. Binol, H.; Moberly, A.C.; Niazi, M.K.K.; Essig, G.; Shah, J.; Elmaraghy, C.; Teknos, T.; Taj-Schaal, N.; Yu, L.; Gurcan, M.N. SelectStitch: Automated Frame Segmentation and Stitching to Create Composite Images from Otoscope Video Clips; Cold Spring Harbor Laboratory Press: Cold Spring Harbor, NY, USA, 2020. [Google Scholar]
  43. Cha, D.; Pae, C.; Seong, S.-B.; Choi, J.Y.; Park, H.-J. Automated diagnosis of ear disease using ensemble deep learning with a big otoendoscopy image database. Ebiomedicine 2019, 45, 606–614. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  44. Cmert, Z. Fusing fine-tuned deep features for recognizing different tympanic membranes. Biocybern. Biomed. Eng. 2019, 40, 40–51. [Google Scholar]
  45. Livingstone, D.; Chau, J. Otoscopic diagnosis using computer vision: An automated machine learning approach. Laryngoscope 2020, 130, 1408–1413. [Google Scholar] [CrossRef] [PubMed]
  46. Lee, J.Y.; Choi, S.H.; Chung, J.W. Automated Classification of the Tympanic Membrane Using a Convolutional Neural Network. Appl. Sci. 2019, 9, 1827. [Google Scholar] [CrossRef] [Green Version]
  47. Seok, J.; Song, J.J.; Koo, J.W.; Kim, H.C.; Choi, B.Y. The semantic segmentation approach for normal and pathologic tympanic membrane using deep learning. BioRxiv 2019, 515007. [Google Scholar]
  48. Tran, T.T.; Fang, T.Y.; Pham, V.T.; Lin, C.; Wang, P.C.; Lo, M.T. Development of an Automatic Diagnostic Algorithm for Pediatric Otitis Media. Otol. Neurotol. 2018, 39, 1060–1065. [Google Scholar] [CrossRef]
  49. Myburgh, H.C.; Jose, S.; Swanepoel, D.W.; Laurent, C. Towards low cost automated smartphone- and cloud-based otitis media diagnosis. Biomed. Signal Process. Control. 2018, 39, 34–52. [Google Scholar] [CrossRef] [Green Version]
  50. Senaras, C.; Moberly, A.C.; Teknos, T.; Essig, G.; Elmaraghy, C.; Taj-Schaal, N.; Yu, L.; Gurcan, M. Autoscope: Automated otoscopy image analysis to diagnose ear pathology and use of clinically motivated eardrum features. In Med Imaging 2017: Compu-Aided Diagnosis; SPIE Medical Imaging: Orlando, Fl, USA, 2017; Volume 10134. [Google Scholar]
  51. Kasher, M.S. Otitis Media Analysis—An Automated Feature Extraction and Image Classification System; Helsinki Metropolia University of Applied Science: Helsinki, Finland, 2018. [Google Scholar]
  52. Myburgh, H.C.; van Zijl, W.H.; Swanepoel, D.W.; Hellström, S.; Laurent, C. Otitis Media Diagnosis for Developing Countries Using Tympanic Membrane Image-Analysis. Ebiomedicine 2016, 5, 156–160. [Google Scholar] [CrossRef] [Green Version]
  53. Shie, C.-K.; Chuang, C.-H.; Chou, C.-N.; Wu, M.-H.; Chang, E.Y. Transfer representation learning for medical image analysis. 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Milan, Italy, 25-29 August; 2015; pp. 711–714. [Google Scholar] [CrossRef]
  54. Shie, C.-K.; Chang, H.-T.; Fan, F.-C.; Chen, C.-J.; Fang, T.-Y.; Wang, P.-C. A hybrid feature-based segmentation and classification system for the computer aided self-diagnosis of otitis media. In Proceedings of the Engineering in Medicine and Biology Society (EMBC), 36th Annual International Conference of the IEEE, Chicago, IL, USA, 27–31 August 2014. [Google Scholar]
  55. Monroy, G.L.; Won, J.; Dsouza, R.; Pande, P.; Hill, M.C.; Porter, R.G.; Novak, M.A.; Spillman, D.R.; Boppart, S.A. Automated classification platform for the identification of otitis media using optical coherence tomography. NPJ Digit. Med. 2019, 2, 22. [Google Scholar] [CrossRef] [Green Version]
  56. Monroy, G.L.; Won, J.; Shi, J.; Hill, M.C.; Porter, R.G.; Novak, M.A.; Hong, W.; Khampang, P.; Kerschner, J.E.; Spillman, D.R.; et al. Automated classification of otitis media with OCT: Augmenting pediatric image datasets with gold-standard animal model data. Biomed. Opt. Express 2022, 13, 3601. [Google Scholar] [CrossRef]
  57. Wang, Y.-M.; Li, Y.; Cheng, Y.-S.; He, Z.-Y.; Yang, J.-M.; Xu, J.-H.; Chi, Z.-C.; Chi, F.-L.; Ren, D.-D. Deep Learning in Automated Region Proposal and Diagnosis of Chronic Otitis Media Based on Computed Tomography. Ear Hear. 2020, 41, 669–677. [Google Scholar] [CrossRef]
  58. Wang, Z.; Song, J.; Su, R.; Hou, M.; Qi, M.; Zhang, J.; Wu, X. Structure-aware deep learning for chronic middle ear disease. Expert Syst. Appl. 2022, 194, 116519. [Google Scholar] [CrossRef]
  59. Eroğlu, O.; Eroğlu, Y.; Yıldırım, M.; Karlıdag, T.; Çınar, A.; Akyiğit, A.; Kaygusuz, I.; Yıldırım, Y.; Keleş, E.; Yalçın, Ş. Is it useful to use computerized tomography image-based artificial intelligence modelling in the differential diagnosis of chronic otitis media with and without cholesteatoma? Am. J. Otolaryngol. 2022, 43, 103395. [Google Scholar] [CrossRef]
  60. Takahashi, M.; Noda, K.; Yoshida, K.; Tsuchida, K.; Yui, R.; Nakazawa, T.; Kurihara, S.; Baba, A.; Motegi, M.; Yamamoto, K.; et al. Preoperative prediction by artificial intelligence for mastoid extension in pars flaccida cholesteatoma using temporal bone high-resolution computed tomography: A retrospective study. PLoS ONE 2022, 17, e0273915. [Google Scholar] [CrossRef]
  61. Duan, B.; Guo, Z.; Pan, L.; Xu, Z.; Chen, W. Temporal bone CT-based deep learning models for differential diagnosis of primary ciliary dyskinesia related otitis media and simple otitis media with effusion. Am. J. Transl. Res. 2022, 14, 4728–4735. [Google Scholar]
  62. Lee, K.J.; Ryoo, I.; Choi, D.; Sunwoo, L.; You, S.-H.; Na Jung, H. Performance of deep learning to detect mastoiditis using multiple conventional radiographs of mastoid. PLoS ONE 2020, 15, e0241796. [Google Scholar] [CrossRef]
  63. Qureshi, S.A.; Hussain, L.; Ibrar, U.; Alabdulkreem, E.; Nour, M.K.; Alqahtani, M.S.; Nafie, F.M.; Mohamed, A.; Mohammed, G.P.; Duong, T.Q. Radiogenomic classification for MGMT promoter methylation status using multi-omics fused feature space for least invasive diagnosis through mpMRI scans. Sci. Rep. 2023, 13, 3291. [Google Scholar] [CrossRef]
  64. Feng, L.; Liu, Z.; Li, C.; Li, Z.; Lou, X.; Shao, L.; Wang, Y.; Huang, Y.; Chen, H.; Pang, X.; et al. Development and validation of a radiopathomics model to predict pathological complete response to neoadjuvant chemoradiotherapy in locally advanced rectal cancer: A multicentre observational study. Lancet Digit. Heal. 2022, 4, e8–e17. [Google Scholar] [CrossRef]
  65. Sundgaard, J.V.; Bray, P.; Laugesen, S.; Harte, J.; Kamide, Y.; Tanaka, C.; Christensen, A.N.; Paulsen, R.R. A Deep Learning Approach for Detecting Otitis Media From Wideband Tympanometry Measurements. IEEE J. Biomed. Heal Inform. 2022, 26, 2974–2982. [Google Scholar] [CrossRef]
  66. Merchant, G.R.; Al-Salim, S.; Tempero, R.M.; Fitzpatrick, D.; Neely, S.T. Improving the Differential Diagnosis of Otitis Media With Effusion Using Wideband Acoustic Immittance. Ear Hear. 2021, 42, 1183–1194. [Google Scholar] [CrossRef]
  67. Grais, E.M.; Wang, X.; Wang, J.; Zhao, F.; Jiang, W.; Cai, Y.; Zhang, L.; Lin, Q.; Yang, H. Analysing wideband absorbance immittance in normal and ears with otitis media with effusion using machine learning. Sci. Rep. 2021, 11, 10643. [Google Scholar] [CrossRef] [PubMed]
  68. Kuruvilla, A.; Shaikh, N.; Hoberman, A.; Kovačević, J. Automated Diagnosis of Otitis Media: Vocabulary and Grammar. Int. J. Biomed. Imaging 2013, 2013, 327515. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  69. Binol, H.; Niazi, M.K.K.; Elmaraghy, C.; Moberly, A.C.; Gurcan, M.N. Automated Video Summarization and Label Assignment for Otoscopy Videos Using Deep Learning and Natural Language Processing. In Imaging Informatics for Healthcare, Research, and Applications; SPIE: Bellingham, WA, USA, 2021. [Google Scholar] [CrossRef]
  70. Herigon, J.C.; Kimia, A.; Harper, M. 1358 Using natural language processing to optimize case ascertainment of acute otitis media in a large, state-wide pediatric practice network. Open Forum Infect. Dis. 2020, 7, S690–S691. [Google Scholar] [CrossRef]
  71. Kuruvilla, A.; Li, J.; Yeomans, P.H.; Quelhas, P.; Shaikh, N.; Hoberman, N.; Kovačević, J. Otitis media vocabulary and grammary. Proc. Int. Conf. Image Proc. 2012, 2012, 2845–2848. [Google Scholar] [PubMed] [Green Version]
  72. Macesic, N.; Bear Don’t Walk IV, O.J.; Pe’er, I.; Tatonetti, N.P.; Peleg, A.Y.; Uhlemann, A.C. Predicting Phenotypic Polymyxin Resistance in Klebsiella pneumoniae through Machine Learning Analysis of Genomic Data. mSystems 2020, 5, e00656-19. [Google Scholar] [CrossRef] [PubMed]
  73. Mansbach, R.A.; Leus, I.V.; Mehla, J.; Lopez, C.A.; Walker, J.K.; Rybenkov, V.V.; Hengartner, N.W.; Hengartner, H.I.; Gnanakaran, S. Machine Learning Algorithm Identifies an Antibiotic Vocabulary for Permeating Gram-Negative Bacteria. J. Chem. Inf. Model. 2020, 60, 2838–2847. [Google Scholar] [CrossRef]
  74. Fields, F.R.; Freed, S.D.; Carothers, K.E.; Hamid, M.N.; Hammers, D.E.; Ross, J.N.; Kalwajtys, V.R.; Gonzalez, A.J.; Hildreth, A.D.; Friedberg, I.; et al. Novel antimicrobial peptide discovery using machine learning and biophysical selection of minimal bacteriocin domains. Drug Dev. Res. 2020, 81, 43–51. [Google Scholar] [CrossRef]
  75. Senior, A.W.; Evans, R.; Jumper, J.; Kirkpatrick, J.; Sifre, L.; Green, T.; Qin, C.; Žídek, A.; Nelson, A.W.R.; Bridgland, A.; et al. Improved protein structure prediction using potentials from deep learning. Nature 2020, 577, 706–710. [Google Scholar] [CrossRef]
  76. Wang, J.; Lisanza, S.; Juergens, D.; Tischer, D.; Watson, J.L.; Castro, K.M.; Ragotte, R.; Saragovi, A.; Milles, L.F.; Baek, M.; et al. Scaffolding protein functional sites using deep learning. Science 2022, 377, 387–394. [Google Scholar] [CrossRef]
  77. Qureshi, S.A.; Hsiao, W.W.-W.; Hussain, L.; Aman, H.; Le, T.-N.; Rafique, M. Recent Development of Fluorescent Nanodiamonds for Optical Biosensing and Disease Diagnosis. Biosensors 2022, 12, 1181. [Google Scholar] [CrossRef]
  78. Livingstone, D.; Talai, A.S.; Chau, J.; Forkert, N.D. Building an Otoscopic screening prototype tool using deep learning. J. Otolaryngol.-Head Neck Surg. 2019, 48, 66. [Google Scholar] [CrossRef] [Green Version]
  79. Wang, X.; Valdez, T.A.; Bi, J. Detecting tympanostomy tubes from otoscopic images via offline and online training. Comput. Biol. Med. 2015, 61, 107–118. [Google Scholar] [CrossRef]
  80. Miwa, T.; Minoda, R.; Yamaguchi, T.; Kita, S.-I.; Osaka, K.; Takeda, H.; Kanemaru, S.-I.; Omori, K. Application of artificial intelligence using a convolutional neural network for detecting cholesteatoma in endoscopic enhanced images. Auris Nasus Larynx 2022, 49, 11–17. [Google Scholar] [CrossRef]
  81. Ding, A.S.; Lu, A.; Li, Z.; Galaiya, D.; Siewerdsen, J.H.; Taylor, R.H.; Creighton, F.X. Automated Registration-Based Temporal Bone Computed Tomography Segmentation for Applications in Neurotologic Surgery. Otolaryngol. Neck Surg. 2022, 167, 133–140. [Google Scholar] [CrossRef]
  82. Dong, B.; Lu, C.; Hu, X.; Zhao, Y.; He, H.; Wang, J. Towards accurate facial nerve segmentation with decoupling optimization. Phys. Med. Biol. 2022, 67, 065007. [Google Scholar] [CrossRef]
  83. Neves, C.A.; Tran, E.D.; Kessler, I.M.; Blevins, N.H. Fully automated preoperative segmentation of temporal bone structures from clinical CT scans. Sci. Rep. 2021, 11, 116. [Google Scholar] [CrossRef]
  84. Nikan, S.; Van Osch, K.; Bartling, M.; Allen, D.G.; Rohani, S.A.; Connors, B.; Agrawal, S.K.; Ladak, H.M. PWD-3DNet: A Deep Learning-Based Fully-Automated Segmentation of Multiple Structures on Temporal Bone CT Scans. IEEE Trans. Image Process. 2021, 30, 739–753. [Google Scholar] [CrossRef]
  85. Lv, Y.; Ke, J.; Xu, Y.; Shen, Y.; Wang, J.; Wang, J. Automatic segmentation of temporal bone structures from clinical conventional CT using a CNN approach. Int. J. Med. Robot. Comput. Assist. Surg. 2021, 17, e2229. [Google Scholar] [CrossRef]
  86. Jeevakala, S.; Sreelakshmi, C.; Ram, K.; Rangasami, R.; Sivaprakasam, M. Artificial intelligence in detection and segmentation of internal auditory canal and its nerves using deep learning techniques. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 1859–1867. [Google Scholar] [CrossRef]
  87. Fauser, J.; Stenin, I.; Bauer, M.; Hsu, W.-H.; Kristin, J.; Klenzner, T.; Schipper, J.; Mukhopadhyay, A. Toward an automatic preoperative pipeline for image-guided temporal bone surgery. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 967–976. [Google Scholar] [CrossRef]
  88. Gare, B.M.; Hudson, T.; Rohani, S.A.; Allen, D.G.; Agrawal, S.K.; Ladak, H.M. Multi-atlas segmentation of the facial nerve from clinical CT for virtual reality simulators. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 259–267. [Google Scholar] [CrossRef] [PubMed]
  89. Powell, K.A.; Liang, T.; Hittle, B.; Stredney, D.; Kerwin, T.; Wiet, G.J. Atlas-Based Segmentation of Temporal Bone Anatomy. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 1937–1944. [Google Scholar] [CrossRef] [PubMed]
  90. Lu, P.; Barazzetti, L.; Chandran, V.; Gavaghan, K.; Weber, S.; Gerber, N.; Reyes, M. Highly Accurate Facial Nerve Segmentation Refinement From CBCT/CT Imaging Using a Super-Resolution Classification Approach. IEEE Trans. Biomed. Eng. 2018, 65, 178–188. [Google Scholar] [CrossRef] [PubMed]
  91. Becker, M.; Kirschner, M.; Sakas, G. Segmentation of risk structures for otologic surgery using the probabilistic active shape model. ProcSPIE. 2014, 9036, 9036–9037. [Google Scholar]
  92. Noble, J.H.; Labadie, R.F.; Majdani, O.; Dawant, B.M. Automatic Segmentation of Intracochlear Anatomy in Conventional CT. IEEE Trans. Biomed. Eng. 2011, 58, 2625–2632. [Google Scholar] [CrossRef] [Green Version]
  93. Dowell, A.; Darlow, B.; Macrae, J.; Stubbe, M.; Turner, N.; McBain, L. Childhood respiratory illness presentation and service utilisation in primary care: A six-year cohort study in Wellington, New Zealand, using natural language processing (NLP) software. BMJ Open 2017, 7, e017146. [Google Scholar] [CrossRef] [Green Version]
  94. Szaleniec, J.; Wiatr, M.; Szaleniec, M.; Składzień, J.; Tomik, J.; Oleś, K.; Tadeusiewicz, R. Artificial neural network modelling of the results of tympanoplasty in chronic suppurative otitis media patients. Comput. Biol. Med. 2013, 43, 16–22. [Google Scholar] [CrossRef]
  95. Koyama, H.; Kashio, A.; Uranaka, T.; Matsumoto, Y.; Yamasoba, T. Application of Machine Learning to Predict Hearing Outcomes of Tympanoplasty. Laryngoscope 2022. [Google Scholar] [CrossRef]
  96. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization. arXiv 2019, arXiv:1610.02391. [Google Scholar]
  97. Saeed, H.; Malik, H.; Bashir, U.; Ahmad, A.; Riaz, S.; Ilyas, M.; Bukhari, W.A.; Khan, M.I.A. Blockchain technology in healthcare: A systematic review. PLoS ONE 2022, 17, e0266462. [Google Scholar] [CrossRef]
  98. Rischke, R.; Schneider, L.; Müller, K.; Samek, W.; Schwendicke, F.; Krois, J. Federated Learning in Dentistry: Chances and Challenges. J. Dent. Res. 2022, 101, 1269–1273. [Google Scholar] [CrossRef]
Figure 1. Flow chart of application of AI in diagnosis, treatment and healthcare of OM. Through collecting large amounts of data from hospitals and daily life and training with AI and ML, it can be realized the diagnosis, risk stratification, disease management and treatment of OM.
Figure 1. Flow chart of application of AI in diagnosis, treatment and healthcare of OM. Through collecting large amounts of data from hospitals and daily life and training with AI and ML, it can be realized the diagnosis, risk stratification, disease management and treatment of OM.
Diagnostics 13 02309 g001
Figure 2. Concepts of artificial intelligence (AI), machine learning (ML), and deep learning (DL). AI is equivalent to the human body, ML equals to the human brain, and DL corresponds to the complex neural circuit consisted of neurons. When AI is applied to vision, it refers to computer vision (CV). And when applied to speech processing, it represents natural language processing (NLP).
Figure 2. Concepts of artificial intelligence (AI), machine learning (ML), and deep learning (DL). AI is equivalent to the human body, ML equals to the human brain, and DL corresponds to the complex neural circuit consisted of neurons. When AI is applied to vision, it refers to computer vision (CV). And when applied to speech processing, it represents natural language processing (NLP).
Diagnostics 13 02309 g002
Figure 3. PRISMA flow diagram depicting the different phases of the review selection process.
Figure 3. PRISMA flow diagram depicting the different phases of the review selection process.
Diagnostics 13 02309 g003
Table 1. Types of machine learning algorithms.
Table 1. Types of machine learning algorithms.
ML AlgorithmsDefinition
Supervised learningThe model is trained based on the labeled data to make the prediction. It is mainly used for classification (continuous data) and regression problems (discrete data).
Ensemble learningGenerate multiple models and then combine these models according to a certain method. It is mainly applied to improve model performance or reduce the possibility of improper model selection.
Bagging-Random forest
(RF)
As a type of ensembled learning, it is based on decision trees. Each of decision trees can produce an independent and de-correlated output. And then the final result is determined by the rule of “majority-vote”. Each decision tree has the equal weight.
Boosting-Gradient boosting machine (GBM)The basic unit is the decision tree. The performance of the first decision tree is only due to random guessing. The decision tree model established later are based on the prior one by adjusting parameters to gradually improve the accuracy of subsequent models. And then the final result is determined by considering comprehensively the predicted values of all decision trees. Base learner (decision tree) with better performance has higher weight.
Support vector machine
(SVM)
In a given set of training samples, with functional distance as constraint condition and geometric distance as objective function, the maximum-margin hyperplane with the best generalization ability and the strongest robust is determined, so as to realize the binary classification of data.
Unsupervised learningThe training data is not labeled and the goal is to find patterns in the data. It is mainly used to solve clustering and dimension reduction problems.
ClusteringThe process of sorting data into different classes or clusters based on mathematical relevance.
Dimensionality reductionIt means converting high-dimensional data into a lower representation with fewer features and is often used for data visualization or data preprocessing.
Reinforcement learningThrough interaction with its environment and trial and error, the training model learns the environment-to-action reflex that maximize cumulative returns.
Transfer learning (TL)Transfer the learned knowledge from a domain to another and save a deal of time and computing resources for the training.
Deep learning (DL)Mimicking the hierarchical network of neurons in the brain and using multiple layers of data processing, the model can automatically detect the required features and predict the result. But it requires larger quantities of data and advanced computational capacity.
Feedforward neural networks
(FNN)
Arranged in layers, each neuron receives only the output of the previous layer and sends it to the next layer. There is no feedback between adjacent layers.
Convolutional neural networks (CNN)It is often used in imaging analysis. Through the convolution layer, activation function and pooling layer, the input image is approximate processed and redundant features are removed to reduce the computational complexity and help prevent overfitting. And then the significant features are combined through the fully connected layer to output prediction or classification.
Natural language processing (NLP)Algorithms are established to organize and interpret human language. The aim is to realize the interactive communication between humans and machines.
Computer vision (CV)Algorithms are established to enable the computer to perceive, observe and understand the environment through images and vision, and finally have the ability to adapt to the environment.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ding, X.; Huang, Y.; Tian, X.; Zhao, Y.; Feng, G.; Gao, Z. Diagnosis, Treatment, and Management of Otitis Media with Artificial Intelligence. Diagnostics 2023, 13, 2309. https://doi.org/10.3390/diagnostics13132309

AMA Style

Ding X, Huang Y, Tian X, Zhao Y, Feng G, Gao Z. Diagnosis, Treatment, and Management of Otitis Media with Artificial Intelligence. Diagnostics. 2023; 13(13):2309. https://doi.org/10.3390/diagnostics13132309

Chicago/Turabian Style

Ding, Xin, Yu Huang, Xu Tian, Yang Zhao, Guodong Feng, and Zhiqiang Gao. 2023. "Diagnosis, Treatment, and Management of Otitis Media with Artificial Intelligence" Diagnostics 13, no. 13: 2309. https://doi.org/10.3390/diagnostics13132309

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop