Next Article in Journal
Off-Label Benralizumab in Severe Non-Necrotizing Eosinophilic Vasculitis following Critical COVID-19 Disease and in DRESS
Next Article in Special Issue
Development of Sound Localization in Infants and Young Children with Cochlear Implants
Previous Article in Journal
Intracoronary Imaging of Vulnerable Plaque—From Clinical Research to Everyday Practice
Previous Article in Special Issue
Cochlear Implantation in Obliterated Cochlea: A Retrospective Analysis and Comparison between the IES Stiff Custom-Made Device and the Split-Array and Regular Electrodes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Web-Based Automated Image Processing Research Platform for Cochlear Implantation-Related Studies

1
Research and Development, KardioMe, 01851 Nova Dubnica, Slovakia
2
Research and Technology Group, Oticon Medical, 2765 Smørum, Denmark
3
Department for Applied Mathematics and Computer Science, Technical University of Denmark, 2800 Kongens Lyngby, Denmark
4
Department of Otolaryngology, Medical University of Hannover, 30625 Hannover, Germany
5
Epione Team, Inria, Université Côte d’Azur, 06902 Sophia Antipolis, France
6
Institut Universitaire de la Face et du Cou, Centre Hospitalier Universitaire de Nice, Université Côte d’Azur, 06100 Nice, France
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2022, 11(22), 6640; https://doi.org/10.3390/jcm11226640
Submission received: 5 October 2022 / Revised: 27 October 2022 / Accepted: 28 October 2022 / Published: 9 November 2022
(This article belongs to the Special Issue Challenges and Opportunities in Application of Cochlear Implantation)

Abstract

:
The robust delineation of the cochlea and its inner structures combined with the detection of the electrode of a cochlear implant within these structures is essential for envisaging a safer, more individualized, routine image-guided cochlear implant therapy. We present Nautilus—a web-based research platform for automated pre- and post-implantation cochlear analysis. Nautilus delineates cochlear structures from pre-operative clinical CT images by combining deep learning and Bayesian inference approaches. It enables the extraction of electrode locations from a post-operative CT image using convolutional neural networks and geometrical inference. By fusing pre- and post-operative images, Nautilus is able to provide a set of personalized pre- and post-operative metrics that can serve the exploration of clinically relevant questions in cochlear implantation therapy. In addition, Nautilus embeds a self-assessment module providing a confidence rating on the outputs of its pipeline. We present a detailed accuracy and robustness analyses of the tool on a carefully designed dataset. The results of these analyses provide legitimate grounds for envisaging the implementation of image-guided cochlear implant practices into routine clinical workflows.

1. Introduction

Cochlear Implants (CI) are, to this day, the most successful neural interfaces ever engineered judging by their functional outcomes benefits, gains in quality of life, or widespread adoption in standard clinical practice [1]. More than 700,000 CI users worldwide have been eligible for and are undergoing CI therapy because of severe or profound deafness [2]. CI systems are neuroprosthetic devices generally composed of two parts. The first part is an external device called the sound processor and is usually worn behind the ear. It is responsible for real-time sensing, processing, and transmitting acoustic information (i.e., sound) to the other, internal, surgically implanted part of the system. This second part is in charge for transmitting the encoded acoustic information content to the auditory nerve by way of trains of electrical impulses delivered through an electrode array placed in the cochlea [2]. CI systems therefore bypass the cochlea altogether and replace the natural hearing mechanism with what is often referred to as “electrical hearing”.
Despite its large overall success, CI therapy still presents significant shortcomings. In particular, documented clinical outcomes remain variable and generally not fully predictable. Additionally, perceptual adaptation to CI hearing, even when functionally successful in terms of speech recognition and communication abilities, often remain unsatisfactory when it comes to real-life scenarios, including complex, spatial, and musical soundscapes [1]. A large body of knowledge points to anatomical factors and our current limited ability to assess patient-specific cochlear anatomy (pre-implantation) and its relation to CI electrode placement (post-implantation) as impediments to the development of more adapted best practices in surgical and audiological CI therapy. The intrinsic inter-individual variability of inner ear anatomy, for instance, compounds the challenge to predict the insertion dynamics of a specific CI electrode, making it difficult to plan and predict how deep a surgeon may expect to insert the CI electrode, which may have consequences on the low-frequency percepts that the implant may be able to elicit—also known as a consequence for the preservation of residual hearing. Likewise, the challenge of assessing where exactly the electrode contacts lay within the cochlea post-operatively prevents a CI device fitting/programming that takes into account the natural tonotopicity of the spiral ganglions lining up the cochlea or the consideration of the fitting parameters set for the contra-lateral ear in bilateral CI users [3,4,5]. A common denominator to these aspects is, therefore, the need for an intimate assessment of individual anatomy and geometry of cochlear structures and CI electrode placement relative to these structures in individuals from various clinical population eligible for CI therapy. Importantly, if some of the mechanisms at play in limiting CI therapy performance outcomes (whichever ones we look at) are known, much obscurity remains as to how to harness individual anatomical information to optimize and personalize CI therapy in relevant clinical populations.
Nautilus is a web-based research-grade tool that allows the automated, accurate, robust, and uncertainty-transparent delineation of the cochlea, scala tympani (ST), scala vestibuli (SV), and of the electrode arrays with tonotopic mapping from conventional computed tomography (CT) and cone-beam computed tomography (CBCT) images (see Figure 1).

Background

The development of an automated imaging pipeline enabling the exploration of cochlear anatomy in clinical populations represents a significant challenge. The cochlear structures relevant to CI therapy, specifically the ST and SV, and the CI electrode array cannot always be easily delineated from clinical CT or CBCT images due to low image contrast and poor resolution. This prevents the manual delineation of ST and SV, which would anyway be a time-consuming, error-prone, and inconsistent process. More reasonably, semi- and fully automatic frameworks have been proposed to segment the cochlear bony labyrinth from pre-operative CT images. Earlier works focused on traditional segmentation techniques, such as level-set and interactive contour algorithms [6,7]. However, these required user input, were computationally time-consuming, and often led to incomplete segmentations. Recent works have focused on designing fully automatic convolutional neural networks capable of handling the intricate anatomy of the bony labyrinth [8,9,10,11]. The bony labyrinth is generally well identifiable in clinical CT or CBCT images, but its robust segmentation remains a challenge if one is to process images acquired with different scanners and image acquisition parameters, which may manifest in ranges of image resolution, contrast, and noise. Provided with a delineation of the bony labyrinth, various techniques permit the estimation of important metrics relevant to CI implantation, such as the cochlear duct length (CDL), which serves as an indicator of general cochlear size and what depth of insertion is reasonable to try to reach for that specific cochlea. The CDL and other metrics also enable the computation of normalized tonotopic frequencies according to Greenwood [12], Stakhovskaya [13], or Helpard et al. [14].
For all the information that can be gained from a segmentation of the bony labyrinth, many clinical questions call for the differentiation of ST from SV within the labyrinth. In this case, the automated image processing task becomes much more complex, since ST and SV are generally not visible in clinical CTs or CBCTs. Consequently, various atlases or shape models derived from temporal bone micro-CTs (µCTs) have been proposed to infer a ST/SV differentiation within the bony labyrinth when exploiting a clinical image [15,16,17,18,19,20]. The delineation of ST and SV is interesting in that CI implantation is preferentially done within ST as implantations or translocations in SV have been associated with observations of auditory pitch reversals and poorer speech intelligibility [21,22].
Post-operatively, CT imaging can provide information about the positioning of each electrode contact within or in the vicinity of the cochlea. However, the exploitation of post-operative CT/CBCT images is often compromised by metal artifacts emanating from the electrodes but generally affecting the region of interest around the electrodes enough so as to prevent the delineation of the bony labyrinth. Therefore, the post-implantation reconstruction of the CI electrode within cochlear structures often requires harnessing both the pre-operative and post-operative scans. Vanderbilt University’s group first proposed to independently segment intra-cochlear structures from pre-operative images using active shape models, followed by detection of the electrode array midline from post-operative imaging before combining pre- and post-operative information through a rigid registration [23]. They also proposed to take advantage of the left/right symmetry of inner-ear anatomy by utilizing the pre-operative image of the normal contra-lateral ear for cochlear structure delineation for cases where pre-operative CT images were not available [24]. Granting the successful reconstruction of electrode placement within cochlear structures, the characteristic frequency (CF) at each contact can legitimately be computed at the estimated corresponding place on the organ of Corti (OC) [12] or at the nearest spiral ganglion (SG) [13,14,25]. The accurate inference of the relative position between an electrode and the basilar membrane (BM) lining up the ST can also enable the assessment of the potential translocation of the electrode in SV or inferential predictions of the degree of traumaticity of the insertion, e.g., if the electrode were to have either elevated or ripped through the BM and entered the SV. Although state-of-the-art research on cochlear imaging has resulted in imaging pipelines that do display accuracy levels that can warrant their use in specific settings, these pipelines have generally not been subject to a strict robustness evaluation: their ability to deal with images of heterogeneous quality as one may expect to have to deal with when working on datasets obtained across different clinical centers. Searching to facilitate the exploration of clinical questions related to the anatomical and geometrical considerations of CI therapy, Nautilus enables the automated, accurate, robust, and transparent-on-uncertainty segmentation of the cochlear bony labyrinth, ST, and SV from pre-operative CT/CBCTs. Post-operatively, Nautilus enables the automated identification and reconstruction of the electrode arrays within the cochlear structures extracted from the pre-operative image. This tool computes a range of metrics relevant to both surgical and audiological research in CI, including the characteristic frequencies at each electrode contact. Nautilus’ predictions have been evaluated against several datasets annotated by experts and demonstrate state-of-the-art accuracy. Importantly, Nautilus was designed and stress-tested against images spanning a range of resolution, contrast, and noise, which results in its robust applicability, especially for a set of image input specifications that promote success, as we discuss later. Finally, the tool intends to transparently notify users of possible processing failures or complications using a set of caution flags to allow for the rejection of data points that may otherwise bias analysis.

2. Methods

Nautilus aims to be a gateway to advanced cochlear analysis. To maximize its availability, it has therefore been designed as a web application accessible via any modern web browser (e.g., Mozilla Firefox, Google Chrome, or Microsoft Edge) with no need for additional installation nor excessive requirements on the hardware. The data processing happens transparently on a cloud computing service. An overview of the processing pipeline can be seen in Figure 2, with Figure 3 illustrating the intermediary outputs of the process.

2.1. Data Upload and Pseudonymization via a Web-Based Frontend

Each user can create their private collection of images and associate each image to a specific case/individual. For each case, a unique anonymous identifier is generated upon creation. Once the image (most of the standard medical imaging formats are admissible (e.g., DICOM, NIFTI, MHA), as they can be loaded by ITK [26]) is loaded on the local browser, the image metadata (if any) are cleared of all personal identifiable information (PII). The user must then inform the laterality of the cochlea (left or right), whether it is a pre- or post-operative scan, and roughly place a cross on the targeted cochlea so as to allow the cropping and upload of a region of interest (ROI) from the original (albeit anonymized) image. After the data are uploaded, a processing job is queued and handled by the backend as soon as required computing resources become available.

2.2. Cochlear Landmarks and Canonical Pose Estimation

Cochlear pose estimation is essential to determine an initial orientation of the cochlea within the image and serves for image visualization in the standardized views [27]. The estimation of cochlear pose is also used for inferring the characteristic equation of the modiolar axis of the cochlea, which, in turn, is used to derive a number of metrics. We estimate the cochlear pose from a set of three automatically estimated landmarks—the center of the basal turn of the cochlea (C), the round window (RW—defined at its center), and the apex (Ap—defined at the helicotrema), as prescribed in [16]. Ap and C form the modiolar axis, which coincides with the z-axis. The basal plane passes through the RW, which defines the direction of the x-axis. The origin of the canonical reference coordinated is the intersection of the basal plane and the modiolar axis. Finally, the remaining axis is chosen such that the angle increases as we follow the cochlear duct starting from 0 deg at the RW. The canonical reference frame allows Nautilus’ users to consistently compare cochleae of different sizes and allows equal treatment for both left and right cochleae.
A number of approaches have been proposed to estimate the landmarks or the pose, including registration and one-shot learning [28] or using regression forests to vote for the location of the landmarks [29]. More recently, reinforcement learning methods [30,31,32] have also been used to efficiently locate landmarks or to generate clinically meaningful image views [33] and, relevantly for our domain of application, to locate cochlear nerve landmarks [34]. Heatmap-based approaches consistently demonstrate robustness, explainability, and computational efficiency and offer an elegant form of uncertainty modelling and failure detection [35]. They do, however, sometimes have difficulties locating landmarks present around the image borders. We employ a conventional U-Net convolutional neural network architecture [36] as implemented in [37] with three output channels, one for each landmark. We modeled each landmark with a Gaussian heatmap and trained the network to map the input image to the three target heatmaps simultaneously. Our network architecture (detailed in the Supplementary Material) has 3 encoding blocks, 8 channels after the first layer and 16 output channels for the final feature map before the final projection onto the 3 heatmap channels (see Figure S1).
Our training set consists of an assortment of 279 pre- and post-operative clinical CT and CBCT images obtained from diverse sources. Our landmark detection block must be capable of handling (and was therefore trained on) both pre- and post-operative images. It is, however, significantly more difficult to accurately annotate C, RW, and Ap on the post-operative images due to the metallic artifacts. As a workaround, the pre-operative images were registered with the post-operative images, and the landmarks from pre-operative images were transported onto the post-operative images.
For training and inference, we resampled the input images to isotropic 0.3 mm spacing and normalized the intensities between the 5–95% percentile to 0–1 with no clipping. To increase the variability of our training set, we randomly sampled from a combination of data augmentations, such as random noise, flipping in all three dimensions, Gaussian blurring, random anisotropy [38], rigid transformations, and small elastic deformations as implemented by the TorchIO library [39]. Similarly to [40], we have observed that focal loss worked particularly well for sufficiently accurate landmark detection. During the inference, we transformed the predictions with the sigmoid activation to normalize them between 0 and 1, and for each output, we pick the mode of the output distribution (the hottest voxel of the heatmap) as the corresponding landmark.

2.3. Segmentation of Cochlear Structures

Nautilus is built with cochlear surgery planning, evaluation, and audiological fitting in mind. Therefore, in the current version, we focus on segmenting the two main cochlear ducts—ST and SV—and compute relevant measurements from these structures as others before us [41]. At a later stage, the delineation of ST and SV serves to relate the placement electrode array placement within the cochlea and infer information such as the characteristic frequency of each electrode contact [23]. An accurate and robust segmentation of ST and SV is therefore critical. Recent approaches based on convolutional neural networks have shown the most promise. Nikan et al., for instance [9], segmented various temporal bone structures including the labyrinth, ossicles, and facial nerve. Most of the cochlear segmentation approaches perform remarkably well on the cochlea and neighboring structures. They do not, however, separate the scalae [8,42], nor do they estimate the position of the BM, the delicate structure responsible for the transduction of mechanical waves within the cochlea into trains of electrical impulses, an essential structure to preserve in anticipation of restorative therapeutic advances. The separation of the scalae on clinical CTs is challenging as ST and SV are not discernible on clinical scans, mainly due to limited image resolution and contrast. To circumvent this issue, a shape model is often used to serve as a priori information on ST/SV distinction within the cochlear labyrinth. Recently, atlases [43] and a hybrid active shape model combined with deep learning [44] have been used with success for the separation of the scalae.
We used a pre-operative image of the implanted cochlea as the reference image for segmentation. Nautilus uses an approach similar to [44], which merges deep learning for appearance modelling with a strong shape prior constraining the final segmentation [45]. Instead of an active shape model, we build on top of a well-validated Bayesian joint appearance and shape inference model [20,46]. The parameters of this shape model were tuned and validated on µCT data. The model can then serve as a strong prior constraining the final output for the lower-resolution clinical CT images. This approach provides a probabilistic separation of ST and SV even in images of poor resolution. We provide an estimate of the BM location from the intersection of ST and SV’s probability maps. Demarcy et al.’s original Bayesian framework proposed to model the foreground and background appearance (i.e., intensity) as mixtures of Student distributions. We observed that this initialization is fairly sensitive to the type of scanner used for image acquisition and to image quality despite using normative Hounsfield units. To achieve better generalization, we therefore replaced the original appearance model with a trained convolutional neural network [36].
Similarly to our landmark detection approach, we used a reference 3D U-Net implementation of MONAI [37] with 6 encoding blocks, 8 output channels after the first layer (see Figure S2), and PReLU as the activation function and trained it on 130 images. We normalized the data by resampling the images to 0.125 mm spacing and rescaled the intensities such that the 5th and 95th percentile of the intensity distribution of each image were mapped to 0 and 1. In addition to augmentations used for landmark detection, we used random patch swapping [47] to increase the robustness to artifacts and force the network to learn a stronger shape prior. The model was trained on 128 × 128 × 128 patches with the AdamW [48] optimizer minimizing the Dice focal loss [37,49].
A large number of the metrics we extract from both pre- and post-operative processes depend on reliable estimation of the cochlear ducts’ centerline. Because our segmentation of ST and SV is based on a parametric shape model [46], extracting an approximate centerline is straightforward. We then refine this curve and estimate ST and SV centerlines from cross-sections of the segmentations along this curve. At each cross-section, we estimate the coordinates of the lateral wall landmark as the furthest point on the ST from the modiolar axis, OC at 80% of the distance to the LW [13], and the SG offset by −0.35 mm both radially and longitudinally from the modiolar wall landmarks (i.e., the point on the ST closest to the modiolus) as an approximation of Rosenthal’s canal.

2.4. Electrode Depth-to-Angular Coverage Prediction

The centerline can be discretized based on angles (in cylindrical coordinates), which can be used to predict a priori the angular coverage an electrode array is expected to reach as a function of the number of electrodes inserted beyond the RW. Shurzig et al. [50,51] proposed an ideal trajectory for the electrode, to be computed by subtracting the radius of the electrode from the radius of the cochlear spiral. A retrospective analysis of our predictions carried out on 98 images from our clinical dataset hinted that, on average, the CI electrode only follows an ideal trajectory after hitting the lateral wall around 150 deg. This observation leads us to propose the following statistical predictive model:
δ i = ρ 1.3 0.007 θ i , if i 150 r i , otherwise
where ρ is the radius of the centerline in cylindrical coordinates, and r and θ represent the radius of the ith electrode. Figure 4 depicts the angular errors based on Equation (1). Our predictions fall, on average, within 20 of the observed insertion angular coverage (n = 58).

2.5. Registration of the Pre- and Post-Operative Images

To evaluate the electrode array placement within the cochlea, we need to be able to fuse the segmentation of the pre-operative scan and the electrode contacts of the post-operative scan to the same reference coordinate system. Although the post-operative scan is deteriorated by the metallic artifacts generated by the electrode contact, it still represents the bony structures somewhat similarly to what is seen in the pre-operative image. Rigid transformation is therefore possible for aligning pre- and post-operative images. We first pre-align the pre- and post-op image pair into their canonical poses with the previously estimated landmarks and fine-tune the final transform using the Elastix package [52,53]. We have observed that even for CT or CBCT images in Hounsfield units, the Advanced Mattes mutual information [54] with 64 histogram bins performs adequately. Invalid voxels (usually found at the boundaries of the image) and metallic artifacts in all voxels with H U > 2500 are masked out and not used for computing the similarity.

2.6. Electrode Array Detection

The electrode array detection starts with the estimation of the 3D coordinates for each of the 20 electrode contacts before the subsequent evaluation of their placement, e.g., with relative distances from relevant cochlear structures such as SG, MW, LW, BM (distances which could presumably be used to infer an indicator of traumaticity [55]). The reconstruction of the electrode array can also help with the visual inspection and assessment of complications such as kinking, tip fold-over, or buckling [56]. Most of these patterns are difficult to identify on 2D images [57], and the 3D processing approaches provide significant advantages. Various approaches can be used to locate electrode contacts. Measuring peaks of an image intensity profile is a straightforward method [58]. When these peaks are less discriminative, modelling intensity and shape with Markov random fields can help [59], and so can morphological or filtering approaches with handcrafted rules [23,60,61] or graph-based approaches [62]. Many of these approaches work well when the image resolution is fine and the contacts are well resolved, with sufficient contrast and limited metallic artifacts, no significant kinking or tip fold-over; they often can be well tuned to a particular set of scanners. With our heterogeneous dataset, the evaluated methods suffer under uneven image quality and artifacts of various appearances. We used machine learning to enhance and detect the electrode contacts of the array and to generalize over differences in appearance and image quality between the different imaging vendors. We have designed a pipeline similar to our landmark estimation similar to [63] and trained a U-Net [36,37] to estimate the likelihood of a voxel being a center of a contact. However, in addition to the contact probability estimation, our network performs two additional tasks, which share a common feature extraction backbone (see Figure S3). For training, we annotated a dataset of 106 post-operative images with ITK-SNAP [64] containing all the individual electrodes (1–20) and lead wires (where visible). From the annotations, we generated 3 different target labels: electrode location heatmap common for all electrodes (with value 1 at the centers of the electrodes and 0 away from them). By connecting electrode coordinates, we constructed a curve, which we turned into a probability map for the electrode array, and lastly, we created a discrete label map with 5 classes (background, proximal electrode, mid-electrode, distal-electrode, and lead wire) used for semantic segmentation of the post-op images.
During the inference, we first estimated the contact probabilities and considered all peaks to be contact candidates. To create an electrode array out of this unsorted set of candidates, we started with the two most central points. We then iteratively fit a cubic B-spline to the already existing set and extrapolated at the two ends to search for the next probable point until no further expansion was plausible. This gave us a sorted array of contacts. To determine the final order, we assumed that the electrode array enters the cochlear around the round window and ascends along the cochlear duct towards the apex, i.e., the signed distance to the basal plane of the first contact should be smaller than that of the last most distal contact. We have observed that this strategy performs well even in the presence of mild to moderate aforementioned electrode array insertion complications. The lead wire is then estimated from the semantic segmentation by fitting a curve to the skeleton of the closest wire-like object near the first contact. This can serve to provide a more reliable estimation of the insertion angle [65].
This electrode array detection block is designed to operate on clinical CT and CBCT images, with, for the best performance, images of resolution of 0.3mm or finer with little anisotropy. The electrode array detector has currently been tuned for and tested with the CLA and EVO electrode array from Oticon Medical (24 mm long with 20 electrodes with 1.2 mm pitch and diameter ranging from 0.5 mm proximally to 0.4 mm distally) [66]. There is, however, no significant limitation to using it for models from other vendors (see Figure 5).

2.7. Extracted Measurements

Both pre- and post-operative processing pipelines output several clinically relevant metrics, some of which are depicted in Figure 6.

2.7.1. Global Pre-Operative Metrics

Global metrics characterize the overall shape and size of the cochlea. These include the volume and surface area of the cochlea along with cochlear dimensions A and B originally proposed by Escude et al. [67], which are defined by the length of the straight line between the round window, passing through the modiolar axis, and reaching the furthest point around the 180 cochlear angle and its perpendicular line, respectively (Figure 6b). Cochlear height h is computed along the modiolar axis. These measurements can be computed for the labyrinth or specifically for ST or SV. Cochlear shape is also defined by its potential “rollercoaster”, which represents the largest deviation in height from a linear fit of the spiral height—or the vertical “dip” of the basal turn before the cochlear spirals upwards around the modiolar axis [68]. Nautilus also supports automatic computations of cochlear, basal and two-turn duct lengths of the labyrinth, ST, and SV along various trajectories within these structures: along the estimated paths of the lateral wall (LW), modiolar wall (MW), organ of Corti (OC) and spiral ganglion (SG) [68] (Figure 6d). The extraction of these metrics allows the computation of the cochlear wrapping factor, which represents the logarithmic spiral angle of the cochlea, and the wrapping ratio, which represents the ratio of the maximum cochlear angle (at the helicotrema) and the lateral wall duct length.

2.7.2. Local Pre- and Post-Operative Metrics

Local metrics characterize cochlear structures at particular places along the cochlear spiral. From pre-operative image processing, cochlear duct cross-sections are extracted at fixed angular displacements based on the labyrinth centerline. Cross-sectional area, radius, height, angle, minor and major axis lengths can then be computed by fitting an ellipse within each specific cross-section [69,70].
Post-operatively, registration parameters and the estimated locations of each electrode allow the computation of other important metrics. Electrode intracochlear positioning is characterized both by distance and angular measures at each electrode contact (where RW relates to 0 and Ap corresponds to the maximum cochlear angle, which is typically around 900 ) cochlear coverage). From these, the characteristic frequencies associated with each electrode are proposed in relation to OC [12] or SG [13,14]. In addition, the distance of each electrode contact to the MW and the estimated BM position are also computed.

2.8. Failure Flagging Mechanisms

Any automated system can occasionally fail. Transparency to the user (e.g., in the form of notifications or flags) in case of such failures is particularly important in order to identify which data point to exclude in any further observation or statistical analysis realized on Nautilus’ outputs. Therefore, Nautilus embeds a self-check flagging module that looks for signs of failures (e.g., detects suspicious segmentation or unexpected electrode array parameters) and explicitly notifies the user that images might not have been successfully processed and that the results should therefore be checked and/or used with caution. Whenever a flag is raised, a corresponding message is shown to the user (see Figure S4 for an example). Specific flags have been implemented at each processing stage. They are presented in Table 1. Figure S5 depicts the receiver operating curve (ROC) for the combined flags, based on which the cutoff values for notifying the user of a potentially faulty processing were chosen.

2.9. Data Export

The user can generate an export bundle containing all the outputs of the analysis in diverse export formats (Parquet, Excel, JSON) allowing further data analysis in their tool of choice. These analysis results are tagged with the unique version identifier for the specific processing pipeline version that was used for processing. Users may generate an export file for each case individually or a group of cases filtered on date. Figure S6 presents distributions of cochlear metrics computed by our pipeline using the export.

3. Results

3.1. Evaluation Datasets

A well-curated multi-centric dataset, comprising both clinical and cadaver bones, was chosen for tye evaluation of Nautilus. CT images acquired from various scanners, using various acquisition parameters, and presenting heterogeneous resolutions, contrasts, and signal-to-noise ratios were included both for training and evaluation (see Figures S7 and S8 in the Supplementary Material). Groundtruth annotations, comprising the C, Ap and RW landmarks, cochlear structures and the electrode center points, were delineated by an expert radiologist using ITK-SNAP [64]. Limited by the poor resolution and imaging conditions of clinical images, only the cochlea could be manually delineated for clinical scans. On the other hand, ST and SV were successfully delineated in cadaver head CT scans since better contrast and resolutions could be achieved. The number of images used for training and evaluation for each process are mentioned in their respective sections. Each part of the pipeline was independently evaluated, as detailed below. A summary of the results is presented in Table 2.

3.2. Accuracy

3.2.1. Landmark Detection

The landmark detection pipeline, utilized both pre- and post-operatively, was evaluated on a dataset of 60 images. The images were passed through the landmark detector, and the distance between the predicted and groundtruth annotation landmarks was computed. Mean detection errors of 0.71 ± 1.0 mm, 0.75 ± 1.14 mm, and 1.30 ± 1.73 mm were observed for C, Ap and RW, respectively. All the individual errors were within a distance of two voxels, with the RW landmark yielding the worst performance.

3.2.2. Segmentation

Nautilus’ segmentation pipeline was evaluated on four different clinical and cadaver datasets. The clinical dataset consisted of 58 pre-operative images with voxel resolutions ranging from 0.1 to 0.4 mm in the x-y plane and slice thickness ranging from 0.1 to 1 mm. The images were uploaded on Nautilus, and the union of ST and SV segmentation masks were obtained and compared with the manually labelled cochlea annotations. All the images were successfully processed, and a mean dice similarity coefficient and average surface error [71] of 86 ± 3 % and 0.14 ± 0.03 mm were, respectively, observed for the clinical dataset. The cadaver datasets comprised 23 temporal bone (TB) µCT images in total. For computational limitations, the CT scans were resampled to an isotropic resolution of 0.1 mm. The images were uploaded on Nautilus, and the segmentation masks were obtained and compared with the manually labelled ST and SV annotations. All the images were successfully processed, and a mean dice similarity coefficient and average surface error of 80 ± 3 % and 0.19 ± 0.04 mm were, respectively, observed for this cadaveric image dataset.
Figure 7 depicts segmentation results for each dataset. For a more thorough analysis, the cochlea was sectioned along its centerline at an 18 angular interval. Dice similarity coefficients were computed for each segment (see Figure S9), where it appears that Dice scores decrease towards the apical area.

3.2.3. Registration

The registration pipeline was evaluated on a dataset containing 15 sets of pre- and post-operative images with resolutions ranging from 0.1 to 0.3 mm. These image pairs did not necessarily have the same resolution. Each post-operative image of each pair was registered together with its pre-operative counterpart, and the average distances between the pre- and post-operative RW, Ap, and C landmarks within the registered coordinate system were computed to quantify the registration error. A mean target registration error of 0.88 ± 0.39 mm was obtained.

3.2.4. Electrode Detection

The electrode detection pipeline was evaluated on a dataset of 60 post-operative images. The electrode coordinates for each image were determined using Nautilus and compared with their corresponding groundtruth coordinates. An average electrode detection distance error of 0.09 ± 0.16 mm was achieved for successfully processed images (those that did not were rated as failures as part of our failure detection analysis—see Section 2.8).

3.3. Robustness

A retrospective robustness analysis was carried out, in which two experts from the Hannover Medical School, Hannover, Germany, and the Institut de la Face et du Cou, Nice, France, independently verified the subjective quality of both pre- and post-operative analysis outputs. A dataset of 156 ears (81 left, 75 right) was used for this study. The reviewers were presented with an assessment sheet in which they reported their subjective evaluations of the quality of the input image (both pre- and post-operative), the quality of the segmentation, and the quality of the reconstruction of the electrode array. Reviewer 1 marked 87 pre- and 59 and post-operative images as being of “good quality”. The remaining pre-operative images were either classified as having poor resolution, being very noisy or already containing an electrode array. A total of 2 out of 156 cases were marked as failures, yielding a pre-operative processing success rate of 98.7%. For the post-operative assessment, 37 cases were marked as failures, yielding a success rate of 76.2%. However, a success rate of 88.3% was realized if out-of-specification images (images that the reviewers judged as being of poor quality) were excluded from the cohort. Reviewer 2 marked 126 pre- and 60 post-operative images as being of good quality. A total of 5 out of 156 cases were marked as failures, yielding a pre-operative success rate of 98.1%. For the post-operative assessment, 33 cases were marked as failures, yielding a success rate of 78.4% or 85.2% if images judged of poor quality by the reviewer herself were excluded from the cohort.

3.4. Failure Detection

The outputs of Nautilus’ flagging system were compared with the qualitative assessment of the two reviewers, as detailed in the previous section. Figure S10 presents a performance summary of each flagging mechanism. An overall pre-operative failure detection sensitivity and specificity 100% and 97.4%, respectively, was achieved, with a corresponding post-operative failure detection sensitivity and specificity of 97.3% and 59.7%, respectively.

3.5. Computational Performances

Average computation times for each process are listed in Table 2. Computation times were obtained for a processing run on a standard Azure cloud VMs (Standard DS3 v2). On average, a complete pre- and post-operative analysis took around 10–12 min, with data storage and shape model adaptation for the segmentation taking the most time. All the other processes take less than two minutes combined. Nautilus is orchestrated with Azure Kubernetes with scalability in mind, and the throughput can be trivially scaled up by increasing the number of worker nodes.

4. Discussion

We present a web-based imaging research platform enabling the segmentation of cochlear structures and reconstruction of a cochlear implant electrode from conventional pre- and post-operative CT scans, respectively. Detailed analyses of accuracy, robustness, and failure detection provide legitimate grounds for using Nautilus for the exploration of clinically relevant questions on cochlear implantation and envisage further developments towards image-guided CI therapy.
Nautilus demonstrates segmentation performances in the range of previously presented academic results. More recent works have reported average cochlear Dice scores and average surface errors in the range of 72–91% and 0.11–0.27 mm, respectively [8,9,10,20,72]. Some of these groups have achieved higher Dice scores on limited datasets with high-resolution CT and µCT images [8,72]. A direct comparison between the works is not possible since our dataset and analysis focused on clinical and downsampled µCT images. Moreover, there is no publicly available benchmark analysis available for a fair comparison between different approaches. Nevertheless, our results on a varied dataset supports our claim of high accuracy and usability with conventional clinical CTs.
Many prior works have focused on inferring cochlea shape from µCT or high-resolution CTs as they offer good contrast and resolution compared to routine clinical CTs [8,72]. Our segmentation approach relies on JASMIN-inspired shape analysis [20], which offers the advantage of more interpretability of the estimated model parameters allowing further statistical studies. However, the same process is the bottleneck of our pipeline in terms of computational efficiency. This process could be adapted to benefit from learned shape models and anatomically inspired post-processing [73,74]. Our analysis also suggests that Nautilus performs better on clinical CT scans compared to cadaver head scans, which might be inherent to the cadaver head preparation process that often results in random air pockets, leading to a different intensity profile [75]. Additionally, our training dataset is comprised of mainly clinical scans. In future, a cadaver-specific pipeline may be developed to support cadaver-based research. Regardless, this is not a limiting factor in the applicability of Nautilus, as the main foreseen applications are in clinical research. Furthermore, our discretized analysis of the segmentation revealed that the performance decreases beyond two turns of the cochlea because of the small diameter of the cochlear ducts relative to image resolutions. This, however, is also not a limiting factor as most of the CI electrode arrays only reach around 450–600 of insertion coverage.
Post-operatively, our electrode detection process outperforms previously reported works, which have reported localization errors in the range of 0.1–0.35 mm [58,61,62]. The electrode contact-BM distances could serve for inferring insertion trauma according to the Eshraghi trauma scales [55]. This would require distance-trauma evaluation against either cadaveric histology samples or high-resolution µCT scans where the various grades of BM trauma would be resolvable. We must note that metallic artifacts emanating from the electrodes do not permit direct segmentation of cochlear structures. This warrants the necessity of a pre-operative CT-scan to infer information about the cochlear structures. The post-operative images can be converted into pseudo-pre-operative images suitable for segmentation using artifact reduction techniques [76], or an atlas can be adapted on the post-operative to segment it directly [77]. The metallic artifacts might have an impact on pre-post registration as well. However, the challenge of post- to pre-operative image conversion can be circumvented by simply using a mirrored version of the contralateral cochlea in the post-operative scan if that contralateral ear is not implanted [24].
Although accuracy is an elementary performance metric for any segmentation pipeline, robustness is key for the usefulness of a tool such as Nautilus, especially given the heterogeneity of image quality expected to be input to the tool. Our subjective quality assessment provides an indication that Nautilus can be used with confidence when dealing with images of various resolutions, contrast, and signal-to-noise ratios. To the best of our knowledge, no other work in this domain has focused on robustness analysis from a comprehensive multi-centric dataset with varying image qualities. Recently, Fan et al. achieved 85% robustness for cochlea segmentation on their 177-image dataset [44]. Contrarily, our qualitative analysis depicts a robustness of around 97% with clinically reasonable performance. Our analysis enabled us to identify a resolution cutoff beyond which robustness seems to drop. The processing of images presenting voxel sizes superior to 0.3 mm does result in a significantly greater number of failures or inadequate outputs. This assessment, therefore, sets input specifications for recommended input image resolutions.
Because the probability of failure of our pipelines is non-zero, especially if out-of-specification images are input to the tool, Nautilus does provide cautionary flagging mechanisms that embody our guiding design principle of transparency. Our current set of flags has been 100 percent sensitive and about 60 percent specific, meaning that processing failures are very unlikely to go unaccounted for and that the system will result in false positives (notified non-failures) in less than half of the time, which we deemed an acceptable threshold for usability, especially as Nautilus is robust. A further observation for failures related to electrode detection in particular is that any failures are hard failures and easily noticed by the user. All in all, our flagging mechanisms should be useful to call for manual verification and potentially discard faulty analyses.
The set of features proposed by Nautilus provides legitimate grounds for exploring many relevant clinical and basic questions related to cochlear anatomy. Nautilus’ statistical model of the electrode insertion trajectory from pre-operative images, for instance, could be used prospectively to aim at a specific insertion angular coverage. The accuracy of these predictions could be validated using Nautilus with the post-operative images. Post-operatively, Nautilus makes possible the exploration of anatomo-physiologically-tuned fitting [78,79] or the exploration of the relationship between electrode geometrical configuration within the cochlea and clinical outcomes, including perhaps residual hearing. For all its utility, Nautilus could in the future be extended with additional features to address a broader spectrum of investigations, such as these related to the prediction of insertion difficulties during surgical planning, including for abnormal anatomies [80,81]. The delineation of other structures, including the facial nerve, chorda tympani, or RW would then be required. Other imaging modalities (e.g., MRI) and electrode arrays could be the subject of future developments. Bridging pre- and post-operative use-cases, an augmented reality setup inspired by [82] could be envisaged for intraoperative guidance.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/jcm11226640/s1, Figure S1: Landmark prediction model architecture; Figure S2: Pre-operative U-Net used for cochlear segmentation; Figure S3: Post-operative U-Net for cochlear implant detection; Figure S4: An example of a failure flag being triggered and shown to caution the user about possible processing failure; Figure S5: ROC curve for failure detection process; Figure S6: Pre-operative statistics from the qualitative assessment cohort automatically computed from the segmentations; Figure S7: Qualitative segmentation performance with respect to image quality criteria; Figure S8: Qualitative registration performance with respect to image quality criteria; Figure S9: Dice scores per cochlear angle for the cadaver bone dataset (n = 23); Figure S10: Quantitative evaluation of the failure detection pipeline with respect to reviewer’s grading.

Author Contributions

Conceptualization, J.M., T.D., D.G. and F.P.; methodology, J.M., R.H., P.L.D., T.D., Z.W. and H.D.; software, J.M., R.H., P.L.D., T.D., Z.W., O.M.M. and H.D.; validation, R.H., A.M., C.V. and N.G.; formal analysis, J.M., R.H., T.D., Z.W. and D.G.; investigation, J.M., R.H., P.L.D., A.M., T.D. and Z.W.; resources, R.H., A.M., T.D., A.B., T.L., F.P. and N.G.; data curation, J.M., R.H., P.L.D., A.M., T.D., C.V., A.B. and N.G.; writing—original draft preparation, J.M., R.H., P.L.D. and F.P.; writing—review and editing, J.M., R.H., P.L.D., A.M., T.D., Z.W., D.G., O.M.M., C.V., H.D., A.B., T.L., F.P. and N.G; visualization, J.M., R.H. and P.L.D.; supervision, J.M., R.H., D.G., H.D., A.B., T.L., F.P. and N.G.; project administration, F.P.; funding acquisition, D.G. and F.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

All clinical CT images used for the development of Nautilus were anonymized. These clinical scans are part of the clinical routine at the Hannover Medical School to pre-operatively evaluate the condition of the cochlea and post-operatively confirm correct intracochlear array placement. The institutional ethics committee at Hannover Medical School approved the use of anonymized imaging data obtained within the clinical routine.

Informed Consent Statement

Informed consent was obtained from all patients, and all experiments were performed in accordance with relevant guidelines and regulations and in accordance with the Declaration of Helsinki.

Data Availability Statement

The datasets analysed within the scope of the current study cannot be made publicly available as they have been made available to the authors under the specific authorization of the Hannover Medical School. The Hannover Medical School has collected the authorization of their patients to share their data anonymously for third-party analyses in the context of clinical research. This authorization does not extend to the public publication and distribution of the data. Access to the tool is, however, available upon reasonable request at [email protected].

Acknowledgments

We would like to thank all beta-testers and early users for critical feedback on the platform. We are also grateful to the developers of the many software tools and packages used for this project, including, but not limited to, PyTorch [83], MONAI [37], TorchIO [39], ITK [26], ITK-SNAP [64], Elastix [53], VTK [84], NumPy [85], SciPy [86], scikit-learn [87], Django, Django REST framework, Celery, Kubernetes, Docker, PostgreSQL, Redis, React, react-vtkjs-viewport [88], React, Chart.js, Plotly.js, Bulma, and PyVista [89].

Conflicts of Interest

J.M. is a consultant for, and at the time of this study, R.H., T.D., O.M.M., D.G. and F.P. worked in the Research & Technology Department at Oticon Medical, manufacturer of the Neuro Zti cochlear implant system. The remaining authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ASSDAverage symmetric surface distance
BMBasilar membrane
BTLBasal turn length
CBCTCone-beam computed tomography
CDLCochlear duct length
HD95Hausdorff distance at the 95th percentile
MRIMagnetic resonance imaging
MWModiolar wall
µCTMicro computed tomography
LWLateral wall
OCOrgan of Corti
RAVDRelative absolute volume difference
ROCReceiver operating characteristic curve
RWRound window
SGSpiral ganglion
STScala tympani
SVScala vestibuli
TBTemporal bone

References

  1. Carlson, M.L. Cochlear Implantation in Adults. N. Engl. J. Med. 2020, 382, 1531–1542. [Google Scholar] [CrossRef] [PubMed]
  2. NIDCD. Cochlear Implants—Who Gets Cochlear Implants? 2021. Available online: https://www.nidcd.nih.gov/health/cochlear-implants (accessed on 22 July 2022).
  3. Kan, A.; Stoelb, C.; Litovsky, R.Y.; Goupell, M.J. Effect of Mismatched Place-of-Stimulation on Binaural Fusion and Lateralization in Bilateral Cochlear-Implant Usersa. J. Acoust. Soc. Am. 2013, 134, 2923. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Goupell, M.J.; Stakhovskaya, O.A.; Bernstein, J.G.W. Contralateral Interference Caused by Binaurally Presented Competing Speech in Adult Bilateral Cochlear-Implant Users. Ear Hear. 2018, 39, 110–123. [Google Scholar] [CrossRef] [PubMed]
  5. Peng, Z.E.; Litovsky, R.Y. Novel Approaches to Measure Spatial Release From Masking in Children with Bilateral Cochlear Implants. Ear Hear. 2022, 43, 101–114. [Google Scholar] [CrossRef] [PubMed]
  6. Yoo, K.S.; Wang, G.; Rubinstein, J.T.; Vannier, M.W. Semiautomatic Segmentation of the Cochlea Using Real-Time Volume Rendering and Regional Adaptive Snake Modeling. J. Digit. Imaging 2001, 14, 173–181. [Google Scholar] [CrossRef] [Green Version]
  7. Xianfen, D.; Siping, C.; Changhong, L.; Yuanmei, W. 3D Semi-automatic Segmentation of the Cochlea and Inner Ear. In Proceedings of the 2005 27th Annual Conference of the IEEE Engineering in Medicine and Biology, Shanghai, China, 31 August–3 September 2005; pp. 6285–6288. [Google Scholar] [CrossRef]
  8. Hussain, R.; Lalande, A.; Girum, K.B.; Guigou, C.; Bozorg Grayeli, A. Automatic Segmentation of Inner Ear on CT-scan Using Auto-Context Convolutional Neural Network. Sci. Rep. 2021, 11, 4406. [Google Scholar] [CrossRef]
  9. Nikan, S.; Van Osch, K.; Bartling, M.; Allen, D.G.; Rohani, S.A.; Connors, B.; Agrawal, S.K.; Ladak, H.M. PWD-3DNet: A Deep Learning-Based Fully-Automated Segmentation of Multiple Structures on Temporal Bone CT Scans. IEEE Trans. Image Process. 2021, 30, 739–753. [Google Scholar] [CrossRef]
  10. Lv, Y.; Ke, J.; Xu, Y.; Shen, Y.; Wang, J.; Wang, J. Automatic Segmentation of Temporal Bone Structures from Clinical Conventional CT Using a CNN Approach. Int. J. Med Robot. Comput. Assist. Surg. 2021, 17, e2229. [Google Scholar] [CrossRef]
  11. Heutink, F.; Koch, V.; Verbist, B.; van der Woude, W.J.; Mylanus, E.; Huinck, W.; Sechopoulos, I.; Caballo, M. Multi-Scale Deep Learning Framework for Cochlea Localization, Segmentation and Analysis on Clinical Ultra-High-Resolution CT Images. Comput. Methods Programs Biomed. 2020, 191, 105387. [Google Scholar] [CrossRef]
  12. Greenwood, D.P. Bandwidth Specification for Adaptive Optics Systems*. JOSA 1977, 67, 390–393. [Google Scholar] [CrossRef]
  13. Stakhovskaya, O.; Sridhar, D.; Bonham, B.H.; Leake, P.A. Frequency Map for the Human Cochlear Spiral Ganglion: Implications for Cochlear Implants. J. Assoc. Res. Otolaryngol. 2007, 8, 220. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Helpard, L.; Li, H.; Rohani, S.A.; Zhu, N.; Rask-Andersen, H.; Agrawal, S.; Ladak, H.M. An Approach for Individualized Cochlear Frequency Mapping Determined From 3D Synchrotron Radiation Phase-Contrast Imaging. IEEE Trans. Biomed. Eng. 2021, 68, 3602–3611. [Google Scholar] [CrossRef] [PubMed]
  15. Gerber, N.; Reyes, M.; Barazzetti, L.; Kjer, H.M.; Vera, S.; Stauber, M.; Mistrik, P.; Ceresa, M.; Mangado, N.; Wimmer, W.; et al. A Multiscale Imaging and Modelling Dataset of the Human Inner Ear. Sci. Data 2017, 4, 170132. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Wimmer, W.; Anschuetz, L.; Weder, S.; Wagner, F.; Delingette, H.; Caversaccio, M. Human Bony Labyrinth Dataset: Co-registered CT and Micro-CT Images, Surface Models and Anatomical Landmarks. Data Brief 2019, 27, 104782. [Google Scholar] [CrossRef] [PubMed]
  17. Sieber, D.; Erfurt, P.; John, S.; Santos, G.R.D.; Schurzig, D.; Sørensen, M.S.; Lenarz, T. The OpenEar Library of 3D Models of the Human Temporal Bone Based on Computed Tomography and Micro-Slicing. Sci. Data 2019, 6, 180297. [Google Scholar] [CrossRef] [Green Version]
  18. Noble, J.H.; Gifford, R.H.; Labadie, R.F.; Dawant, B.M. Statistical Shape Model Segmentation and Frequency Mapping of Cochlear Implant Stimulation Targets in CT. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI, Nice, France, 1–5 October 2012; Ayache, N., Delingette, H., Golland, P., Mori, K., Eds.; Springer: Berlin/Heidelberg, Germany, 2012; pp. 421–428. [Google Scholar] [CrossRef] [Green Version]
  19. Noble, J.H.; Labadie, R.F.; Majdani, O.; Dawant, B.M. Automatic Segmentation of Intra-Cochlear Anatomy in Conventional CT. IEEE Trans. BioMed Eng. 2011, 58, 2625–2632. [Google Scholar] [CrossRef] [Green Version]
  20. Wang, Z.; Demarcy, T.; Vandersteen, C.; Gnansia, D.; Raffaelli, C.; Guevara, N.; Delingette, H. Bayesian Logistic Shape Model Inference: Application to Cochlear Image Segmentation. Med. Image Anal. 2022, 75, 102268. [Google Scholar] [CrossRef]
  21. Finley, C.C.; Holden, T.A.; Holden, L.K.; Whiting, B.R.; Chole, R.A.; Neely, G.J.; Hullar, T.E.; Skinner, M.W. Role of Electrode Placement as a Contributor to Variability in Cochlear Implant Outcomes. Otol. Neurotol. 2008, 29, 920–928. [Google Scholar] [CrossRef]
  22. Macherey, O.; Carlyon, R.P. Place-Pitch Manipulations with Cochlear Implants. J. Acoust. Soc. Am. 2012, 131, 2225–2236. [Google Scholar] [CrossRef] [Green Version]
  23. Schuman, T.A.; Noble, J.H.; Wright, C.G.; Wanna, G.B.; Dawant, B.; Labadie, R.F. Anatomic Verification of a Novel Method for Precise Intrascalar Localization of Cochlear Implant Electrodes in Adult Temporal Bones Using Clinically Available Computed Tomography. Laryngoscope 2010, 120, 2277–2283. [Google Scholar] [CrossRef] [Green Version]
  24. Reda, F.A.; McRackan, T.R.; Labadie, R.F.; Dawant, B.M.; Noble, J.H. Automatic Segmentation of Intra-Cochlear Anatomy in Post-Implantation CT of Unilateral Cochlear Implant Recipients. Med. Image Anal. 2014, 18, 605–615. [Google Scholar] [CrossRef] [Green Version]
  25. Dillon, M.T.; Canfarotta, M.W.; Buss, E.; O’Connell, B.P. Comparison of Speech Recognition with an Organ of Corti versus Spiral Ganglion Frequency-to-Place Function in Place-Based Mapping of Cochlear Implant and Electric-Acoustic Stimulation Devices. Otol. Neurotol. Off. Publ. Am. Otol. Soc. Am. Neurotol. Soc. Eur. Acad. Otol. Neurotol. 2021, 42, 721–725. [Google Scholar] [CrossRef] [PubMed]
  26. Johnson, H.J.; McCormick, M.; Ibáñez, L.; Consortium, T.I.S. The ITK Software Guide, 3rd ed.; Kitware, Inc.: Clifton Park, NY, USA, 2013. [Google Scholar]
  27. Verbist, B.M.; Joemai, R.M.S.; Briaire, J.J.; Teeuwisse, W.M.; Veldkamp, W.J.H.; Frijns, J.H.M. Cochlear Coordinates in Regard to Cochlear Implantation: A Clinically Individually Applicable 3 Dimensional CT-Based Method. Otol. Neurotol. 2010, 31, 738–744. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Wang, Z.; Vandersteen, C.; Raffaelli, C.; Guevara, N.; Patou, F.; Delingette, H. One-Shot Learning for Landmarks Detection. In Deep Generative Models, and Data Augmentation, Labelling, and Imperfections; Engelhardt, S., Oksuz, I., Zhu, D., Yuan, Y., Mukhopadhyay, A., Heller, N., Huang, S.X., Nguyen, H., Sznitman, R., Xue, Y., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2021; pp. 163–172. [Google Scholar] [CrossRef]
  29. Criminisi, A.; Shotton, J.; Robertson, D.; Konukoglu, E. Regression Forests for Efficient Anatomy Detection and Localization in CT Studies. In Medical Computer Vision. Recognition Techniques and Applications in Medical Imaging; Menze, B., Langs, G., Tu, Z., Criminisi, A., Eds.; Springer: Berlin/Heidelberg, Germany, 2011; Volume 6533, pp. 106–117. [Google Scholar] [CrossRef]
  30. Ghesu, F.C.; Georgescu, B.; Mansi, T.; Neumann, D.; Hornegger, J.; Comaniciu, D. An Artificial Agent for Anatomical Landmark Detection in Medical Images. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016; Ourselin, S., Joskowicz, L., Sabuncu, M.R., Unal, G., Wells, W., Eds.; Springer International Publishing: Cham, Switzerland, 2016; Volume 9902, pp. 229–237. [Google Scholar] [CrossRef]
  31. Alansary, A.; Oktay, O.; Li, Y.; Folgoc, L.L.; Hou, B.; Vaillant, G.; Kamnitsas, K.; Vlontzos, A.; Glocker, B.; Kainz, B.; et al. Evaluating Reinforcement Learning Agents for Anatomical Landmark Detection. Med. Image Anal. 2019, 53, 156–164. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Leroy, G.; Rueckert, D.; Alansary, A. Communicative Reinforcement Learning Agents for Landmark Detection in Brain Images. In Machine Learning in Clinical Neuroimaging and Radiogenomics in Neuro-Oncology; Kia, S.M., Mohy-ud-Din, H., Abdulkadir, A., Bass, C., Habes, M., Rondina, J.M., Tax, C., Wang, H., Wolfers, T., Rathore, S., et al., Eds.; Springer International Publishing: Cham, Switzerland, 2020; pp. 177–186. [Google Scholar] [CrossRef]
  33. Alansary, A.; Folgoc, L.L.; Vaillant, G.; Oktay, O.; Li, Y.; Bai, W.; Passerat-Palmbach, J.; Guerrero, R.; Kamnitsas, K.; Hou, B.; et al. Automatic View Planning with Multi-Scale Deep Reinforcement Learning Agents. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2018; Frangi, A.F., Schnabel, J.A., Davatzikos, C., Alberola-López, C., Fichtinger, G., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 277–285. [Google Scholar] [CrossRef] [Green Version]
  34. López Diez, P.; Sundgaard, J.V.; Patou, F.; Margeta, J.; Paulsen, R.R. Facial and Cochlear Nerves Characterization Using Deep Reinforcement Learning for Landmark Detection. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 12904, pp. 519–528. [Google Scholar] [CrossRef]
  35. McCouat, J.; Voiculescu, I. Contour-Hugging Heatmaps for Landmark Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–20 June 2022; pp. 20597–20605. [Google Scholar]
  36. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar] [CrossRef] [Green Version]
  37. MONAI Consortium. MONAI Consortium. MONAI: Medical Open Network for AI (1.0.0). Zenodo. 2022. Available online: https://zenodo.org/record/7086266 (accessed on 22 September 2022). [CrossRef]
  38. Billot, B.; Robinson, E.; Dalca, A.V.; Iglesias, J.E. Partial Volume Segmentation of Brain MRI Scans of Any Resolution and Contrast. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2020; Martel, A.L., Abolmaesumi, P., Stoyanov, D., Mateus, D., Zuluaga, M.A., Zhou, S.K., Racoceanu, D., Joskowicz, L., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2020; pp. 177–187. [Google Scholar] [CrossRef]
  39. Pérez-García, F.; Sparks, R.; Ourselin, S. TorchIO: A Python Library for Efficient Loading, Preprocessing, Augmentation and Patch-Based Sampling of Medical Images in Deep Learning. Comput. Methods Programs Biomed. 2021, 208, 106236. [Google Scholar] [CrossRef]
  40. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar] [CrossRef] [Green Version]
  41. Schurzig, D.; Timm, M.E.; Majdani, O.; Lenarz, T.; Rau, T.S. The Use of Clinically Measurable Cochlear Parameters in Cochlear Implant Surgery as Indicators for Size, Shape, and Orientation of the Scala Tympani. Ear Hear. 2021, 42, 1034–1041. [Google Scholar] [CrossRef]
  42. Fauser, J.; Stenin, I.; Bauer, M.; Hsu, W.H.; Kristin, J.; Klenzner, T.; Schipper, J.; Mukhopadhyay, A. Toward an Automatic Preoperative Pipeline for Image-Guided Temporal Bone Surgery. Int. J. Comput. Assist. Radiol. Surg. 2019, 14, 967–976. [Google Scholar] [CrossRef]
  43. Powell, K.A.; Wiet, G.J.; Hittle, B.; Oswald, G.I.; Keith, J.P.; Stredney, D.; Andersen, S.A.W. Atlas-Based Segmentation of Cochlear Microstructures in Cone Beam CT. Int. J. Comput. Assist. Radiol. Surg. 2021, 16, 363–373. [Google Scholar] [CrossRef]
  44. Fan, Y.; Zhang, D.; Banalagay, R.; Wang, J.; Noble, J.H.; Dawant, B.M. Hybrid Active Shape and Deep Learning Method for the Accurate and Robust Segmentation of the Intracochlear Anatomy in Clinical Head CT and CBCT Images. J. Med. Imaging 2021, 8, 064002. [Google Scholar] [CrossRef]
  45. Margeta, J.; Demarcy, T.; Lopez Diez, P.; Hussain, R.; Vandersteen, C.; Guevarra, N.; Delingette, H.; Gnansia, D.; Kamaric Riis, S.; Patou, F. Nautilus: A Clinical Tool for the Segmentation of Intra-Cochlear Structures and Related Applications. In Proceedings of the Conference on Implantable Auditory Prostheses (CIAP), Lake Tahoe, CA, USA, 12–16 July 2021. [Google Scholar]
  46. Demarcy, T. Segmentation and Study of Anatomical Variability of the Cochlea from Medical Images. Ph.D. Thesis, Université Côte d’Azur, Nice, France, 2017. [Google Scholar]
  47. Chen, L.; Bentley, P.; Mori, K.; Misawa, K.; Fujiwara, M.; Rueckert, D. Self-Supervised Learning for Medical Image Analysis Using Image Context Restoration. Med. Image Anal. 2019, 58, 101539. [Google Scholar] [CrossRef]
  48. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. arXiv 2019, arXiv:1711.05101. [Google Scholar]
  49. Yeung, M.; Sala, E.; Schönlieb, C.B.; Rundo, L. Unified Focal Loss: Generalising Dice and Cross Entropy-Based Losses to Handle Class Imbalanced Medical Image Segmentation. Comput. Med. Imaging Graph. 2022, 95, 102026. [Google Scholar] [CrossRef] [PubMed]
  50. Schurzig, D.; Timm, M.E.; Batsoulis, C.; Salcher, R.; Sieber, D.; Jolly, C.; Lenarz, T.; Zoka-Assadi, M. A Novel Method for Clinical Cochlear Duct Length Estimation toward Patient-Specific Cochlear Implant Selection. OTO Open 2018, 2, 2473974X18800238. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  51. Mertens, G.; Van Rompaey, V.; Van de Heyning, P.; Gorris, E.; Topsakal, V. Prediction of the Cochlear Implant Electrode Insertion Depth: Clinical Applicability of Two Analytical Cochlear Models. Sci. Rep. 2020, 10, 3340. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Shamonin, D. Fast Parallel Image Registration on CPU and GPU for Diagnostic Classification of Alzheimer’s Disease. Front. Neuroinform. 2013, 7, 50. [Google Scholar] [CrossRef] [Green Version]
  53. Klein, S.; Staring, M.; Murphy, K.; Viergever, M.; Pluim, J. Elastix: A Toolbox for Intensity-Based Medical Image Registration. IEEE Trans. Med. Imaging 2010, 29, 196–205. [Google Scholar] [CrossRef]
  54. Mattes, D.; Haynor, D.; Vesselle, H.; Lewellen, T.; Eubank, W. PET-CT Image Registration in the Chest Using Free-Form Deformations. IEEE Trans. Med. Imaging 2003, 22, 120–128. [Google Scholar] [CrossRef]
  55. Eshraghi, A.A.; Van De Water, T.R. Cochlear Implantation Trauma and Noise-Induced Hearing Loss: Apoptosis and Therapeutic Strategies. Anat. Rec. Part A Discov. Mol. Cell. Evol. Biol. 2006, 288A, 473–481. [Google Scholar] [CrossRef]
  56. Ishiyama, A.; Risi, F.; Boyd, P. Potential Insertion Complications with Cochlear Implant Electrodes. Cochlear Implant. Int. 2020, 21, 206–219. [Google Scholar] [CrossRef]
  57. McClenaghan, F.; Nash, R. The Modified Stenver’s View for Cochlear Implants—What Do the Surgeons Want to Know? J. Belg. Soc. Radiol. 2020, 104, 37. [Google Scholar] [CrossRef]
  58. Bennink, E.; Peters, J.P.; Wendrich, A.W.; Vonken, E.j.; van Zanten, G.A.; Viergever, M.A. Automatic Localization of Cochlear Implant Electrode Contacts in CT. Ear Hear. 2017, 38, e376–e384. [Google Scholar] [CrossRef] [PubMed]
  59. Hachmann, H.; Krüger, B.; Rosenhahn, B.; Nogueira, W. Localization Of Cochlear Implant Electrodes From Cone Beam Computed Tomography Using Particle Belief Propagation. In Proceedings of the 2021 IEEE 18th International Symposium on Biomedical Imaging (ISBI), Nice, France, 13–16 April 2021; pp. 593–597. [Google Scholar] [CrossRef]
  60. Zhao, Y.; Dawant, B.M.; Labadie, R.F.; Noble, J.H. Automatic Localization of Closely Spaced Cochlear Implant Electrode Arrays in Clinical CTs. Med. Phys. 2018, 45, 5030–5040. [Google Scholar] [CrossRef] [PubMed]
  61. Zhao, Y.; Dawant, B.M.; Labadie, R.F.; Noble, J.H. Automatic Localization of Cochlear Implant Electrodes in CT. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2014; Golland, P., Hata, N., Barillot, C., Hornegger, J., Howe, R., Eds.; Springer International Publishing: Cham, Switzerland, 2014; Volume 8673, pp. 331–338. [Google Scholar] [CrossRef] [Green Version]
  62. Zhao, Y.; Chakravorti, S.; Labadie, R.F.; Dawant, B.M.; Noble, J.H. Automatic Graph-Based Method for Localization of Cochlear Implant Electrode Arrays in Clinical CT with Sub-Voxel Accuracy. Med. Image Anal. 2019, 52, 1–12. [Google Scholar] [CrossRef]
  63. Chi, Y.; Wang, J.; Zhao, Y.; Noble, J.H.; Dawant, B.M. A Deep-Learning-Based Method for the Localization of Cochlear Implant Electrodes in CT Images. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; IEEE: Venice, Italy, 2019; pp. 1141–1145. [Google Scholar] [CrossRef]
  64. Yushkevich, P.A.; Piven, J.; Hazlett, H.C.; Smith, R.G.; Ho, S.; Gee, J.C.; Gerig, G. User-Guided 3D Active Contour Segmentation of Anatomical Structures: Significantly Improved Efficiency and Reliability. NeuroImage 2006, 31, 1116–1128. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  65. Torres, R.; Jia, H.; Drouillard, M.; Bensimon, J.L.; Sterkers, O.; Ferrary, E.; Nguyen, Y. An Optimized Robot-Based Technique for Cochlear Implantation to Reduce Array Insertion Trauma. Otolaryngol. Head Neck Surg. 2018, 159, 019459981879223. [Google Scholar] [CrossRef]
  66. Bento, R.; Danieli, F.; Magalhães, A.; Gnansia, D.; Hoen, M. Residual Hearing Preservation with the Evo® Cochlear Implant Electrode Array: Preliminary Results. Int. Arch. Otorhinolaryngol. 2016, 20, 353–358. [Google Scholar] [CrossRef]
  67. Escudé, B.; James, C.; Deguine, O.; Cochard, N.; Eter, E.; Fraysse, B. The Size of the Cochlea and Predictions of Insertion Depth Angles for Cochlear Implant Electrodes. Audiol. Neurotol. 2006, 11, 27–33. [Google Scholar] [CrossRef]
  68. Pietsch, M.; Aguirre Dávila, L.; Erfurt, P.; Avci, E.; Lenarz, T.; Kral, A. Spiral Form of the Human Cochlea Results from Spatial Constraints. Sci. Rep. 2017, 7, 7500. [Google Scholar] [CrossRef] [Green Version]
  69. Fitzgibbon, A.; Pilu, M.; Fisher, R. Direct Least Square Fitting of Ellipses. IEEE Trans. Pattern Anal. Mach. Intell. 1999, 21, 476–480. [Google Scholar] [CrossRef] [Green Version]
  70. Burger, W.; Burge, M.J. Principles of Digital Image Processing: Core Algorithms; Springer Science & Business Media: New York, NY, USA, 2010. [Google Scholar]
  71. Maier-Hein, L.; Reinke, A.; Christodoulou, E.; Glocker, B.; Godau, P.; Isensee, F.; Kleesiek, J.; Kozubek, M.; Reyes, M.; Riegler, M.A.; et al. Metrics Reloaded: Pitfalls and Recommendations for Image Analysis Validation. arXiv 2022, arXiv:2206.01653. [Google Scholar]
  72. Ruiz Pujadas, E.; Kjer, H.M.; Piella, G.; Ceresa, M.; González Ballester, M.A. Random Walks with Shape Prior for Cochlea Segmentation in Ex Vivo µCT. Int. J. Comput. Assist. Radiol. Surg. 2016, 11, 1647–1659. [Google Scholar] [CrossRef] [PubMed]
  73. Girum, K.B.; Lalande, A.; Hussain, R.; Créhange, G. A Deep Learning Method for Real-Time Intraoperative US Image Segmentation in Prostate Brachytherapy. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 1467–1476. [Google Scholar] [CrossRef] [PubMed]
  74. Painchaud, N.; Skandarani, Y.; Judge, T.; Bernard, O.; Lalande, A.; Jodoin, P.M. Cardiac Segmentation With Strong Anatomical Guarantees. IEEE Trans. Med. Imaging 2020, 39, 3703–3713. [Google Scholar] [CrossRef] [PubMed]
  75. Soldati, E.; Pithioux, M.; Guenoun, D.; Bendahan, D.; Vicente, J. Assessment of Bone Microarchitecture in Fresh Cadaveric Human Femurs: What Could Be the Clinical Relevance of Ultra-High Field MRI. Diagnostics 2022, 12, 439. [Google Scholar] [CrossRef]
  76. Wang, Z.; Vandersteen, C.; Demarcy, T.; Gnansia, D.; Raffaelli, C.; Guevara, N.; Delingette, H. Inner-Ear Augmented Metal Artifact Reduction with Simulation-Based 3D Generative Adversarial Networks. Comput. Med. Imaging Graph. 2021, 93, 101990. [Google Scholar] [CrossRef]
  77. Wang, J.; Su, D.; Fan, Y.; Chakravorti, S.; Noble, J.H.; Dawant, B.M. Atlas-Based Segmentation of Intracochlear Anatomy in Metal Artifact Affected CT Images of the Ear with Co-trained Deep Neural Networks. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2021; de Bruijne, M., Cattin, P.C., Cotin, S., Padoy, N., Speidel, S., Zheng, Y., Essert, C., Eds.; Springer International Publishing: Cham, Switzerland, 2021; Volume 12904, pp. 14–23. [Google Scholar] [CrossRef]
  78. Mertens, G.; Van de Heyning, P.; Vanderveken, O.; Topsakal, V.; Van Rompaey, V. The Smaller the Frequency-to-Place Mismatch the Better the Hearing Outcomes in Cochlear Implant Recipients? Eur. Arch. Oto-Rhino 2022, 279, 1875–1883. [Google Scholar] [CrossRef]
  79. Canfarotta, M.W.; Dillon, M.T.; Buss, E.; Pillsbury, H.C.; Brown, K.D.; O’Connell, B.P. Frequency-to-Place Mismatch: Characterizing Variability and the Influence on Speech Perception Outcomes in Cochlear Implant Recipients. Ear Hear. 2020, 41, 1349–1361. [Google Scholar] [CrossRef]
  80. López Diez, P.; Sørensen, K.; Sundgaard, J.V.; Diab, K.; Margeta, J.; Patou, F.; Paulsen, R.R. Deep Reinforcement Learning for Detection of Inner Ear Abnormal Anatomy in Computed Tomography. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Singapore, 18–22 September 2022; Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S., Eds.; Springer Nature Switzerland: Cham, Switzerland, 2022. Lecture Notes in Computer Science. pp. 697–706. [Google Scholar] [CrossRef]
  81. López Diez, P.; Juhl, K.A.; Sundgaard, J.V.; Diab, H.; Margeta, J.; Patou, F.; Paulsen, R.R. Deep Reinforcement Learning for Detection of Abnormal Anatomies. In Proceedings of the Northern Lights Deep Learning Workshop, North Pole, Norway, 10–12 January 2022; Volume 3. [Google Scholar] [CrossRef]
  82. Hussain, R.; Lalande, A.; Guigou, C.; Bozorg-Grayeli, A. Contribution of Augmented Reality to Minimally Invasive Computer-Assisted Cranial Base Surgery. IEEE J. Biomed. Health Inform. 2020, 24, 2093–2106. [Google Scholar] [CrossRef]
  83. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Number 721. Curran Associates Inc.: Red Hook, NY, USA, 2019; pp. 8026–8037. [Google Scholar]
  84. Schroeder, W.; Martin, K.; Lorensen, B. The Visualization Toolkit—An Object-Oriented Approach to 3D Graphics, 4th ed.; Kitware, Inc.: Clifton Park, NY, USA, 2006. [Google Scholar]
  85. Harris, C.R.; Millman, K.J.; van der Walt, S.J.; Gommers, R.; Virtanen, P.; Cournapeau, D.; Wieser, E.; Taylor, J.; Berg, S.; Smith, N.J.; et al. Array Programming with NumPy. Nature 2020, 585, 357–362. [Google Scholar] [CrossRef]
  86. Virtanen, P.; Gommers, R.; Oliphant, T.E.; Haberland, M.; Reddy, T.; Cournapeau, D.; Burovski, E.; Peterson, P.; Weckesser, W.; Bright, J.; et al. SciPy 1.0: Fundamental Algorithms for Scientific Computing in Python. Nat. Methods 2020, 17, 261–272. [Google Scholar] [CrossRef] [Green Version]
  87. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-Learn: Machine Learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  88. Ziegler, E.; Urban, T.; Brown, D.; Petts, J.; Pieper, S.D.; Lewis, R.; Hafey, C.; Harris, G.J. Open Health Imaging Foundation Viewer: An Extensible Open-Source Framework for Building Web-Based Imaging Applications to Support Cancer Research. JCO Clin. Cancer Inform. 2020, 4, 336–345. [Google Scholar] [CrossRef] [PubMed]
  89. Sullivan, C.B.; Kaszynski, A. PyVista: 3D Plotting and Mesh Analysis through a Streamlined Interface for the Visualization Toolkit (VTK). J. Open Source Softw. 2019, 4, 1450. [Google Scholar] [CrossRef]
Figure 1. Nautilus offers a comprehensive set of research tools for pre- and post-operative cochlear image analysis for CI implantation and interactive visualization via a web browser. A number of metrics and additional outputs are generated by the pipeline and are made available for data export (e.g., spreadsheet of metrics for all cochleae in a user’s collection or STL models of the cochlear meshes) for further data analysis and applications (e.g., simulation or 3D printing, novel electrode array development).
Figure 1. Nautilus offers a comprehensive set of research tools for pre- and post-operative cochlear image analysis for CI implantation and interactive visualization via a web browser. A number of metrics and additional outputs are generated by the pipeline and are made available for data export (e.g., spreadsheet of metrics for all cochleae in a user’s collection or STL models of the cochlear meshes) for further data analysis and applications (e.g., simulation or 3D printing, novel electrode array development).
Jcm 11 06640 g001
Figure 2. Nautilus pipeline overview. After the images are dropped onto a web browser window, the user moves a cross-hair roughly to the cochlea’s center and selects the side (left/right) and whether it is a pre- or a post-operative scan. A crop (10 × 10 × 10 mm) centered on that landmark is then rid of personally identifiable information and uploaded for processing. First, relevant landmarks (the center, round window, and apex) are estimated and used for initial cochlear pose (reference coordinate system) computation. Segmentation of the cochlear bony labyrinth (CO) is obtained through a convolutional neural network, whereas subsequently, the scala tympani (ST) and scala vestibuli (SV) are obtained using Bayesian inference. From the post-operative image, electrode array contact coordinates and lead wire are extracted and fit to the Oticon Medical EVO electrode CAD model. An interactive visualization as well as pre- and post-operative metrics are available directly on the web browser. A number of additional outputs are generated by the pipeline and made available for data export for further processing and applications. The segmentations can be exported in STL format for 3D printing, for instance. An estimate of electrode trajectory is also provided from the pre-operative image to estimate the equivalent angular coverage for a given electrode insertion depth in millimeters.
Figure 2. Nautilus pipeline overview. After the images are dropped onto a web browser window, the user moves a cross-hair roughly to the cochlea’s center and selects the side (left/right) and whether it is a pre- or a post-operative scan. A crop (10 × 10 × 10 mm) centered on that landmark is then rid of personally identifiable information and uploaded for processing. First, relevant landmarks (the center, round window, and apex) are estimated and used for initial cochlear pose (reference coordinate system) computation. Segmentation of the cochlear bony labyrinth (CO) is obtained through a convolutional neural network, whereas subsequently, the scala tympani (ST) and scala vestibuli (SV) are obtained using Bayesian inference. From the post-operative image, electrode array contact coordinates and lead wire are extracted and fit to the Oticon Medical EVO electrode CAD model. An interactive visualization as well as pre- and post-operative metrics are available directly on the web browser. A number of additional outputs are generated by the pipeline and made available for data export for further processing and applications. The segmentations can be exported in STL format for 3D printing, for instance. An estimate of electrode trajectory is also provided from the pre-operative image to estimate the equivalent angular coverage for a given electrode insertion depth in millimeters.
Jcm 11 06640 g002
Figure 3. Steps of the image analysis pipeline in Nautilus. Regions of interest (10 × 10 × 10 mm) around a manually placed center (blue sphere) are cropped from both pre-operative (a) and post-operative (f) images. Landmark heatmaps are estimated (b,g) for the center (green), round window (blue), and apex (red). Images are aligned with rigid registration (c,h) as shown in cochlear view. Segmentation of the cochlear bony labyrinth (CO) (d) is subsequently split into the scala tympani (ST) and scala vestibuli (SV) (e). From the post-operative image, electrode array contact coordinates and lead wire are extracted (i), and an Oticon Medical EVO electrode CAD model is fit (j).
Figure 3. Steps of the image analysis pipeline in Nautilus. Regions of interest (10 × 10 × 10 mm) around a manually placed center (blue sphere) are cropped from both pre-operative (a) and post-operative (f) images. Landmark heatmaps are estimated (b,g) for the center (green), round window (blue), and apex (red). Images are aligned with rigid registration (c,h) as shown in cochlear view. Segmentation of the cochlear bony labyrinth (CO) (d) is subsequently split into the scala tympani (ST) and scala vestibuli (SV) (e). From the post-operative image, electrode array contact coordinates and lead wire are extracted (i), and an Oticon Medical EVO electrode CAD model is fit (j).
Jcm 11 06640 g003
Figure 4. Angular insertion depth estimation based on the number of electrodes inserted inside the cochlea. The comparison graph shows (left), contrarily to our approach, how the performance of state-of-the-art approaches decreases exponentially with insertion depth. On the (right) is an example of a predicted trajectory (blue) and inserted electrode.
Figure 4. Angular insertion depth estimation based on the number of electrodes inserted inside the cochlea. The comparison graph shows (left), contrarily to our approach, how the performance of state-of-the-art approaches decreases exponentially with insertion depth. On the (right) is an example of a predicted trajectory (blue) and inserted electrode.
Jcm 11 06640 g004
Figure 5. Electrode array detection in Nautilus has been developed and validated with Oticon Medical EVO electrode arrays (left) in mind. However, the same approach can be used with other electrode arrays. Example detection outputs for Cochlear Nucleus CI622 (middle), and MED-EL FLEX24 (right) cochlear implants.
Figure 5. Electrode array detection in Nautilus has been developed and validated with Oticon Medical EVO electrode arrays (left) in mind. However, the same approach can be used with other electrode arrays. Example detection outputs for Cochlear Nucleus CI622 (middle), and MED-EL FLEX24 (right) cochlear implants.
Jcm 11 06640 g005
Figure 6. Cochlea and cochlear implant in the reference coordinate frame (a) and representation of different global (b,c) or cross-sectional metrics (d,e) that can be obtained using Nautilus. Examples include A, B, and the basal turn length (BTL) along various paths within the bony labyrinth (here, BTL LW and BTL MW are the 360-degree lengths covered while following the lateral wall (LW) or the modiolar wall (MW), respectively).
Figure 6. Cochlea and cochlear implant in the reference coordinate frame (a) and representation of different global (b,c) or cross-sectional metrics (d,e) that can be obtained using Nautilus. Examples include A, B, and the basal turn length (BTL) along various paths within the bony labyrinth (here, BTL LW and BTL MW are the 360-degree lengths covered while following the lateral wall (LW) or the modiolar wall (MW), respectively).
Jcm 11 06640 g006
Figure 7. Segmentation output for different patients. (A) Clinical dataset, (B) cadaver dataset 1, (C) cadaver dataset 2, (D) cadaver dataset 3, blue: Nautilus estimation, orange: ground truth, green: overlap between the two.
Figure 7. Segmentation output for different patients. (A) Clinical dataset, (B) cadaver dataset 1, (C) cadaver dataset 2, (D) cadaver dataset 3, blue: Nautilus estimation, orange: ground truth, green: overlap between the two.
Jcm 11 06640 g007
Table 1. Description of different flags from the self-check module as implemented in Nautilus. The failure flags trigger a decreased level of confidence in the processed results when extra attention is needed.
Table 1. Description of different flags from the self-check module as implemented in Nautilus. The failure flags trigger a decreased level of confidence in the processed results when extra attention is needed.
CategoryFlags Implemented
Imagepoor image quality (resolution)
Segmentationlow cochlear volume
low segmentation reliability
irregular cochlear centerline
irregular voxel intensities within segmented region
Registrationlow correlation between pre-op and post-op
large difference between registered landmarks
too many electrode detected outside cochlea
too many electrodes detected outside scala tympani
non-basal electrodes detected outside cochlea
Electrode detectionincorrect number of electrodes detected
irregular electrode ordering
incorrect intensity at electrode locations
irregular electrode pitch
detected electrodes clustered together
incorrect distance to modiolar axis
electrodes detected near image boundaries
Table 2. Accuracy and robustness analysis for each pipeline process. ASSD: average symmetric surface distance, RAVD: relative absolute volume difference, HD95: 95% Hausdorff distance.
Table 2. Accuracy and robustness analysis for each pipeline process. ASSD: average symmetric surface distance, RAVD: relative absolute volume difference, HD95: 95% Hausdorff distance.
Landmark Detection
DatasetApex (mm)Center (mm)Round Window (mm)
Clinical (n = 60)0.710.751.30
Segmentation
DatasetDice (%)ASSD (mm)RAVDHD95 (mm)
StructureCOSTSVCOSTSVCOSTSVCOSTSV
TB set 1 (n = 9)8367640.170.210.18−0.10−0.02−0.200.430.610.43
TB set 2 (n = 9)7764580.210.230.24−0.100.23−0.380.760.770.99
TB set 3 (n = 5)7964560.190.220.20−0.21−0.04−0.400.620.710.64
Clinical (n = 58)86 0.14 −0.13 0.35
Mean8465600.150.220.20−0.140.02−0.320.410.680.63
Electrode Detection
DatasetElectrode Distance (mm)
Clinical (n = 60)0.09
Registration
DatasetMutual InformationMean Registration Error (mm)
Clinical (n = 15)0.150.88
Robustness Analysis
DatasetReviewer 1 (%)Reviewer 2 (%)
Pre-operative (n = 156)98.798.1
Post-operative (n = 156)88.3 (76.2)85.2 (78.4)
Failure Detection
DatasetSensitivity (%)Specificity (%)Accuracy (%)
Pre-operative (n = 156)10097.497.4
Post-operative (n = 156)97.357.768.6
Computational Time
ProcessApproximate Time (s)
Landmark estimation5.9
Cochlear view generation12.5
Segmentation and pre-operative analysis468.9
Electrode detection and post-operative analysis148.2
Registration49.8
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Margeta, J.; Hussain, R.; López Diez, P.; Morgenstern, A.; Demarcy, T.; Wang, Z.; Gnansia, D.; Martinez Manzanera, O.; Vandersteen, C.; Delingette, H.; et al. A Web-Based Automated Image Processing Research Platform for Cochlear Implantation-Related Studies. J. Clin. Med. 2022, 11, 6640. https://doi.org/10.3390/jcm11226640

AMA Style

Margeta J, Hussain R, López Diez P, Morgenstern A, Demarcy T, Wang Z, Gnansia D, Martinez Manzanera O, Vandersteen C, Delingette H, et al. A Web-Based Automated Image Processing Research Platform for Cochlear Implantation-Related Studies. Journal of Clinical Medicine. 2022; 11(22):6640. https://doi.org/10.3390/jcm11226640

Chicago/Turabian Style

Margeta, Jan, Raabid Hussain, Paula López Diez, Anika Morgenstern, Thomas Demarcy, Zihao Wang, Dan Gnansia, Octavio Martinez Manzanera, Clair Vandersteen, Hervé Delingette, and et al. 2022. "A Web-Based Automated Image Processing Research Platform for Cochlear Implantation-Related Studies" Journal of Clinical Medicine 11, no. 22: 6640. https://doi.org/10.3390/jcm11226640

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop