Next Article in Journal
Using a Xenogeneic Acellular Dermal Matrix Membrane to Enhance the Reparability of Bone Marrow Mesenchymal Stem Cells for Cartilage Injury
Next Article in Special Issue
Force-Controlled Biomechanical Simulation of Orthodontic Tooth Movement with Torque Archwires Using HOSEA (Hexapod for Orthodontic Simulation, Evaluation and Analysis)
Previous Article in Journal
Clinical Stability of Bespoke Snowman Plates for Fixation following Sagittal Split Ramus Osteotomy of the Mandible
Previous Article in Special Issue
A Proposal for the Classification of Temporomandibular Joint Disc Deformity in Hemifacial Microsomia
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Models for Automatic Upper Airway Segmentation and Minimum Cross-Sectional Area Localisation in Two-Dimensional Images

1
Orthodontics, Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China
2
Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Hong Kong SAR, China
3
Division of Oral and Maxillofacial Surgery, Faculty of Dentistry, The University of Hong Kong, Hong Kong SAR, China
4
Department of Orthodontics, The State Key Laboratory Breeding Base of Basic Science of Stomatology (Hubei-MOST), Key Laboratory of Oral Biomedicine Ministry of Education, School & Hospital of Stomatology, Wuhan University, Wuhan 430072, China
*
Author to whom correspondence should be addressed.
Bioengineering 2023, 10(8), 915; https://doi.org/10.3390/bioengineering10080915
Submission received: 5 June 2023 / Revised: 19 July 2023 / Accepted: 31 July 2023 / Published: 2 August 2023
(This article belongs to the Special Issue Application of Bioengineering to Clinical Orthodontics)

Abstract

:
Objective: To develop and validate convolutional neural network algorithms for automatic upper airway segmentation and minimum cross-sectional area (CSAmin) localisation in two-dimensional (2D) radiographic airway images. Materials and Methods: Two hundred and one 2D airway images acquired using cone-beam computed tomography (CBCT) scanning were randomly assigned to a test group (n = 161) to train artificial intelligence (AI) models and a validation group (n = 40) to evaluate the accuracy of AI processing. Four AI models, UNet18, UNet36, DeepLab50 and DeepLab101, were trained to automatically segment the upper airway 2D images in the test group. Precision, recall, Intersection over Union, the dice similarity coefficient and size difference were used to evaluate the performance of the AI-driven segmentation models. The CSAmin height in each image was manually determined using three-dimensional CBCT data. The nonlinear mathematical morphology technique was used to calculate the CSAmin level. Height errors were assessed to evaluate the CSAmin localisation accuracy in the validation group. The time consumed for airway segmentation and CSAmin localisation was compared between manual and AI processing methods. Results: The precision of all four segmentation models exceeded 90.0%. No significant differences were found in the accuracy of any AI models. The consistency of CSAmin localisation in specific segments between manual and AI processing was 0.944. AI processing was much more efficient than manual processing in terms of airway segmentation and CSAmin localisation. Conclusions: We successfully developed and validated a fully automatic AI-driven system for upper airway segmentation and CSAmin localisation using 2D radiographic airway images.

1. Introduction

Upper airway obstruction is a major cause of sleep-disordered breathing and obstructive sleep apnoea (OSA); this has increased the importance of airway assessment [1,2]. OSA caused by upper airway obstruction is common in the modern global population and has been associated with increased incidences of hypertension, atrial fibrillation, coronary heart disease and stroke [3,4]. OSA reportedly affects 17% of women and 34% of men in the US and has shown a similar prevalence in other countries; however, approximately 85% of patients remain undiagnosed [3,5].
Because the upper airway is anatomically located just behind the oral cavity, airway diseases are often first detected upon radiographic examination by dentists, especially orthodontists. However, the superimposition of the upper airway upon adjacent structures in 2D images, combined with the lack of quantitative measurement standards, makes the preliminary diagnosis of upper airway obstruction using lateral cephalograms (LCs) quite challenging and sometimes inaccurate, even though LC is usually the ‘first-hand’ record for airway evaluation [6,7].
Three-dimensional (3D) computed tomography (CT) and cone-beam CT (CBCT) are more promising techniques than LCs for upper airway assessment. However, only patients with severe craniomaxillofacial deformities are required to undergo large-scale CT scans that cover the upper airway area, owing to the relatively higher radiation dose than 2D radiography. The risks of exposure to high radiation and high costs render 3D screening tools impractical for detecting potential airway problems [8,9]. Therefore, albeit challenging, obtaining sufficient and accurate upper airway information through 2D images, i.e., ‘first-hand’ image records, is important for the early detection of potential airway problems.
Currently, both 2D and 3D airway image analysis are primarily conducted manually, with the processing performed by professional dental/radiographic experts [10,11,12]. These methods are time-consuming, tedious and expert-dependent, which negatively affects the efficiency and accuracy and can lead to missed diagnoses of the disease; all of the above can have significant impacts on human health [13,14]. Recently, artificial intelligence (AI) and deep learning techniques, which employ computers or machines to imitate human logic and cognition to complete a series of intelligent tasks [15], have been widely applied in medical imaging. AI processing is highly efficient because it can extract dental features and swiftly make decisions based on big data; studies have reported its superiority and efficacy in imaging detection compared with manual processing. In dentistry, AI algorithms based on convolutional neural networks (CNNs) have shown remarkable capabilities in numbering and classifying teeth, detecting root fractures and periodontal bone loss and diagnosing dental caries [16,17,18]. However, until now, the effects of AI techniques on upper airway assessment have not been thoroughly explored. Therefore, we aimed to explore the possibilities of applying AI techniques to automatic upper airway assessment to achieve more accurate and efficient evaluations, thereby providing valuable information for the preliminary diagnosis of upper airway obstruction.
The basic step of upper airway assessment is to detect the airway contour and divide it into three segments: the nasopharynx, retropalatal pharynx and retroglossal pharynx [13,19]. The morphology and features of the different segments serve as anatomical data that indicate airway obstruction, if present [9,14]. Automatic and accurate segmentation of the upper airway is fundamental for the application of AI techniques to airway assessment. Furthermore, the minimum cross-sectional area (CSAmin), which denotes the region with the greatest constriction, is a crucial parameter for airway assessment [20,21,22,23]. But CSAmin localisation requires clinicians to reconstruct the upper airway structure in three dimensions and compare the area values across different planes. In recent AI studies, CNNs were used to successfully infer or reconstruct 3D structures based on 2D data after realising large amounts of 3D features and data [24,25]; this would have been impossible to achieve using manual processing (owing to a lack of data from the third dimension).
Therefore, we hypothesised that the development of AI models based on deep learning algorithms can facilitate automatic upper airway segmentation and CSAmin localisation in 2D radiographic airway images, and AI processing can improve the efficiency greatly.

2. Materials and Methods

This study was approved by the Institutional Review Board of the University of Hong Kong (IRB reference no: UW 21-519). The methods were conducted in accordance with approved guidelines and regulations. All CBCT scans were obtained for diagnosis and treatment planning. Written informed consent was obtained from patients, and patients were informed that their clinical images may be used for clinical teaching and paper publishing.

2.1. Data Collection

CBCT scans from 201 orthodontic patients (76 men and 125 women) aged over 18 years (43.45 ± 19.28 years) were collected from a consecutive sample of the Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, the University of Hong Kong, between 1 September, 2016 and 31 December, 2021. The CBCT scan field view covered the whole skull and upper airway contour, with the superior border above the calvarium and inferior border below the epiglottic base. The regions of interest (ROIs) of training data in this study are three segmented regions (the nasopharynx, retropalatal pharynx and retroglossal pharynx) of upper airway in the annotated segmentation map. The exclusion criteria were as follows: (1) patients who presented severe skeletal deformities, facial asymmetry, cleft lip and palate and craniofacial syndrome, which might affect the upper airway morphology and (2) patients with defective CBCT scans (low resolution, improper head posture or unclear airway structures).
The sample ratio of the training group to the test group was set at 4:1 [16,17,18]; the 201 images were randomly assigned to the training group (n = 161) to train AI models or the test group (n = 40) to evaluate the efficacy of AI processing. Standard 4-fold cross-validation was conducted on the training set for basic hyperparameter search, and the test data were kept blind to the model until the final evaluation. The data collection and AI processing flow are presented in Figure 1.

2.2. Consistency Tests

A solo examiner (CHU G), who is a well-trained dentist with sufficient experience in dental imaging detection, conducted all the manual image annotation work including midsagittal plane capture, upper airway segmentation and CSAmin level localisation. To ensure data validity and reliability, two experienced examiners (CHU G and Ng CH) conducted consistency tests. Intra- and interexaminer reliability were assessed for midsagittal plane capture, image segmentation and CSAmin level detection.
To assess intraexaminer reliability, 10 randomly selected images from all sample data of 201 images were analysed by the examiner (CHU G); after 1 month, the assessment was repeated to test reliability and repeatability. Precision (the ratio of correctly predicted pixels between the two measurements) was used as the main metric to evaluate consistency.
To assess interexaminer reliability, 10 randomly selected images from all sample data of 201 images were analysed by two examiners (CHU G and Ng CH) to test the repeatability of the midsagittal plane capture and CSAmin level detection, respectively.

2.3. Determination of the Midsagittal Plane and Image Capture

All of the CBCT scans were obtained using the CBCT device ProMax 3D Mid (Planmeca Oy, Helsinki, Finland). The clinical staff of the Oral and Maxillofacial Radiology Unit of the Faculty of Dentistry, the University of Hong Kong, performed CBCT scanning using a reference ear plug and head posture aligner to maintain a natural head position. Raw CBCT data were saved and extracted using proprietary software (ROMEXIS, v4.4.0.R; Planmeca, Helsinki, Finland).
Then, CBCT scans were converted to a DICOM image format and evaluated using Mimics v19.0 (Materialise, Leuven, Belgium). All the scans were then reconstructed with slices of 0.5 mm thickness and 0.4 mm voxel size. Five reference points (posterior nasal spine, posterior point of vomer and first, second and fourth cervical vertebrae) were manually obtained for each scan to determine the midsagittal planes of the CBCT images (Figure 2). Each of these reference points is defined in Table 1. The determined midsagittal plane images were captured and saved in PNG format (32 bits, 910 × 910 pixels).

2.4. Upper Airway Segmentation

2.4.1. Manual Segmentation of the Airway and Data Augmentation

A well-trained dentist (CHU G, solo examiner) segmented the upper airway from the 2D images using the image processing tool Microsoft Paint in Windows 10 system. The airway structure in each image was segmented into three parts [13,19]: the nasopharynx, extending from the nasal turbinate level to the hard palate; the retropalatal pharynx, extending from the hard palate to the margin of the soft palate; and the retroglossal pharynx, extending from the soft palate to the epiglottic base. Each segment of the airway is shown using different colours in Figure 3.
All of the images were resized to obtain a unified resolution of 400 × 400 pixels. Manually annotated segmentation maps were converted into label maps, in which all pixel values were integers between 0 and 3; 0 represented the background region, whereas 1–3 represented the three segmented regions (the nasopharynx, retropalatal pharynx and retroglossal pharynx) in the annotated segmentation map. During the training process, data augmentation techniques were used to increase variability in the training dataset by resizing, cropping, flipping and randomly rotating the images.

2.4.2. AI Segmentation Models

The segmentation architectures UNet and DeepLab v3 were selected as the training models. The UNet and DeepLab network architectures were each implemented at two depths: UNet18, UNet36, DeepLab50 and DeepLab101. These four AI models were then trained for the segmentation tasks. All the models were trained using the Adam optimisation algorithm at 200 epochs. The architecture was trained in the Pytorch framework (learning rate, 0.0001; batch size, 8) to achieve the best possible validation loss. All the processing work was conducted on a PC with an Intel i7-8700 CPU, 32GB RAM and a single Nvidia RTX 2080 Ti GPU with 12G VRAM (Jumbo computer supplies, Hong Kong, China). Figure 4 presents the network architecture of the UNet model.

2.4.3. Evaluation Metrics for Airway Segmentation

The following metrics were used to evaluate the performance of the segmentation models: precision, recall, Intersection over Union (IoU), dice similarity coefficient (DSC) and size difference.
Precision indicated how many of all the predicted pixel results had been correctly predicted.
P r e c i s i o n = T P T P + F P
Recall measured how many of all the correct pixel results had been correctly predicted.
R e c a l l = T P T P + F N
IoU has been commonly used to evaluate image segmentation tasks; it represents the similarity of segmentation results between the predicted image and ground truth.
I o U ( A , B ) = A B A B = T P T P + F N + F P
DSC was calculated as twice the overlapping area divided by the sum of the areas of A and B.
D S C = 2 ( A B ) A + B
Size difference measured the gap between the predicted image and ground truth in pixels, i.e., the number of ground truth segmentation pixels subtracted from the number of pixels of the predicted segmentation results.

2.5. CSAmin Localisation Task

2.5.1. Manual Determination of CSAmin

To determine the exact CSAmin level in the entire upper airway passage, each 3D upper airway model was reconstructed using Mimics v19.0. The inferred CSAmin could be manually detected through the midsagittal view in reconstructed 3D images. The height level was marked in each image to illustrate the specific localisation of the inferred CSAmin and was recorded as HM (manually determined height of CSAmin). The work of manual determination of CSAmin was conducted by the solo examiner (CHU G) as stated previously.

2.5.2. AI-Driven Determination of CSAmin

Forty images each, after manual and AI-driven segmentation, were acquired as the two test datasets. The nonlinear mathematical morphological operations comprising erosion and dilation were employed as the postprocessing procedure for the segmentation images. CSAmin was automatically determined after applying erosion and dilation to extract the binary image boundaries.

Erosion and Dilation

Erosion is generally used to extract the inner boundary, which refers to the set of pixels comprising boundary points and belonging to a part of the original regions. The inner boundary was extracted using erosion processing to obtain a contraction of the original image. An exclusive OR operation was then conducted between the contraction results and target image, thereby realising the difference set extraction. After erosion processing, the undesired noise was efficiently removed, and the main area was smoothed in the predicted segmentation map.
Dilation is generally used to extract the outer boundary, which refers to the set of pixels adjacent to the boundary points outside the region and is a part of the background. The outer boundary was extracted using dilation processing. An exclusive OR operation was then performed between the dilation result and original target image to obtain the difference set.

Computation and Prediction of CSAmin

After erosion and dilation processing, 40 manual segmentation maps (GT1) and 40 AI-driven segmentation maps (GT2) were considered as the two types of ground truth. The sum of the number of the pixels in each row in GT1 and GT2 was calculated to obtain the width of the upper airway at different levels. The narrowest slice with the minimum number of pixels was selected and subjected to the original coordinate system. The narrowest heights in the GT1 and GT2 groups were recorded as H 1 and H 2 , respectively. The height difference (L) between the predicted ground truth and manually determined heights was calculated to compare the performance of CSAmin prediction as follows:
L 1 = | H 1 H M | L 2 = | H 2 H M |

2.6. Time Comparison

To test the efficiency of the AI models, the time consumed for manual work (CHU G) and AI processing was recorded when performing the segmentation and CSAmin localisation tasks in the test set (40 images). The average time spent on each image was compared between the two processing methods.

2.7. Statistical Analysis

Precision (the ratio of correctly predicted pixels between two measures) was assessed to evaluate the consistency in manual segmentation. The intra- and interclass correlation coefficients (ICCs) were calculated to evaluate the intra- and inter-rater agreements of the sagittal view coordinates and HM, respectively.
To evaluate the performance of the AI-driven segmentation, the chi-square test was then used to compare precision, recall, IoU and DSC across the different segments. And the Kruskal–Wallis test was used to compare the size differences.
To evaluate the CSAmin height prediction performance, the kappa test was used to evaluate the consistency of CSAmin determinations across specific segments and compare it between manual and AI processing. The paired t-test and Bland–Altman analysis were used to compare HM, H1 and HM, H2. The two-sample t-test was used to compare the height differences L1 and L2. p < 0.05 indicated statistical significance. All the data were analysed using Statistical Product and Service Solutions (SPSS) v27.0.

3. Results

The average precision reliability values of manual segmentation for the 10 paired images were 0.971, 0.967 and 0.969 in the nasopharynx, retropalatal pharynx and retroglossal pharynx, respectively. The ICCs of the midsagittal plane coordinate were 0.973 and 0.936 for the intra- and inter-rater agreements, respectively. The ICC values of HM determination for the intra- and inter-rater agreements were 0.989 and 0.954, respectively. All the consistency tests indicated a high degree of intra- and interobserver agreement.

3.1. Accuracy Analysis for AI-Driven Upper Airway Segmentation

The accuracy results of the different AI models are presented in Table 2 and Table 3. Overall, the precision and recall reached 90.0–90.6% and 88.0–89.2%, respectively. In general, no significant differences in accuracy were noted for any of the four AI models (p = 0.476, 0.562, 0.433, 0.552 and 0.283 for precision, recall, IoU, DSC and size difference, respectively). In terms of the segmentation results, a difference was detected in the size difference only when using the DeepLab50 model. As the size difference results of the UNet36 model were not normally distributed after testing using the Shapiro–Wilk test (p = 0.037), the Kruskal–Wallis test was used to compare the size differences. The size difference in the nasopharynx was significantly larger than those in the retropalatal pharynx and retroglossal pharynx (p = 0.039). Figure 5 shows the airway segmentation results obtained from manual processing and the UNet18 model.

3.2. Accuracy Analysis for AI-Driven CSAmin Localisation

The accuracy of CSAmin localisation in specific segments for the manual (GT1) and full AI (GT2) processing methods is presented in Table 4. The kappa test revealed that the consistency between the two results was 0.944. Figure 6 shows the CSAmin localisation results for the GT1 and GT2 groups.
The detailed results of AI performance in the two test sets are presented Table 5; no significant differences were noted between H1 and HM (p = 0.780) or between H2 and HM (p = 0.295). The height difference between the HM of manual work and H1 of the first test set was 1.95 ± 2.21 mm, whereas that between the HM of manual work and H2 of the second test set was 2.53 ± 2.21 mm. Furthermore, Bland–Altman analysis was conducted to compare the difference between HM, H1 and HM, H2, and showed that 95.00% (38/40) and 90.00% (36/40) of the points were within the 95% conformance level, respectively (Figure 7 and Figure 8). The height error L1 across the 40 subjects ranged from 0 to 12.31 mm; the proportion of errors <3 mm was 82.5% (33 out of 40). The height error L2 ranged from 0 to 9.89 mm; the proportion of errors <3 mm was 70.0% (28 out of 40). The two-sample t-test showed no significant differences between L1 and L2 (p = 0.786).

3.3. Time Comparison

The time consumed was recorded for both manual and AI processing methods. In each image, the average time for manual upper airway segmentation was 6 min 28 s. The time to average manual CSAmin determination was 21 min 10 s. With respect to AI processing, fully automatic algorithms of all four models can complete the segmentation and localisation tasks in the test group (all 40 images) within 1 s.

4. Discussion

With this study, we successfully explored the possibility that the AI models we developed were able to predict 3D information (CSAmin) from 2D data to solve clinical challenges. Regarding AI applications, airway segmentation and CSAmin localisation tasks are basic and initial steps in the development of AI systems for pathology detection and disease diagnosis in airway assessment. From the clinical perspective, this study provides an innovative upper airway analysis method that will help improve clinicians’ efficiency.
Nowadays, a few studies have explored and discussed the possibilities of applying AI techniques in upper airway analysis [26,27,28,29,30]. For example, Çağla et al. [31] developed an automatic pharyngeal airway segmentation in CBCT images by using a CNN system; a dice ratio of 0.919 and a weighted IoU of 0.993 were achieved. Antonio et al. [32] developed a fully automatic and end-to-end airway segmentation method based on the U-Net architecture, and the results showed the EXACT’09 public dataset achieved the highest sensitivity among three different test sets. And Rosalia et al. [33] applied a deep learning-based fully CNN technique to segment the sinonasal cavity and pharyngeal airway. A mean volume difference of 1.93 ± 0.73 cm3 was found between manual segmentation and AI processing.
However, most studies used 3D CT or CBCT images as training data to develop algorithms and AI models. In dentistry, only patients with severe craniomaxillofacial deformities are required to undergo CT or CBCT scans owing to the relatively higher cost and radiation dose than 2D radiography. Compared with 3D techniques, 2D images are more commonly applied in daily dental practice, which makes them an important tool for the early detection of potential airway problems. There have never been similar studies reported to infer CSAmin information from CBCT scans based on 2D data. Our study is the first to establish a fully automatic AI-driven system to evaluate the upper airway using 2D images. Compared with the detection of craniofacial hard tissues such as the skull, teeth and alveolar bone, airway structure detection is quite complicated owing to the complexity of the structures, uncertainty of landmarks and slight differences in the greyscale between air and soft tissue [34,35,36]. Given the sharing of convolution kernel parameters and sparsity of connections between layers, the CNN deep learning algorithms developed in this study allowed computer models comprising multiple convolutional layers to efficiently extract the boundaries and features of upper airway regions.
It is also extremely time-consuming to manually determine CSAmin, a crucial parameter in airway assessment, in three dimensions; the judgement cannot be based solely on 2D images because data from the third dimension are lacking. However, CNNs can predict 3D information from 2D data after realising a large amount of 3D feature data [24,25]. We utilised this strength of CNN deep learning algorithms and successfully developed a fully automatic AI-driven system for upper airway segmentation and CSAmin localisation using 2D radiographic airway images.

4.1. AI-Driven Segmentation Accuracy

The developed AI system was accurate and rapid in terms of upper airway segmentation. Regarding segmentation accuracy, all the four AI models achieved a precision of >90.0% and recall of >88.0% in the entire upper airway. Because of the lack of comparable studies on airway segmentation, we compared our accuracy with that reported in other studies on tooth detection involving deep learning methods and CNNs. Chen et al. [37] developed a fast region CNN method for detecting and numbering teeth in dental periapical films and reported a precision and recall of approximately 90%. Leite et al. [16] evaluated the performance of a new AI-driven system for tooth detection and segmentation and reported a sensitivity and precision of up to 90%. In addition to the high consistency with the above-mentioned studies, our segmentation models were ‘fully automatic’, whereas their AI segmentation systems required manual modifications after AI processing.
Owing to the irregular morphology and high variability in the nasopharynx region, the accuracy of segmentation in the nasopharynx was relatively lower than that in the retropalatal pharynx and retroglossal pharynx. Because the retropalatal and retroglossal segments are the main obstruction sites in the upper airway [38], the accuracy of segmentation in these two segments is critical for upper airway assessment. Our results showed that the precision of segmentation in the retropalatal pharynx and retroglossal pharynx was 92.0%, providing reasonably accurate anatomical information for clinical reference.
No differences were noted in the accuracies of the four AI segmentation models. The AI models with more convolutional layers (DeepLab101 and UNet36) did not show significantly better performance than the DeepLab50 and UNet18 models, perhaps because 2D data are relatively ‘simple’, with little heterogeneity and diversity. Therefore, the limited convolutional layers were sufficient for segmentation tasks and achieving good performance. For more comprehensive tasks, algorithms with a higher number of convolutional layers may show obvious superiority.

4.2. AI-Driven CSAmin Localisation Accuracy

Automatic CSAmin localisation was achieved using the established segmentation systems. In spite of the significance of CSAmin, the automatic localisation of CSAmin from 2D images is challenging because of a lack of 3D data, which renders manual determination impossible. In this study, we developed a new AI-driven algorithm to automatically predict the CSAmin height using 2D airway images, thereby facilitating the evaluation of CSAmin in LCs. The kappa test revealed a high consistency between manual processing using 3D images and AI prediction from 2D images (0.944). Our study therefore successfully validated the AI algorithms developed to accurately determine the CSAmin localisation based on 2D images.
Based on the results of our test set of 40 images, in 27 (67.5%) and 13 (32.5%) images, CSAmin localised to the retropalatal region and retroglossal region, respectively, consistent with previous retrospective clinical studies that demonstrated that the retropalatal region was the most common site for airway collapse in patients with OSA [39,40].
The calculated height errors of CSAmin localisation obtained from the 2D images were merely 1.95 and 2.53 mm for the manual and AI processing methods, respectively. The height error differences between the two segmentation ground truths demonstrated that the airway structure boundaries were slightly overestimated or underestimated during the AI processing. However, a height error of 2 mm was equivalent to 4–6 slices out of over 500 in a CBCT scan (0.8–1.2% of error), which did not affect the accuracy of CSAmin determination by the AI models in specific segments (kappa value: 0.944).

4.3. Limitations and Prospects

The purpose of using 3D CBCT scans in this study was to obtain exact CSAmin data. Considering ethical considerations in clinical practice that CBCT scans and LCs cannot be ethically obtained simultaneously from the same patient, we used 2D upper airway midsagittal images obtained from CBCT scans as the training dataset to establish CNNs. Previous studies have reported that both LCs and CBCT scans are reliable for the evaluation of upper airway structures and that there are no significant differences in the dimension measurements between the two techniques [23,41]. Although 2D midsagittal plane images are not an ideal source of training data, our study successfully established AI algorithms and systems with reasonably high accuracy. Further studies should consider the application of LCs to AI upper airway assessment to verify the practicality of the method.
In future, a larger number of subjects with different ethnicities, genders and ages should be recruited to develop more robust algorithms. In terms of future clinical applications, various learning material datasets, including those with clinical information, should be used to develop customised airway assessment systems for individual subjects.

5. Conclusions

In conclusion, we successfully developed an accurate and efficient fully automatic AI-driven system for upper airway segmentation and CSAmin localisation based on 2D radiographic upper airway images. The AI models we developed are able to predict 3D CSAmin information from 2D images for screening out potential airway problems, which provides an innovative upper airway analysis method that can improve clinicians’ efficiency greatly.

Author Contributions

G.C.: designed the experiment and wrote the main manuscript; R.Z. and Y.H.: designed and supervised the convolutional neural network system; C.H.N.: completed the consistency tests; M.G., Y.Y.L. and H.H.: provided important instructions and constructive suggestions for the experiment; Y.Y.: provided instruction, supervision and funding. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Seed fund for basic research of the University of Hong Kong university research committee [grant no. 202111159228].

Institutional Review Board Statement

This study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Institutional Review Board of the University of Hong Kong (IRB reference no: UW 21-519).

Informed Consent Statement

We added and provided soft copy of written informed consent.

Data Availability Statement

All data of the study are available from the Division of Paediatric Dentistry and Orthodontics, Faculty of Dentistry, the University of Hong Kong, and were used under license and approved by the Institutional Review Board of the University of Hong Kong/Hospital Authority Hong Kong West Cluster (IRB reference no: UW 21-519). The data are, however, available from the corresponding author upon reasonable request and with permission of the University of Hong Kong. Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Alsufyani, N.A.; Al-Saleh, M.A.; Major, P.W. CBCT assessment of upper airway changes and treatment outcomes of obstructive sleep apnoea: A systematic review. Sleep. Breath. 2013, 17, 911–923. [Google Scholar] [CrossRef]
  2. Dultra, F.; Tavares, A.; Dultra, J.A.; Salles, C.; Crusoe-Rebelo, I.M.; Barbosa, I.; Souza-Machado, A. Pharyngeal airspace of asthmatic individuals and those suffering from obstructive sleep apnea syndrome: Study by CBCT. Eur. J. Radiol. 2017, 95, 342–348. [Google Scholar] [CrossRef]
  3. Gottlieb, D.J.; Punjabi, N.M. Diagnosis and Management of Obstructive Sleep Apnea: A Review. JAMA 2020, 323, 1389–1400. [Google Scholar] [CrossRef]
  4. Javaheri, S.; Barbe, F.; Campos-Rodriguez, F.; Dempsey, J.A.; Khayat, R.; Javaheri, S.; Malhotra, A.; Martinez-Garcia, M.A.; Mehra, R.; Pack, A.I.; et al. Sleep Apnea: Types, Mechanisms, and Clinical Cardiovascular Consequences. J. Am. Coll. Cardiol. 2017, 69, 841–858. [Google Scholar] [CrossRef]
  5. Lugaresi, E.; Cirignotta, F.; Coccagna, G.; Piana, C. Some epidemiological data on snoring and cardiocirculatory disturbances. Sleep 1980, 3, 221–224. [Google Scholar] [CrossRef]
  6. Ahlqvist, J.; Eliasson, S.; Welander, U. The effect of projection errors on cephalometric length measurements. Eur. J. Orthod. 1986, 8, 141–148. [Google Scholar] [CrossRef]
  7. Vig, P.S.; Hall, D.J. The inadequacy of cephalometric radiographs for airway assessment. Am. J. Orthod. 1980, 77, 230–233. [Google Scholar] [CrossRef]
  8. Zondervan, R.L.; Hahn, P.F.; Sadow, C.A.; Liu, B.; Lee, S.I. Body CT scanning in young adults: Examination indications, patient outcomes, and risk of radiation-induced cancer. Radiology 2013, 267, 460–469. [Google Scholar] [CrossRef]
  9. Eslami, E.; Katz, E.S.; Baghdady, M.; Abramovitch, K.; Masoud, M.I. Are three-dimensional airway evaluations obtained through computed and cone-beam computed tomography scans predictable from lateral cephalograms? A systematic review of evidence. Angle Orthod. 2017, 87, 159–167. [Google Scholar] [CrossRef] [Green Version]
  10. Chen, H.; van Eijnatten, M.; Wolff, J.; de Lange, J.; van der Stelt, P.F.; Lobbezoo, F.; Aarab, G. Reliability and accuracy of three imaging software packages used for 3D analysis of the upper airway on cone beam computed tomography images. Dentomaxillofac Radiol. 2017, 46, 20170043. [Google Scholar] [CrossRef]
  11. Torres, H.M.; Evangelista, K.; Torres, E.M.; Estrela, C.; Leite, A.F.; Valladares-Neto, J.; Silva, M.A.G. Reliability and validity of two software systems used to measure the pharyngeal airway space in three-dimensional analysis. Int. J. Oral. Maxillofac. Surg. 2020, 49, 602–613. [Google Scholar] [CrossRef] [PubMed]
  12. Burkhard, J.P.; Dietrich, A.D.; Jacobsen, C.; Roos, M.; Lubbers, H.T.; Obwegeser, J.A. Cephalometric and three-dimensional assessment of the posterior airway space and imaging software reliability analysis before and after orthognathic surgery. J. Cranio Maxillofac. Surg. 2014, 42, 1428–1436. [Google Scholar] [CrossRef]
  13. Guijarro-Martinez, R.; Swennen, G.R. Three-dimensional cone beam computed tomography definition of the anatomical subregions of the upper airway: A validation study. Int. J. Oral. Maxillofac. Surg. 2013, 42, 1140–1149. [Google Scholar] [CrossRef]
  14. Katyal, V.; Pamula, Y.; Martin, A.J.; Daynes, C.N.; Kennedy, J.D.; Sampson, W.J. Craniofacial and upper airway morphology in pediatric sleep-disordered breathing: Systematic review and meta-analysis. Am. J. Orthod. Dentofacial Orthop. 2013, 143, 20–30e3. [Google Scholar] [CrossRef]
  15. Wong, S.H.; Al-Hasani, H.; Alam, Z.; Alam, A. Artificial intelligence in radiology: How will we be affected? Eur. Radiol. 2019, 29, 141–143. [Google Scholar] [CrossRef] [Green Version]
  16. Leite, A.F.; Gerven, A.V.; Willems, H.; Beznik, T.; Lahoud, P.; Gaeta-Araujo, H.; Vranckx, M.; Jacobs, R. Artificial intelligence-driven novel tool for tooth detection and segmentation on panoramic radiographs. Clin. Oral. Investig. 2021, 25, 2257–2267. [Google Scholar] [CrossRef]
  17. Casalegno, F.; Newton, T.; Daher, R.; Abdelaziz, M.; Lodi-Rizzini, A.; Schurmann, F.; Krejci, I.; Markram, H. Caries Detection with Near-Infrared Transillumination Using Deep Learning. J. Dent. Res. 2019, 98, 1227–1233. [Google Scholar] [CrossRef] [Green Version]
  18. Umer, F.; Habib, S. Critical Analysis of Artificial Intelligence in Endodontics: A Scoping Review. J. Endod. 2022, 48, 152–160. [Google Scholar] [CrossRef]
  19. De Vos, W.; Casselman, J.; Swennen, G.R. Cone-beam computerized tomography (CBCT) imaging of the oral and maxillofacial region: A systematic review of the literature. Int. J. Oral. Maxillofac. Surg. 2009, 38, 609–625. [Google Scholar] [CrossRef]
  20. Kim, H.S.; Kim, G.T.; Kim, S.; Lee, J.W.; Kim, E.C.; Kwon, Y.D. Three-dimensional evaluation of the pharyngeal airway using cone-beam computed tomography following bimaxillary orthognathic surgery in skeletal class III patients. Clin. Oral. Investig. 2016, 20, 915–922. [Google Scholar] [CrossRef]
  21. Chen, H.; Aarab, G.; de Ruiter, M.H.; de Lange, J.; Lobbezoo, F.; van der Stelt, P.F. Three-dimensional imaging of the upper airway anatomy in obstructive sleep apnea: A systematic review. Sleep. Med. 2016, 21, 19–27. [Google Scholar] [CrossRef] [PubMed]
  22. Schwab, R.J.; Pasirstein, M.; Pierson, R.; Mackley, A.; Hachadoorian, R.; Arens, R.; Maislin, G.; Pack, A.I. Identification of upper airway anatomic risk factors for obstructive sleep apnea with volumetric magnetic resonance imaging. Am. J. Respir. Crit. Care Med. 2003, 168, 522–530. [Google Scholar] [CrossRef]
  23. Hsu, W.E.; Wu, T.Y. Comparison of upper airway measurement by lateral cephalogram in upright position and CBCT in supine position. J. Dent. Sci. 2019, 14, 185–191. [Google Scholar] [CrossRef] [PubMed]
  24. Wang, C.W.; Peng, C.C. 3D Face Point Cloud Reconstruction and Recognition Using Depth Sensor. Sensors 2021, 21, 2587. [Google Scholar] [CrossRef]
  25. Ikeda, T.; Okano, S.; Hashimoto, N.; Kimura, K.; Kudo, K.; Tsutsumi, R.; Sasaki, S.; Kawasaki, J.; Miyashita, Y.; Wada, H. Histomorphological investigation of intrahepatic connective tissue for surgical anatomy based on modern computer imaging analysis. J. Hepatobiliary Pancreat. Sci. 2021, 28, 76–85. [Google Scholar] [CrossRef]
  26. Soares, J.M.; Cançado, H.R.; Sousa, A.A.; Moraes, G.M.; Carvalho, J.R.; Silva, G.S.; Porcaro-Salles, J.M. Prognostic factors for surgically resected squamous cell carcinoma of the lower oral cavity. Minerva Stomatol. 2014, 63, 341–349. [Google Scholar]
  27. Neelapu, B.C.; Kharbanda, O.P.; Sardana, V.; Gupta, A.; Vasamsetti, S.; Balachandran, R.; Rana, S.S.; Sardana, H.K. A pilot study for segmentation of pharyngeal and sino-nasal airway subregions by automatic contour initialization. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 1877–1893. [Google Scholar] [CrossRef]
  28. Neelapu, B.C.; Kharbanda, O.P.; Sardana, H.K.; Gupta, A.; Vasamsetti, S.; Balachandran, R.; Rana, S.S.; Sardana, V. The reliability of different methods of manual volumetric segmentation of pharyngeal and sinonasal subregions. Oral. Surg. Oral. Med. Oral. Pathol. Oral. Radiol. 2017, 124, 577–587. [Google Scholar] [CrossRef]
  29. Wang, M.; Jin, R.; Jiang, N.; Liu, H.; Jiang, S.; Li, K.; Zhou, X. Automated labeling of the airway tree in terms of lobes based on deep learning of bifurcation point detection. Med. Biol. Eng. Comput. 2020, 58, 2009–2024. [Google Scholar] [CrossRef]
  30. Maken, P.; Gupta, A.; Gupta, M.K. A systematic review of the techniques for automatic segmentation of the human upper airway using volumetric images. Med. Biol. Eng. Comput. 2023, 61, 1901–1927. [Google Scholar] [CrossRef]
  31. Sin, Ç.; Akkaya, N.; Aksoy, S.; Orhan, K.; Öz, U. A deep learning algorithm proposal to automatic pharyngeal airway detection and segmentation on CBCT images. Orthod. Craniofac Res. 2021, 2 (Suppl. S24), 117–123. [Google Scholar] [CrossRef]
  32. Garcia-Uceda, A.; Selvan, R.; Saghir, Z.; Tiddens, H.A.W.M.; de Bruijne, M. Automatic airway segmentation from computed tomography using robust and efficient 3-D convolutional neural networks. Sci. Rep. 2021, 11, 16001. [Google Scholar] [CrossRef]
  33. Leonardi, R.; Lo Giudice, A.; Farronato, M.; Ronsivalle, V.; Allegrini, S.; Musumeci, G.; Spampinato, C. Fully automatic segmentation of sinonasal cavity and pharyngeal airway based on convolutional neural networks. Am. J. Orthod. Dentofacial Orthop. 2021, 159, 824–835. [Google Scholar] [CrossRef] [PubMed]
  34. De Water, V.R.; Saridin, J.K.; Bouw, F.; Murawska, M.M.; Koudstaal, M.J. Measuring upper airway volume: Accuracy and reliability of Dolphin 3D software compared to manual segmentation in craniosynostosis patients. J. Oral. Maxillofac. Surg. 2014, 72, 139–144. [Google Scholar] [PubMed]
  35. Alsufyani, N.A.; Hess, A.; Noga, M.; Ray, N.; Al-Saleh, M.A.; Lagravere, M.O.; Major, P.W. New algorithm for semiautomatic segmentation of nasal cavity and pharyngeal airway in comparison with manual segmentation using cone-beam computed tomography. Am. J. Orthod. Dentofacial Orthop. 2016, 150, 703–712. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  36. Weissheimer, A.; Menezes, L.M.; Sameshima, G.T.; Enciso, R.; Pham, J.; Grauer, D. Imaging software accuracy for 3-dimensional analysis of the upper airway. Am. J. Orthod. Dentofacial Orthop. 2012, 142, 801–813. [Google Scholar] [CrossRef]
  37. Chen, H.; Zhang, K.; Lyu, P.; Li, H.; Zhang, L.; Wu, J.; Lee, C.H. A deep learning approach to automatic teeth detection and numbering based on object detection in dental periapical films. Sci. Rep. 2019, 9, 3840. [Google Scholar] [CrossRef] [Green Version]
  38. Campos, L.D.; Trindade, I.E.K.; Yatabe, M.; Trindade, S.H.K.; Pimenta, L.A.; Kimbell, J.; Drake, A.F.; Trindade-Suedam, I.K. Reduced pharyngeal dimensions and obstructive sleep apnea in adults with cleft lip/palate and Class III malocclusion. Cranio 2021, 39, 484–490. [Google Scholar] [CrossRef]
  39. Nandi, P.R.; Charlesworth, C.H.; Taylor, S.J.; Nunn, J.F.; Dore, C.J. Effect of general anaesthesia on the pharynx. Br. J. Anaesth. 1991, 66, 157–162. [Google Scholar] [CrossRef]
  40. Isono, S.; Remmers, J.E.; Tanaka, A.; Sho, Y.; Sato, J.; Nishino, T. Anatomy of pharynx in patients with obstructive sleep apnea and in normal subjects. J. Appl. Physiol. 1997, 82, 1319–1326. [Google Scholar] [CrossRef]
  41. Heinz, J.; Stewart, K.; Ghoneima, A. Evaluation of two-dimensional lateral cephalogram and three-dimensional cone beam computed tomography superimpositions: A comparative study. Int. J. Oral. Maxillofac. Surg. 2019, 48, 519–525. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The dataset collection and flow of AI processing.
Figure 1. The dataset collection and flow of AI processing.
Bioengineering 10 00915 g001
Figure 2. Landmark points on midsagittal plane images (PNS, VP, CV1, CV2 and CV4).
Figure 2. Landmark points on midsagittal plane images (PNS, VP, CV1, CV2 and CV4).
Bioengineering 10 00915 g002
Figure 3. Three parts of the upper airway were annually marked in different colours. The red, yellow and green colour regions correspond to the nasopharynx, the retropalatal pharynx and the retroglossal pharynx, respectively.
Figure 3. Three parts of the upper airway were annually marked in different colours. The red, yellow and green colour regions correspond to the nasopharynx, the retropalatal pharynx and the retroglossal pharynx, respectively.
Bioengineering 10 00915 g003
Figure 4. The architecture of the employed UNet model. Each convolution block contains 2 and 4 convolution layers (as well as batch-norm layers and ReLU activation units) for UNet18 and UNet36, respectively.
Figure 4. The architecture of the employed UNet model. Each convolution block contains 2 and 4 convolution layers (as well as batch-norm layers and ReLU activation units) for UNet18 and UNet36, respectively.
Bioengineering 10 00915 g004
Figure 5. Manual segmentation results (left) and UNet18 model segmentation result (right). The red, yellow and green colour regions correspond to the nasopharynx, the retropalatal pharynx and the retroglossal pharynx, respectively.
Figure 5. Manual segmentation results (left) and UNet18 model segmentation result (right). The red, yellow and green colour regions correspond to the nasopharynx, the retropalatal pharynx and the retroglossal pharynx, respectively.
Bioengineering 10 00915 g005
Figure 6. AI-driven results of the CSAmin localisation in GT1 (left) and GT2 groups (right).
Figure 6. AI-driven results of the CSAmin localisation in GT1 (left) and GT2 groups (right).
Bioengineering 10 00915 g006
Figure 7. Bland–Altman figure of CSAmin height level between HM and H1.
Figure 7. Bland–Altman figure of CSAmin height level between HM and H1.
Bioengineering 10 00915 g007
Figure 8. Bland–Altman figure of CSAmin height level between HM and H2.
Figure 8. Bland–Altman figure of CSAmin height level between HM and H2.
Bioengineering 10 00915 g008
Table 1. Definitions and abbreviations of upper airway landmarks in midsagittal plane.
Table 1. Definitions and abbreviations of upper airway landmarks in midsagittal plane.
Reference PointsExplanation
PNSMost posterior point of palate
VPMost posterior point of vomer
CV1Most anterior inferior point of anterior arch of atlas
CV2Most anterior inferior point of anterior arch of second vertebra
CV4Most anterior inferior point of anterior arch of fourth vertebra
PNS = posterior nasal spine, VP = posterior point of vomer, CV1 = the first cervical vertebra, CV2 = the second cervical vertebra, CV4 = the fourth cervical vertebra.
Table 2. Performance (accuracy metrics) of the DeepLab system for segmenting the upper airway in 2D images. IoU = Intersection over Union, DSC = dice similarity coefficient.
Table 2. Performance (accuracy metrics) of the DeepLab system for segmenting the upper airway in 2D images. IoU = Intersection over Union, DSC = dice similarity coefficient.
ModelDeepLab50 DeepLab101
PrecisionRecallIoUDSCSize DifferencePrecisionRecallIoUDSCSize Difference
Nasopharynx93.179.975.485.5146.192.782.177.086.7122.3
Retropalatal pharynx87.390.479.888.6102.189.189.580.689.1101.1
Retroglossal pharynx89.793.684.291.3100.490.094.385.091.892.2
Overall90.088.079.888.4116.290.688.680.989.2105.2
p value0.3690.3810.4150.4880.039 *0.4510.4670.4340.4510.322
Table 3. Performance (accuracy metrics) of the UNet system for segmenting the upper airway in 2D images. IoU = Intersection over Union, DSC = dice similarity coefficient.
Table 3. Performance (accuracy metrics) of the UNet system for segmenting the upper airway in 2D images. IoU = Intersection over Union, DSC = dice similarity coefficient.
ModelUNet18 UNet36
PrecisionRecallIoUDSCSize DifferencePrecisionRecallIoUDSCSize Difference
Nasopharynx90.885.578.887.983.090.385.978.687.790.5
Retropalatal pharynx87.690.880.488.9115.088.889.880.589.0102.8
Retroglossal pharynx92.091.384.491.3101.591.391.083.690.7101.8
Overall90.289.281.289.499.890.188.980.989.198.4
p value0.4690.3980.4160.4330.2700.4160.4880.4980.4330.807
Table 4. Consistency of CSAmin localisation in specific segments of upper airway between manual work and AI prediction in test group.
Table 4. Consistency of CSAmin localisation in specific segments of upper airway between manual work and AI prediction in test group.
CSAmin LocalisationManual Work Total
Retropalatal pharynxRetroglossal pharynx
AI prediction
Retropalatal pharynx27027
Retroglossal pharynx11213
Total281240
Table 5. Performance of AI-driven CSAmin localisation results in GT1 and GT2 groups. SD, standard deviation.
Table 5. Performance of AI-driven CSAmin localisation results in GT1 and GT2 groups. SD, standard deviation.
Patient IDHT (Pixel)H1 (Pixel)H2 (Pixel)L1 (Pixel)L2 (Pixel)L1 (mm)L2 (mm)
170168573716363.527.91
2642633649971.981.54
3612668576563612.317.91
4547551541460.881.32
5597599598210.440.22
658960157412152.643.3
7587583583440.880.88
8640645634561.11.32
974276372621164.623.52
106626606512110.442.42
115505595379131.982.86
1261562960414113.082.42
136406406370300.66
1459060157511152.423.3
15571569579280.441.76
16578583576521.10.44
17633628630531.10.66
185695685571120.222.64
195585635475111.12.42
2064565863313122.862.64
2167566369112162.643.52
225815715731082.21.76
2374071777123315.056.81
246316256206111.322.42
256326296223100.662.2
266306396139171.983.74
276776796672100.442.2
2862662663701102.42
29621613625841.760.88
3054653156215163.33.52
31596599596300.660
325705655565141.13.08
33602596600621.320.44
34683679676470.881.54
356776776700701.54
36698699696120.220.44
3765963070429456.379.89
3866167364412172.643.74
39586581587511.10.22
40646651649531.10.66
Mean625.13625.73622.588.8511.501.952.53
SD50.2450.3856.2710.0710.082.212.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chu, G.; Zhang, R.; He, Y.; Ng, C.H.; Gu, M.; Leung, Y.Y.; He, H.; Yang, Y. Deep Learning Models for Automatic Upper Airway Segmentation and Minimum Cross-Sectional Area Localisation in Two-Dimensional Images. Bioengineering 2023, 10, 915. https://doi.org/10.3390/bioengineering10080915

AMA Style

Chu G, Zhang R, He Y, Ng CH, Gu M, Leung YY, He H, Yang Y. Deep Learning Models for Automatic Upper Airway Segmentation and Minimum Cross-Sectional Area Localisation in Two-Dimensional Images. Bioengineering. 2023; 10(8):915. https://doi.org/10.3390/bioengineering10080915

Chicago/Turabian Style

Chu, Guang, Rongzhao Zhang, Yingqing He, Chun Hown Ng, Min Gu, Yiu Yan Leung, Hong He, and Yanqi Yang. 2023. "Deep Learning Models for Automatic Upper Airway Segmentation and Minimum Cross-Sectional Area Localisation in Two-Dimensional Images" Bioengineering 10, no. 8: 915. https://doi.org/10.3390/bioengineering10080915

APA Style

Chu, G., Zhang, R., He, Y., Ng, C. H., Gu, M., Leung, Y. Y., He, H., & Yang, Y. (2023). Deep Learning Models for Automatic Upper Airway Segmentation and Minimum Cross-Sectional Area Localisation in Two-Dimensional Images. Bioengineering, 10(8), 915. https://doi.org/10.3390/bioengineering10080915

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop