Next Article in Journal
Clinical and Novel Biomarkers in Penile Carcinoma: A Prospective Review
Previous Article in Journal
Racial/Ethnic Differences in the Pharmacokinetics of Antipsychotics: Focusing on East Asians
Previous Article in Special Issue
Prediction of All-Cause Mortality Based on Stress/Rest Myocardial Perfusion Imaging (MPI) Using Deep Learning: A Comparison between Image and Frequency Spectra as Input
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Position Classification of the Endotracheal Tube with Automatic Segmentation of the Trachea and the Tube on Plain Chest Radiography Using Deep Convolutional Neural Network

1
School of Medicine, College of Medicine, Hanyang University, Seoul 04763, Korea
2
Department of Computer Science, Hanyang University, Seoul 04763, Korea
3
Department of Emergency Medicine, College of Medicine, Hanyang University, Seoul 04763, Korea
4
Machine Learning Research Center for Medical Data, Hanyang University, Seoul 04763, Korea
5
Department of Otolaryngology—Head and Neck Surgery, College of Medicine, Hanyang University, Seoul 04763, Korea
6
Department of HY-KIST Bio-Convergence, College of Medicine, Hanyang University, Seoul 04763, Korea
7
Department of Emergency Medicine, Seoul National University Bundang Hospital, Seongnam-si 13620, Korea
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work as first authors.
J. Pers. Med. 2022, 12(9), 1363; https://doi.org/10.3390/jpm12091363
Submission received: 9 June 2022 / Revised: 20 August 2022 / Accepted: 22 August 2022 / Published: 24 August 2022
(This article belongs to the Special Issue Application of Artificial Intelligence in Personalized Medicine)

Abstract

:
Background: This study aimed to develop an algorithm for multilabel classification according to the distance from carina to endotracheal tube (ETT) tip (absence, shallow > 70 mm, 30 mm ≤ proper ≤ 70 mm, and deep position < 30 mm) with the application of automatic segmentation of the trachea and the ETT on chest radiographs using deep convolutional neural network (CNN). Methods: This study was a retrospective study using plain chest radiographs. We segmented the trachea and the ETT on images and labeled the classification of the ETT position. We proposed models for the classification of the ETT position using EfficientNet B0 with the application of automatic segmentation using Mask R-CNN and ResNet50. Primary outcomes were favorable performance for automatic segmentation and four-label classification through five-fold validation with segmented images and a test with non-segmented images. Results: Of 1985 images, 596 images were manually segmented and consisted of 298 absence, 97 shallow, 100 proper, and 101 deep images according to the ETT position. In five-fold validations with segmented images, Dice coefficients [mean (SD)] between segmented and predicted masks were 0.841 (0.063) for the trachea and 0.893 (0.078) for the ETT, and the accuracy for four-label classification was 0.945 (0.017). In the test for classification with 1389 non-segmented images, overall values were 0.922 for accuracy, 0.843 for precision, 0.843 for sensitivity, 0.922 for specificity, and 0.843 for F1-score. Conclusions: Automatic segmentation of the ETT and trachea images and classification of the ETT position using deep CNN with plain chest radiographs could achieve good performance and improve the physician’s performance in deciding the appropriateness of ETT depth.

Graphical Abstract

1. Introduction

Endotracheal tube (ETT) intubation is an important procedure for airway management and ventilation of emergent and critical patients [1]. The distal tip of the ETT should not be placed in the esophagus, but in the trachea, and the correct depth position is approximately 30–70 mm from the carina, which would be in the middle third of the trachea in adult patients [2,3,4]. If the tip of the ETT is inserted too deeply, it could cause bronchial intubation, which leads to hyperinflation of the ipsilateral lung with atelectasis of the other, and barotrauma and hypoxemia could occur, resulting in an increased risk of mortality [5,6,7]. If the tip is inserted too shallowly in the first third of the trachea, it could cause accidental extubation, laryngeal spasm, and aspiration pneumonia, depending on changes in the patient’s head position [7,8,9].
Methods for ensuring correct positioning of the ETT include physical examination methods (auscultation of chest and epigastrium, visualization of thoracic movement, fogging in the tube), pulse oximetry, chest radiography, capnography, ultrasound, bronchoscopy, etc. [7,10,11,12,13,14]. Physical examination methods, as well as pulse oximetry and chest radiography, are not sufficiently reliable. Capnography is one of the best methods to confirm ETT placement in the trachea and can be used in patients who have adequate tissue perfusion [10,11,12,13,14,15]. Ultrasound is one of the portable and repeatable screening methods for the same purpose and may be used by someone who is experienced in this technique [14]. However, these methods cannot measure the distance from the carina to the tip of the ETT. Bronchoscopy is the gold standard for confirming the correct position of ETT and is always not available [15]. The chest radiograph is not the gold standard for this purpose, but certainly forms an important part of the overall assessment because it can be measured on an image [2,7]. However, measuring the distance on an image from chest radiography may be difficult for a novice clinician, or even for an expert, when working with poor image quality or under work overload conditions [16].
Recently, several studies reported on computer-aided assessment of ETT tube presence and position on chest radiography images using deep convolutional neural networks (CNN) [17,18,19,20]. They almost all showed that the network performed well in assessing ETT presence or absence with more than 0.97~0.99 accuracy, whereas its performance in assessing the proper positioning of the ETT tip was lacking or poorer. The technique of segmentation or localization of the carina and the ETT tip could determine malposition of the ETT or measure the distance between them and might improve performance in the classification of the ETT position [20]. There has been no research on the automatic segmentation of trachea and ETT images or on classification of proper positioning of the ETT tip. This study aims to develop an algorithm for the multi-label classification of ETT tip position (absence, shallow, proper, and deep position of ETT) with the application of automatic segmentation of the trachea and the ETT on chest radiographs using deep CNN.

2. Methods

2.1. Study Design

This is a retrospective study using plain chest anteroposterior (AP) radiography for segmentation of the trachea and the ETT tube and using CNN to determine proper positioning between them. The study was conducted at one regional emergency center of the tertiary hospital in Seoul (Seoul, Korea) between June and December 2021, and used data from January 2018 to March 2021. This study was approved by the Institutional Review Boards (IRB) of Hanyang University Hospital (ref. no. 2021-05-040) and the requirements for informed consent were waived by the IRBs of Hanyang University Hospital. All methods and procedures were carried out in accordance with the Declaration of Helsinki.

2.2. Dataset of Participants

A flowchart of data collection and analysis is shown in Figure 1.
Using the airway registry, we sorted and gathered chest AP images from a list of patients who underwent ETT intubation in the emergency room between January 2018 and March 2021 [21]. Images with the ETT were extracted from chest radiographs that had been used to check for confirmation or follow-up after intubation, and images without ETT were extracted from chest radiographs that had been taken for examination before intubation or after extubation of the ETT. We excluded images with tracheostomy tubes or images of poor quality due to severe damage to anatomical structures or massive subcutaneous emphysema by trauma. However, we did not exclude images with electrodes or nasogastric tubes or with central vein catheters. Two authors (Jung H.C., and Oh J.) reached consensus on each position of the carina and the distal tip of ETT and measured the distance between them in millimeters (mm) using the scale in the picture archiving and communication system (PACS, Centricity, GE Healthcare, Milwaukee, WI, USA). These images were categorized into four classes according to the ETT position, as follows: (1) absence (absence of the ETT on the image), (2) shallow position (more than 70 mm), (3) proper position (between 30 mm and 70 mm), (4) deep position (less than 30 mm). These images were extracted and stored as digital imaging and communication and medicine (DICOM)-format images using the PACS system. No personal information was included when saving images for data collection, and data were obtained without personal identifying data. In addition, arbitrary numbers were assigned to images, which were then coded and managed.

2.3. Segmentation of the Trachea and the ETT on Images and Data Processing

The author (Jung H.C.) performed a binary mask by painting the entire trachea area and ETT on images in consensus with another author (Oh J.) for the segmentation of the trachea and ETT. Both the trachea and ETT were segmented on images with ETT, whereas only the trachea was segmented on images without ETT, and these segmented images were stored as Nifty (Neuroimaging Informatics Technology Initiative) using the AVIEW program (AVIEW-Research, Corelinesoft, Seoul, Korea) (Figure 2). Moreover, we converted the segmented binary mask into COCO Dataset format [22]. To be specific, the binary mask was converted to a 2D polygon according to the COCO Dataset format. To obtain a 2D polygon, we computed the contour coordinates of the mask using the marching squares algorithm [23] and linearly interpolated them.

2.4. Proposed Models for Classification of Four Classes of ETT Position with Application of Automatic Segmentation Using Deep CNN

Classification of the ETT position on chest radiograph images is a typical multi-label task. Therefore, we used a classifier based on EfficientNet [24], which took only a single chest radiograph input image as our baseline. To further elevate the performance of the depth classification task, we proposed a 2-stage approach, which first segmented the trachea and the ETT areas, then used the segmentation results for the classification task. For the segmentation, we used the fully pretrained Mask R-CNN [25] as initial reference and fine-tuned it with our own dataset to segment out the trachea and the ETT regions. For the classification, the input image and binary masks of trachea and tube detected by our segmentation network were concatenated, and they were fed into the classification network (i.e., EfficientNet) based on inputs. The overall flow of the proposed model is illustrated in Figure 3.

2.4.1. Network for the Segmentation of the Trachea and ETT

For the segmentation of the trachea and ETT areas, we employed conventional Mask R-CNN with the ResNet50 [26] feature pyramid network (FPN) [27] pretrained on the ImageNet dataset [28]. By using our own training dataset, we further updated the initial parameters of Mask R-CNN. Specifically, we fine-tuned the Mask R-CNN to segment the trachea and the ETT regions for 80 k iterations with the stochastic gradient descent (SGD) optimizer with momentum. The learning rate was set to 2.5 × 10−4, and we used a batch size of 128 for training.

2.4.2. Network for Multi-Label Classification for Proper Positioning of the ETT in the Trachea

To detect proper positioning of the ETT in the trachea, we used the conventional EfficientNet B0 model for the multi-label classification problem. Unlike the original EfficientNet, which takes a colored image as an input, we utilized the segmentation results to elevate the classification performance. To be specific, we generated a 3-channel input including a gray-scale radiograph image and two segmentation results of trachea and tube from the segmentation network. In addition, we cropped the 3-channel image using the bounding boxes predicted by Mask R-CNN, and then resized the cropped image to 224 × 224 resolution, as shown in Figure 3. We used the initially pretrained EfficientNet on the ImageNet dataset and fine-tuned the network parameters with our own dataset. During training, we minimized the cross-entropy loss with the Adam optimizer [29]. The learning rate was fixed to 1 × 10−5 and the batch size was set to 32.

2.5. Experiments

We separately trained the proposed networks for the segmentation and classification tasks, and five-fold cross-validation was used for the performance evaluation with the training and validation set consisting of segmented images. Chest radiograph images were randomly divided into five parts; four out of the five images were used for training and the other was used for validation. Using this strategy, we could train the five models with five different training and validation datasets. Finally, we trained the optimized model with a total set consisting of segmented images and tested the performance of the ETT position classification using a test set consisting of images without segmentation mask.

2.6. Primary Outcomes

Our primary outcome was a favorable performance for automatic segmentation and four-label classification of the relationship between the trachea and the tip of the ETT. To evaluate the performance of the segmentation model, we used the Dice coefficient (Dice) to estimate the degree of overlap between the ground truth area and the predicted area, as follows:
Dice = 2 TP 2 TP + FP + FN
To evaluate the performance of the classification model, we used the module with accuracy, precision, sensitivity, specificity, and F1-score in five-fold validation and a test. These values were analyzed according to each label performance. The prediction label was selected according to the class of which the prediction probability was highest among four classes. Accuracy is the fraction of correct predictions over total predictions. Precision is the positive predictive value that is the fraction of true positives (TP) among true and false positive (FP). Sensitivity (recall) is the fraction of TP among TP and false negative (FN). Specificity is the fraction of TN among TN and FP. F1-score is the harmonic mean of precision and recall as follows:
F 1   score   = 2 × precision × recall precision + recall

2.7. Statistical Analysis

Data were compiled using a standard spreadsheet application (Excel 2016; Microsoft, Redmond, WA, USA) and analyzed using NCSS 12 (Statistical Software 2018, NCSS, LLC, Kaysville, UT, USA, ncss.com/software/ncss (accessed on 30 August 2021)). Kolmogorov–Smirnov tests were performed to demonstrate normal distribution of all datasets. We generated descriptive statistics and presented them as frequency and percentage for categorical data and as either median and interquartile range (IQR) (non-normal distribution) or mean and standard deviation (SD) (normal distribution) for continuous data. Student t-tests or Mann–Whitney tests were used to compare the characteristics of the segmented and non-segmented data. p-values < 0.05 were considered statistically significant.

3. Results

A total of 2092 images, including 932 images without the ETT and 1160 images with the ETT, were collected from 934 patients. These 934 patients consisted of 588 males (63%) with a mean [SD] age of 63.5 [18.3] years. A total of 107 images were excluded according to the exclusion criteria: 94 were images with tracheostomy tubes, and 13 images of poor quality by trauma (such as severe damage to anatomical structures or massive subcutaneous emphysema). Finally, a total of 1985 images, including 891 images without the ETT and 1094 images with the ETT were included in our study. Of the 1094 images with ETT, there were 129 images with shallow ETT position, 817 images with proper ETT position, and 148 images with deep ETT position.
Out of the total number of images, 298 images without ETT, 97 images with shallow ETT position, 100 images with the ETT in the proper position, and 101 images with the ETT in deep position were segmented for training and five-fold validation. The distances between the carina of the trachea and the tip of the ETT were not significantly different between segmented and non-segmented datasets in all labels (all p > 0.05) (Table 1).

3.1. Five-Fold Validation for Automatic Segmentation of the Trachea and the ETT and Classification of Four Labels According to the ETT Position

The results for the five-fold validation of automatic segmentation and classification of four labels are shown in Table 2. Dice [mean (SD)] between segmented and predicted masks were 0.841 (0.063) for the trachea and 0.893 (0.078) for the ETT through five-fold validation. Values of the ETT according to position of the ETT were 0.879 (0.113) for shallow position, 0.908 (0.055) for proper position, and 0.892 (0.045) for deep position. Overall accuracy, F1 score [mean (SD)] of five-fold validation was 0.889 (0.034) (Table 2).

3.2. Performance Test of Each Label for Classification of Four Labels with Non-Segmented Images after Training with All Segmented Images

The confusion matrix and outcomes of the test for classification are shown in Table 3. Overall values of accuracy, precision, sensitivity, specificity, and F1-score were 0.922, 0.843, 0.843, 0.922, and 0.843, respectively. All values were highest in the label of absence among the four labels. Accuracy had the lowest value of 0.853 in the label of the proper position, and F1-score had the lowest value of 0.325 in the label of the shallow position. Sensitivity was higher in the shallow and deep positions than in the proper position, whereas precision values were the reverse.

4. Discussion

In the review article on computer-aided assessment of catheter and tube positions on radiographs, the authors recommended answering five-stage research questions covering presence, detection of the tip, course, type, and satisfactory position, as a system of evaluating catheter placement [30]. Compared to a general binary classifier, which may predict ETT absence/presence or ETT proper/improper position, our proposed method with automatic segmentation and classification of ETT and trachea using deep learning could provide four labels: absent, shallow, proper, and deep placement of ETT tip relative to the carina.
In a study on segmentation of ETT and trachea images using deep learning and chest radiographs, the mean Dice scores of the ETT and the carina were 0.861 and 0.574, respectively, in the validation set [20]. The authors manually segmented the distal two-thirds of the ETT and carina, drawing a triangle that outlined the visible border of the main stem bronchi. In our study, the mean (SD) Dice for segmentation of the ETT and trachea were 0.841 (0.063) and 0.893 (0.078), respectively, in five-fold validation. We segmented the entire ETT and trachea shown in the chest radiograph.
Previous studies have almost all used computerized methods to detect the presence of an ETT with results of approximately 0.93 (sensitivity) but did not assess the proper position of the ETT, which is a more crucial clinical judgment [16,31,32]. Recently, several studies have reported detection of the proper position including the presence of the ETT, or prediction of the distance from the tip of the ETT to the carina, using chest radiographs. Kara et al. reported that the performance of binary classification of presence or absence improved to 0.971 (accuracy), 0.971 (precision), and 0.974 (sensitivity) after identification and localization of the ETT using 16,000 images on the open-source MIMIC chest radiographs (MIMIC-CXR) dataset [19]. They also showed that the distance between the carina and distal ETT tip was predicted within a median error of 6 mm from manual ground-truth annotations in the overall final CNN assessment. Frid-Adar et al. reported that performance in detection of the ETT also reached an area under the curve (AUC) of 0.987, sensitivity of 0.955, and specificity of 0.965 with a test set of National Institutes of Health (NIH) dataset after training with several thousand synthetic and real images. However, these studies did not train the classification of proper ETT positions [32]. Our proposed method showed favorable performance for the detection of ETT absence among four labels, achieving 0.985 (accuracy), 0.986 (precision), 0.978 (sensitivity), and 0.990 (specificity) in the test set with non-segmented images after training with about 600 segmented images.
Lakhani trained a GoogLeNet convolutional neural network (CNN) to determine the presence/absence and low/normal position of the ETT on 300 labeled chest radiographs without automatic segmentation of the ETT and trachea, and achieved an AUC of 0.989 and 0.809, respectively [17]. In a study using approximately 23,000 chest radiographs and Inception V3 deep learning architecture, authors classified images into 12 categories, according to the distance from the carina and the tip of the ETT, at 1.0 cm intervals. The sensitivity and specificity for detecting shallow positioning of the ETT (i.e., distance > 70 mm) were 0.665 and 0.924; the sensitivity and specificity for detecting deep positioning of the ETT (i.e., distance < 20 mm) were 0.901 and 0.924 [18]. The authors found that their network misclassified the ground truth of 70–80 mm above the carina to a prediction of 60–70 mm, which negatively impacted sensitivity. We also achieved high-performance outcomes of 0.844 (sensitivity) and 0.921 (specificity) for detecting shallow positioning of the ETT (i.e., distance > 70 mm), and 0.894 (sensitivity) and 0.930 (specificity) for detecting deep positioning of the ETT (i.e., the distance < 30 mm), while the precision and F1-score were low, at about 0.3~0.5 for shallow and deep positioning of the ETT. In this study, we decided the proper depth of the ETT was in the range of 30 mm and 70 mm. The precision value might be lower than other values because the mean (SD) of the distance in the images was 78.56 mm (6.79 mm) in the shallow position and 19.09 mm (8.32 mm) in the deep position, which is near the proper depth of the ETT. Our proposed method could be helpful as a screening tool to detect excessively shallow and deep positions of the ETT, which could be critical for the patient.
There were several limitations in this study. First, the data of the chest radiographs and the patients originated from a single center, and our proposed model might not be suitable for the environment of other hospitals. Second, we decided that the distance for proper positioning of the ETT was 30–70 mm. This standard could be different according to other hospitals. Third, this method could not detect esophageal intubation, because there were no images of esophageal intubation in our dataset. Fourth, we did not assess how different lung conditions (such as pleural effusion, pneumonic infiltration, pneumothorax, and the underlying pathology of the respiratory insufficiency that led to intubation) influence the results of our proposed CNN method. Finally, we did not compare the performance of our model to that of physicians with respect to key factors such as clinical outcomes, the time required to reach a diagnosis, and the equipment required to use the model as a screening tool.

5. Conclusions

Automatic segmentation of ETT and trachea images and classification of the ETT position using CNN with plain chest radiographs could achieve good performance and improve the physician’s performance in deciding the appropriateness of the ETT depth.

Author Contributions

J.O. and T.H.K. conceived the study and designed the trial. J.O., H.C.J., H.B., J.H.C., J.L. and D.K.L. supervised the trial process and were involved in data collection. T.H.K., C.K., B.K. and M.S.Y. analyzed all images and data. H.C.J., C.K., J.O. and T.H.K. drafted the manuscript, and all authors substantially contributed to its revision. J.O. and T.H.K. take responsibility for the content of the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the National Research Foundation of Korea (NRF-2022R1A2C1012627 and Institute of Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korean government (MSIT) (No. 2020-0-01373, Artificial Intelligence Graduate School Program (Hanyang University)).

Institutional Review Board Statement

This study was approved by the Institutional Review Boards of Hanyang University Hospital (ref. No. 2020-01-005). All methods and procedures were carried out in accordance with the Declaration of Helsinki.

Informed Consent Statement

Informed consent was waived by the IRB due to the retrospective study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

ETTendotracheal tube
CNNconvolution neural network
APanteroposterior
PACSpicture archiving and communication system
DICOMdigital imaging and communication and medicine
NiftyNeuroimaging Informatics Technology Initiative
FPNfeature pyramid network
SGDstochastic gradient descent
SDstandard deviation
TPtrue positive
FPfalse positive
FNfalse negative
NIHNational Institutes of Health

References

  1. Sitzwohl, C.; Langheinrich, A.; Schober, A.; Krafft, P.; I Sessler, D.; Herkner, H.; Gonano, C.; Weinstabl, C.; Kettner, S.C. Endobronchial intubation detected by insertion depth of endotracheal tube, bilateral auscultation, or observation of chest movements: Randomised trial. BMJ 2010, 341, c5943. [Google Scholar] [CrossRef] [PubMed]
  2. Goodman, L.R.; Conrardy, P.A.; Laing, F.; Singer, M.M. Radiographic evaluation of endotracheal tube position. AJR Am. J. Roentgenol. 1976, 127, 433–434. [Google Scholar] [CrossRef] [PubMed]
  3. Peitzman, A.B.; Yealy, D.M.; Fabian, T.C.; Schwab, C.W. The Trauma Manual: Trauma and Acute Care Surgery; Wolters Kluwer: Philadelphia, PA, USA, 2019. [Google Scholar]
  4. Allan, E.; Giggens, R.; Ali, T.; Bhuva, S. The ICU Chest Radiograph: Line and Tube Essentials for Radiologists and ICU Physicians 2019: European Congress of Radiology-ECR 2019. Available online: https://epos.myesr.org/poster/esr/ecr2019/C-3024 (accessed on 1 April 2022).
  5. Owen, R.L.; Cheney, F.W. Endobronchial intubation: A preventable complication. Anesthesiology 1987, 67, 255–257. [Google Scholar] [CrossRef] [PubMed]
  6. Szekely, S.M.; Webb, R.K.; Williamson, J.A.; Russell, W.J. The Australian Incident Monitoring Study. Problems related to the en-dotracheal tube: An analysis of 2000 incident reports. Anaesth. Intensiv. Care 1993, 21, 611–616. [Google Scholar] [CrossRef]
  7. Brunel, W.; Coleman, D.L.; Schwartz, D.E.; Peper, E.; Cohen, N.H. Assessment of routine chest roentgenograms and the physical examination to confirm endotracheal tube position. Chest 1989, 96, 1043–1045. [Google Scholar] [CrossRef]
  8. Kim, J.T.; Kim, H.J.; Ahn, W.; Kim, H.S.; Bahk, J.H.; Lee, S.C.; Kim, C.S.; Kim, S.D. Head rotation, flexion, and extension alter endotra-cheal tube position in adults and children. Can. J. Anaesth. 2009, 56, 751–756. [Google Scholar] [CrossRef]
  9. Zwillich, C.W.; Pierson, D.J.; Creagh, C.; Sutton, F.D.; Schatz, E.; Petty, T.L. Complications of assisted ventilation. A prospective study of 354 consecutive episodes. Am. J. Med. 1974, 57, 161–170. [Google Scholar] [CrossRef]
  10. Stone, D.J.; Gal, T.J. Airway management. In Anesthesia, 5th ed.; Miller, R.D., Ed.; Churchill Livingstone: Philadephia, PA, USA, 2000; pp. 1431–1432. [Google Scholar]
  11. Bozeman, W.P.; Hexter, D.; Liang, H.K.; Kelen, G.D. Esophageal detector device versus detection of end-tidal carbon dioxide level in emergency intubation. Ann. Emerg. Med. 1996, 27, 595–599. [Google Scholar] [CrossRef]
  12. Grmec, S. Comparison of three different methods to confirm tracheal tube placement in emergency intubation. Intensive Care Med. 2002, 28, 701–704. [Google Scholar] [CrossRef]
  13. Chou, E.H.; Dickman, E.; Tsou, P.Y.; Tessaro, M.; Tsai, Y.M.; Ma, M.H.; Lee, C.C.; Marshall, J. Ultrasonography for confirmation of en-dotracheal tube placement: A systematic review and meta-analysis. Resuscitation 2015, 90, 97–103. [Google Scholar] [CrossRef]
  14. Song, Y.; Oh, J.; Chee, Y.; Lim, T.; Kang, H.; Cho, Y. A novel method to position an endotracheal tube at the correct depth using an infrared sensor stylet. Can. J. Anaesth. 2013, 60, 444–449. [Google Scholar] [CrossRef] [PubMed]
  15. Mitra, A.; Gave, A.; Coolahan, K.; Nguyen, T. Confirmation of endotracheal tube placement using disposable fiberoptic bronchoscopy in the emergent setting. World J. Emerg. Med. 2019, 10, 210–214. [Google Scholar] [CrossRef] [PubMed]
  16. Chen, S.; Zhang, M.; Yao, L.; Xu, W. Endotracheal tubes positioning detection in adult portable chest radiography for intensive care unit. Int. J. Comput. Assist. Radiol. Surg. 2016, 11, 2049–2057. [Google Scholar] [CrossRef] [PubMed]
  17. Lakhani, P. Deep Convolutional Neural Networks for Endotracheal Tube Position and X-ray Image Classification: Chal-lenges and Opportunities. J. Digit. Imaging 2017, 30, 460–468. [Google Scholar] [CrossRef] [PubMed]
  18. Lakhani, P.; Flanders, A.; Gorniak, R. Endotracheal Tube Position Assessment on Chest Radiographs Using Deep Learning. Radiol. Artif. Intell. 2020, 3, e200026. [Google Scholar] [CrossRef]
  19. Kara, S.; Akers, J.Y.; Chang, P.D. Identification and Localization of Endotracheal Tube on Chest Radiographs Using a Cascaded Convolutional Neural Network Approach. J. Digit. Imaging 2021, 34, 898–904. [Google Scholar] [CrossRef]
  20. Schultheis, W.G.; Lakhani, P. Using Deep Learning Segmentation for Endotracheal Tube Position Assessment. J. Thorac. Imaging 2022, 37, 125–131. [Google Scholar] [CrossRef]
  21. Kim, C.; Kang, H.; Lim, T.H.; Choi, B.Y.; Shin, Y.J.; Choi, H.J. What factors affect the success rate of the first attempt at endotracheal intubation in emergency departments? Emerg. Med. J. 2013, 30, 888–892. [Google Scholar] [CrossRef]
  22. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft coco: Common objects in context. In Proceedings of the European Conference on Computer Vision, Zürich, Switzerland, 5–12 September 2014; pp. 740–755. [Google Scholar] [CrossRef]
  23. Lorensen, W.E.; Cline, H.E. Marching cubes: A high resolution 3D surface construction algorithm. ACM Siggraph Comput. Graph. 1987, 21, 163–169. [Google Scholar] [CrossRef]
  24. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  25. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  26. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Caesars Palace, NV, USA, 1–26 June 2016; pp. 770–778. [Google Scholar]
  27. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  28. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  29. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar] [CrossRef]
  30. Yi, X.; Adams, S.J.; Henderson, R.D.E.; Babyn, P. Computer-aided Assessment of Catheters and Tubes on Radiographs: How Good Is Artificial Intelligence for Assessment? Radiol. Artif. Intell. 2020, 2, e190082. [Google Scholar] [CrossRef] [PubMed]
  31. Kao, E.F.; Jaw, T.S.; Li, C.W.; Chou, M.C.; Liu, G.C. Automated detection of endotracheal tubes in paediatric chest radiographs. Comput. Methods Programs Biomed. 2015, 118, 1–10. [Google Scholar] [CrossRef] [PubMed]
  32. Frid-Adar, M.; Amer, R.; Greenspan, H. (Eds.) Endotracheal tube detection and segmentation in chest radiographs using synthetic data. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Cham, Switzerland, 2019. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Flowchart of data collection for training/validation with segmented images and a test set with non-segmented images in the study.
Figure 1. Flowchart of data collection for training/validation with segmented images and a test set with non-segmented images in the study.
Jpm 12 01363 g001
Figure 2. Segmentation of the trachea and the ETT on images according to four labels. (A) Absence of the ETT. (B) Shallow position of the ETT. (C) Proper position of the ETT. (D) Deep position of the ETT. Both the trachea and ETT were segmented on images with ETT, whereas only the trachea was segmented on images without ETT, and these segmented images were stored as Nifty (Neuroimaging Informatics Technology Initiative).
Figure 2. Segmentation of the trachea and the ETT on images according to four labels. (A) Absence of the ETT. (B) Shallow position of the ETT. (C) Proper position of the ETT. (D) Deep position of the ETT. Both the trachea and ETT were segmented on images with ETT, whereas only the trachea was segmented on images without ETT, and these segmented images were stored as Nifty (Neuroimaging Informatics Technology Initiative).
Jpm 12 01363 g002
Figure 3. Overall flow of network architecture for segmentation of the trachea and ETT and for multi-label classification for proper position of the ETT in trachea. Mask R-CNN outputs segmentation results and bounding boxes. These results were fed into Efficientnet as an additional input to improve the multi-label classification accuracy. FC, fully connected layers.
Figure 3. Overall flow of network architecture for segmentation of the trachea and ETT and for multi-label classification for proper position of the ETT in trachea. Mask R-CNN outputs segmentation results and bounding boxes. These results were fed into Efficientnet as an additional input to improve the multi-label classification accuracy. FC, fully connected layers.
Jpm 12 01363 g003
Table 1. Comparison of the distances from the carina of the trachea to the tip of the ETT between segmented and non-segmented datasets according to the labels.
Table 1. Comparison of the distances from the carina of the trachea to the tip of the ETT between segmented and non-segmented datasets according to the labels.
The Distance from the Carina to Tip of the ETT, Mean (SD)
Segmented Data
For the Training and 5-Fold Validation
Non-Segmented Data
For the Test
p-Value
LabelsAbsence---
Shallow, mm78.76 (9.83)78.56 (6.79)0.915
Proper, mm48.68 (11.79)47.60 (10.53)0.343
Deep, mm20.56 (6.85)19.09 (8.32)0.261
ETT, endotracheal tube; SD, standard deviation.
Table 2. Outcomes of five-fold validation for automatic segmentation of the trachea and the ETT and classification of four labels according to the ETT position.
Table 2. Outcomes of five-fold validation for automatic segmentation of the trachea and the ETT and classification of four labels according to the ETT position.
1-Fold2-Fold3-Fold4-Fold5-FoldAverage of 5-Fold, Mean (SD)
Segmentation
Dice of the trachea, mean (SD)0.840 (0.069)0.843 (0.044)0.847 (0.057)0.846 (0.070)0.832 (0.068)0.841 (0.063)
Dice of the ETT, mean (SD)0.895 (0.065)0.881 (0.132)0.886 (0.060)0.908 (0.041)0.896 (0.056)0.893 (0.078)
ShallowProperDeep
0.879
(0.113)
0.908
(0.055)
0.892
(0.045)
Classification
Overall accuracy (F1 score)0.9420.8670.8820.8990.8560.889 (0.034)
ETT, endotracheal tube; Dice, Dice coefficient; SD, standard deviation.
Table 3. Confusion matrix (A) and outcomes (B) of performance test for classification of four labels with non-segmented images after training with all segmented images.
Table 3. Confusion matrix (A) and outcomes (B) of performance test for classification of four labels with non-segmented images after training with all segmented images.
(A) Confusion MatrixPrediction
AbsenceShallowProperDeepSum
LabelsAbsence580544593
Shallow3272032
Proper310252290717
Deep2034247
Sum588134531136
(B) OutcomesPrediction
AbsenceShallowProperDeepTotal (n)
True Positive, n58027522421171
False Positive, n8107994218
True Negative, n788125066312483949
False Negative, n1351955218
Accuracy0.9850.9190.8530.9290.922
Precision0.9860.2010.9830.3090.843
Sensitivity0.9780.8440.7280.8940.843
Specificity0.9900.9210.9870.9300.922
F1 score0.9820.3250.8370.4590.843
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Jung, H.C.; Kim, C.; Oh, J.; Kim, T.H.; Kim, B.; Lee, J.; Chung, J.H.; Byun, H.; Yoon, M.S.; Lee, D.K. Position Classification of the Endotracheal Tube with Automatic Segmentation of the Trachea and the Tube on Plain Chest Radiography Using Deep Convolutional Neural Network. J. Pers. Med. 2022, 12, 1363. https://doi.org/10.3390/jpm12091363

AMA Style

Jung HC, Kim C, Oh J, Kim TH, Kim B, Lee J, Chung JH, Byun H, Yoon MS, Lee DK. Position Classification of the Endotracheal Tube with Automatic Segmentation of the Trachea and the Tube on Plain Chest Radiography Using Deep Convolutional Neural Network. Journal of Personalized Medicine. 2022; 12(9):1363. https://doi.org/10.3390/jpm12091363

Chicago/Turabian Style

Jung, Heui Chul, Changjin Kim, Jaehoon Oh, Tae Hyun Kim, Beomgyu Kim, Juncheol Lee, Jae Ho Chung, Hayoung Byun, Myeong Seong Yoon, and Dong Keon Lee. 2022. "Position Classification of the Endotracheal Tube with Automatic Segmentation of the Trachea and the Tube on Plain Chest Radiography Using Deep Convolutional Neural Network" Journal of Personalized Medicine 12, no. 9: 1363. https://doi.org/10.3390/jpm12091363

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop