Next Article in Journal
A Novel Three-Point Localization Method for Bladder Volume Estimation
Previous Article in Journal
Spatial and Temporal Variation of Urban Heat Islands in French Guiana
Previous Article in Special Issue
Metric Learning in Histopathological Image Classification: Opening the Black Box
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning-Based Multi-Class Segmentation of the Paranasal Sinuses of Sinusitis Patients Based on Computed Tomographic Images

1
Department of Computer Science, Wesleyan University, Middletown, CT 06459, USA
2
Medical Devices R&D Center, Gachon University Gil Medical Center, Incheon 21565, Republic of Korea
3
Department of Health Sciences and Technology, Gachon Advanced Institute for Health & Sciences and Technology (GAIHST), Gachon University, Incheon 21565, Republic of Korea
4
Department of Otolaryngology-Head and Neck Surgery, Gachon University Gil Hospital, Incheon 21565, Republic of Korea
5
Department of Biomedical Engineering, College of IT Convergence, Gachon University, Seongnam-si 13120, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(6), 1933; https://doi.org/10.3390/s24061933
Submission received: 19 January 2024 / Revised: 7 March 2024 / Accepted: 15 March 2024 / Published: 18 March 2024

Abstract

:
Accurate paranasal sinus segmentation is essential for reducing surgical complications through surgical guidance systems. This study introduces a multiclass Convolutional Neural Network (CNN) segmentation model by comparing four 3D U-Net variations—normal, residual, dense, and residual-dense. Data normalization and training were conducted on a 40-patient test set (20 normal, 20 abnormal) using 5-fold cross-validation. The normal 3D U-Net demonstrated superior performance with an F1 score of 84.29% on the normal test set and 79.32% on the abnormal set, exhibiting higher true positive rates for the sphenoid and maxillary sinus in both sets. Despite effective segmentation in clear sinuses, limitations were observed in mucosal inflammation. Nevertheless, the algorithm’s enhanced segmentation of abnormal sinuses suggests potential clinical applications, with ongoing refinements expected for broader utility.

1. Introduction

In 1994, around 200,000 sinus surgeries were conducted in the United States [1]. By 1996, 12 percent of Americans under the age of 45 reported symptoms indicative of chronic sinusitis [2]. This widespread condition imposes a substantial societal burden, manifesting in frequent office visits, absenteeism from work, and missed school days [1]. When medicinal treatments fail to alleviate the condition, patients are often referred for sinus surgery. Many physicians refer to Computed Tomography (CT) scans when evaluating patients referred for sinus surgery [1,3]. Radiologists report anatomic variants, that can affect operative techniques, and critical variants, that can complicate surgery [4]. Identification of these anatomical variants affords the opportunity to avoid surgical complications [5]. Segmentation data can be used for the diagnosis, surgical planning, or workspace definition of robot-assisted systems. However, manual and semiautomatic segmentation of the paranasal sinuses has been evaluated as impractical in clinical settings because of the amount of time required for both systems [6,7]. The application of machine learning in this process warrants attention due to its potential to substantially mitigate the time and labor costs associated with manual segmentation. Ultimately, this holds promise for making the segmentation process feasible and practical in clinical settings.
Artificial intelligence is gaining popularity in the medical imaging field for developing models that produce human-interpretable results [8,9,10]. Because of the clustered arrangement of regions, including the frontal, ethmoid, and sphenoid sinuses, developing models that can produce practical results for the paranasal sinuses is an ongoing challenge. Two published studies focused on processing cone-beam computed tomography images to achieve segmentation of the maxillary sinus. In 2022, Choi et al. [11] trained a U-Net model to segment maxillary sinuses. The segmented results were refined using post-processing techniques to isolate and remove disconnected false positives. The trained model made predictions with a Dice similarity coefficient (DSC) value of 0.90 ± 0.19 before post-processing and 0.90 ± 0.19 after post-processing. Morgan et al. [12] trained two U-Net models to segment the maxillary sinus. The first model suggested crop boxes in the original image of the maxillary sinus, which were used to train the second part of the model to produce high-resolution segmentation results. The final segmentation results achieved a DSC score of 0.98 for the first model and 0.99 for the second model. Kuo et al. [13] proposed a 6-class segmentation model that segmented four different areas of the paranasal sinuses, treating the ethmoid sinus as two different areas: the anterior and posterior ethmoid sinus. A secondary model was trained to generate pseudo-labels on the unlabeled datasets. The model used in this study was an adaptation of the U-Net model [14] with the addition of depth-wise separable convolution, squeeze-and-excitation networks, and residual connections. The model was able to make predictions with a DSC value of 0.90. The approaches proposed by Choi et al. [11] and Morgan et al. [12] exhibited performance adequate for clinical applications. However, the aim of both studies was limited to the binary segmentation of the maxillary sinus.
We proposed a 5-class segmentation model for the four regions of the paranasal sinus: frontal sinus, ethmoid sinus, sphenoid sinus, and maxillary sinus. Training and validation were conducted on clinical-level CT scans sourced from patients exhibiting high degrees of genetic and biological variations. The objective was to develop a model capable of generating clinical data with sufficient accuracy to be practically applicable in clinical settings.

2. Materials and Methods

This study was approved by the Institutional Review Board (IRB) of Gachon University Gil Medical Center (GAIRB2022-182) and was conducted in accordance with the relevant guidelines and ethical regulations.
A total of 39,605 paranasal CT scans were collected from 201 patients with varying degrees of chronic sinusitis, including 3821 images from 20 patients without sinusitis. A total of 40 datasets were randomly selected as the hold-out test set, with 20 datasets originating from the patient group without sinusitis. These subsets were then labeled as “normal” and “abnormal” to reflect the respective patient group characteristics. Training was performed on the remaining 161 datasets with 5-fold cross validation, where 128 datasets were used for training and 33 for validation. In summary, the dataset was divided into sets comprising 128 patients for training, 33 patients for validation, and 40 patients for testing. Demographic information of the participating patients is summarized in Table 1.
Data collection and storage were performed using Excel (version 16.83, Microsoft, Redmond, WA, USA) and statistical analyses were performed using MedCalc (version 22, MedCalc Software Ltd., Ostend, Belgium). Training was performed on an Ubuntu server (version 20.04.6 LTS) with four Nvidia A100 80Gb GPUs (NVIDIA, Santa Clara, CA, USA), an AMD EPYC 7452 32-Core Processor (AMD, Santa Clara, CA, USA), and 1,031,900 Mb of RAM. The following libraries were used for training: Python (version 3.7), TensorFlow (version 2.6.0), and Keras (version 2.6.0).
Using the collected sinus data, we meticulously curated a ground truth dataset by labeling the sinus region for each patient. The oversight and guidance of two experienced otorhinolaryngologists was integral to this process, ensuring the utmost quality and accuracy of the dataset. The final ground truth data were congregated through a consensus between the two physicians. The ground truth was labeled along the axial, sagittal, and coronal axes, as visually depicted in Figure 1. The volumetric reconstruction (Figure 1d) presents the data in its authentic form, providing insight into how it is inputted into the deep learning model. The axial view (Figure 1a) shows the maxillary and sphenoid sinuses beneath the ethmoid sinuses. The sagittal view (Figure 1b) shows the left maxillary sinus and part of the sphenoid sinus. The coronal view (Figure 1c) shows the frontal and maxillary sinuses surrounding the ocular area.
To facilitate the extraction of features within the CT scans, the datasets underwent several enhancements (Figure 2A), including window setting adjustments, isotropic voxel reconstruction, contrast-limited adaptive histogram equalization (CLAHE), and region of interest (ROI) cropping. The preprocessed images were used to train the segmentation model (Figure 2B) to produce segmentation results (Figure 2C). The overall training process is presented in Figure 2. A bone window with a width of 2,000 and a level of 0 was set and converted into 8-bit encoding. This setting has been established as the imaging technique of choice for examining patients before functional endoscopic sinus surgery [15,16].
Depending on the acquisition process, CT images can have varying slice thickness and pixel spacing within the protocol range [17,18]. The acquired images exhibited a consistent 1 mm slice thickness but varying pixel spacings, resulting in the disproportionate volumetric ratio of planar CT images. To eliminate unwanted ratio variations among the dataset, an isotropic voxel reconstruction algorithm was applied across the dataset to equalize the slice thickness to pixel spacing ratio. The ratio of slice thickness to pixel spacing was calculated to downsample the images accordingly using cubic spline interpolation [19] such that the volumetry of the resized images matched real proportions.
Adaptive histogram algorithms are commonly used in medical imaging to create images with equal intensity levels, thereby generating an image with an increased dynamic range, leading to an increase in contrast [20,21]. CLAHE [22,23] was employed in this study to restrict amplification and prevent overamplification of noise in areas with relatively homogeneous contrast.
To equalize the image dimensions for training, a cropping algorithm was used to crop images based on the region of interest. To guarantee the comprehensive inclusion of the region of interest, specific dimensions were set, with a target depth of 192, a height of 128, and a width of 128. The dimensions were chosen based on an analysis of the ground truth data in the entire dataset. The algorithm used in the analysis calculated the 3-dimensional coordinates of the edges for the largest ground truth data. As the voxel reconstruction algorithm resized the CT scans in accordance with the actual proportions of the paranasal sinuses, a greater amount of ground truth data became available along the depth axis.
The U-Net architecture is commonly used for medical image segmentation models because of its reliable performance on medical images [24,25,26]. Furthermore, its utilization of depth-wise 3D convolution operations allows for the simultaneous extraction of features along the 3 axes: axial, sagittal, and coronal. Three variants of the 3D U-Net architecture, each deeper than the last, were trained and compared: 3D U-Net with residual connections [27], 3D U-Net with dense blocks [28], and 3D U-Net with dense blocks and residual connections [29]. The 3D U-Net architecture, which served as the basis for constructing our model, is presented in Figure 3.
The 3D U-Net used in this study comprised 18 convolutional layers with 5,644,981 trainable parameters. The residual 3D U-Net comprised 63 convolutional layers and 2,350,989 trainable parameters. The dense 3D U-Net comprised 28 convolutional layers and 10,960,437 trainable parameters. The residual dense 3D U-Net comprised 34 convolutional layers and 47,078,117 trainable parameters. A summary of the parameter and layer counts for each model is provided in Table 2, along with the kernel-wise feature map details summarized in Table 3. All models were trained on the same hyperparameters. The Adam [30] optimizer was used with an initial learning rate of 0.0001. Categorical cross-entropy loss was used to monitor validation loss, and accuracy was used as the evaluation metric. Learning rates on plateaus, early stoppers, and model checkpoints were used to prevent issues such as overfitting and plateauing. The tolerance for learning rate reduction was configured to 20 epochs, while the early stopper tolerance was set at 30 epochs.

3. Results

Each model was tested against the hold-out test set to generate segmentation results. The segmentation results were evaluated using the following five performance metrics: intersection over union (IoU), accuracy, recall, precision, and F1 score. The results are expressed as the mean ± 95% confidence interval, with statistical significance set at p < 0.05.
The segmentation results from the normal test set were evaluated using the performance metrics and summarized in Table 4. Overall, the models were able to make predictions with an F1 score in the range of 0.843–0.785, of which the 3D U-Net model achieved the highest F1 score with a value of 0.843. Conversely, the residual 3D U-Net model recorded the lowest F1 score, standing at 0.785.
The segmentation results from the abnormal test set are summarized in Table 5. In the abnormal test set, the segmentation results were evaluated to record a lower overall F1 score in the range of 0.793–0.740. The 3D U-Net model made predictions with the highest F1 score of 0.793, whereas the predictions made by the residual-dense 3D U-Net model recorded the lowest F1 score of 0.741.
A comparative plot of IoU values across the models in the normal and abnormal test set is presented in Figure 4. The average IoU difference across the models was 0.082 ± 0.034 (mean ± 95% confidence interval). Paired t-tests of the IoU across the models showed statistically insignificant differences in IoU values between the four models (p < 0.05). The average F1 score difference, encompassing both test sets, between the 3D U-Net and the other three models were as follows: 0.067 ± 0.016 for the residual model, 0.069 ± 0.028 for the dense model, and 0.082 ± 0.037 for the residual-dense 3D U-Net. Paired t-tests of the F1 scores between the models showed statistically insignificant F1 score variation across the models (p < 0.05). The average differences in F1 scores between the two test sets (normal and abnormal) were as follows: 0.170 ± 0.067 for the 3D U-Net, 0.188 ± 0.064 for the residual 3D U-Net, 0.206 ± 0.072 for the dense 3D U-Net, and 0.257 ± 0.099 for the residual-dense 3D U-Net. Statistical analysis using paired t-tests showed a statistically significant difference in the F1 scores between the normal and abnormal test sets (p > 0.05).
Visual overviews of the segmentation results for the normal and abnormal test sets are shown in Figure 5 and Figure 6, respectively. The figures show the segmentation results for the ethmoid sinus, maxillary sinus, and sphenoid sinus; each area is color-coded for better visual representation. The images were chosen randomly from the fold with the best mIoU score. Each row represents predictions from different models. From left to right, the three columns represent the ground truth, prediction, and overlay comparison of the ground truth and prediction.
Normalized true positive (TP) distribution per class as a heatmap for the 3D U-Net is shown in Figure 6. For the normal dataset, the sphenoid sinus showed the highest TP rate of 0.95, whereas the ethmoid sinus showed the lowest at 0.82. For the abnormal dataset, the sphenoid sinus reported the highest TP rate at 0.88, and the lowest for the frontal sinus at 0.67.

4. Discussion

In this study, a 3D segmentation model for the four areas of the paranasal sinus based on CT images was developed and evaluated. Four models based on the 3D U-Net were trained and evaluated on a hold-out test set of 40 datasets, comprising 20 datasets from patients without sinusitis and 20 datasets from patients with sinusitis. Prediction results were further validated using 5-fold cross validation. In the normal test set, the models showed performances in the range of 0.843–0.785 with an average F1 score of 0.805. In the abnormal test set, the models performed in the range of 0.793–0.740 with an average F1 score of 0.755. In both test sets, the base 3D U-Net was able to make predictions with the highest F1 score of 0.843, and 0.793, respectively, in the normal test set and the abnormal test set. Statistical analysis of performance metrics was performed across the four models between normal and abnormal test sets with statistical significance set at p = 0.05. Performance metrics across the models exhibited statistically insignificant variations. However, mucosal inflammation had a greater impact on the performance metrics across the models.
The method proposed by Choi et al. [11] reported an F1 score of 0.972 in normal sinuses and 0.912 in sinuses with mucosal inflammation. Morgan et al. [12] reported an F1 score of 0.984 and 0.996, respectively, for normal and abnormal sinuses. Note that these studies were limited to binary segmentation of the maxillary sinus, manifesting in the higher F1 score. The study by Kuo et al. [13] trained multiple models with the aim of multi-class segmentation of the sinus, in which the U-Net model reported an average F1 score of 0.896. This is within 6.2% of the highest performing model in our study, the base 3D U-Net.
We performed a thorough analysis of prediction accuracy for the 3D U-Net model across the four main sinus regions, focusing on true positive rates. The outcomes underscored notable limitations in the precise prediction of the frontal and ethmoid sinus regions. The abnormal test set showed lower prediction metrics, overall, in comparison to the normal test set. The frontal and ethmoid sinuses showed particularly lower TP rates in the abnormal test set, at 0.67 and 0.75, respectively. The frontal and ethmoid sinuses are anatomically adjacent structures, and both have smaller volumes than the sphenoid and maxillary sinuses [31]. In sinus cavities with mucosal inflammation, the cavities of the ethmoid and frontal sinuses had much less pronounced features compared to other areas of the paranasal sinuses. This limitation is evident in Figure 7 of the right ethmoid sinus, where the contrast between the sinus bone and cavity appears less pronounced compared to the left ethmoid sinus.
Despite the substantial size of the dataset collected for this study, the clinical nature of the CT scans led to an uneven distribution of data between patients with sinusitis and those without the condition. Moreover, training data was obtained solely from a single institution, suggesting the possibility that the trained models could exhibit limited generalization capabilities on external datasets. A comprehensive follow-up study should encompass a well-balanced dataset, including an equal distribution of data from patients with sinusitis and those without the condition. It would be advantageous to source this data from multiple institutes to enable internal and external validations.
Accurate segmentation of the paranasal sinuses is crucial for the preoperative evaluation of patients undergoing sinus surgery. To this end, this study aimed to evaluate the segmentation efficacy in patients with mucosal inflammation. While limitations do exist in the segmentation of paranasal sinuses with mucosal inflammation, the proposed method exhibited promising results. With minor refinements, our segmentation model has the potential to enhance surgical accuracy when integrated into guidance systems. Such integration can aid surgeons in avoiding healthy mucosal tissue, thereby reducing the risk of complications.

Author Contributions

J.W., J.L., Y.J.K., S.T.K. and K.G.K. contributed to the conception and design of this study. Material preparation, data collection, and analyses were performed by J.L. The first draft of this manuscript was written by J.W., and all authors commented on the previous versions of this manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Technology Innovation Program (K_G012001187801, “Development of Diagnostic Medical Devices with Artificial intelligence Based Image Analysis Technology”) funded By the Ministry of Trade, Industry & Energy (MOTIE, Korea), and by Gachon University (GCU-202205980001).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, approved by the Institutional Review Board (IRB) of Gachon University Gil Medical Center (GAIRB2022-182), and was conducted in accordance with relevant guidelines and ethical regulations.

Informed Consent Statement

As this study is retrospective in nature, the Institutional Review Board (IRB) of Gachon University Gil Medical Center has granted a waiver for the requirement of patient informed consent. The retrospective design involves the analysis of existing data and poses minimal risk to the participants. This waiver ensures the ethical conduct of the study while respecting the rights and privacy of the individuals involved.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Hamilos, D.L. Chronic Sinusitis. J. Allergy Clin. Immunol. 2000, 106, 213–227. [Google Scholar] [CrossRef] [PubMed]
  2. Kaliner, M.; Osguthorpe, J.; Fireman, P.; Anon, J.; Georgitis, J.; Davis, M.; Naclerio, R.; Kennedy, D. Sinusitis: Bench to BedsideCurrent Findings, Future Directions. J. Allergy Clin. Immunol. 1997, 99, S829–S847. [Google Scholar] [CrossRef] [PubMed]
  3. Bhattacharyya, N. Clinical and Symptom Criteria for the Accurate Diagnosis of Chronic Rhinosinusitis. Laryngoscope 2006, 116, 1–22. [Google Scholar] [CrossRef] [PubMed]
  4. Hoang, J.K.; Eastwood, J.D.; Tebbit, C.L.; Glastonbury, C.M. Multiplanar Sinus CT: A Systematic Approach to Imaging Before Functional Endoscopic Sinus Surgery. Am. J. Roentgenol. 2010, 194, W527–W536. [Google Scholar] [CrossRef] [PubMed]
  5. O’Brien, W.T.; Hamelin, S.; Weitzel, E.K. The Preoperative Sinus CT: Avoiding a “CLOSE” Call with Surgical Complications. Radiology 2016, 281, 10–21. [Google Scholar] [CrossRef]
  6. Tingelhoff, K.; Moral, A.I.; Kunkel, M.E.; Rilk, M.; Wagner, I.; Eichhorn, K.W.G.; Wahl, F.M.; Bootz, F. Comparison between Manual and Semi-Automatic Segmentation of Nasal Cavity and Paranasal Sinuses from CT Images. In Proceedings of the 2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, Lyon, France, 22–26 August 2007; pp. 5505–5508. [Google Scholar]
  7. Pirner, S.; Tingelhoff, K.; Wagner, I.; Westphal, R.; Rilk, M.; Wahl, F.M.; Bootz, F.; Eichhorn, K.W.G. CT-Based Manual Segmentation and Evaluation of Paranasal Sinuses. Eur. Arch. Oto-Rhino-Laryngol. 2009, 266, 507–518. [Google Scholar] [CrossRef]
  8. Varoquaux, G.; Cheplygina, V. Machine Learning for Medical Imaging: Methodological Failures and Recommendations for the Future. NPJ Digit. Med. 2022, 5, 48. [Google Scholar] [CrossRef]
  9. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A Survey on Deep Learning in Medical Image Analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef]
  10. Cheplygina, V.; de Bruijne, M.; Pluim, J.P.W. Not-so-Supervised: A Survey of Semi-Supervised, Multi-Instance, and Transfer Learning in Medical Image Analysis. Med. Image Anal. 2019, 54, 280–296. [Google Scholar] [CrossRef] [PubMed]
  11. Choi, H.; Jeon, K.J.; Kim, Y.H.; Ha, E.-G.; Lee, C.; Han, S.-S. Deep Learning-Based Fully Automatic Segmentation of the Maxillary Sinus on Cone-Beam Computed Tomographic Images. Sci. Rep. 2022, 12, 14009. [Google Scholar] [CrossRef] [PubMed]
  12. Morgan, N.; Van Gerven, A.; Smolders, A.; de Faria Vasconcelos, K.; Willems, H.; Jacobs, R. Convolutional Neural Network for Automatic Maxillary Sinus Segmentation on Cone-Beam Computed Tomographic Images. Sci. Rep. 2022, 12, 7523. [Google Scholar] [CrossRef]
  13. Kuo, C.-F.J.; Liu, S.-C. Fully Automatic Segmentation, Identification and Preoperative Planning for Nasal Surgery of Sinuses Using Semi-Supervised Learning and Volumetric Reconstruction. Mathematics 2022, 10, 1189. [Google Scholar] [CrossRef]
  14. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2016: 19th International Conference, Athens, Greece, October 17–21, 2016, Proceedings, Part II 19; Springer International Publishing: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar]
  15. Som, P.M. CT of the Paranasal Sinuses. Neuroradiology 1985, 27, 189–201. [Google Scholar] [CrossRef]
  16. Melhem, E.R.; Oliverio, P.J.; Benson, M.L.; Leopold, D.A.; Zinreich, S.J. Optimal CT Evaluation for Functional Endoscopic Sinus Surgery. AJNR Am. J. Neuroradiol. 1996, 17, 181–188. [Google Scholar]
  17. Huang, K.; Rhee, D.J.; Ger, R.; Layman, R.; Yang, J.; Cardenas, C.E.; Court, L.E. Impact of Slice Thickness, Pixel Size, and CT Dose on the Performance of Automatic Contouring Algorithms. J. Appl. Clin. Med. Phys. 2021, 22, 168–174. [Google Scholar] [CrossRef]
  18. Cantatore, A.; Müller, P. Introduction to Computed Tomography; DTU Mechanical Engineering: Lyngby, Denmark, 2011. [Google Scholar]
  19. Dyer, S.A.; Dyer, J.S. Cubic-Spline Interpolation. 1. IEEE Instrum. Meas. Mag. 2001, 4, 44–46. [Google Scholar] [CrossRef]
  20. Salem, N.; Malik, H.; Shams, A. Medical Image Enhancement Based on Histogram Algorithms. Procedia Comput. Sci. 2019, 163, 300–311. [Google Scholar] [CrossRef]
  21. Rahman, T.; Khandakar, A.; Qiblawey, Y.; Tahir, A.; Kiranyaz, S.; Abul Kashem, S.B.; Islam, M.T.; Al Maadeed, S.; Zughaier, S.M.; Khan, M.S.; et al. Exploring the Effect of Image Enhancement Techniques on COVID-19 Detection Using Chest X-Ray Images. Comput. Biol. Med. 2021, 132, 104319. [Google Scholar] [CrossRef] [PubMed]
  22. Pizer, S.M.; Amburn, E.P.; Austin, J.D.; Cromartie, R.; Geselowitz, A.; Greer, T.; ter Haar Romeny, B.; Zimmerman, J.B.; Zuiderveld, K. Adaptive Histogram Equalization and Its Variations. Comput. Vis. Graph. Image Process. 1987, 39, 355–368. [Google Scholar] [CrossRef]
  23. Sørensen, T.; Sørensen, T.; Biering-Sørensen, T.; Sørensen, T.; Sorensen, J.T. A Method of Establishing Group of Equal Amplitude in Plant Sociobiology Based on Similarity of Species Content and Its Application to Analyses of the Vegetation on Danish Commons. Biol. Skr. 1948, 5, 1–34. [Google Scholar]
  24. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015, Proceedings, Part III 18; Springer International Publishing: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  25. Azad, R.; Aghdam, E.K.; Rauland, A.; Jia, Y.; Avval, A.H.; Bozorgpour, A.; Karimijafarbigloo, S.; Cohen, J.P.; Adeli, E.; Merhof, D. Medical Image Segmentation Review: The Success of U-Net. arXiv 2022, arXiv:2211.14830. [Google Scholar]
  26. Siddique, N.; Paheding, S.; Elkin, C.P.; Devabhaktuni, V. U-Net and Its Variants for Medical Image Segmentation: A Review of Theory and Applications. IEEE Access 2021, 9, 82031–82057. [Google Scholar] [CrossRef]
  27. Lee, K.; Zung, J.; Li, P.H.; Jain, V.; Seung, H.S. Superhuman Accuracy on the SNEMI3D Connectomics Challenge. arXiv 2017, arXiv:1706.00120. [Google Scholar]
  28. Kolařík, M.; Burget, R.; Uher, V.; Dutta, M.K. 3D Dense-U-Net for MRI Brain Tissue Segmentation. In Proceedings of the 2018 41st International Conference on Telecommunications and Signal Processing (TSP), Athens, Greece, 4–6 July 2018; pp. 1–4. [Google Scholar]
  29. Sarica, B.; Seker, D.Z.; Bayram, B. A Dense Residual U-Net for Multiple Sclerosis Lesions Segmentation from Multi-Sequence 3D MR Images. Int. J. Med. Inform. 2023, 170, 104965. [Google Scholar] [CrossRef] [PubMed]
  30. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  31. Mynatt, R.G.; Sindwani, R. Surgical Anatomy of the Paranasal Sinuses. In Rhinology and Facial Plastic Surgery; Springer: Berlin/Heidelberg, Germany, 2009; pp. 13–33. [Google Scholar]
Figure 1. CT image of the paranasal sinuses with ground truth data overlayed. (a) Axial view, (b) sagittal view, (c) coronal view, and (d) volumetric reconstruction.
Figure 1. CT image of the paranasal sinuses with ground truth data overlayed. (a) Axial view, (b) sagittal view, (c) coronal view, and (d) volumetric reconstruction.
Sensors 24 01933 g001
Figure 2. Flowchart of the multiclass sinus segmentation training process.
Figure 2. Flowchart of the multiclass sinus segmentation training process.
Sensors 24 01933 g002
Figure 3. Architecture of the 3D U-Net.
Figure 3. Architecture of the 3D U-Net.
Sensors 24 01933 g003
Figure 4. mIoU comparison of each model in the normal and abnormal test set (gray: normal dataset results, black: abnormal dataset results, ****: statistical significance (p < 0.05)).
Figure 4. mIoU comparison of each model in the normal and abnormal test set (gray: normal dataset results, black: abnormal dataset results, ****: statistical significance (p < 0.05)).
Sensors 24 01933 g004
Figure 5. 3D U-Net segmentation comparison for the normal test case. Color legend: green, orange—maxillary sinus; blue, pink—ethmoid sinus; red, yellow—sphenoid sinus; (a) ground truth data, (b) model prediction, (c) overlay comparison of ground truth and prediction.
Figure 5. 3D U-Net segmentation comparison for the normal test case. Color legend: green, orange—maxillary sinus; blue, pink—ethmoid sinus; red, yellow—sphenoid sinus; (a) ground truth data, (b) model prediction, (c) overlay comparison of ground truth and prediction.
Sensors 24 01933 g005
Figure 6. 3D U-Net segmentation comparison for the abnormal test case. Color legend: green, orange—maxillary sinus; blue, pink—ethmoid sinus; red, yellow—sphenoid sinus; (a) ground truth data, (b) model prediction, (c) overlay comparison of ground truth and prediction.
Figure 6. 3D U-Net segmentation comparison for the abnormal test case. Color legend: green, orange—maxillary sinus; blue, pink—ethmoid sinus; red, yellow—sphenoid sinus; (a) ground truth data, (b) model prediction, (c) overlay comparison of ground truth and prediction.
Sensors 24 01933 g006
Figure 7. 3D U-Net prediction heatmap; (a) prediction results for the normal test set, (b) prediction results for the abnormal test set.
Figure 7. 3D U-Net prediction heatmap; (a) prediction results for the normal test set, (b) prediction results for the abnormal test set.
Sensors 24 01933 g007
Table 1. Patient distribution by age group and gender.
Table 1. Patient distribution by age group and gender.
Age GroupMaleGender RatioFemaleGender RatioTotalRatio by Age
10–20640.00%960.00%157.58%
20–301568.18%731.82%2211.11%
30–402184.00%416.00%2512.63%
40–501463.64%836.36%2211.11%
50–603462.96%2037.04%5427.27%
60–703271.11%1328.89%4522.73%
70–80872.73%327.27%115.56%
80–250.00%250.00%42.02%
Total13266.67%6633.33%198100%
Table 2. Parameter and layer count by model.
Table 2. Parameter and layer count by model.
3D U-NetResidualDenseResidual-Dense
CountParameterLayerParameterLayerParameterLayerParameterLayer
block 17376444561342,3521084,48017
block 241,536419,84013125,0889388,41617
block 3166,016478,59213499,00891,550,97617
block 4663,8084312,832131,993,34496,198,52817
block 52,654,72031,248,256122,657,664824,783,36016
block 61,589,6324517,632134,117,504910,622,20818
block 7397,5044129,792131,063,42492,656,89618
block 899,424432,64013266,4969664,89618
block 924,880482561373,8729138,62418
Output851165185143251
Total5,644,981362,352,46111710,838,8378247,092,709157
Table 3. Layer-by-layer kernel-wise details of each model. The 3D U-Net and dense 3D U-Net models share feature map details, while the residual 3D U-Net and residual-dense 3D U-Net models also share feature map details.
Table 3. Layer-by-layer kernel-wise details of each model. The 3D U-Net and dense 3D U-Net models share feature map details, while the residual 3D U-Net and residual-dense 3D U-Net models also share feature map details.
3D U-Net/Dense 3D U-NetResidual 3D U-Net/Residual-Dense 3D U-Net
NameFeat Maps (Input)Feat Maps (Output)Feat Maps (Input)Feat Maps (Output)
Encoding pathconv3d_block_1192 × 128 × 128 × 1192 × 128 × 128 × 16192 × 128 × 128 × 1192 × 128 × 128 × 32
maxpool3d_1192 × 128 × 128 × 1696 × 64 × 64 × 16192 × 128 × 128 × 3296 × 64 × 64 × 32
conv3d_block_296 × 64 × 64 × 1696 × 64 × 64 × 3296 × 64 × 64 × 3296 × 64 × 64 × 64
maxpool3d_296 × 64 × 64 × 3248 × 32 × 32 × 3296 × 64 × 64 × 6448 × 32 × 32 × 64
conv3d_block_348 × 32 × 32 × 3248 × 32 × 32 × 6448 × 32 × 32 × 6448 × 32 × 32 × 128
maxpool3d_348 × 32 × 32 × 6424 × 16 × 16 × 6448 × 32 × 32 × 12824 × 16 × 16 × 128
conv3d_block_424 × 16 × 16 × 6424 × 16 × 16 × 12824 × 16 × 16 × 128 24 × 16 × 16 × 256
maxpool3d_424 × 16 × 16 × 12812 × 8 × 8 × 12824 × 16 × 16 × 25612 × 8 × 8 × 256
Bridge 12 × 8 × 8 × 12812 × 8 × 8 × 25612 × 8 × 8 × 25612 × 8 × 8 × 512
Decoding pathconv3d_trans_112 × 8 × 8 × 25624 × 16 × 16 × 12812 × 8 × 8 × 51224 × 16 × 16 × 256
conv3d_block_524 × 16 × 16 × 12824 × 16 × 16 × 12824 × 16 × 16 × 25624 × 16 × 16 × 256
conv3d_trans_224 × 16 × 16 × 12848 × 32 × 32 × 6424 × 16 × 16 × 25648 × 32 × 32 × 128
conv3d_block_648 × 32 × 32 × 6448 × 32 × 32 × 6448 × 32 × 32 × 12848 × 32 × 32 × 128
conv3d_trans_348 × 32 × 32 × 6496 × 64 × 64 × 3248 × 32 × 32 × 12896 × 64 × 64 × 64
conv3d_block_796 × 64 × 64 × 3296 × 64 × 64 × 3296 × 64 × 64 × 6496 × 64 × 64 × 64
conv3d_trans_496 × 64 × 64 × 32192 × 128 × 128 × 1696 × 64 × 64 × 64192 × 128 × 128 × 32
conv3d_block_8192 × 128 × 128 × 16192 × 128 × 128 × 5192 × 128 × 128 × 32192 × 128 × 128 × 5
Table 4. Prediction results obtained on the normal test set, reported in performance metrics per model.
Table 4. Prediction results obtained on the normal test set, reported in performance metrics per model.
MetricsBaseResidualDenseResidual-Dense
F1 score0.843 ± 0.6990.785 ± 0.0660.790 ± 0.0730.802 ± 0.093
Accuracy0.995 ± 0.0030.992 ± 0.0010.993 ± 0.0020.993 ± 0.003
Precision0.857 ± 0.0560.789 ± 0.0590.801 ± 0.0600.822 ± 0.073
Recall0.854 ± 0.0640.821 ± 0.0600.822 ± 0.0680.836 ± 0.078
Mean IoU0.787 ± 0.0710.703 ± 0.0670.714 ± 0.0740.742 ± 0.092
Table 5. Prediction results obtained on the abnormal test set, reported in performance metrics per model.
Table 5. Prediction results obtained on the abnormal test set, reported in performance metrics per model.
MetricsBaseResidualDenseResidual-Dense
F1 score0.793 ± 0.0630.741 ± 0.0690.747 ± 0.0740.740 ± 0.095
Accuracy0.994 ± 0.0020.991 ± 0.0020.992 ± 0.0020.991 ± 0.003
Precision0.839 ± 0.0570.779 ± 0.0670.785 ± 0.0710.793 ± 0.089
Recall0.785 ± 0.0670.755 ± 0.0760.756 ± 0.0680.745 ± 0.092
Mean IoU0.717 ± 0.0610.653 ± 0.0630.666 ± 0.0740.670 ± 0.089
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Whangbo, J.; Lee, J.; Kim, Y.J.; Kim, S.T.; Kim, K.G. Deep Learning-Based Multi-Class Segmentation of the Paranasal Sinuses of Sinusitis Patients Based on Computed Tomographic Images. Sensors 2024, 24, 1933. https://doi.org/10.3390/s24061933

AMA Style

Whangbo J, Lee J, Kim YJ, Kim ST, Kim KG. Deep Learning-Based Multi-Class Segmentation of the Paranasal Sinuses of Sinusitis Patients Based on Computed Tomographic Images. Sensors. 2024; 24(6):1933. https://doi.org/10.3390/s24061933

Chicago/Turabian Style

Whangbo, Jongwook, Juhui Lee, Young Jae Kim, Seon Tae Kim, and Kwang Gi Kim. 2024. "Deep Learning-Based Multi-Class Segmentation of the Paranasal Sinuses of Sinusitis Patients Based on Computed Tomographic Images" Sensors 24, no. 6: 1933. https://doi.org/10.3390/s24061933

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop