Next Article in Journal
Adaptive High-Precision 3D Reconstruction of Highly Reflective Mechanical Parts Based on Optimization of Exposure Time and Projection Intensity
Previous Article in Journal
Towards the Performance Characterization of a Robotic Multimodal Diagnostic Imaging System
Previous Article in Special Issue
Evolutionary-Driven Convolutional Deep Belief Network for the Classification of Macular Edema in Retinal Fundus Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

3D-NASE: A Novel 3D CT Nasal Attention-Based Segmentation Ensemble

by
Alessandro Pani
*,
Luca Zedda
*,
Davide Antonio Mura
,
Andrea Loddo
* and
Cecilia Di Ruberto
Department of Mathematics and Computer Science, University of Cagliari, Via Ospedale 72, 09124 Cagliari, Italy
*
Authors to whom correspondence should be addressed.
J. Imaging 2025, 11(5), 148; https://doi.org/10.3390/jimaging11050148
Submission received: 7 March 2025 / Revised: 12 April 2025 / Accepted: 2 May 2025 / Published: 7 May 2025

Abstract

:
Accurate segmentation of the nasal cavity and paranasal sinuses in CT scans is crucial for disease assessment, treatment planning, and surgical navigation. It also facilitates the advanced computational modeling of airflow dynamics and enhances endoscopic surgery preparation. This work presents a novel ensemble framework for 3D nasal CT segmentation that synergistically combines CNN-based and transformer-based architectures, 3D-NASE. By integrating 3D U-Net, UNETR, Swin UNETR, SegResNet, DAF3D, and V-Net with majority and soft voting strategies, our approach leverages both local details and global context to improve segmentation accuracy and robustness. Results on the NasalSeg dataset demonstrate that the proposed ensemble method surpasses previous state-of-the-art results by achieving a 35.88 % improvement in the DICE score and reducing the standard deviation by 4.53 % . These promising results highlight the potential of our method to advance clinical workflows in diagnosis, treatment planning, and surgical navigation while also promoting further research into computationally efficient and highly accurate segmentation techniques.

1. Introduction

An accurate understanding of nasal cavity and paranasal sinus morphology is vital for clinical applications, disease detection, treatment planning, endonasal surgery simulations, or surgical navigation. For instance, the precise delineation of regions such as the nasal cavity and paranasal sinuses in computed tomography (CT) scans is vital for assessing sinus-related diseases [1], planning endoscopic surgeries [2], and enabling advanced computational models of airflow dynamics [3,4].
Segmentation data are key in diagnosis, surgical preparation, and robotic system navigation. Volume and surface area quantification help assess nasal health and the impact of inhaled substances [5]. Also, detailed anatomical knowledge supports endoscopic sinus surgery, guiding structural reconstruction and outcome evaluation. Although manual segmentation has long been considered the gold standard in clinical practice, it is inherently labor-intensive and prone to both inter- and intra-observer variability. These limitations have fueled the search for robust, automated segmentation approaches that can deliver consistent and reproducible results. In this context, a precise, automated segmentation of anatomical structures in 3D medical imaging becomes critical for effective diagnosis, treatment planning, and surgical navigation [5,6].
Furthermore, the recent surge in radiomics research has underscored the importance of high-quality segmentation as a foundation for extracting reliable quantitative imaging features [7]. Radiomics seeks to convert medical images into high-dimensional data by extracting features that capture subtle tissue heterogeneity and underlying pathophysiological processes. The spectrum of features ranges from classical handcrafted descriptors, such as shape, first-order intensity statistics, and texture measures, to sophisticated representations learned via deep neural networks, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), and transformer-based architectures. Such diversity in feature representation plays a crucial role in enhancing diagnostic accuracy, prognostication, and the prediction of treatment responses.
Moreover, the introduction of novel datasets, such as the NasalSeg [8] dataset, has begun to address the scarcity of large-scale, annotated imaging data necessary for training and benchmarking advanced segmentation models. This work proposes an ensemble-based segmentation framework that synergistically combines multiple 3D architectures, 3D-NASE. Our approach integrates CNN-based models with transformer-based networks that exploit self-attention mechanisms. By fusing these complementary architectures, our method aims to enhance segmentation robustness and improve generalization across anatomical variations and heterogeneous imaging conditions. It aims not only to refine segmentation accuracy but also to provide a reliable basis for subsequent radiomic analysis. In summary, the research question we aim to address in this work is how can an ensemble-based segmentation framework that integrates both CNN-based and transformer-based architectures be designed to enhance the accuracy, robustness, and reproducibility of 3D nasal CT segmentation for improved clinical diagnosis and treatment planning?
The remainder of the paper is organized as follows. In Section 2, we review the state-of-the-art radiomic segmentation by discussing both 2D and 3D approaches and the emerging trends and challenges in the field. Section 3 details the materials and methods used in our study, including the architectures of U-Net, 3D U-Net, UNETR, and Swin UNETR, along with the description of the NasalSeg dataset and the evaluation metrics. In Section 4, we introduce 3D-NASE and outline the experimental setup. Section 5 presents the quantitative experimental results, while Section 6 offers a qualitative analysis of the segmentation outputs. We discuss the limitations of our approach in Section 7 and conclude the paper with final remarks and directions for future work in Section 8.

2. Related Work

Over the past decade, significant progress has been made in radiomic segmentation through contributions from traditional image processing, machine learning, and deep learning methodologies. This section provides a comprehensive review of the state-of-the-art segmentation techniques in radiomics. The discussion is organized into four subsections, with a focus on the 2D (Section 2.1) and 3D (Section 2.2) segmentation approaches, along with a discussion of their distinctions and complementarities (Section 2.3), and future developments (Section 2.4).

2.1. Two-Dimensional Radiomic Segmentation

Two-dimensional segmentation techniques have been extensively explored due to their simplicity and relatively low computational burden. Early methods predominantly relied on classical image processing techniques such as thresholding, region growing, edge detection, and active contour models [9,10]. These methods proved effective when target structures exhibited high contrast and well-defined boundaries.
More recently, the emergence of deep learning models has shifted the common approaches toward more automatic and fast architectures capable of segmenting clinically relevant regions in different imaging techniques [11,12].
However, despite their computational efficiency and ease of implementation, 2D segmentation methods inherently lack the ability to fully exploit the volumetric context present in modern imaging modalities [13]. This shortcoming limits their utility in applications for which three-dimensional structural continuity [14] is critical, such as in tumor delineation or the assessment of complex anatomical regions.

2.2. Three-Dimensional Radiomic Segmentation

High-resolution volumetric imaging has spurred a paradigm shift toward 3D segmentation methods. Unlike 2D approaches, 3D segmentation takes advantage of the complete spatial context, resulting in a more comprehensive and accurate delineation of anatomical structures [13,15]. Volumetric methods are particularly indispensable in modalities like CT and magnetic resonance (MR) imaging, where the intricate spatial relationships between tissues must be preserved.
Deep learning has become the cornerstone of modern 3D segmentation [16,17]. The 3D U-Net architecture [18], for example, has emerged as a benchmark due to its encoder–decoder structure, which facilitates the capture of both local and global contexts across multiple scales. In addition to CNN-based models, transformer-based architectures have recently been introduced to model long-range dependencies across the volume [19]. With their self-attention mechanisms, these models are adept at identifying subtle variations and heterogeneities within the data, which are often missed when conventional convolutional filters are used [20].
Hybrid approaches have also been proposed, combining the rapid processing of 2D methods with the spatial accuracy of 3D segmentation [21,22]. Typically, these frameworks commence with a 2D segmentation step to quickly identify candidate regions, followed by a refined 3D segmentation to ensure volumetric consistency. This two-tiered strategy effectively balances the trade-offs between computational efficiency and segmentation accuracy [23], making it highly attractive for real-time clinical applications [24].

2.3. Comparative Analysis and Challenges

A comparative analysis between 2D and 3D segmentation techniques reveals a range of trade-offs. While 2D methods offer simplicity and speed [25,26], they are limited by their inability to capture inter-slice contextual information. On the other hand, 3D approaches, despite their higher computational demands, provide a more faithful representation of anatomical structures [27,28]. One of the primary challenges with 3D segmentation is the substantial computational overhead associated with processing volumetric data. High-resolution images require large amounts of memory and extended training times, often necessitating specialized hardware or cloud-based resources.
Another significant challenge is the variability in imaging protocols and acquisition parameters across different clinical centers. This variability can lead to discrepancies in image quality and intensity distributions, thereby affecting the generalizability of segmentation models [29]. To address these issues, recent research has focused on developing standardized preprocessing pipelines and data harmonization techniques essential for mitigating inter-scanner and inter-center variations.
Furthermore, the integration of segmentation outputs into radiomic pipelines poses additional challenges. High-quality segmentation is a prerequisite for reliable feature extraction, as errors in delineation can propagate and degrade the performance of downstream predictive models [30]. As such, the development of explainable and robust segmentation methods is a critical area of ongoing research.

2.4. Emerging Trends and Future Directions

The field of radiomic segmentation is rapidly evolving, driven by innovations that bridge the gap between traditional methods and modern deep learning techniques. One emerging trend is the integration of self-supervised learning and foundation models [31,32], which aim to leverage large-scale unannotated data to improve model robustness and reduce the reliance on extensively labeled datasets. These approaches are particularly promising for addressing the limitations posed by data scarcity in medical imaging [33,34].
Another exciting direction is the development of multi-modal segmentation frameworks. By fusing complementary information from different imaging modalities such as CT, MR, and positron emission tomography (PET), researchers are beginning to overcome the limitations inherent in single-modality analysis. Multi-modal fusion techniques can enhance the accuracy and reliability of segmentation, particularly in complex clinical scenarios where different modalities provide unique and complementary insights [35,36].

3. Materials and Methods

In this section, we present the materials and methods employed in our study, focusing on the architectures used for medical image segmentation (Section 3.1, Section 3.2, Section 3.3 and Section 3.4). We detail these networks’ theoretical foundations and practical adaptations to handle volumetric data, which is essential for accurately delineating complex anatomical structures. In addition, we provide a dataset description in Section 3.9 and the metrics definition in Section 3.10.

3.1. U-Net

U-Net [37] is a popular biomedical image segmentation architecture with an encoder–decoder structure. In the encoding stage, the network progressively reduces the spatial dimensions of the input through a series of convolutional layers followed by a downsampling operation. For instance, if f l denotes the feature map at layer l, a typical downsampling can be expressed as
f l + 1 = D ( f l ) ,
where D might represent a max pooling operation (e.g., with a stride of two) or a convolution with a stride greater than one, the selected U-Net architecture uses strided convolution as the downsampling operator. This reduction not only decreases computational complexity but also helps in capturing contextual information over larger regions. In the decoding stage, the spatial resolution is gradually recovered using an upsampling operation. Mathematically, if g l + 1 is the feature map at a deeper layer in the decoder, then upsampling can be described as
g l = U ( g l + 1 ) ,
where U may be implemented via transposed convolutions or interpolation, followed by convolution. The network further refines the segmentation by incorporating skip connections that merge the high-resolution features from the encoder with the upsampled features in the decoder, thereby preserving fine structural details.

3.2. Three-Dimensional U-Net

Three-Dimensional U-Net [18] extends the principles of U-Net to volumetric data by replacing two-dimensional operations with their three-dimensional counterparts. In this architecture, the downsampling operator is adapted to process volumetric feature maps. If f l is the 3D feature map at layer l, the volumetric downsampling is represented as
f l + 1 = D 3 D ( f l ) ,
where D 3 D aggregates information along the depth, height, and width dimensions simultaneously. Similarly, the upsampling operator in the decoder is defined as
g l = U 3 D ( g l + 1 ) ,
which reconstructs the spatial dimensions of the volume, again using transposed convolutions or interpolation methods adapted to three dimensions. The integration of skip connections in 3D U-Net allows the network to effectively combine coarse, context-rich features with finer details, thereby enabling the accurate segmentation of complex anatomical structures in modalities such as CT and MRI.

3.3. UNETR

UNETR [19] integrates a transformer-based [38] encoder into the U-Net framework to capture long-range dependencies in volumetric data. The input volume x is first divided into non-overlapping patches, each of which is embedded into a token via a linear projection, i.e.,
t i = ϕ ( x i ) ,
where x i is the ith patch, and ϕ denotes the embedding function. The transformer encoder then processes the sequence of tokens using self-attention, which is computed as
Attention ( Q , K , V ) = softmax Q K d k V ,
with Q, K, and V being the query, key, and value matrices, respectively, and d k the key dimension. The resulting encoded representation is reshaped and fed into a U-Net style decoder that recovers spatial resolution through an upsampling operation
g l = U ( g l + 1 ) ,
while incorporating skip connections to merge high-resolution features from the encoder. This design enables UNETR to effectively combine global context with local details. In our work, we follow the original UNETR implementation [19]. Specifically, feature maps are extracted from five distinct layers of the transformer encoder. These multi-scale features are fused in the upsampling branch by first applying deconvolution to match the spatial dimensions of both the backbone and the target feature maps. This is subsequently refined through a sequence of convolutional operations, batch normalization, and ReLU activations, which collectively improve the training process.

3.4. Swin UNETR

Swin UNETR [20] extends the UNETR framework by employing the Swin Transformer [39] architecture, which utilizes a hierarchical structure with window-based self-attention. In this model, the input volume is partitioned into patches, and self-attention is computed locally within non-overlapping windows. The attention mechanism is the same as defined in [19] applied independently over each window. A subsequent shift in the window partitioning enables the capture of cross-window dependencies, ensuring that global context is effectively modeled. Following the transformer encoding, a decoder analogous to that of UNETR performs upsampling. This approach synergizes local attention with global context, thereby improving the segmentation of complex volumetric data [20].

3.5. SegResNet

SegResNet [40] is a modification of the traditional ResNet architecture tailored to semantic segmentation tasks. This architecture incorporates residual connections, which help mitigate the vanishing gradient problem by allowing gradients to flow through the network without degradation. The encoding path consists of a series of extensive residual blocks that learn multiscale features from the input images, progressively reducing the spatial dimensions. The unique aspect of SegResNet is its autoencoder portion of the architecture, which improves the final segmentation prediction according to [40]. In our implementation, we only rely on the segmentation loss and not on the reconstruction.

3.6. DAF3D

DAF3D [41] introduces a novel approach for robust 3D medical image segmentation by utilizing deep attentive features within a convolutional neural network framework. The architecture leverages attention mechanisms to automatically focus on relevant parts of the input volumes, enhancing segmentation performance by prioritizing critical features while suppressing less informative ones. The network operates through a series of convolutional layers that capture intricate patterns in the volumetric data, supported by a dedicated attention module that refines feature representations based on channel-wise and spatial relationships.

3.7. V-Net

V-Net [42] employs a fully convolutional neural network architecture specifically designed for volumetric segmentation tasks. The structure follows a similar encoder–decoder paradigm while utilizing 3D convolutions to process volumetric data directly. In the encoder portion, V-Net incorporates residual learning blocks that facilitate the training of deeper networks by enhancing gradient flow. The decoder then upsamples the feature maps via transposed convolution and combines them with high-resolution features from the encoder via skip connections, allowing for precise localization.

3.8. Advantages and Disadvantages of the Proposed Networks

Our study utilizes six different 3D segmentation architectures: 3D U-Net, UNETR, Swin UNETR, SegResNet, DAF3D, and V-Net. Among these, the 3D U-Net, V-Net, SegResNet, and DAF3D are CNN-based models that rely on convolutional operations. These models are renowned for their robust ability to extract fine-grained local features and are computationally efficient. However, their dependence on convolutions also results in a limited receptive field, which can restrict the capture of long-range contextual information. This limitation may lead to challenges such as over-segmentation, particularly in complex anatomical regions.
In contrast, UNETR incorporates transformer-based techniques that excel at modeling global contexts. By leveraging self-attention mechanisms, UNETR can capture long-range dependencies that are crucial for understanding subtle variations and complex spatial relationships in volumetric data. Although this approach improves the representation of broader contextual information, it comes at the cost of increased computational complexity and demands a larger training dataset. The process of tokenizing volumetric data adds further preprocessing intricacies, which can be challenging compared to traditional CNNs [43].
Swin UNETR builds upon the strengths of both convolutional and transformer architectures by incorporating a hierarchical structure with window-based self-attention. This design allows the model to effectively capture local patterns within non-overlapping windows, while also integrating global context through a shifting mechanism. Swin UNETR thus achieves a balance between detailed feature extraction and comprehensive contextual integration. However, this balance introduces additional model complexity and requires careful tuning of window sizes and shift parameters to optimize performance [43].
The ensemble framework we propose—3D-NASE—combines these six networks to take advantage of their complementary benefits, also following consolidated literature works that previously combined the characteristics of both model typologies [44]. By integrating the local feature precision and global feature from both CNN and transformer-based models, our ensemble is designed to counteract the individual limitations of each architecture. The combination, implemented via majority and soft voting strategies, ensures improved segmentation accuracy and robustness across diverse anatomical variations and imaging conditions, ultimately enhancing the overall performance of the segmentation task.

3.9. Dataset

The NasalSeg dataset [8] is a large-scale, open-access resource developed for the automatic segmentation of the nasal cavity and paranasal sinuses from 3D CT images. It comprises 130 head CT scans collected from independent patients, including 74 males and 56 females, with an age range of 24 to 82 years (mean 54.6 ± 12.1 years). All scans were acquired using a Biograph 64 scanner (Siemens, Erlangen, Germany) at the Department of Nuclear Medicine/PET Center, Huashan Hospital, and they typically exhibit a volume of 148 × 512 × 512 pixels, with an in-plane resolution of 0.586 × 0.586 mm and a slice spacing of 1.5 mm.
Each scan is accompanied by pixel-wise annotations of five anatomical structures: the left nasal cavity, right nasal cavity, nasal pharynx, left maxillary sinus, and right maxillary sinus. These annotations were initially performed by three skilled annotators with one to five years of experience and subsequently refined and verified by senior experts, with five to ten years of experience, to ensure high accuracy and consistency. The manual annotation and refined segmentation procedure was adopted from 3D Slicer 5.6.2 (3D Slicer: www.slicer.org/, accessed on 6 March 2025). The air threshold was set to −400 to 1000 Hounsfield units (HU) for all scans. The Laplacian filter was then used to highlight regions with rapid intensity changes to enhance edges and facilitate the segmentation. The annotations were verified by four experienced experts specializing in radiology and otolaryngology (with over 20 years of experience).
Furthermore, the dataset has been released as organized into five pre-defined folds to support cross-validation and robust benchmarking of segmentation algorithms. Each set consists of 104 cases for training and 26 cases for validation.
We report samples of the NasalSeg dataset along with over-imposed annotations in Figure 1.

3.10. Metrics

We evaluate our 3D segmentation results using several quantitative metrics that capture different aspects of performance. For a binary segmentation task, we define the following basic terms: T P represents true positives (the number of correctly identified positive elements), T N represents true negatives (the correctly identified negative elements), F P represents false positives (negative elements incorrectly labeled as positive), and F N represents false negatives (positive elements that were missed). In multi-class segmentation, these metrics are computed for each class individually and then averaged to yield an overall performance score without weighting the number of items in each class. This strategy ensures that performance is fairly assessed even when class distributions vary, and, in the context of this work, this decision was made to ensure that each anatomical structure is treated equally during evaluation.
The quality of the segmentation overlap is measured by the DICE coefficient, which is calculated as
DICE = 2 T P 2 T P + F P + F N
This metric is widely adopted in medical image segmentation because it effectively captures the degree of overlap between the predicted segmentation and the ground truth. A DICE score of 1 indicates perfect overlap, while scores closer to 0 reflect poor agreement. The formulation penalizes both FPs and FNs equally, making it a robust indicator of segmentation quality.
Another critical measure is the intersection over union (IoU), which, for a single class, is defined as
IoU = T P T P + F P + F N
The IoU quantifies the ratio of the intersection between the predicted and actual segmentation regions relative to their union, offering an intuitive measure of error when extra regions are included or parts of the target are missed. For multi-class segmentation, the mean intersection over union (mIoU) is calculated by averaging the IoU values across all classes:
mIoU = 1 C c = 1 C T P c T P c + F P c + F N c
where C is the total number of classes, and the subscript c indicates that the metric is computed for each class separately. This aggregate metric is particularly useful in evaluating the overall performance across diverse anatomical structures.
To further analyze segmentation performance, we employ sensitivity (or recall) to assess the proportion of actual positive elements that are correctly identified:
Sensitivity = T P T P + F N
A high sensitivity is crucial, especially in medical applications, because it ensures that most of the true positives are captured, thereby reducing the risk of missing critical regions.
Complementary to sensitivity, specificity measures the proportion of actual negative elements that are correctly identified:
Specificity = T N T N + F P
High specificity is important for minimizing the inclusion of irrelevant regions in the segmentation, which is particularly critical when false positives can lead to misinterpretations.
Precision, also known as positive predictive value (PPV), evaluates the accuracy of the positive predictions:
PPV = T P T P + F P
This metric indicates the proportion of predicted positive elements that are indeed correct. High precision reflects the reliability of the segmentation in not overestimating the presence of the target structure.
Similarly, the negative predictive value (NPV) assesses the correctness of the negative predictions:
NPV = T N T N + F N
A high NPV ensures that the segmentation method reliably identifies non-target areas, which is essential for maintaining overall segmentation integrity.
Finally, the overall performance of the segmentation is summarized by the accuracy metric:
Accuracy = T P + T N T P + T N + F P + F N
Accuracy measures the proportion of correctly classified elements (both positive and negative) out of all elements. Although it provides a global performance view, it should be interpreted alongside the other metrics, especially in scenarios with class imbalance.
Together, these metrics offer a comprehensive framework for assessing segmentation performance. They not only quantify the spatial overlap between the predicted and ground truth segmentations but also provide insights into the balance between correctly identifying positive elements and avoiding erroneous classifications. This multi-faceted evaluation is critical for refining segmentation algorithms, particularly in high-stakes applications such as medical imaging.

4. The Proposed Framework: 3D-NASE

Ensemble learning has emerged as a promising strategy for 3D segmentation in medical imaging due to its ability to reduce model variance, improve generalization, and alleviate artifacts commonly encountered in high-dimensional volumetric data. In our proposed framework, 3D-NASE, we harness the complementary strengths of both CNN-based and transformer-based architectures by integrating six state-of-the-art segmentation models: 3D U-Net, UNETR, Swin UNETR, SegResNet, DAF3D, and V-Net. Each of these models is specifically designed to capture distinct aspects of the data—while the CNN-based networks 3D U-Net, SegResNet, DAF3D, and V-Net excel at extracting fine-grained local details, the transformer-based models UNETR and Swin UNETR are adept at modeling long-range dependencies and global contextual information via self-attention mechanisms. By processing the same CT volume in parallel, these models generate segmentation outputs that are subsequently aggregated using both majority and soft voting strategies. This dual ensembling approach leverages diverse feature representations across different architectures, effectively balancing their individual strengths and limitations. Figure 2 schematically illustrates our pipeline, where the input volume is processed by the selected models, their outputs concatenated, and an ensemble strategy is applied to generate the final prediction.
Experimental Setup. We conducted our experiments on a workstation equipped with an RTX 4060 Ti GPU, 16GB VRAM, and an Intel Core i5-13400 processor. All models were trained for 800 epochs with a learning rate of 3 × 10 4 , and the best models were selected based on the highest DICE score on the validation set. 3D-NASE was built upon the MONAI repository [45]. For optimization, we employed AdamW, together with the DICE loss function.
Data augmentation was an integral part of our training process. Specifically, we utilized two common strategies: random cropping and intensity normalization. For random cropping, a random volume crop with a region of interest (ROI) of 96 × 96 × 32 voxels was extracted from each input image. After an analysis performed over the size of the images, it was identified that an ROI of 96 × 96 × 32 allows a significant portion of the anatomical structure of interest to be captured while reducing the variability introduced by irrelevant or background regions. Intensity normalization was performed on a per-channel basis by considering only non-zero voxels, scaling their intensities to have a mean of 0 and a standard deviation of 1. We do not consider other augmentation strategies in order to focus our results analysis on the impact of the ensemble techniques for a more stable prediction framework, rather than artificially improving upon the quantity of the data towards the same goal.

5. Experimental Results

This section presents the experimental evaluation conducted and the results obtained. We evaluated the employed architectures as stand-alone and the proposed ensemble-based framework across different folds since every experiment uses a five-fold cross-validation approach to ensure a fair assessment and mitigate potential biases. The complete set of experiments across folds is presented in Table 1, where we also compared the results with the baseline provided in [8]. As can be seen, the proposed soft voting approach improves the DICE score by 35.88 % with respect to the baseline ( 94.43 % against 58.55 % ) and reduces the standard deviation by 4.52 % (from 5.43 % of the baseline to 0.91 % ), thereby highlighting enhanced adaptability and stability under varying training conditions. Moreover, the ensemble strategy achieves a 0.37 % improvement over Swin UNETR, the best stand-alone model. Although the improvement is marginal, these results illustrate that leveraging the diverse knowledge acquired by different models leads to more reliable outcomes.
Notably, the current state of the art for the NasalSeg dataset [8] exhibits lower performance than individual models and ensemble strategies regarding the DICE score and demonstrates higher variability across folds.
Extending the evaluation to a broader range of segmentation metrics, Table 2 demonstrates that 3D-NASE precisely segments the relevant regions, as indicated by high accuracy, specificity, and NPV; however, it tends to under-segment the analyzed volumes, except for the nasal pharynx, which suffers from over-segmentation. These insights are further supported by the results provided in Table 3, which shows the DICE score for the five different classes.
Among the classes with similar performance, the right and left nasal cavities are the most challenging to segment, likely due to intra-observer differences observed in the model predictions. This aspect is evidenced by the average DICE score and the higher standard deviation for these two classes, which is an order of magnitude greater than that for the other classes.

6. Qualitative Results

The qualitative results are presented in Figure 3. All models exhibit a tendency toward over-segmentation for the nasal pharynx. This issue is particularly evident in 3D U-Net and UNETR, which also often predict small, scattered label fragments across different image regions. These fragmented predictions result in artifacts and inconsistencies in the final segmentation masks.
The proposed ensemble methods effectively mitigate this issue by filtering out small-scale imperfections, leading to more refined and coherent segmentation. However, larger over-segmentation artifacts tend to persist even in the ensemble predictions, indicating that, while ensemble strategies improve overall segmentation quality, they may not entirely eliminate over-segmentation when it occurs on a larger scale.
The qualitative results presented in Figure 3 align with the observations made during the discussion of the quantitative results. Specifically, the regions most affected by segmentation errors are the right and left nasal cavities. These structures are either misclassified as other anatomical classes, as seen in the U-Net prediction for the second image, or other classes are mistakenly predicted as nasal cavities, as observed in the U-Net prediction for the third image.
These findings further reinforce the quantitative evaluation, highlighting the challenges associated with segmenting the nasal cavity regions. This suggests that their complex shape, small size, and similarity in intensity to surrounding structures contribute to the models’ difficulty in accurately distinguishing them.

7. Limitations and Future Aspects

While our proposed framework for nasal segmentation demonstrates significant improvements over the current state of the art, it also presents some limitations.
The first limitation of our approach is the modest improvement provided by the ensemble techniques over the Swin UNETR architecture. While the soft voting ensemble achieves a DICE score of 94.43%, an incremental gain of 0.54 compared to Swin UNETR—a difference that is not statistically significant, but, as shown in Figure 3, the ensemble notably enhances the prediction quality by substantially reducing artifacts.
It still exhibits artifacts and instances of over-segmentation and under-segmentation that vary across different anatomical classes. Such inconsistencies can compromise the reliability of the segmentation outputs, particularly when delineating complex or subtle structures. To improve these aspects, we plan further investigations. For instance, implementing post-processing steps such as conditional random fields (CRFs) or connected-component analysis can be helpful in removing scattered label fragments and refining the borders of segmented regions. Additionally, experimenting with loss functions like Tversky loss may better balance false positives and negatives during the training phase.
Another critical limitation is the necessity for a large number of training epochs to achieve satisfactory results. This high computational demand can hinder accessibility, especially for smaller institutions or research groups with limited resources in a fine-tuning scenario. The extensive training requirements not only prolong the development cycle but also increase the overall cost of deployment. To take this issue into account, we plan to explore techniques such as progressive learning rate scheduling and early stopping to reduce training duration without compromising performance.
Moreover, while the ensemble methodology effectively combines CNN- and transformer-based architectures, it requires all six selected models to be trained on the same dataset and perform inference on the same volume. This integrated approach further compounds computational and memory resource requirements, posing a significant challenge in scenarios where real-time performance is essential. Moreover, the marginal performance improvement warrants further exploration. For instance, we intend to assess alternative ensembling methods, such as stacked generalization or weighted voting schemes, for which model predictions are weighted based on performance per anatomical region. This could lead to a more balanced segmentation performance across different regions.
Finally, since CT images can vary due to differences in acquisition protocols or patient positioning and the absence of further public datasets dealing with this task, we will simulate varying imaging conditions during training and testing. Introducing synthetic variations (e.g., changes in resolution or noise levels) will assess the model’s robustness under different scenarios.
Future work should focus on optimizing training strategies to reduce the number of required epochs, as well as exploring more resource-efficient ensemble methods. By addressing these issues, it may be possible to maintain or even enhance segmentation performance while reducing the computational burden, thus broadening the applicability of the framework in resource-constrained environments.

8. Conclusions

In this work, we have introduced 3D-NASE, an ensemble-based segmentation framework that integrates the complementary strengths of CNN and attention-based architectures for 3D nasal CT segmentation. By merging predictions from 3D U-Net, UNETR, Swin UNETR, SegResNet, DAF3D, and V-Net through majority and soft voting, our method effectively captures both fine-grained details and long-range contextual information. Experimental results on the NasalSeg dataset reveal that our ensemble approach not only outperforms individual models but also improves the overall state-of-the-art segmentation performance, with a notable 35.88 % increase in the DICE score and a 4.52 % reduction in variability across cross-validation folds. Despite these significant gains, challenges such as over-segmentation artifacts and high computational demands persist. Future work will focus on optimizing the training pipeline and exploring more resource-efficient ensemble strategies to further enhance segmentation accuracy and efficiency, thus broadening the clinical applicability of the proposed framework.

Author Contributions

Conceptualization, L.Z. and A.P.; methodology, L.Z., A.L., A.P., and D.A.M.; Investigation, L.Z., A.L., A.P., D.A.M., and C.D.R.; software, L.Z. and A.P.; writing—original draft, L.Z., A.L., and A.P.; writing—review and editing, L.Z., A.L., A.P., and C.D.R.; supervision, L.Z., A.L., and C.D.R. All authors have read and agreed to the published version of the manuscript.

Funding

We acknowledge financial support under the National Recovery and Resilience Plan (NRRP), Mission 4 Component 2 Investment 1.5—Call for tender No.3277 published on 30 December 2021, by the Italian Ministry of University and Research (MUR), funded by the European Union—NextGenerationEU. Project Code ECS0000038—Project Title eINS Ecosystem of Innovation for Next Generation Sardinia—CUP F53C22000430001—Grant Assignment Decree No. 1056, adopted on 23 June 2022, by the Italian Ministry of University and Research (MUR). This work was also partially supported by the project DEMON, “Detect and Evaluate Manipulation of ONline information”, funded by MIUR under the PRIN 2022 grant 2022BAXSPY (CUP F53D23004270006, NextGenerationEU), and by the SERICS (PE00000014) project under the NRRP MUR program, funded by the EU—NGEU (NextGenerationEU).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the material used and developed for this work is available at the following GitHub repository (https://github.com/snarci/YOLO-SPAM, accessed on 6 March 2025).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CTComputed Tomography
MRMagnetic Resonance
PETPositron Emission Tomography
CNNConvolutional Neural Network
RNNRecurrent Neural Network
TPTrue Positive
FPFalse Positive
FNFalse Negative
IoUIntersection over Union
PPVPositive Predictive Value
NPVNegative Predictive Value
ROIRegion of Interest

References

  1. Siu, J.; Dong, J.; Inthavong, K.; Shang, Y.; Douglas, R.G. Quantification of airflow in the sinuses following functional endoscopic sinus surgery. Rhinology 2020, 58, 257–265. [Google Scholar] [CrossRef] [PubMed]
  2. Lee, S.; Fernandez, J.; Mirjalili, S.A.; Kirkpatrick, J. Pediatric paranasal sinuses—Development, growth, pathology, & functional endoscopic sinus surgery. Clin. Anat. 2022, 35, 745–761. [Google Scholar] [PubMed]
  3. Shang, Y.; Inthavong, K. Numerical assessment of ambient inhaled micron particle deposition in a human nasal cavity. Exp. Comput. Multiph. Flow 2019, 1, 109–115. [Google Scholar] [CrossRef]
  4. Pourmehran, O.; Psaltis, A.; Vreugde, S.; Zarei, K.; Shang, Y.; Inthavong, K.; Wormald, P. Evaluating nebulisation and nasal irrigation efficiency in post-operative chronic rhinosinusitis patients through computational fluid dynamics simulation. Comput. Methods Programs Biomed. 2025, 264, 108697. [Google Scholar] [CrossRef]
  5. Jing, H.; Cui, X. A review on numerical studies of airflow dynamics and particle deposition in human respiratory system. Particuology 2024, 91, 351–377. [Google Scholar] [CrossRef]
  6. Laura, C.O.; Hartwig, K.; Distergoft, A.; Hoffmann, T.; Scheckenbach, K.; Brüsseler, M.; Wesarg, S. Automatic segmentation of the structures in the nasal cavity and the ethmoidal sinus for the quantification of nasal septal deviations. In Proceedings of the Medical Imaging 2021: Computer-Aided Diagnosis, Online, 15–20 February 2021; Mazurowski, M.A., Drukker, K., Eds.; SPIE: Bellingham, WA, USA, 2021; Volume 11597. [Google Scholar] [CrossRef]
  7. Avrunin, O.G.; Tymkovych, M.Y.; Abdelhamid, I.Y.; Shushliapina, N.O.; Nosova, Y.V.; Semenets, V.V. Features of image segmentation of the upper respiratory tract for planning of rhinosurgical surgery. In Proceedings of the 2019 IEEE 39th International Conference on Electronics and Nanotechnology (ELNANO), Kyiv, Ukraine, 16–18 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 485–488. [Google Scholar]
  8. Zhang, Y.; Wang, J.; Pan, T.; Jiang, Q.; Ge, J.; Guo, X.; Jiang, C.; Lu, J.; Zhang, J.; Liu, X.; et al. NasalSeg: A Dataset for Automatic Segmentation of Nasal Cavity and Paranasal Sinuses from 3D CT Images. Sci. Data 2024, 11, 1329. [Google Scholar] [CrossRef]
  9. Homayounieh, F.; Singh, R.; Nitiwarangkul, C.; Lades, F.; Schmidt, B.; Sedlmair, M.; Saini, S.; Kalra, M.K. Semiautomatic segmentation and radiomics for dual-energy CT: A pilot study to differentiate benign and malignant hepatic lesions. Am. J. Roentgenol. 2020, 215, 398–405. [Google Scholar] [CrossRef]
  10. Parmar, C.; Rios Velazquez, E.; Leijenaar, R.; Jermoumi, M.; Carvalho, S.; Mak, R.H.; Mitra, S.; Shankar, B.U.; Kikinis, R.; Haibe-Kains, B.; et al. Robust radiomics feature quantification using semiautomatic volumetric segmentation. PLoS ONE 2014, 9, e102107. [Google Scholar] [CrossRef] [PubMed]
  11. Holbrook, M.D.; Blocker, S.; Mowery, Y.; Badea, A.; Qi, Y.; Xu, E.; Kirsch, D.; Johnson, G.; Badea, C.T. Mri-based deep learning segmentation and radiomics of sarcoma in mice. Tomography 2020, 6, 23–33. [Google Scholar] [CrossRef]
  12. Lin, Y.C.; Lin, G.; Pandey, S.; Yeh, C.H.; Wang, J.J.; Lin, C.Y.; Ho, T.Y.; Ko, S.F.; Ng, S.H. Fully automated segmentation and radiomics feature extraction of hypopharyngeal cancer on MRI using deep learning. Eur. Radiol. 2023, 33, 6548–6556. [Google Scholar] [CrossRef]
  13. Avesta, A.; Hossain, S.; Lin, M.; Aboian, M.; Krumholz, H.M.; Aneja, S. Comparing 3D, 2.5D, and 2D approaches to brain image auto-segmentation. Bioengineering 2023, 10, 181. [Google Scholar] [CrossRef] [PubMed]
  14. Shusharina, N.; Söderberg, J.; Edmunds, D.; Löfman, F.; Shih, H.; Bortfeld, T. Automated delineation of the clinical target volume using anatomically constrained 3D expansion of the gross tumor volume. Radiother. Oncol. 2020, 146, 37–43. [Google Scholar] [CrossRef]
  15. Shivdeo, A.; Lokwani, R.; Kulkarni, V.; Kharat, A.; Pant, A. Evaluation of 3D and 2D Deep Learning Techniques for Semantic Segmentation in CT Scans. In Proceedings of the 2021 International Conference on Artificial Intelligence, Big Data, Computing and Data Communication Systems (icABCD), Durban, South Africa, 5–6 August 2021; pp. 1–8. [Google Scholar] [CrossRef]
  16. Wu, W.; Gao, L.; Duan, H.; Huang, G.; Ye, X.; Nie, S. Segmentation of pulmonary nodules in CT images based on 3D-UNET combined with three-dimensional conditional random field optimization. Med. Phys. 2020, 47, 4054–4063. [Google Scholar] [CrossRef] [PubMed]
  17. Feng, X.; Tustison, N.J.; Patel, S.H.; Meyer, C.H. Brain Tumor Segmentation Using an Ensemble of 3D U-Nets and Overall Survival Prediction Using Radiomic Features. Front. Comput. Neurosci. 2020, 14, 25. [Google Scholar] [CrossRef]
  18. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2016—19th International Conference, Athens, Greece, 17–21 October 2016; Proceedings, Part II. Ourselin, S., Joskowicz, L., Sabuncu, M.R., Ünal, G.B., III, Wells, W., Eds.; Springer: Cham, Switzerland, 2016; Volume 9901, pp. 424–432. [Google Scholar] [CrossRef]
  19. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.A.; Roth, H.R.; Xu, D. UNETR: Transformers for 3D Medical Image Segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, WACV 2022, Waikoloa, HI, USA, 3–8 January 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1748–1758. [Google Scholar] [CrossRef]
  20. Cai, Y.; Long, Y.; Han, Z.; Liu, M.; Zheng, Y.; Yang, W.; Chen, L. Swin Unet3D: A three-dimensional medical image segmentation network combining vision transformer and convolution. BMC Med. Inform. Decis. Mak. 2023, 23, 33. [Google Scholar] [CrossRef] [PubMed]
  21. Hu, P.; Zhou, H.; Yan, T.; Miu, H.; Xiao, F.; Zhu, X.; Shu, L.; Yang, S.; Jin, R.; Dou, W.; et al. Deep learning-assisted identification and quantification of aneurysmal subarachnoid hemorrhage in non-contrast CT scans: Development and external validation of Hybrid 2D/3D UNet. NeuroImage 2023, 279, 120321. [Google Scholar] [CrossRef] [PubMed]
  22. Gu, H.; Gan, W.; Zhang, C.; Feng, A.; Wang, H.; Huang, Y.; Chen, H.; Shao, Y.; Duan, Y.; Xu, Z. A 2D–3D hybrid convolutional neural network for lung lobe auto-segmentation on standard slice thickness computed tomography of patients receiving radiotherapy. BioMed. Eng. OnLine 2021, 20, 94. [Google Scholar] [CrossRef]
  23. Shirokikh, B.; Shevtsov, A.; Dalechina, A.; Krivov, E.; Kostjuchenko, V.; Golanov, A.; Gombolevskiy, V.; Morozov, S.; Belyaev, M. Accelerating 3D Medical Image Segmentation by Adaptive Small-Scale Target Localization. J. Imaging 2021, 7, 35. [Google Scholar] [CrossRef]
  24. de Araújo, A.S.; Pinho, M.S.; da Silva, A.M.M.; Fiorentini, L.F.; Becker, J. A 2.5D Self-Training Strategy for Carotid Artery Segmentation in T1-Weighted Brain Magnetic Resonance Images. J. Imaging 2024, 10, 161. [Google Scholar] [CrossRef]
  25. Satpute, N.; Naseem, R.; Palomar, R.; Zachariadis, O.; Gómez-Luna, J.; Cheikh, F.A.; Olivares, J. Fast parallel vessel segmentation. Comput. Methods Programs Biomed. 2020, 192, 105430. [Google Scholar] [CrossRef]
  26. Yan, J.; Zhuang, T.G.; Zhao, B.; Schwartz, L.H. Lymph node segmentation from CT images using fast marching method. Comput. Med. Imaging Graph. 2004, 28, 33–38. [Google Scholar] [CrossRef]
  27. El Badaoui, R.; Coll, E.B.; Psarrou, A.; Asaturyan, H.A.; Villarini, B. Enhanced CATBraTS for Brain Tumour Semantic Segmentation. J. Imaging 2025, 11, 8. [Google Scholar] [CrossRef] [PubMed]
  28. Jin, Z.; Gong, J.; Deng, M.; Zheng, P.; Li, G. Deep Learning-Based Diagnosis Algorithm for Alzheimer’s Disease. J. Imaging 2024, 10, 333. [Google Scholar] [CrossRef] [PubMed]
  29. Vasiliuk, A.; Frolova, D.; Belyaev, M.; Shirokikh, B. Limitations of Out-of-Distribution Detection in 3D Medical Image Segmentation. J. Imaging 2023, 9, 191. [Google Scholar] [CrossRef]
  30. Poirot, M.G.; Caan, M.; Ruhe, H.G.; Bjørnerud, A.; Groote, I.; Reneman, L.; Marquering, H. Robustness of radiomics to variations in segmentation methods in multimodal brain MRI. Sci. Rep. 2022, 12, 16712. [Google Scholar] [CrossRef] [PubMed]
  31. Wang, C.; Jiang, L.; Wu, X.; Tian, Z.; Peng, B.; Zhao, H.; Jia, J. GroupContrast: Semantic-Aware Self-Supervised Representation Learning for 3D Understanding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, CVPR 2024, Seattle, WA, USA, 16–22 June 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 4917–4928. [Google Scholar] [CrossRef]
  32. Qi, L.; Jiang, Z.; Shi, W.; Qu, F.; Feng, G. GMIM: Self-supervised pre-training for 3D medical image segmentation with adaptive and hierarchical masked image modeling. Comput. Biol. Med. 2024, 176, 108547. [Google Scholar] [CrossRef]
  33. Zhang, S.; Metaxas, D.N. On the challenges and perspectives of foundation models for medical image analysis. Med. Image Anal. 2024, 91, 102996. [Google Scholar] [CrossRef]
  34. Pai, S.; Bontempi, D.; Hadzic, I.; Prudente, V.; Sokač, M.; Chaunzwa, T.L.; Bernatz, S.; Hosny, A.; Mak, R.H.; Birkbak, N.J.; et al. Foundation model for cancer imaging biomarkers. Nat. Mach. Intell. 2024, 6, 354–367. [Google Scholar] [CrossRef]
  35. Xiong, L.; Tang, X.; Jiang, X.; Chen, H.; Qian, B.; Chen, B.; Lin, X.; Zhou, J.; Li, L. Automatic segmentation-based multi-modal radiomics analysis of US and MRI for predicting disease-free survival of breast cancer: A multicenter study. Breast Cancer Res. 2024, 26, 157. [Google Scholar] [CrossRef]
  36. Zhan, F.; He, L.; Yu, Y.; Chen, Q.; Guo, Y.; Wang, L. A multimodal radiomic machine learning approach to predict the LCK expression and clinical prognosis in high-grade serous ovarian cancer. Sci. Rep. 2023, 13, 16397. [Google Scholar] [CrossRef]
  37. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015—18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III. Navab, N., Hornegger, J., III, Wells, W.M., Frangi, A.F., Eds.; Springer: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef]
  38. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All you Need. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems 2017, Long Beach, CA, USA, 4–9 December 2017; Guyon, I., von Luxburg, U., Bengio, S., Wallach, H.M., Fergus, R., Vishwanathan, S.V.N., Garnett, R., Eds.; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 5998–6008. [Google Scholar]
  39. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision, ICCV 2021, Montreal, QC, Canada, 10–17 October 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
  40. Myronenko, A. 3D MRI Brain Tumor Segmentation Using Autoencoder Regularization. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries—4th International Workshop, BrainLes 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 16 September 2018; Revised Selected Papers, Part II. Crimi, A., Bakas, S., Kuijf, H.J., Keyvan, F., Reyes, M., van Walsum, T., Eds.; Springer: Cham, Switzerland, 2018; Volume 11384, pp. 311–320. [Google Scholar] [CrossRef]
  41. Wang, Y.; Ni, D.; Dou, H.; Hu, X.; Zhu, L.; Yang, X.; Xu, M.; Qin, J.; Heng, P.; Wang, T. Deep Attentive Features for Prostate Segmentation in 3D Transrectal Ultrasound. IEEE Trans. Med. Imaging 2019, 38, 2768–2778. [Google Scholar] [CrossRef] [PubMed]
  42. Milletari, F.; Navab, N.; Ahmadi, S. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the Fourth International Conference on 3D Vision, 3DV 2016, Stanford, CA, USA, 25–28 October 2016; IEEE Computer Society: Piscataway, NJ, USA, 2016; pp. 565–571. [Google Scholar] [CrossRef]
  43. Li, X.; Ding, H.; Yuan, H.; Zhang, W.; Pang, J.; Cheng, G.; Chen, K.; Liu, Z.; Loy, C.C. Transformer-Based Visual Segmentation: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 46, 10138–10163. [Google Scholar] [CrossRef] [PubMed]
  44. Yuan, F.; Zhang, Z.; Fang, Z. An effective CNN and Transformer complementary network for medical image segmentation. Pattern Recognit. 2023, 136, 109228. [Google Scholar] [CrossRef]
  45. Cardoso, M.J.; Li, W.; Brown, R.; Ma, N.; Kerfoot, E.; Wang, Y.; Murrey, B.; Myronenko, A.; Zhao, C.; Yang, D.; et al. MONAI: An open-source framework for deep learning in healthcare. arXiv 2022, arXiv:2211.02701. [Google Scholar]
Figure 1. Samples from the NasalSeg dataset along with corresponding labels.
Figure 1. Samples from the NasalSeg dataset along with corresponding labels.
Jimaging 11 00148 g001
Figure 2. The proposed framework, 3D-NASE: the input volume is processed in parallel by the selected models, their outputs are concatenated, and an ensemble strategy is applied to generate the final prediction.
Figure 2. The proposed framework, 3D-NASE: the input volume is processed in parallel by the selected models, their outputs are concatenated, and an ensemble strategy is applied to generate the final prediction.
Jimaging 11 00148 g002
Figure 3. Qualitative results: original images, ground-truth labels, and predictions from 3D U-Net, UNETR, Swin UNETR, DAF3D, V-Net, SegResNet, majority voting, and soft voting. The first two rows represent the first example, while the last two rows represent the second one.
Figure 3. Qualitative results: original images, ground-truth labels, and predictions from 3D U-Net, UNETR, Swin UNETR, DAF3D, V-Net, SegResNet, majority voting, and soft voting. The first two rows represent the first example, while the last two rows represent the second one.
Jimaging 11 00148 g003
Table 1. DICE scores across folds demonstrate that 3D-NASE with soft voting approach outperforms the current state-of-the-art on the NasalSeg dataset by 35.88 % . Notably, ensemble methods achieve superior segmentation performance compared to individual networks. The best results are emphasized in bold.
Table 1. DICE scores across folds demonstrate that 3D-NASE with soft voting approach outperforms the current state-of-the-art on the NasalSeg dataset by 35.88 % . Notably, ensemble methods achieve superior segmentation performance compared to individual networks. The best results are emphasized in bold.
Fold# 0↑1↑2↑3↑4↑AVG↑
Baseline [8]56.2565.7062.3158.7349.7658.55 ± 5.43
3D U-Net94.2193.2290.9392.6194.7593.15 ± 1.49
UNETR93.5192.4892.4992.0194.0792.92 ± 0.85
Swin UNETR94.6593.5793.5892.9894.6793.89 ± 0.74
DAF3D92.4791.5991.4591.0393.1191.93 ± 0.84
SegResNet94.4293.2391.2192.4794.6293.19 ± 1.41
VNet92.9592.1989.5691.6793.4891.98 ± 1.51
3D-NASE (Majority voting)95.1793.9294.0793.4195.4794.41 ± 0.88
3D-NASE (Soft voting)95.2193.9493.9593.4695.5894.43 ± 0.91
Table 2. Quantitative comparison of segmentation performance across different models on the NasalSeg dataset. The soft voting ensemble achieves the highest DICE, mIoU, and Sensitivity, outperforming individual networks. Additionally, ensemble methods maintain high specificity and accuracy, demonstrating their effectiveness in producing reliable segmentation results. The best results are emphasized in bold.
Table 2. Quantitative comparison of segmentation performance across different models on the NasalSeg dataset. The soft voting ensemble achieves the highest DICE, mIoU, and Sensitivity, outperforming individual networks. Additionally, ensemble methods maintain high specificity and accuracy, demonstrating their effectiveness in producing reliable segmentation results. The best results are emphasized in bold.
MethodDICE↑mIoU↑Sensitivity↑Specificity↑Accuracy↑PPV↑NPV↑
3D U-Net93.15 ± 1.4988.57 ± 1.6192.75 ± 2.0298.82 ± 0.3699.64 ± 0.0493.95 ± 0.8199.03 ± 0.14
UNETR92.92 ± 0.8587.89 ± 0.8693.61 ± 0.8898.97 ± 0.0899.62 ± 0.0392.51 ± 0.9598.81 ± 0.14
Swin UNETR93.89 ± 0.7489.91 ± 0.5994.54 ± 0.9098.92 ± 0.0499.12 ± 0.0293.58 ± 0.6498.65 ± 0.15
DAF3D91.93 ± 0.8486.32 ± 0.8492.79 ± 0.9398.76 ± 0.1799.54 ± 0.0291.29 ± 0.8798.50 ± 0.06
SegResNet93.19 ± 1.4188.59 ± 1.5794.06 ± 1.6999.05 ± 0.2399.63 ± 0.0592.57 ± 1.2498.78 ± 0.28
VNet91.98 ± 1.5186.63 ± 1.6492.25 ± 1.5998.68 ± 0.3599.57 ± 0.0592.05 ± 1.3798.66 ± 0.28
3D-NASE (majority voting)94.41 ± 0.8890.43 ± 0.8395.10 ± 1.0499.23 ± 0.1199.70 ± 0.0293.94 ± 0.6999.02 ± 0.09
3D-NASE (soft voting)94.43 ± 0.9190.46 ± 0.9495.10 ± 1.0899.14 ± 0.1499.70 ± 0.0294.51 ± 0.7299.13 ± 0.08
Table 3. DICE scores for each anatomical class segmented by different models on the NasalSeg dataset. The soft voting ensemble achieves the highest DICE scores across all classes, demonstrating superior segmentation performance for the maxillary sinuses, nasal cavities, and nasal pharynx. The best results are emphasized in bold.
Table 3. DICE scores for each anatomical class segmented by different models on the NasalSeg dataset. The soft voting ensemble achieves the highest DICE scores across all classes, demonstrating superior segmentation performance for the maxillary sinuses, nasal cavities, and nasal pharynx. The best results are emphasized in bold.
MethodBackground↑Maxillary Sinus (R)↑Maxillary Sinus (L)↑Nasal Cavity (R)↑Nasal Cavity (L)↑Nasal Pharynx↑
3D U-Net99.48 ± 0.6994.91 ± 1.7294.21 ± 2.6888.33 ± 3.0688.20 ± 2.7793.77 ± 2.15
UNETR99.44 ± 0.4193.78 ± 0.8394.13 ± 1.9788.42 ± 2.9388.07 ± 2.8993.65 ± 1.06
Swin UNETR99.59 ± 0.4394.71 ± 1.0594.28 ± 1.8989.89 ± 2.7589.82 ± 2.7895.03 ± 1.06
DAF3D99.30 ± 0.3895.32 ± 0.6794.52 ± 1.7183.91 ± 2.7683.66 ± 2.7094.90 ± 0.91
SegResNet99.46 ± 0.8694.81 ± 2.0894.08 ± 2.7588.40 ± 3.0788.34 ± 3.0794.07 ± 1.57
VNet99.35 ± 0.9594.02 ± 2.8193.14 ± 2.9186.19 ± 2.9186.06 ± 2.6793.09 ± 2.18
3D-NASE (Majority voting)99.56 ± 0.8795.93 ± 2.0895.10 ± 2.9190.23 ± 2.6490.17 ± 2.7895.51 ± 1.18
3D-NASE (Soft voting)99.56 ± 0.9195.94 ± 2.9795.09 ± 2.8690.28 ± 2.8390.17 ± 2.7795.54 ± 1.83
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pani, A.; Zedda, L.; Mura, D.A.; Loddo, A.; Di Ruberto, C. 3D-NASE: A Novel 3D CT Nasal Attention-Based Segmentation Ensemble. J. Imaging 2025, 11, 148. https://doi.org/10.3390/jimaging11050148

AMA Style

Pani A, Zedda L, Mura DA, Loddo A, Di Ruberto C. 3D-NASE: A Novel 3D CT Nasal Attention-Based Segmentation Ensemble. Journal of Imaging. 2025; 11(5):148. https://doi.org/10.3390/jimaging11050148

Chicago/Turabian Style

Pani, Alessandro, Luca Zedda, Davide Antonio Mura, Andrea Loddo, and Cecilia Di Ruberto. 2025. "3D-NASE: A Novel 3D CT Nasal Attention-Based Segmentation Ensemble" Journal of Imaging 11, no. 5: 148. https://doi.org/10.3390/jimaging11050148

APA Style

Pani, A., Zedda, L., Mura, D. A., Loddo, A., & Di Ruberto, C. (2025). 3D-NASE: A Novel 3D CT Nasal Attention-Based Segmentation Ensemble. Journal of Imaging, 11(5), 148. https://doi.org/10.3390/jimaging11050148

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop