Next Article in Journal
Promoting Re-Epithelialization in Diabetic Foot Wounds Using Integrative Therapeutic Approaches
Previous Article in Journal
Voice-Based Early Diagnosis of Parkinson’s Disease Using Spectrogram Features and AI Models
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

LG-UNet Based Segmentation and Survival Prediction of Nasopharyngeal Carcinoma Using Multimodal MRI Imaging

1
College of Electronic Engineering (College of Artificial Intelligence), South China Agricultural University, Guangzhou 510642, China
2
State Key Laboratory of Oncology in South China, Guangdong Key Laboratory of Nasopharyngeal Carcinoma Diagnosis and Therapy, Guangdong Provincial Clinical Research Center for Cancer, Sun Yat-sen University Cancer Center, Guangzhou 510060, China
3
State Key Laboratory of Oncology in South China, Department of Radiology, Sun Yat-sen University Cancer Center, Guangzhou 510060, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Bioengineering 2025, 12(10), 1051; https://doi.org/10.3390/bioengineering12101051
Submission received: 8 August 2025 / Revised: 14 September 2025 / Accepted: 23 September 2025 / Published: 29 September 2025
(This article belongs to the Section Biosignal Processing)

Abstract

Image segmentation and survival prediction for nasopharyngeal carcinoma (NPC) are crucial for clinical diagnosis and treatment decisions. This study presents an improved 3D-UNet-based model for NPC GTV segmentation, referred to as LG-UNet. The encoder introduces deep strip convolution and channel attention mechanisms to enhance feature extraction while avoiding spatial feature loss and anisotropic constraints. The decoder incorporates Dynamic Large Convolutional Kernel (DLCK) and Global Feature Fusion (GFF) modules to capture multi-scale features and integrate global contextual information, enabling precise segmentation of the tumor GTV in NPC MRI images. Risk prediction is performed on the segmented multi-modal MRI images using the Lung-Net model, with output risk factors combined with clinical data in the Cox model to predict metastatic probabilities for NPC lesions. Experimental results on 442 NPC MRI scans from Sun Yat-sen University Cancer Center showed DSC of 0.8223, accuracy of 0.8235, recall of 0.8297, and HD95 of 1.6807 mm. Compared to the baseline model, the DSC improved by 7.73%, accuracy increased by 4.52%, and recall improved by 3.40%. The combined model’s risk prediction showed C-index values of 0.756, with a 5-year AUC value of 0.789. This model can serve as an auxiliary tool for clinical decision-making in NPC.

1. Introduction

Nasopharyngeal carcinoma (NPC) has a high incidence, particularly in southern China [1,2]. Distant metastasis in NPC poses a significant challenge in clinical treatment, making the effective identification and screening of patients at high risk of metastasis crucial for developing precise treatment plans and improving patient survival rates [1,2]. In recent years, deep learning techniques have been widely applied in NPC risk prediction. Neural network-based models enable automatic identification, precise segmentation, and feature extraction of NPC lesions from MRI images [3]. Building on this capability, these models facilitate the prediction of distant metastasis risk in NPC.
Currently, U-Net and its variants, based on convolutional neural networks (CNNs) [3,4,5,6,7,8,9], are widely adopted for medical image segmentation. CNNs effectively capture local image features through convolutional kernels of varying sizes [7]. However, their limited receptive fields, constrained by small kernel sizes, restrict their ability to model global patterns and contextual relationships. On the other hand, Vision Transformers (ViTs) [9,10,11], although originally developed for image classification, have recently been adapted for medical image segmentation to expand the receptive field; however, they exhibit weaker local feature extraction capabilities, often requiring additional strategies such as attention mechanisms to enhance segmentation performance [12,13,14,15,16].
To address these limitations, CNNs integrated with large convolutional kernels (LCKs) have been proposed. LCKs leverage surrounding tissue information to ensure a large receptive field while capturing large-scale features and contextual information that traditional CNNs struggle to obtain [17,18]. This makes LCK-based models particularly suitable for segmenting larger pathological structures [19]. In tasks involving audio, image, and other modalities [20,21], specifically adapted large-kernel [22,23,24] CNNs have demonstrated superior performance compared to ViTs. Furthermore, LCK modules with dynamic mechanisms enhance the modeling capability of tumor heterogeneity regions through adaptive weight adjustments.
However, the effectiveness of tumor image segmentation depends not only on the backbone network but also on the segmentation approach. Methods based on 2D [25,26], 3D [27,28,29], and 2.5D [30,31] segmentation all face challenges such as small spatial representation of lesions, background interference, and boundary ambiguity. The LVPA-UNet segmentation model [32], which integrates 2D and 3D parallel processing along with layer-channel attention mechanisms, addresses these issues to a significant extent. Nevertheless, its suboptimal skip connections and decoder design limit its ability to fully leverage contextual information.
To effectively capture large-scale features of nasopharyngeal carcinoma (NPC), this study proposes an improved LG-UNet segmentation model that expands the receptive field, leverages comprehensive contextual information, and addresses challenges such as ensuring adequate spatial representation of lesions, mitigating background interference, and resolving boundary ambiguity. The model integrates a dynamically cascaded large convolutional kernel (DLCK) mechanism [33] with a 2D and 3D parallel processing approach. The proposed LG-UNet model was applied to achieve precise segmentation of MRI images from 442 NPC cases obtained from the Sun Yat-sen University Cancer Center.
Although several models (e.g., DA-DSUNet, NPCNet, LVPA-UNet) [8,32,34] have improved NPC segmentation [35,36,37,38], and MRI-based approaches have been explored for survival prediction [39,40,41,42,43,44,45], most studies treat these tasks separately. Our work provides an integrated framework.
Building on the segmentation results, risk coefficients were extracted using the deep learning network Lung-Net [39]. These risk coefficients were then combined with clinical diagnostic information to construct a survival prediction model for NPC patients based on the Cox proportional hazards (CoxPH) model [40]. This approach provides a scientific foundation for developing individualized treatment strategies.
The main contributions of this study are as follows:
(a) Based on a 2D-3D hybrid network architecture, we propose the LG-UNet model for accurate gross tumor volume (GTV) segmentation of nasopharyngeal carcinoma (NPC) using multi-modality MRI images. This model assists clinicians in precise tumor staging.
(b) To complement the single T1-modality image dataset, T2 and T1-weighted images are introduced to enable multi-modality joint decision-making, thereby enriching the information available to the model.
(c) Leveraging the Lung-Net and Cox proportional hazards (CoxPH) models, transfer risk coefficients are extracted from high-dimensional features within the segmented regions to facilitate survival prediction.

2. Materials and Methods

2.1. Network Encoder

The LG-UNet employs the encoder architecture described in reference [1], organized in a four-stage cascading manner. In Figure 1, these stages work together to gradually reduce the resolution of feature maps while extracting richer semantic features. This structured encoder cleverly integrates the Layer-Volume Parallel Attention(LVPA) module with the Overlapping Patch Embedding down-sampling operation, significantly improving the network’s performance in medical image segmentation tasks. The LVPA module, comprising 2D-MSCA, 3D-MSCA, and channel-wise attention modules, enables the simultaneous extraction of 2D slice and 3D volume features. The channel-wise attention module emphasizes tumor-related features, enhancing the network’s ability to identify lesion areas. Meanwhile, the Overlapping Patch Embedding down-sampling operation divides the image into multiple patches for processing using a sliding window approach. Furthermore, the Skip Connection strategy in the encoder not only preserves feature information at different scales but also ensures effective fusion of these features during the decoding stage, enabling comprehensive utilization of both detailed and contextual information. This encoder design allows the network to generate more discriminative feature maps, achieving finer tumor segmentation.

2.2. Network Decoder

The decoder of LG-UNet consists of four stages, each containing a group of GDU modules. These modules work synergistically to effectively extract multi-scale features and integrate global contextual information, providing rich feature representations for the decoder stage. The construction of feature dimensions in the decoder corresponds to that of the encoder and are designed to retain as much image feature information as possible, considering the relatively low D average layer depth in the dataset used in this study.

2.3. The Architecture of the GDU Module

The specific architecture is shown in Figure 2. The GDU module is a key component of LG-UNet, composed of convolutional layers such as Conv3 and Conv1, Group-Norm batch normalization layer, RELU activation function, and the GFF and DLCK modules.

2.3.1. The Overview of GFF Module

Based on the feature fusion module described in reference [2], the number of feature channels and convolution layers has been optimized, leading to the development of a Global Feature Fusion (GFF) module. In Figure 3, this module adaptively fuses multi-scale local features based on global information. In traditional U-Net upsampling, due to the limitations of the operation itself, feature maps often experience dilution of information and loss of details as the resolution increases. This leads to increased sparsity of feature maps and blurred details, which in turn affects the accuracy of segmentation results and the clarity of edges. Global information carries more complete contextual information, not only complementing fine details that may be lost during the upsampling process, such as texture and edge information, but also providing a deeper understanding of the overall structure of the image. Therefore, the incorporation of global information can effectively address the issue of feature information loss and detail blurring caused by upsampling.
Specifically, the GFF module receives feature maps x u p and x s k i p from the skip connections and upsampling, respectively, for feature concatenation. The module then applies cascaded average pooling (AvgPool), two 1 × 1 × 1 convolution operations (Conv1), and two Sigmoid activations to first expand and then compress the feature channels, thereby extracting more detailed global channel descriptors. This enables the model to assign attention weights based on the overall performance of each channel across the entire spatial domain. This operation can be succinctly expressed as:
α g c   =   Sigmoid   ( Conv 1   ( Sigmoid   ( Conv 1   ( AvgPool   ( [ x u p ;   x s k i p ] ) ) ) ) )
Based on the obtained channel information weights, key features in the concatenated feature map are enhanced, while relatively less important features are suppressed. This helps the model to focus on the more relevant features for the task. Subsequently, a 1 × 1 × 1 convolution (Conv1) operation is applied again to extract more discriminative feature maps. This operation can be succinctly expressed as:
x α =   Con v 1   ( α g c [ x u p ; x s k i p ] )
At the same time as the above two steps, 1 × 1 × 1 convolution (Conv1) operations are applied to the feature maps x u p and x s k i p , followed by feature multiplication to obtain the global spatial location feature α g s . α g s reflects the importance of each spatial position within the entire feature representation, including both x u p and x s k i p . This operation can be succinctly expressed as:
α g s =   Sigmoid   ( Con v 1   ( x u p )     Con v 1   ( x s k i p ) )
x o u t = α g s     x α

2.3.2. The Overview of DLCK Block

Traditional convolutional neural networks (CNNs) typically employ small convolutional kernels, such as 3 × 3 × 3 or 5 × 5 × 5, which limit the receptive field and hinder the ability to capture comprehensive contextual information. This study incorporates the Dynamic Large Convolutional Kernel (DLCK) block, as referenced in [2], introducing larger convolutional kernels, such as 5 × 5 × 5 and 7 × 7 × 7, to expand the receptive field. Since the upsampling operation reduces feature map resolution, applying the DLCK block before upsampling in the decoder helps the model better retain fine-grained details that would otherwise be lost. Unlike traditional CNNs that employ parallel multi-kernel structures, the DLCK block adopts a cascaded approach, progressively increasing kernel size and dilation rates. This cascaded design offers two distinct advantages: (a) recursively aggregating contextual information within the receptive field, continuously expanding the effective receptive field. (b) Assigning different weights to features based on their respective receptive field sizes, enabling deeper features from larger receptive fields to contribute more significantly to the final DLCK output.
Figure 4 illustrates the overall structure of the DLCK block. Given an input feature map F l R C × D × H × W at layer l , a 5 × 5 × 5 depth-wise convolution with a dilation rate of 1, denoted as DWConv ( 5,1 ) , is first applied, followed by a 7 × 7 × 7 depth-wise convolution with a dilation rate of 3, denoted as DWConv ( 7 , 3 ) . This sequential operation assimilates local information while progressively expanding the effective receptive field. This process can be concisely expressed as:
F 1 ι =   DWCon v   ( 5 , 1 ) ( F l )
F 2 ι = DWCon v   ( 7 , 3 ) ( F 1 ι )
The cascaded design enables the DLCK block to achieve an effective receptive field equivalent to a 23 × 23 × 23 convolution kernel, significantly reducing computational costs. Subsequently, F 1 ι , F 2 ι are concatenated and processed separately using max pooling and average pooling to obtain dimension-reduced features. A 7 × 7 × 7 convolution layer, denoted as C o n v 7 , followed by a Sigmoid activation, is then applied to generate dynamic weights ω 1 , ω 2 , which adaptively determine the relative importance assigned to different features.
ω a v g   = A v g p o o l ( [ F 1 ι ;   F 2 ι ] )
ω m a p   = M a x p o o l ( [ F 1 ι ;   F 2 ι ] )
[ ω 1 ;   ω 2 ] = Sigmoid   ( Con v 7   ( [ ω a v g ;   ω m a p ] ) )
The dynamic weights guide the calibration of features extracted from different convolution kernels. Finally, a residual connection is applied to ensure that no original information is lost and to mitigate the risk of overfitting.
F l   = ( ( ω 1 F 1 ι )     ( ω 2 F 2 ι ) ) + F l
The DLCK module is implemented by integrating the DLCK block into two convolutional layers with GELU activation and residual connections. It comprises feature extraction, global spatial relationship modeling, dynamic selection value generation, and feature calibration output. The operations performed by the DLCK module can be concisely expressed as:
F l =   Con v 1     ( F l 1 )
F l = DLCK   ( G E L U   ( F l ) )
F ^ ι   = Con v 1   ( F l ) + F l 1  
LG-UNet is built upon a CNN architecture, incorporating a DLCK module and an MLP (Multilayer Perceptron) module following the global feature fusion operation in the encoder. This enables the network to capture multi-scale contextual information within the fused features, thereby enhancing feature representation.

2.4. Survival Prediction Task

LungNet is a deep learning model specifically designed for medical image segmentation tasks, particularly in lung imaging. In this study, the model was fine-tuned based on the task scenario, utilizing images segmented by LG-UNet. A multilayer perceptron (MLP), composed of a series of fully connected layers, batch normalization layers, ReLU activation functions, and dropout layers, was employed to extract and transform key features while effectively implementing regularization. Through the combined effect of these layers, the model precisely captures and identifies critical information from the high-dimensional features within the segmented regions that are decisive for metastasis risk assessment. Ultimately, the model outputs a quantitative risk score, which, when integrated with clinical information, facilitates further survival prediction.
The Cox proportional hazards (CoxPH) regression model is a commonly used semi-parametric model in survival analysis, designed to analyze the impact of covariates on survival time while accounting for censoring. It operates under assumptions such as proportional hazards, linearity, and independence. The regression coefficients are typically estimated using the maximum likelihood estimation (MLE) method. The fundamental form of the model is expressed as:
h ( t X ) = h 0 ( t ) e x p ( β 1 X 1 + β 2 X 2 + + β p X p )
After obtaining the risk scores from the model, the five-year ROC curve is plotted by computing the true positive rate (sensitivity) and false positive rate at different risk score thresholds. Additionally, the Youden index is calculated based on these values. By iterating through all possible thresholds, the threshold corresponding to the maximum Youden index is determined as the optimal threshold.
Based on the identified optimal threshold, patients are classified into high-risk and low-risk groups. The Kaplan–Meier method is then used to generate survival curves. The log-rank test is performed to assess the statistical significance of survival differences between the high-risk and low-risk groups, thereby evaluating the effectiveness of the risk score combined with clinical information in survival prediction.

3. Data

3.1. Image Segmentation Task Data

This study retrospectively analyzed 495 stage II nasopharyngeal carcinoma (NPC) patients from Sun Yat-sen University Cancer Center, with data collected between May 2010 and July 2017 in Figure 5. Based on the inclusion and exclusion criteria and the selection process, a total of 442 NPC patients with corresponding diagnostic MRI scans (median age: 44 years; 310 males, 132 females) were included in this study. The dataset was randomly divided into training, testing, and validation sets in a 3:1:1 ratio. Each patient’s MRI data consisted of three modalities: axial T1-weighted (T1W), T2-weighted (T2W), and contrast-enhanced T1-weighted (T1W) images. Two experienced radiation oncologists manually delineated the gross tumor volume (GTV) on all MRI slices following institutional guidelines. Discrepancies were resolved by consensus with a senior oncologist, and the final contours served as the ground truth masks for training and evaluation.
Prior to the experiments, all three-modal MRI datasets underwent uniform preprocessing. In addition to pixel value normalization, data augmentation techniques, including random flipping and rotation, were applied to enhance the model’s generalization ability. Furthermore, to ensure consistent voxel spacing and image dimensions, the MRI images were resampled using trilinear interpolation, with voxel spacing set to 0.5 × 0.5 × 6 mm3 and dimensions adjusted to 32 × 256 × 256, which adequately covering all tumor lesions. For samples exceeding the specified dimensions, boundaries along the x, y, and z axes were cropped, ensuring that the cropping did not affect the lesion areas.

3.2. Survival Prediction Task Data

The clinical and treatment information collected in this study is shown in Table 1, including age, gender, T stage, N stage, lactate dehydrogenase (LDH), Epstein–Barr virus deoxyribonucleic acid (EBV DNA), and distant metastasis-free survival (DMFS), among others. For missing LDH and EBV DNA data, the missing values were imputed using the mean value. Regarding the follow-up plan, patients were followed up every three months during the first two years after admission, then every six months, and finally annually, in order to monitor the occurrence of distant metastasis. Distant metastasis-free survival (DMFS) was used as the primary clinical outcome.

Survival Prediction Data Preprocessing

To balance the number of positive and negative samples, a subset of 388 patients’ clinical information and corresponding deep learning image features was randomly selected from the original dataset for survival prediction. To address missing information in the clinical data, data cleaning was performed to remove obvious input errors and anomalous values, ensuring dataset accuracy and consistency. For missing numerical data, mean imputation was applied, filling missing values with the average of each feature to maintain the overall data distribution. Given the relatively small sample size, five-fold cross-validation was employed in survival prediction experiments to fully utilize the limited data and enhance the model’s reliability and generalization capability.

3.3. Performancce Metrics

3.3.1. Segmentation Task Evaluation Metrics

The Dice Similarity Coefficient (DSC) is a region-based metric used to evaluate the overlap between the predicted values (P) and the ground truth values (G). The definition of DSC is as follows:
D S C ( P , G ) = 2 P G P + G
DSC ranges from 0 to 1, with a higher DSC indicating better performance.
HD95 measures the boundary discrepancy between the prediction and the ground truth, where lower values indicate better segmentation performance. Let Z p r e d represent the boundary of the predicted mask, and Z t r u e represent the boundary of the ground truth mask. The maximum HD is defined as:
m a x h Z p r e d , Z t r u e , h Z t r u e , Z p r e d
h Z p r e d , Z t r u e = m a x a Z p r e d m i n b Z t r u e a b
h Z t r u e , Z p r e d = m a x a Z t r u e m i n b Z p r e d a b
The 95% Hausdorff Distance (HD95) is adopted to reduce the impact of a small number of outliers.

3.3.2. Survival Prediction Task Evaluation Metrics

The Time-Dependent Receiver Operating Characteristic (TD-ROC) curve is primarily constructed by calculating sensitivity and specificity at a specific time point t .
S e n s i t i v i t y t = T P t T P t + F N t
S p e c i f i c i t y t = T N t T N t + F P t
In this context, TP(t) represents the number of individuals correctly predicted as experiencing the event by time t or earlier (true positives), while FN(t) denotes the number of individuals for whom the event actually occurred but was not predicted within this timeframe (false negatives). Similarly, TN(t) indicates the number of individuals accurately predicted as not experiencing the event (true negatives), whereas FP(t) refers to the count of individuals incorrectly predicted to experience an event that did not occur (false positives).
The Area Under the Curve (AUC) serves as a comprehensive metric for evaluating the predictive accuracy of a model. At a specific time point t, the AUC can be expressed as:
A U C t = 0 1 S e n s i t i v i t y t F P   R a t e d F P   R a t e
where the False Positive Rate (FP Rate) is defined as 1−Specificity1−Specificity. The AUC ranges from 0.5, indicating no predictive ability, to 1.0, representing perfect predictive performance.
The Concordance Index (C-Index), quantifies the ability of a given model to correctly predict the order of survival times. The C-Index ranges from 0.5 to 1.0, where a value of 0.5 indicates purely random predictions, signifying that the model has no predictive power, and a value of 1.0 denotes perfect concordance between the model’s predictions and the observed outcomes. The formula for the C-Index is as follows:
C I n d e x = Σ i , j 1 T i > T j 1 r i > r j δ j Σ i , j T i > T j δ j
Here, 1 T i > T j denotes a function that returns 1 if T i > T j ,and 0 otherwise; r i represents the risk score of sample i ; and 1 r i > r j is a function that returns 1 when r i > r j , and 0 otherwise.

3.4. Implementation Details

In this study, LG-UNet and subsequent survival prediction were implemented using PyTorch 1.11.0 + cu113 on a Windows Server equipped with a 12-core Intel Xeon CPU E5-2650 v4, 24 GB of RAM, and an NVIDIA GeForce RTX 3090. For the segmentation task, the Adam optimizer and a cosine scheduler were used with a batch size of 1. The initial learning rate was set to 1e-4, and training was performed for 300 epochs. To ensure reproducibility, we set the random seed to 1234, and experiments were run with Python 3.9.11, PyTorch 1.11.0 + cu113, CUDA 11.3, and cuDNN 8.2. The model contains 6.97 M parameters and 276.6 GFLOPs. Training and validation loss curves have also been included (see Figure A1 in the Appendix A).During the image testing process, the sliding window method from MONAI was applied, with a window size of 32 × 256 × 256 and an overlap rate of 50%. Validation was performed after each training epoch, and the model weights that achieved the highest validation performance were retained. In the segmentation experiment, the time for one training iteration was approximately 1200 s, and the time for one validation iteration was approximately 1700 s. All model parameters in both the image segmentation and survival prediction tasks remained consistent throughout this study.

4. Results

4.1. Analysis of Segmentation Experiment Results

4.1.1. Comparison of Model Experiments

The proposed LG-UNet model was evaluated in comparison to a range of widely used medical image segmentation models, including 3D UNet [40], TransBTS [20], VT-UNet [19], UNETR [18], and LVPA-UNet [1]. As demonstrated in Table 2, LG-UNet consistently outperformed all competing models across multiple evaluation metrics, including the currently state-of-the-art LVPA-UNet.
Specifically, LG-UNet achieved an average Dice Similarity Coefficient (DSC) of 82.22%, surpassing LVPA-UNet’s 80.09% by nearly 2 percentage points. Additionally, LG-UNet exhibited superior performance in terms of HD95 (1.68 mm vs. 1.76 mm), precision (82.36% vs. 79.06%), and recall (82.97% vs. 82.29%) compared to LVPA-UNet. These results underscore LG-UNet’s enhanced accuracy in segmentation and classification tasks.
Figure 6a,b present real segmentation examples comparing LG-UNet with five other representative models. In these figures, the blue contours represent the ground truth annotations of the tumor GTV, while the red contours denote the segmentation predictions generated by the models. Each set of images, from top to bottom, displays T1, T1C, and T2 modality MRI slices of the same scan.
In Figure 6a, the segmentation results from 3D-UNet and TransBTS exhibit discontinuities and distortions in morphology. UNETR, VT-UNet, and LVPA-UNet also demonstrate varying degrees of inaccuracy, with noticeable discrepancies between their predictions and the ground truth. In Figure 6b, Areas of under- and over-segmentation by the baseline models are highlighted with green arrows. In contrast, yellow arrows point out the superior agreement of the LG-UNet boundaries with the ground truth. In contrast, Figure 6c shows that the predicted shapes from all six models generally align with the ground truth. However, 3D-UNet, TransBTS, UNETR, and VT-UNet display limited precision in capturing edge details. While LVPA-UNet performs better in edge detail preservation, its predictions extend beyond the ground truth boundaries. Notably, LG-UNet’s predictions in both slices closely match the ground truth, with superior edge detail representation. This observation highlights LG-UNet’s enhanced capability in processing edge details and deep features.

4.1.2. Comparison of Module Ablation Experiments

To elucidate the specific contributions of each key component in LG-UNet, we conducted a series of ablation studies by progressively removing or modifying modules within the GDU (Gated Deformable Unit). The impact of these changes on the overall segmentation performance was quantitatively evaluated. As illustrated in Table 3, LVPA-UNet(U) [1] serves as the baseline model for the ablation experiments, where “U” denotes a cascaded decoder module containing only upsampling and its associated auxiliary components. “DU” represents a cascaded module incorporating the DLCK (Dynamic Large Convolutional Kernel), upsampling, and auxiliary components. “GU” refers to a cascaded module that includes the GDU, upsampling, and auxiliary components. Finally, “GDU” represents the complete GDU module.
Figure 7 presents real segmentation examples from the ablation experiments conducted on the individual modules of LG-UNet. Each set of images, from top to bottom, displays T1, T1C, and T2 modality MRI slices of the same scan. In the segmentation experiments performed on the same case, all ablated models demonstrated relatively strong performance, achieving Dice scores of 0.8936, 0.8873, 0.8978, and 0.916, respectively. Compared to the baseline model, the DU model exhibited a slight performance decline, as the use of large convolutional kernels alone led to the omission of certain local details in the images. In contrast, the GU and GDU models showed improved accuracy in predicting the extent of the GTV lesion and better alignment with edge details compared to the baseline model. These results highlight the contributions of the GDU module to enhancing segmentation precision and detail preservation.

4.1.3. Comparison of Multimodal and Single-Modality Inputs

The multimodal setting yielded higher Dice scores and lower HD95 compared to single-modality inputs, confirming that complementary anatomical and contrast information improves segmentation accuracy and clinical applicability, as detailed in Table 4.

4.2. Analysis of Survival Prediction Experiment Results

In this study, a number of widely used medical picture segmentation models based on the Lung-Net framework and images processed by LG-Unet, were subjected to risk coefficient extraction. The extracted features were combined with clinical information for further analysis. Due to the limited availability of clinical data, a 5-fold cross-validation approach was employed to train the Cox proportional hazards (CoxPH) model. In Table 5, the optimal Concordance Index (C-Index) achieved was 0.756, demonstrating the model’s predictive performance.
As illustrated in the TD-ROC curve, in Figure 8, LG-UNet achieves the highest predictive performance among all compared models, with its ROC trajectory consistently positioned closest to the upper-left corner and an AUC of 0.789. This substantially outperforms other architectures, underscoring its superior ability in time-dependent risk stratification and outcome discrimination. Additionally, LVPA-UNet and UNETR also demonstrate good performance, with AUC values ranging from 0.76 to 0.77. In contrast, 3D-UNet and the model based on clinical features exhibit poorer predictive capability, with AUC values of 0.625 and 0.640, respectively, and curves that are significantly offset from the top-left corner, demonstrating subpar predictive ability and a reduced capacity to differentiate between high-risk and low-risk populations.
The use of clinical data alone (Clinical) and clinical data combined with risk ratings obtained from LG-UNet (LG-UNet + Clinical) were compared in the Kaplan–Meier (KM) curve analysis. Both strategies showed a substantial difference in distant metastasis-free survival (DMFS) between the low-risk and high-risk groups (p-value: <0.001), as shown in Figure 9. The survival disparities between the risk groups were more clearly defined by the KM curve that included LG-UNet risk scores. This suggests that the risk scores produced by LG-UNet are an effective way to supplement clinical data and offer more discriminative information for patient survival prediction. In conclusion, LG-UNet demonstrates strong performance and potential clinical utility in survival prediction tasks.

5. Discussion

This study utilized a tri-modality MRI dataset comprising 442 cases of Stage II nasopharyngeal carcinoma from a single medical center. The gross tumor volume (GTV) annotations were meticulously delineated by two experienced radiation oncologists. Based on this dataset, we developed LG-UNet, a novel tumor segmentation model with multiple advantages. The Lung-Net prediction network was then used to obtain risk coefficients by utilizing the segmentation findings from LG-UNet. A survival prediction model was created by further integrating these risk coefficients with clinical data.
LG-UNet is an improved model based on the LCK-CNN framework, employing a 2D and 3D parallel processing approach for segmentation. In order to ensure precise spatial representation of lesions while reducing background interference and border ambiguity, it improves feature fusion operations and integrates a dynamic method for LCK (Large Convolutional Kernel) to enlarge the receptive field. LG-UNet achieves a Dice Similarity Coefficient (DSC) of 0.8223, precision of 0.8235, recall of 0.8297, and an HD95 of 1.6807 mm. Compared to the state-of-the-art LVPA-UNet model, LG-UNet demonstrates significant improvements, with a 7.73% increase in DSC, a 4.52% increase in precision, a 3.40% increase in recall, and a reduction of 0.0796 mm in HD95. Ablation experiments confirm that the model’s performance progressively improves with the integration of GFF (Global Feature Fusion) and DLCK (Dynamic Large Convolutional Kernel).
In comparison to other representative segmentation networks, LG-UNet exhibits superior performance in both segmentation results and accuracy, particularly in capturing edge details and precise delineation of tumor GTV. Deep learning models excel at extracting complex, disease-relevant features from medical images, which are often challenging to discern through traditional visual inspection. The survival prediction model built using the segmentation findings from LG-UNet has a five-year average AUC of 0.789 and a C-index of 0.756. LG-UNet has more accurate prediction performance when compared to models that use segmentation data from other networks. These experiments validate that the incorporation of deep learning features enhances the precision of metastasis prediction, offering potential benefits for the development of personalized treatment strategies.
This study has several limitations that should be acknowledged:
(a) Recently, the Mamba neural network architecture [41,42,43,44,45] has emerged as a promising approach due to its flexibility, enhanced feature representation capabilities, and computational efficiency. By incorporating hardware-aware parallelization design, this architecture achieves efficient modeling of long-range spatial dependencies while maintaining linear computational complexity, representing a significant breakthrough in architectural design. In the future, the integration of this architecture could serve as a potential solution for addressing challenges in medical image segmentation and research on distant metastasis.
(b) LG-UNet incorporates a dynamic large convolutional kernel mechanism and global information fusion at the decoder end, primarily optimizing skip connections and upsampling stages. Due to limited computational resources, the size and number of convolutional kernels in the large kernel mechanism were constrained. In the future, leveraging greater computational resources, dynamic large convolutional kernels or more sophisticated convolutional and fusion strategies could be introduced at the encoder end. Such improvements would enable more comprehensive utilization of image features and enhance the model’s ability to focus on critical features.
(c) The multi-modality MRI dataset used in this study is relatively limited in size, and a small portion of the corresponding clinical information is missing. These factors may constrain the accuracy of the statistical analysis for survival prediction. Additionally, the Lung-Net model employed in the prediction phase has limited capability in mapping deep features. In the future, the proposed methodology could be further validated using more extensive datasets and more advanced deep learning models.
(d) This study evaluated segmentation performance using DSC, HD95, precision, and recall. While additional metrics such as Surface Dice, ASSD, volumetric similarity, and boundary F-score, as well as stratified analyses by tumor size, contrast use, and imaging artifacts, could provide deeper insights, these were beyond the current scope and will be explored in future work.

6. Conclusions

An improved LG-UNet model is presented in this work for accurate GTV segmentation in nasopharyngeal carcinoma (NPC). A Global Feature Fusion (GFF) module optimizes skip connections and minimizes information loss during upsampling, while a Dynamic Large Convolution Kernel (DLCK) module broadens the receptive field and improves contextual feature representation. Furthermore, to handle spatial anisotropy and combine multimodal MRI data, a 2.5 D parallel processing approach is applied.
The LungNet model was modified for task-specific adjustments in terms of survival prediction. Using a multi-layer perceptron, high-dimensional characteristics were taken out of the LG-UNet segmentation findings and transformed into quantifiable risk coefficients. For the purpose of analyzing survival, these risk coefficients were input into a Cox proportional hazards model together with clinical factors. ROC curves and the Youden index were used to determine the ideal risk threshold, and patients were then split into high- and low-risk groups. Survival differences were evaluated using Log-Rank tests and Kaplan–Meier curves, confirming the model’s efficacy and potential for clinical use in predicting metastasis risk.

Author Contributions

Conceptualization, Y.Z.; methodology, Y.Y. and J.W.; validation, Y.Y. and J.W.; formal analysis, J.W.; investigation, T.W. and J.D.; resources, Y.Z.; data curation, J.W.; writing—original draft preparation, J.W. and Y.Y.; writing—review and editing, Y.Z. and Y.X.; visualization, J.W. and J.D.; supervision, Y.Z. and Y.X.; project administration, Y.X. and Y.Z.; funding acquisition, Y.X. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the 2024 Key Research and Development Project (Industry University Research Cooperation Project) and the Development of Intelligent Algorithms and Systems Based on Machine Vision (SL2024B01F00004).

Informed Consent Statement

This study received ethics approval from the Ethics Committee of Sun Yat-sen University Cancer Center, approval number B2021-229-02. This study is a retrospective research project that analyzes medical records acquired from past diagnoses and treatments. The privacy and personal identity information of the subjects will be safeguarded. Consequently, this study is exempt from obtaining informed consent from patients.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing does not apply to this article.

Acknowledgments

We would thank the reviewers for their valuable feedback, which helped improve the manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
LVPALayer Volume Parallel Attention
CNNConvolutional Neural Network
NPCNasopharyngeal Carcinoma
AJCC/UICCAmerican Joint Committee on Cancer/Union for International Cancer Control
MRIMagnetic Resonance Imaging
DMFSDistant Metastasis Free Survival
LDHLactate Dehydrogenase
EBV DNAEpstein–Barr Virus Deoxyribonucleic Acid
DSCDice Similarity Coefficient
C-IndexConcordance Index
TD-ROCTime-Dependent Receiver Operating Characteristic
DLCKDynamic Large Convolutional Kernel
GFFGlobal Feature Fusion
GELUGaussian Error Linear Unit
DWConvDepthwise Convolution

Appendix A

Figure A1. Training and validation loss curves of LG-UNet on the NPC MRI dataset.training/validation loss curves.
Figure A1. Training and validation loss curves of LG-UNet on the NPC MRI dataset.training/validation loss curves.
Bioengineering 12 01051 g0a1

References

  1. De Martel, C.; Georges, D.; Bray, F.; Ferlay, J.; Clifford, G.M. Global burden of cancer attributable to infections in 2018: A worldwide incidence analysis. Lancet Glob. Health 2020, 8, e180–e190. [Google Scholar] [CrossRef]
  2. Chen, Y.-P.; Chan, A.T.C.; Le, Q.-T.; Blanchard, P.; Sun, Y.; Ma, J. Nasopharyngeal carcinoma. Lancet 2019, 394, 64–80. [Google Scholar] [CrossRef]
  3. Isensee, F.; Jaeger, P.F.; Kohl, S.A.A.; Petersen, J.; Maier-Hein, K.H. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat. Methods 2021, 18, 203–211. [Google Scholar] [CrossRef]
  4. Tang, P.; Zu, C.; Hong, M.; Yan, R.; Peng, X.; Xiao, J.; Wu, X.; Zhou, J.; Zhou, L.; Wang, Y. Da-dsunet: Dual attention-based dense su-net for automatic head-and-neck tumor segmentation in mri images. Neurocomputing 2021, 435, 103–113. [Google Scholar] [CrossRef]
  5. Huang, H.; Lin, L.; Tong, R.; Hu, H.; Zhang, Q.; Iwamoto, Y.; Han, X.; Chen, Y.-W.; Wu, J. UNet 3+: A Full-Scale Connected UNet for Medical Image Segmentation. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [Google Scholar] [CrossRef]
  6. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. arXiv 2018. [Google Scholar] [CrossRef]
  7. Qin, X.; Zhang, Z.; Huang, C.; Dehghan, M.; Zaiane, O.R.; Jagersand, M. U2-Net: Going deeper with nested U-structure for salient object detection. Pattern Recognit. 2020, 106, 107404. [Google Scholar] [CrossRef]
  8. Li, Y.; Dan, T.; Li, H.; Chen, J.; Peng, H.; Liu, L.; Cai, H. Npcnet: Jointly segment primary nasopharyngeal carcinoma tumors and metastatic lymph nodes in mr images. IEEE Trans. Med. Imaging 2022, 41, 1639–1650. [Google Scholar] [CrossRef] [PubMed]
  9. Hatamizadeh, A.; Nath, V.; Tang, Y.; Yang, D.; Roth, H.R.; Xu, D. Swin unetr: Swin transformers for semantic segmentation of brain tumors in mri images. In International MICCAI Brainlesion Workshop; Springer: Berlin/Heidelberg, Germany, 2021; pp. 272–284. [Google Scholar]
  10. Hatamizadeh, A.; Xu, Z.; Yang, D.; Li, W.; Roth, H.; Xu, D. UNetFormer: A Unified Vision Transformer Model and Pre-Training Framework for 3D Medical Image Segmentation. arXiv 2022. [Google Scholar] [CrossRef]
  11. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar] [CrossRef]
  12. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation. arXiv 2021. [Google Scholar] [CrossRef]
  13. Wang, W.; Chen, C.; Ding, M.; Yu, H.; Zha, S.; Li, J. Transbts: Multimodal Brain Tumor Segmentation Using Transformer; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
  14. Chen, S.; Luo, C.; Liu, S.; Li, H.; Liu, Y.; Zhou, H.; Liu, L.; Chen, H. Ld-unet: A long-distance perceptual model for segmentation of blurred boundaries in medical images. Comput. Biol. Med. 2024, 171, 108120. [Google Scholar] [CrossRef] [PubMed]
  15. Chowdary, G.J.A.Y. Diffusion transformer u-net for medical image segmentation. In Proceedings of the 26th International Conference, Vancouver, BC, Canada, 8–12 October 2023. [Google Scholar]
  16. Azad, R.; Niggemeier, L.; Hüttemann, M.; Kazerouni, A.; Aghdam, E.K.; Velichko, Y.; Bagci, U.; Merhof, D. Beyond self-attention: Deformable large kernel attention for medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2024; pp. 1287–1297. [Google Scholar]
  17. Lee, H.H.; Bao, S.; Huo, Y.; Landman, B.A. 3d ux-net: A large kernel volumetric convnet modernizing hierarchical transformer for medical image segmentation. arXiv 2022, arXiv:2209.15076. [Google Scholar]
  18. Ding, X.; Zhang, X.; Zhou, Y.; Han, J.; Ding, G.; Sun, J. Scaling Up Your Kernels to 31x31: Revisiting Large Kernel Design in CNNs. arXiv 2022. [Google Scholar] [CrossRef]
  19. Chen, H.; Chu, X.; Ren, Y.; Zhao, X.; Huang, K. PeLK: Parameter-efficient Large Kernel ConvNets with Peripheral Convolution. arXiv 2024. [Google Scholar] [CrossRef]
  20. Ding, X. UniRepLKNet: A Universal Perception Large-Kernel ConvNet for Audio, Video, Point Cloud, Time-Series and Image Recognition. arXiv 2023, arXiv:2311.15599. [Google Scholar] [CrossRef]
  21. Saeed, N.; Sobirov, I.; Al Majzoub, R.; Yaqub, M. TMSS: An End-to-End Transformer-Based Multimodal Network for Segmentation and Survival Prediction. In Medical Image Computing and Computer Assisted Intervention—MICCAI 2022. MICCAI 2022; Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S., Eds.; Lecture Notes in Computer Science; Springer: Cham, Switzerland, 2022; Volume 13437. [Google Scholar] [CrossRef]
  22. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  23. Wang, W.; Li, S.; Shao, J.; Jumahong, H. LKC-Net: Large kernel convolution object detection network. Sci. Rep. 2023, 13, 9535. [Google Scholar] [CrossRef]
  24. Cai, J.; Tang, Y.; Lu, L.; Harrison, A.P.; Yan, K.; Xiao, J.; Yang, L.; Summers, R.M. Accurate weakly-supervised deep lesion segmentation using large-scale clinical annotations: Slice-propagated 3d mask generation from 2d recist. In Proceedings of the Medical Image Computing and Computer Assisted Intervention–MICCAI 2018: 21st International Conference, Granada, Spain, 16–20 September 2018; Proceedings, Part IV 11. Springer: Berlin/Heidelberg, Germany, 2018; pp. 396–404. [Google Scholar]
  25. Poudel, R.P.K.; Lamata, P.; Montana, G. Recurrent fully convolutional neural networks for multi-slice MRI cardiac segmentation. In Proceedings of the Reconstruction, Segmentation, and Analysis of Medical Images: First International Workshops, RAMBO 2016 and HVSMR 2016, Held in Conjunction with MICCAI 2016, Athens, Greece, 17 October 2016; Revised Selected Papers 1. Springer: Berlin/Heidelberg, Germany, 2017; pp. 83–94. [Google Scholar]
  26. Li, J.; Chen, J.; Tang, Y.; Wang, C.; Landman, B.A.; Zhou, S.K. Transforming medical imaging with Transformers? A comparative review of key properties, current progresses, and future perspectives. Med. Image Anal. 2023, 85, 102762. [Google Scholar] [CrossRef]
  27. Bilic, P.; Christ, P.; Li, H.B.; Vorontsov, E.; Ben-Cohen, A.; Kaissis, G.; Szeskin, A.; Jacobs, C.; Mamani, G.E.H.; Chartrand, G.; et al. The liver tumor segmentation benchmark (lits). Med. Image Anal. 2023, 84, 102680. [Google Scholar] [CrossRef] [PubMed]
  28. Xia, F.; Peng, Y.; Wang, J.; Chen, X. A 2.5D multi-path fusion network framework with focusing on z-axis 3D joint for medical image segmentation. Biomed. Signal Process. Control. 2024, 91, 106049. [Google Scholar] [CrossRef]
  29. Zhao, J.; Xing, Z.; Chen, Z.; Wan, L.; Han, T.; Fu, H.; Zhu, L. Uncertainty-Aware Multi-Dimensional Mutual Learning for Brain and Brain Tumor Segmentation. IEEE J. Biomed. Health Inform. 2023, 27, 43624372. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Xu, H.R.; Wen, J.H.; Hu, Y.J.; Diao, Y.L.; Chen, J.L.; Xia, Y.F. A novel LVPA-UNet network for target volume automatic delineation: An MRI case study of nasopharyngeal carcinoma. Heliyon 2024, 10, e30763. [Google Scholar] [CrossRef] [PubMed] [PubMed Central]
  31. Yang, J.; Qiu, P.; Zhang, Y.; Marcus, D.S.; Sotiras, A. D-Net: Dynamic Large Kernel with Dynamic Feature Fusion for Volumetric Medical Image Segmentation. arXiv 2024, arXiv:2403.10674. [Google Scholar] [CrossRef]
  32. Zhang, L.; Wu, X.; Liu, J.; Zhang, B.; Mo, X.; Chen, Q.; Fang, J.; Wang, F.; Li, M.; Chen, Z.; et al. Mri-based deep-learning model for distant metastasis-free survival in locoregionally advanced nasopharyngeal carcinoma. J. Magn. Reson. Imaging 2021, 53, 167–178. [Google Scholar] [CrossRef]
  33. Huang, S.H.; O’Sullivan, B. Overview of the 8th edition TNM classification for head and neck cancer. Curr. Treat. Options Oncol. 2017, 18, 40. [Google Scholar] [CrossRef]
  34. Hui, E.; Li, W.; Ma, B.; Lam, W.; Chan, K.; Mo, F.; Ai, Q.; King, A.; Wong, C.; Guo, R.; et al. Integrating postradiotherapy plasma epstein-barr virus DNA and TNM stage for risk stratification of nasopharyngeal carcinoma to adjuvant therapy. Ann. Oncol. 2020, 31, 769–779. [Google Scholar] [CrossRef]
  35. Demirjian, N.L.; Varghese, B.A.; Cen, S.Y.; Hwang, D.H.; Aron, M.; Siddiqui, I.; Fields, B.K.K.; Lei, X.; Yap, F.Y.; Rivas, M.; et al. CT-based radiomics stratification of tumor grade and TNM stage of clear cell renal cell carcinoma. Eur. Radiol. 2022, 32, 2552–2563. [Google Scholar] [CrossRef] [PubMed]
  36. Gu, B.; Meng, M.; Xu, M.; Feng, D.D.; Bi, L.; Kim, J.; Song, S. Multi-task deep learning-based radiomic nomogram for prognostic prediction in locoregionally advanced nasopharyngeal carcinoma. Eur. J. Nucl. Med. Mol. Imaging 2023, 50, 3996–4009. [Google Scholar] [CrossRef]
  37. Delfan, N.; Abrishami Moghaddam, H.; Modaresi, M.; Afshari, K.; Nezamabadi, K.; Pak, N.; Ghaemi, O.; Forouzanfar, M. CT-LungNet: A Deep Learning Framework for Precise Lung Tissue Segmentation in 3D Thoracic CT Scans. arXiv 2022, arXiv:2212.13971. [Google Scholar] [CrossRef]
  38. Tibshirani, R. The lasso method for variable selection in the cox model. Stat. Med. 1997, 16, 38595. [Google Scholar] [CrossRef]
  39. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar] [CrossRef]
  40. Zhang, H.; Zhu, Y.; Wang, D.; Zhang, L.; Chen, T.; Wang, Z.; Ye, Z. A survey on visual mamba. Appl. Sci. 2024, 14, 5683. [Google Scholar] [CrossRef]
  41. Xu, R.; Yang, S.; Wang, Y.; Du, B.; Chen, H. A survey on vision mamba: Models, applications and challenges. arXiv 2024, arXiv:2404.18861. [Google Scholar] [CrossRef]
  42. Liu, X.; Zhang, C.; Zhang, L. Vision mamba: A comprehensive survey and taxonomy. arXiv 2024, arXiv:2405.04404. [Google Scholar] [CrossRef]
  43. Li, J.; Wang, W.; Chen, C.; Zhang, T.; Zha, S.; Wang, J.; Yu, H. Transbtsv2: Towards better and more efficient volumetric segmentation of medical images. arXiv 2022, arXiv:2201.12785. [Google Scholar] [CrossRef]
  44. Zhang, Y.; Liao, Q.; Ding, L.; Zhang, J. Bridging 2d and 3d segmentation networks for computation-efficient volumetric medical image segmentation: An empirical study of 2.5 d solutions. Comput. Med. Imaging Graph. 2022, 99, 102088. [Google Scholar] [CrossRef] [PubMed]
  45. Zheng, H.; Lin, Z.; Zhou, Q.; Peng, X.; Xiao, J.; Zu, C.; Jiao, Z.; Wang, Y. Multi-Transsp: Multimodal Transformer for Survival Prediction of Nasopharyngeal Carcinoma Patients; Springer: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
Figure 1. The proposed overall architecture of LG-UNet. Abbreviation: LVPA = Layer-Volume Parallel Attention.
Figure 1. The proposed overall architecture of LG-UNet. Abbreviation: LVPA = Layer-Volume Parallel Attention.
Bioengineering 12 01051 g001
Figure 2. The architecture of the proposed GDU module in detail. Abbreviation: Conv = convolution; RELU = rectified linear unit; Group Norm = group normalization.
Figure 2. The architecture of the proposed GDU module in detail. Abbreviation: Conv = convolution; RELU = rectified linear unit; Group Norm = group normalization.
Bioengineering 12 01051 g002
Figure 3. Detailed architecture of the Global Feature Fusion (GFF) module, illustrating the process of integrating multi-scale features through skip connections and feature upsampling techniques.
Figure 3. Detailed architecture of the Global Feature Fusion (GFF) module, illustrating the process of integrating multi-scale features through skip connections and feature upsampling techniques.
Bioengineering 12 01051 g003
Figure 4. Detailed architecture of the Dynamic Large Conventional Kernel (DLCK) module, illustrating the process of enhancing model performance through dynamically adjusting large-sized convolution kernels and integrating multi-scale feature extraction techniques.
Figure 4. Detailed architecture of the Dynamic Large Conventional Kernel (DLCK) module, illustrating the process of enhancing model performance through dynamically adjusting large-sized convolution kernels and integrating multi-scale feature extraction techniques.
Bioengineering 12 01051 g004
Figure 5. Flowchart of patient enrollment. NPC = nasopharyngeal carcinoma, AJCC/UICC = American Joint Committee on Cancer/Union for International Cancer.
Figure 5. Flowchart of patient enrollment. NPC = nasopharyngeal carcinoma, AJCC/UICC = American Joint Committee on Cancer/Union for International Cancer.
Bioengineering 12 01051 g005
Figure 6. Real segmentation examples comparing LG-UNet with five other representative models. Sub-figure (b) provides an enlarged view of the region outlined in (a).
Figure 6. Real segmentation examples comparing LG-UNet with five other representative models. Sub-figure (b) provides an enlarged view of the region outlined in (a).
Bioengineering 12 01051 g006aBioengineering 12 01051 g006b
Figure 7. Real segmentation examples from the ablation experiments conducted on the individual modules of LG-UNet.
Figure 7. Real segmentation examples from the ablation experiments conducted on the individual modules of LG-UNet.
Bioengineering 12 01051 g007
Figure 8. Comparison of TD-ROC curves for different models.
Figure 8. Comparison of TD-ROC curves for different models.
Bioengineering 12 01051 g008
Figure 9. Comparison of Kaplan–Meier curves for different combination models.
Figure 9. Comparison of Kaplan–Meier curves for different combination models.
Bioengineering 12 01051 g009
Table 1. Statistical information of the dataset.
Table 1. Statistical information of the dataset.
CharacteristicsDataset (n = 388)
Age (years), median (IQR) 644 (38.51)
Sex, No (%)
Male271 (69.9)
Female117 (30.1)
Histological type (WHO) 6, No (%)
WHO I0
WHO II20 (5.2)
WHO III368 (94.8)
T a, No (%)
T1197 (50.8)
T2191 (49.2)
N a, No (%)
N048 (12.4)
N1340 (87.6)
LDH b 6, No (%)
Normal369 (95.1)
Abnormal19 (4.9)
Unknown0
EBV DNA c 6, No (%)
Undetectable208 (53.6)
Detectable161 (41.5)
Unknown 19 (4.9)
DMFS 6, No (%)
Metastasis-free353 (91.0)
Metastasis35 (9.0)
a According to the 8th edition of the American Joint Committee on Cancer/Union for international Cancer Control cancer staging manual. b Abnormal, center 1: >245 U/L. c Detectable thresholds, center 1: <1000 copy/mL. d Data are represented as median (IQR). 6 Abbreviations: IQR: interquartile range; WHO, World Health Organization; LDH, lactate dehydrogenase; EBV DNA, Epstein–Barr virus deoxyribonucleic acid; DMFS, distant metastasis-free survival.
Table 2. Comparison of evaluation metrics of various typical models.
Table 2. Comparison of evaluation metrics of various typical models.
ModelsDSC (%) ↑HD95 (mm) ↓Precision (%) ↑Recall (%) ↑
3D UNet76.31 ± 7.29 2.26 ± 1.41 76.88 ± 11.0577.68 ± 10.74
TransBTS77.85 ± 6.40 2.45 ± 1.92 78.45 ± 9.0178.65 ± 10.10
UNETR78.76 ± 6.872.19 ± 1.54 77.94 ± 8.84 79.71 ± 8.43
VT-UNet78.67 ± 7.83 2.04 ± 2.1378.36 ± 10.48 80.52 ± 10.49
LVPA-UNet80.09 ± 7.251.76 ± 1.0279.06 ± 10.4682.29 ± 8.73
LG-UNet82.22 ± 6.471.68 ± 0.9782.36 ± 7.9682.97 ± 9.08
Table 3. Comparison of evaluation metrics of various typical models.
Table 3. Comparison of evaluation metrics of various typical models.
ModelsDSC (%) ↑HD95 (mm) ↓Precision (%) ↑Recall (%) ↑
LVPA-UNet (U)80.09 ± 7.251.76 ± 1.0279.06 ± 10.4682.29 ± 8.73
DU79.31 ± 6.832.11 ± 2.5579.02 ± 10.7880.65 ± 10.51
GU80.42 ± 6.891.75 ± 0.9681.04 ± 9.8181.01 ± 9.35
GDU82.22 ± 6.471.68 ± 0.9782.36 ± 7.9682.97 ± 9.08
Table 4. Comparison of evaluation metrics of various typical models.
Table 4. Comparison of evaluation metrics of various typical models.
Input ChannelsDSC (%) ↑HD95 (mm) ↓Precision (%) ↑Recall (%) ↑
T177.18 ± 6.311.89 ± 1.0279.05 ± 7.1080.21 ± 6.55
T1C80.05 ± 6.551.81 ± 0.9680.10 ± 7.5181.33 ± 7.48
T278.76 ± 6.711.93 ± 1.1378.60 ± 6.6879.94 ± 6.61
T1 + T1C + T282.22 ± 6.471.68 ± 0.9782.36 ± 7.9682.97 ± 9.08
Table 5. AUC and C-index values of different combined models.
Table 5. AUC and C-index values of different combined models.
ModelsAUC (95%CI) C-Index (95%CI)
CoxPH (3D UNet + Clinical)0.625 (0.571–0.679)0.584 (0.528–0.641)
CoxPH (TransBTS + Clinical)0.762 (0.715–0.807)0.726 (0.678–0.775)
CoxPH (UNETR + Clinical)0.769 (0.725–0.812)0.716 (0.667–0.784)
CoxPH (VT-UNet + Clinical)0.703 (0.652–0.753)0.694 (0.673–0.762)
CoxPH (LVPA-UNet + Clinical)0.776 (0.728–0.820)0.734 (0.687–0.781)
CoxPH (LG-UNet + Clinical)0.789 (0.746–0.829)0.756 (0.699–0.804)
CoxPH (Clinical)0.640 (0.589–0.694)0.636 (0.577–0.684)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, Y.; Wen, J.; Wu, T.; Dong, J.; Xia, Y.; Zhang, Y. LG-UNet Based Segmentation and Survival Prediction of Nasopharyngeal Carcinoma Using Multimodal MRI Imaging. Bioengineering 2025, 12, 1051. https://doi.org/10.3390/bioengineering12101051

AMA Style

Yang Y, Wen J, Wu T, Dong J, Xia Y, Zhang Y. LG-UNet Based Segmentation and Survival Prediction of Nasopharyngeal Carcinoma Using Multimodal MRI Imaging. Bioengineering. 2025; 12(10):1051. https://doi.org/10.3390/bioengineering12101051

Chicago/Turabian Style

Yang, Yuhao, Junhao Wen, Tianyi Wu, Jinrang Dong, Yunfei Xia, and Yu Zhang. 2025. "LG-UNet Based Segmentation and Survival Prediction of Nasopharyngeal Carcinoma Using Multimodal MRI Imaging" Bioengineering 12, no. 10: 1051. https://doi.org/10.3390/bioengineering12101051

APA Style

Yang, Y., Wen, J., Wu, T., Dong, J., Xia, Y., & Zhang, Y. (2025). LG-UNet Based Segmentation and Survival Prediction of Nasopharyngeal Carcinoma Using Multimodal MRI Imaging. Bioengineering, 12(10), 1051. https://doi.org/10.3390/bioengineering12101051

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop