Next Article in Journal
ST-CRMF: Compensated Residual Matrix Factorization with Spatial-Temporal Regularization for Graph-Based Time Series Forecasting
Next Article in Special Issue
Impact of Anatomical Variability on Sensitivity Profile in fNIRS–MRI Integration
Previous Article in Journal
Heuristic Method for Minimizing Model Size of CNN by Combining Multiple Pruning Techniques
Previous Article in Special Issue
CAFS: An Attention-Based Co-Segmentation Semi-Supervised Method for Nasopharyngeal Carcinoma Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Lightweight Compound Scaling Network for Nasopharyngeal Carcinoma Segmentation from MR Images

1
School of Biomedical Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen 518107, China
2
Sun Yat-sen University, Guangzhou 510275, China
3
School of Information Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450046, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(15), 5875; https://doi.org/10.3390/s22155875
Submission received: 30 May 2022 / Revised: 23 July 2022 / Accepted: 30 July 2022 / Published: 5 August 2022

Abstract

:
Nasopharyngeal carcinoma (NPC) is a category of tumours with a high incidence in head-and-neck. To treat nasopharyngeal cancer, doctors invariably need to perform focal segmentation. However, manual segmentation is time consuming and laborious for doctors and the existing automatic segmentation methods require large computing resources, which makes some small and medium-sized hospitals unaffordable. To enable small and medium-sized hospitals with limited computational resources to run the model smoothly and improve the accuracy of structure, we propose a new LW-UNet network. The network utilises lightweight modules to form the Compound Scaling Encoder and combines the benefits of UNet to make the model both lightweight and accurate. Our model achieves a high accuracy with a Dice coefficient value of 0.813 with 3.55 M parameters and 7.51 G of FLOPs within 0.1 s (testing time in GPU), which is the best result compared with four other state-of-the-art models.

1. Introduction

Among cancers of the head and neck [1], nasopharyngeal carcinoma is one of the most common types [2]. Nasopharyngeal carcinoma (NPC) is a highly invasive neoplasia that spreads early to regional lymph nodes. NPC is common in southern China, the Middle East, and North Africa. It has significant geographical variation and gender differences [3]. NPC has the highest incidence in Southeast Asia in up to 6.4/100,000 males and 2.4/100,000 females in these regions [4]. In 2012, the mortality of nasopharyngeal carcinoma (NPC) reached 58.6%, and in 2018 there were about 130,000 incidents in the world, including more than 73,000 people dying from nasopharyngeal carcinoma [5].
Nasopharyngeal carcinoma is mainly treated with radiotherapy. Currently, the delineation of the tumour area in radiotherapy is generally performed by manual segmentation. Doctors segment the MR images of nasopharyngeal carcinoma to match the target area and tumour extent as accurately as possible, so that radiotherapy can obtain better effects. However, manual segmentation costs a lot of time and energy for doctors [6], and the accuracy of manual segmentation is also affected by the experience of doctors. Although some automatic segmentation models have emerged to assist physicians in nasopharyngeal carcinoma segmentation, the existing models have a large number of parameters and consume many computing resources. Small and medium-sized hospitals cannot afford the large resource consumption and therefore still use manual segmentation for nasopharyngeal carcinoma treatment. To solve the above problems caused by manual segmentation, we proposed a lightweight automatic segmentation method for nasopharyngeal carcinoma.
With the advancement of convolutional neural networks, the research on medical image processing based on semantic segmentation has developed rapidly. Semantic segmentation is divided into two categories. Traditional image segmentation uses grayscale, color, texture, shape, and other features to divide the image into regions so that there are obvious differences and similarities between regions. Tatanun et al. [7] and Huang et al. [8] used the threshold method, region growth method, statistical theory, and other traditional image segmentation methods to segment nasopharyngeal carcinoma tumours. In addition, machine learning methods such as SVM [9,10,11] and SOM [12] are also used to segment nasopharyngeal carcinoma tumours. However, the traditional methods mentioned above usually require manual intervention processes such as feature extraction and dimension reduction; this has disadvantages such as poor model robustness and noise sensitivity.
The emergence of deep learning-based methods has led to significant changes in the approach to the field of computer vision [13,14,15,16,17,18,19,20,21]. Deep learning is used in a wide variety of tasks in computer vision and medicine [22,23,24,25,26], where neural networks consisting of an encoder-decoder framework have become one of the dominant models in deep learning. Ji et al. [27] reviewed a CNN-based encoder-decoder framework for a significant object detection neural network in recent years, showing its achievements and great potential in the field of salient object detection. Chen et al. [28] built PAD-Net by using an encoder-decoder framework, which has achieved great success in the field of stereoscopic image quality measurement (SIQM). In addition, the encoder-decoder framework has also achieved success in damage detection [29], scene independent evaluation [30], material capture [31], handwriting recognition [32], etc. The automatic segmentation method of nasopharyngeal carcinoma is developed based on the second method to achieve the effect of assisting doctors in treatment and diagnosis.
However, the existing automatic segmentation methods for nasopharyngeal carcinoma face three challenges: First, nasopharyngeal carcinoma is adjacent to a few normal tissues and even infiltrates, such as mucosa, and its intensity range is almost the same as that of nasopharyngeal carcinoma [33]. In addition, the shape and size of lesions vary from patient to patient and the blurred boundaries of head and neck tumours on MRI are also a pain point (as shown in Figure 1) [8]. Second, unlike other medical image processing tasks that have a large number of similar images and have the same image quality, for the same NPC patient, the complicated shapes and remote location of the tumour may create more difficulties for segmentation. Third, the different hospitals provide different operating environment resources for the model. Some hospitals can only provide limited resources, so it is unavailable for them to run larger models.
Currently some classical convolutional neural networks are proposed and used for medical image segmentation of nasopharyngeal carcinoma, such as VGGNet [34], ResNet [35], FCN [36], SegNet [37]. In 2015, Olaf proposed a U-shaped network framework named UNet [38], it has greatly influenced the field of semantic segmentation, especially the field of medical image segmentation. Since UNet [38] can combine low-resolution information, which is conducive to the recognition of targets, and at the same time combines high-resolution information, it solves the disadvantages of blurred boundaries and complex gradients in medical images. After that, many UNet-based convolutional neural networks were proposed, such as Att-UNet [39], AttR2U-Net [40], Res-UNet [35] and DA-DSUNet [41] etc.
In addition to the UNet-based method, many other deep learning methods have been proposed and used for nasopharyngeal carcinoma medical image segmentation tasks. Li et al. [42] used the full convolution encoding-decoding neural network of 27 patients with NPC MR image segmentation and adopted it to realize the NPC automatic segment. Ma et al. [43] used a convolutional neural network to segment the focal regions of T1W mode MR images of 30 NPC patients, improved the segmentation results with a 3D image cutting algorithm, and finally achieved a relatively good result. However, the networks of the above two methods have the disadvantages of a simple structure, fewer experimental data, and the generality and accuracy of the model need to be improved. Moreover, the deployment of the network is too troublesome and not light enough for small and medium-sized hospitals.
In this paper, we propose a lightweight model named LW-UNet to solve the problems of the resource-constrained situation of small and medium-sized hospitals. We build the Compound Scaling Encoder inspired by EfficientNet [44]; this encoder improves the accuracy and reduces the parameters of nasopharyngeal carcinoma segmentation by compound scaling depth, width, and resolution. It enables small and medium-sized hospitals to run smoothly under resource constraints. The decoder is similar to UNet. Compared with other NPC segmentation models, our model has the following two advantages:
(1) Our model uses a fixed mixing coefficient to uniformly scale depth, width, and resolution to improve the accuracy and phase rate of the network;
(2) Our model uses lightweight modules to enable it to run under resource constraints.
The rest of the paper is structured as follows. In Section 2, we will introduce our proposed method. The dataset used in this research and the experimental details are described in Section 3. The results are discussed in Section 4.

2. Method

LW-UNet is a UNet-like model; the detailed network is shown in Figure 2. Our model consists of two parts: Compound Scaling Encoder and UNet-like decoder. Inspired by EfficientNet [44], we propose the Compound Scaling Encoder. It introduces scale coefficient to scale width, depth and resolution to reduce network parameters and FLOPs (See Section 2.1). We refer to UNet [38] and propose a decoder with a similar structure (See Section 2.2). Firstly our model utilizes the Compound Scaling Encoder to capture image information to obtain the feature maps. Then, the feature maps of the encoder are upsampled by up-convolution. The output of the 2, 3, 5, 6, 7 blocks of the encoder and up-convolution are concatenated by a skip connection. Finally, as the decoder continues to recover image resolution and detail, we obtain the segmentation maps with the same resolution as the original image.

2.1. Compound Scaling Encoder

The development of the convolutional neural network is usually based on a fixed resource budget [44]. If more resources are available, the size of the network can be expanded. A traditional CNN network typically scales the depth and width of the network and the size of the input images resolution to improve the accuracy, such as ResNet [35], GPipe [45]. However, the traditional method generally scales the width, depth and resolution randomly. It is difficult to adjust with limited resources.
In 2020, Tan et al. [44] proposed a new model scaling method, which utilizes a coefficient to uniformly scale the depth, width and resolution, rather than scale the single dimension of the network as in the traditional method. Tan introduced α , β , γ to measure the specific weights of depth, width, resolution, and θ as a scaling coefficient. He considers that if depth is doubled, the amount of calculation is doubled. However, if width or resolution are doubled, then the amount of calculation increases by four times. In other words, the amount of calculation is proportional to depth, width, and resolution. By referring to the method of Tan et al. [44], we utilize a compound scaling coefficient to uniformly scale the width, depth, and the resolution of the input image of the network. The compound scaling coefficients are utilized as in Equation (1):
D = k i × d W = L i × ω R = r × γ ,
where D, W, R represent the depth, width and resolution, k i , L i , r represent the i-th block kernel size, the i-th layer and the input size resolution respectively, and d, ω , γ represent the scaling coefficient. By changing the scaling coefficient, models with different depth, width, and resolution are obtained. We refer to the method in EfficientNet [44] and use NAS (neural architecture search) to search for the best parameters corresponding to depth, width, and resolution, respectively, and fix them to obtain the baseline model LW-UNet 0. On this basis, the fixed parameters are scaled uniformly to obtain the scaling coefficients shown in Table 1. The fixed mixing coefficients determine the number of MBConv modules per block within the composite scaling encoder (as shown in Figure 2) in the baseline model (LW-UNet 0). With the scaling coefficients, we vary the number of MBConv modules in each block to obtain LW-UNet 1–5.
In addition to the compound coefficient for depth, width, and resolution, the Compound Scaling Encoder utilizes the Mobile Inverted Bottleneck Convolution module to make the model more lightweight. MBConv [46] is a lightweight module designed specifically for resource-constrained environments. Unlike normal convolution modules, MBConv utilizes depthwise separable convolution. Standard convolution takes a h i × w i × d i input tensor L i , and applies convolutional kernel K R k × k × d i × d i to produce a h i × w i × d j output tensor L j . So the standard convolutional layer has the computational cost of h i · w i · d i · d j · k · k . Compared with normal convolution, depthwise separable convolutions have almost the same effect but only cost h i · w i · d i ( k 2 + d j ) .
MBConv also constructs the Bottleneck Residual Block by using a linear bottleneck and an inverted residual mechanism. First, in the layer with a small number of channels, the linear bottleneck replaces the ReLU activation layer with the linear transform, which reduces the large information spoilage caused by ReLU. Second, the inverted residuals extend the low-dimensional input to higher dimensions by pointwise, and then subsequently proceed to depthwise. It can enhance the transfer of information and gradients without increasing the computational cost excessively. Our model utilizes MBConv modules of different sizes to constitute the encoder. The encoder composed of MBConv can reduce the computational resources efficiently.

2.2. UNet-like Decoder

UNet is proposed to solve the problem of medical image segmentation [38]. It is frequently used in the baseline of the medical image segmentation task because of its excellent performance. UNet consists of two parts: encoder and decoder. In the conventional UNet, the decoder is nearly symmetric to the encoder. However, unlike UNet, our network is asymmetric and the encoder is deeper than the decoder. The detailed network is revealed in Figure 2. In our work, we utilize the feature map of the last block of the encoder to upsample, then we concatenate with the same spatial resolution feature map from encoder. Through up-convolution and concatenation with a corresponding feature map from the encoder, the decoder can combine the spatial information. Before upsampling again, the image has to go through 3 × 3 convolutional layers first. This process is repeated until the segmentation map is the same size as the original input image.

2.3. Evaluation Method

2.3.1. Accuracy Evaluation

In the training process, we utilize binary cross-entropy loss to evaluate the performance of the model, as shown in Equation (2).
L o s s = 1 N ( y n × l o g ( z n ) + ( 1 y n ) × l o g ( 1 z n ) ) ,
where N represents the total number of pixels, Z n represents the probability of predicting the n-th sample as a positive example, and Y n represents the label of the n-th sample.
In the test process, we utilise the Dice coefficient as the metric to evaluate the accuracy of the model [47]. The Dice coefficient measures the consistency between the two regions, which is defined as Equation (3).
D S C = 2 T P 2 T P + F P + F N ,
where T P , F P , F N represent the number of true positive, false positive and false negative pixels respectively.
We also use IoU, Jaccard similarity [48], precision, specificity and sensitivity to evaluate the segmentation effect. We will discuss our model performance in Section 5.

2.3.2. Parameters Evaluation

Deep neural networks are widely used in machine vision tasks such as image classification and object detection with great success. However, due to the limitation of memory and computing power, deploying a neural network on embedded devices is still an enormous challenge. In order to evaluate the lightweight degree of the model, we use Pytorch-OpCounter to calculate the parameters and flops of the model compared with four other state-of-the-art models.

3. Experiments

3.1. Dataset Description

Our model uses the NPC dataset from a total of 92 patients diagnosed with nasopharyngeal carcinoma, which contain 735 MR images. All images are MR T1W+C images of head and neck scanned by the Siemens Aera MRI system (approximately 100 slices per patient). The resulting MR images are stored in Digital Imaging and Communications in Medicine (DICOM) file format. To utilize the raw data for 2D image segmentation, we cropped each image to the region of interest (ROI) to reduce unnecessary computing workload. The cropped image includes the nasopharyngeal carcinoma area and the rest of the head region. Given the low number of nasopharyngeal carcinoma data obtained, we applied data augmentation to the dataset. We used HorizontalFlip, ShiftScaleRotate and other methods to transform the original nasopharyngeal carcinoma data. Finally we obtained 3678 images. Eighty percent of the images were used for training, 10% for validation, and 10% for testing.

3.2. Data Preprocessing

Considering the difference in image quality of MR images captured by different imaging equipment, we normalized the data inputted to the model. The specific formula is as in Equation (4).
X o u t p u t = X i X m i n X m a x X m i n ,
where X o u t p u t is the normalised data, X i is the original data, X m a x and X m i n are the maximum and minimum values of the original data set respectively.
Several previous studies have shown that data normalisation can make the data distribution more uniform and make the model converge faster, resulting in an improved model performance [49,50]. Images after processing have a mean of 0 and a variance of 1, conforming to the standard normal distribution. It effectively solves the problem of different picture quality. After the data were normalised, we scaled our input images to 256 × 256 and randomly assigned them to the training, validation and test sets. Finally our model was trained with a hyperparameter value of 1 for batch_size.

3.3. Implementation Details

In the training process of LW-UNet, we used the Adam optimizer to implement the gradient descent method [51]. We also used LRfinder to find the best learning rate and used ReduceLROnPlateau provided by PyTorch to adjust the learning rate properly [52]. In order to solve the overfitting problem to some extent, we used dropout and other mechanisms. We trained the model with the shuffled nasopharyngeal carcinoma images with segmentation labels [53]. We promptly tested the model on the validation set at the end of each training epoch to adjust the hyper-parameter.

4. Result

4.1. Ablation Study

To verify the effectiveness of our model, we perform the ablation study by comparing four variants of our model and the baseline UNet. The results in Table 2 demonstrate the efficiency and accuracy of our model. It proves that our structure is effective for nasopharyngeal carcinoma segmentation. The test results of our model are as follows: The average Dice coefficient value obtained from the test is 0.813, the average Jaccard similarity is 0.695, the average IoU is 0.696, the average specificity is 0.998 , the average precision is 0.787 and the average sensitivity is 0.824. We select four models LW-UNet 0, LW-UNet 1, LW-UNet 2 and LW-UNet 3 for comparison with UNet [38]. The results show that with the increasing width, depth and resolution of the model, the accuracy of the model increases continuously. Compared with LW-UNet 4, the DSC, JC, and IoU values of LW-UNet 3 are 0.02, 0.04, and 0.02 lower than those of LW-UNet 4, respectively, and the SE values are 0.1 higher than those of LW-UNet 4. Compared with LW-UNet 5, the DSC, JC, IoU, and SE values of LW-UNet 3 are respectively higher by 0.09, 0.06, 0.08, and 0.04. The LW-UNet 3 we use is superior to the LW-UNet 0, 1, 2 and 5 in terms of accuracy. The values of LW-UNet 4 are essentially the same as LW-UNet 3 at DSC, and some of the metrics are lower than LW-UNet 3. From the point of view of lightweight and accuracy, we finally chose LW-UNet 3 as our segmentation model. Starting from LW-UNet 1, our model has exceeded UNet with a higher accuracy. Compared with UNet, the LW-UNet-3 model increases DSC by 5%, IoU, JC and PC by 12%, 9% and 10% respectively.
Additional to this, we add experiments with the model to the original dataset. We processed MR images obtained from 92 patients diagnosed with nasopharyngeal carcinoma in the same way as the dataset obtained by data augmentation. We conducted the same experiments on both datasets using LW-UNet 0, 1, 2 and 3. By comparing the results with the experiments under data augmentation, we demonstrated the effectiveness of the data obtained by data augmentation in the case of fewer nasopharyngeal carcinoma data. The detailed experimental results are shown in Table 3. Compared to the values of the model in the original dataset, the DSC increased by 0.064, 0.073, 0.078 and 0.078 for LW-UNet 0, 1, 2 and 3, respectively, and the JC values increased by 0.06, 0.081, 0.124 and 0.112. In addition, for IoU, the values of LW-UNet 0, 1, 2 and 3 increased by 16%, 21% , 22% and 21%. 21%, 22%, 21%, for SE, the values of LW-UNet 0, 1, 2 and 3 increased by 20%, 19%, 10% and 6%, respectively.

4.2. Comparison with State-of-the-Art Models

4.2.1. Comparison of Accuracy

We compare the accuracy of our model with four state-of-the-art models and carried out the Kruskal–Wallis test. Five of them are conventional medical image segmentation models (Att-UNet [39], FCN [36], DeeplabV3 [54], TransNet [55], FastTransNet [56]), and the other is a nasopharyngeal cancer image segmentation model (RendUNet [57]). Figure 3 shows the segmentation results for each model. Figure 4 shows the performance of each model on the DSC and JC metrics and Kruskal–Wallis test result. We achieved the highest DSC and JC value in the nasopharyngeal carcinoma segmentation test.
In DSC and JC metrics, our model outperforms the above four comparison models for nasopharyngeal carcinoma segmentation. For DeepLabV3, the average DSC value is 0.788 ± 0.045; we are 3.17% higher than it, and the average Jaccard similarity value is 0.654 ± 0.059; we are 6.26% higher than it. For Att-UNet, the average DSC value is 0.787 ± 0.047; we are 3.30% higher than it, the average Jaccard similarity value is 0.661 ± 0.061, we are 4.23% higher than it. For FCN, the average DSC value is 0.735 ± 0.072, we are 10.61% higher than it; the average Jaccard similarity value is 0.586 ± 0.087, we are 17.60% higher than it. For RendUNet, the average DSC value is 0.789 ± 0.054, we are 3.04% higher than it; the average Jaccard similarity value is 0.643 ± 0.058, 8.08% lower than ours. For TransNet, the average DSC value is 0.807 ± 0.036, ours 0.7% higher than it, the average Jaccard similarity value is 0.689 ± 0.051, ours 0.87% higher than ours. For FastTransNet, the average DSC value is 0.810 ± 0.033, ours 0.37% higher. The average Jaccard similarity value is 0.698 ± 0.045.
We also compare other accuracy indicators of our model with four state-of-the-art models. Figure 5 shows the performance of each model on IoU and SE metrics. As we can see in Figure 5, we achieve the highest value in the nasopharyngeal carcinoma segmentation test.
As for other accuracy metrics, our model outperforms the above four comparison models for nasopharyngeal carcinoma segmentation. For DeepLabV3, the average IoU value is 0.653 ± 0.060, we are 6.58% higher than that, and the average sensitivity value is 0.787 ± 0.053, 4.70% lower than ours. For Att-UNet, the average IoU value is 0.671 ± 0.050, we are 2.68% higher; the average sensitivity value is 0.789 ± 0.046, we are 4.43% higher. For FCN, the average IoU value is 0.586 ± 0.088, we are 18.77% higher; the average sensitivity value is 0.735 ± 0.071, we are 12.10% higher. For RendUNet, the average IoU value is 0.643 ± 0.057, we are 8.24% higher; the average sensitivity value is 0.792 ± 0.045, we are 4.04% higher. For TransNet, the average IoU value is 0.691 ± 0.039, ours 0.71% higher; the average sensitivity value is 0.826 ± 0.074. For FastTransNet, the average IoU value is 0.692 ± 0.031, ours 0.57% higher than it, the average sensitivity value is 0.810 ± 0.033, ours 1.69% higher than it.

4.2.2. Comparison of Parameters and FLOPs

As mentioned in Section 2.1 above, the LW-UNet model uses the MBConv module that enables the model to run under resource constraints. We compare the parameters and the FLOPs between our model and some advanced models, as shown in Table 4. For UNet [38], the parameters of our model are reduced by 872.67% and the FLOPs are reduced by 771.77%. For DeepLabV3 [54], the parameters of our model are reduced by 330.34% and the FLOPs are reduced by 114.91%. For Att-UNet [39], the parameters of our model are reduced by 882.53% and the FLOPs are reduced by 786.41%. For FCN [36], our model’s parameters are reduced by 314.64% and the FLOPs are reduced by 167.24%. For RendUNet [57], the parameters of our model are reduced by 1190.14% and the FLOPs are reduced by 533.55%. For TransNet, the parameters of our model are reduced by 2865.33% and the FLOPs are reduced by 228.09%. For FastTransNet, the parameters of our model are reduced by 740.84% and the FLOPs are reduced by 172.97%. Compared with the above advanced models, it can be seen that our proposed model achieves the best performance. Under the circumstance of limited hospital resources and environment, our model can still run efficiently and provide doctors with a more accurate delineation of target areas.

5. Discussion

Deep learning has achieved success on a variety of computer vision tasks [58,59,60,61,62,63]. LW-UNet is a lightweight automatic segmentation algorithm for nasopharyngeal carcinoma (NPC) based on deep learning. The lightweight network is applied in two main applications. First, it is applied to batch prediction tasks that require a high speed response [64], such as tumour segmentation tasks for nasopharyngeal cancer, lung cancer etc. This task requires a high segmentation speed of the model, and the LW-UNet has small parameters, consumes less computational power and can provide a low latency model to meet the needs of the hospital task. Secondly, the lightweight network is also applied to complete real-time low-latency segmentation tasks with limited computational storage resources or even on the mobile side [65]. Especially for small and medium-sized hospitals, it is able to complete real-time segmentation tasks at high speed with low equipment configuration, reducing the resource consumption of hospitals. LW-UNet has a low dependence on computational resources through a uniform scaling network of width, depth and resolution to improve the accessibility of the algorithm to small and medium-sized hospitals.
In previous research on medical imaging [66,67,68,69,70,71], deep learning technology is generally applied to simple organs such as the pancreas [72], liver [73] and lung [74]. However, the segmentation of nasopharyngeal carcinoma is more complex. Nasopharyngeal carcinoma tumour cells are generally located in the nasopharynx. The surrounding tissue structure is relatively complex, so the high segmentation accuracy is imperative. Because slightly inaccurate segmentation may cause damage to the patient’s brain.
In Figure 3, in the case of some simple nasopharyngeal carcinoma tumour, the results of the automatic segmentation of LW-UNet are the most similar to manual segmentation. However, LW-UNet is slightly limited in some complex cases because of unusual tumour shape and remote tumour location. In comparison with four other advanced models (including RendUNet [57], which specializes in nasopharyngeal carcinoma segmentation), our model achieved the best segmentation performance. In addition, our models are more effective with the lower parameters and FLOPs; this enables our model to operate in some small and medium-sized hospitals with limited resources. Considering the accuracy requirement of the segmentation, our model is temporarily unable to complete the segmentation task independently, so doctors need to check and modify it.

6. Conclusions

LW-UNet is an automatic segmentation model of nasopharyngeal carcinoma (NPC) proposed to solve the problems and difficulties of the problems of high computing resource requirements and low accuracy. By unifying the width, depth, and resolution of the scaling network, the model reduces the parameters and flops to improve the precision and efficiency of segmentation. Compared with other advanced models, our model achieve the best performance. It is expected to play a role in the future of nasopharyngeal carcinoma treatment.

Author Contributions

Conceptualization, Y.L., G.H. and X.L.; methodology, Y.L.; validation, Y.L.; formal analysis, Y.L.; resources, X.L.; visualization, Y.L.; supervision, X.L. and G.H.; project administration, X.L.; funding acquisition, G.H. and X.L.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L., G.H. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (61901533, 62101610), the Shenzhen Fundamental Research Program, China (JCYJ20190807154601663), and the China Postdoctoral Science Foundation (2021M693673).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki. Ethical review and approval were waived for this study, due to the retrospective nature of the survey.

Informed Consent Statement

Patient consent was waived due to the retrospective design of this study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to patient privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lee, A.W.; Ng, W.; Chan, Y.; Sze, H.; Chan, C.; Lam, T. The battle against nasopharyngeal cancer. Radiother. Oncol. 2012, 104, 272–278. [Google Scholar] [CrossRef] [PubMed]
  2. Tsao, S.W.; Lo, K.W.; Huang, D.P. Nasopharyngeal carcinoma. In Epstein-Barr Virus; CRC Press: Boca Raton, FL, USA, 2006. [Google Scholar]
  3. Mimi, C.Y.; Yuan, J.M. Epidemiology of nasopharyngeal carcinoma. Semin. Cancer Biol. 2002, 12, 421–429. [Google Scholar]
  4. Hamid, G.A. Epidemiology and Outcomes of Nasopharyngeal Carcinoma. In Pharynx-Diagnosis and Treatment; IntechOpen: London, UK, 2021. [Google Scholar]
  5. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Das, I.J.; Moskvin, V.; Johnstone, P.A. Analysis of treatment planning time among systems and planners for intensity-modulated radiation therapy. J. Am. Coll. Radiol. 2009, 6, 514–517. [Google Scholar] [CrossRef] [PubMed]
  7. Tatanun, C.; Ritthipravat, P.; Bhongmakapat, T.; Tuntiyatorn, L. Automatic segmentation of nasopharyngeal carcinoma from CT images: Region growing based technique. In Proceedings of the 2010 2nd International Conference on Signal Processing Systems, Dalian, China, 5–7 July 2010; p. V2-537. [Google Scholar]
  8. Huang, K.W.; Zhao, Z.Y.; Gong, Q.; Zha, J.; Chen, L.; Yang, R. Nasopharyngeal carcinoma segmentation via HMRF-EM with maximum entropy. In Proceedings of the 2015 37th annual international conference of the IEEE engineering in medicine and biology society (EMBC), Milan, Italy, 25–29 August 2015; pp. 2968–2972. [Google Scholar]
  9. Wu, B.; Khong, P.L.; Chan, T. Automatic detection and classification of nasopharyngeal carcinoma on PET/CT with support vector machine. Int. J. Comput. Assist. Radiol. Surg. 2012, 7, 635–646. [Google Scholar] [CrossRef] [Green Version]
  10. Zhou, J.; Chan, K.L.; Xu, P.; Chong, V.F. Nasopharyngeal carcinoma lesion segmentation from MR images by support vector machine. In Proceedings of the 3rd IEEE International Symposium on Biomedical Imaging: Nano to Macro, Arlington, VA, USA, 6–9 April 2006; pp. 1364–1367. [Google Scholar]
  11. Mohammed, M.A.; Abd Ghani, M.K.; Hamed, R.I.; Ibrahim, D.A.; Abdullah, M.K. Artificial neural networks for automatic segmentation and identification of nasopharyngeal carcinoma. J. Comput. Sci. 2017, 21, 263–274. [Google Scholar] [CrossRef]
  12. Chanapai, W.; Ritthipravat, P. Adaptive thresholding based on SOM technique for semi-automatic NPC image segmentation. In Proceedings of the 2009 International Conference on Machine Learning and Applications, Miami, FL, USA, 13–15 December 2009; pp. 504–508. [Google Scholar]
  13. Gao, Z.; Liu, X.; Qi, S.; Wu, W.; Hau, W.K.; Zhang, H. Automatic segmentation of coronary tree in CT angiography images. Int. J. Adapt. Control. Signal Process. 2019, 33, 1239–1247. [Google Scholar] [CrossRef]
  14. Boudiaf, M.; Kervadec, H.; Masud, Z.I.; Piantanida, P.; Ben Ayed, I.; Dolz, J. Few-shot segmentation without meta-learning: A good transductive inference is all you need? In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13979–13988. [Google Scholar]
  15. Cho, J.H.; Mall, U.; Bala, K.; Hariharan, B. Picie: Unsupervised semantic segmentation using invariance and equivariance in clustering. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16794–16804. [Google Scholar]
  16. Xu, C.; Xu, L.; Gao, Z.; Zhao, S.; Zhang, H.; Zhang, Y.; Du, X.; Zhao, S.; Ghista, D.; Liu, H.; et al. Direct delineation of myocardial infarction without contrast agents using a joint motion feature learning architecture. Med. Image Anal. 2018, 50, 82–94. [Google Scholar] [CrossRef]
  17. Douillard, A.; Chen, Y.; Dapogny, A.; Cord, M. Plop: Learning without forgetting for continual semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4040–4050. [Google Scholar]
  18. Guzov, V.; Mir, A.; Sattler, T.; Pons-Moll, G. Human poseitioning system (hps): 3d human pose estimation and self-localization in large scenes from body-mounted sensors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4318–4329. [Google Scholar]
  19. Xu, C.; Xu, L.; Ohorodnyk, P.; Roth, M.; Chen, B.; Li, S. Contrast agent-free synthesis and segmentation of ischemic heart disease images using progressive sequential causal GANs. Med. Image Anal. 2020, 62, 101668. [Google Scholar] [CrossRef]
  20. Mok, T.C.; Chung, A. Fast symmetric diffeomorphic image registration with convolutional neural networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4644–4653. [Google Scholar]
  21. Sriram, A.; Zbontar, J.; Murrell, T.; Zitnick, C.L.; Defazio, A.; Sodickson, D.K. GrappaNet: Combining parallel imaging with deep learning for multi-coil MRI reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 14315–14322. [Google Scholar]
  22. Luo, X.; Liao, W.; Chen, J.; Song, T.; Chen, Y.; Zhang, S.; Chen, N.; Wang, G.; Zhang, S. Efficient semi-supervised gross target volume of nasopharyngeal carcinoma segmentation via uncertainty rectified pyramid consistency. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Virtual, 27–1 September–October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 318–329. [Google Scholar]
  23. Tao, G.; Li, H.; Liu, L.; Cai, H. Detection-and-Excitation Neural Network Achieves Accurate Nasopharyngeal Carcinoma Segmentation in Multi-modality MR Images. In Proceedings of the 2021 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Houston, TX, USA, 9–12 December 2021; pp. 1063–1068. [Google Scholar]
  24. Tang, P.; Zu, C.; Hong, M.; Yan, R.; Peng, X.; Xiao, J.; Wu, X.; Zhou, J.; Zhou, L.; Wang, Y. DSU-net: Dense SegU-net for automatic head-and-neck tumour segmentation in MR images. arXiv 2020, arXiv:2006.06278. [Google Scholar]
  25. Jin, Z.; Li, X.; Shen, L.; Lang, J.; Li, J.; Wu, J.; Xu, P.; Duan, J. Automatic Primary Gross Tumor Volume Segmentation for Nasopharyngeal carcinoma using ResSE-UNet. In Proceedings of the 2020 IEEE 33rd International Symposium on Computer-Based Medical Systems (CBMS), Rochester, MN, USA, 28–30 July 2020; pp. 585–590. [Google Scholar]
  26. Gao, Z.; Chung, J.; Abdelrazek, M.; Leung, S.; Hau, W.K.; Xian, Z.; Zhang, H.; Li, S. Privileged modality distillation for vessel border detection in intracoronary imaging. IEEE Trans. Med. Imaging 2019, 39, 1524–1534. [Google Scholar] [CrossRef] [PubMed]
  27. Ji, Y.; Zhang, H.; Zhang, Z.; Liu, M. CNN-based encoder-decoder networks for salient object detection: A comprehensive review and recent advances. Inf. Sci. 2021, 546, 835–857. [Google Scholar] [CrossRef]
  28. Xu, J.; Zhou, W.; Chen, Z.; Ling, S.; Le Callet, P. Binocular Rivalry Oriented Predictive Autoencoding Network for Blind Stereoscopic Image Quality Measurement. IEEE Trans. Instrum. Meas. 2021, 70, 1–13. [Google Scholar]
  29. Bang, S.; Park, S.; Kim, H.; Kim, H. Encoder–decoder network for pixel-level road crack detection in black-box images. Comput.-Aided Civ. Infrastruct. Eng. 2019, 34, 713–727. [Google Scholar] [CrossRef]
  30. Mandal, M.; Vipparthi, S.K. Scene independency matters: An empirical study of scene dependent and scene independent evaluation for CNN-based change detection. IEEE Trans. Intell. Transp. Syst. 2020, 23, 2031–2044. [Google Scholar] [CrossRef]
  31. Deschaintre, V.; Aittala, M.; Durand, F.; Drettakis, G.; Bousseau, A. Single-image svbrdf capture with a rendering-aware deep network. ACM Trans. Graph. (ToG) 2018, 37, 1–15. [Google Scholar] [CrossRef] [Green Version]
  32. Sueiras, J.; Ruiz, V.; Sanchez, A.; Velez, J.F. Offline continuous handwriting recognition using sequence to sequence neural networks. Neurocomputing 2018, 289, 119–128. [Google Scholar] [CrossRef]
  33. Chong, V.; Fan, Y.F. Detection of recurrent nasopharyngeal carcinoma: MR imaging versus CT. Radiology 1997, 202, 463–470. [Google Scholar] [CrossRef]
  34. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Cconference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  36. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  37. Badrinarayanan, V.; Kendall, A.; Cipolla, R. Segnet: A deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  38. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  39. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.J.; Heinrich, M.P.; Misawa, K.; Mori, K.; McDonagh, S.G.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv 2018, arXiv:1804.03999. [Google Scholar]
  40. Zhang, J.; Gu, L.; Han, G.; Liu, X. AttR2U-Net: A Fully Automated Model for MRI Nasopharyngeal Carcinoma Segmentation Based on Spatial Attention and Residual Recurrent Convolution. Front. Oncol. 2021, 11, 816672. [Google Scholar] [CrossRef] [PubMed]
  41. Tang, P.; Zu, C.; Hong, M.; Yan, R.; Peng, X.; Xiao, J.; Wu, X.; Zhou, J.; Zhou, L.; Wang, Y. DA-DSUnet: Dual attention-based dense SU-net for automatic head-and-neck tumour segmentation in MRI images. Neurocomputing 2021, 435, 103–113. [Google Scholar] [CrossRef]
  42. Li, Q.; Xu, Y.; Chen, Z.; Liu, D.; Feng, S.T.; Law, M.; Ye, Y.; Huang, B. Tumor segmentation in contrast-enhanced magnetic resonance imaging for nasopharyngeal carcinoma: Deep learning with convolutional neural network. BioMed Res. Int. 2018, 2018, 9128527. [Google Scholar] [CrossRef] [Green Version]
  43. Ma, Z.; Wu, X.; Song, Q.; Luo, Y.; Wang, Y.; Zhou, J. Automated nasopharyngeal carcinoma segmentation in magnetic resonance images by combination of convolutional neural networks and graph cut. Exp. Ther. Med. 2018, 16, 2511–2521. [Google Scholar] [CrossRef] [Green Version]
  44. Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–5 June 2019; pp. 6105–6114. [Google Scholar]
  45. Huang, Y.; Cheng, Y.; Bapna, A.; Firat, O.; Chen, D.; Chen, M.; Lee, H.; Ngiam, J.; Le, Q.V.; Wu, Y.; et al. Gpipe: Efficient training of giant neural networks using pipeline parallelism. Adv. Neural Inf. Process. Syst. 2019, 32, 103–112. [Google Scholar]
  46. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  47. Dice, L.R. Measures of the Amount of Ecologic Association Between Species. Ecology 1945, 26, 297–302. [Google Scholar] [CrossRef]
  48. Bertels, J.; Eelbode, T.; Berman, M.; Vandermeulen, D.; Maes, F.; Bisschops, R.; Blaschko, M.B. Optimizing the dice score and jaccard index for medical image segmentation: Theory and practice. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 92–100. [Google Scholar]
  49. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  50. Wu, Y.; He, K. Group normalization. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  51. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2015, arXiv:1412.6980. [Google Scholar]
  52. Smith, L.N. Cyclical learning rates for training neural networks. In Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, CA, USA, 24–31 March 2017; pp. 464–472. [Google Scholar]
  53. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  54. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef] [Green Version]
  55. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. Transunet: Transformers make strong encoders for medical image segmentation. arXiv 2021, arXiv:2102.04306. [Google Scholar]
  56. Xu, G.; Wu, X.; Zhang, X.; He, X. Levit-unet: Make faster encoders with transformer for medical image segmentation. arXiv 2021, arXiv:2107.08623. [Google Scholar] [CrossRef]
  57. Li, Y.; Peng, H.; Dan, T.; Hu, Y.; Tao, G.; Cai, H. Coarse-to-fine Nasopharyngeal carcinoma Segmentation in MRI via Multi-stage Rendering. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Korea, 16–19 December 2020; pp. 623–628. [Google Scholar]
  58. Wu, C.; Zhang, H.; Chen, J.; Gao, Z.; Zhang, P.; Muhammad, K.; Del Ser, J. Vessel-GAN: Angiographic reconstructions from myocardial CT perfusion with explainable generative adversarial networks. Future Gener. Comput. Syst. 2022, 130, 128–139. [Google Scholar] [CrossRef]
  59. Graham, B.; El-Nouby, A.; Touvron, H.; Stock, P.; Joulin, A.; Jegou, H.; Douze, M. LeViT: A Vision Transformer in ConvNet’s Clothing for Faster Inference. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 12259–12269. [Google Scholar]
  60. Islam, M.A.; Kalash, M.; Bruce, N.D. Revisiting salient object detection: Simultaneous detection, ranking, and subitizing of multiple salient objects. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7142–7150. [Google Scholar]
  61. Guo, S.; Xu, L.; Feng, C.; Xiong, H.; Gao, Z.; Zhang, H. Multi-level semantic adaptation for few-shot segmentation on cardiac image sequences. Med. Image Anal. 2021, 73, 102170. [Google Scholar] [CrossRef] [PubMed]
  62. Ashraf, M.W.; Sultani, W.; Shah, M. Dogfight: Detecting drones from drones videos. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 7067–7076. [Google Scholar]
  63. Salehi, M.; Sadjadi, N.; Baselizadeh, S.; Rohban, M.H.; Rabiee, H.R. Multiresolution knowledge distillation for anomaly detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14902–14912. [Google Scholar]
  64. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  65. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. Shufflenet: An extremely efficient convolutional neural network for mobile devices. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6848–6856. [Google Scholar]
  66. Gao, Z.; Wu, S.; Liu, Z.; Luo, J.; Zhang, H.; Gong, M.; Li, S. Learning the implicit strain reconstruction in ultrasound elastography using privileged information. Med. Image Anal. 2019, 58, 101534. [Google Scholar] [CrossRef]
  67. Kim, E.; Kim, S.; Seo, M.; Yoon, S. XProtoNet: Diagnosis in chest radiography with global and local explanations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 15719–15728. [Google Scholar]
  68. Gamper, J.; Rajpoot, N. Multiple instance captioning: Learning representations from histopathology textbooks and articles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 16549–16559. [Google Scholar]
  69. Gao, Z.; Xiong, H.; Liu, X.; Zhang, H.; Ghista, D.; Wu, W.; Li, S. Robust estimation of carotid artery wall motion using the elasticity-based state-space approach. Med. Image Anal. 2017, 37, 1–21. [Google Scholar] [CrossRef]
  70. Lee, H.J.; Kim, J.U.; Lee, S.; Kim, H.G.; Ro, Y.M. Structure boundary preserving segmentation for medical image with ambiguous boundary. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4817–4826. [Google Scholar]
  71. Mahapatra, D.; Bozorgtabar, B.; Shao, L. Pathological retinal region segmentation from oct images using geometric relation based augmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9611–9620. [Google Scholar]
  72. Roth, H.R.; Lu, L.; Farag, A.; Shin, H.C.; Liu, J.; Turkbey, E.B.; Summers, R.M. Deeporgan: Multi-level deep convolutional networks for automated pancreas segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 556–564. [Google Scholar]
  73. Gotra, A.; Sivakumaran, L.; Chartrand, G.; Vu, K.N.; Vandenbroucke-Menu, F.; Kauffmann, C.; Kadoury, S.; Gallix, B.; de Guise, J.A.; Tang, A. Liver segmentation: Indications, techniques and future directions. Insights Imaging 2017, 8, 377–392. [Google Scholar] [CrossRef]
  74. Skourt, B.A.; El Hassani, A.; Majda, A. Lung CT image segmentation using deep neural networks. Procedia Comput. Sci. 2018, 127, 109–113. [Google Scholar] [CrossRef]
Figure 1. MRI slices of head and neck and tumour location of nasopharyngeal carcinoma. The area delineated by the red line is the result of the doctor’s manual segmentation.
Figure 1. MRI slices of head and neck and tumour location of nasopharyngeal carcinoma. The area delineated by the red line is the result of the doctor’s manual segmentation.
Sensors 22 05875 g001
Figure 2. Network of LW-UNET: The Compound Scaling Encoder of LW-UNet consists of seven MBConv blocks of different sizes. Decoder is composed of a series of up-convolutions, and through skip connection and up-convolution can obtain the final segmentation map of nasopharyngeal carcinoma.
Figure 2. Network of LW-UNET: The Compound Scaling Encoder of LW-UNet consists of seven MBConv blocks of different sizes. Decoder is composed of a series of up-convolutions, and through skip connection and up-convolution can obtain the final segmentation map of nasopharyngeal carcinoma.
Sensors 22 05875 g002
Figure 3. Examples of NPC segmentation results: We select four typical MRI images of nasopharyngeal carcinoma and present the segmentation results of our model and six models used for comparison.
Figure 3. Examples of NPC segmentation results: We select four typical MRI images of nasopharyngeal carcinoma and present the segmentation results of our model and six models used for comparison.
Sensors 22 05875 g003
Figure 4. Box plots of the DSC and JC values obtained from the tests on the test set and its Kruskal–Wallis results. The results show that our model achieves the highest DSC and JC values in the test of nasopharyngeal carcinoma segmentation and is significantly different from other models.
Figure 4. Box plots of the DSC and JC values obtained from the tests on the test set and its Kruskal–Wallis results. The results show that our model achieves the highest DSC and JC values in the test of nasopharyngeal carcinoma segmentation and is significantly different from other models.
Sensors 22 05875 g004
Figure 5. Box plots of the SE and IoU values obtained from the tests on the test set and its Kruskal–Wallis results. The results show that our model achieves the highest SE and IoU values in the test of nasopharyngeal carcinoma segmentation and is significantly different from other models.
Figure 5. Box plots of the SE and IoU values obtained from the tests on the test set and its Kruskal–Wallis results. The results show that our model achieves the highest SE and IoU values in the test of nasopharyngeal carcinoma segmentation and is significantly different from other models.
Sensors 22 05875 g005
Table 1. The scaling coefficient of LW-UNet 0–3.
Table 1. The scaling coefficient of LW-UNet 0–3.
ModelsDepth Scaling CoefficientWidth Scaling CoefficientResolution Scaling Coefficient
LW-UNet 01.01.01.0
LW-UNet 11.41.21.3
LW-UNet 22.21.62.0
LW-UNet 33.12.02.7
LW-UNet 43.62.23.0
LW-UNet 54.12.43.4
Table 2. The ablation analysis validates the effectiveness of our model configuration.
Table 2. The ablation analysis validates the effectiveness of our model configuration.
ModelsDSCIoUJCSEPCSP
UNet0.769 ± 0.0630.618 ±0.0580.632 ±0.0760.858 ± 0.0760.713 ± 0.0890.996 ± 0.002
LW-UNet-00.696 ± 0.0350.516 ± 0.0430.542 ± 0.0440.674 ± 0.0350.623 ± 0.0670.986 ± 0.001
LW-UNet-10.771 ± 0.0410.621 ± 0.0580.634 ± 0.0580.801 ± 0.0580.765 ± 0.0590.997 ± 0.001
LW-UNet-20.796 ± 0.0600.685 ±0.0350.685 ±0.0520.815 ± 0.0480.767 ±0.0530.998 ± 0.001
LW-UNet-3 (Our)0.813 ± 0.0390.696± 0.0550.695 ± 0.0550.824 ± 0.0440.787± 0.0430.998 ± 0.001
LW-UNet-40.815 ± 0.0750.698 ± 0.0430.699 ± 0.0430.814 ± 0.0110.787± 0.0560.998 ± 0.001
LW-UNet-50.806 ± 0.0540.688 ± 0.0270.689 ± 0.0760.820 ± 0.0580.774± 0.0840.998 ± 0.001
Table 3. The ablation analysis validates the effectiveness of our data augmentation.
Table 3. The ablation analysis validates the effectiveness of our data augmentation.
ModelsDSCIoUJCSEPCSP
LW-UNet-00.632 ± 0.0810.443 ± 0.0280.482 ± 0.0560.561 ± 0.0480.545 ± 0.0590.976 ± 0.001
LW-UNet-0
(with dataset augmentation)
0.696 ± 0.0350.516 ± 0.0430.542 ± 0.0440.674 ± 0.0350.623 ± 0.0670.986 ± 0.001
LW-UNet-10.698 ± 0.0710.513 ± 0.0980.553 ± 0.0780.668 ± 0.0380.626 ± 0.0790.984 ± 0.002
LW-UNet-1
(with dataset augmentation)
0.771 ± 0.0410.621 ± 0.0580.634 ± 0.0580.801 ± 0.0580.765 ± 0.0590.997 ± 0.001
LW-UNet-20.718 ± 0.0530.561 ± 0.0760.561 ±0.0550.736 ± 0.0880.647 ±0.0430.988 ± 0.001
LW-UNet-2
(with dataset augmentation)
0.796 ± 0.0600.685 ±0.0350.685 ±0.0520.815 ± 0.0480.767 ±0.0530.998 ± 0.001
LW-UNet-30.735 ± 0.0530.574 ± 0.0680.583 ±0.0960.775 ± 0.0640.686 ± 0.0850.995 ± 0.003
LW-UNet-3
(with dataset augmentation)
0.813 ± 0.0390.696± 0.0550.695 ± 0.0550.824 ± 0.0440.787± 0.0430.998 ± 0.001
Table 4. Comparison Parameters and FLOPs with Other Models.
Table 4. Comparison Parameters and FLOPs with Other Models.
CategoryModelsParameters (M)FLOPs (G)
Ablation Study ModelsLW-UNet 57.369.01
LW-UNet 45.328.38
LW-UNet 3 (Our)3.557.51
LW-UNet 22.224.75
LW-UNet 11.272.99
LW-UNet 00.851.85
UNet34.5365.47
State-of-the-art ModelsDeepLabV315.3116.14
Att-UNet34.8866.57
FCN3214.7220.07
RendUNet45.8047.58
FastTransNet29.8520.50
TransNet105.2824.64
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Liu, Y.; Han, G.; Liu, X. Lightweight Compound Scaling Network for Nasopharyngeal Carcinoma Segmentation from MR Images. Sensors 2022, 22, 5875. https://doi.org/10.3390/s22155875

AMA Style

Liu Y, Han G, Liu X. Lightweight Compound Scaling Network for Nasopharyngeal Carcinoma Segmentation from MR Images. Sensors. 2022; 22(15):5875. https://doi.org/10.3390/s22155875

Chicago/Turabian Style

Liu, Yi, Guanghui Han, and Xiujian Liu. 2022. "Lightweight Compound Scaling Network for Nasopharyngeal Carcinoma Segmentation from MR Images" Sensors 22, no. 15: 5875. https://doi.org/10.3390/s22155875

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop