Next Article in Journal
Wearable Sensors and Systems in the IoT
Next Article in Special Issue
Two-Stage Segmentation Framework Based on Distance Transformation
Previous Article in Journal
The Effects of Over-Ground Robot-Assisted Gait Training for Children with Ataxic Cerebral Palsy: A Case Report
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

DCNet: Densely Connected Deep Convolutional Encoder–Decoder Network for Nasopharyngeal Carcinoma Segmentation

1
School of Mathematics, Sun Yat-sen University, Guangzhou 510275, China
2
School of Biomedical Engineering, Sun Yat-sen University, Shenzhen 518107, China
3
School of Information Engineering, North China University of Water Resources and Electric Power, Zhengzhou 450046, China
*
Author to whom correspondence should be addressed.
Co-first authors.
Sensors 2021, 21(23), 7877; https://doi.org/10.3390/s21237877
Submission received: 8 November 2021 / Revised: 22 November 2021 / Accepted: 23 November 2021 / Published: 26 November 2021

Abstract

:
Nasopharyngeal Carcinoma segmentation in magnetic resonance imagery (MRI) is vital to radiotherapy. Exact dose delivery hinges on an accurate delineation of the gross tumor volume (GTV). However, the large-scale variation in tumor volume is intractable, and the performance of current models is mostly unsatisfactory with indistinguishable and blurred boundaries of segmentation results of tiny tumor volume. To address the problem, we propose a densely connected deep convolutional network consisting of an encoder network and a corresponding decoder network, which extracts high-level semantic features from different levels and uses low-level spatial features concurrently to obtain fine-grained segmented masks. Skip-connection architecture is involved and modified to propagate spatial information to the decoder network. Preliminary experiments are conducted on 30 patients. Experimental results show our model outperforms all baseline models, with improvements of 4.17%. An ablation study is performed, and the effectiveness of the novel loss function is validated.

1. Introduction

Nasopharyngeal Carcinoma (NPC) is a type of malignant tumor with high incidence in East and Southeast Asia. In 2018, about 129,000 cases were reported, and more than 70% of the new cases were diagnosed in the region mentioned. Since Nasopharyngeal Carcinoma is sensitive to ionizing radiation, the treatment of Nasopharyngeal Carcinoma is mainly based on radiotherapy [1]. There are three stages of radiotherapy technique improvement: conventional 2D, 3D conformal, and then intensity-modulated. Presently, intensity-modulated radiotherapy (IMRT) is widely used and has reduced the 5-year occurrence rates of locoregional failure for newly diagnosed and non-metastatic Nasopharyngeal Carcinoma to 7.4% [2,3]. Despite many significant radiotherapy technique improvements, precise delineation of the gross tumor volume (GTV) plays an important role in all techniques proposed [1]. IMRT is an advanced type of high-precision radiotherapy that uses a computer-controlled linear accelerator to deliver modulated radiation doses to malignant tumors. A precise 3D segmentation of the malignant tumor is needed to determine the dose intensity pattern conforming to the tumor shape best to maximize tumor dose while also minimizing the dose to adjacent normal tissues. Hence, dose delivery accuracy hinges on the precision of the delineation of the GTV of the tumor. However, manually delineating the GTV is labor-intensive and time-consuming since manual segmentation of a brain or neck tumor typically costs 2.7 h [4]. Moreover, inter-observer variability always remains high due to variations in oncologists’ experience. Therefore, methods of automatic GTV segmentation are in demand to relieve oncologists from heavy work and elevate the consistency and accuracy.
Although some methods for automatic GTV segmentation have been proposed (See Section 2), a problem of significant intraclass variation remains not solved for medical image segmentation. Therefore, we are inspired by [5] and propose a densely connected deep convolutional encoder–decoder network for Nasopharyngeal Carcinoma segmentation. A tumor is a lump composed of Carcinoma cells, usually, with no particular shape instead of a regular cylindrical shape [6]. Hence, when it comes to Nasopharyngeal Carcinoma segmentation, the areas of segmented masks always vary depending on anterior coordinates of the medical image slices (See in Figure 1). Therefore, using multi-scale semantics in the feature extraction (encoding) stage is important for refining segmentation results. U-Net [7] extracts multi-scale semantics through skip-connection architecture. It is simply a concatenation of feature maps in the encoding stage and maps generated in the corresponding decoding (up-sampling) process. However, we are unable to bring some optimality or rationality to the concatenation operator. So, we assume that some kinds of operators acting on feature maps in the encoding stage before concatenation can obtain better results. In addition, convolution approximates the optimal operator. Additionally, to expand the receptive fields, a Pyramids Pooling Module (PPM) [8] is added. Since too many layers may cause the gradient vanishing problem during training, we modify supervised loss via integrating loss functions plugged at every feature fusion level. The novel loss function also contributes to the ability of the model to achieve an equal performance of multi-scale tumor regions detection because outputs from all decoding levels of various scales are fused and leveraged. The innovations of the work are summarized as follows:
  • We proposed a deep densely connected convolutional network with a pyramid pooling module followed to establish a multi-level representation of multi-scale feature maps. The size of GTV probably changes dramatically as the z-axis of the transverse slice changes. To remedy the large-scale variation, we deploy five densely connected convolutional blocks with expansive receptive fields to detect multi-scale regions of tumors and leverage a pyramid pooling module to fuse the information from different levels. The high-level feature maps with larger receptive fields contribute to locating tumors, and the low-level feature maps with smaller receptive fields with more spatial information contribute to refining segmentation.
  • We involved and modified skip-connection architecture to our network and brought three-fold benefits: the relatively low-level spatial information is propagated to the corresponding level in the decoder network; the feature maps from the decoder network are convoluted before concatenation which outperforms direct concatenation; the convolutional layers added to compress the channels of feature maps to reduce the number of parameters and accelerate training steps.
  • The multi-scale and multi-level loss functions contribute to realizing multi-scale supervision. It enhances the performance of multi-scale GTV detection.
The remainder of this paper is organized as follows. We give a review of related work in Section 2. In Section 3 we introduce the proposed DCNet in detail. Section 4 describes our experimental setup, performance metrics, and discusses experimental results. Finally, the conclusions and future work are laid out in Section 5.

2. Related Work

Previous studies on Nasopharyngeal Carcinoma segmentation are roughly categorized into traditional methods and deep learning methods, which are to be introduced, respectively.

2.1. Traditional Methods

Traditional visual segmentation algorithms before learning-based methods always extract image features manually and classify voxels or pixels based on extracted features, which are perceived as semi-automatic methods. Fitton et al. [9] proposed an operator-guided delineation tool to refine physician-drawn contours of the GTV assisted by a snake algorithm using defined attraction force as an image feature. Because of demands for automatic segmentation methods, machine learning is prevalent and widely used in clustering and classification tasks. As segmentation is equal to a pixel/voxel-wise classification or clustering task, many learning-based methods have been applied to NPC segmentation. Since support vector machine (SVM) [10] is a typical learning-based classifier, it is widely used. Zhou et al. [11] presented a two-class SVM based method. SVM can learn the actual distribution of magnetic resonance (MR) image data without prior knowledge, and segmentations are performed by a trained SVM classifier. Huang et al. [12] introduced two region-based methods with parameters learning. One is based on metric learning by defining a spatially weighted metric-based similarity function, and the other is based on SVM. Moreover, some probabilistic graphical models are also exploited. Tatanun et al. [13] employed a probabilistic model to select potential seeds for region growing process subsequently. Huang et al. [14] proposed a hidden Markov random field model with maximum entropy to model the spatial information with prior knowledge and achieved good results. Despite precise segmentation obtained from specialized datasets, most conventional machine learning methods require handcrafted feature extraction from the datasets to be trained and their consistency in heterogeneous datasets, where considerable variation in medical image modality (CT/MR), image position (patient), spacing, etc., probably exists. Hence their generalization ability cannot be ensured.

2.2. Deep Learning Methods

In recent years, many deep learning methods have been applied to medical image segmentation and outperform many conventional methods such as level set [15], watershed [16], and K-means [17]. Since convolutional neural networks (CNN) [18] has proved to be a precise and efficient method of semantic image segmentation since it is end-to-end and able to capture contextual semantics through computing high-level feature maps, SegNet [19] referring to a fully convolutional network (FCN) [20] as the encoder stage was first proposed. Then Olaf Ronneberger proposed U-Net [7], one of the most successful networks based on FCN structure, which has been a baseline network architecture in medical image segmentation due to its powerful effect and few training parameters [21]. Many of other network architectures are its modifications such as 3D U-Net [22], V-Net [23], U-Net 3+ [24] etc.
The excellent performance of CNNs and U-Net has prompted some studies on applying them to NPC segmentation. Most investigations are innovations in network architectures, and the modifications are categorized into decoder components, input layers, and connection architecture. In terms of decoding stages, Men et al. [25] replaces up-sampling-convolution layers with deconvolution layers. The extension of input layers also is considered by some investigators. 3D U-Net [22] takes 3D medical volume data instead of 2D slices as inputs to generate dense volumetric segmentation of medical images. Beyond expanding dimensions, multi-modality fusion networks also have been experimented with. MMFNet [26] takes 3D images from three MR modalities as inputs to capture cross-modality and modality-specific features. Zhao et al. [27] invented and auxiliary paths on dual-modality PET-CT images, which takes PET and CT slices as inputs. Additionally, multi-view learning is also supposed to be a potential way [28] for biomedical image segmentation. In addition, Ma et al. [29] proposed three CNN networks taking a single view as input, respectively. As mentioned, concatenation introduced as “skip-connection” to U-Net may not be the best choice for propagating original information to a high-level feature map. Therefore, Lin et al. [30] adopts residual block to deliver context information to higher resolution layers. In addition to a conventional CNN structure, attention mechanism has been investigated in recent years, and experiments prove it is promising for improvements of the performance of CNN. CA-Net [31] involves attention mechanism by adding a spatial attention module, a novel channel attention module, and a scale attention module.

3. Methodology

In this section, we provide details of our DCNet in terms of the preprocessing method, the architecture of the network, and the integrated loss function. In the section about preprocessing, we mainly introduce data augmentation. Then in the next section, we present the architecture of DCNet, focusing on Dense Block, Average Pooling layer, and the skip-connection architecture. Finally, we propose an integrated loss function for multi-level supervision learning.

3.1. Preprocessing

The voxel values of MR scans are affected by the configuration and many complicated external factors, which consistently show high variability between scans. Hence normalization is performed initially to let all the values fall within the fixed interval [ 0 ,   1 ] .
Since few transverse slices are labeled, data augmentation has been applied to generate training data from original slices to alleviate over-fitting. Augmentation methods include random rotation from −10 to 10 degrees, flipping, and translation. The proportion of each method accords with the work from Hao et al. [32].

3.2. Network Architecture

The proposed model is inspired by [5], since it performs well in locating salient objects and detecting the edges of them. They parallel the two main tasks (location of tumor regions and boundary-awareness) of Nasopharyngeal Carcinoma segmentation. The whole model is composed of the encoder network and the decoder network. The architecture of the proposed model is shown in Figure 2. In the encoder network, densely convolutional blocks (referred to as DB) are designed to extract high-level semantic information (i.e., locate tumors in MR slices). Pyramid Pooling Module (PPM) introduced is intended to represent the highest-level semantic feature further and render global contextual information in the largest receptive field to locate tumor signals roughly easier. The encoder network contains five Dense Blocks (DBs), the architecture of which is similar to the dense layer in work from Huang et al. [33]. Each DB has 12 convolutional layers, each of which is composed of a BN+ReLU+Conv [34] layer. BN+ReLU+Conv is a layer consisting of a Batch Normalization layer, a ReLU, and a convolutional operator. A BN+ReLU+Conv layer is placed between every two DBs (referred to as transition layer in DenseNet [33]) to compress the channels of the outputs of DBs. To expand receptive fields by down-sampling, a 2 × 2 average pooling layer is placed after the first two DBs. The numbers of the channels of the feature maps generated by DB1, DB2, DB3, DB4, and DB5 are 80, 112, 118, 136, and 16, respectively. The sizes of them are 256, 128, 64, 64, and 64, respectively. The architecture of the pyramid pooling module (PPM) is the same as the pyramid pooling module in PSPNet [8]. Feature maps with the size of 64 × 64 are down-sampled to four levels and then recovered to 64 × 64 . The decoder network containing up-sampling is aimed to recover the feature maps to the size of 256 × 256 at different levels ( S 0 S 5 ) Skip-connection architecture is involved and modified to restore the low-level spatial information and propagate it to the decoder network for segmentation refinement since low-level spatial information contributes to detecting the edge of tumors. In the decoder network, feature maps are up-sampled to the same resolution as input images 256 × 256 . S 1 S 5 are the five feature maps produced at different levels, concatenated to generate the final segmentation S 0 . The Activation Module is a BN+ReLU+Conv layer. In addition, S 0 S 5 have corresponding loss functions J 0 J 5 as mentioned.

3.2.1. The Encoder Network

The encoder network includes the following three types of main components: Dense Blocks (DBs) is employed to extract features from input images; BN+ReLU+Conv, consisting of a Batch Normalization layer, a Rectified Linear Unit, and a convolution operator, is involved as a processing sequence to compress the channels of the outputs of DBs; Average Pooling Module is embedded to enlarge the receptive fields of subsequent Dense Blocks, through which global semantic (or contextual) feature be extracted to locate tumors. Additionally, it downsizes the training parameters.

Dense Block

Densely connected convolutional network, referred to as Dense Block, is inspired by DenseNet [33]. The number of convolutional layer and feature maps each convolutional layer producing (The growth rate [33]) are both 12, The detailed architecture is shown in Figure 3. Each convolutional layer is a combination of Batch Normalization, Rectified Linear Unit (ReLU), and a convolutional operator. The Batch Normalization layer introduced is intended to remedy Internal Covariate Shift by fixing the distribution of the inputs of the ReLU [34]. As a modification, we substitute dilated convolution for general convolution for the extension of receptive fields, which is the same as [5]. Since dilated convolution expands the receptive field without increasing the number of operations [35]. At the same time, the sizes of feature maps are invariant to preserve spatial information. The feature maps produced by DBs are designed to be compressed to 16 and propagated to the decoder network as low-level (spatial detailed) information via skip-connection architecture (see Section Skip-Connection). The forward model of the k th layer in the DB can be formulated as:
y k = H ( y 1 y 2 y k 1 )
where y k is input and output of k th layer, respectively. y 1 y 2 y k 1 represents the concatenation in channel dimension, and the convolutional layer (more specifically, it is a combination of a Batch Normalization layer, a ReLU, and a convolutional layer) is denoted by H ( ) . Additionally, x is padded with zeros to preserve the size of feature maps. (referred to as same padding in [36]). Outside the boundaries of MR slices, no organs or tissues exist, so the corresponding signal intensities are zeros, being a rational explanation for padding.
Figure 2. The architecture of DCNet. The growth rate [33] and the number of convolution layers in each dense block (DB) are both set to 12. The Dense Block (Conv+Pool) refers to a Dense Block (see Section Dense Block), of which the transition layer is a combination of an Average Pooling layer and a convolutional layer.
Figure 2. The architecture of DCNet. The growth rate [33] and the number of convolution layers in each dense block (DB) are both set to 12. The Dense Block (Conv+Pool) refers to a Dense Block (see Section Dense Block), of which the transition layer is a combination of an Average Pooling layer and a convolutional layer.
Sensors 21 07877 g002
The transition layers of the first two DBs are a combination of a convolutional layer and an average pooling layer. In addition, the transition layers of DB3 and DB4 are a convolutional layer to compress the channels of the outputs. Finally, the transition layer of DB5 is a simple convolutional layer (no BN layers and ReLUs).
The 2 × 2 average pooling layer embedded in the first two DBs (DB1 and DB2 in Figure 2) to enlarge receptive fields and reduce trained parameters. The number of feature maps is invariant in pooling layers, while the sizes reduce by one half in both width and height.

Skip-Connection

Similar to U-Net [22], we propagate low-level (spatial detailed) feature to the decoder network. The feature maps rendered in each layer are compressed to 16 and connected to the corresponding layer in the decoder network (referred to as concatenation in Figure 2). Hence, the segmentation mask each layer in the decoder network produced is based on not only high-level semantic information extracted from all higher-level feature maps but also low-level fine-grained feature from the corresponding level in the decoder network. The principle will be explained in detail in Section 3.2.2.

3.2.2. The Decoder Network

The decoder network is designed to perceive the segmentation maps from the feature maps in different abstract levels. In addition, all the segmentation maps produced will be concatenated and fused to generate the final segmentation S 0 . Up-sampling is adapted to recover the sizes of feature maps to 256 × 256 . S 1 S 5 are segmentation based on only higher-level information. The forward model of the k th level of the decoder network can be formulated as:
t k 1 = C o n v ( y k 1 )
S k = S i g m o i d H ( S k 1 t k 1 )
where C o n v represents a convolution operator compressing the channels of feature maps to 16. y k 1 is the same feature map as it in Equation (1). H is a convolutional layer and S i g m o i d is a sigmoid function which is defined as:
S i g m o i d ( x ) = 1 1 + e x
which maps R 1 to [ 0 ,   1 ] . Hence the value at each pixel of the outputs of sigmoid functions shows the probability that the corresponding pixel within the input image belongs to tumors. Finally, the outputs are binarized to obtain S 0 S 5 and the classification threshold is 0.5 .
From Equation (2), we derive that S k only depends on y k 1 and S k 1 . y k 1 preserves low-level spatial information extracted by the first k DBs. So each level of the decoder network leverages the spatial information through skip-connection architecture.

3.3. Multi-Level Integrated Loss Function

Similar to [5], an integrated loss function is defined to realize multi-level and multi-scale supervision. Since S 0 S 5 are segmentation obtained from different abstract levels, we include the difference between S 0 S 5 and ground truth as a part of the defined loss. Furthermore, we involve dice loss as a typical metric of medical image segmentation. Hence, our loss function for one segmentation mask produced can be formulated as:
J ( P ) = C E ( P ) + D L ( P )
C E is pixel-wised weighted cross entropy (Equation (1) in [7]) loss function, which is defined as:
C E ( P ) = i , j λ ω G i , j log P i , j + ( 1 G i , j ) log ( 1 P i , j ) V
where G i , j is the value of pixel within ground truth (manual Nasopharyngeal Carcinoma segmentation by experienced oncologists). P i , j is that within the predicted segmentation mask. V is the area (i.e., the number of pixels) of the ground truth or the prediction. Since the tumor region always occupies a small part of a MR slice, it raises the problem of imbalanced foreground and background, which will lower predictive performance, especially for the minority class (referred to tumor area in the proposed model) ω is a weight to balance minor tumor and major normal organs, which is defined by ω = V V P 1 V P 1 . V P 1 is the area of predicted tumors in MR slices (the number of pixels labeled one within the prediction). λ is a coefficient to control the influence of the weight w.
D L is generalized weighted dice loss [37]:
D L ( P ) = 1 2 ω 1 | P 1 G 1 | + ω 0 | P 0 G 0 | ω 1 ( | P 1 | + | G 1 | ) + ω 0 ( | P 0 | + | G 0 | )
where P 1 is the same as P and P 0 is 1 P . Similarly, G 1 is the same as G and G 0 is 1 G . The weights ω 1 and ω 0 are defined as 1 | G 1 | and 1 | G 0 | respectively. As mentioned above, the multi-level losses for S 0 S 5 are to be summed up to enhance segmentation performance (see Section 3.3).

4. Experimental Evaluation

To evaluate our model, we conducted preliminary experiments. The model is built up on PyTorch platform. We performed the experiments using a server with four Tesla P100 GPU with 16GB RAM. The hyper-parameters to be tuned include epoch, drop rate, and batch size. In addition, the number of trainable parameters is 0.28 M. The epoch, drop rate, growth rate, and batch size are set to 300, 0.2 , 12, and 8 for all the following experiments.

4.1. Data Acquisitions

Forty T1-weighted (T1) MR scans have been collected from 30 patients diagnosed with NPC. Thirty of them are selected randomly as the training set, and the other is used for testing. The data are re-sampled to 256 × 256 with the spatial resolutions fixed to 1 mm using cubic spline interpolation. In addition, all scans have been re-oriented to R A S + coordinate system.

4.2. Training Details

We trained the model through the stochastic gradient descent algorithm (SGD). The 40 training scans are divided into a training set composed of 30 composed of 10 scans. The learning rate is 0.0001 , and the batch size is 8. Our training steps involve three stages to enhance the performance of DCNet: First, we freeze all the layers in the decoder network, and backward gradient propagation only acts on the training parameters of the encoder network. Then, the layers in the encoder network are frozen, and the decoder network is trained by SGD. Finally, we do not freeze any layers, and the whole model is integrally trained. In each stage, the model is trained for equal one hundred epochs. Concurrently, the learning rate is set to be multiplied by 0.1 if the Dice Similarity Coefficient evaluated on the validation set has not risen for ten epochs.
The average time of learning of one epoch is 141.14 s. In addition, the average time of processing a single MR slice is 1.03 s.

4.3. Experimental Results

The proposed model is compared with universal and representative models as for tumor segmentation, including FCN [38], U-Net [22] and U-Net++ [39] as baseline models. The models are mainly evaluated according to Dice Similarity Coefficient (DSC). It is a typical metric of medical image segmentation. It can be formulated as:
D ( P , G ) = | P G | | P | + | G |
where P and G is the predicted segmentation and ground truth, respectively. | P | represents the number of pixels within the region P. The experimental results are summarized in Table 1.
Concurrently, five other widely used metrics are selected for auxiliary evaluation:
Table 1. Quantitative segmentation results of DCNet and the baseline models. The learning rate is 0.0001 and the batch size is 8. All models have been trained for 300 epochs on the same training set, and are evaluated on the same testing set in every cross-validation trial. The number of cross-validation trials is 4. ↑ means the higher value is better, while ↓ means the lower value is better. The best results are shown in boldface.
Table 1. Quantitative segmentation results of DCNet and the baseline models. The learning rate is 0.0001 and the batch size is 8. All models have been trained for 300 epochs on the same training set, and are evaluated on the same testing set in every cross-validation trial. The number of cross-validation trials is 4. ↑ means the higher value is better, while ↓ means the lower value is better. The best results are shown in boldface.
NetDSC 1 SEN ↑SPE ↑PPV ↑VOE ↓RVD ↓
U-Net [22]0.7060.5830.9980.9290.4520.358
U-Net++ [39]0.7150.7440.9950.7280.4380.332
FCN [38]0.7420.7320.9960.7670.4050.139
DCNet (ours)0.7730.8540.9950.7320.3630.321
1 The p value was calculated by performing Mann–Whitney U Test on the main metric DSC. Two-tailed p < 0.05 indicates a significant difference. In addition, the p values of the paired tests between DCNet and FCN, U-Net++, and U-Net are 0.200 , 0.029 , and 0.029 respectively. So, our method outperforms U-Net and U-Net++ significantly.

4.3.1. Sensitivity (SEN)

The S E N shows the percentage of tumors detected. It can be formulated as:
S E N = | P G | | G |

4.3.2. Specificity (SPE)

The S P E shows the percentage of normal region correctly segmented. It can be formulated as:
S P E = | P 1 P G | | P 1 G |
where P 1 is a three-dimensional tensor with the same size as P and G, all of which elements labeled 1.

4.3.3. Positive Predictive Value (PPV)

The P P V is the precision of segmentation results. It can be formulated as the following.
P P V = | P G | | P |

4.3.4. Volumetric Overlap Error (VOE)

The V O E [40] reflects the error rate of segmentation results, which is defined as the following.
V O E = 1 | V G | | V G |

4.3.5. Relative Volume Difference (RVD)

The R V D [40] reflects the difference between segmentation results and Ground Truth. It is defined as the following.
R V D = | P | | G | | G |
According to Table 1, in terms of the main metric DSC, our proposed model outperforms any other baseline model with a value of 0.773. In terms of the main metric DSC, a significant improvement of 4.17% on FCN, the baseline model with the best performance with a value of 0.742, can be observed. DCNet also performs top three in terms of all the auxiliary metrics. So, its effectiveness is proven.
To verify the interpretation of the proposed model (see Section 3), Figure 4 illustrates an example of segmentation results produced by different levels of the decoder network. Please note that the outputs at low levels ( S 1 , S 2 ) preserve more spatial information since the edges of the left half of the results are rougher and fit ground truth better than that of results in higher levels ( S 4 , S 5 ). However, there are a few dots outside and around the tumor region labeled. We suppose that the mislabeling is caused by a lack of context information, which accords with the interpretation made in Section 3.
Figure 5 illustrates a visual comparison of segmentation results of tumors region with different sizes among DCNet and other baseline models. The results show our model performs better than the baseline models and gives a consistently precise result. Hence our model can handle the scale variation problem mentioned in Section 1.

4.4. Ablation Analysis

To demonstrate the effectiveness of the network we designed, we conducted four ablation studies to validate the performance of main designed components, including the PPM, the skip-connection architecture, and the integrated loss function. Additionally, we validate that the number of DBs, which is five, is most effective. It keeps a balance between the ability to extract semantic features and the scale of training parameters. The DCNet we proposed is referred to as DCNet-A, and its variants for ablation studies are DCNet-B, DCNet-C, DCNet-D, DCNet-E, DCNet-F, and DCNet-G. All the variants are trained and validated using the same procedure as described in Table 1 Each ablation experiment is performed 3 times and the results are averaged and shown in the following tables and figures.

4.4.1. Effectiveness of Key Modules of DCNet

Pyramid Pooling Module

To measure the contribution of PPM, we design a baseline model without PPM: DCNet-B in Table 2. From Table 2, DCNet-A outperforms the baseline model in terms of all the metrics provided. Its architecture is illustrated in Figure 6. The quantitative results in the table show PPM improves the segmentation performance of the model.

Skip-Connection Architecture

As mentioned above, we suppose the skip-connection architecture functions as a pathway to convey the low-level spatial feature in the encoder network to the decoder network. Hence it boosts the performance of the model. To verify the argument, we derive a comparative model: DCNet-C in Table 2. Its structure is summarized in Figure 6. From the table, we observe DCNet-A provides the best results in terms of all the metrics. It has a 15.7% improvement in terms of DSC compared to DCNet-C. So, the skip-connection architecture is necessary for boosting the performance of the model.

Integrated Loss Function

We have made the argument that the novel loss function plugged in contributes to the improvements of multi-scale GTV detection. To substantiate it, we applied only one loss function (the same formulation as Equation (5)) to evaluate the difference between the segmentation result generated at the highest level of the decoder network S 0 and the ground truth in the training process. It is referred to as DCNet-F in Figure 6. Table 3 shows integrated loss functions can enhance the performance of segmentation. DCNet-A with an integrated loss function outperforms DCNet-F on the main metric DSC and most of the auxiliary metrics. The only metric where DCNet-F obtains a better result is RVD. Hence, the argument made is strongly proven.

The Number of Dense Blocks

We investigate whether five is the optimal value of the number of DBs. So, we design two baseline models: DCNet-D with four DBs and DCNet-E with six DBs. A significant decrease in performance is clearly observed from Table 4. The DSC of DCNet-D with four DBs is 47.6% lower than DCNet-A with five DBs. The DSC of DCNet-E with six DBs is 40.4% lower than DCNet-A. In addition, DCNet-A performs better than the other two comparative models in terms of all the auxiliary metrics except RVD. A direct interpretation is that the depth of DCNet-D with four DBs is not adequate for dealing with large intraclass variation [41] of scale and shape of tumors, while DCNet-E with six DBs is exceedingly deep, causing the gradient vanishing problem. Collectively, five is the optimal choice.

Dense Block

To validate the effectiveness of the Dense Blocks, we remove the dense connection in the five DBs (referred to as DCNet-G in Figure 6). The comparative model includes five blocks. Each block is composed of 12 BN + ReLU + Conv layers. In addition, each block produces a feature map with the same number of channels as the outputs of the corresponding DB. The quantitative results are summarized in Figure 7. The comparative model performs significantly worse than DCNet-A in terms of all metrics. The values of DSC, SEN, SPE, and PPV decrease by 53.8%, 63.0%, 0.3%, and 46.4% respectively. In addition, the increase rates of VOE and RVD are 116.5% and 18.7%, respectively. The quantitative results are shown in Table 5.

4.4.2. Comparison of the Influence of Modules

As is illustrated in Figure 7, a comprehensive comparison of the performance of DCNet and its variants is intended to investigate the relative contribution of the designed modules. The figure shows the performance of DCNet-B without PPM is worst. Hence PPM is the vital component of DCNet to boost the performance of segmentation. The modules proposed sorted by contribution, from major to minor, are PPM, DB, the integrated loss function, and the skip-connection architecture, according to the comparison of most metrics. In addition, the number of DBs also has significant influence on the segmentation performance, more than the loss function and connection.

5. Conclusions and Future Work

In this paper, a densely connected deep convolutional encoder–decoder network is proposed for Nasopharyngeal Carcinoma segmentation in MR slices. The network is composed of the encoder network and the decoder network for location of tumor regions and boundary-awareness (the ability to detect and label the edges of tumors precisely). In the encoder network, we use densely connected block and pyramid pooling modules to extract high-level semantic features to locate tumors. In addition, the decoder network is designed to obtain masks. Skip-connection architecture is involved and modified to propagate low-level spatial features to the decoder network. Hence it integrates spatial information propagated and semantic information fed by pyramid pooling module. Additionally, masks generated at each level are fused (concatenated and convoluted) to acquire fine-grained segmentation. In terms of training strategy, we also propose an integrated loss function to remedy the gradient vanishing problem by strengthening the gradient signal at each level of the decoder network.
Although our method shows satisfactory results, we suppose there are some limitations and potential improvements. The limitations and future work are summarized as follows:
1.
The influence of the value set of the hyper-parameters has not been investigated.
2.
The network only processes a single 2D slice at one time. So, we think the network can take 3D image series as inputs by redesigning parts of the network, to speed up the prediction.
3.
We consider using transfer learning methods, by training the network on the other data containing other tumors in MR images, to solve the problem of few data.
4.
We have noticed Generative Adversarial Network (GAN) is used in web data to resolve the class imbalance problem [42]. In addition, we suppose it is a potential way to resolve the problem of intraclass variation and lack of sufficient data in our research. We consider involving a GAN in the preprocessing stage.
5.
We have noticed transformer is used for segmentation tasks for other types of tumors and organs such as MedT [43] and UTNet [44]. These models also outperform baseline models. Therefore, we plan to compare our method to new transformer methods to seek potential modifications.
6.
We doubt whether it performs well in cases with other types of tumors, and we plan to investigate in the future.

Author Contributions

Conceptualization, X.L.; methodology, Y.L., G.H. and X.L.; software, Y.L.; validation, G.H.; formal analysis, Y.L.; investigation, Y.L.; resources, X.L.; data curation, Y.L.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L., G.H. and X.L.; visualization, Y.L.; supervision, X.L.; project administration, G.H.; funding acquisition, G.H. and X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (61901533, 62101610), the Shenzhen Fundamental Research Program, China (JCYJ20190807154601663), and the China Postdoctoral Science Foundation (2021M693673).

Institutional Review Board Statement

The study was conducted according to the guidelines of the Declaration of Helsinki. Ethical review and approval were waived for this study, due to the retrospective nature of the survey.

Informed Consent Statement

Patient consent was waived due to the retrospective design of this study.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to patient privacy.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, Y.; Chan, A.; Le, Q.; Blanchard, P.; Sun, Y.; Ma, J. Nasopharyngeal Carcinoma. Lancet 2019, 394, 64–80. [Google Scholar] [CrossRef]
  2. Mao, Y.; Tang, L.; Chen, L.; Sun, Y.; Qi, Z.; Zhou, G.; Liu, L.; Li, L.; Lin, A.; Ma, J. Prognostic Factors and Failure Patterns in Non-metastatic Nasopharyngeal Carcinoma after Intensity-modulated Radiotherapy. Chin. J. Cancer 2016, 35, 103. [Google Scholar] [CrossRef] [Green Version]
  3. Pow, E.; Kwong, D.; McMillan, A.; Wong, M.; Sham, J.; Leung, L.; Leung, K. Xerostomia and Quality of Life after Intensity-modulated Radiotherapy vs. Conventional Radiotherapy for Rarly-stage Nasopharyngeal Carcinoma: Initial Report on a Randomized Controlled Clinical Trial. Int. J. Radiat. Oncol. Biol. Phys. 2006, 66, 981–991. [Google Scholar] [CrossRef]
  4. Das, I.; Moskvin, V.; Johnstone, P. Analysis of Treatment Planning Time among Systems and Planners for Intensity-modulated Radiation Therapy. J. Am. Coll. Radiol. 2019, 6–7, 514. [Google Scholar] [CrossRef] [PubMed]
  5. Dong, S.; Gao, Z.; Sun, S.; Wang, X.; Li, M.; Zhang, H.; Yang, G.; Liu, H.; Li, S. Holistic and Deep Feature Pyramids for Saliency Detection. In Proceedings of the 29th British Machine Vision Conference, BMVC 2018, Newcastle, UK, 3–6 September 2018; p. 67. [Google Scholar]
  6. Hesamian, M.; Jia, W.; He, X.; Kennedy, P. Deep Learning Techniques for Medical Image Segmentation: Achievements and Challenges. J. Digit. Imaging 2019, 32, 582–596. [Google Scholar] [CrossRef] [Green Version]
  7. Olaf, R.; Philipp, F.; Thomas, B. U-net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention, MICCAI 2015, Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  8. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2016; pp. 6230–6239. [Google Scholar]
  9. Fitton, I.; Cornelissen, S.; Duppen, J.; Steenbakkers, R.; Peeters, S.; Hoebers, F.; Kaanders, J.; Nowak, P.; Rasch, C.; Van Herk, M. Semi-automatic Delineation Using Weighted CT-MRI Registered Images for Radiotherapy of Nasopharyngeal Cancer. Med. Phys. 2011, 38, 4662–4666. [Google Scholar] [CrossRef] [PubMed]
  10. Cortes, C.; Vapnik, V. Support-vector Networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  11. Zhou, J.; Chan, K.; Xu, P.; Chong, F. Nasopharyngeal Carcinoma Lesion Segmentation from MR Images by Support Vector Machine. In Proceedings of the 3rd IEEE International Symposium on Biomedical Imaging, ISBI 2006, Arlington, VA, USA, 6–9 April 2006; pp. 1364–1367. [Google Scholar]
  12. Huang, W.; Chan, K.; Zhou, J. Region-based Nasopharyngeal Carcinoma Lesion Segmentation from MRI Using Clustering and Classification-based Methods with Learning. J. Digit. Imaging 2013, 26, 472–482. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Tatanun, C.; Ritthipravat, P.; Bhongmakapat, T.; Tuntiyatorn, L. Automatic Segmentation of Nasopharyngeal Carcinoma from CT images: Region Growing Based Technique. In Proceedings of the 2nd International Conference on Signal Processing Systems, ICSPS 2010, Dalian, China, 5–7 July 2010; Volume 2, pp. 537–541. [Google Scholar]
  14. Huang, K.; Zhao, Z.; Gong, Q.; Zha, J.; Chen, L.; Yang, R. Nasopharyngeal Carcinoma Segmentation via HMRF-EM with Maximum Entropy. In Proceedings of the 37th Annual International Conference of the IEEE Engineering in Medicine and Biology Society, EMBC 2015, Milan, Italy, 25–29 August 2015; pp. 2968–2972. [Google Scholar]
  15. Xu, A.; Wang, L.; Feng, S.; Qu, Y. Threshold-Based Level Set Method of Image Segmentation. In Proceedings of the 3rd International Conference on Intelligent Networks and Intelligent Systems, ICINIS 2010, Shenyang, China, 1–3 November 2010; pp. 703–706. [Google Scholar]
  16. Chen, J.; Liu, S. A Medical Image Segmentation Method Based on Watershed Transform. In Proceedings of the 5th International Conference on Computer and Information Technology, CIT 2005, Shanghai, China, 21–23 September 2005; pp. 634–638. [Google Scholar]
  17. Ng, H.; Ong, S.; Foong, K.; Goh, P.; Nowinski, W. Medical Image Segmentation Using K-Means Clustering and Improved Watershed Algorithm. In Proceedings of the 6th IEEE Southwest Symposium on Image Analysis and Interpretation, SSIAI 2006, Denver, CO, USA, 26–28 March 2006; pp. 61–65. [Google Scholar]
  18. Sudha, S.; Jayanthi, K.; Rajasekaran, C.; Sunder, T. Segmentation of RoI in Medical Images Using CNN-A Comparative Study. In Proceedings of the 34th IEEE Region 10 Conference, TENCON 2019, Kochi, India, 17–20 October 2019; pp. 767–771. [Google Scholar]
  19. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef]
  20. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef]
  21. Jiang, H.; Diao, Z.; Yao, Y. Deep Learning Techniques for Tumor Segmentation: A Review. J. Supercomput. 2021, 77. [Google Scholar] [CrossRef]
  22. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention, MICCAI 2016, Athens, Greece, 17 October 2016; pp. 424–432. [Google Scholar]
  23. Men, K.; Chen, X.; Zhu, J.; Yang, B.; Zhang, Y.; Yi, J.; Dai, J. Continual Improvement of Nasopharyngeal Carcinoma Segmentation with Less Labeling Effort. Phys. Medica 2020, 80, 347–351. [Google Scholar] [CrossRef]
  24. Huang, H.; Lin, L.; Tong, R.; Hu, H.; Zhang, Q.; Iwamoto, Y.; Han, X.; Chen, Y.; Wu, J. UNet 3+: A Full-Scale Connected UNet for Medical Image Segmentation. In Proceedings of the 45th IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP 2020, Barcelona, Spain, 4–8 May 2020; pp. 1055–1059. [Google Scholar]
  25. Men, K.; Chen, X.; Zhang, Y.; Zhang, T.; Dai, J.; Yi, J.; Li, Y. Deep Deconvolutional Neural Network for Target Segmentation of Nasopharyngeal Cancer in Planning Computed Tomography Images. Front. Oncol. 2017, 7, 7–15. [Google Scholar] [CrossRef] [Green Version]
  26. Chen, H.; Qi, Y.; Yin, Y.; Li, T.; Liu, X.; Li, X.; Gong, G.; Wang, L. MMFNet: A Multi-modality MRI Fusion Network for Segmentation of Nasopharyngeal Carcinoma. Neurocomputing 2020, 394, 27–40. [Google Scholar] [CrossRef] [Green Version]
  27. Zhao, L.; Lu, Z.; Jiang, J.; Zhou, Y.; Wu, Y.; Feng, Q. Automatic Nasopharyngeal Carcinoma Segmentation Using Fully Convolutional Networks with Auxiliary Paths on Dual-Modality PET-CT Images. J. Digit. Imaging 2019, 32, 462–470. [Google Scholar] [CrossRef]
  28. Zhang, W.; Yang, G.; Zhang, N.; Xu, L.; Wang, X.; Zhang, Y.; Zhang, H.; Javier, D.; de Albuquerque, V.H.C. Multi-task Learning with Multi-view Weighted Fusion Attention for artery-specific calcification analysis. Inf. Fusion 2021, 71, 64–76. [Google Scholar] [CrossRef]
  29. Ma, Z.; Wu, X.; Zhou, J. Automatic Nasopharyngeal Carcinoma Segmentation in MR Images with Convolutional Neural Networks. In Proceedings of the International Conference on the Frontiers and Advances in Data Science, FADS 2017, Xi’an, China, 23–25 October 2017; pp. 147–150. [Google Scholar]
  30. Lin, L.; Dou, Q.; Jin, Y.; Zhou, G.; Tang, Y.; Chen, W.; Su, B.; Liu, F.; Tao, C.; Jiang, N.; et al. Deep Learning for Automated Contouring of Primary Tumor Volumes by MRI for Nasopharyngeal Carcinoma. Radiology 2019, 291, 677–686. [Google Scholar] [CrossRef] [PubMed]
  31. Gu, R.; Wang, G.; Song, T.; Huang, R.; Aertsen, M.; Deprest, J.; Ourselin, S.; Vercauteren, T.; Zhang, S. CA-Net: Comprehensive Attention Convolutional Neural Networks for Explainable Medical Image Segmentation. IEEE Trans. Med. Imaging 2021, 40, 699–711. [Google Scholar] [CrossRef] [PubMed]
  32. Dong, H.; Yang, G.; Liu, F.; Mo, Y.; Guo, Y. Automatic Brain Tumor Detection and Segmentation Using U-Net Based Fully Convolutional Networks. Commun. Comput. Inf. Sci. 2017, 723, 506–517. [Google Scholar]
  33. Huang, G.; Liu, Z.; Maaten, L.; Weinberger, K. Densely Connected Convolutional Networks. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2016; pp. 2261–2269. [Google Scholar]
  34. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In Proceedings of the 32nd International Conference on Machine Learning, ICML 2015, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  35. Lopes, A.; Ribeiro, A.; Silva, C. Dilated Convolutions in Retinal Blood Vessels Segmentation. In Proceedings of the 6th IEEE Portuguese Meeting on Bioengineering, ENBENG 2019, Lisbon, Portugal, 22–23 February 2019; pp. 1–4. [Google Scholar]
  36. Dumoulin, V.; Visin, F. A Guide to Convolution Arithmetic for Deep Learning. arXiv 2016, arXiv:1603.07285. [Google Scholar]
  37. Lin, T.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2016; pp. 936–944. [Google Scholar]
  38. Shen, H.; Wang, R.; Zhang, J.; McKenna, S. Boundary-Aware Fully Convolutional Network for Brain Tumor Segmentation. In Proceedings of the 20th Medical Image Computing and Computer-Assisted Intervention, MICCAI 2017, Quebec City, QC, Canada, 11–13 September 2017; pp. 433–441. [Google Scholar]
  39. He, Y.; Yu, X.; Liu, C.; Zhang, J.; Hu, K.; Zhu, H. A 3D Dual Path U-Net of Cancer Segmentation Based on MRI. In Proceedings of the 3rd IEEE International Conference on Image, Vision and Computing, ICIVC 2018, Chongqing, China, 27–29 June 2018; pp. 268–272. [Google Scholar]
  40. Tobias, H.; Bram, G.; Martin, S.; Yulia, A.; Volker, A.; Christian, B.; Andreas, B.; Christoph, B.; Reinhard, B.; György, B.; et al. Comparison and Evaluation of Methods for Liver Segmentation from CT Datasets. IEEE Trans. Med. Imaging 2009, 28, 1251–1265. [Google Scholar]
  41. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P. Automated Melanoma Recognition in Dermoscopy Images via Very Deep Residual Networks. IEEE Trans. Med. Imaging 2017, 36, 994–1004. [Google Scholar] [CrossRef] [PubMed]
  42. Hao, J.; Wang, C.; Yang, G.; Gao, Z.; Zhang, J.; Zhang, H. Annealing Genetic GAN for Imbalanced Web Data Learning. IEEE Trans. Multimed. 2021. [Google Scholar] [CrossRef]
  43. Valanarasu, J.; Oza, P.; Hacihaliloglu, I.; Patel, V. Medical Transformer: Gated Axial-Attention for Medical Image Segmentation. In Proceedings of the 24th Medical Image Computing and Computer Assisted Intervention, MICCAI 2021, Virtual, 27 September–1 October 2021; pp. 36–46. [Google Scholar]
  44. Gao, Y.; Zhou, M.; Metaxas, D. UTNet: A Hybrid Transformer Architecture for Medical Image Segmentation. In Proceedings of the 24th Medical Image Computing and Computer Assisted Intervention, MICCAI 2021, Virtual, 27 September–1 October 2021; pp. 61–71. [Google Scholar]
Figure 1. A sample of the significant variation of the scales of tumors. Most current methods are not able to perform consistently precisely on such slices with noticeable different scales.
Figure 1. A sample of the significant variation of the scales of tumors. Most current methods are not able to perform consistently precisely on such slices with noticeable different scales.
Sensors 21 07877 g001
Figure 3. A four-layer Dense Block. The growth rate is 12. The node is a combination of a BN layer, a ReLU, and a convolutional operator. K means the node takes k inputs. The transition layer depends on the specific DB. In DCNet, the transition layers of the first two DBs (DB1 and DB2) are a combination of a convolutional operator and an average pooling layer, and the transition layers of the other DBs (DB3, DB4, and DB5) are a convolutional operator.
Figure 3. A four-layer Dense Block. The growth rate is 12. The node is a combination of a BN layer, a ReLU, and a convolutional operator. K means the node takes k inputs. The transition layer depends on the specific DB. In DCNet, the transition layers of the first two DBs (DB1 and DB2) are a combination of a convolutional operator and an average pooling layer, and the transition layers of the other DBs (DB3, DB4, and DB5) are a convolutional operator.
Sensors 21 07877 g003
Figure 4. An example of outputs at each level of the decoder network. S i represents the same as that in Figure 2. G is ground truth. Please note that the boundaries of S 4 and S 5 are smoother than the boundaries of S 1 and S 2 caused by the loss of low-level spatial information. However, S 1 and S 2 have a few dots mislabeled outside the main tumor region labeled resulting from the lack of context information. In addition, S 0 fuses them, so its boundary is relatively clear without noticeable mislabeled dots.
Figure 4. An example of outputs at each level of the decoder network. S i represents the same as that in Figure 2. G is ground truth. Please note that the boundaries of S 4 and S 5 are smoother than the boundaries of S 1 and S 2 caused by the loss of low-level spatial information. However, S 1 and S 2 have a few dots mislabeled outside the main tumor region labeled resulting from the lack of context information. In addition, S 0 fuses them, so its boundary is relatively clear without noticeable mislabeled dots.
Sensors 21 07877 g004
Figure 5. Visual comparison of segmentation results of DCNet and baseline methods. Please note that the segmentation results of our DCNet are more sharper when compared to U-Net and U-Net++. Additionally, the boundaries of the results of DCNet also fit ground truth better than the baseline models.
Figure 5. Visual comparison of segmentation results of DCNet and baseline methods. Please note that the segmentation results of our DCNet are more sharper when compared to U-Net and U-Net++. Additionally, the boundaries of the results of DCNet also fit ground truth better than the baseline models.
Sensors 21 07877 g005
Figure 6. The structures of DCNet and its variants for ablation studies. 5 DBs represent five Dense Blocks and GT meaning ground truth. 5 CNNs means five DBs without dense connection. Blue lines are skipped connections. PPM is the pyramid pooling module. The architecture of the decoder is provided in Section 3.2.2. The models with five red arrows towards decoder have the integrated loss function.
Figure 6. The structures of DCNet and its variants for ablation studies. 5 DBs represent five Dense Blocks and GT meaning ground truth. 5 CNNs means five DBs without dense connection. Blue lines are skipped connections. PPM is the pyramid pooling module. The architecture of the decoder is provided in Section 3.2.2. The models with five red arrows towards decoder have the integrated loss function.
Sensors 21 07877 g006
Figure 7. Comparison of the evaluation metrics of DCNet-A and its variants. This is a summary of Table 2, Table 3, Table 4 and Table 5. The results show PPM contributes most to enhancing the performance of DCNet.
Figure 7. Comparison of the evaluation metrics of DCNet-A and its variants. This is a summary of Table 2, Table 3, Table 4 and Table 5. The results show PPM contributes most to enhancing the performance of DCNet.
Sensors 21 07877 g007
Table 2. Quantitative evaluation metrics of DCNet and its variants. The learning rate is 0.0001 , and the batch size is 8. The architecture of the models is illustrated in Figure 6. All the models have been trained for 300 epochs on the same training set and are evaluated on the same testing set. The best results are shown in boldface. The results show DCNet-A outperforms the other comparative models in terms of the majority of all the metrics except SPE. So, the effectiveness of PPM and skip-connection is validated.
Table 2. Quantitative evaluation metrics of DCNet and its variants. The learning rate is 0.0001 , and the batch size is 8. The architecture of the models is illustrated in Figure 6. All the models have been trained for 300 epochs on the same training set and are evaluated on the same testing set. The best results are shown in boldface. The results show DCNet-A outperforms the other comparative models in terms of the majority of all the metrics except SPE. So, the effectiveness of PPM and skip-connection is validated.
MethodsDSC ↑SEN ↑SPE ↑ PPV ↑VOE ↓RVD ↓
DCNet-A (original)0.7730.8540.9950.7320.3630.321
DCNet-B (no PPM)0.2900.2550.9920.4120.8250.490
DCNet-C (no skip-connection)0.6680.6730.9950.7050.4930.338
Table 3. Quantitative evaluation metrics of DCNet-A with the integrated loss function and DCNet-F with the single loss function. The learning rate is 0.0001 , and the batch size is 8. The loss function of DCNet-F is defined as 6 J ( S 0 ) , while the loss function of DCNet-A is defined as Equation (5). Both models have been trained for 300 epochs on the same training set and are evaluated on the same testing set. The better results are shown in boldface. The results show DCNet-A outperforms the other comparative models in terms of the majority of all the metrics except RVD. Hence the integrated loss function contributes to boosting the performance of DCNet.
Table 3. Quantitative evaluation metrics of DCNet-A with the integrated loss function and DCNet-F with the single loss function. The learning rate is 0.0001 , and the batch size is 8. The loss function of DCNet-F is defined as 6 J ( S 0 ) , while the loss function of DCNet-A is defined as Equation (5). Both models have been trained for 300 epochs on the same training set and are evaluated on the same testing set. The better results are shown in boldface. The results show DCNet-A outperforms the other comparative models in terms of the majority of all the metrics except RVD. Hence the integrated loss function contributes to boosting the performance of DCNet.
MethodsDSC ↑SEN ↑SPE ↑PPV ↑VOE ↓RVD ↓
DCNet-A (integrated loss function)0.7730.8540.9950.7320.3630.321
DCNet-F (single loss function)0.4100.4190.9900.4210.7360.281
Table 4. Quantitative evaluation metrics of DCNet and its variants with different numbers of DBs. The learning rate is 0.0001 , and the batch size is 8. The encoder network of DCNet-D has four DBs. The transition layers of the DBs remain the same as the transition layers of the first four DBs in DCNet-A. A DB is added between DB5 and PPM in DCNet-F. The sizes of the outputs of the additional DB remain 64 × 64 , while the number of channels is changed to 140. All models have been trained for 300 epochs on the same training set and are evaluated on the same testing set. The best results are shown in boldface. The results validate the optimal number of DBs is five.
Table 4. Quantitative evaluation metrics of DCNet and its variants with different numbers of DBs. The learning rate is 0.0001 , and the batch size is 8. The encoder network of DCNet-D has four DBs. The transition layers of the DBs remain the same as the transition layers of the first four DBs in DCNet-A. A DB is added between DB5 and PPM in DCNet-F. The sizes of the outputs of the additional DB remain 64 × 64 , while the number of channels is changed to 140. All models have been trained for 300 epochs on the same training set and are evaluated on the same testing set. The best results are shown in boldface. The results validate the optimal number of DBs is five.
MethodsDSC ↑SEN ↑SPE ↑PPV ↑VOE ↓RVD ↓
DCNet-D (Four DBs)0.4050.4060.9890.4640.7390.377
DCNet-A (Five DBs, original)0.7730.8540.9950.7320.3630.321
DCNet-E (Six DBs)0.4610.4520.9910.5070.6950.282
Table 5. Quantitative evaluation metrics of DCNet-A with dense connection and DCNet-G without dense connection. The learning rate is 0.0001 , and the batch size is 8. Both models have been trained for 300 epochs on the same training set and are evaluated on the same testing set. The better results are shown in boldface. The results show DCNet-A outperforms the comparative model in terms of all the metrics. Hence the dense connection contributes to boosting the performance of DCNet.
Table 5. Quantitative evaluation metrics of DCNet-A with dense connection and DCNet-G without dense connection. The learning rate is 0.0001 , and the batch size is 8. Both models have been trained for 300 epochs on the same training set and are evaluated on the same testing set. The better results are shown in boldface. The results show DCNet-A outperforms the comparative model in terms of all the metrics. Hence the dense connection contributes to boosting the performance of DCNet.
MethodsDSC ↑SEN ↑SPE ↑PPV ↑VOE ↓RVD ↓
DCNet-A (dense connection)0.7730.8540.9950.7320.3630.321
DCNet-G (without dense connection)0.3340.3160.9920.3920.7860.381
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, Y.; Han, G.; Liu, X. DCNet: Densely Connected Deep Convolutional Encoder–Decoder Network for Nasopharyngeal Carcinoma Segmentation. Sensors 2021, 21, 7877. https://doi.org/10.3390/s21237877

AMA Style

Li Y, Han G, Liu X. DCNet: Densely Connected Deep Convolutional Encoder–Decoder Network for Nasopharyngeal Carcinoma Segmentation. Sensors. 2021; 21(23):7877. https://doi.org/10.3390/s21237877

Chicago/Turabian Style

Li, Yang, Guanghui Han, and Xiujian Liu. 2021. "DCNet: Densely Connected Deep Convolutional Encoder–Decoder Network for Nasopharyngeal Carcinoma Segmentation" Sensors 21, no. 23: 7877. https://doi.org/10.3390/s21237877

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop