Next Article in Journal
The Power of Images and the Logics of Discovery in Psychiatric Care
Previous Article in Journal
Effects of Transcranial Direct Current Stimulation Combined with Treadmill Training on Kinematics and Spatiotemporal Gait Variables in Stroke Survivors: A Randomized, Triple-Blind, Sham-Controlled Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Axial Attention Convolutional Neural Network for Brain Tumor Segmentation with Multi-Modality MRI Scans

1
Shandong Key Laboratory of Medical Physics and Image Processing, Shandong Institute of Industrial Technology for Health Sciences and Precision Medicine, School of Physics and Electronics, Shandong Normal University, Jinan 250358, China
2
School of Environment and Energy, South China University of Technology, Guangzhou 510006, China
*
Author to whom correspondence should be addressed.
Brain Sci. 2023, 13(1), 12; https://doi.org/10.3390/brainsci13010012
Submission received: 15 November 2022 / Revised: 13 December 2022 / Accepted: 18 December 2022 / Published: 21 December 2022
(This article belongs to the Section Computational Neuroscience and Neuroinformatics)

Abstract

:
Accurately identifying tumors from MRI scans is of the utmost importance for clinical diagnostics and when making plans regarding brain tumor treatment. However, manual segmentation is a challenging and time-consuming process in practice and exhibits a high degree of variability between doctors. Therefore, an axial attention brain tumor segmentation network was established in this paper, automatically segmenting tumor subregions from multi-modality MRIs. The axial attention mechanism was employed to capture richer semantic information, which makes it easier for models to provide local–global contextual information by incorporating local and global feature representations while simplifying the computational complexity. The deep supervision mechanism is employed to avoid vanishing gradients and guide the AABTS-Net to generate better feature representations. The hybrid loss is employed in the model to handle the class imbalance of the dataset. Furthermore, we conduct comprehensive experiments on the BraTS 2019 and 2020 datasets. The proposed AABTS-Net shows greater robustness and accuracy, which signifies that the model can be employed in clinical practice and provides a new avenue for medical image segmentation systems.

Graphical Abstract

1. Introduction

Brain tumors, defined as abnormal cells that grow and multiply uncontrollably inside the brain, are not only dangerous to health but also lead to death as the tumor spreads malignantly. Brain tumors can be grouped into secondary and primary forms [1]. The former refers to tumors that grow elsewhere in the body and metastasize to the brain; the latter refers to tumors that originate from the brain tissues or the immediate surroundings of the brain and have a higher incidence. Glioma, the most prevalent primary cancer, accounting for 78 percent of malignant brain tumors, is usually treated with surgical resection, radiotherapy, and chemotherapy. The World Health Organization (WHO) classifies gliomas into two categories: low-grade gliomas (LGG) and high-grade gliomas (HGG), including grades I to IV [2]. Grade I refers to benign tumors that can be resected from the skull and are curable. Grades II and III gradually become malignant, while grade IV refers to highly malignant tumors with the worst prognosis. Early detection and diagnosis of gliomas can help experts to develop accurate treatment plans to prolong survival. Magnetic resonance imaging (MRI) [3] is the examination method of choice for the diagnosis and treatment of gliomas, mainly due to its presenting detailed brain structural information by enabling tomography in any direction and generating non-invasive multi-modality images [4]. Four MRI modalities (Figure 1) are most commonly used to label brain tumors. Brain tumor segmentation aims to mark the tumor regions using an MRI of the brain, playing an essential role in clinical diagnostics. Accurately identifying tumors from MRI images relies on professional skills [5] and the experience level of radiologists [6]. Furthermore, the manual diagnosis of massive MRI images by radiologists is a time-consuming task. Therefore, an efficient and accurate automatic method is crucial for assisting doctors in segmenting brain tumors from a mass of MRI data.
Taking the above into consideration, we propose an axial attention brain tumor segmentation network (AABTS-Net) with a hybrid loss function, automatically segmenting tumor subregions based on multi-modality MRI. An axial attention mechanism (AAM) has been employed, enhancing the ability to model global dependencies. It considers the relationships of pixels to each other on different dimensions, using long-range cues to guide the segmentation while having fewer parameters and taking less GPU memory. The AAM consists of a depth self-attention module, a height self-attention module, and a width self-attention module. The depth self-attention module enhances the relationships of pixels to each other on the depth dimension. The height self-attention module and width self-attention module capture the relationships of pixels to each other on the height and width dimensions. A deep supervision (DS) mechanism is employed in the AABTS-Net, addressing the gradient vanishing, and guiding the generation of better feature representations at each layer of the network. The training loss of the network is the weighted sum of the loss of the intermediate and last outputs. The hybrid loss function is applied for the purposes of increasing the sensitivity of segmenting tumors, including a batch-based Dice loss for improving training stability on the imbalanced samples and a binary cross-entropy loss for handling issues arising from varying tumor sizes. The proposed hybrid loss function can direct our model toward paying more attention to the tumor region and achieving ideal segmentation results, overcoming the problem of tumors of different sizes and even in the case of small target tumors. Regarding segmentation performance, the proposed AABTS-Net can make reliable results on the BraTS 2021 validation dataset for segmenting three subregions (ET, TC, and WT), achieving 0.830, 0.861, and 0.922 in DSC, respectively. Furthermore, we conducted comparative experiments involving other state-of-the-art methods on the BraTS 2019 and 2021 validation datasets, showing that our method had greater robustness and accuracy.
Specifically, the main contributions of this study include:
(1) Applying AAM to improve segmentation accuracy makes full use of the relationships of pixels on different dimensions and enhances the ability to model global dependencies. Notably, an axial attention mechanism, presenting a novel method to solve computational complexity, is employed to capture plenty of the semantic information that is essential for segmenting the boundary of the tumor and retaining more detailed tumor information.
(2) Handling the vanishing gradient and improving the feature expression ability of the model by applying the deep supervision mechanism on the decoder to guide the model to generate better feature representations.
(3) Suppressing the effect of the dataset class imbalance by introducing a hybrid loss function to emphasize the lesion area. The hybrid loss function can guide our model to pay closer attention to the tumor region to achieve ideal segmentation results, overcoming the problem of tumors of different sizes and even small target tumors.

2. Related Work

2.1. DL-Based Methods for Brain Tumor Segmentation

The convolutional neural network (CNN), an effective deep learning (DL)-based architecture, has attracted more and more attention from researchers for image-processing tasks, such as image classification and semantic image segmentation. Compared with ML-based methods, the most significant advantage of a CNN is that it automatically learns representative and complex feature representations from the raw images, forming a feature-learning model with high robustness and adaptability. A dual-path architecture, combining local and global information, was designed by Kamnitsas et al. [7] to process multi-scale images simultaneously and extend the fully connected CRF to three dimensions (3D) for handling arbitrarily large neighborhoods. The disadvantage of this method is that the post-processing process is complicated, and the accuracy will decrease after removing the post-processing operation. Havaei et al. [8] proposed a two-channel CNN for segmentation, which includes convolutional and fully connected layers. However, the method has more parameters, and the complex structure of the model makes the network hard to train. Pereira et al. [9] adopted a deeper CNN structure and multiple 3 × 3 convolution kernels to replace the 7 × 7 and 5 × 5 convolution kernels. The 3 × 3 convolution kernel was applied to improve the operation speed of CNN and enhance the ability to extract features, making the segmentation accuracy reach about 87%. Although the method simplifies the network structure, it misses a massive quantity of local–global information, leading to low segmentation accuracy. Zhao et al. [10] used image slices with FCNN parameters to train three models and adopted a voting-based strategy to fuse the three models for final segmentation. Chen et al. [11] introduced an FCNN-based multi-scale receptive field and a densely connected blocks-based hierarchical architecture to consider the different types of brain tumors, using a block-level training strategy to alleviate the class imbalance. The two-stage method alleviates the class imbalance, while the segmentation accuracy shows no significant improvement. CNN, a widely adopted architecture for segmentation, only obtained limited accuracy improvement due to the smaller training dataset available. To address this problem, Ronneberger et al. [12] proposed the U-Net via modifying and extending the FCN, enabling it to use fewer training images and produce more accurate segmentation. Wang et al. established a 3D U-Net-based model using the strategy of brain normalization and patching [13], achieving 0.737, 0.807, and 0.894 in terms of DSC, respectively. To reduce computational complexity and enhance feature representation, Cheng et al. [14] further proposed MECU-Net by applying multi-scale feature fusion modules, fewer down-sampling channels, and cascaded architecture to 3D U-Net. Chen et al. proposed a separable 3D U-Net architecture using separable 3D convolutions [15]. Jiang et al. [16] proposed an end-to-end two-stage cascaded U-Net architecture to segment brain tumor substructures from coarse to fine and won the first prize in the BraTS2019. These methods show the effectiveness of the U-shaped structure; therefore, we chose the U-shaped structure as our network framework.

2.2. The Attention-Based Module for Brain Tumor Segmentation

Various attention-based modules have been proposed by researchers, facilitating segmentation accuracy by enhancing the ability to identify relevant feature representations. Zhou et al. [17] applied an attention mechanism to the encoder–decoder structure to fuse features from cross-modality to emphasize the most closely related features. Xu et al. [18] proposed a CH-UNet for brain tumor segmentation by applying a corner attention module (CAM) and high-dimensional perceptual loss (HDPL) to U-Net. The CAM can capture rich contextual dependencies and extract complementary inter- and intra-slice information. The HDPL preserves local consistency and explores perceptual similarities to fine the predicted boundary. These two methods show that the attention mechanism plays an important role in focusing on task-related features. Zhang et al. combined the residual module and attention gate with the original U-Net architecture [19] to extract rich semantic information. However, this method ignores the local–global information and leads to poor performance. Mazumdar et al. proposed an efficient spatial attention (ESA) block containing a depth-wise separable convolutional layer and a lightweight spatial attention module, using it to build an efficient spatial attention network (ESA-Net) [20], an improved variant of the popular U-Net. Kong et al. proposed a 3D, fully convolutional network (FCN) with a dual attention mechanism [21], segmenting different gliomas simultaneously. The multi-inception residual attention U-Net (MIRAU-Net), developed by AboElenein et al. [22], is based on an encoder–decoder architecture. Inception Residual is applied to connect the encoder–decoder, to reduce the distance between their feature maps for further enhancing segmentation performance. However, due to the limitations of convolution operations, CNN-based attention has a weak ability to learn global dependencies, resulting in poor performance in the context of brain tumor segmentation. To solve the problem, Wang et al. proposed a novel TransBTS, based on the encoder–decoder structure [23], applying a transformer to 3D CNN, improving the performance by combining the CNN and transformer while exploiting and modeling local–global information. In addition, Jia et al. proposed a combined CNN-transformer model called BiTr-UNet, with specific modifications [24] for multi-modality MRI segmentation. Although transformer-based models can capture global correlation, transformer-based models are not kind to GPU memory. Moreover, the computational complexity of using transformers is relatively high when handling 3D medical images. Hence, we propose an efficient AABTS-Net, which combines axial attention and CNN to automatically segment tumor subregions by taking into account the required computational complexity.

3. Methodology

3.1. The Structure of the AABTS-Net

The suggested U-shaped encoder–decoder structure with fewer parameters shows good segmentation performance in many image segmentation tasks. Similar to the 3D U-Net (39), three-dimensional operations were employed in the AABTS-Net (Figure 2), focusing on detailed and local feature representations that are suitable for brain tumor (small object) segmentation. In detail, AABTS-Net consists of an encoding path, a decoding path, and skip connections. The encoding path, including five basic convolutional blocks and five down-samplings, is used to extract feature representations at different resolution levels and reduce the resolution of feature representations, respectively. The basic convolutional block contains two consecutive convolution layers, applying Leaky ReLU (LReLU) with a slope of 0.01 as the activation function after each convolution operation. Stridden convolution operations have been considered a popular method applied in the encoding path for down-sampling. The initial size of the feature representation is 128 × 128 × 128, and the resolution of the feature representation doubles for each down-sampling operation, with a minimum of 4 × 4 × 4. The initial number of filters is 32, and the number of filters doubles with each down-sampling operation, up to a maximum of 320. A skip connection is employed to fuse feature representations from the same resolution levels of the encoding and decoding paths. Similar to the encoding path, the decoding path consists of five basic convolutional blocks and five up-samplings that employ transposed convolutions to restore the resolution of the feature representation. The sigmoid activation has replaced the softmax nonlinearity of the last layer. The input requires four-channel 3D images (4 × 128 × 128 × 128) merged by four modalities, while the output comprises three-channel predicted segmentation maps (3 × 128 × 128 × 128).

3.2. The Axial Attention Mechanism

As an effective modeling method for capturing global context information, global feature interaction is one of the greatest advantages of self-attention [25]. The self-attention module captures global context information, resulting in a large memory capacity and high computational complexity. When the self-attention mechanism is applied to three-dimensional data, it shows higher computational complexity. Most previous attention modules reduce the ability to model position-dependent interactions without considering the positional information that is crucial for capturing spatial structure or shape. Axial attention [26], applying self-attention to each axis of the input independently, has been proposed to reduce the computational complexity. Therefore, we employed the axial attention mechanism for three-dimensional medical image segmentation, capturing the long-term dependencies of the brain MRI and computing a representation of the local–global information. In detail, axial attention is applied independently on each axis (depth, height, and width). The AAM receives two inputs; one is the feature representation (with all contextual and spatial information) of the same resolution level, passed from the skip connection, and the other is the feature representation from the previous resolution level.
Figure 3 shows the AAM. Specifically, the features of the skip connection are represented as F s c D C H W , where D, C, H, and W denote the depth, channel number, height, and width of the feature map, respectively. The feature of the previous layer is represented as F p r e D 2 C H 2 W 2 . We first perform a transposed convolution on the F p r e to obtain the feature F D C H W . With respect to self-attention, it performs the attention process with each input pixel without considering the input position information, which will lose the original position information. Therefore, we perform axial position embedding on the feature map to ensure position information. The depth self-attention module (Figure 3) is first applied to compute the attention maps representing the relationships of pixels to each other on the different slices, which calculates a similarity matrix along the depth dimension. In detail, we obtain three feature spaces q ( f ) D C ^ H W , k ( f ) D C ^ H W ,and v ( f ) D C H W by conducting linear operations to transform the input feature map F , which can be represented by Equations (1)–(3).
q ( f ) = L i n e a r ( F ) = F W q
k ( f ) = L i n e a r ( F ) = F W k
v ( f ) = L i n e a r ( F ) = F W v
Here, the W q , W k , and W v are matrixes to improve the fitting ability, which can be obtained by training.
Then, q ( f ) and k ( f ) are transposed from D C ^ H W to ( H W ) D C ^ . The inner product is employed to compute the correlation matrix, which is shown in Equation (4). The depth correlation matrix δ j , i is normalized by the softmax function shown in Equation (5).
δ j , i = q ( f i ) T k ( f j )
r j , i = S o f t m a x ( δ j , i ) = e δ j , i i = 1 N e δ j , i
The deep value of δ j , i , with a dimension of D D , denotes the depth correlation matrix.
Finally, the depth correlation matrix r j , i and the v ( f ) were performed by the inner product to obtain the depth self-attention feature maps, which can be formulated as in Equation (6). To restore the original feature space, F s a D is performed by a linear operation.
F s a D = i = 1 N v ( f i ) × r j , i
Similarity, the height self-attention is obtained by a height self-attention module and is present as F s a D H , while the width self-attention is present as F s a D H W . The attention features output by the axis attention fusion module are denoted as Equation (7) and the output of the axial attention mechanism is represented in Equation (8):
F a t t = F s a D H W + F
F o u t = F a t t + F s c
Since the method cannot be applied at the highest resolution feature representation (128 × 128 × 128), the AAM is employed at four lower resolutions. The number of attention heads and the dimension of each head, initially set to 4 and 16, were doubled as the resolution decreased.

3.3. The Deep Supervision (DS) Mechanism

A deep supervised learning scheme [27,28,29,30] is beneficial when training a CNN, enabling the network to collect gradients from the last and intermediate layers and propagate the error information layer by layer during the training process. DS, including three auxiliary output branches, is employed in the AABTS-Net, addressing the gradient’s vanishing and guiding the generation of better feature representations at each layer of the network. In detail, the auxiliary branch, achieved by three sigmoid outputs, is added in the middle three resolution levels (Figure 2). Notably, AABTS-Net contains four sigmoid outputs. Only the last output is the predicted segmentation image. The intermediate output, after being restored to the highest resolution level by up-sampling, is calculated in the auxiliary loss branch. The training loss of the network is the weighted sum of the loss of the intermediate and last outputs, giving each output a weight that decreases exponentially (divided by 2) with lower resolution, which makes higher-resolution outputs more weighted in the loss. The proposed deep supervision mechanism enables the network to generate better feature representation by collecting error information from more layers and by propagating misinformation to earlier layers, improving the expressiveness of the network and its segmentation performance.

3.4. The Hybrid Loss Function

The ratio between tumor area and background varies widely in the BraTS2021 training dataset, representing the imbalance of samples, making it difficult to precisely segment the lesion area [20,31]. The objective loss function, which also affects the segmentation performance of the network, is widely employed for handling the severe class imbalance of the datasets. A reasonable loss function should be designed for network optimization while retaining the complete information of the image, achieving the purpose of accurately segmenting small tumors. Many loss functions, such as Dice loss, cross-entropy loss, and focal loss, have been proposed for employment in segmentation, handling the class imbalance of samples. In this paper, a suitable hybrid loss function, combining the Dice and binary cross-entropy losses, is employed for solving class imbalance during the training process; it is computed at the lower-resolution auxiliary branch output and the final output (Equation (11)). Notably, the Dice loss (Equation (9)), referring to the batch-based Dice used here, is calculated by treating all samples in a batch as one larger sample. The common Dice loss, exhibiting gradient instability on highly imbalanced samples, is modified to the batch-based Dice to improve training stability. The cross-entropy (CE) loss function, aiming to handle issues arising from varying tumor sizes, is widely applied for increasing the sensitivity of segmenting tumor images. Since the activation function of the last layer is sigmoid, the corresponding output is the binary classification label. Then, a binary cross-entropy (BCE) loss function is employed (Equation (10)). Overall, the proposed hybrid loss function can guide our model to pay more attention to the tumor region to achieve ideal segmentation results, overcoming the problem of mapping tumors of different sizes and even small target tumors. These metrics are defined as follows:
D i c e = 1 2 y p y t y p + y t
B C E = λ y t log ( y p ) ( 1 y t ) log ( 1 y p )
T o t a l = H 1 + α H 2 + β H 3 + γ H 4
where y t represents the ground truth and y p represents the prediction result. α , β , and γ are the weight coefficient of the three auxiliary branches, and they are set to 0.5, 0.25, and 0.125, respectively (the higher the resolution, the greater the weight coefficient).

3.5. Experiments

3.5.1. Datasets and Pre-Processing

In this work, the BraTS benchmark dataset [32,33,34] is applied to evaluate the proposed AABTS-Net. The BraTS dataset (Table 1) includes two sub-datasets, which comprise the training dataset and the validation dataset. In detail, the BraTS 2019 training dataset includes 335 cases, while the validation dataset includes 125 cases. The BraTS 2021 training dataset includes 1251 cases, and the validation dataset includes 219 cases. Each case of the dataset includes four MRI modalities, while a ground truth model included three different tumor classes (non-enhanced and necrotic tumor cores—label 1, peritumoral edema—label 2, and GD-enhanced tumors—label 4). The ground truth annotations of the training dataset are publicly available and are created by expert neuroradiologists. Since the annotations of the validation dataset are unavailable, we tested the performance by uploading the predicted segmentation results to the official platform provided by the organizer.
All cases in the datasets, having undergone basic preprocessing by the organizers, were interpolated to 1 × 1 × 1   mm 3 , co-registered, and skull-stripped. On this basis, we only performed relatively simple data preprocessing. First, the 3D raw data were cropped to a nonzero region. One case in the BraTS dataset showed relatively more black background (gray value 0). Since this area has no information, cropping these areas did not affect the final segmentation result. However, the process can significantly reduce the image size, avoid useless calculations, and improve computational efficiency. Specifically, it is necessary to find a minimum three-dimensional bounding box in the image, assign the value outside the bounding box area as 0, and use this bounding box to crop the raw data. Moreover, we used normalization in a standard space for each case in the dataset. In detail, employing a z-score ( X z s c o r e = X μ σ , where μ and σ are mean and standard deviation) for the normalization operation accelerates the training. Notably, under the conditions that the image size is reduced by more than 1/4, the normalization operation is only performed on the non-zero regions. Dynamic data augmentation, a technique intended to prevent overfitting while training a CNN, is also employed, improving the robustness and the generalization ability of the AABTS-Net. Specifically, five types of data augmentation, comprising random rotation, random scaling, elastic deformation, gamma augmentation, and additional brightness enhancement, with a probability of 0.3, were applied during the training. The random scaling factor is (0.65, 1.6), using a single scaling factor for each axis.

3.5.2. Evaluation Metrics

Three evaluation metrics, including Dice similarity coefficients (DSC), the Hausdorff distance (HD), and Sensitivity, were employed to evaluate the performance of the models. The DSC is a widely considered criterion, quantitatively assessing the 3D predicted segmentation of the model. It provides the similarity or overlap of two samples, with a value in the range [0–1], and is defined as in Equation (12). The higher the value, the more similar the two samples are, indicating more accurate performance between the predicted segmentation and manual segmentation. Additionally, when affected by uncertain factors, such as shape, location, and size, the model performance also depends on the accuracy of segmenting tumor boundaries. Considering the fact that the DSC is sensitive to the inner filling of the mask, the HD, computing the distance between the surface of the prediction regions and the ground truth, is employed for assessing the performance of the boundary segmentation. The HD is defined as follows (Equation (13)). Sensitivity measures the ability of the model to segment the region of interest by calculating the proportions of true positives in the predicted segmentation results and ground truth and is given by Equation (14).
S C = 2 T P 2 T P + F P + F N
H D ( G , P ) = m a x   { s u p s ϵ S i n f r ϵ R d ( s , r ) ,   s u p r ϵ R i n f s ϵ S d ( s , r ) }  
S e n s i t i v i t y = T P T P + F N
where F N , F P , and T P are the number of false negative voxels, false positive voxels and true positive voxels. Here, d ( · , · ) is the function that computes the distance between points s and r , while s u p represents the supremum and i n f denotes the infimum; s and r denote the points on surface S of the ground-truth regions and surface R of the predicted regions, respectively.

3.5.3. The Training Details

The model, established on the Pytorch framework, was implemented using a single NVIDIA Tesla P100 GPU, with 16 GB of memory. The Adam optimizer [35], an essential optimization algorithm, was employed for gradient descent. The initial learning rate of 10−4 is progressively multiplied by ( 1 e p o c h max e p o c h   ) 0.9 during the training. The k-fold cross-validation procedure was used to estimate the performance of the model on new data. In our experiments, five-fold cross-validation is set up during the training of the model. Specifically, the training dataset was divided (randomly) into five parts, training for four parts and testing for one part. Each epoch consisted of 250 mini-batches, and training did not end until the maximum epoch (300) was achieved (Table 2).

4. Results

4.1. Comparison with State-of-the-Art Methods

4.1.1. The Experiment Results on the BraTS 2019 Dataset

The BraTS 2019 dataset was applied to evaluate the effectiveness and robustness of the proposed AABTS-Net. The qualitative comparison is shown in Figure 4, from left to right: four MRI modalities, corresponding ground truth, 3D-UNet, Attention-UNet, and our method. From Figure 4, it can be seen that these methods can roughly segment the tumor region. However, detailed information on the tumor boundary is missing. In detail, the segmentation results generated by 3D-UNet have an obviously inaccurate tumor boundary. The segmentation results generated by Attention-UNet improved the segmentation, but the tumor boundary still demonstrates the problem of missing information. The segmentation results produced by our proposed AABTS-Net are more accurate and are similar to the ground truth results, effectively refining the boundaries of brain tumors and enhancing the detailed features.
In addition, we evaluated the model on the BraTS 2019 validation dataset. The BraTS 2019 validation dataset contains 125 cases. It is worth noting that the BraTS 2019 training dataset and the BraTS 2019 validation dataset are two independent datasets provided by the organizer. Since the ground truth of the BraTS 2019 validation dataset is unavailable, we uploaded the segmentation results to the official website and obtained the quantitative results from the platform. We also performed comparative experiments (Table 3) between our method and other state-of-the-art methods. From Table 3, it is clear that our method achieves DSC values of 0.777, 0.838, and 0.911, at ET, TC, and WT, respectively, which values are higher than those of other methods. The DSC increased by 3.7% to 24.7% in ET, and the DSC increased by 0.8% to 15.6% and 1% to 6.9% in TC and WT, respectively. Compared with the other methods, our method shows great superiority in both metrics, with significant improvements. The comparison results show that our method achieves superior performance.

4.1.2. The Experiment Results on the Brats 2021 Dataset

Comparison of the segmentation performance between our method and other state-of-the-art methods is conducive to analyzing the advantages and disadvantages of the model and achieving improvements in segmentation performance. The predicted results (see Table 4) on the BraTS 2021 validation dataset, showing the three tumor subregions, were analyzed by employing two evaluation metrics for assessing the performance of the model. As shown in Table 4, the DSC at three tumor sub-regions was significantly distinct for the DSC 0.830, 0.861, and 0.922 at ET, TC, and WT, showing a clear upward trend that represents an accurate segmentation: the larger the value, the more precise the segmentation. The HD, being different from the DSC, posed a downward trend at three tumor sub-regions, with the maximum value of 17.728 observed at ET and the minimum value of 3.996 observed at WT, indicating that the AABTS-Net method can precisely segment the boundary of the tumor. Compared with other methods [20,42,43,44,45,46], the DSC increased by 0.4% to 5.9% in WT. The DSC increased by 1.3% to 13.7% and 1% to 14.6% in ET and TC, respectively. This demonstrates that our method achieved a promotive performance in terms of tumor sub-region segmentation. In addition, compared with other state-of-the-art UNet-based models [24,47,48], our model still achieved promising segmentation performance by analyzing the different metrics. Our model has fewer parameters and refines the boundaries of the tumor subregions by enhancing the local and global feature information.
Furthermore, we segmented the cases (shown in Figure 5) selected from the BraTS 2021 dataset, visualizing the predicted segmentation results. Specifically, the first case segmented by our model retained detailed information and showed many similarities with the GT. The predicted segmentation results of AABTS-Net were not significantly affected, as the input MRI modality is blurred, as shown in Figure 5 (the fourth row). In conclusion, the results predicted by AABTS-Net, effectively refining the boundaries of the tumor subregions and enhancing the local and global feature information, are not significantly different from those of the GT. This phenomenon suggests that the AABTS-Net model can learn better feature representations to localize tumors and guide the model to produce accurate segmentation results. In general, the AABTS-Net model achieved accurate brain tumor segmentation and can be employed to assist doctors in diagnosis.

4.2. The Analysis of Key Components

4.2.1. Analysis of the Axial Attention Mechanism

To verify the effectiveness of the AAM, we conducted ablation experiments (see Table 5) to compare the proposed AABTS-Net with the same network without employing the AAM. The DSC of subregions ET, TC, and WT exhibited an upward trend. The reason for this may be that the structural distribution of the ET region is complex, and the model found it difficult to distinguish between the edema region and the tumor core and the surrounding area, enhancing the tumors. Segmentation of the different tumor subregions improved as the AAM was added to the network. The network with the AAM improved DSC by 4.6%, 2.6%, and 3.9% in the WT, TC, and ET regions, respectively. According to the results (Figure 6 of the two models on the three metrics, the corresponding DSC and Sensitivity values were increased significantly by applying the AAM to the model. Furthermore, we visualized the segmentation results generated by AABTS-Net and AABTS-Net, removing the AAM. From the predicted segmentation results (Figure 7) of the two cases, it can be seen that both models can segment the WT region. However, the AABTS-Net’s removal of the AAM, segmenting the ET and TC, obtained a rough edge (red arrows). By removing the AAM, the model’s ability to recognize fuzzy boundaries was weakened, thereby reducing the segmentation accuracy of ET and TC. The segmentation results predicted by AABTS-Net were similar to ground truth, which shows that the proposed AAM improved the segmentation performance. In addition, we have drawn a DSC boxplot to analyze the contribution of the AAM. From Figure 8, we observed that the AAM improved the result significantly in terms of the different tumor regions. In general, the values of the three metrics and the visualized result, showing the better values of the tumor subregions and the predicted result with more detailed information, indicated that the segmentation performance was improved by employing the AAM. This can be attributed to AAM’s ability to model local–global feature information and the resulting enhanced capability of boundary recognition.

4.2.2. Analysis of the Deep Supervision Mechanism

To evaluate the contribution of the DS to the AABTS-Net, we conducted comparative experiments on the BraTS2021 dataset (Table 5). Specifically, Figure 9 shows the experimental results of AABTS-Net with and without the DS on three metrics (DSC, HD, and Sensitivity). It is clear that the DSC and Sensitivity are significantly increased in the scenario where DS is applied to AABTS-Net, which indicates that DS assists the network, to improve the segmentation performance. The DSC value increased by 2.0%, 0.9%, and 2.1% in the WT, TC, and ET regions, respectively. From the WT region, the DSC value (see Table 5) after adding DS was smaller than that after adding AAM. This was mainly due to the fact that the DS was employed for gradually refining the segmentation and retaining more detailed tumor information. However, this detailed information has less of a contribution than the tumor masses to the DSC value. The HD value (the smaller the HD value, the better the performance) decreased by 35.2%, 28.9%, and 26.0% in the WT, TC, and ET regions, respectively. This large reduction in the HD value also indicates that more accurate edge segmentation and more detailed information can be obtained when using DS. The Sensitivity value increased by 1.6%, 2.0%, and 1.7% in the WT, TC, and ET regions, respectively. From Figure 8, it can be seen that the model with the DS showed better segmentation results in the WT and ET regions. Figure 10 shows the segmentation results produced by the AABTS-Net and the AABTS-Net without the DS. From Figure 10, it can be seen that the tumor regions (red arrows) segmented by AABTS-Net have more similarities with the ground truth. This demonstrates that the DS not only guides the model to generate better feature representations from multi-modality but also makes full use of the different feature representations.

4.2.3. Analysis of the Different Loss Functions

Additionally, for class-imbalanced data, choosing an appropriate loss function fulfills an important role in segmentation accuracy. The DSC variations of different loss functions were observed. However, using Dice alone is not conducive to backpropagation, and it is easy to cause parameter oscillation during training, which is not conducive to convergence. Therefore, we considered applying a combined loss function for training. As shown in Figure 6, the segmentation performance metrics at four loss functions (i.e., Focal, Dice, Dice + CE, and Dice + BCE) fluctuated significantly, and three notable changes occurred when segmenting the different tumor subregions. The DSC of the ET sub-region in four loss functions was relatively low. However, the DSC values increased at the Focal, Dice, Dice + CE, and Dice + BCE loss functions. In detail, the average DSC values increased from 0.819 to 0.863 at the Focal and the Dice + BCE loss functions, respectively. In addition, the maximum DSC values of all loss functions are shown, with the DSC values of 0.828, 0.859, and 0.901 in the ET, TC, and WT subregions, respectively. From the Dice + BCE loss function, the DSC values at ET, TC, and WT sub-region also increased significantly. According to the average DSC values (see Figure 11), the DSC values of all loss functions (Focal, Dice, Dice + CE, and Dice + BCE) exhibited an upward trend, while the DSC value of the Dice + BCE loss function was significantly higher than the other loss functions. Overall, the combination of Dice and BCE, which achieved the best segmentation results, was employed for optimizing the network during the training process. The segmentation accuracy of improvement can benefit from the BCE loss function and the modified batch-based Dice loss function, as when handling the class imbalance from the BCE loss function and improving the backpropagation from the combination loss function. Moreover, the hybrid loss function, which played a vital role in segmentation performance, was an innovative and effective strategy for solving the segmenting of tumors of different sizes.

4.3. Analysis of the Bad Tumor Segmentation Results

To further analyze the segmentation performance, we selected five bad segmentation cases and visualized the segmentation results of the proposed AABTS-Net. As shown in Figure 12, each row contained four different modalities for a case, the ground truth, and the segmentation result generated by the AABTS-Net. From the first two cases, we observed that the ET and TC regions segmented by the AABTS-Net are similar to the ground truth. However, the similar characteristics of normal and tumor regions led to the WT region exhibiting over-segmentation. From the third and fourth cases, we observed that the tumor regions generated by the AABTS-Net were significantly different from the ground truth. This may be due to the ambiguity of the MRI modality, making it difficult for the network to identify tumor regions from the MRI modality. In addition, we quantified the cases of bad tumor segmentation in terms of the DSC (Table 6). By analyzing the bad segmentation cases, we realized that the quality of the MRI modality can affect the segmentation performance. Therefore, we considered the impact of data quality on network performance when designing the model and must collect more clear data in a different way to further improve the accuracy by optimizing the tumor margins.

5. Discussion

In this paper, AABTS-Net, an axial attention convolutional neural network, was established for multi-modality MRI brain tumor segmentation. The axial attention mechanism, employed to decrease the high computational complexity of applying the attention mechanism to three-dimensional medical images, capturing plenty of semantic information, achieved accurate segmentation and improved segmentation performance. The deep supervision mechanism guided the AABTS-Net to generate better feature representations, handling the problem of the vanishing gradient easily and improving the stability of the network. The hybrid loss model was employed for dealing with the class imbalance of the data. The AABTS-Net exhibited superior performance on the BraTS 2021 validation dataset, with DSC values of 0.830, 0.861, and 0.922, and HD values of 17.728, 11.178, and 3.996, respectively, at the three tumor subregions (ET, TC, and WT). Furthermore, the quantitative and qualitative experiments demonstrated the accuracy of the AABTS-Net. Experiments using two datasets demonstrated the effectiveness of our model compared to previous work. Our model achieved better robustness and accuracy, which indicated that the AABTS-Net can be employed to assist doctors in diagnosis. However, we observed that our model has some disadvantages in terms of TC region segmentation when we compared our method with the Extending nn-UNet in the context of the HD metric. To resolve this issue, we applied a simple feature extraction method that ignored the shallow feature fusion. In the future, we will consider employing a deeper CNN to extract the features and applying a more flexible feature fusion strategy to improve the accuracy by fusing multi-modality feature representations.

Author Contributions

Methodology, software, writing—original draft, W.T.; data curation, P.H.; project administration, D.L.; visualization, M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China, grant number 61971271, the Jinan City-School Integration Development Strategy Project, grant number JNSX2021023, and the Shandong Province Major Technological Innovation Project, grant number 2022CXGC010502.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and was approved by the Institutional Review Board of the RSNA-ASNR-MICCAI BraTS 2021 challenge for studies involving humans.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The datasets are provided by BraTS Challenges (2019, 2020, and 2021) and are allowed for personal academic research. The specific link to the dataset is https://ipp.cbica.upenn.edu/ (accessed on 3 November 2021).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

References

  1. Holland, E.C. Progenitor Cells and Glioma Formation. Curr. Opin. Neurol. 2001, 14, 683–688. [Google Scholar] [CrossRef] [PubMed]
  2. Kleihues, P.; Burger, P.C.; Scheithauer, B.W. The New Who Classification of Brain Tumours. Brain Pathol. 1993, 3, 255–268. [Google Scholar] [CrossRef] [PubMed]
  3. Bauer, S.; Wiest, R.; Nolte, L.-P.; Reyes, M. A Survey of Mri-Based Medical Image Analysis for Brain Tumor Studies. Phys. Med. Biol. 2013, 58, R97. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Liang, Z.-P.; Lauterbur, P.C. Principles of Magnetic Resonance Imaging: A Signal Processing Perspective; The Institute of Electrical and Electronics Engineers Press: Manhattan, NY, USA, 2000. [Google Scholar]
  5. Jayadevappa, D.; Srinivas Kumar, S.; Murty, D. Medical Image Segmentation Algorithms Using Deformable Models: A Review. IETE Tech. Rev. 2011, 28, 248–255. [Google Scholar] [CrossRef]
  6. Mansourvar, M.; Ismail, M.A.; Herawan, T.; Gopal Raj, R.; Abdul Kareem, S.; Nasaruddin, F.H. Automated Bone Age Assessment: Motivation, Taxonomies, and Challenges. Comput. Math Methods Med. 2013, 2013, 391626. [Google Scholar] [CrossRef] [Green Version]
  7. Kamnitsas, K.; Ledig, C.; Newcombe, V.F.; Simpson, J.P.; Kane, A.D.; Menon, D.K.; Rueckert, D.; Glocker, B. Efficient Multi-Scale 3d Cnn with Fully Connected Crf for Accurate Brain Lesion Segmentation. Med. Image Anal. 2017, 36, 61–78. [Google Scholar] [CrossRef]
  8. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.M.; Larochelle, H. Brain Tumor Segmentation with Deep Neural Networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef] [Green Version]
  9. Pereira, S.; Pinto, A.; Alves, V.; Silva, C.A. Brain Tumor Segmentation Using Convolutional Neural Networks in Mri Images. IEEE Trans. Med. Imaging 2016, 35, 1240–1251. [Google Scholar] [CrossRef]
  10. Zhao, X.; Wu, Y.; Song, G.; Li, Z.; Zhang, Y.; Fan, Y. A Deep Learning Model Integrating Fcnns and Crfs for Brain Tumor Segmentation. Med. Image Anal. 2018, 43, 98–111. [Google Scholar] [CrossRef]
  11. Chen, L.; Wu, Y.; DSouza, A.M.; Abidin, A.Z.; Wismüller, A.; Xu, C. (Eds.) Mri Tumor Segmentation with Densely Connected 3d Cnn. In Medical Imaging 2018: Image Processing; SPIE: Bellingham, WA, USA, 2018. [Google Scholar]
  12. Ronneberger, O.; Fischer, P.; Brox, T. (Eds.) U-Net: Convolutional Networks for Biomedical Image Segmentation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Manhattan, NY, USA, 2015. [Google Scholar]
  13. Wang, F.; Jiang, R.; Zheng, L.; Meng, C.; Biswal, B. (Eds.) 3d U-Net Based Brain Tumor Segmentation and Survival Days Prediction. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2019. [Google Scholar]
  14. Cheng, X.; Jiang, Z.; Sun, Q.; Zhang, J. (Eds.) Memory-Efficient Cascade 3d U-Net for Brain Tumor Segmentation. In International Miccai Brainlesion Workshop; Springer: Manhattan, NY, USA, 2019. [Google Scholar]
  15. Chen, W.; Liu, B.; Peng, S.; Sun, J.; Qiao, X. (Eds.) S3d-Unet: Separable 3d U-Net for Brain Tumor Segmentation. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2018. [Google Scholar]
  16. Jiang, Z.; Ding, C.; Liu, M.; Tao, D. (Eds.) Two-Stage Cascaded U-Net: 1st Place Solution to Brats Challenge 2019 Segmentation Task. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2019. [Google Scholar]
  17. Zhou, T.; Canu, S.; Vera, P.; Ruan, S. Latent Correlation Representation Learning for Brain Tumor Segmentation with Missing Mri Modalities. IEEE Trans. Image Processing 2021, 30, 4263–4274. [Google Scholar] [CrossRef]
  18. Xu, W.; Yang, H.; Zhang, M.; Cao, Z.; Pan, X.; Liu, W. Brain Tumor Segmentation with Corner Attention and High-Dimensional Perceptual Loss. Biomed. Signal Processing Control. 2022, 73, 103438. [Google Scholar] [CrossRef]
  19. Zhang, J.; Jiang, Z.; Dong, J.; Hou, Y.; Liu, B. Attention Gate Resu-Net for Automatic Mri Brain Tumor Segmentation. IEEE Access 2020, 8, 58533–58545. [Google Scholar] [CrossRef]
  20. Mazumdar, I.; Mukherjee, J. Fully Automatic Mri Brain Tumor Segmentation Using Efficient Spatial Attention Convolutional Networks with Composite Loss. Neurocomputing 2022, 500, 243–254. [Google Scholar] [CrossRef]
  21. Kong, D.; Liu, X.; Wang, Y.; Li, D.; Xue, J. 3d Hierarchical Dual-Attention Fully Convolutional Networks with Hybrid Losses for Diverse Glioma Segmentation. Knowl.-Based Syst. 2022, 237, 107692. [Google Scholar] [CrossRef]
  22. AboElenein, N.M.; Piao, S.; Noor, A.; Ahmed, P.N. Mirau-Net: An Improved Neural Network Based on U-Net for Gliomas Segmentation. Signal Processing: Image Commun. 2022, 101, 116553. [Google Scholar] [CrossRef]
  23. Wang, W.; Chen, C.; Ding, M.; Yu, H.; Zha, S.; Li, J. (Eds.) Transbts: Multimodal Brain Tumor Segmentation Using Transformer. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Manhattan, NY, USA, 2021. [Google Scholar]
  24. Jia, Q.; Shu, H. (Eds.) Bitr-Unet: A Cnn-Transformer Combined Network for Mri Brain Tumor Segmentation. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2022. [Google Scholar]
  25. Zhou, Z.; Zhou, Y.; Wang, D.; Mu, J.; Zhou, H. Self-Attention Feature Fusion Network for Semantic Segmentation. Neurocomputing 2021, 453, 50–59. [Google Scholar] [CrossRef]
  26. Wang, H.; Zhu, Y.; Green, B.; Adam, H.; Yuille, A.; Chen, L.-C. (Eds.) Axial-Deeplab: Stand-Alone Axial-Attention for Panoptic Segmentation. In European Conference on Computer Vision; Springer: Manhattan, NY, USA, 2020. [Google Scholar]
  27. Liu, P.; Dou, Q.; Wang, Q.; Heng, P.-A. An Encoder-Decoder Neural Network with 3d Squeeze-and-Excitation and Deep Supervision for Brain Tumor Segmentation. IEEE Access 2020, 8, 34029–34037. [Google Scholar] [CrossRef]
  28. Zhang, Y.; Lu, Y.; Chen, W.; Chang, Y.; Gu, H.; Yu, B. Msmanet: A Multi-Scale Mesh Aggregation Network for Brain Tumor Segmentation. Appl. Soft Comput. 2021, 110, 107733. [Google Scholar] [CrossRef]
  29. Zhou, Z.; He, Z.; Shi, M.; Du, J.; Chen, D. 3d Dense Connectivity Network with Atrous Convolutional Feature Pyramid for Brain Tumor Segmentation in Magnetic Resonance Imaging of Human Heads. Comput Biol. Med. 2020, 121, 103766. [Google Scholar] [CrossRef]
  30. Chen, H.; Qin, Z.; Ding, Y.; Tian, L.; Qin, Z. Brain Tumor Segmentation with Deep Convolutional Symmetric Neural Network. Neurocomputing 2020, 392, 305–313. [Google Scholar] [CrossRef]
  31. Yang, Q.; Guo, X.; Chen, Z.; Woo, P.Y.; Yuan, Y. D2-Net: Dual Disentanglement Network for Brain Tumor Segmentation with Missing Modalities. IEEE Trans. Med. Imaging 2022, 41, 2953–2964. [Google Scholar] [CrossRef] [PubMed]
  32. Baid, U.; Ghodasara, S.; Mohan, S.; Bilello, M.; Calabrese, E.; Colak, E.; Farahani, K.; Kalpathy-Cramer, J.; Kitamura, F.C.; Pati, S.; et al. The Rsna-Asnr-Miccai Brats 2021 Benchmark on Brain Tumor Segmentation and Radiogenomic Classification. arXiv preprint 2021, arXiv:210702314. [Google Scholar]
  33. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The Multimodal Brain Tumor Image Segmentation Benchmark (Brats). IEEE Trans. Med Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef]
  34. Bakas, S.; Akbari, H.; Sotiras, A.; Bilello, M.; Rozycki, M.; Kirby, J.S.; Freymann, J.B.; Farahani, K.; Davatzikos, C. Advancing the Cancer Genome Atlas Glioma Mri Collections with Expert Segmentation Labels and Radiomic Features. Sci. Data 2017, 4, 1–13. [Google Scholar] [CrossRef] [Green Version]
  35. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXivpreprint 2014, arXiv:14126980. [Google Scholar]
  36. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. (Eds.) 3d U-Net: Learning Dense Volumetric Segmentation from Sparse Annotation. In International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Manhattan, NY, USA, 2016. [Google Scholar]
  37. Oktay, O.; Schlemper, J.; Folgoc, L.L.; Lee, M.; Heinrich, M.; Misawa, K.; Mori, K.; McDonagh, S.; Hammerla, N.Y.; Kainz, B.; et al. Attention U-Net: Learning Where to Look for the Pancreas. arXiv preprint 2018, arXiv:180403999. [Google Scholar]
  38. Shi, W.; Pang, E.; Wu, Q.; Lin, F. (Eds.) Brain Tumor Segmentation Using Dense Channels 2d U-Net and Multiple Feature Extraction Network. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2019. [Google Scholar]
  39. Zhao, Y.-X.; Zhang, Y.-M.; Liu, C.-L. (Eds.) Bag of Tricks for 3d Mri Brain Tumor Segmentation. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2019. [Google Scholar]
  40. Ribalta Lorenzo, P.; Marcinkiewicz, M.; Nalepa, J. (Eds.) Multi-Modal U-Nets with Boundary Loss and Pre-Training for Brain Tumor Segmentation. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2019. [Google Scholar]
  41. Ahmad, P.; Qamar, S.; Hashemi, S.R.; Shen, L. (Eds.) Hybrid Labels for Brain Tumor Segmentation. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2019. [Google Scholar]
  42. Choi, Y.; Al-masni, M.A.; Kim, D.-H. (Eds.) 3d Cmm-Net with Deeper Encoder for Semantic Segmentation of Brain Tumors in Brats 2021 Challenge; Springer International Publishing: Cham, Switzerland, 2022. [Google Scholar]
  43. Jiang, Y.; Zhang, Y.; Lin, X.; Dong, J.; Cheng, T.; Liang, J. Swinbts: A Method for 3d Multimodal Brain Tumor Segmentation Using Swin Transformer. Brain Sci. 2022, 12, 797. [Google Scholar] [CrossRef]
  44. Singh, G.; Phophalia, A. (Eds.) Multimodal Brain Tumor Segmentation Using Modified Unet Architecture; Springer International Publishing: Cham, Switzerland, 2022. [Google Scholar]
  45. Akbar, A.S.; Fatichah, C.; Suciati, N. (Eds.) Unet3d with multiple Atrous Convolutions Attention Block for Brain Tumor Segmentation; Springer International Publishing: Cham, Switzerland, 2022. [Google Scholar]
  46. Singh, H.S. (Ed.) Brain Tumor Segmentation Using Attention Activated U-Net with Positive Mining; Springer International Publishing: Cham, Switzerland, 2022. [Google Scholar]
  47. Luu, H.M.; Park, S.-H. (Eds.) Extending Nn-Unet for Brain Tumor Segmentation. In International MICCAI Brainlesion Workshop; Springer: Manhattan, NY, USA, 2022. [Google Scholar]
  48. Bukhari, S.T.; Mohy-Ud-Din, H. E1d3 U-Net for Brain Tumor Segmentation: Submission to the Rsna-Asnr-Miccai Brats 2021 Challenge. In International MICCAI Brainlesion Workshop 2022; Springer: Cham, Switzerland, 2021. [Google Scholar]
Figure 1. A visualized case is chosen from the BraTS2021 training dataset. (Different colors present different tumor types. Whole Tumors (WTs) are shown in blue, purple, and green, Tumor Cores (TCs) are shown in blue and purple and Enhanced Tumors (ETs) are shown in blue.)
Figure 1. A visualized case is chosen from the BraTS2021 training dataset. (Different colors present different tumor types. Whole Tumors (WTs) are shown in blue, purple, and green, Tumor Cores (TCs) are shown in blue and purple and Enhanced Tumors (ETs) are shown in blue.)
Brainsci 13 00012 g001
Figure 2. The structure of the proposed AABTS-Net.
Figure 2. The structure of the proposed AABTS-Net.
Brainsci 13 00012 g002
Figure 3. The structure of the proposed axial attention mechanism.
Figure 3. The structure of the proposed axial attention mechanism.
Brainsci 13 00012 g003
Figure 4. The visualized segmentation results of the proposed AABTS-Net and other methods. (ae) represents different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth, the result segmented by 3D-UNet, the result segmented by Attention-UNet, and the result segmented by our model.
Figure 4. The visualized segmentation results of the proposed AABTS-Net and other methods. (ae) represents different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth, the result segmented by 3D-UNet, the result segmented by Attention-UNet, and the result segmented by our model.
Brainsci 13 00012 g004
Figure 5. The visualized segmentation results of the proposed AABTS-Net. (ae) represents different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth and the result segmented by AABTS-Net.
Figure 5. The visualized segmentation results of the proposed AABTS-Net. (ae) represents different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth and the result segmented by AABTS-Net.
Brainsci 13 00012 g005
Figure 6. The performance indexes of the proposed AABTS-Net and the AABTS-Net without AAM. (a) represents the DSC value, (b) represents the Sensitivity value and (c) represents the HD value.
Figure 6. The performance indexes of the proposed AABTS-Net and the AABTS-Net without AAM. (a) represents the DSC value, (b) represents the Sensitivity value and (c) represents the HD value.
Brainsci 13 00012 g006
Figure 7. The visualized segmentation results of the proposed AABTS-Net and the AABTS-Net without AAM. (a,b) represent the different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth, the result segmented by AABTS-Net with the AAM and the result segmented by AABTS-Net without the AAM.
Figure 7. The visualized segmentation results of the proposed AABTS-Net and the AABTS-Net without AAM. (a,b) represent the different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth, the result segmented by AABTS-Net with the AAM and the result segmented by AABTS-Net without the AAM.
Brainsci 13 00012 g007
Figure 8. The DSC boxplot of the ablation study in three tumor regions.
Figure 8. The DSC boxplot of the ablation study in three tumor regions.
Brainsci 13 00012 g008
Figure 9. The performance indexes of the proposed AABTS-Net and the AABTS-Net without DS. (a) represents the DSC value, (b) represents the Sensitivity value and (c) represents the HD value.
Figure 9. The performance indexes of the proposed AABTS-Net and the AABTS-Net without DS. (a) represents the DSC value, (b) represents the Sensitivity value and (c) represents the HD value.
Brainsci 13 00012 g009
Figure 10. The visualized segmentation results of the proposed AABTS-Net and the AABTS-Net without DS. (a,b) represent the different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth, the result segmented by AABTS-Net with the DS and the result segmented by AABTS-Net without the DS.
Figure 10. The visualized segmentation results of the proposed AABTS-Net and the AABTS-Net without DS. (a,b) represent the different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth, the result segmented by AABTS-Net with the DS and the result segmented by AABTS-Net without the DS.
Brainsci 13 00012 g010
Figure 11. The DSC values for the different losses.
Figure 11. The DSC values for the different losses.
Brainsci 13 00012 g011
Figure 12. The visualized bad tumor segmentation cases of the proposed AABTS-Net. (ae) represents different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth, the result segmented by AABTS-Net.
Figure 12. The visualized bad tumor segmentation cases of the proposed AABTS-Net. (ae) represents different visualized cases. From left to right: T1, T2, T1c, Flair, Ground truth, the result segmented by AABTS-Net.
Brainsci 13 00012 g012
Table 1. A description of the datasets in this work.
Table 1. A description of the datasets in this work.
DatasetTraining DatasetValidation Dataset *
BraTS 2019335 cases125 cases
BraTS 20211251 cases219 cases
* Without ground truth.
Table 2. The training parameter of the proposed AABTS-Net.
Table 2. The training parameter of the proposed AABTS-Net.
NameParameter
Batch_size2
Epoch300
Learning rate0.0001
GPU memory16 GB
Patch size128 × 128 × 128
OptimizerAdam
FrameworkPytorch
Table 3. Objective performance comparison with the state-of-the-art methods on the BraTS2019 validation dataset, in terms of their DSC and HD values. (Bold font indicates best results.)
Table 3. Objective performance comparison with the state-of-the-art methods on the BraTS2019 validation dataset, in terms of their DSC and HD values. (Bold font indicates best results.)
MethodsDSCHD
ETTCWTAVGETTCWTAVG
3D UNet [36]0.7090.7250.8740.7695.0628.7199.4327.738
Attention UNet [37]0.7600.7720.8880.8075.2027.7568.2587.072
Shi et al. [38]0.6910.7700.8870.7835.88812.19221.19013.090
Zhao et al. [39]0.7020.8000.8930.7984.7666.4725.0785.439
Lorenzo et al. [40]0.6630.7510.8900.768////
Ahmad et al. [41]0.6230.7580.8520.7448.46810.6749.0089.383
Our Method0.7770.8380.9110.8423.2466.0283.9884.421
Table 4. Objective performance comparison with the state-of-the-art methods on the BraTS2021 validation dataset, in terms of their DSC and HD values. (Bold font indicates best results.)
Table 4. Objective performance comparison with the state-of-the-art methods on the BraTS2021 validation dataset, in terms of their DSC and HD values. (Bold font indicates best results.)
MethodsDSCHD
ETTCWTAVGETTCWTAVG
ESA-Net [20]0.8120.8520.9070.85726.47013.8305.94015.413
3D CMM-Net [42]0.7320.7510.8740.78635.00124.63810.16123.267
SwinBTS [43]0.8320.8480.9180.86616.03014.5103.65011.397
Singh et al. [44]0.7300.7600.8700.78730.50014.7006.29017.163
Akbar et al. [45]0.7800.8070.8910.82625.82021.17011.78019.590
Singh et al. [46]0.7530.8080.8990.82021.80012.5006.45013.583
Extending nn-UNet [47]0.8450.8780.9280.88420.7307.6233.47010.608
E1D3-UNet [48]0.8180.8630.9230.86818.2409.6204.34010.733
Bitr-UNet [24]0.8190.8430.9100.85717.84716.6894.50813.015
Our Method0.8300.8610.9220.87117.72811.1783.99610.967
Table 5. The ablation studies of the different methods on the BraTS 2021 training dataset (1251 cases). All experiments were performed via 5-fold cross-validation on the training cases (no external data were used). (Bold font indicates best results.)
Table 5. The ablation studies of the different methods on the BraTS 2021 training dataset (1251 cases). All experiments were performed via 5-fold cross-validation on the training cases (no external data were used). (Bold font indicates best results.)
MethodsDSCHDSensitivity
ETTCWTAVGETTCWTAVGETTCWTAVG
Baseline0.8280.8590.9010.86319.70311.3105.74312.2520.8400.8650.9120.872
Baseline + AAM0.8600.8810.9420.89413.7287.1783.9968.3010.8620.8900.9220.891
Baseline + DS0.8450.8670.9190.87714.5838.0423.7238.7830.8540.8820.9270.887
AABTS-Net0.8900.9090.9400.91310.1327.1323.3436.8690.9050.9150.9520.924
Table 6. The DSCs of the bad segmentation cases.
Table 6. The DSCs of the bad segmentation cases.
CasesDSC
ETTCWTAVG
(a) BraTS2021_004930.8850.9240.2700.693
(b) BraTS2021_004940.9640.9900.7300.895
(c) BraTS2021_0166600.7380.9160.551
(d) BraTS2021_011791.00000.7690.590
(e) BraTS2021_012930.5410.9290.8740.781
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tian, W.; Li, D.; Lv, M.; Huang, P. Axial Attention Convolutional Neural Network for Brain Tumor Segmentation with Multi-Modality MRI Scans. Brain Sci. 2023, 13, 12. https://doi.org/10.3390/brainsci13010012

AMA Style

Tian W, Li D, Lv M, Huang P. Axial Attention Convolutional Neural Network for Brain Tumor Segmentation with Multi-Modality MRI Scans. Brain Sciences. 2023; 13(1):12. https://doi.org/10.3390/brainsci13010012

Chicago/Turabian Style

Tian, Weiwei, Dengwang Li, Mengyu Lv, and Pu Huang. 2023. "Axial Attention Convolutional Neural Network for Brain Tumor Segmentation with Multi-Modality MRI Scans" Brain Sciences 13, no. 1: 12. https://doi.org/10.3390/brainsci13010012

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop