A Novel Data Augmentation-Based Brain Tumor Detection Using Convolutional Neural Network

: Brain tumor is a severe cancer and a life-threatening disease. Thus, early detection is crucial in the process of treatment. Recent progress in the ﬁeld of deep learning has contributed enormously to the health industry medical diagnosis. Convolutional neural networks (CNNs) have been intensively used as a deep learning approach to detect brain tumors using MRI images. Due to the limited dataset, deep learning algorithms and CNNs should be improved to be more efﬁcient. Thus, one of the most known techniques used to improve model performance is Data Augmentation. This paper presents a detailed review of various CNN architectures and highlights the characteristics of particular models such as ResNet, AlexNet, and VGG. After that, we provide an efﬁcient method for detecting brain tumors using magnetic resonance imaging (MRI) datasets based on CNN and data augmentation. Evaluation metrics values of the proposed solution prove that it succeeded in being a contribution to previous studies in terms of both deep architectural design and high detection success.


Introduction
In 2020, it was estimated that 308,102 people were diagnosed with a primary brain or spinal cord tumor in the world [1]. Brain tumors are the 10th leading cause of death worldwide [2]. It is caused by tissue abnormality that develops within the brain or the central spine. As a result, it disrupts the proper operation of the brain. The causes of brain tumors are unknown; nevertheless, the risk can be enhanced by exposure to radiation and family history [3]. Consequently, detection and identification of brain tumors at an early phase is key to successful treatment. Indeed, it plays an indispensable role in improving treatment and ensuring a higher gain of survival possibility. There are several medical imaging techniques and diagnostic methods used to acquire information about tumors, such as Computed Tomography (CT) scans and Magnetic Resonance Imaging (MRI) scans that can distinguish between normal and abnormal cells that grow in the brain [4]. The medical science field has, in the past few years, seen striking progress leading to accurate classification of brain tumors thanks to AI and deep learning. CNN is used in image processing techniques to segment, identify, and classify MRI images as well as to classify and detect brain tumors. These image processing techniques can be based on the image content analysis described in [5][6][7], which plays a dynamic role in various computer vision applications. Recent advances in AI, and in particular in machine learning and deep learning, have contributed to the development of autonomous objects, such as robots, drones, and cars. This has allowed it to become the most important innovation driving force in the fields of technology and industry. The last few years have been marked by the growing interest in the healthcare sector and diseases detection to enhance the implementations of E-Health services. Deep Learning has recently become an active field of interest that attracts researchers, mostly in the field of medical sciences. It has significantly impacted the study of diseases in numerous ways: in the detection, prediction, and diagnosis of diseases. In [8,9], the author's proposed solutions and new techniques to impact image reconstruction and recognition performance. Computer science scientists have developed many deep learning algorithms to detect and diagnose diseases such as cancer, lung diseases, diabetes, heart diseases, Alzheimer's disease, hepatitis, liver disease, among others. The attentiveness to deep learning is raised to convolutional neural networks (CNN), a powerful way to learn useful representations mainly of images and other structured data. Convolutional neural networks (CNN) are deep artificial neural networks majorly used in image classification, image segmentation, and objection detection. CNN has shown significant advantages in image recognition [10,11]. Currently, it is attracting interest in a variety of domains and has achieved a huge advancement in various fields. Recently, new technologies have also taken an interest in other medical fields, such as neurosurgery. In [12,13], authors showed that Augmented Reality (AR) and mobile devices could help in the operating room. In [14], authors developed a new approach based on deep learning techniques to classify White Blood Cells for disease diagnosing. Experimental results showed that the classification of the modified images is more significant than the classification of the original ones. Authors, in [15,16], proposed to identify and classify liver diseases by using a deep supervised learning method based on CNN architecture. A classification framework was proposed in [15] and consists of improving the processing images and a segmentation of the liver lesions. In [16], the authors developed a two-step classification approach. The first step is the collection of a sufficient number of isolated training samples. The second step is to train two CNN with the same architecture but employing different optimization algorithms. The architectures described in [15,16] have reached a classification accuracy of 95%. Recently, with the COVID-19 pandemic, the world is facing a virus with unknown behavior. Therefore, several studies have been initiated to detect people attacked by this virus [17]. In [18], the author introduced a study to identify the presence or absence of malaria parasites in the blood smears of people by using a deep learning algorithm. The Convolutional Neural Network algorithm has successfully achieved an accuracy rate of 96%. As for Ghulam [19], he suggested a study based on deep learning to develop an accurate classification model to classify Breast Cancer into eight subtypes. In [20], authors stated a deep learning survey for detecting lung disease.

Contribution
Convolutional Neural Networks (CNNs) have demonstrated indisputable effectiveness in detecting many diseases and are widely used in medical image analysis. These networks are especially being used for the detection, classification, and segmentation of brain tumors in MRI datasets. The main objectives of our approach are • Detecting brain tumors from MRI datasets using deep learning and convolutional neural networks. • Sometimes, we face issues like limited data, so we are extremely interested in the data augmentation technique. This technique allows us to implement the detection algorithms we plan to develop. • In our paper, we used data augmentation techniques to improve the detection of brain tumors by using the VGG-16 model.

•
Experimental results showed that expanding a dataset by using flipping, rotation, and translation techniques is very useful to train the VGG model.
In fact, when we have limited data, deep learning algorithms and CNNs should be improved to be more efficient. The data augmentation technique exploits various transfor-mations of the original data, such as affine image transformations, elastic transformations, and pixel-level transformations. In the literature, data augmentation approaches have been applied to enrich the size of training sets to allow developers to benefit from more representative training data.

Related Works
In [21,22], the authors provided an overview of some potential clinical use cases using deep learning techniques by defining the steps to undertake a deep learning project in radiology. The main idea of these two papers is to discuss opportunities and challenges for incorporating deep learning in the radiology practice of the future. The effectiveness of existing applications in radiology are not yet encouraging to say that the DL can replace a radiologist in all of his diagnostic work. However, radiologists and DL can help each other to give better results. Hence, several works have been done on the classification and segmentation of the brain using MRI images. El Abbadi et al. proposed a new method using SVD as a classifier to classify brain tumors. At the first level, the algorithm had been trained with normal brain MR images. Then, at the second level, it became capable of classifying the brain images into healthy and non-healthy images. The accuracy of this method reached up to 97%. In [23], Sheikh Basheera et al. focused on brain tumor classification in MRI images using a classifier based on Convolutional Neural Networks (CNN). The main idea of the proposed approach is based on two steps. The first one is the tumor region segmentation using an ICA mixture mode model (Independent Component Analysis). The second step is the extraction of deep features. In [24], Muhammad Sajjad et al. proposed a novel convolutional neural network (CNN) based multi-grade brain tumor classification system. The first step consists of segmenting the tumor regions from an MR image using a deep learning technique. After that, they employed extensive data augmentation to train the system effectively. Finally, a pre-trained VGG-19 CNN model is fine-tuned using augmented data for brain tumor grade classification. Sunanda Das et al. [25] trained a CNN model with an image processing technique to identify various brain tumor types and achieved 94.39% accuracy with an average precision of 93.33%. In [26], Muhammed Talo et al. used deep transfer learning to classify normal and abnormal brain MR images automatically. The proposed model that used ResNet34 has achieved a 5-fold classification accuracy of 100% on 613 MR images. Ahmet Inner et al., in [27], used the ResNet50 pre-trained model, and they removed the last 5 layers of the model, then they added 8 new layers. Then, comparing its accuracy with other pre-trained models such as GoogleNet, AlexNet, and ResNet50. The modified ResNet50 model showed effective results by achieving 97.2% accuracy. He obtained a 90% accuracy in the classified images as normal and abnormal in his proposed machine learning method. The authors in [28], proposed a modified AlexNet for the detection and classification of brain tumor images and obtained 91.6% of average classification accuracy. Another approach based on a modified ResNet50 model for brain tumor detection was developed in [29]. The proposed architecture is based on the ResNet50 model with a modified layer model including five convolutional layers and three fully connected layers. In [30], researchers proposed a brain tumor detection and classification. The main idea of their approach is to use a biologically inspired orthogonal wavelet transform and deep learning techniques. Techniques of graph theory were used [31] to detect abnormalities in brains. A VGG16 architecture was the main model to classify brain images in [32]. In this research paper, the authors described their approach based on the Mask R-CNN model to detect and identify brain tumors with improved precision.
Limited datasets are a particularly common challenge in medical image analysis. Most computer vision tasks could use more data and data augmentation is one of the techniques often used to enhance the performance of computer vision systems. To overcome this limitation, many approaches based on deep learning have been proposed and detailed in the literature. One of the first applications of data augmentation was proposed in LeNet-5 [33] to classify the handwritten digit. In 2012, Krizhevsky et al. [34] boosted image classification by the data augmentation techniques on the ImageNet dataset. The goal of the proposed approach is to increase the dataset size. The authors used in their experiments random cropping patches from the original images, flipping them horizontally, and changing the pixel intensity. Experimental results showed that the data augmentation reduced the error rate by over 1%. After the appearance of several research works using different data augmentation techniques, we can categorize them into two main categories [35]. (1): Traditional transformations, which are based on the combination of the affine image transformation and color modification. (2): Generative Adversarial Networks (GANs), a tool based on an unsupervised generation of new images using min-max strategy [36]. GANs were introduced in 2014 in [37] and it consists of generating a new dataset. The new dataset is indistinguishable from the original one. In [38], authors combined data augmentation with min-max normalization to increase the contrast of tumor cells. In the experimental results part, the proposed model was 99.97% accurate during training and 98.78% accurate during testing. A novel generative adversarial model based on cancer genes was developed in [39]. A deep multi-scale 3D Convolutional Neural Network for MRI Gliomas brain tumor classification was developed in [40]. Researchers showed that using data augmentation techniques enhanced the proposed approach, which achieved 96.49% of accuracy.

LeNet
The LeNet model is a classic CNN model proposed by Yann LeCun et al. [41]. It has a wide range of applications in image classification [42][43][44]. The LeNet-5 usually uses the ReLU function or the Sigmoid function as an activation function. It consists of an input layer, two convolutional layers, two pooling layers, two fully connected layers, and an output layer.

AlexNet
This architecture was developed by Alex Krizhevsky, Ilya Sutskever, and Geoff Hinton, and it is considered the first convolutional network to popularize it in the field of computer vision [34]. The AlexNet architecture consists of five convolutional layers (conv), three pooling layers (Pool) which are followed by three fully connected layers (FC). Compared to LeNet, this network is much bigger and deeper.

GoogleNet
In 2015, Google released GoogleNet, a deep neural network, which is a convolutional neural network that is 22 layers deep. Parallelization was introduced in this architecture. Indeed, it is characterized by an inception block that comprises a 1 × 1, 3 × 3, and a 5 × 5 convolution filter in addition to a 3 × 3 max-pooling layer [45].

ResNet
He et al. initialized ResNet models that rely on deep architectures that have demonstrated convincing precision and convergence behaviors of high quality. ResNet was conceived through numerous stacked residual units and evolved using different numbers of layers: 18, 34, 50, 101, 152, and 1202. The main disadvantage of this network is that it is very expensive to evaluate due to a large number of parameters [46].

VGGNet
VGGNet is an abbreviation of Visual Geometry Group; it is a convolutional neural network architecture proposed by Karen Simonyan and Andrew Zisserman of the University of Oxford in 2014 [47]. Its main contribution was to show that the depth of the network is a critical component to achieve better recognition or classification accuracy in CNNs.

DenseNet
In 2017, Huang et al. developed DenseNet [48]. DenseNet uses dense connections between layers via dense blocks [49][50][51][52]. DenseNet basically connects every layer to every other layer. This is extremely powerful. The entry of a layer in DenseNet is the concatenation of feature maps from previous layers. By connecting in this way, DenseNet requires fewer parameters than an equivalent traditional CNN, as there is no need to learn redundant feature maps.

SqueezeNet
SqueezeNet was designed as a more compact replacement for AlexNet. It is a smaller network that has almost 50 times fewer parameters than AlexNet, but it runs 3 times faster [53]. To reduce the size of the model, SqueezeNet was designed with three strategies: • Reduction of the filter size with the use of 1 × 1 filter instead of 3 × 3. • Reduction of the input channels to 3 × 3 filters. • Downsampling at the end of the array so that the convolutional layers have large activation maps.

MobileNet
MobileNet is an architecture of CNN. It is efficient for mobile and embedded vision systems [54]. Its model is designed to be used in mobile applications and it is the first mobile computer vision model based on TensorFlow. In MobileNet, the convolution is replaced by a "Depthwise Separable Convolution" which is carried out in two stages: • Depthwise Convolution or Convolution in depth. • Pointwise Convolution or Point Convolution.
The Depthwise Convolution applies a filter to each channel, unlike conventional convolution, which applies a filter to all channels. The Pointwise Convolution consists of combining the outputs of the Depthwise Convolution. It is also called 1 × 1 convolution.

Deep Convolutional Neural Network
A Convolution Neural Network contains neurons with some weights and biases. These neurons capture inputs from the anterior layers ( Figure 1). CNN gives a high-speed and accurate algorithm that displays excellent performance in detection and classification compared to classical neural networks [55,56]. The classification of the most well-known and used image databases, such as MNIST [57,58] and CIFAR 10 [59,60] has been improved by the use of CNNs.

Convolution Layer
The basic architecture of a CNN consists of different convolutional layers, in addition to pooling layers and fully connected layers. The convolutional layer aims at taking or extracting features from the input data. To achieve the featured maps, we reiterate the process starting from the input image and then calculating the dot product considering the weights and biases. The formula for computing a single output matrix is described as follows in Equation (1): where I is an input vector, and K is the corresponding convolution kernel with the size of B j × n. N is the input size and B j is the bias value. f is a non-linear activation function such as Sigmoid, Tanh, ReLu and Leaky ReLu (Figure 2). The activation function used in our work is the one used in [61]. Its formula is as follows (Equation (2)).

Back Propagation
The main objective of our experiment is to study the efficiency of the chosen model in the classification of brain tumors. To minimize the loss function, we need to calculate optimal parameter values in the backpropagation phase. Kernels and biases are the main parameters in a Convolutional Neural Network. To find optimal values of parameters, we will apply the Stochastic Gradient Descent algorithm. The model is based on a very small convolutional filter the size of (3 × 3) to deal with large-scale images. Each block of the model is a sequence of convolutional layers. These layers are followed by a max-pooling layer. We applied a kernel of size (3 × 3) overall to the model. Then, a max-pooling of size 2 × 2 with strides of 2 is also applied to divide equally the resolution after each block. A VGG model has two fully connected hidden layers and one fully connected output layer.
The structure of the selected model is described in Table 1.

DataBase Collection
In this work, we suggest a classification model that would allow us to consider MRI images of the patient as input and compute to detect whether there is a tumor in the brain or not, as an output. We relied on Kaggle, which publicly provides brain MRI images. The dataset we selected contains 253 brain MRI images. The yes folder contains 155 tumor brain MRI images and the no folder contains 98 non-tumor MRI images. This shows that 61% (

Database Augmentation
Data augmentation is a solution to increase the quantity and complexity of existing data artificially [62]. Data augmentation approaches have been applied to enrich the size of training sets, to allow developers to benefit from more representative training data [63,64]. The main principle is to increase, artificially, the number of training examples. It can act as a regularizer in preventing overfitting in neural networks. In literature, we can classify data augmentation techniques into three types: • Dataset generation and expanding an existing dataset ( Figure 5) • In-place/on-the-fly data augmentation • Combining dataset generation and in-place augmentation. Translation: involves moving the image along the X or Y direction or both.
In our case, we applied changes using flipping, rotation, and translation techniques ( Figure 6 and Algorithm 1).

Algorithm 1: Data Augmentation
Input : DataSet (DS) Output : Augmented images 1 DataAugmentation 2 n ← number of images in DS In Table 2, we detailed the performances of three models that were initially trained on a dataset without augmentation and then on a dataset with augmentation. This study was proposed in [65] and the main idea was that Cascaded Net segments the tumor following three stages: (1) locate and pick out the whole brain tumor area. (2) Remove the useless surrounding tissue area and crop a square tumor region as the input to the next network to segment the tumor core. (3) The third network divides the tumor core into an enhanced region and a non-enhanced region. The mean value is the Dice score that is mainly used to quantify the performance of image segmentation methods [66]. As shown in Table 2, the mean values were enhanced by 2% or 3% for each of the three models. Thus, we can conclude that models with data augmentation outperformed those without augmentation.

Results and Discussions
VGG-16 is a very good architecture for benchmarking a particular task. Also, pretrained networks for VGG are available freely on the internet, so it is commonly used out of the box for various applications. Although VGG-16 represented the acquired features effectively, deep structure and supervised learning may cause overproduction when the quantity of training data is restricted. It is the case with many medical situations where there is limited data, and the VGG-16 model, which is characterized by many parameters, may lead to over-fitting. In the VGG-16 model, there are thirteen convolution layers, five pooling layers, and always uses a 3 × 3 Kernel for convolution and 2 × 2 size of the maxpooling layer. In our work, we create a CNN architecture using Tensorflow, Keras, and jupyter. We train the model on an MRI brain image database. A block diagram of our solution is described in Figure 7. The output of the model is "YES" for the abnormal brain ( Figure 8) or "NO" for the normal brain ( Figure 9).   The running time is about 300 (s). The results demonstrate the model's ability and accuracy in classifying images ( Figure 10). To confirm the obtained results, we evaluated our model on a second Kaggle dataset. Its accuracy is shown in Figure 11.  As shown in Table 3, the used dataset of brain tumors consists of 253 real brain images (Kaggle platform). Data are divided into training (185 images), validation (48 images), and testing (20 images). Then, this dataset of 253 images was augmented to produce 3700 new images. The proposed model was trained using the augmented dataset and the results are described in Table 4. Table 3. Dataset before and after augmentation.

Number of Images
Original dataset 253 Tumor brain MRI images 155 Non-tumor MRI images  98  After augmentation  3700  Training  185  Validation  48  Test  20 The number of epochs = 15 and the Batch size = 32. To evaluate the proposed model's efficiency, we relied on ten models of machine learning methods. The comparison between the proposed VGG-16 model and the other models was based on the following values [67,68]: For comparison purposes, true positive (TP), true negative (TN), false positive (FP), and false negative (FN) are used to assess the performance of the proposed model, on the one hand, and ten machine learning models, on the other. Table 4 illustrates the comparison between the different models.

Conclusions
In this paper, a brain tumor classification was implemented based on the Convolutional Neural Network and data augmentation technique. We presented a detailed review of various CNN architectures and their limitation if we have a limited dataset. The goal is to overcome this problem. Then, we presented that we can improve performances on limited brain tumor datasets by enriching them using data augmentation. Experimental results showed that the model's ability and accuracy in classifying images are very motivating. Our data augmentation-based solution has shown high detection efficiency and good evaluation metrics value even in a limited MRI dataset. In future work, we plan to explore more complex architecture, more varied datasets, and more data augmentation techniques.  Data Availability Statement: Dataset of brain tumors: https://www.kaggle.com/datasets/navoneel/ brain-mri-images-for-brain-tumor-detection, accessed on 20 February 2022.

Conflicts of Interest:
The authors declare no conflict of interest.