Next Article in Journal
A Novel Method for Lung Image Processing Using Complex Networks
Previous Article in Journal
The Diagnostic Performance of Multi-Detector Computed Tomography (MDCT) in Depiction of Acute Spondylodiscitis in an Emergency Department
Previous Article in Special Issue
AI Denoising Improves Image Quality and Radiological Workflows in Pediatric Ultra-Low-Dose Thorax Computed Tomography Scans
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Multi-Scale Convolutional Neural Network Based Multi-Class Brain MRI Classification for SaMD

1
Department of Computer Science, Attock Campus, COMSATS University Islamabad, Attock 43600, Pakistan
2
Big Data Research Center, Jeju National University, Jeju 63243, Korea
3
Department of Computer Engineering, Jeju National University, Jeju 63243, Korea
4
Advanced Technology Research Institute, Jeju National University, Jeju 63243, Korea
*
Authors to whom correspondence should be addressed.
Tomography 2022, 8(4), 1905-1927; https://doi.org/10.3390/tomography8040161
Submission received: 10 April 2022 / Revised: 28 June 2022 / Accepted: 13 July 2022 / Published: 26 July 2022

Abstract

:
A brain tumor is the growth of abnormal cells in certain brain tissues with a high mortality rate; therefore, it requires high precision in diagnosis, as a minor human judgment can eventually cause severe consequences. Magnetic Resonance Image (MRI) serves as a non-invasive tool to detect the presence of a tumor. However, Rician noise is inevitably instilled during the image acquisition process, which leads to poor observation and interferes with the treatment. Computer-Aided Diagnosis (CAD) systems can perform early diagnosis of the disease, potentially increasing the chances of survival, and lessening the need for an expert to analyze the MRIs. Convolutional Neural Networks (CNN) have proven to be very effective in tumor detection in brain MRIs. There have been multiple studies dedicated to brain tumor classification; however, these techniques lack the evaluation of the impact of the Rician noise on state-of-the-art deep learning techniques and the consideration of the scaling impact on the performance of the deep learning as the size and location of tumors vary from image to image with irregular shape and boundaries. Moreover, transfer learning-based pre-trained models such as AlexNet and ResNet have been used for brain tumor detection. However, these architectures have many trainable parameters and hence have a high computational cost. This study proposes a two-fold solution: (a) Multi-Scale CNN (MSCNN) architecture to develop a robust classification model for brain tumor diagnosis, and (b) minimizing the impact of Rician noise on the performance of the MSCNN. The proposed model is a multi-class classification solution that classifies MRIs into glioma, meningioma, pituitary, and non-tumor. The core objective is to develop a robust model for enhancing the performance of the existing tumor detection systems in terms of accuracy and efficiency. Furthermore, MRIs are denoised using a Fuzzy Similarity-based Non-Local Means (FSNLM) filter to improve the classification results. Different evaluation metrics are employed, such as accuracy, precision, recall, specificity, and F1-score, to evaluate and compare the performance of the proposed multi-scale CNN and other state-of-the-art techniques, such as AlexNet and ResNet. In addition, trainable and non-trainable parameters of the proposed model and the existing techniques are also compared to evaluate the computational efficiency. The experimental results show that the proposed multi-scale CNN model outperforms AlexNet and ResNet in terms of accuracy and efficiency at a lower computational cost. Based on experimental results, it is found that our proposed MCNN2 achieved accuracy and F1-score of 91.2% and 91%, respectively, which is significantly higher than the existing AlexNet and ResNet techniques. Moreover, our findings suggest that the proposed model is more effective and efficient in facilitating clinical research and practice for MRI classification.

1. Introduction

Brain tumors are masses formed due to the abnormal growth of lesions inside human brains that directly affect the functionality of brains for controlling voluntary and involuntary processes [1]. It is a life-threatening and leading disease toward cancer mortality worldwide [2]. According to data from the International Agency for Research on Cancer 2012, brain tumors are the 22nd most common form of tumor. However, it is 12th in terms of the mortality rate [3]. According to Cancer Statistics 2019, there has been an increase in deaths caused by different types of cancer, including brain cancer [4]. A report published by World Health Organization (WHO) estimated that 9.6 million fatalities were caused by diagnosed cancer around the globe in 2019 [5]. In [6], 296,851 new cases of brain tumors and other spinal tumors were reported worldwide in 2018. In 2017, among all types of tumors, brain and other spinal cord tumors were the leading cause of death in men below 40 and women below 20 years of age [7]. Furthermore, the survival rate of the patient decreases with age. The survival rate for adults above 40 is about 21 percent [8].
Neuroimaging is a powerful non-invasive tool for finding abnormalities in the brain. Computerized Tomography (CT) scans and MRIs are the two most commonly used neuroimaging techniques for brain tumor diagnosis [9,10]. The advantage of using MRI over a CT scan in brain imaging is that not only does it provide a better tissue contrast, but it also does so without the use of radiation [11]. MRI uses strong magnetic fields and radio waves to analyze the anatomy and physiological processes of a body. Therefore, MRI is a strong tool for detecting diseases and anatomical anomalies in a body. It has a wide range of applications in medical imaging. For example, cardiovascular MRI is used for structural analysis of the heart. Musculoskeletal MRI is used to assess the spine, joint diseases, and soft tissue tumors. Similarly, brain MRI is used to detect neurological diseases such as tumors, clots, etc. [12]. However, detecting a tumor in a brain MRI is difficult and requires an expert to examine the images to determine the presence of a tumor. It is also imperative to detect and identify the type of tumor. Therefore, Machine Learning (ML) has become a widely used paradigm for detecting tumors in several body parts, including brain tumor detection in MRI [13], liver tumor classification [14], breast tumor detection [15], etc. Hence, a robust solution is needed to utilize a Computer Vision (CV) technology that can accurately detect and identify the presence of a tumor in an MRI.
MRIs suffer from the inherited problem of being vulnerable to noise [16,17]. Getting the desired image resolution decreases the Signal to Noise Ratio (SNR) [18]. Noise can occur in any listed processing phase: acquisition, compression, pre-processing, storage, transmission, and/or reproduction. The transformation of MRI images into magnitude images changes the Gaussian distribution into a Rician distribution [19]. The presence of noise in the MRIs makes it difficult to perform any further image processing techniques on these images [20]. Therefore, there is a need for noise removal as pre-processing to pass the pre-processed images to the ML classifier for accurate tumor detection. Several techniques have been used to denoise the MRIs [21]. However, denoising filters also affect the level of detail in the images, which affects the classification process. Hence, FSNLM [22] is employed to perform noise removal, but it has a comparatively low processing cost. This technique uses fuzzy logic to find segments in the image that are similar to the noisy pixel and use these segments to find the noise-less pixel.
ML techniques have been proven promising for identifying and classifying a brain tumor in MRI images [23]. Therefore, developing an accurate classification model to detect a brain tumor in MRI images is quite possible. However, conventional ML techniques: Support Vector Machine (SVM) [24], k-Nearest Neighbor (KNN) [25], Decision Tree (DT) [26], Random Forest (RF) [27], etc., require a lot of knowledge of the input images so that useful features can be extracted from them. Deep Learning (DL) techniques have recently revolutionized ML by automatically extracting features from MRI images [28]. Convolutional Neural Networks (CNNs) are widely used DL techniques that perform automatic feature extraction, making them suitable for image classification and object detection problems [29,30]. A CNN performs a series of convolutional, non-linearity, and pooling operations on the input images to extract the useful features from given images. These features are then used for object detection, segmentation, or image classification. Moreover, the traditional CNN model requires a huge amount of training samples to produce effective classification results. Therefore, our research study focuses on transfer learning that allows pre-training a CNN model on a vast image dataset and then reusing the resulting parameters of a model for a similar task with specific samples.
The transfer learning approach is very effective for training a DL architecture from scratch when training samples are small, causing a biased and over-fitted DL model. Several pre-trained models, such as GoogleNet, AlexNet, ResNet, and VGG, were originally designed for different brain tumor classification tasks [31]. However, these architectures have a large number of trainable parameters and hence require a lot of computational resources. Furthermore, the size of the convolution kernel governs the features extracted by the CNN. Using a single kernel restricts the extractions of a wider range of features from the input images. Therefore, utilizing more than one kernel for a single convolution is imperative. Hence, our study proposes a multi-scale CNN model to enhance the performance and efficiency of traditional CNN architectures.
The notable contributions of the proposed research study are as follows:
  • The impact of convolutional scaling is evaluated for the varying tumor size, shape, and location on the classification performance of CNN.
  • An empirical scale-based multi-scale CNN architecture is proposed to outperform the state-of-the-art CNN architectures in terms of effectiveness and efficiency.
  • The proposed multi-scale CNN model is tested on noisy and denoised images, concluding that noise affects the CNN classifier accuracy.
  • Furthermore, a detailed comparative analysis is presented to evaluate the accuracy and efficiency of the proposed research study.
The remaining structure of the proposed study is ordered as follows: Section 2 describes the theoretical background of the computer-aided diagnosis of brain tumor MRIs. Section 3 discusses the proposed model for brain tumor detection. Section 4 consists of the experimental environment, experimental results, and performance analysis. Section 5 presents a comparative analysis to evaluate the accuracy and efficiency of proposed and existing approaches. Section 6 concludes the proposed research study, and possible future directions are also suggested for research in this field.

2. Related Work

This section presents existing studies related to brain tumors and the general methods used to diagnose them. It also discusses the various ML techniques that have been used to develop Computer-Aided Diagnosis (CAD) systems for brain tumors. Furthermore, different CNN architectures are discussed for brain tumor detection, classification, and segmentation.

2.1. Brain Tumor

The brain is undoubtedly an essential organ of the human body. It consists of a complex mechanism in which a network of billions of neurons controls all the body’s functions. The human body creates billions of cells per day. A cell will eventually die after completing its life cycle and be replaced by a new one in a normal situation. However, sometimes this cycle is disrupted. Some cells do not die; instead, they continue to grow in an uncontrolled manner [32]. A group of such cells forms a tumor. A tumor can form in any part of the body. A tumor in the brain is called a brain tumor [33]. The brain is enclosed inside a rigid skull. Tumor growth in such a restricted area can press against the brain. This can not only cause brain damage but possibly be fatal.
Brain tumors can be classified into different types based on their origin, aggressiveness, and location. A brain tumor may be primary or secondary. Primary brain tumors originate in the brain, while secondary brain tumors originate in another part of the body, such as the lungs, and then metastasize to the brain. Based on the nature and aggressiveness of their growth, tumors have three categories: benign, pre-malignant, and malignant [34]. A benign tumor is not cancerous, meaning it will not metastasize to other body parts. A pre-malignant tumor is not yet cancerous but risks becoming cancerous. However, a malignant tumor is cancerous. It grows over time in size and can even metastasize to other organs. Malignant tumors can eventually become fatal. Therefore, they must be surgically removed and possibly treated with chemotherapy, radiation, or immunotherapy. In terms of location [35], brain tumors have different types: glioma, pituitary, and meningioma. Glioma is cancer formed in the central or peripheral nervous system due to glial cells. Two-thirds of all primary cancerous tumors are categorized as glioma [36]. Three layers of membrane called meninges cover the brain and the spinal cord. A benign tumor formed in the interior of the pituitary gland is called a pituitary tumor. The pituitary gland is a small gland responsible for secreting hormones vital for regulating body functions. Pituitary tumors account for 8% of all primary brain tumors [37]. Meningioma is one of the most common forms of brain tumor that form in the meninges. Out of all the cases of brain tumors, nearly 20 percent are meningioma. Meningioma can be malignant but are mostly (90%) benign. They either grow at a very slow rate or do not grow at all. In most cases, they do not require any immediate treatment [38]. However, it is important to detect the presence of a tumor in neuroimaging and identify the type of tumor detected.
Furthermore, CT scan and MRI are the two main neuroimaging techniques for brain tumor diagnosis without any invasive method. MRI is preferable to a CT scan for two main reasons: it does not use radiation and provides better contrast than the CT scan [39]. A brain tumor diagnosis is initially conducted using MRI. Once the presence of a tumor has been confirmed, other tests such as tissue analysis can be conducted to get more details about the tumor [40].

2.2. Noise Removal from MRIs

Noise in MRIs makes it challenging to apply image processing techniques to detect brain tumors. It also negatively influences the classification accuracy of the brain tumor detection systems [41]. Therefore, different noise removal filters have been developed to enhance the quality of MRIs for diagnosis. In [42], the authors proposed a denoising technique for MRIs to perform classification for brain tumor detection. The technique was developed by combining Low-Rank Matrix Decomposition (LRMD) and SVM. An adaptive hexagonal fuzzy hybrid filter proposed by Kala et al. [43] removes noise from MRI images using local and non-local filters to improve SNR. Zeng et al. [44] proposed a denoising technique for MRIs. Their method consisted of three stages. In the first stage, an MRI image is segmented into cartoon, texture, and residual parts using Morphological Component Analysis (MCA). In the second stage, the three segments are denoised using the wiener filter, hard wavelet threshold, and soft wavelet threshold. In the third stage, all the denoised segments are combined for a denoised MRI. Dictionary learning-based algorithms have proven to be quite effective for noise removal while preserving necessary details, but at a very high computational cost [45]. However, noise removal filters affect MRI information (pixels) that negatively affect the brain tumor classification processes. Therefore, it is required to preserve MRI information while performing the denoising process. Hence, our research study utilizes the FSNLM filter [22] to remove noise from MRIs at a low computational cost compared to other noise removal filters.

2.3. Traditional CADs for Brain Tumor

There is a need for a CAD system to classify MR images into tumor and non-tumor. A lot of work has been performed in brain tumor detection and classification. In [46], the authors designed a model for brain tumor MRI classification based on offline and online phases. In the offline phase, MRIs are processed in a sequence of steps: tumor segmentation, feature extraction, and distance metric learning. The online phase processes MRIs to extract features and compare them with the learned distanced metrics stored in the online database. In [47], the authors proposed a hybrid technique based on Fuzzy C-mean clustering and multi-object optimization for brain tumor tissue segmentation. A study presented in [48] performed segmentation on MRI images using the Skippy greedy snake algorithm. In [49], the authors used watershed and thresholding-based techniques for tumor segmentation. Similarly, a threshold-based tumor detection technique was used in [50]. Another detailed review study was presented in [51] for different brain tumor MRI segmentation techniques. The four techniques studied were K-Mean, Fuzzy C-mean, region growing, and Otsu segmentation. Based on the results, it was concluded that K-mean outperformed other segmentation techniques. In [52], the authors used Fuzzy C-mean clustering for tumor detection in MRIs. Furthermore, a study presented in [53] used SVM-based techniques for tumor classification. The presented study consisted of four steps. First, the region of interest was manually extracted from the images. Second, the tumor was segmented from the images using thresholding. Third, a Genetic Algorithm (GA) extracted features from the images. Finally, an SVM with a linear kernel function was used to classify the images.

2.4. Traditional CNN Architecture

In ML, DL is an emerging and prominent research topic used in different domains. CNNs are a type of Deep Neural Network (DNN) mostly used in computer vision problems. Traditional supervised learning techniques required hand-crafted feature extractors. The quality of the model depended on how accurately these feature extractors suit the input data. Designing these extractors is time-consuming and requires a lot of prior knowledge about the input data. Another disadvantage of these designed feature extractors is that they seldomly reflect real-world data [54]. The core advantage of the CNN model is to extract the significant features without human supervision [55]. Therefore, it enables the automatic representation of promising features compared to the conventional ML. It also eliminates the need for manual feature extraction as pre-processing and enables us to work on large and unstructured datasets. This automatic feature extraction in CNN makes it preferable over other ML techniques in image segmentation, classification, and object recognition tasks [56].

2.5. CNN in CAD Systems

Due to the automatic representation of promising features, CNN has become popular in image classification, segmentation, and object detection problems. Therefore, it has been widely used in CADs for several diseases such as skin cancer, breast cancer, and lung cancer [57]. In [58], the authors used a simple CNN architecture for successfully detecting skin cancer in images. In [59], the authors achieved 98.94% accuracy in detecting breast cancer from mammogram images. Similarly, a lung cancer detection system was designed to detect lung cancer in low-dose computer tomography (LDCT) scans with an AUC of 0.913 [60]. In [61], the authors used a simple CNN for detecting Alzheimer’s disease in brain MRIs.
Similarly, CNN has also been very effective for brain tumor detection, classification, and segmentation [62,63,64]. In [62], a CNN was developed based on clustering in MRI images. In [63], the authors used a CNN for MRI image segmentation while using a small 3 × 3 kernel. In [65], the authors employed an Artificial Neural Network (ANN) in MR image classification. A study presented in [66] performed classification on MRI images for tumor detection using KNN, SVM, Linear Discriminant Analysis (LDA), and DNN and concluded that DNN outperformed the other classification techniques. In [67], the authors designed a simple CNN for brain tumor classification. They used three different types of input: original images, cropped images, and segmented regions. Gray Level Co-occurrence Matrix (GLCM) performed feature extraction, and then those features were used by CNN to perform classification [68]. Similarly, DNN was used with auto-encoders to classify brain tumors. The Discrete Wavelet Transform (DWT) and GLCM were used in the first phase to extract texture and intensity-based features from brain MRIs. These features were then passed on to a DNN, which consisted of two auto-encoders and one SoftMax layer to perform classification. Similarly, in [69], the authors used GLCM and DWT for feature extraction and then used a Probabilistic Neural Network (PNN) to classify the tumor images. In [70], the authors performed multi-class brain tumor classification on T1-weighted contrast-enhanced images on two different datasets using a simple CNN model. First, MRIs from the first dataset was classified into three tumor types: meningioma, glioma, and pituitary tumor. Second, the images from a different dataset were classified into different grades of glioma. In a study presented in [34], the authors performed tumor classification using CNN architectures with different depths. The proposed study was used to classify MRIs into three different tumor types: meningioma, glioma, and pituitary. The results concluded that increasing the depth of the architecture does not always improve accuracy. However, increasing the number of convolution filters does improve performance.
Some studies introduced randomized ensemble CNN for detecting brain tumors and early stages of Alzheimer’s disease. In [71], the authors suggested a novel 2D-CNN randomized ensemble approach to detect the early stages of Alzheimer’s disease using magnetoencephalography (MEG) activity. The proposed randomized ensemble approach handles both noisy and sparse data. Similarly, in [72], the authors developed a multi-path lightweight deep CNN model using randomized dilated convolutions. The proposed model consists of the two multi-path networks, where the features map of the first network is used as an input for the second network. The resulting analysis indicates that the proposed multi-path reduced the classification error by 0.8% compared to the conventional techniques.

2.6. Transfer Learning

Transfer learning has also shown great promise in medical applications. Transfer learning allows a CNN model trained for a separate related problem to solve the current problem. CNN models such as AlexNet [73], GoogleNet [74], and ResNet [75], which were originally trained on an ImageNet dataset consisting of 1000 different classes, can be retrained on a different dataset. The presented study in [76] performed a classification of CT images into benign and malignant renal tumors by using a pre-trained Inception V3 model. In [77], the authors used a pre-trained GoogleNet to classify MRIs into different types of tumors and achieved an accuracy of 98% and outperformed other state-of-the-art methods. Another study presented in [78] performed image classification using different pre-trained models: GoogleNet, ResNet, AlexNet, and SqeezeNet. While the AlexNet was able to give the highest classification accuracy, it had a very high computation cost compared to the other networks. In another research study [79], the authors used nine different pre-trained CNN models for brain tumor classification. The results indicated that VGG16, VGG19, and AlexNet perform better than deeper models such as ResNet, GoogleNet, and SeNEt.

2.7. Summary

To the best of our knowledge, all aforementioned studies attempted to utilize conventional ML and serial DL architectures to identify and classify tumors in brain MRIs and CT. However, the traditional ML techniques are ineffective because these models require extensive domain knowledge and time to manually extract features from the MRIs. In contrast, serial DL architectures utilized a uniform kernel size for the convolutional layers that restrict the feature extraction process. Furthermore, it is imperative to analyze small and large kernel sizes for extracting relevant features from brain MRI to enhance the performance of the brain tumor detection systems. Furthermore, the existing studies did not analyze the impact of magnetically injected Rician noise on classification results. Hence, our study is the first attempt to propose multi-scale CNN architectures with parallel convolution paths having different kernel sizes and analyze the impact of Rician noise in brain MRIs on the classification performance of CNN. Moreover, this study focuses on optimizing/shrinking trainable parameters of the proposed and existing DL architectures to enhance the efficiency of tumor detection in brain MRIs.

3. Proposed Methodology

This section presents the proposed methodology of a multi-scale CNN model for identifying and classifying brain tumors in four classes such as glioma, meningioma, pituitary, and non-tumor MRI. The proposed multi-scale model aims to perform multi-class classification, where brain tumors are classified into four classes, as shown in Figure 1.

3.1. Overview of Proposed Model

An overview of the proposed multi-scale CNN model is presented in Figure 2. The proposed multi-scale CNN model consists of multiple parallel convolutions paths having different filter sizes. The main objective of the proposed multi-scale CNN architecture is to analyze the impact of convolutional filters having different sizes on the classification of brain tumors. Therefore, different filter sizes are considered, such as 3 × 3, 5 × 5, 7 × 7, and 9 × 9. First, the pre-processed MRI is passed to each convolution path to extract features. Each convolutional path has a different filter size with the same number of filters to extract features from pre-processed MRIs. Furthermore, each convolutional path consists of a uniform number of convolutional layers. In addition, the activation function and max pooling layers are used for each convolutional layer to map inputs into output and reduce computational complexity. Next, a concentration layer is deployed to integrate parallel convolutional path feature maps. The integrated features map is then passed to the fully connected layer, where each neuron of the fully connected layer is connected with neurons of the previous layer. Finally, a SoftMax method is used as an activation function that uses a probabilistic approach to predict the best class label for the given inputs. It is used to determine the probability for each class label to predict the best class label.

3.2. General Flow of Proposed Model

In this subsection, a general flow of the CNN model is discussed. CNN architecture takes inspiration from the biological processes of a real eye. The pattern of connectivity among the neurons resembles that of the visual cortex of animals. A CNN performs several linear and nonlinear transformations on the input data to get more useful abstract representations [80]. In addition, CNN tries to learn the relationship between the pixels of the input images. The typical architecture of a CNN has an input layer, followed by alternating convolution and max pool layers. The output of these layers is then passed onto fully-connected hidden layers. Figure 3 shows the architecture of a generic CNN architecture.

3.2.1. Convolutional Layers

The convolution layers consist of several convolutional filters that are used to extract features from the images. First, the input data is multiplied with a two-dimensional array of weights in this layer, called a kernel or a filter. Each kernel is designed to extract a specific type of feature. Then, the kernel is applied over the entire image. The result of this operation is a two-dimensional matrix called the feature map. Equation (1) gives the formula for calculating the feature map.
G [ m , n ] = j k f [ j , k ] I [ m j , n k ]
where G is the output feature map, I is the input images, and f is the kernel. The network uses more than one filter to extract different types of features. Each filter gives a two-dimensional matrix as an output. The output matrices of all the filters are stacked to form a three-dimensional tensor. Equation (2) expresses the dimensions of the output tensor.
( n , n , n c ) ( f , f , f c ) = n + 2 p f s + 1 , n + 2 p f s , n f
where n represents the size of the images, f represents the size of the filter, n c indicates the number of channels, p indicates padding, s represents stride, and n f represents the total number of filters.
These filters do not need to be designed by hand. Instead, the values of these filters can be learned during the training to determine what type of feature needs to be extracted. The first layer extracts simple, low-level features from the images. The later successor layers, however, extract more complex high-level features. Figure 4 describes a convolution operation.

3.2.2. Non-Linearity Layer

The non-linearity layer is used to introduce non-linearity in the network. It uses a set of functions that are called the activation function. These functions are used to remove redundant data while preserving useful data [81]. The most commonly used activation function is the Rectified Linear Unit (ReLU) function [82]. Different variants of ReLU are available, such as Leaky ReLU (LReLU) [83], Parametric ReLU (PReLU) [84], and Randomized ReLU (RReLU) [85]. A simple description of the ReLU function is given in Equation (3).
f ( z x , y ) = z x , y , i f   z x , y > 0 0 , i f   z x , y 0
where z x , y is the input value at position ( x , y ) .

3.2.3. Pooling Layers

Pooling layers are used for down-sampling. They reduce the resolution of the output from the convolution layer. They combine a cluster of output neurons into a single neuron before sending the image to the next layer. Pooling can be max or average.
Figure 5 describes the working of max pooling and average pooling operations using a 2 × 2 filter. The max pooling selects the maximum of the cluster, while the average pooling selects the cluster’s average. The purpose of pooling is to reduce the number of parameters and avoid over-fitting. The output size M of a pooling layer can be calculated by using Equation (4).
m = n + 2 p f s + 1
where n is the input image size, f is the size of the pooling filter, p is the padding, and s is stride.
Unlike the convolution layer, the pooling layer does not have any learnable parameters [86].

3.2.4. Fully Connected Layer

The fully connected layer is like a Multi-layer perceptron. As the name suggests, in a fully connected layer, every single neuron in one layer is connected to each neuron in another layer. The 2D matrix from the previous layers is flattened and converted into a vector before passing to the fully-connected layer. This layer is usually used as an output layer. It uses the one-dimensional vector representation of the feature map of the image outputs the class label. Figure 6 describes the basic structure of the fully connected dense layer. It is the final layer where all the computations and reasoning are performed on the features received from the concatenation layer.

3.3. Proposed Multi-Scale CNN Architecture

The architecture of the proposed multi-scale CNN model is developed based on parallel convolutions paths. The architecture of the proposed multi-scale CNN model consists of the following main steps: a collection of MRI data, pre-processing of the acquired data, a configuration of multiple parallel convolutional paths, and performance analysis of multi-classification outcomes. First, MRI data are collected from Kaggle, consisting of 3264 images of four classes: glioma, meningioma, pituitary, and non-tumor images. Second, a pre-processing module is developed to pre-process raw MRI by removing noise to enhance the quality of the given MRIs. Third, a CNN model is developed based on multiple parallel convolutional paths to extract features from the pre-processed MRIs to detect and identify brain tumors. Finally, performance analyses are considered to evaluate the accuracy and efficiency of the proposed multi-scale CNN model. The detailed flow of the proposed multi-scale CNN model is as follows.

3.3.1. Collection of MRI Data

In this research study, an open-source, publicly available MRI dataset is acquired from Kaggle [87]. The dataset contains 3264 MRIs from four classes: glioma, meningioma, pituitary, and non-tumor images and the combination of T1, T2, and Flair types. The acquired data consists of the commonly used MRI sequences, such as T1, T2, and Fluid Attenuated Inversion Recovery (Flair). T1 images are generated based on short Time to Echo (TE) and Repetition Time (RT). T1 properties of the tissue are used to determine the contrast and brightness of the MRI. Similarly, T2 images are generated using long TE and TR times. In T2 images, T2 tissue properties are used for contrast and brightness. Furthermore, the Flair sequence is similar to the T2 sequence but uses very long TE and TR times to produce the MRI. The class distribution of the dataset is given in Table 1. Out of 3264 MRIs, 926 MRIs belong to glioma, 937 to meningioma, 500 are from pituitary, and the remaining 901 are from non-tumor patients.
Furthermore, Figure 7 shows data distribution analysis according to class labels. It found that 29% of data instances belong to class meningioma. Similarly, 926 (28%) and 500 (15%) MRI samples belong to glioma and pituitary, respectively. Furthermore, 29% of MRI samples belong to the class non-tumor.

3.3.2. Pre-Processing of MRI Data

MRI images suffer from Rician noise. The presence of noise can affect the accuracy of the classification model. Therefore, this noise needs to be removed before any classification can be performed in the pre-processing stage. For this purpose, an FSNLM filter was used. This technique uses fuzzy logic to find segments in the image that are homogeneous to the noisy pixel. Once these segments have been identified, they can be used to find the noise-less pixel. The FSNLM filter’s advantage over other denoising techniques is its low computational cost and its ability to retain the level of details in the MRIs. The FSNLM has been reported with a mean PSNR of 19.7 for T1 weighted images and 17.5 for T2 weighted images [22]. Figure 8 shows the actual noisy and denoised MRI.

3.3.3. Proposed Multi-Scale CNN Architecture

This research study develops a CNN model based on a multi-scale architecture to detect and classify brain tumors from prepared MRIs. The CNN performs a convolution operation on the input images to extract promising features. However, the extracted features depend on the size of the kernel. A traditional approach uses only one kernel size. The size of the kernel restricts the features that can be extracted from the images. The multi-scale approach uses more than one kernel size, which allows the extraction of a wider range of features from the input images. Each parallel path extracts distinct features from the images. The output of each path is passed to the depth concatenation layer. The depth concatenation layer is deployed to concatenate the features map and pass them to the fully connected layer.
Furthermore, three different multi-scale architectures are used for the proposed CNN model. These three architectures are named MCNN1, MCNN2, and MCNN3, and they have 2, 3, and 4 parallel convolutional paths, respectively. The reason for testing more than one multi-scale architecture is to understand the effect of adding more parallel paths to the performance. Figure 9 presents the detailed architecture of the proposed CNN (MCNN1) based on a multi-scale architecture with two parallel paths. The proposed MCNN1 consists of two parallel convolutional paths to extract features from the prepared MRIs. The same input image is passed to both parallel paths to extract different feature sets according to kernel size. Furthermore, it can be seen that two different kernel sizes are used, such as 3 × 3 and 5 × 5, to extract features and pass them to the concatenation layer for integration purposes. In addition, integrated features are flattened from a 2D matrix to a 1D vector and given to the fully-connected layer. Moreover, a fully-connected layer is an output layer that employs a Feed-Forward Neural Network (FFNN) to map inputs received from the previous layer into a classification outcome using a SoftMax method.
Similarly, Figure 10 shows the detailed architecture of the proposed MCNN2 based on a multi-scale architecture with three parallel convolutional paths. The proposed MCNN2 uses three parallel convolutional paths with different kernel sizes, such as 3 × 3, 5 × 5, and 7 × 7, to extract rich features from the input MRIs. The variation in kernel sizes for each convolutional path extracts a distinct set of features. The reason behind using different kernel sizes for each convolutional path is to analyze the effect of multi-scale architecture in terms of accuracy and efficiency.
Furthermore, Figure 11 presents a detailed architecture for the proposed MCNN3. In the MCNN3 model, parallel convolutional paths are increased to analyze the effect on the performance of the proposed multi-scale architecture. It uses four parallel convolutional paths with different kernel sizes, such as 3 × 3, 5 × 5, 7 × 7, and 9 × 9, to extract underlying features from input MRIs. Thus, three different multi-scale architectures are proposed to analyze the accuracy and efficiency of the proposed CNN model. Finally, the performance of the proposed CNN models is compared with state-of-the-art DL architectures, such as AlexNet and ResNet.

4. Experimental Results and Analysis

This section presents the experimental environment, evaluation metrics, experimental results, and analysis. First, an experimental environment of the proposed research study is presented. Second, different evaluation metrics are discussed to evaluate the performance of proposed and conventional DL architectures. Third, experimental results of proposed and state-of-the-art techniques such as AlexNet and Renset18 are discussed. Finally, comparative analyses show the impact of the proposed multi-scale CNN model over traditional CNN architectures.

4.1. Experimental Environment

The experiments are performed using Matlab R2020b on two MRI datasets: original and denoised images, to understand the effect of noise on the performance of the proposed multi-scale CNN models. Table 2 gives the specification of the systems on which all the experiments were performed. Furthermore, a dataset is divided into training and testing sample sets. The k-fold validation strategy is utilized to ensure the quality of the results. Moreover, the k-fold (i.e., k = 10) cross-validation method is used to ensure the generalization and fast convergence of the proposed multi-scale CNN models.

4.2. Evaluation Measures

For performance evaluation of the proposed model, four evaluation measures were used: accuracy, precision, recall, and F1-score [88]. These measures are calculated with the help of the confusion matrix. Table 3 shows an example of a confusion matrix for a binary classification problem.
Where T g , T m , T p , and T n t indicate instances that were actually true and correctly predicted true; T n shows instances that were actually negative and correctly predicted negative. Similarly, F n indicates instances that were actually positive but falsely predicted negative; F p shows the instance that was actually negative but falsely predicted positive.
Based on the confusion matrix, different evaluation measures are constructed, such as accuracy, precision, recall, and F1-score.
Accuracy is defined as the ratio between the correctly predicted instances (true positive) and the total number of instances N. Accuracy is expressed for the aforementioned multi-classification problem as shown in Equation (5).
Accuracy = T g + T m + T p + T n t N
Precision is defined as “out of all the actual positives, how many were correctly predicted as positive”. For multi-classification, it is defined for class label i as the sum over the ith row of the confusion matrix. Precision is expressed in Equation (6).
Precision i = M i i j M i j
where M represents a matrix, i is defined for rows (predicted label), and j is defined for columns (actual label).
Similarly, recall is defined as “out of all the predicted positives, how many are actually positive”. For the multi-classification problem, it is calculated for the class label ith as the sum over columns of the confusion matrix. The recall is expressed in Equation (7).
Recall i = M i i j M j i
Furthermore, the F1-score is defined as the harmonic mean of precision and recall. The F1-score is expressed in Equation (8).
F1-score = 2 Precision Recall Precision + Recall

4.3. Experimental Results and Performance Analysis

This section presents the experimental results of the proposed and state-of-the-art DL architectures. In this research study, experiments were performed in four different stages.
  • In the first stage, the raw images were used for classification by all models.
  • In the second stage, the images were denoised using an FSNLM filter and were classified using all models.
  • In the third stage, an artificial noise of magnitude 0.05 was added to the original images. These noisy images were then classified using all models.
  • Finally, images with artificial noise were denoised using an FSNLM filter and sent to the classifiers.

4.3.1. Original (Raw) MRI Data

In this stage, the proposed and existing DL architectures are employed on an original set of MRI data. Table 4 shows the accuracy, precision, recall (aka sensitivity), specificity, and F1-score of the five classification models on the original MRI dataset. The classification results indicate that the proposed MCNN2 model yields higher accuracy than other listed DL architectures. Furthermore, the classification results of the proposed MCNN2 model are significantly higher than the existing AlexNet architecture. Similarly, the classification performance of the proposed MCNN2 model is slightly higher compared to the ResNet architecture. Moreover, it shows that the proposed MCNN2 gives high specificity compared to other listed models. Based on the evaluation results, it is concluded that the proposed MCNN2 model gives an accurate classification rate on original MRI data compared to other DL architectures.

4.3.2. Denoised Images Dataset

Next, Table 5 shows the classification results of the five classification models on the denoised images dataset. The classification results indicate that the proposed MCNN2 model gives superior performance compared to the AlexNet, ResNet, and other CNN architectures. Furthermore, the classification results of the proposed MCNN2 using denoised images are much better than the existing AlexNet architecture. It is also found that the performance of MCNN2 in terms of specificity is better than conventional CNN architectures. Similarly, it is also evident that the proposed MCNN2 model performed slightly better than the ResNet architecture. The evaluation results concluded that the proposed MCNN2 model gives high classification performance on denoised image data compared to other DL architectures.

4.3.3. Synthetic Noise Based Dataset

Although some noise is present in the original images, the level of noise is very low. To deeply observe the effect on noise, a synthetic noise of 0.05 magnitude was added to the original images. These images were then classified using all five classifiers. Table 6 shows the classification results of all the classifiers on images with synthetic noise. The classification results using synthetic noise indicate that the proposed MCNN2 model performed significantly well compared to the AlexNet architecture. The resulting analysis shows that the proposed MCCN2 model improved accuracy by 6.73%, precision by 8.34%, recall by 5.44%, specificity by 7.65%, and F1-score by 6.91% compared to the AlexNet architecture. Furthermore, the classification results of the proposed MCNN2 were slightly better than the existing AlexNet architecture. The classification results using artificial noise data concluded that the proposed MCNN2 model gives high classification performance compared to the listed DL architectures.

4.3.4. Synthetic Noise Removed

In the final stage, the images from dataset 3 are denoised using an FSNLM filter to remove the synthetic noise. Table 7 shows the classification results of these images. The evaluation analysis indicates that the MCNN2 gives higher accuracy, recall, and F1-score than AlexNet, ResNet, and other multi-scale CNN architectures. Furthermore, the ResNet architecture gives a high precision of 0.92301 compared to other DL architectures. Moreover, MCNN2 gives a slightly better specificity of 0.89671 than ResNet. Hence, our MCNN2 architecture yields high performance and outperforms state-of-the-art DL techniques.

5. Comparative Analysis

This section presents a comparative analysis to show the effectiveness and significance of the proposed research study. This research study proposes three different multi-scale architectures for brain MRI classification. The proposed architectures are named MCNN1, MCNN2, and MCNN3 and have 2, 3, and 4 parallel convolution paths, respectively. For performance evaluation, the proposed architectures’ classification results were compared with state-of-the-art CNN models, AlexNet and ResNet, using measures of accuracy, precision, recall, and F1-score. Accuracy was normalized between 0 and 1 to make it easier to present. The experiments were performed on four different datasets, each having a different level of noise.
Figure 12 shows the average accuracy, precision, recall, and F1-score of the five classification models across the four different datasets. The comparison of the results indicated that the proposed architecture, MCNN2, performed better than state-of-the-art methods in all the evaluation measures. MCNN2 was able to classify the MRIs with an accuracy of 91.2% and an F1-score of 0.91%, which is higher than not only the other two proposed models but also the state-of-the-art techniques such as ResNet and AlexNet. Not only does the proposed model give better classification, but it does so at a lower computational cost. AlexNet and ResNet 18 have approximately 61 million and 11 million trainable parameters, respectively [89]; MCNN2 uses only around 1.8 million trainable parameters.
Furthermore, the research study also studied the effect of noise on the classification results. For this purpose, all five classifiers were applied to the dataset in four different stages. Each stage had a different level of noise in the images. The original images with naturally existing noise were used in the first stage. In the second stage, images were denoised using an FSNLM filter. In the third stage, artificial Rician noise was added to the original image dataset. Finally, the images with artificial noise were denoised using an FSNLM filter in the fourth stage. Figure 13 shows the average classification results of all the classifiers on each of the datasets. The results indicate that the presence of noise adversely affects the classification results. While the original images produced decent results, they had some noise. However, when the images were denoised, it improved performance. Furthermore, adding artificial noise to the images resulted in a decline in performance of the classifiers. However, removing the noise from these images recovered the performance.
Moreover, Table 8 presents a comparative analysis to evaluate the efficiency of the proposed multi-scale approaches with state-of-the-art architectures, such as ResNet and AlexNet. The efficiency of the proposed and existing models is evaluated based on trainable parameters (weights and biases). The comparative analysis shows that our proposed multi-scale approaches require a small number of trainable parameters to extract features compared to ResNet and AlexNet. Our proposed multi-scale approaches drastically reduce the number of trainable parameters to extract features. The comparative analysis shows that the total number of trainable parameters for the proposed MCNN1, MCNN2, and NCNN3 is 3.443%, 8.191%, and 15.966% of the ResNet architecture. Similarly, it is also found that the total number of trainable parameters for the proposed MCNN1, MCNN2, and MCNN3 is 0.635%, 1.511%, and 2.946% of AlexNet. Hence, our proposed multi-scale approaches enhance efficiency for tumor detection in brain MRIs compared to the existing models such as ResNet and AlexNet [89].

6. Conclusions

The advancements in machine learning have paved the way for reliable computer-aided diagnosis systems, especially for MRI image classification. This study presents an effective and efficient deep learning model for brain tumor MRI classification. The model is compared with state-of-the-art techniques using accuracy, precision, recall, and F1-score. This research study presented a multi-scale CNN architecture for brain tumor MRI classification. The proposed model consisted of three parallel convolutional paths using kernel sizes of 3, 5, and 7. The model was tested on noisy and denoised images, and the results were compared with state-of-the-art CNN architectures: ResNet and AlexNet. The results portrayed that the proposed model MCNN2 not only gave better accuracy but also had less trainable parameters, decreasing its complexity. However, increasing the number of paths from three to four caused an immediate decline in the performance. The architecture with three parallel paths produced the best classification results. This study also analyzed the effect of Rician noise in MRIs on the classification performance of the deep learning model. The results indicated that the presence of noise in the MRIs caused a decline in the performance of the classification models; however, the effectiveness of FSNLM showed that MCNN2 can perform reasonably better once the image is denoised with FSNLM. The experimental results showed that increasing the filter size had a negative impact on the performance of the proposed architecture. Furthermore, the number of filters is empirically decided. The experiments are carried out without the supervision of a radiologist, and the findings are purely based on the dataset.

Author Contributions

Conceptualization, S.A.Y.; Data curation, S.A.Y.; Formal analysis, N.I., A.R., A.N.K. and D.-H.K.; Funding acquisition, D.-H.K.; Methodology, S.A.Y. and R.A.; Project administration, R.A.; Resources, R.A.; Software, N.I. and A.N.K.; Supervision, R.A. and D.-H.K.; Validation, N.I. and D.-H.K.; Visualization, S.A.Y., A.R. and A.N.K.; Writing—original draft, S.A.Y.; Writing—review and editing, R.A., N.I., A.R., A.N.K. and D.-H.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Energy Cloud R&D Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science, ICT (2019M3F2A1073387), and this work was supported by Electronics and Telecommunications Research Institute (ETRI) grant funded by the Korean government. [22ZD1140, Development of ICT Convergence Technology for Daegu-Gyeongbuk Regional Industry].

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare that there is no conflict of interest related to this publication.

References

  1. Li, Y.Q.; Chiu, K.S.; Liu, X.R.; Hsiao, T.Y.; Zhao, G.; Li, S.J.; Lin, C.P.; Sun, C.W. Polarization-sensitive optical coherence tomography for brain tumor characterization. IEEE J. Sel. Top. Quantum Electron. 2019, 25, 1–7. [Google Scholar] [CrossRef]
  2. Harish, P.; Baskar, S. MRI based detection and classification of brain tumor using enhanced faster R-CNN and Alex Net model. Mater. Today Proc. 2020; in press. [Google Scholar]
  3. Ferlay, J.; Soerjomataram, I.; Dikshit, R.; Eser, S.; Mathers, C.; Rebelo, M.; Parkin, D.M.; Forman, D.; Bray, F. Cancer incidence and mortality worldwide: Sources, methods and major patterns in GLOBOCAN 2012. Int. J. Cancer 2015, 136, E359–E386. [Google Scholar] [CrossRef]
  4. Siegel, R.L.; Miller, K.D.; Jemal, A. Cancer statistics, 2019. CA Cancer J. Clin. 2019, 69, 7–34. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Shakeel, P.M.; Tobely, T.E.E.; Al-Feel, H.; Manogaran, G.; Baskar, S. Neural network based brain tumor detection using wireless infrared imaging sensor. IEEE Access 2019, 7, 5577–5588. [Google Scholar] [CrossRef]
  6. Bray, F.; Ferlay, J.; Soerjomataram, I.; Siegel, R.L.; Torre, L.A.; Jemal, A. Global cancer statistics 2018: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2018, 68, 394–424. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Siegel, R.L.; Miller, K.D.; Goding Sauer, A.; Fedewa, S.A.; Butterly, L.F.; Anderson, J.C.; Cercek, A.; Smith, R.A.; Jemal, A. Colorectal cancer statistics, 2020. CA Cancer J. Clin. 2020, 70, 145–164. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  8. Miller, K.D.; Ostrom, Q.T.; Kruchko, C.; Patil, N.; Tihan, T.; Cioffi, G.; Fuchs, H.E.; Waite, K.A.; Jemal, A.; Siegel, R.L.; et al. Brain and other central nervous system tumor statistics, 2021. CA Cancer J. Clin. 2021, 71, 381–406. [Google Scholar] [CrossRef] [PubMed]
  9. Westbrook, C.; Talbot, J. MRI in Practice; John Wiley & Sons: Hoboken, NJ, USA, 2018. [Google Scholar]
  10. Comelli, A.; Stefano, A.; Bignardi, S.; Coronnello, C.; Russo, G.; Sabini, M.G.; Ippolito, M.; Yezzi, A. Tissue Classification to Support Local Active Delineation of Brain Tumors. In Medical Image Understanding and Analysis; Springer International Publishing: Cham, Switzerland, 2020; pp. 3–14. [Google Scholar]
  11. Bauer, S.; Wiest, R.; Nolte, L.P.; Reyes, M. A survey of MRI-based medical image analysis for brain tumor studies. Phys. Med. Biol. 2013, 58, R97. [Google Scholar] [CrossRef] [Green Version]
  12. Abd-Ellah, M.K.; Awad, A.I.; Khalaf, A.A.; Hamed, H.F. A review on brain tumor diagnosis from MRI images: Practical implications, key achievements, and lessons learned. Magn. Reson. Imaging 2019, 61, 300–318. [Google Scholar] [CrossRef]
  13. Lotlikar, V.S.; Satpute, N.; Gupta, A. Brain Tumor Detection Using Machine Learning and Deep Learning: A Review. Curr. Med. Imaging 2022, 18, 604–622. [Google Scholar] [CrossRef]
  14. Hussain, M.; Saher, N.; Qadri, S. Computer Vision Approach for Liver Tumor Classification Using CT Dataset. Appl. Artif. Intell. 2022, 36, 1–23. [Google Scholar] [CrossRef]
  15. Mao, Y.J.; Lim, H.J.; Ni, M.; Yan, W.H.; Wong, D.W.C.; Cheung, J.C.W. Breast tumour classification using ultrasound elastography with machine learning: A systematic scoping review. Cancers 2022, 14, 367. [Google Scholar] [CrossRef] [PubMed]
  16. Angulakshmi, M.; Lakshmi Priya, G. Automated brain tumour segmentation techniques—A review. Int. J. Imaging Syst. Technol. 2017, 27, 66–77. [Google Scholar] [CrossRef] [Green Version]
  17. Saladi, S.; Amutha Prabha, N. Analysis of denoising filters on MRI brain images. Int. J. Imaging Syst. Technol. 2017, 27, 201–208. [Google Scholar] [CrossRef]
  18. Kale, S.C.; Chen, X.J.; Henkelman, R.M. Trading off SNR and resolution in MR images. NMR Biomed. Int. J. Devoted Dev. Appl. Magn. Reson. In Vivo 2009, 22, 488–494. [Google Scholar] [CrossRef]
  19. Aja-Fernández, S.; Alberola-López, C.; Westin, C.F. Noise and signal estimation in magnitude MRI and Rician distributed images: A LMMSE approach. IEEE Trans. Image Process. 2008, 17, 1383–1398. [Google Scholar] [CrossRef] [Green Version]
  20. Dakua, S.; Sahambi, J. Noise effect on LV image segmentation. In Proceedings of the 2009 Annual IEEE India Conference, Ahmedabad, India, 18–20 December 2009; pp. 1–4. [Google Scholar]
  21. Suganthi, P.D.M.; Deepa, P. Performance evaluation of various denoising filters for medical image. Int. J. Comput. Sci. Inf. Technol. 2014, 5, 4205–4209. [Google Scholar]
  22. Sharif, M.; Hussain, A.; Jaffar, M.A.; Choi, T.S. Fuzzy similarity based non local means filter for rician noise removal. Multimed. Tools Appl. 2015, 74, 5533–5556. [Google Scholar] [CrossRef]
  23. Ahmadvand, A.; Daliri, M.R. Brain MR image segmentation methods and applications. OMICS J. Radiol. 2014, 3, e130. [Google Scholar] [CrossRef]
  24. Edalati-rad, A.; Mosleh, M. Improving Brain Tumor Diagnosis Using MRI Segmentation Based on Collaboration of Beta Mixture Model and Learning Automata. Arab. J. Sci. Eng. 2019, 44, 2945–2957. [Google Scholar] [CrossRef]
  25. Shelke, S.M.; Mohod, S.W. Automated segmentation and detection of brain tumor from MRI. In Proceedings of the 2018 International Conference on Advances in Computing, Communications and Informatics (ICACCI), Bangalore, India, 19–22 September 2018; pp. 2120–2126. [Google Scholar]
  26. Naik, J.; Patel, S. Tumor detection and classification using decision tree in brain MRI. Int. J. Comput. Sci. Netw. Secur. IJCSNS 2014, 14, 87. [Google Scholar]
  27. Anitha, R.; Siva Sundhara Raja, D. Development of computer-aided approach for brain tumor detection using random forest classifier. Int. J. Imaging Syst. Technol. 2018, 28, 48–53. [Google Scholar] [CrossRef]
  28. Srikrishna, M.; Pereira, J.B.; Heckemann, R.A.; Volpe, G.; van Westen, D.; Zettergren, A.; Kern, S.; Wahlund, L.O.; Westman, E.; Skoog, I.; et al. Deep learning from MRI-derived labels enables automatic brain tissue classification on human brain CT. NeuroImage 2021, 244, 118606. [Google Scholar] [CrossRef]
  29. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Into Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef] [Green Version]
  30. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  31. Toğaçar, M.; Ergen, B.; Cömert, Z. BrainMRNet: Brain tumor detection using magnetic resonance images with a novel convolutional neural network model. Med. Hypotheses 2020, 134, 109531. [Google Scholar] [CrossRef]
  32. Brain Tumor. Available online: https://www.healthline.com/health/brain-tumor/ (accessed on 20 May 2021).
  33. Brain Tumor: Classifications, Symptoms, Diagnosis and Treatments. Available online: https://www.aans.org/en/Publications/AANS-Neurosurgeon (accessed on 22 May 2021).
  34. Abiwinanda, N.; Hanif, M.; Hesaputra, S.T.; Handayani, A.; Mengko, T.R. Brain tumor classification using convolutional neural network. In World Congress on Medical Physics and Biomedical Engineering 2018; Springer: Berlin/Heidelberg, Germany, 2019; pp. 183–189. [Google Scholar]
  35. Classification of Brain Tumors. Available online: https://www.aans.org/ (accessed on 28 May 2021).
  36. Brain Cancer & Brain Tumor: Symptoms, Causes & Treatments. Available online: https://my.clevelandclinic.org/health/diseases/6149-brain-cancer-brain-tumor/ (accessed on 29 May 2021).
  37. Brain Tumor Types, Glioblastoma, Meningioma & More. Available online: https://miamineurosciencecenter.com/en/conditions/brain-tumors/types/ (accessed on 29 May 2021).
  38. Meningioma Brain Tumor: Causes, Symptoms, and Treatment. Available online: https://www.webmd.com/cancer/brain-cancer/meningioma-causes-symptoms-treatment/ (accessed on 29 May 2021).
  39. Zhang, Y.D.; Wu, L.; Wang, S. Magnetic resonance brain image classification by an improved artificial bee colony algorithm. Prog. Electromagn. Res. 2011, 116, 65–79. [Google Scholar] [CrossRef] [Green Version]
  40. Brain Tumor: Diagnosis. Available online: https://www.cancer.net/cancer-types/brain-tumor/diagnosis/ (accessed on 2 June 2021).
  41. Vaishali, S.; Rao, K.K.; Rao, G.S. A review on noise reduction methods for brain MRI images. In Proceedings of the 2015 International Conference on Signal Processing and Communication Engineering Systems, Guntur, India, 2–3 January 2015; pp. 363–365. [Google Scholar]
  42. Bhukra, M.J.; Sharma, K.K. Image Denoising and Rician Noise Reduction by LRMD, SVM and Iterative Bilateral Filter in different type of Medical Images. Image 2018, 5, 303–307. [Google Scholar]
  43. Kala, R.; Deepa, P. Adaptive hexagonal fuzzy hybrid filter for Rician noise removal in MRI images. Neural Comput. Appl. 2018, 29, 237–249. [Google Scholar] [CrossRef]
  44. Zeng, Y.; Zhang, B.; Zhao, W.; Xiao, S.; Zhang, G.; Ren, H.; Zhao, W.; Peng, Y.; Xiao, Y.; Lu, Y.; et al. Magnetic resonance image denoising algorithm based on cartoon, texture, and residual parts. Comput. Math. Methods Med. 2020, 2020, 1405647. [Google Scholar] [CrossRef]
  45. Ananthi, V.; Balasubramaniam, P. A new image denoising method using interval-valued intuitionistic fuzzy sets for the removal of impulse noise. Signal Process. 2016, 121, 81–93. [Google Scholar] [CrossRef]
  46. Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Enhanced performance of brain tumor classification via tumor region augmentation and partition. PLoS ONE 2015, 10, e0140381. [Google Scholar] [CrossRef]
  47. Rodríguez-Méndez, I.A.; Ureña, R.; Herrera-Viedma, E. Fuzzy clustering approach for brain tumor tissue segmentation in magnetic resonance images. Soft Comput. 2019, 23, 10105–10117. [Google Scholar] [CrossRef]
  48. Nabizadeh, N.; Kubat, M. Automatic tumor segmentation in single-spectral MRI using a texture-based and contour-based algorithm. Expert Syst. Appl. 2017, 77, 1–10. [Google Scholar] [CrossRef]
  49. Mustaqeem, A.; Javed, A.; Fatima, T. An efficient brain tumor detection algorithm using watershed & thresholding based segmentation. Int. J. Image Graph. Signal Process. 2012, 4, 34. [Google Scholar]
  50. Ilhan, U.; Ilhan, A. Brain tumor segmentation based on a new threshold approach. Procedia Comput. Sci. 2017, 120, 580–587. [Google Scholar] [CrossRef]
  51. Said, A.M.; Ibrahim, F.S. Comparative Study of Segmentation Techniques for Detection of Tumors Based on MRI Brain Images. Int. J. Biosci. Biochem. Bioinform. 2017, 8. [Google Scholar] [CrossRef] [Green Version]
  52. Gopal, N.N.; Karnan, M. Diagnose brain tumor through MRI using image processing clustering algorithms such as Fuzzy C Means along with intelligent optimization techniques. In Proceedings of the 2010 IEEE International Conference on Computational Intelligence and Computing Research, Coimbatore, India, 28–29 December 2010; pp. 1–4. [Google Scholar]
  53. Sharif, M.; Tanvir, U.; Munir, E.U.; Khan, M.A.; Yasmin, M. Brain tumor segmentation and classification by improved binomial thresholding and multi-features selection. J. Ambient. Intell. Humaniz. Comput. 2018, 1–20. [Google Scholar] [CrossRef]
  54. Chahal, P.K.; Pandey, S.; Goel, S. A survey on brain tumor detection techniques for MR images. Multimed. Tools Appl. 2020, 79, 21771–21814. [Google Scholar] [CrossRef]
  55. Suzuki, K. Overview of deep learning in medical imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef]
  56. LeCun, Y.; Haffner, P.; Bottou, L.; Bengio, Y. Object recognition with gradient-based learning. In Shape, Contour and Grouping in Computer Vision; Springer: Berlin/Heidelberg, Germany, 1999; pp. 319–345. [Google Scholar]
  57. Agnes, S.A.; Anitha, J.; Pandian, S.I.A.; Peter, J.D. Classification of mammogram images using multiscale all convolutional neural network (MA-CNN). J. Med. Syst. 2020, 44, 1–9. [Google Scholar] [CrossRef] [PubMed]
  58. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef] [PubMed]
  59. Chougrad, H.; Zouaki, H.; Alheyane, O. Deep convolutional neural networks for breast cancer screening. Comput. Methods Programs Biomed. 2018, 157, 19–30. [Google Scholar] [CrossRef] [PubMed]
  60. Perez, G.; Arbelaez, P. Automated lung cancer diagnosis using three-dimensional convolutional neural networks. Med. Biol. Eng. Comput. 2020, 58, 1803–1815. [Google Scholar] [CrossRef]
  61. Islam, J.; Zhang, Y. Brain MRI analysis for Alzheimer’s disease diagnosis using an ensemble system of deep convolutional neural networks. Brain Inform. 2018, 5, 1–14. [Google Scholar] [CrossRef]
  62. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.M.; Larochelle, H. Brain tumor segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [Google Scholar] [CrossRef] [Green Version]
  63. Pereira, S.; Pinto, A.; Alves, V.; Silva, C.A. Brain tumor segmentation using convolutional neural networks in MRI images. IEEE Trans. Med. Imaging 2016, 35, 1240–1251. [Google Scholar] [CrossRef]
  64. Thaha, M.M.; Kumar, K.P.M.; Murugan, B.; Dhanasekeran, S.; Vijayakarthick, P.; Selvi, A.S. Brain tumor segmentation using convolutional neural networks in MRI images. J. Med. Syst. 2019, 43, 1–10. [Google Scholar] [CrossRef]
  65. El Abbadi, N.K.; Kadhim, N.E. Brain cancer classification based on features and artificial neural network. Brain 2017, 6, 123–134. [Google Scholar] [CrossRef]
  66. Mohsen, H.; El-Dahshan, E.S.A.; El-Horbaty, E.S.M.; Salem, A.B.M. Classification using deep learning neural networks for brain tumors. Future Comput. Inform. J. 2018, 3, 68–71. [Google Scholar] [CrossRef]
  67. Alqudah, A.M.; Alquraan, H.; Qasmieh, I.A.; Alqudah, A.; Al-Sharu, W. Brain Tumor Classification Using Deep Learning Technique–A Comparison between Cropped, Uncropped, and Segmented Lesion Images with Different Sizes. arXiv 2020, arXiv:2001.08844. [Google Scholar] [CrossRef]
  68. Widhiarso, W.; Yohannes, Y.; Prakarsah, C. Brain tumor classification using gray level co-occurrence matrix and convolutional neural network. IJEIS Indones. J. Electron. Instrum. Syst. 2018, 8, 179–190. [Google Scholar] [CrossRef]
  69. Shree, N.V.; Kumar, T. Identification and classification of brain tumor MRI images with feature extraction using DWT and probabilistic neural network. Brain Inform. 2018, 5, 23–30. [Google Scholar] [CrossRef] [Green Version]
  70. Sultan, H.H.; Salem, N.M.; Al-Atabany, W. Multi-classification of brain tumor images using deep neural network. IEEE Access 2019, 7, 69215–69225. [Google Scholar] [CrossRef]
  71. Lopez-Martin, M.; Nevado, A.; Carro, B. Detection of early stages of Alzheimer’s disease based on MEG activity with a randomized convolutional neural network. Artif. Intell. Med. 2020, 107, 101924. [Google Scholar] [CrossRef]
  72. Park, S.; Chang, D.E. Multipath Lightweight Deep Network Using Randomly Selected Dilated Convolution. Sensors 2021, 21, 7862. [Google Scholar] [CrossRef]
  73. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  74. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  75. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  76. Zhou, L.; Zhang, Z.; Chen, Y.C.; Zhao, Z.Y.; Yin, X.D.; Jiang, H.B. A deep learning-based radiomics model for differentiating benign and malignant renal tumors. Transl. Oncol. 2019, 12, 292–300. [Google Scholar] [CrossRef]
  77. Deepak, S.; Ameer, P. Brain tumor classification using deep CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef]
  78. Mehrotra, R.; Ansari, M.; Agrawal, R.; Anand, R. A Transfer Learning approach for AI-based classification of brain tumors. Mach. Learn. Appl. 2020, 2, 100003. [Google Scholar] [CrossRef]
  79. Chelghoum, R.; Ikhlef, A.; Hameurlaine, A.; Jacquir, S. Transfer learning using convolutional neural network architectures for brain tumor classification from MRI images. In IFIP International Conference on Artificial Intelligence Applications and Innovations; Springer: Berlin/Heidelberg, Germany, 2020; pp. 189–200. [Google Scholar]
  80. Basheera, S.; Ram, M.S.S. Classification of brain tumors using deep features extracted using CNN. J. Phys. Conf. Ser. 2019, 1172, 012016. [Google Scholar] [CrossRef]
  81. Lin, G.; Shen, W. Research on convolutional neural network based on improved Relu piecewise activation function. Procedia Comput. Sci. 2018, 131, 977–984. [Google Scholar] [CrossRef]
  82. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier nonlinearities improve neural network acoustic models. Proc. ICML CiteSeerX 2013, 30, 3. [Google Scholar]
  83. Dubey, A.K.; Jain, V. Comparative study of convolution neural network’s relu and leaky-relu activation functions. In Applications of Computing, Automation and Wireless Systems in Electrical Engineering; Springer: Berlin/Heidelberg, Germany, 2019; pp. 873–880. [Google Scholar]
  84. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 11–18 December 2015; pp. 1026–1034. [Google Scholar]
  85. Xu, B.; Wang, N.; Chen, T.; Li, M. Empirical evaluation of rectified activations in convolutional network. arXiv 2015, arXiv:1505.00853. [Google Scholar]
  86. Brain Tumor Classification (MRI) Dataset. Available online: https://missinglink.ai/guides/convolutional-neural-networks/fully-connected-layers-convolutional-neural-networks-complete-guide/ (accessed on 10 June 2021).
  87. Fully Connected Layers in Convolutional Neural Networks: The Complete Guide. Available online: https://www.kaggle.com/sartajbhuvaji/brain-tumor-classification-mri/ (accessed on 25 May 2021).
  88. Iqbal, N.; Ahmad, R.; Jamil, F.; Kim, D.H. Hybrid features prediction model of movie quality using Multi-machine learning techniques for effective business resource planning. J. Intell. Fuzzy Syst. 2021, 40, 9361–9382. [Google Scholar] [CrossRef]
  89. Anwar, A. Difference between AlexNet, VGGNet, ResNet and Inception. Medium-Towards Data Sci. 2019. [Google Scholar]
Figure 1. Classification of brain tumors from MRIs.
Figure 1. Classification of brain tumors from MRIs.
Tomography 08 00161 g001
Figure 2. Block model of the proposed multi-scale CNN architecture.
Figure 2. Block model of the proposed multi-scale CNN architecture.
Tomography 08 00161 g002
Figure 3. General architecture of a conventional CNN model.
Figure 3. General architecture of a conventional CNN model.
Tomography 08 00161 g003
Figure 4. Convolutional example using a 2 × 2 filter.
Figure 4. Convolutional example using a 2 × 2 filter.
Tomography 08 00161 g004
Figure 5. Comparison of max and average pooling.
Figure 5. Comparison of max and average pooling.
Tomography 08 00161 g005
Figure 6. Basic structure of the fully connected layer.
Figure 6. Basic structure of the fully connected layer.
Tomography 08 00161 g006
Figure 7. Data distribution analysis according to class labels.
Figure 7. Data distribution analysis according to class labels.
Tomography 08 00161 g007
Figure 8. Pre-processed MRI data (Before and After Denoising).
Figure 8. Pre-processed MRI data (Before and After Denoising).
Tomography 08 00161 g008
Figure 9. Detailed architecture of the proposed MCNN1.
Figure 9. Detailed architecture of the proposed MCNN1.
Tomography 08 00161 g009
Figure 10. Detailed architecture of the proposed MCNN2.
Figure 10. Detailed architecture of the proposed MCNN2.
Tomography 08 00161 g010
Figure 11. Detailed architecture of the proposed MCNN3.
Figure 11. Detailed architecture of the proposed MCNN3.
Tomography 08 00161 g011
Figure 12. Comparative analysis of proposed and state-of-the-art DL techniques.
Figure 12. Comparative analysis of proposed and state-of-the-art DL techniques.
Tomography 08 00161 g012
Figure 13. Comparative analysis of proposed and state-of-the-art DL techniques using noisy and denoised images.
Figure 13. Comparative analysis of proposed and state-of-the-art DL techniques using noisy and denoised images.
Tomography 08 00161 g013
Table 1. Data distribution as per class label.
Table 1. Data distribution as per class label.
#ClassNumber of Images
1Glioma926
2Meningioma937
3Pituitary500
4Non-Tumor901
5Total3264
Table 2. Implementation environment for the proposed multi-scale CNN architectures.
Table 2. Implementation environment for the proposed multi-scale CNN architectures.
System SpecificationsDescription
Operating SystemWindows 10
ProcessorIntel Core i5, 2nd Generation
Main Memory10 GB
IDEMatlab R2020b
GPUNvidia GTX 750 TI
Table 3. Confusion matrix.
Table 3. Confusion matrix.
Predicted Label
GliomaMeningiomaPituitaryNon-Tumor
Actual LabelGlioma T g F m F p F n t
Meningioma F g T m F p F n t
Pituitary F g F m T p F n t
Non-Tumor F g F m F p T n t
Table 4. Classification results using original (raw) MRIs data.
Table 4. Classification results using original (raw) MRIs data.
ClassifierAccuracyPrecisionRecallSpecificityF1-Score
AlexNet0.845570.862060.842610.857310.83727
ResNet0.909170.912050.908090.878740.90906
MCNN10.827180.826980.829650.831430.82073
MCNN20.911320.9140.911530.907980.9122
MCNN30.83180.840520.815730.823690.82157
Table 5. Classification results using denoised image data.
Table 5. Classification results using denoised image data.
ClassifierAccuracyPrecisionRecallSpecificityF1-Score
AlexNet0.878900.889860.878590.854210.88032
ResNet0.919810.918760.91440.89790.91593
MCNN10.892660.898510.891540.86910.89405
MCNN20.941900.944540.937410.926170.94057
MCNN30.896720.901820.892370.883510.89488
Table 6. Classification results using added synthetic noise to image data.
Table 6. Classification results using added synthetic noise to image data.
ClassifierAccuracyPrecisionRecallSpecificityF1-Score
AlexNet0.81040.809130.813260.794610.80584
ResNet0.874230.872230.87120.852780.86941
MCNN10.74540.759640.749780.725840.73574
MCNN20.877680.892510.867610.878210.87493
MCNN30.816510.807280.818640.794310.80437
Table 7. Classifications results for denoised images.
Table 7. Classifications results for denoised images.
ClassifierAccuracyPrecisionRecallSpecificityF1-Score
AlexNet0.892970.896290.889860.879210.89283
ResNet0.917430.923010.881720.892830.91920
MCNN10.899080.848480.903230.869910.875
MCNN20.923550.919290.923510.896710.92122
MCNN30.905200.912970.907580.872310.90905
Table 8. Comparative analysis of trainable parameters for proposed multi-scale and existing DL architectures.
Table 8. Comparative analysis of trainable parameters for proposed multi-scale and existing DL architectures.
#ClassifierTrainable Parameters
1MCNN1396,356
2MCNN2942,948
3MCNN31,837,956
4ResNet11,511,784
5AlexNet62,378,334
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yazdan, S.A.; Ahmad, R.; Iqbal, N.; Rizwan, A.; Khan, A.N.; Kim, D.-H. An Efficient Multi-Scale Convolutional Neural Network Based Multi-Class Brain MRI Classification for SaMD. Tomography 2022, 8, 1905-1927. https://doi.org/10.3390/tomography8040161

AMA Style

Yazdan SA, Ahmad R, Iqbal N, Rizwan A, Khan AN, Kim D-H. An Efficient Multi-Scale Convolutional Neural Network Based Multi-Class Brain MRI Classification for SaMD. Tomography. 2022; 8(4):1905-1927. https://doi.org/10.3390/tomography8040161

Chicago/Turabian Style

Yazdan, Syed Ali, Rashid Ahmad, Naeem Iqbal, Atif Rizwan, Anam Nawaz Khan, and Do-Hyeun Kim. 2022. "An Efficient Multi-Scale Convolutional Neural Network Based Multi-Class Brain MRI Classification for SaMD" Tomography 8, no. 4: 1905-1927. https://doi.org/10.3390/tomography8040161

APA Style

Yazdan, S. A., Ahmad, R., Iqbal, N., Rizwan, A., Khan, A. N., & Kim, D. -H. (2022). An Efficient Multi-Scale Convolutional Neural Network Based Multi-Class Brain MRI Classification for SaMD. Tomography, 8(4), 1905-1927. https://doi.org/10.3390/tomography8040161

Article Metrics

Back to TopTop