Next Article in Journal
Clinicopathological and Genetic Profiles of Cases with Myocyte Disarray—Investigation for Establishing the Autopsy Diagnostic Criteria for Hypertrophic Cardiomyopathy
Next Article in Special Issue
Developing a Diagnostic Decision Support System for Benign Paroxysmal Positional Vertigo Using a Deep-Learning Model
Previous Article in Journal
Characteristics of ADHD Symptom Response/Remission in a Clinical Trial of Methylphenidate Extended Release
Previous Article in Special Issue
Global Evolution of Research in Artificial Intelligence in Health and Medicine: A Bibliometric Study
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effective Diagnosis and Treatment through Content-Based Medical Image Retrieval (CBMIR) by Using Artificial Intelligence

Division of Electronics and Electrical Engineering, Dongguk University, 30 Pildong-ro 1-gil, Jung-gu, Seoul 04620, Korea
*
Author to whom correspondence should be addressed.
J. Clin. Med. 2019, 8(4), 462; https://doi.org/10.3390/jcm8040462
Submission received: 7 March 2019 / Revised: 2 April 2019 / Accepted: 3 April 2019 / Published: 5 April 2019
(This article belongs to the Special Issue The Future of Artificial Intelligence in Clinical Medicine)

Abstract

:
Medical-image-based diagnosis is a tedious task‚ and small lesions in various medical images can be overlooked by medical experts due to the limited attention span of the human visual system, which can adversely affect medical treatment. However, this problem can be resolved by exploring similar cases in the previous medical database through an efficient content-based medical image retrieval (CBMIR) system. In the past few years, heterogeneous medical imaging databases have been growing rapidly with the advent of different types of medical imaging modalities. Recently, a medical doctor usually refers to various types of imaging modalities all together such as computed tomography (CT), magnetic resonance imaging (MRI), X-ray, and ultrasound, etc of various organs in order for the diagnosis and treatment of specific disease. Accurate classification and retrieval of multimodal medical imaging data is the key challenge for the CBMIR system. Most previous attempts use handcrafted features for medical image classification and retrieval, which show low performance for a massive collection of multimodal databases. Although there are a few previous studies on the use of deep features for classification, the number of classes is very small. To solve this problem, we propose the classification-based retrieval system of the multimodal medical images from various types of imaging modalities by using the technique of artificial intelligence, named as an enhanced residual network (ResNet). Experimental results with 12 databases including 50 classes demonstrate that the accuracy and F1.score by our method are respectively 81.51% and 82.42% which are higher than those by the previous method of CBMIR (the accuracy of 69.71% and F1.score of 69.63%).

1. Introduction

Over the past few decades, computer-aided diagnosis (CAD) tools and techniques have been widely adopted for better medical treatment [1,2]. These modern tools support medical experts in many areas such as medical diagnosis and treatment for any specific disease or injury. In the current era of medical science, many computer-aided tools provide visual information for diagnosis and treatment such as magnetic resonance imaging (MRI), X-ray, angiography, computed tomography (CT), digital mammography, optical projection tomography (OPT), colonoscopy, ultrasonography, optical endoscopy, nuclear medical imaging‚ and positron-emission tomography (PET) [3,4]. These various medical imaging modalities provide visual insight into different hidden body organs, thus enabling better diagnosis and treatment. Medical image analysis is a challenging task due to the complex structure of body organs‚ and medical experts are required for accurate interpretation [4]. To arrive at a sound decision about a serious medical condition, past relevant cases are explored by many medical experts. This practice can facilitate better diagnosis and treatment. However, due to the enormous number of medical visual records generated by many different imaging modalities, it is very difficult and time-consuming to retrieve relevant cases. This problem can be resolved by using a computer-based medical image retrieval system (MIRS), which helps medical experts in retrieving past relevant cases from previous patients’ databases.
To develop such a MIRS, image classification is the key challenge due to the existence of highly correlated visual features among different classes, which ultimately results in low retrieval performance. This problem can be solved by using advanced machine learning tools and techniques that may result in better classification performance. Better performance can be achieved by exploring hidden features that the human visual system (HVS) finds very difficult to identify. In the past few years, significant advances have been made in the area of machine learning and artificial intelligence (AI) including deep learning framework [5]. The key idea behind deep learning is analogous to the operation of human brain, in which information is also processed through multiple layers of transformation [5,6].
Deep learning methods have shown significant performance in general content-based image retrieval (CBIR) applications [7]. In the past few years, deep learning models have made significant contributions in various medical domains [8,9] including brain tumor detection [1], blood flow quantification and visualization, diabetic retinopathy (DR), and many cancer detection applications. However, these methods are still in the developmental phase for content-based medical image retrieval (CBMIR) tasks, due to the rapid growth in medical imaging technology [10]. This paper mainly focuses on the analysis of different deep learning models used in medical image classification and retrieval. We analyze in depth the performance of the most recent convolutional neural network (CNN) models from the following standpoints, by considering: (1) different configuration modes of CNN models, (2) feature selection from different layers within a network, (3) training from scratch and fine-tuning, and (4) modification of the pre-trained model. In this way, we proposed the best CNN model after modifying the existing model to obtain the best classification accuracy. Finally, we provide a first pre-trained model for a heterogeneous medical database including the number of classes captured by different modalities, which is our main contribution, and we have also made our pre-trained model and image indices of experimental images publicly available for other researchers.
CBMIR is an active area of research with significant applications in routine clinical diagnostic aid, medical education, and research. Many solved cases related to different diseases can be stored in a picture archiving and communication system (PACS) or in CBMIR systems with comprehensive patient record and treatment details. In the future, similar cases can be diagnosed in less time by exploring such previous records. In this way, medical experts can save precious time and improve diagnosis and treatment. Moreover, CBMIR is also helpful in medical teaching and research areas.
The rest of the paper is organized as follows: In Section 2, we describe the related studies. Section 3 summarizes the main contribution of this paper. The proposed CNN-based classification method for medical image retrieval is described in Section 4. In Section 5 and Section 6, the experimental setup, performance analysis, and discussions are presented. Finally, Section 7 concludes our research.

2. Related Works

The present era of digital technology has made a significant contribution to medical science. The number of medical imaging modalities is growing rapidly with improvements in biomedical sensors and high-throughput image acquisition technologies. These devices generate an enormous collection of heterogeneous medical images that make a significant contribution to disease analysis and treatment. A medical expert can make a better diagnosis related to a similar situation in the past by retrieving relevant cases from this enormous collection of medical images. Before the advent of machine learning (ML) and AI algorithms, it was considered a tedious task to explore the huge multimodal database for getting assistance related to any complex problem. Hence, it is important to evolve an efficient MIRS that will support medical experts and thus improve diagnosis and treatment.
Conventional text-based image retrieval systems use certain textual tags that images are often manually annotated with as search keywords. Due to the enormous collection of heterogeneous medical image databases, this manual annotation task is very tedious and time-consuming. In many hospitals, the PACS [11] is deployed to manage a very large collection of medical images that is compatible with the digital imaging and communications in medicine (DICOM) file format [12]. This framework utilizes the textual information stored in the DICOM header for image retrieval; the header contains a patient identifier (ID), name, date, modality, body parts examined, etc. This header information is lost when a DICOM image is converted into another image format for efficient storage and communication such as tagged image file format (TIFF), joint photographic experts group (JPEG), portable network graphics (PNG), etc. To resolve this problem, CBMIR systems have been proposed by many researchers to assist medical experts. However, these systems are application-specific and can store or retrieve a specific type of medical image, e.g., a retrieval system for X-ray images of the chest as proposed in [13].
Although many researchers have studied the CBMIR by using handcrafted features [14,15,16,17,18,19,20,21,22,23,24,25,26], the overall performance of the existing systems is still low due to the growing heterogeneous medical images of multiclass database and conventional ML techniques. These techniques are unable to decrease the “semantic gap,” which is the information lost by converting an image (i.e., a high-level representation) into its visual features (i.e., a low-level representation) [27]. Recently, a significant breakthrough has occurred in the ML domain with the advent of the deep learning framework, which comprises many efficient ML algorithms that can show high-level abstractions in visual data with a minimum semantic gap [28]. Ultimately, these layers extract the complex deep features from the input data in a fully systematic way. Finally, the deep network learns from these features without using other handcrafted features.
In recent studies, a significant breakthrough in deep learning has been done in the medical domain, and they are classified into two categories of single modality-based [29,30,31,32,33,34,35,36] and multiple modalities-based methods [28] of imaging.
As the single modality-based method, a two-stage CBMIR framework is presented for automatic retrieval of radiographic images [29]. In the first stage, the main class label is assigned by using CNN-based features, and in the second stage, outlier images are filtered out from the predicted class on the basis of low-level edge histogram features. Another CNN-based system is presented in [30] for categorization of interstitial lung diseases (ILDs) patterns by extraction of ILD features from the selected dataset. In [31], a convolutional classification restricted Boltzmann machine (RBM)-based framework is proposed for analyzing the lung CT scan by combining both generative and discriminative representation learning. A CNN-based automatic classification of peri-fissural nodules (PFN) is presented in [32], which has high relevance in the context of lung cancer screening. In [33], a two-stage multi-instance deep learning framework is presented for the classification of different body organs. In the first stage, a CNN is trained on local patches to separate discriminative and non-informative patches from training data samples. The network is then fine-tuned on extracted discriminative patches for the classification task. A detailed analysis of deep learning in CAD is presented in [37]. Three main characteristics (i.e., different CNN architectures, dataset scale, and transfer learning) of CNN are explored in this work. A deep CNN model pre-trained on the general dataset is then fine-tuned for a large collection of multimodal medical image databases. A fully automatic 3D CNN framework to detect cerebral microbleeds (CMBs) from MRI is proposed in [34]. CMBs are small hemorrhages near blood vessels whose detection provides deep insight into many cerebrovascular diseases and cognitive dysfunctions. In [35], an efficient CNN training method is proposed by dynamically choosing negative samples (misclassified) during the training process, which shows better performance in hemorrhage detection within a color fundus image. A multiview convolutional network (ConvNets)-based CAD system is proposed [36] for detecting pulmonary nodules from lung CT scan images.
As the multiple modalities-based method, a deep-learning-based framework for multiclass CBMIR is recently proposed in [28] that can classify multimodal medical images. In this framework, an intermodal dataset that contains twenty-four classes with five modalities (CT, MRI, fundus camera, PET, and OPT) is used to train the network.
The maximum numbers of classes can usually increase the usability of CBMIR system in healthcare medical application [28]. In addition, it is reported that a large number of classes can help the medical expert in exploring the specific class of disease from a huge collection of medical record according to [38] and healthcare professional. Nevertheless, in previous researches, the maximum numbers of classes to be dealt with were limited as 31 [20,29], and we increased the numbers of classes as 50 in our research. For this purpose, we propose a deep-feature-based medical image classification and retrieval framework by using the enhanced residual network (ResNet) for CBMIR of large numbers of classes with nine modalities (CT, MRI, fundus camera, PET, OPT, X-ray, ultrasound, endoscopy, and visible light camera). The strengths and weaknesses of our proposed and existing methods are summarized in Table 1.

3. Contribution

Our research is novel in the following six ways compared to previous works. The brief definitions of the closed-world, open-world, and mixed-world configurations in the 1st contribution are as follows. The closed-world configuration means the case that the classes in training are same as those used in testing whereas the open-world configuration represents the case that the classes in training are different from those used in testing and the classes in testing are unknown. The mixed-world configuration is the combination of the closed-world and open-world configurations, and it means the case that some parts of testing data are not known in training process whereas the others of testing data are known in training process. Detail definitions of these three configurations are explained in Section 5.3.5.
-
This is the first approach toward classifying the large collection of multiclass medical image databases with multiple modalities based on the deep residual network in the closed-world, open-world, and mixed-world configurations. Different from our research, most of the previous studies [10,28,29,30,31,32,33,34,35,36] have been conducted only in a closed-world configuration.
-
In general, the problem for classification with larger numbers of classes is more difficult than that with fewer numbers of classes. Based on the theories in pattern recognition, the inter-distance between classes in case of larger numbers of classes becomes smaller than that in case of fewer numbers of classes. This increases the possibility of overlapping of data from different classes, and consequent classification error is increased [39,40,41]. It is also experimentally confirmed that the previous method [28] shows the accuracy of F1.score as 69.63% with 50 classes whereas it presents the accuracy of F1.score as 99.76% with 24 classes [28].
-
In our proposed medical image classification and retrieval framework, we modified the conventional ResNet50 [42] CNN model by replacing its last 7 × 7 average pooling layer with a 7 × 7 × 2048 convolutional layer. Finally, the number of nodes in the last fully connected (FC) layer is also adjusted according to the number of classes in our dataset.
-
We deeply analyze the characteristics of various CNNs for multiclass medical images, and then check how a specific CNN structure can influence the classification performance of multiclass medical images.
-
We compare the performance of state-of-the-art CNN models, not only through fine-tuning and tuning from scratch but also against different handcrafted approaches. Our analysis is more detailed, in contrast to previous studies [10,28], which provided only a limited performance comparison for a small number of databases.
-
We analyze the performance of a CNN model based on feature selection from the different layers of the network.
-
We have made our trained model and image indices of experimental images publicly available through [43], so that other researchers can evaluate and compare its performance.

4. Proposed Method

4.1. An Overview of the Proposed Approach

Figure 1 presents a brief flowchart for the classification of medical images by using our modified deep residual CNN framework. In the first step, the given medical image was resized into 224 × 224 × 3 for input to our CNN model. The resized image was then passed to a deep residual CNN model for feature extraction from the last convolutional layer. In this way, a deep feature vector ( 1 × 2048 ) was obtained, which represents the complex hidden structure of the given input image (i.e., high-level representation) as a feature vector (i.e., low-level representation). This extracted feature vector was compared one by one with the labeled feature vectors in the database by measuring the Euclidean distance. Finally, a class label was assigned to the given input image on the basis of the minimum distance score. A detailed explanation of our proposed model is provided in subsequent sections.

4.2. The Structure of our Modified Deep Residual CNN

In our proposed medical image classification and retrieval framework, we modified the conventional ResNet50 [42] CNN model by replacing its last 7 × 7 average pooling layer with a 7 × 7 × 2048 convolutional layer. The reasons for using 7 × 7 × 2048 convolutional layer are as follows. Compared to the classification of general images, the classification of medical images has the problems of high inter-class similarity. Therefore, more features which can be useful for the classification should be extracted from the CNN. The original ResNet50 [42] obtains the feature map of 1 × 1 × 2048 from the previous feature map of 7 × 7 × 2048 by using average pooling layer including one filter of 7 × 7, which can cause the loss of useful features. To solve this problem, our revised ResNet50 obtained the feature map of 1 × 1 × 2048 from the previous feature map of 7 × 7 × 2048 by using the additional convolution layer (Conv6 in Table 2) including 2048 filters of 7 × 7 × 2048, which can reduce the loss of useful features. In addition, the filter coefficients of average pooling layer in original ResNet50 are fixed ones, whereas the optimal filter coefficients of the additional convolutional layer in our revised ResNet50 can be obtained by training. In order to prove this, we experimentally compared the accuracies by original ResNet50 [42] using the average pooling layer including 1 filter of 7 × 7 with those by our revised ResNet50 using the additional convolution layer including 2048 filters of 7 × 7 × 2048. The accuracies by our method are higher than those by original ResNet50 [42].
Finally, the number of nodes in the last FC layer is adjusted according to the number of classes in our dataset. The modified structure of the CNN with the complete layer configuration is presented in Figure 2 and Table 2.
Our modified deep residual CNN network was made up of multiple residual units that can be considered as a basic building block. These residual units included both identity-mapping-based and 1 × 1 convolutional-mapping-based shortcut connections [42]. The shortcut connection in identity-mapping-based residual unit mapped the input feature map as it was, without changing its size and depth. On the other hand, the shortcut connection in the 1 × 1 convolutional-mapping-based residual unit increased the depth of the input feature map. Our deep residual network contained a total of 16 residual units, in which there were 12 identity mapping units and four convolutional mapping units of 1 × 1, as shown in Figure 3. Using more residual units as identity mapping decreases the complexity and training time. Furthermore, both identity and 1 × 1 convolutional shortcut connections make information propagation smooth in both forward and backward directions [44].
The detailed layer configuration for our model is given in Table 2. Conv1–Conv6 was the convolutional layers stack in which Conv2–Conv5 represent the group of convolutional layers. Each individual group comprised multiple residual units including only one 1 × 1 convolutional-mapping-based residual unit and multiple identity-mapping-based residual units. The number of identity mapping units was different in each group, which is represented by the number of iterations in Table 2. In addition, max pool was a subsampling layer that was used to select the maximum value in a subregion of the feature map defined by the kernel size. Its main purpose was to reduce the feature map size by preserving information on key features. Finally, the FC layer, SoftMax, and the classification layer were used to classify the features extracted from the previous convolutional layers.

4.2.1. Feature Extraction

The convolutional layer stack was used for feature extraction by applying a traditional 2D convolution operation using a different number of filters with different sizes. These filters contained learnable parameters that were determined during the training procedure. When convolution was applied, the output feature map size changed depending upon the filter size, number of filters, the stride values for the horizontal and vertical directions, and the range of filter movement using padding options. All these parameters, known as hyperparameters, were defined during the network construction phase. Therefore, they were very important for constructing an efficient model. In our deep residual CNN, Conv1 had 64 filters of 7 × 7 × 3 and it explored the given input image X 1 of 224 × 224 × 3 , in both the horizontal and vertical directions with a stride of two pixel units, and a padding of three pixel units in both directions. The max pool layer had one filter of 3 × 3 pixels that explored the output feature map X 2 of Conv1 in both the horizontal and vertical directions with a stride of two pixel units for each input channel, and generated a down-sampled feature map of 56 × 56 × 64 .
Conv2–conv5 was the group of multiple convolutional layers that comprise multiple residual units. In each group, there was only one convolutional mapping unit at the start, followed by multiple identity mapping units. As given in Table 2, in the first group (Conv2), Conv2–1 presents the first convolutional-mapping-based residual unit, consisting of four convolutional layers with filters w 2 , 1 , w 2 , 2 , w 2 , 3 , w 2 , 4 . In this residual unit, the first three and fourth convolutional layers were connected in a parallel fashion as shown in Figure 3b. The first three layers performed the convolution operation in sequential order for a given input X 2 by applying filters w 2 , 1 , w 2 , 2 ,   w 2 , 3 and generating the intermediate feature map as F ( X 2 , W 2 ) . The fourth layer converts the given input X 2 as H ( X 2 , W 2 ) by applying a 1 × 1 filter w 2 , 4 to equalize the depth size of X 2 according to F ( X 2 , W 2 ) . Finally, the output feature map X 3 of 56 × 56 × 256 was obtained by adding F ( X 2 , W 2 ) and H ( X 2 , W 2 ) . Conv2–2 presents the first identity-mapping-based residual unit including three convolutional layers, as shown in Figure 3a. These three layers further processed the output X 3 in sequential order by applying three different filters w 3 , 1 , w 3 , 2 , w 3 , 3 and generated the intermediate feature map F ( X 3 , W 3 ) . The final feature map X 4 of 56 × 56 × 256 was generated by adding F ( X 3 , W 3 ) and the previous output feature map X 3 . Similarly, Conv2–3 was the second identity-mapping-based residual unit, which performsed the same operation as in Conv2–2 and generated the feature map X 5 of 56 × 56 × 256 .
Similarly, all the other convolutional-mapping-based and identity-mapping-based residual units in groups Conv3, Conv4, and Conv5 performed the same operation as in Conv2. The only difference was the different number of filters and identity-mapping-based residual units in each group. Due to the different number of filters in each group, the input feature map depth also increased. Furthermore, the input feature map size decreased by a factor of two after passing through each successive group. The reason is that in each group, the first convolutional-mapping-based residual unit considers a unit stride of two pixels. In conclusion, each group, Conv3, Conv4, and Conv5, generated output feature maps ( X 9 ,   X 15 , and X 18 ) of 28 × 28 × 512 , 14 × 14 × 1024 , and 7 × 7 × 2048 , respectively, as shown in Table 2. Finally, the optimal feature vector x of 1 × 2048 was obtained after convolving the output X 18 of Conv5 with the last convolutional layer Conv6 using a filter of 7 × 7 × 2048 . Batch normalization and the rectified linear unit (ReLU) activation function were also applied after each convolutional layer on the basis of the mean and standard deviation of the data. The final feature vector x was further used as the input to the FC layer.

4.2.2. Classification

In our deep residual CNN method, we considered two classification architectures separately for both the training and testing phases. The 1st classification architecture used the 50 output nodes of classification layer in Table 2. For example, if the 2nd output node of classification layer with one input image showed the higher value than those from the other 49 nodes, the input image was determined as the class 2. The 1st classification architecture was used only for closed-world configuration.
The 2nd classification architecture determines the class of input image based on the Euclidean distance matching with the 2048 features extracted by Conv6 of Table 2. For example, if the distance between the 2048 features of input and 2048 mean feature vector of class 3 was the smallest, the input image was determined as class 3. The 2nd classification architecture was used for both open-world and mixed-world configurations.
For example, in open-world configuration, the data of C1, C2, … C25 were used for the training of CNN whereas those of C26, C27, … C50 were used for testing. Although the data of C26, C27, … C50 were not seen during the training, the 25 sets of 2048 mean feature vectors from C26, C27, … C50 were calculated and stored at our database in advance (the offline phase of Figure 4) for the Euclidean distance matching. For example, in mixed-world configuration, half of the data of C1, C2, … C40 were used for training of CNN whereas the other half data of C1, C2, … C40 and the whole data of C41, C42, … C50 were used for testing. In this case, the 10 sets of 2048 mean feature vectors from C41, C42, … C50 were calculated and stored at our database in advance (the offline phase of Figure 4) for the Euclidean distance matching.
The reason is that a variable number of testing classes can be considered in open-world and mixed-world configurations as compared to the closed-world configuration. In the closed-word configuration, the number of training and testing classes remained the same, and therefore there was no need to use a separate classifier for this operational mode. In this way, the FC part of our modified CNN model was used during the training phase. In the testing phase, our proposed deep-feature-based variable node classification (VNC) framework was deployed for class-prediction-based retrieval, whereas the feature extraction part remained the same in both the training and testing phases. In the training phase, a fully connected part mainly comprised the stack of the FC layer, the SoftMax layer, and the classification layer as shown in Figure 2. The FC layer (including the number of nodes which is equal to the total number of classes) was followed by the last convolutional layer Conv6 of the feature extraction part. The main purpose of this layer was to identify the larger patterns by combining all the features learned by the previous layers across the image. It multiplied the input feature vector obtained from Conv6 by a weight matrix and then added a bias vector. The next SoftMax layer of the classification part converted the output of the FC layer in terms of probability by applying the softmax function [45]. Finally, the classification layer took the output from the SoftMax layer and assigned each input to one of the 50 mutually exclusive classes using the cross-entropy function [45].
In the testing phase, the deep-feature-based VNC framework was implemented to classify the input query image in all configuration modes (i.e., closed-world, open-world, and mixed-world). A simple flow diagram of the proposed VNC framework was represented in Figure 4. There were two operational phases in the proposed VNC framework as in any general CBIR system, called the offline and online phases. In the offline phase, a database of K mean feature vectors was built from the available training dataset. For each individual class, a single mean feature vector was obtained. In this way, a set consisting of a total of 50 mean feature vectors was obtained from the entire training dataset. In our proposed work, the offline phase was mostly used in the open-world and mixed-world configurations in which the reference feature database can be updated for new classes without repeating the training process. In the online phase, the classification was performed for a given input query image by extracting and comparing its feature vector with the set of mean feature vectors by taking the L 2 -norm. Ultimately, the final class label was assigned on the basis of the minimum distance. In this way, efficient class-prediction-based image retrieval was performed by retrieving the required images from the selected class label rather than by exploring the entire dataset.

5. Experimental Setup and Performance Analysis

Several experiments were performed to evaluate our deep-feature-based method from various perspectives. Different comparisons were made with many deep CNN and handcrafted feature-based methods. For better performance analysis in a real-world scenario, the settings of the experiments were made in three different configuration modes (i.e., closed-world, open-world, and mixed-world). In this section, we describe the details of the selected dataset, experimental configurations as well as observations, and analysis of the results.

5.1. Dataset and Experimental Protocol

In our research, we focused on the classification-based retrieval of medical image having multiple classes with multiple imaging modalities. Medical image computing and computer assisted intervention (MICCAI) grand challenges share the medical images [46], but most of these images were used for the purpose of detection and segmentation instead of classification-based retrieval. Therefore, we did not use the benchmark of MICCAI grand challenges, but we categorized 12 different publicly available databases [47,48,49,50,51,52,53,54,55,56,57,58] into 50 different classes (i.e., C1 to C50) on the basis of different medical imaging modalities, body organs, and disease types. Because our experimental images were not collected by us, we cannot make them open to other researchers. Instead, we made the websites [47,48,49,50,51,52,53,54,55,56,57,58] with the image indices of our experimental images and our trained CNN models available to other researchers through [43] in order for fair comparisons with our method.
In this multimodal dataset, we randomly selected a maximum of 1000 images for each class. In this way, a dataset of 45,464 images was selected for 50 different classes in our experiments as shown in Figure 5. Figure 6 shows the examples of experimental data of Figure 5 according to the anatomical district (row) and imaging modality (column). The challenges in our data were high intra-class variance and high inter-class similarity caused by using multiple classes with multiple imaging modalities as shown in Figure 7 and Figure 8. We performed two-fold cross-validation by randomly dividing the whole dataset into almost 50% training and almost 50% testing. In other words, we used half images of C 1 ,   C 2 ,   C 3 , C 50 as training and the remaining half of C 1 ,   C 2 ,   C 3 , C 50 as testing in the closed-world configuration. In case of open-world configuration, we used all the images of C 1 ,   C 2 ,   C 3 , C 25 as training and those of C 26 ,   C 27 ,   C 28 , C 50 as testing. In the closed-world configuration, the training dataset mostly contained 500 images per class but in few classes, the number of training images was less than 500, which results in the class imbalance problem [59]. To avoid this problem, we generated some images by data augmentation using image translation and cropping, and in-plane rotation. This data augmentation was performed for only the training dataset.
A similar data augmentation procedure was adopted in the open-world and mixed-world configurations for the training dataset. A detailed description of the training and testing dataset in two-fold cross-validation is given in Table 3, and class imbalance details with augmented images are shown in Table 4.
All the images from each class were resized to 224 × 224 and converted into a standard bitmap (BMP) file format due to the different size and format of the collected dataset. We used the class label provided in the original datasets for supervised learning. Example images from the selected classes including actual class labels are shown in Figure 5.
Figure 7 and Figure 8 show the degree of intra-class variance and inter-class similarity in our collected dataset, respectively. A significant intra-class variation occurs among different images of a single class as shown in Figure 7. In addition, high inter-class similarity can be observed among different classes as shown in Figure 8. For example, in Figure 8b (cervix MRI) and Figure 8f (bladder MRI), a significant structural correlation can be observed between these two classes. This high degree of intra-class variance and inter-class similarity helps to analyze the performance of different models in a challenging scenario.
In our research study, all the experiments were performed by using a desktop computer with the following specifications: 3.50 GHz Intel® (Santa Clara, CA, USA) Core™ i7-3770K CPU [60] with 12 GB RAM, and NVIDIA (Santa Clara, CA, USA) GeForce GTX 1070 graphics card [61]. This graphics card provides parallel processing capability for both training and testing phase. All the training and testing CNN algorithms are implemented by MATLAB R2018b (MathWorks, Inc., Natick, MA, USA) [62] on the Windows 10 operating system.

5.2. Training of CNN Model

Before starting the training process, all the images in the dataset were resized to 224 × 224 × 3 . In case of images having a single channel such as CT, MRI, X-ray and ultrasound, we made the 3-channel image by copying the image of the 1st channel into those in the 2nd and 3rd channels. The same procedure of copying was performed in case of testing, also.
Our deep residual CNN was then trained by using the stochastic gradient descent (SGD) algorithm [63]. The SGD is the most commonly used algorithm for optimal training of CNNs, and it is very efficient in learning of discriminative linear classifiers with a convex loss function. Its main purpose is to optimize model learnable parameters such as filter weights and biases by taking the derivative of the loss function. In the training process, the correctly labeled data samples are used for the extraction of optimal features. These labeled training data samples passed through the feed-forward stage in the CNN, and then the loss between each actual and desired label is calculated. If the loss value was still greater than a certain threshold, the SGD further optimized the loss function by updating the parameters. The SGD method split the training dataset into mini-batches, performed an iteration for each mini-batch, and then proceeded to learn by defining the time taken for all iterations to complete as one epoch.
The hyperparameters selected for the SGD method in this study were as follows: mini-batch size = 10, learning rate = 0.001, learning rate drop factor = 0.1, learning rate drop period = 10, L2 regularization = 0.0001, and momentum = 0.9. The detailed explanation of each parameter can be found in [64]. During the training process, training data samples were shuffled, and the learning rate was multiplied by the learning rate drop factor for each 10-epoch period. The initial weights used in the FC layer were randomly initialized by using a Gaussian distribution with zero mean and 0.001 standard deviation, and the biases were initialized to zero.
Figure 9 shows the training loss and accuracy for each epoch from both folds of cross-validations. In all configurations, the loss approaches approximately zero while the training accuracy approaches 100% after a certain number of training epochs, which shows that our deep residual CNN is sufficiently trained with training data. In addition, after performing a number of training experiments for different CNN models, the conclusion is that the fine-tuning of our model results in faster convergence than with conventional training from scratch.
There existed only training and testing data in our experiments, and there was no validation dataset. The best CNN model was selected as follows. As shown in Figure 9, during the total epochs, all the weights of CNN model (whose training loss and accuracy were respectively lower and higher than those of previous model) were updated and stored at each iteration of training. Then, the weights of model which showed the minimum training loss and maximum training accuracy was finally selected as our CNN model.
Figure 10a,b visualize the significant differences in the learned filters from the first convolutional layer Conv1 of Table 2 after training from scratch and fine-tuning, respectively. The learned filters in Figure 10b after fine-tuning were more distinctive as compared to those extracted after training from scratch, as shown in Figure 10a, which shows that more useful features for classification can be extracted by fine-tuning.

5.3. Testing and Performance Analysis

The performance of the proposed method was evaluated in term of the average accuracy, average F1.score, mean average prevision (mAP), and mean average recall (mAR) [65], which were calculated as:
Accuracy = 1 K k = 1 K TP k + TN k TP k + TN k + FP k + FN k
F 1 . score = 2 × mAP × mAR mAP + mAR
mAP = 1 K k = 1 K TP k TP k + FP k
mAR = 1 K k = 1 K TP k TP k + TN k   ,
where TP k is the true positive, which denotes the correctly classified number of images from class k . FP k is the false positive, which shows the number of images misclassified as class k . TN k is the true negative, which indicates the number of images correctly classified as not belonging to class k . FN k is the false negative, which denotes the number of misclassified images that actually belong to class k . K represents the total number of classes, which equals 50 in our research.

5.3.1. Comparisons of Classification Accuracies by Proposed Modified Residual CNN with Various CNN Models

To evaluate the performance of the proposed deep CNN-based framework for medical image classification, a comparison was made with the most recent deep-learning-based medical image classification and retrieval framework [28,66]. In order to make a fair comparison, the performance of this existing framework was evaluated for our selected dataset. Our proposed method showed a significant performance gain in comparison with [28,66] as shown in Table 5. We also compared the performance of our model with the state-of-the-art methods of CNN models [42,66,68,69,70,71]. The main target of these comparisons was to evaluate the impact of the existing state-of-the-art CNN models in the medical domain. Finally, the impact of transfer learning was explored by training the selected CNN models in two different ways. In the first method, all the models were trained from scratch for our selected dataset. The experimental results for different baseline models are shown in Table 5 without using transfer learning. These results confirm that our modified deep residual model showed the highest average accuracy, and all other selected models also showed performance that was comparable with the existing framework [28].
In the second method, the impact of transfer learning was explored by fine-tuning the top three CNN models on the basis of the results of Table 5 and our modified deep residual model. These selected models were already pre-trained by the ImageNet dataset [72]. For transfer learning, the last few convolutions and all the FC layers (30% of the layers of the complete network) were fine-tuned by our selected dataset, and the filter weights for the initial convolutional layers (70% of the layers of the complete network) were optimized by the ImageNet dataset. The results of transfer learning are reported in Table 6. It can be observed that our modified deep residual model outperformed the other models after applying transfer learning in term of average accuracy, F1.score, mAP, and mAR.
Furthermore, a Monte Carlo simulation setup [73] was created to evaluate the robustness of the various trained CNN models. A detailed analysis is performed in this simulation setup. The performance of each model is iteratively evaluated for a random selection of the testing dataset. In each iteration, 20% of the testing images are selected randomly from the testing dataset, and a total of 20 iterations are performed for both folds. Finally, we calculate the standard deviation and average performance (i.e., accuracy and F1.score) for each model. Figure 11 shows the overall sensitivity performance in term of the average accuracy and F1.score. From the plots, we can see that our modified deep residual model has the best robustness among all the selected models. The second-best model is ResNet50, which also shows comparable performance.
The significance of our modified model was further explored in comparison with the second-best model, ResNet50 [42], by performing a t-test analysis [74]. Figure 12 shows the t-test performance for our modified and the second-best model. The t-test analysis was based on a null hypothesis, in which it was supposed that there was no performance difference between our modified model and the second-best model. After performing a t-test, the experimental results in Figure 12 show that the p-values of accuracy and F1.score for this test were 0.0488 (less than 0.05) and 0.0287 (less than 0.05), respectively. These results show that the null hypothesis for the overall average accuracy was rejected at a 95% confidence level, which indicates that there was a significant difference between the accuracy of our model and that of the second-best model. In addition, the null hypothesis for F1.score was also rejected at a 95% confidence level, which demonstrates the effective performance gain of our modified model in comparison with the second-best model.
The more detailed classification performance of our modified deep residual model in terms of the confusion matrix is shown in Figure 13. It can be observed from these results that there were only a few classes that showed a low classification performance, due to the significant structural similarity of the different neighboring body sections. For example, class 10 (i.e., bladder CT in Figure 5a-10) showed low performance because classes 10, 21 (i.e., kidney, renal CT in Figure 5c-1), and 26 (i.e., stomach CT in Figure 5c-6) belonged to neighboring body sections with a significant visual correlation. Similarly, the performance of class 4 (i.e., cervix MRI in Figure 5a-4) was also low due to structure overlapping with class 9 (i.e., bladder MRI in Figure 5a-9) and class 24 (i.e., pancreas MRI in Figure 5c-4). However, the overall performance of our proposed model was good for a heterogeneous dataset with a large number of classes.

5.3.2. Comparisons of Classification Accuracies according to the Features from Different Layers

To investigate whether our modified model can discover the required discriminative features at some intermediate layers, we performed additional experiments. The performance of our proposed and the second-best model (i.e., ResNet50) was analyzed in these experiments. There were two main reasons for comparing our model only with ResNet50 model. The first reason was that both models were similar in layer-wise structure with a small difference (i.e., the difference between the convolutional layer and average pooling layer as explained at the beginning of Section 4.2). The second reason was that all the other models have shown lower performance, which can be seen in the previous performance comparisons in Section 5.3.1.
A total of seven different layers (i.e., Conv2-1, Conv3-1, Conv4-1, Conv5-1, Conv6/AvgPool, FC layer, and Classification layer) were considered for extracting the hidden activation features with the sizes of 802816; 401408; 200704; 100352; 2048; 50; and 50, respectively. These selected features are classified by considering the VNC framework, which we have explained in Figure 4. To make a fair comparison, the same hidden activation features are also extracted from ResNet50. Finally, the average performance is computed for both models (i.e., the proposed model and ResNet50 [42]), which is shown in Table 7. On the basis of the overall performance, we conclude that (1) deeper features are better for the classification task, and (2) the extracted features from the last four layers (i.e., Conv5-1, Conv6, FC layer, and Classification layer) for our modified model are more representative and discriminative in comparison with those from ResNet50.

5.3.3. Comparisons of Classification Accuracies with or without Principal Component Analysis

The discriminative nature of our modified method and the second-best method was further explored by applying principal component analysis (PCA) [75] as a post-processing step. The features of 1 × 2048 extracted from the last convolutional layer (Conv6 of Table 2) and AvgPool of ResNet50 [42] were projected to the eigenspace separately by applying PCA. A total of 2048 eigenvectors and eigenvalues were obtained. The features obtained after PCA for the testing dataset were classified by using our VNC framework. Various PCA features were selected by considering the different number of eigenvectors for performance analysis.
Figure 14 shows the performance of both models according to the number of eigenvectors. On the basis of PCA performance as shown in Table 8, we conclude that the classification performance of PCA was not as good as when using original high-dimension features (extracted from the last convolutional layer Conv6) directly. This shows that the features extracted by our modified model are already diverse. In addition, we can find that the overall PCA performance of our modified method was also high in comparison with ResNet50, as shown in Table 8.

5.3.4. Performance Comparison with Handcrafted Feature-Based Methods

The performance of our modified deep residual method was also compared with that of conventional handcrafted feature-based methods. For a fair comparison, the same dataset was used for all these selected methods. Two known handcrafted feature extraction methods, called LBP [76] and the histogram of oriented gradients (HoG) [77], were considered for feature extraction. Finally, these extracted features were classified by using four different classifiers (i.e., adaptive boosting (AdaBoostM2), multiclass support vector machine (multi-SVM), random forest (RF), and k-nearest neighbor (KNN)) for both feature extraction methods. The classification performance of these selected feature extractions and classification methods is given in Table 9.
It is evident that the proposed deep-CNN-based classification method also outperformed the various handcrafted feature-based methods. There was a significant performance difference between our proposed model (i.e., 81.51%; 82.42%; 83.15%; 81.71%) and the best handcrafted feature-based method (i.e., HoG-KNN shows 70.84%; 70.98%; 71.69%; 70.28%) in terms of classification accuracy, F1.score, mAP, and mAR, respectively.

5.3.5. Closed-World vs. Open-World vs. Mixed-World Configurations

The performance of different classification models (i.e., CNN-based, or handcrafted feature-based) can show a significant performance disparity in different configuration modes. Therefore, it is important to analyze the performance of a model in all possible working scenarios. In this way, the true discriminative nature of a model can be evaluated. Any classification model can be expected to work in the following three configuration modes: closed-world, open-world, and mixed-world. Most of the previous studies [10,28,29,30,31,32,33,34,35,36] have been conducted in a closed-world configuration, which always shows the best performance in comparison to the other modes. The reason for this is that the closed-world configuration is subject to the constraint that all the image categories to be classified in the deployment phase are already known and used in the training phase (the classes in training are same as those used in testing). On the other hand, the open-world configuration is more challenging and is often used in the real environment. In this configuration, the image categories are enrolled during the deployment phase rather than during the training phase, after which classification is performed (the classes in training are different from those in testing). This configuration mode shows the scalable nature of a classification model because the number of image categories can be increased in the deployment phase. Finally, the mixed-world mode includes both open-world and closed-world configurations. In this configuration mode, the categories to be classified in the deployment process may be used during the training phase, and the unseen categories in the training phase can also be enrolled in the deployment phase. In this performance analysis part, the scalable nature of our modified model was explored for all possible configuration modes. Two-fold cross-validation was also performed for both open-world and mixed-world configurations. For each configuration mode, the training and testing dataset was divided in a different way. In the closed-world configuration, the half dataset of C 1 ,   C 2 ,   C 3 , , C 50 was used in training and the remaining half in testing. On the other hand, in the open-world configuration, the testing dataset was unseen in the training phase, and thus the dataset was divided in 50% training as C 1 ,   C 2 ,   C 3 , , C 25 and in 50% testing as C 26 ,   C 27 ,   C 28 , , C 50 . Finally, in the mixed-world configuration, the 50% dataset of C 1 ,   C 2 ,   C 3 , , C 40 was used in training, and the remaining dataset such as half of C 1 ,   C 2 ,   C 3 , , C 40 (i.e., similar to closed-world splitting) and full C 41 ,   C 42 ,   C 43 , , C 50 (i.e., similar to open-world splitting) was used in testing.
Table 10 presents the experimental results of our modified model and the second-best method (i.e., ResNet50 [42]) for all these configuration modes. The closed-world and mixed-world configuration results in Table 10 reveal that our modified model outperforms ResNet50. It should also be noted that our model showed the best performance in the open-world configuration, which was more challenging than the mixed-world and closed-world configuration modes. On the other hand, the performance of ResNet50 in the open-world configuration is lower than that in the other configuration modes, which shows the low performance of ResNet50 in real-world situations.

6. Discussion

In general, the efficient image classification is the key part of any CBIR system. In recent few years, the deep learning-based algorithms have shown significant performance gain in image classification tasks. In this proposed work, our main goal is to utilize the strength of deep learning in medical image classification for the CBMIR system. For this purpose, we analyze the performance of different state-of-the-art deep learning models in medical image classification task. In this way, we proposed an enhanced version of existing deep learning model (i.e., ResNet50) which shows the best classification performance in comparison with other models. Finally, based on our enhanced deep learning model, a class-prediction-based CBMIR system is proposed for medical image retrieval as shown in Figure 15.
In this proposed class-prediction-based CBMIR, image retrieval is performed based on class prediction rather than exploring the whole dataset without class prediction. In our proposed class prediction-based retrieval, the key step is to predict the actual class label for the given query image by measuring the similarity score of query image feature vector with the class mean features. In this way, a class label is predicted for the given query image by using the similarity score. Finally, the image retrieval is done by exploring the desired image in predicted class as shown in Figure 15. On the other hand, in without class-prediction-based retrieval, the whole dataset is being explored for a given input query image which is more time taking. A performance comparison is made for both retrieval methods (i.e., with class prediction and without class prediction) by using our modified model and the second-best model (i.e., ResNet50 [42]). It can be observed from Table 11 that class-prediction-based retrieval for our method shows better performance.
It should be noted that our class-prediction-based retrieval method significantly reduces the retrieval time also. The total retrieval time for both methods can be calculated as:
τ with   class   pred = τ f . e + τ f . c ( K + n )
τ without   class   pred = τ f . e + τ f . c ( Kn ) ,
where τ with   class   pred and τ without   class   pred present the retrieval time for class prediction and without class prediction, respectively. τ f . e is the feature extraction time for the input query image, and τ f . c presents the feature comparison time for two feature vectors (i.e., those extracted from the query image and the database image). The entire features database comprises K classes including a total of n feature vectors in each class. From Equations (5) and (6), it can be concluded that the total feature comparison time in the case of class prediction is approximately K times lower than without class-prediction-based retrieval, as K < < n . On the other hand, the total feature extraction time (i.e., τ f . e ) remains the same in both cases. To check the validity of Equations (5) and (6) for our proposed class-prediction-based retrieval framework, the total feature extraction and comparison time is measured in both cases. The average feature extraction time τ f . e for a single query image is obtained as 955   ms in both cases. The total feature comparison times in the case of class prediction and without class prediction are 15.4   ms and 824   ms , respectively. The total feature comparison time in the case of class prediction is approximately 53.5 times lower than that without class-prediction-based retrieval. Finally, we obtain τ with   class   pred = 971   ms and τ without   class   pred = 1779   ms , which shows that the overall performance of our proposed class-prediction-based retrieval system is much better.
Figure 16 presents a few examples of correctly retrieved images obtained with our proposed method for different input query images. It can be observed in Figure 16 that the retrieved images have varying illumination, contrast, and high intra-class variance. Despite this challenging nature of the dataset, our method still outperforms with 100% retrieval performance for the selected query images. This shows that our method can be robust to the high intra-class variance of a dataset with the significant performance gain.
A few classes in our collected dataset exhibit low retrieval performance, as shown in Figure 17. The main reason for this performance degradation is the high inter-class similarity among these classes. In Figure 17a, the given query image belongs to the bladder CT scan class, but in the retrieval results, some samples have been misclassified as kidney, renal, and stomach CT scans. This misclassification occurred due to the significant structural overlapping of these three classes, which can also be observed visually in Figure 17. Figure 17b also visualizes the similar structural overlapping of the cervix, bladder, and pancreas MRI scans. However, such misclassification cases can be resolved by adding a feedback mechanism in the proposed retrieval framework. This feedback mechanism will allow the user to explore the given query image in other relevant classes in case of misclassification.

7. Conclusions

In this paper, a medical image classification framework is proposed for retrieving heterogeneous medical images by utilizing recent deep learning techniques. The proposed deep-learning-based framework bridges the semantic gap by exploring the discriminative features (i.e., all low-level and high-level features) directly from the images. These extracted features are used to perform class-prediction-based image retrieval tasks. The performance of the proposed system is evaluated on various multimodal databases for all possible real-world configuration modes (i.e., closed-world, open-world, and mixed-world). Our proposed system significantly outperforms the existing retrieval systems used in the medical domain. Moreover, our enhanced ResNet solved the problem of high intra-class variance and inter-class similarity in a medical database, and it improved the classification accuracies. The retrieval performance of the proposed system demonstrates its applicability to various clinical situations, education, and research. Our trained model and image indices of experimental images have been made publicly available to permit other researchers to make performance comparisons. In previous researches on our research topic [28,66], they did not use a validation set for determining the optimal CNN model, but used only training and testing sets like our experiments. In order to maintain fair experimental conditions and comparisons with [28,66] of Table 5, we used only training and testing sets.
In future work, we would compare the accuracies based on the optimal model selection using the additional validation set with our accuracies. We would also study the method which can deal with the case of more classes than 50 classes. In addition, we intended to implement a video-based MIRS for exploring moving sequences. Also, we would further optimize the network by reducing the number of layers and other parameters to make it more efficient.

Author Contributions

M.O. and K.R.P. designed the overall system. In addition, they wrote and revised the paper. M.A. and J.C. helped to design the comparative analysis and experiments.

Acknowledgments

This research was supported by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2018R1D1A1B07041921), by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2017R1D1A1B03028417), and by the National Research Foundation of Korea (NRF) grant funded by the Korea government (Ministry of Science and ICT) (NRF-2017R1C1B5074062).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cheng, C.-H.; Liu, W.-X. Identifying degenerative brain disease using rough set classifier based on wavelet packet method. J. Clin. Med. 2018, 7, 124. [Google Scholar] [CrossRef] [PubMed]
  2. Purcaru, M.A.P.; Repanovici, A.; Nedeloiu, T. Non-invasive assessment method using thoracic-abdominal profile image acquisition and mathematical modeling with Bezier curves. J. Clin. Med. 2019, 8, 65. [Google Scholar] [CrossRef]
  3. Tang, J.; Again, S.; Thompson, I. Guest editorial: Computer-aided detection or diagnosis (CAD) systems. IEEE Syst. J. 2014, 8, 907–909. [Google Scholar] [CrossRef]
  4. Miranda, E.; Aryuni, M.; Irwansyah, E. A survey of medical image classification techniques. In Proceedings of the IEEE International Conference on Information Management and Technology, Bandung, Indonesia, 16–18 November 2016; pp. 56–61. [Google Scholar]
  5. Wan, J.; Wang, D.; Hoi, S.C.H.; Wu, P.; Zhu, J.; Zhang, Y.; Li, J. Deep learning for content-based image retrieval: A comprehensive study. In Proceedings of the 22nd ACM International Conference on Multimedia, Orlando, FL, USA, 3–7 November 2014; pp. 157–166. [Google Scholar]
  6. Bengio, Y.; Courville, A.; Vincent, P. Unsupervised feature learning and deep learning: A review and new perspectives. arXiv 2012, arXiv:1206.5538v1. [Google Scholar]
  7. Yu, D.; Deng, L. Deep learning and its applications to signal and information processing. IEEE Signal Process. Mag. 2011, 28, 145–154. [Google Scholar] [CrossRef]
  8. Moccia, S.; De Momi, E.; El Hadji, S.; Mattos, L.S. Blood vessel segmentation algorithms—Review of methods, datasets and evaluation metrics. Comput. Meth. Programs Biomed. 2018, 158, 71–91. [Google Scholar] [CrossRef]
  9. Litjens, G.; Kooi, T.; Bejnordi, B.E.; Setio, A.A.A.; Ciompi, F.; Ghafoorian, M.; van der Laak, J.A.W.M.; van Ginneken, B.; Sánchez, C.I. A survey on deep learning in medical image analysis. Med. Image Anal. 2017, 42, 60–88. [Google Scholar] [CrossRef] [PubMed]
  10. Tajbakhsh, N.; Shin, J.Y.; Gurudu, S.R.; Hurst, R.T.; Kendall, C.B.; Gotway, M.B.; Liang, J. Convolutional neural networks for medical image analysis: Full training or fine tuning? IEEE Trans. Med. Imaging 2016, 35, 1299–1312. [Google Scholar] [CrossRef]
  11. Choplin, R.H.; Boehme, J.M., II; Maynard, C.D. Picture archiving and communication systems: An overview. Radiographics 1992, 12, 127–129. [Google Scholar] [CrossRef] [PubMed]
  12. Graham, R.N.J.; Perriss, R.W.; Scarsbrook, A.F. DICOM demystified: A review of digital file formats and their use in radiological practice. Clin. Radiol. 2005, 60, 1133–1140. [Google Scholar] [CrossRef]
  13. Avni, U.; Greenspan, H.; Konen, E.; Sharon, M.; Goldberger, J. X-ray categorization and retrieval on the organ and pathology level, using patch-based visual words. IEEE Trans. Med. Imaging 2011, 30, 733–746. [Google Scholar] [CrossRef] [PubMed]
  14. Orphanoudakis, S.C.; Chronaki, C.; Kostomanolakis, S. I2C: A system for the indexing, storage, and retrieval of medical images by content. Med. Inform. 1994, 19, 109–122. [Google Scholar] [CrossRef]
  15. Chu, W.W.; Hsu, C.-C.; Cardenas, A.F.; Taira, R.K. Knowledge-based image retrieval with spatial and temporal constructs. IEEE Trans. Knowl. Data Eng. 1998, 10, 872–888. [Google Scholar] [CrossRef]
  16. El-Kwae, E.A.; Xu, H.; Kabuka, M.R. Content-based retrieval in picture archiving and communication systems. J. Digit. Imaging 2000, 13, 70–81. [Google Scholar] [CrossRef] [PubMed]
  17. Muller, H.; Rosset, A.; Garcia, A.; Vallee, J.-P.; Geissbuhler, A. Benefits of content-based visual data access in radiology. Radiographics 2005, 25, 849–858. [Google Scholar] [CrossRef]
  18. Rahman, M.M.; Bhattacharya, P.; Desai, B.C. A framework for medical image retrieval using machine learning and statistical similarity matching techniques with relevance feedback. IEEE Trans. Inf. Technol. Biomed. 2007, 11, 58–69. [Google Scholar] [CrossRef]
  19. Rahman, M.M.; Bhattacharya, P.; Desai, B.C. A unified image retrieval framework on local visual and semantic concept-based feature spaces. J. Vis. Commun. Image Represent. 2009, 20, 450–462. [Google Scholar] [CrossRef]
  20. Rahman, M.M.; You, D.; Simpson, M.S.; Antani, S.K.; Demner-Fushman, D.; Thoma, G.R. Multimodal biomedical image retrieval using hierarchical classification and modality fusion. Int. J. Multimed. Inf. Retr. 2013, 2, 159–173. [Google Scholar] [CrossRef]
  21. Sudhakar, M.S.; Bagan, K.B. An effective biomedical image retrieval framework in a fuzzy feature space employing phase congruency and GeoSOM. Appl. Soft Comput. 2014, 22, 492–503. [Google Scholar] [CrossRef]
  22. Jyothi, B.; MadhaveeLatha, Y.; Mohan, P.G.K. An effective multiple visual features for content based medical image retrieval. In Proceedings of the IEEE 9th International Conference on Intelligent Systems and Control, Coimbatore, India, 9–10 January 2015; pp. 1–5. [Google Scholar]
  23. Ramamurthy, B.; Chandran, K.R. CBMIR: Content based medical image retrieval using multilevel hybrid approach. Int. J. Comput. Commun. Control 2015, 10, 382–389. [Google Scholar] [CrossRef]
  24. Bedo, M.V.N.; dos Santos, D.P.; Ponciano-Silva, M.; de Azevedo-Marques, P.M.; de Carvalho, A.P.D.L.F.; Traina, C., Jr. Endowing a content-based medical image retrieval system with perceptual similarity using ensemble strategy. J. Digit. Imaging 2016, 29, 22–37. [Google Scholar] [CrossRef]
  25. Malviya, N.; Choudhary, N.; Jain, K. Content based medical image retrieval and clustering based segmentation to diagnose lung cancer. Adv. Comput. Sci. Technol. 2017, 10, 1577–1594. [Google Scholar]
  26. Kumar, M.; Singh, K.M. Content based medical image retrieval system (CBMIRS) to diagnose hepatobiliary images. In Proceedings of the International Conference on Next Generation Computing Technologies, Dehradun, India, 30–31 October 2017; pp. 663–676. [Google Scholar]
  27. Kumar, K.K.; Gopal, T.V. A novel approach to self order feature reweighting in CBIR to reduce Semantic gap using relevance feedback. In Proceedings of the IEEE International Conference on Circuit, Power and Computing Technologies, Nagercoil, India, 20–21 March 2014; pp. 1437–1442. [Google Scholar]
  28. Qayyum, A.; Anwar, S.M.; Awais, M.; Majid, M. Medical image retrieval using deep convolutional neural network. Neurocomputing 2017, 266, 8–20. [Google Scholar] [CrossRef]
  29. Chowdhury, M.; Bulò, S.R.; Moreno, R.; Kundu, M.K.; Smedby, Ö. An efficient radiographic image retrieval system using convolutional neural network. In Proceedings of the IEEE 23rd International Conference on Patteren Recognition, Cancun, Mexico, 4–8 December 2016; pp. 3134–3139. [Google Scholar]
  30. Anthimopoulos, M.; Christodoulidis, S.; Ebner, L.; Christe, A.; Mougiakakou, S. Lung pattern classification for interstitial lung diseases using a deep convolutional neural network. IEEE Trans. Med. Imaging 2016, 35, 1207–1216. [Google Scholar] [CrossRef]
  31. Van Tulder, G.; de Bruijne, M. Combining generative and discriminative representation learning for lung CT analysis with convolutional restricted Boltzmann machines. IEEE Trans. Med. Imaging 2016, 35, 1262–1272. [Google Scholar] [CrossRef]
  32. Ciompi, F.; de Hoop, B.; van Riel, S.J.; Chung, K.; Scholten, E.T.; Oudkerk, M.; de Jong, P.A.; Prokop, M.; van Ginneken, B. Automatic classification of pulmonary peri-fissural nodules in computed tomography using an ensemble of 2D views and a convolutional neural network out-of-the-box. Med. Image Anal. 2015, 26, 195–202. [Google Scholar] [CrossRef]
  33. Yan, Z.; Zhan, Y.; Peng, Z.; Liao, S.; Shinagawa, Y.; Zhang, S.; Metaxas, D.N.; Zhou, X.S. Multi-instance deep learning: Discover discriminative local anatomies for bodypart recognition. IEEE Trans. Med. Imaging 2016, 35, 1332–1343. [Google Scholar] [CrossRef] [PubMed]
  34. Dou, Q.; Chen, H.; Yu, L.; Zhao, L.; Qin, J.; Wang, D.; Mok, V.C.; Shi, L.; Heng, P.-A. Automatic detection of cerebral microbleeds from MR images via 3D convolutional neural networks. IEEE Trans. Med. Imaging 2016, 35, 1182–1195. [Google Scholar] [CrossRef]
  35. Van Grinsven, M.J.J.P.; van Ginneken, B.; Hoyng, C.B.; Theelen, T.; Sánchez, C.I. Fast convolutional neural network training using selective data sampling: Application to hemorrhage detection in color fundus images. IEEE Trans. Med. Imaging 2016, 35, 1273–1284. [Google Scholar] [CrossRef] [PubMed]
  36. Setio, A.A.A.; Ciompi, F.; Litjens, G.; Gerke, P.; Jacobs, C.; van Riel, S.J.; Wille, M.M.W.; Naqibullah, M.; Sánchez, C.I.; van Ginneken, B. Pulmonary nodule detection in CT images: False positive reduction using multi-view convolutional networks. IEEE Trans. Med. Imaging 2016, 35, 1160–1169. [Google Scholar] [CrossRef]
  37. Shin, H.-C.; Roth, H.R.; Gao, M.; Lu, L.; Xu, Z.; Nogues, I.; Yao, J.; Mollura, D.; Summers, R.M. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans. Med. Imaging 2016, 35, 1285–1298. [Google Scholar] [CrossRef]
  38. Jiang, F.; Jiang, Y.; Zhi, H.; Dong, Y.; Li, H.; Ma, S.; Wang, Y.; Dong, Q.; Shen, H.; Wang, Y. Artificial intelligence in healthcare: Past, present and future. Stroke Vasc. Neurol. 2017, 2, 230–243. [Google Scholar] [CrossRef]
  39. Gupta, M.R.; Bengio, S.; Weston, J. Training highly multiclass classifiers. J. Mach. Learn. Res. 2014, 15, 1461–1492. [Google Scholar]
  40. Brucker, F.; Benites, F.; Sapozhnikova, E. Multi-label classification and extracting predicted class hierarchies. Pattern Recognit. 2011, 44, 724–738. [Google Scholar] [CrossRef]
  41. Silva-Palacios, D.; Ferri, C.; Ramirez-Quintana, M.J. Improving performance of multiclass classification by inducing class hierarchies. In Proceedings of the International Conference on Computational Science, Zurich, Switzerland, 12–14 June 2017; pp. 1692–1701. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  43. Dongguk CNN Model and Image Indices of Open Databases for CBMIR. Available online: http://dm.dgu.edu/link.html (accessed on 15 February 2019).
  44. He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 630–645. [Google Scholar]
  45. Heaton, J. Artificial Intelligence for Humans; Deep learning and neural networks; Heaton Research Inc.: St. Louis, MO, USA, 2015; Volume 3. [Google Scholar]
  46. MICCAI Grand Challenges. Available online: https://grand-challenge.org/challenges/ (accessed on 29 March 2019).
  47. Chest X-rays Database. Available online: https://nihcc.app.box.com/v/ChestXray-NIHCC (accessed on 1 February 2019).
  48. Decenciere, E.; Zhang, X.; Cazuguel, G.; Lay, B.; Cochener, B.; Trone, C.; Gain, P.; Ordonez-Varela, J.-R.; Massin, P.; Erginay, A.; et al. Feedback on a publicly distributed image database: The Messidor database. Image Anal. Stereol. 2014, 33, 231–234. [Google Scholar] [CrossRef]
  49. Suckling, J.; Parker, J.; Dance, D.R.; Astley, S.; Hutt, I.; Boggis, C.; Ricketts, I.; Stamatakis, E.; Cerneaz, N.; Kok, S.L.; et al. The mammographic image analysis society digital mammogram database. In Proceedings of the 2nd International Workshop on Digital Mammography, York, UK, 10–12 July 1994; pp. 375–378. [Google Scholar]
  50. Brain Tumor Database. Available online: https://figshare.com/articles/brain_tumor_dataset /1512427 (accessed on 1 February 2019).
  51. Bones X-rays Database. Available online: https://sites.google.com/site/mianalysis16/ (accessed on 1 February 2019).
  52. Neck Nerve Structure Database. Available online: https://www.kaggle.com/c/ultrasound-nerve-segmentation/data (accessed on 1 February 2019).
  53. Radau, P.; Lu, Y.; Connelly, K.; Paul, G.; Dick, A.J.; Wright, G.A. Evaluation Framework for Algorithms Segmenting Short Axis Cardiac MRI. Available online: https://www.midasjournal.org/browse/publication/658 (accessed on 5 April 2019).
  54. Visible Human Project CT Datasets. Available online: https://mri.radiology.uiowa.edu/visible_human_datasets.html (accessed on 1 February 2019).
  55. Baby Ultrasound Videos. Available online: https://youtu.be/SrUoXkKoREE (accessed on 1 February 2019).
  56. Clark, K.; Vendt, B.; Smith, K.; Freymann, J.; Kirby, J.; Koppel, P.; Moore, S.; Phillips, S.; Maffitt, D.; Pringle, M.; et al. The cancer imaging archive (TCIA): Maintaining and operating a public information repository. J. Digit. Imaging 2013, 26, 1045–1057. [Google Scholar] [CrossRef] [PubMed]
  57. Endoscopy Videos. Available online: http://www.gastrolab.net/ni.htm (accessed on 1 February 2019).
  58. Skin’s Diseases Database. Available online: https://www.dermnetnz.org/image-licence/#use (accessed on 1 February 2019).
  59. Wong, S.C.; Gatt, A.; Stamatescu, V.; McDonnell, M.D. Understanding data augmentation for classification: When to warp? In Proceedings of the IEEE International Conference on Digital Image Computing: Techniques and Applications, Gold Coast, Australia, 30 November–2 December 2016; pp. 1–6. [Google Scholar]
  60. Intel® Core™ i7-3770K Processor. Available online: https://ark.intel.com/content/www/us/en/ark/products/65523/intel-core-i7-3770k-processor-8m-cache-up-to-3-90-ghz.html (accessed on 1 February 2019).
  61. GeForce GTX 1070. Available online: https://www.geforce.com/hardware/desktop-gpus/geforce-gtx-1070/specifications (accessed on 1 February 2019).
  62. MATLAB R2018b. Available online: https://ch.mathworks.com/products/new_products/latest_features.html (accessed on 1 February 2019).
  63. Bottou, L. Stochastic gradient descent tricks. In Neural Networks: Tricks of the Trade; Springer: Berlin, Germany, 2012; pp. 421–436. [Google Scholar]
  64. Training Options. Available online: http://kr.mathworks.com/help/nnet/ref/trainingoptions.html (accessed on 1 February 2019).
  65. Hossin, M.; Sulaiman, M.N. A review on evaluation metrics for data classification evaluations. Int. J. Data Min. Knowl. Manag. Process 2015, 5, 1–11. [Google Scholar]
  66. Takiyama, H.; Ozawa, T.; Ishihara, S.; Fujishiro, M.; Shichijo, S.; Nomura, S.; Miura, M.; Tada, T. Automatic anatomical classification of esophagogastroduodenoscopy images using deep convolutional neural networks. Sci. Rep. 2018, 8, 7497. [Google Scholar] [CrossRef] [PubMed]
  67. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <0.5 MB model size. arXiv 2016, arXiv:1602.07360v4. [Google Scholar]
  68. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
  69. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  70. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  71. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A.A. Inception-v4, Inception-ResNet and the impact of residual connections on learning. In Proceedings of the 31st AAAI Conference on Artificial Intelligence, San Francisco, CA, USA, 4–9 February 2017; pp. 4278–4284. [Google Scholar]
  72. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. ImageNet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
  73. Raychaudhuri, S. Introduction to Monte Carlo simulation. In Proceedings of the IEEE Winter Simulation Conference, Miami, FL, USA, 7–10 December 2008; pp. 91–100. [Google Scholar]
  74. Student’s t-Test. Available online: https://en.wikipedia.org/wiki/Student%27s_t-test (accessed on 1 February 2019).
  75. Ilin, A.; Raiko, T. Practical approaches to principal component analysis in the presence of missing values. J. Mach. Learn. Res. 2010, 11, 1957–2000. [Google Scholar]
  76. Subrahmanyam, M.; Maheshwari, R.P.; Balasubramanian, R. Local maximum edge binary patterns: A new descriptor for image retrieval and object tracking. Signal Process. 2012, 92, 1467–1479. [Google Scholar] [CrossRef]
  77. Velmurugan, K.; Baboo, L.D.S.S. Image retrieval using Harris corners and histogram of oriented gradients. Int. J. Comput. Appl. 2011, 24, 6–10. [Google Scholar]
Figure 1. Overall procedure of the proposed method for classification.
Figure 1. Overall procedure of the proposed method for classification.
Jcm 08 00462 g001
Figure 2. Overview of the proposed deep convolutional neural network (CNN) architecture used for feature extraction and classification.
Figure 2. Overview of the proposed deep convolutional neural network (CNN) architecture used for feature extraction and classification.
Jcm 08 00462 g002
Figure 3. The residual building block of our modified ResNet50 with (a) identity-mapping-based residual unit, and (b) 1 × 1 convolutional-mapping-based residual unit.
Figure 3. The residual building block of our modified ResNet50 with (a) identity-mapping-based residual unit, and (b) 1 × 1 convolutional-mapping-based residual unit.
Jcm 08 00462 g003
Figure 4. Proposed deep-feature-based variable node classification (VNC) framework.
Figure 4. Proposed deep-feature-based variable node classification (VNC) framework.
Jcm 08 00462 g004
Figure 5. Examples from each class of total 50 classes. (a-1) Hip computed tomography (CT); (a-2) knee CT; (a-3) ultrasound; (a-4) cervix magnetic resonance imaging (MRI); (a-5) 4D lung CT; (a-6) head, neck MRI; (a-7) head, neck CT; (a-8) lung CT; (a-9) bladder MRI; (a-10) bladder CT; (b-1) chest X-Ray; (b-2) urography CT; (b-3) breast MRI; (b-4) prostate MRI; (b-5) uterus CT; (b-6) rectum MRI; (b-7) ovary CT; (b-8) liver MRI; (b-9) liver CT; (b-10) kidney, renal MRI; (c-1) kidney, renal CT; (c-2) retina; (c-3) CT topogram; (c-4) pancreas MRI; (c-5) pancreas CT; (c-6) stomach CT; (c-7) colonography CT; (c-8) esophagus CT; (c-9) malignant tumors; (c-10) sigmoid colon; (d-1) rectum; (d-2) colon; (d-3) breast mammogram; (d-4) caecum; (d-5) duodenal bulb; (d-6) normal esophagus; (d-7) benign tumors; (d-8) Crohn’s disease; (d-9) gastric fundus; (d-10) ulcerative colitis; (e-1) upper endoscopy; (e-2) facial acne; (e-3) hand, foot allergies; (e-4) brain MRI; (e-5) legs, arms allergies; (e-6) bones X-rays; (e-7) neck nerves; (e-8) cardiac MRI; (e-9) shoulder CT; (e-10) ankle CT.
Figure 5. Examples from each class of total 50 classes. (a-1) Hip computed tomography (CT); (a-2) knee CT; (a-3) ultrasound; (a-4) cervix magnetic resonance imaging (MRI); (a-5) 4D lung CT; (a-6) head, neck MRI; (a-7) head, neck CT; (a-8) lung CT; (a-9) bladder MRI; (a-10) bladder CT; (b-1) chest X-Ray; (b-2) urography CT; (b-3) breast MRI; (b-4) prostate MRI; (b-5) uterus CT; (b-6) rectum MRI; (b-7) ovary CT; (b-8) liver MRI; (b-9) liver CT; (b-10) kidney, renal MRI; (c-1) kidney, renal CT; (c-2) retina; (c-3) CT topogram; (c-4) pancreas MRI; (c-5) pancreas CT; (c-6) stomach CT; (c-7) colonography CT; (c-8) esophagus CT; (c-9) malignant tumors; (c-10) sigmoid colon; (d-1) rectum; (d-2) colon; (d-3) breast mammogram; (d-4) caecum; (d-5) duodenal bulb; (d-6) normal esophagus; (d-7) benign tumors; (d-8) Crohn’s disease; (d-9) gastric fundus; (d-10) ulcerative colitis; (e-1) upper endoscopy; (e-2) facial acne; (e-3) hand, foot allergies; (e-4) brain MRI; (e-5) legs, arms allergies; (e-6) bones X-rays; (e-7) neck nerves; (e-8) cardiac MRI; (e-9) shoulder CT; (e-10) ankle CT.
Jcm 08 00462 g005
Figure 6. Examples of experimental data of Figure 5 according to the anatomical district (row) and imaging modality (column).
Figure 6. Examples of experimental data of Figure 5 according to the anatomical district (row) and imaging modality (column).
Jcm 08 00462 g006aJcm 08 00462 g006bJcm 08 00462 g006c
Figure 7. Selected example images for showing high intra-class variance; (a) chest X-ray in Figure 5b-1; (b) urography CT in Figure 5b-2; (c) breast MRI in Figure 5b-3; (d) uterus CT in Figure 5b-5; (e) CT topogram in Figure 5c-3; (f) rectum in Figure 5d-1; (g) colon in Figure 5d-2; (h) bones X-rays in Figure 5e-6.
Figure 7. Selected example images for showing high intra-class variance; (a) chest X-ray in Figure 5b-1; (b) urography CT in Figure 5b-2; (c) breast MRI in Figure 5b-3; (d) uterus CT in Figure 5b-5; (e) CT topogram in Figure 5c-3; (f) rectum in Figure 5d-1; (g) colon in Figure 5d-2; (h) bones X-rays in Figure 5e-6.
Jcm 08 00462 g007
Figure 8. Selected example images for showing high inter-class similarity; (a,e) hip CT in Figure 5a-1 and knee CT in Figure 5a-2, respectively; (b,f) cervix MRI in Figure 5a-4 and bladder MRI in Figure 5a-9, respectively; (c,g) benign tumors in Figure 5d-7 and Rectum in Figure 5d-1, respectively; (d,h) urography CT in Figure 5b-2 and uterus CT in Figure 5b-5, respectively.
Figure 8. Selected example images for showing high inter-class similarity; (a,e) hip CT in Figure 5a-1 and knee CT in Figure 5a-2, respectively; (b,f) cervix MRI in Figure 5a-4 and bladder MRI in Figure 5a-9, respectively; (c,g) benign tumors in Figure 5d-7 and Rectum in Figure 5d-1, respectively; (d,h) urography CT in Figure 5b-2 and uterus CT in Figure 5b-5, respectively.
Jcm 08 00462 g008aJcm 08 00462 g008b
Figure 9. Plot for training loss and accuracy: (a) 1-fold cross-validation; (b) 2-fold cross-validation.
Figure 9. Plot for training loss and accuracy: (a) 1-fold cross-validation; (b) 2-fold cross-validation.
Jcm 08 00462 g009
Figure 10. Visualization of learned filters from the first convolutional layer in the case of (a) training from scratch with random initialization, and (b) fine-tuning with transfer learning.
Figure 10. Visualization of learned filters from the first convolutional layer in the case of (a) training from scratch with random initialization, and (b) fine-tuning with transfer learning.
Jcm 08 00462 g010
Figure 11. Sensitivity analysis plot of our proposed and various baseline models in terms of (a) average accuracy and (b) average F1.score (unit: %).
Figure 11. Sensitivity analysis plot of our proposed and various baseline models in terms of (a) average accuracy and (b) average F1.score (unit: %).
Jcm 08 00462 g011
Figure 12. T-test plot of our proposed and the second-best model (ResNet50 [42]) in terms of (a) average accuracy and (b) average F1-Score (unit: %).
Figure 12. T-test plot of our proposed and the second-best model (ResNet50 [42]) in terms of (a) average accuracy and (b) average F1-Score (unit: %).
Jcm 08 00462 g012
Figure 13. Confusion matrix of the proposed method. The entry in the i -th row and j -th column corresponds to the percentage of samples from class i that were classified as class j .
Figure 13. Confusion matrix of the proposed method. The entry in the i -th row and j -th column corresponds to the percentage of samples from class i that were classified as class j .
Jcm 08 00462 g013
Figure 14. Principal component analysis (PCA)-based performance analysis. (a) ResNet50 [42]: features selected from the last average pooling layer; (b) proposed model: features selected from the last convolutional layer.
Figure 14. Principal component analysis (PCA)-based performance analysis. (a) ResNet50 [42]: features selected from the last average pooling layer; (b) proposed model: features selected from the last convolutional layer.
Jcm 08 00462 g014
Figure 15. The class-prediction-based content-based medical image retrieval (CBMIR) system by using our proposed deep CNN model.
Figure 15. The class-prediction-based content-based medical image retrieval (CBMIR) system by using our proposed deep CNN model.
Jcm 08 00462 g015
Figure 16. Examples of good retrieval performance of our proposed system for input query image. (a) colonography CT; (b) liver CT; (c) upper endoscopy; (d) facial acne.
Figure 16. Examples of good retrieval performance of our proposed system for input query image. (a) colonography CT; (b) liver CT; (c) upper endoscopy; (d) facial acne.
Jcm 08 00462 g016aJcm 08 00462 g016b
Figure 17. Class-prediction-based false retrieval performance of our proposed system. (a) Bladder CT scan as a query image; (b) cervix MRI scan as the query image.
Figure 17. Class-prediction-based false retrieval performance of our proposed system. (a) Bladder CT scan as a query image; (b) cervix MRI scan as the query image.
Jcm 08 00462 g017
Table 1. Comparison of our proposed and existing methods for medical image classification and retrieval.
Table 1. Comparison of our proposed and existing methods for medical image classification and retrieval.
Imaging ModalitiesMethodNumber of ClassesStrengthWeakness
Single modalityCTPre-trained CNN [32]2Classification performance reaches that of a human observerClassify only lung cancer CT scan images rather than multiclass images
X-rayCNN + edge histogram features are selected [29]31High classification performanceLimited dataset (i.e., 1550 images) related to 31 classes (i.e., 50 images in each class) and only 10 images per class are selected for calculating system performance
CTDeep CNN model [30]7High CAD sensitivity performance with less computation timeOnly classify infected and non-infected lung CT scans
CTTwo-stage multiple instance CNN [33]12High classification accuracyLimited dataset and number of classes
MRI3D CNN-based discrimination model [34]2High sensitivityOnly classify infected and non-infected brain MRI scans (only 2 classes)
Fundus cameraCNN + selective sampling (SeS) [35]2High average classification accuracyUses the reference guide from a single expert
CTRestricted Boltzmann machine (RBM) [31]5High average classification accuracySuitable for smaller representations learning with smaller filters or hidden nodes
CTMultiview convolutional network (ConvNets) [36]2False positive error is reducedThe CAD sensitivity performance should be enhanced
Multiple modalitiesCT, MRI, fundus camera, PET, OPTContent-based medical image retrieval system (CBMIR) by using CNN [28]24High classification accuracy- A limited number of experimental images
- Performance was measured only by closed-world configuration
CT, MRI, fundus camera, PET, OPT, X-ray, ultrasound, endoscopy, visible light cameraProposed method50- High classification performance for multiple modalities data.
- Number of classes is much larger than that in the previous work
Using deeper CNN requires more training time
Table 2. Layer configuration details of our deep residual convolutional neural network (CNN) architecture.
Table 2. Layer configuration details of our deep residual convolutional neural network (CNN) architecture.
Layer NameFeature Map SizeNumber of FiltersKernel SizeStrideNumber of PaddingNumber of Iterations
Image input layer224 × 224 × 3
Conv1112 × 112 × 64647 × 7 × 3231
Max pool56 × 56 × 6413 × 3201
Conv2Conv2-1
(1 × 1 Convolutional Mapping)
56 × 56 × 64641 × 1 × 64101
56 × 56 × 64643 × 3 × 6411
56 × 56 × 2562561 × 1 × 6410
56 × 56 × 2562561 × 1 × 6410
Conv2-2–Conv2-3
(Identity Mapping)
56 × 56 × 64641 × 1 × 256102
56 × 56 × 64643 × 3 × 6411
56 × 56 × 2562561 × 1 × 6410
Conv3Conv3-1
(1 × 1 Convolutional Mapping)
28 × 28 × 1281281 × 1 × 256201
28 × 28 × 1281283 × 3 × 12811
28 × 28 × 5125121 × 1 × 12810
28 × 28 × 5125121 × 1 × 25620
Conv3-2–Conv3-4
(Identity Mapping)
28 × 28 × 1281281 × 1 × 512103
28 × 28 × 1281283 × 3 × 12811
28 × 28 × 5125121 × 1 × 12810
Conv4Conv4-1
(1 × 1 Convolutional Mapping)
14 × 14 × 2562561 × 1 × 512201
14 × 14 × 2562563 × 3 × 25611
14 × 14 × 102410241 × 1 × 25610
14 × 14 × 102410241 × 1 × 51220
Conv4-2–Conv4-6
(Identity Mapping)
14 × 14 × 2562561 × 1 × 1024105
14 × 14 × 2562563 × 3 × 25611
14 × 14 × 102410241 × 1 × 25610
Conv5Conv5-1
(1 × 1 Convolutional Mapping)
7 × 7 × 5125121 × 1 × 1024201
7 × 7 × 5125123 × 3 × 51211
7 × 7 × 204820481 × 1 × 51210
7 × 7 × 204820481 × 1 × 102420
Conv5-2–Conv5-3
(Identity Mapping)
7 × 7 × 5125121 × 1 × 2048102
7 × 7 × 5125123 × 3 × 51211
7 × 7 × 204820481 × 1 × 51210
Conv61 × 1 × 204820487 × 7 × 2048101
FC layer50 1
SoftMax50 1
Classification layer50 1
Table 3. Summary of training and testing dataset in two-fold cross-validation (unit: images).
Table 3. Summary of training and testing dataset in two-fold cross-validation (unit: images).
ConfigurationsValidationTrainingTestingTotal
OriginalAugmented
Closed-world1st fold22,732226822,73247,732
2nd fold22,732226822,73247,732
Open-world1st fold21,870313023,59448,594
2nd fold23,594140621,87046,870
Mixed-world1st fold18,435156527,02947,029
2nd fold18,435156527,02947,029
Table 4. Class imbalance details with augmented images for the classes which contain less than 500 images in training (unit: images).
Table 4. Class imbalance details with augmented images for the classes which contain less than 500 images in training (unit: images).
Caption Detail as in Figure 5Class NameClass Imbalance Details
OriginalAugmentedTotalImbalance Ratio (%)
d-3Breast mammogram16133950067.8
e-6Bones X-rays16933150066.2
e-9Shoulder CT455455009
e-10Ankle CT7542550085
a-1Hip CT40010050020
a-2Knee CT17532550065
e-2Facial acne487135002.6
e-3Hand, foot allergies23826250052.4
e-5Legs, arms allergies7242850085.6
Table 5. Classification performance of proposed and different baseline CNN models after the training from scratch (unit: %).
Table 5. Classification performance of proposed and different baseline CNN models after the training from scratch (unit: %).
CNN ModelAccuracyF1.scoremAPmAR
Fold1Fold2Avg.Fold1Fold2Avg.Fold1Fold2Avg.Fold1Fold2Avg.
AlexNet [28]71.0168.4169.7171.4367.8469.6372.4268.2170.3170.4767.4768.97
SqueezeNet [66]73.2171.4372.3274.4573.7974.1276.6475.5076.0772.3772.1672.27
VGG16 [68]77.3877.3377.3678.0178.4478.2278.8379.2979.0677.2177.6077.41
VGG19 [68]77.1077.8277.4677.9878.5378.2579.0179.1479.0876.9777.9277.45
GoogLeNet [66,69]79.9477.3778.6680.9078.1179.5182.3978.0880.2379.4778.1578.81
ResNet101 [42]81.0879.1680.1281.8180.5481.1782.8580.8781.8680.7980.2080.50
ResNet50 [42]81.2979.5480.4282.1880.6581.4183.2980.7482.0181.0980.5680.83
InceptionV3 [70]81.1779.6980.4382.2481.0281.6382.9881.2882.1381.5380.7681.14
InceptionResNetV2 [71]81.1180.0580.5882.2881.2581.7783.4681.4282.4481.1381.0981.11
Proposed81.8479.3980.6282.8480.1681.5084.4180.0782.2481.3380.2580.79
Table 6. Classification performance of proposed and different baseline CNN models in the case of transfer learning (unit: %).
Table 6. Classification performance of proposed and different baseline CNN models in the case of transfer learning (unit: %).
CNN ModelAccuracyF1.scoremAPmAR
Fold1Fold2Avg.Fold1Fold2Avg.Fold1Fold2Avg.Fold1Fold2Avg.
InceptionV3 [70]79.9979.4679.7280.8280.3880.6082.0380.2681.1479.6680.4980.07
InceptionResNetV2 [71]80.4578.7379.5982.0679.7780.9282.8879.6781.2881.2579.8880.56
ResNet50 [42]82.4879.3380.9083.1880.6281.9084.1480.9082.5282.2480.3381.28
Proposed82.6080.4281.5183.6081.2482.4285.1081.2083.1582.1581.2781.71
Table 7. Performance comparison of our proposed and the second-best model (Resnet50 [42]) on the basis of feature extraction from different layers (unit: %) (* in our modified layer, average pooling (AvgPool) is replaced by Conv6).
Table 7. Performance comparison of our proposed and the second-best model (Resnet50 [42]) on the basis of feature extraction from different layers (unit: %) (* in our modified layer, average pooling (AvgPool) is replaced by Conv6).
Layer NameFeature Dim.ResNet50 [42]Proposed
AccuracyF1.scoremAPmARAccuracyF1.scoremAPmAR
Conv2-180281657.9258.9459.3958.5257.9258.9559.4058.52
Conv3-140140861.0662.5463.2561.8961.0762.5563.2661.90
Conv4-120070468.7069.9870.6169.3668.6969.9770.6069.36
Conv5-110035273.3775.1676.0674.2874.2876.2777.5275.07
AvgPool/* Conv6204879.2580.2680.7679.7779.8981.3082.3180.34
FC layer5080.1781.2181.8080.6381.2682.4383.3381.55
Classification layer5080.9081.9082.5281.2881.5182.4283.1581.71
Table 8. Principal component analysis (PCA) performance comparisons of our proposed and the second-best CNN model (ResNet50 [42]) (unit: %).
Table 8. Principal component analysis (PCA) performance comparisons of our proposed and the second-best CNN model (ResNet50 [42]) (unit: %).
OptionResNet50 (No. of Eigenvectors = 170) [42]Proposed (No. of Eigenvectors = 160)
AccuracyF1.scoremAPmARAccuracyF1.scoremAPmAR
With PCA79.1479.9280.1479.7180.0181.3282.2480.45
Without PCA80.9081.9082.5281.2881.5182.4283.1581.71
Table 9. Comparison of classification performance of the proposed method with different handcrafted feature-based methods (unit: %).
Table 9. Comparison of classification performance of the proposed method with different handcrafted feature-based methods (unit: %).
MethodClassifierAccuracyF1.scoremAPmAR
LBP [76]AdaBoostM235.9435.9736.0235.91
Multi-SVM45.6245.4845.3845.58
RF61.3661.2861.5261.05
KNN59.7159.3159.3959.24
HOG [77]AdaBoostM241.3741.2541.9440.58
Multi-SVM65.6667.4769.5165.55
RF69.5470.0671.3268.86
KNN70.8470.9871.6970.28
Proposed81.5182.4283.1581.71
Table 10. Closed-world, open-world, and mixed-world performance comparisons of our modified model and ResNet50 [42] (unit: %).
Table 10. Closed-world, open-world, and mixed-world performance comparisons of our modified model and ResNet50 [42] (unit: %).
Configuration ModeResNet50 [42]Proposed
Accuracy F1.scoremAPmARAccuracyF1.scoremAPmAR
Closed-World80.9081.9082.5281.2881.5182.4283.1581.71
Open-World78.5678.9579.3378.5682.9883.3183.6382.98
Mixed-World79.5579.4979.9179.0881.3381.3581.8780.84
Table 11. Retrieval performance of our proposed model and the second-best model (i.e., ResNet50 [42]) for both methods (i.e., with class and without class prediction) (unit: %).
Table 11. Retrieval performance of our proposed model and the second-best model (i.e., ResNet50 [42]) for both methods (i.e., with class and without class prediction) (unit: %).
CNN ModelWithout Class PredictionWith Class Prediction
Accuracy F1.scoremAPmARAccuracyF1.scoremAPmAR
ResNet50 [42]80.4681.5882.3180.8680.9081.9082.5281.28
Proposed80.9081.8782.6081.1781.5182.4283.1581.71

Share and Cite

MDPI and ACS Style

Owais, M.; Arsalan, M.; Choi, J.; Park, K.R. Effective Diagnosis and Treatment through Content-Based Medical Image Retrieval (CBMIR) by Using Artificial Intelligence. J. Clin. Med. 2019, 8, 462. https://doi.org/10.3390/jcm8040462

AMA Style

Owais M, Arsalan M, Choi J, Park KR. Effective Diagnosis and Treatment through Content-Based Medical Image Retrieval (CBMIR) by Using Artificial Intelligence. Journal of Clinical Medicine. 2019; 8(4):462. https://doi.org/10.3390/jcm8040462

Chicago/Turabian Style

Owais, Muhammad, Muhammad Arsalan, Jiho Choi, and Kang Ryoung Park. 2019. "Effective Diagnosis and Treatment through Content-Based Medical Image Retrieval (CBMIR) by Using Artificial Intelligence" Journal of Clinical Medicine 8, no. 4: 462. https://doi.org/10.3390/jcm8040462

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop