Next Article in Journal
Diagnostic Accuracy of Imaging Findings in Pleural Empyema: Systematic Review and Meta-Analysis
Previous Article in Journal
Nearly Exact Discrepancy Principle for Low-Count Poisson Image Restoration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Effectiveness of Learning Systems from Common Image File Types to Detect Osteosarcoma Based on Convolutional Neural Networks (CNNs) Models

by
Chanunya Loraksa
1,*,
Sirima Mongkolsomlit
2,
Nitikarn Nimsuk
1,
Meenut Uscharapong
3 and
Piya Kiatisevi
3
1
Medical Engineering, Faculty of Engineering, Thammasat University, Pathum Thani 12121, Thailand
2
Faculty of Public Health, Thammasat University, Pathum Thani 12121, Thailand
3
Department of Medical Services, Lerdsin Hospital, Ministry of Public Health in Thailand, Bangkok 10500, Thailand
*
Author to whom correspondence should be addressed.
J. Imaging 2022, 8(1), 2; https://doi.org/10.3390/jimaging8010002
Submission received: 28 October 2021 / Revised: 13 December 2021 / Accepted: 23 December 2021 / Published: 27 December 2021
(This article belongs to the Topic Medical Image Analysis)

Abstract

:
Osteosarcoma is a rare bone cancer which is more common in children than in adults and has a high chance of metastasizing to the patient’s lungs. Due to initiated cases, it is difficult to diagnose and hard to detect the nodule in a lung at the early state. Convolutional Neural Networks (CNNs) are effectively applied for early state detection by considering CT-scanned images. Transferring patients from small hospitals to the cancer specialized hospital, Lerdsin Hospital, poses difficulties in information sharing because of the privacy and safety regulations. CD-ROM media was allowed for transferring patients’ data to Lerdsin Hospital. Digital Imaging and Communications in Medicine (DICOM) files cannot be stored on a CD-ROM. DICOM must be converted into other common image formats, such as BMP, JPG and PNG formats. Quality of images can affect the accuracy of the CNN models. In this research, the effect of different image formats is studied and experimented. Three popular medical CNN models, VGG-16, ResNet-50 and MobileNet-V2, are considered and used for osteosarcoma detection. The positive and negative class images are corrected from Lerdsin Hospital, and 80% of all images are used as a training dataset, while the rest are used to validate the trained models. Limited training images are simulated by reducing images in the training dataset. Each model is trained and validated by three different image formats, resulting in 54 testing cases. F1-Score and accuracy are calculated and compared for the models’ performance. VGG-16 is the most robust of all the formats. PNG format is the most preferred image format, followed by BMP and JPG formats, respectively.

1. Introduction

The global bone cancer rate is 0.2% among all types of cancer. Approximately 3600 patients were diagnosed with bone cancer and around 1720 patients passed away in the year 2020 [1]. The most common type of bone cancer is osteosarcoma, which comprises 28% of adult and 56% of adolescent bone cancer found [2]. Tumors can be found when touched and are painful when pressed. Bones can easily break from moving the body during daily routines. One of the issues in the orthopedic nursing is that the prevalence is so low. As the result, doctors are not familiar with this disease and find it difficult to diagnose. Osteosarcoma is a primary bone malignancy with a particularly high incidence rate in children and adolescents relative to other age groups [3,4]. Within one to two years after surgery, osteosarcoma patients often return with a skinny body and dyspnoea because cancer cells have spread to other organs, especially the lungs. Unfortunately, between 50 and 75% of patients with osteosarcoma will present with clinically detectable metastases from bone to lung [4,5,6], a proportion that has increased with sophisticated methods of detection such as computed tomography (CT).
Computer-aided diagnosis (CAD) of osteosarcoma metastasizes is a popular research topic because it can help doctors to detect the nodule in a patient’s lung at the early state. There are several methods that have been proposed in the past few years. The result from the CAD system can separate nodule and non-nodule images out of the large CT-scanned image series from the Digital Imaging and Communications in Medicine (DICOM) format. The most promising machine learning tools are the Convolutional Neural Networks (CNNs), since they are trainable by using labelled data or images which are then fed into the networks to get the output categories. CNNs are the most popular methods for classifying images [7,8]. They can classify images into suitable categories. A CNN extracts features and learns the important features in the Convolution layers, then classifies the output as a positive or negative diagnosis of osteosarcoma metastasizes using the fully connected layers. This learning algorithm, called a supervised learning algorithm, is used to train a CNN’s model to achieve good performance by using quality images in the dataset, such as balanced dataset images or enhanced images. Then, the quantity of the dataset, as well as the suitable pretrained neural network, all have an influence on how well the learning techniques function [9,10,11,12]. The CNN is a trainable machine learning algorithm and can classify with a high accuracy when provided with a good images dataset. In most cases, the dataset consists of huge images to learn, however, since osteosarcoma metastatic disease is very rare disease, the CT-scanned images are limited. This situation is more severe for transferred patients because of the patient privacy law and government policy in Thailand. Only the CT-scanned image file in CD-ROM media can be sent to the new hospital. The transmission of patient data or image files via the internet is prohibited for safety control. Generally, CD-ROM media has a capacity of 703 MB, therefore, the Digital Imaging and Communications in Medicine (DICOM) file from the CT-Scan machine cannot be written onto one CD-ROM. Due to this situation, the source hospital has to convert the DICOM file to another image file format such as JPG, PNG or BMP to reduce the file size before writing onto CD-ROM and transferring to the destination hospital. Then, the quality of image files is changed due to their format and may affect the accuracy of the network that was used for detecting the nodule in the lung.
One popular image file format was raster format, which consisted of grid pixels combined into a big image. When these images were zoomed in, there would be plenty of visible squares. Therefore, raster was a complex image file format that had a smooth color radiant. However, if size of the image was reduced or enlarged, the lines between pixels would start showing. The difference between pixels could be used in image processing to learn how to detect the edge of the object in an image. In addition, there was another type of image, called a vector image, from mathematical calculation. Vector images displayed in geometric shapes that were created by a graphic program. The image file format was defined by a set of mathematical parameters. The examples of vector files included AI and EPS. Presently, this type of image is not applied to medical use due to the huge file size which affected the storage space and delayed the display time, furthermore, the software in each hospital still cannot support this type of image [13,14,15].
The DICOM file does not contain only pixel image data from the CT scan machine, but it contains extra data such as a patient record, pixel density, sliced size, medical protocol, hospital information and machine specification. Moreover, the pixel data in DICOM file are recorded in a long binary format without being separated into an individual CT sliced image, therefore, the pixel data must be extracted, separated and converted into a series of common CT-scanned image files before they can be used in any machine learning algorithms. Due to the conversion, there are a variety of the image file formats with different qualities.
The common image file formats are BMP, JPG and PNG. Therefore, according to the limited images and different file formats, this research will create the CNNs for osteosarcoma metastatic disease detection and investigate the effect of the number of images in the training dataset when using different image formats, as well as evaluate the performance index F1-Score of the trained networks from different image formats with all formats.
Therefore, the aim of this study is to analyze the effectiveness of learning systems from common image file types to detect osteosarcoma based on CNNs’ models to be a guideline for a radiologist to choose a robust model regardless of image format and dataset size.

2. Methods

This study design was conducted as an experimental design in Lerdsin Hospital, Department of Medical Services, Ministry of Public Health in Thailand. According to Lerdsin Hospital, there are were total of 202 cases of the osteosarcoma metastatic disease between 2016 and 2021. The total CT-scanned images from all cases are 269,025 images, but only 2212 images were classified as positive for osteosarcoma metastatic disease. The software was written by Python language version 3.7, based on the PyTorch 1.6.0 platform, and run in the Windows 10 operating system with Nvidia CUDA 10.1 library and GTX1070 hardware acceleration GPU.

2.1. Dataset Preparation

2.1.1. Data Acquisition

These experiments were approved by the Research Ethics Committee of the Lerdsin Hospital, Department of Medical Services, Ministry of Public Health in Thailand and The Human Research Ethics Committee of Thammasat University (Science), (HREC-TUSc) for consideration. CT-scanned image data were recorded from SIEMENS SOMATOM DEFINITION 64. Specifications: 78 cm Gantry aperture, OIL/AIR, OIL/WATER tube cooling, 100–80 kW output, range: 28–665 mA (optional 800), max load capacity 227 (299 optional), reconstruction matrices 512 × 512. Siemens CARE Dose 4D low dose technology reduces patient radiation by up to 70%.

2.1.2. Ground Truth

The study included 202 patients with osteosarcoma from the Lerdsin Hospital, Department of Medical Services, Ministry of Public Health in Thailand and had 269,025 image files in Digital Imaging and Communications in Medicine (DICOM) format. Patients’ Lung lungs were scanned. There were 2212 image files indicating the presence of an abnormal nodule. To indicate the correct ground truth, one radiologist and two oncology specialists, a total of three individuals, read the CT-scanned image and identified abnormal images separately. If at least two of the three people identified the same abnormal images, then these images are the positive class images for osteosarcoma metastatic disease, which means this abnormality is possibly caused by bone cancer cells that have migrated to the lungs. The DICOM file cannot be directly applied to the CNNs since it contains both image information and other non-image information, such as hospital name and patient name. All DICOM files must be converted to BMP format using the program that comes with the hospital’s CT-scanner as part of the pre-processing step. Then, the Open Computer Vision (OpenCV) Library is used to convert the image files from BMP to JPG and PNG formats, which are the most common picture file formats used in the medical field and in CNNs’ teaching process. There are 269,025 image files extracted from the patients’ DICOMs, and 2212 positive and 2212 negative class image files are used for training the networks in this study.

2.1.3. Datasets

The quantity and shape of the dataset have a big influence on how well machine learning and deep learning approaches work. As a result, in this study, the total 2212 positive and 2212 negative class image files are separated into train images and test images. Eighty percent of the images in this dataset are used to train the model (1769 images), while twenty percent of the data (443 images) is utilized to evaluate the suggested CAD system’s resilience. In order to simulate the limited training images in a dataset situation and in order to evaluate CNNs as a realistic diagnosis and management tool for osteosarcoma metastasizes, the second dataset is generated. The small dataset contains only ten percent of the training images of the original large dataset and equals 177 images for both positive and negative images. The number of images in the small dataset was selected by considering the average of CT-scanned images that contains nodules of one patient in one CT scan to simulate the situation where a hospital has only one patient with an osteosarcoma disease case.

2.2. Pretrained Neural Networks

CNNs have been widely utilized to solve many issues in various fields, but their performance in image processing for health applications is outstanding. There is a lot of research that proposes CAD-based detection for the lesion of disease [16,17]. CNNs are currently the most widely used for the computer vision application using the deep learning (DL) approaches. They can also apply new and powerful regularization techniques to a limited training dataset, which is one of the most important factors in their success [17]. VGG architecture is modest and commonly used in the image classification network because it is easy to construct using the repeating structure when it grows in depth [18]. ResNet is also a popular and commonly used CNN that constructs based on a residue deep structure with skip connections [19]. This structure allows information to flow and hop over a set of layers, therefore resolving the problem of vanishing gradients. The vanishing gradients are the big issue for deep networks because it is difficult for deep networks to transmit the error function backwards. The gradients which are used for training weights and biases in the layers become too tiny beyond a certain point, resulting in a network training stop. MobileNet is proposed by Google for low computational devices to be able to use CNNs. This network is most generally used on mobile devices, such as a mobile phone, or on a mini-computer such as Raspberry Pi [20,21]. MobileNet becomes a popular and commonly used CNN due to its performance and small calculation power required.
Generally, the common CNN networks are provided as pretrained neural networks [9,10,11,12]. The pretrained network is the network that comes with all weights and biases in every layer and provides the best output result. The most important characteristics of a pretrained neural network are the accuracy, speed and size. Greater accuracy increases the specificity and sensitivity for detection. According to previous research, the performance of sixteen popular pretrained neural networks are ranked in descending order based on the overall accuracy: ResNet-50—97.18%, DarkNet-53—95.99%, VGG-19—95.35%, DenseNet-201—95.11%, ResNet-18—94.23%, DarkNet-19—94.08%, SqueezeNet—93.84%, ResNet-101—91.43%, GoogLeNet—90.71%, SuffleNet—90.07%, MobileNet-V2—88.07%, Place365-GoogLeNet—85.69%, Inception-V3—80.37%, Xception—74.62%, RasNet-Mobile—73.75% and Inception-ReNet-V28—72.93% [22].
In this research, VGG-16, ResNet-50 and MobileNet-V2 CNN models are selected due to the accuracy and simplicity. VGG-16, ResNet-50 and MobileNet-V2 models are the representative of high, medium and low parameter models that were applied and compared in a medical image application [23]. These models are trained for the classification of lesions of osteosarcoma metastasizes from bone to lungs and used to investigate the effect of the different train and test image formats. The pretrained networks are used and modified using transfer learning techniques that replace the fully connected layers into a suitable structure for the osteosarcoma metastasizes’ classification, as shown in Figure 1 [22,24]. There are two classes of output of the networks which are positive and negative classes for nodules.

2.2.1. VGG Model

VGG is the CNNs’ deep network proposed to solve image classification problems [18,25]. This network demonstrated that, by increasing the depth from 16 to 19 levels of layers, it can produce a substantially good result. The convolutional layer receives a fixed-size 224 × 224 pixels picture as an input. The picture is processed through a stack of convolutional layers with ReLU activations and filters with extremely tiny receptive fields. The stride of the convolution is likewise set to 1. Spatial pooling is performed after part of the convolutional layers by five max-pooling layers. The advantage of VGG is that, by stacking several convolutional layers with small-sized kernels, the network’s effective receptive field is enhanced while the number of parameters is reduced, as compared to utilizing fewer convolutional layers with bigger kernels for the same receptive field. The authors experimented with a variety of combinations of different depth (9, 11, 16 and 19 layers) [18,25]. According to the original article, the greatest results were obtained for depths ranging from 16 to 19. The pretrained VGG is modified to fit to our research by replacing the original fully connected layers with new fully connected layers that have two output classes (positive and negative classes) and one dropout layer. The dropout layer is added to prevent the overfit issue [26]. Figure 2 shows the VGG-16 architecture of this research. Normally, the convolutional layers in the pretrained CNNs are frozen but, in our case, the CT-scanned training images are represented in a gray scale format. The pretrained model is trained by RGB format images, and they are different from the CT-scanned gray scale format that was used in this research. Therefore, these convolutional layers are not frozen in this research to allow weights and biases in these layers to adapt themselves for a better result.

2.2.2. Residual Networks Model

Residual networks (ResNets) [19] are made up of reformed convolutional layers that learn residual functions based on the input images. According to the original article, this network is easy to optimize and can have a much greater depth. The unique structure of ResNets is a “residual block”, as described in [19,27]. It creates a simple “shortcut connection” for every few convolutional layer that runs parallel to these layers and performs the identity mapping enhancement. Aside from the usage of shortcut links, the network architecture is mostly influenced by the VGG network concept. Small kernels of size 3 × 3 are used in all convolutional layers, which follow two basic design rules. First, the layers have the same number of filters for the same output feature map size. Second, when the feature map size is half (with stride 2 convolutional layers), the number of filters is doubled to maintain the time complexity per layer [25]. ResNet-50′s core principle is to utilize identity mapping to anticipate the need to obtain the final prediction of the preceding layer outputs. ResNet-50 reduces the vanishing gradient effect by using a different shortcut path to circumvent it. The identity mapping enables the model to bypass the superfluous layers. This assists the model in overcoming the overfitting issue to the training set [27]. In this research, the pretrained ResNet-50 CNN model is selected [24,27]. Like the VGG model, the fully connected layers are replaced based on the transfer learning technique by the desired fully connect layer that has two outputs classed with a dropout layer, as shown in Figure 3. The convolutional layers are not frozen as they are in the VGG network.

2.2.3. MobileNet Model

MobileNets are efficient models for mobile and embedded vision applications, especially for hardware with limited resources. MobileNets are built on a simplified design that uses depthwise separable convolutions to construct low weight deep neural networks instead of regular convolutions, except for the first layer, which is a complete convolution [21]. MobileNet is simpler than VGG-16 and ResNet-50 [8,20,21]. MobileNet-V2 expands on the principles of MobileNet-V1 by employing depthwise separable convolution as efficient building pieces. However, MobileNet-V2 adds two new architectural features [28], which are linear bottlenecks between layers and shortcut connections between bottlenecks. The overview of MobileNet-V2 Architecture is shown in Figure 4, where blocks represent composite convolutional building. In this research, the fully connected layers of MobileNet-V2 are also transferred learning to suit for two classes of output. A dropout layer is also added to the network, similar to VGG-16 and ResNet-50.

2.3. Train and Validate the CNNs Networks

There are three possible common CT-scanned image formats that submit to Lerdsin Hospital during the patient transfer procedure, which are BMP, JPG and PNG. In order to obtain the impact from the image file formats and two different size datasets, 18 networks must be created and trained, as shown in Table 1. The CT-scanned small dataset consists of 177 positive and 177 negative class images, each of JPG, PNG and BMP. Training is performed via the three CNN models of VGG-16, ResNet-50 and MobileNet-V2. Thus, there are nine training sets for nine networks (B1 to B9). The CT-scanned large dataset consists of 1769 positive class images and 1769 negative class images, each of JPG, PNG and BMP. Training is conducted via the following three models: VGG-16, ResNet-50 and MobileNet-V2. Thus, we have nine training sets for nine networks (B10 to B18) to ensure that unbiased data lead to proper conclusions [29].
The performance of the trained CNN is evaluated by unseen test images that were separated when the datasets are created. Each trained network is evaluated by three different common image formats to obtain the effect when the training image formats are different from testing image formats. Moreover, in order to make the comparison between networks trained by large and small datasets fair, the same number of unseen test images (443 images) are fed into these networks. For an overview of cases in this research, there are 18 trained networks, and each network is tested with 3 different image formats, thus, there are 54 testing cases, T1 to T54, in total. The evaluation of networks’ performance is performed by calculation of the F1-Score.

2.4. Loss Function

The loss function is essential for the neural network training process, in order to modify the weights and biases of a neural network to build a better result. During forward propagation, the neural network is run on training set data and outputs are created that reflect the probability, or confidence in probable labels in the case of classification. These probabilities are compared to the target labels, and the loss function computes a penalty for any difference between the target label and the neural network outputs. The partial derivative of the loss function is generated for each trainable weight of the neural network during backpropagation [30,31,32]. These partial derivatives are used to modify the weights. Backpropagation iteratively changes the trainable weights of a neural network to generate a model with reduced loss under normal conditions. Thus, the suitable loss function must be selected carefully to suit with the output, since the output of the considered CNNs are positive and negative classes for nodules. As a result, the Binary Cross-Entropy (BCE) loss function is introduced for determining classification goodness-of-fit [23]. The BCE is suitable because it is good for a training network with two possible classes of results (positive or negative). The BCE loss function is suitable for our three selected networks and is given by (1) [30,32].
L o s s = 1 N n = 1 N [ y n l o g y ^ n + ( 1 y n ) log ( 1 y ^ n ) ]
where
  • N number of training examples
  • y ^ n predicted value
  • y n expected value

2.5. Performance Benchmark Scores

In the engineering field, the accuracy, precision, recall and F1-Score [33] are four important metrics that are typically required for measuring the performance of a neural network model [12]. The accuracy, precision and recall scores can be calculated by using Equations (2)–(4), respectively. The calculations are based on four qualitative values, which are True Positive (TP), True Negative (TN), False Positive (FP) and False Negative (FN). TP is the correctly predicted positive value, which means that the value of the actual class is positive, and the value of the predicted class is also positive. TN is the correctly predicted negative value, which means that the value of the actual class is negative, and the value of the predicted class is also negative. FP and FN are the values that occur when the predicted classes are different from the actual classes. If the predicted class is positive but the actual class is negative, this is the FP. On the other hand, if the predicted class is negative but the actual class is positive, this is the FN.
Accuracy   ( % ) = TP + TN TP + FP + FN + TN × 100 %
Precision = TP TP + FP
Recall = TP TP + FN
The F1-Score is a metric for how accurate a model is on a given dataset. It is used to assess binary classification algorithms that categorize examples as either positive or negative. The F1-Score, which is defined as the harmonic mean of the model’s precision and recall, is a technique of combining the model’s precision and recall. F1-Score is commonly used to evaluate binary or multiclass classification models’ performance [9,12,33] because this F1-Score takes both false positives and false negatives into account. Accuracy works best if false positives and false negatives have a similar cost. If the cost of false positives and false negatives is very different, the F1-Score is the better measurement benchmark. The F1-Score is defined by Equation (5).
F 1   Score = 2 × Precision   ×   Recall Precision + Recall
For the medical field, other scores are commonly used, because it reflexes the medical diagnose information better than the engineering score. These scores are the sensitivity, specificity, F1-Score and Receiver Operating Characteristic (ROC), because the agreement between a proposed (index) test and a reference standard for diagnosing a target condition is examined in diagnostic accuracy studies. The main method is to look at a set of well-defined patients who are blinded to both the index and reference tests in a sequential order. Diagnostic accuracy refers to the degree of agreement between the index test and the reference standard. The reference standard is shown in Table 2. The percentage of malignant nodules that are accurately diagnosed as metastatic nodules is measured by sensitivity, which is the proportion of real positive samples that are appropriately identified. As a result, it is calculated using Equation (6). The sensitivity and recall values are the same. In this case, TP is the number of nodules successfully detected, and FN represents the number of positive nodules that are undetected. Specificity, on the other hand, assesses the percentage of detected negative samples, in which the percentage of nodules that are not malignant is accurately categorized as non-cancerous. Specificity is calculated by Equation (7). TN is the number of non-cancer patients correctly categorized, and FP denotes the number of non-cancer patients incorrectly diagnosed as cancer. F1-Score, on the other hand, calculates the average score, as already explained in Equation (5) [34,35].
Sensitivity = TP TP + FN
Specificity = TN TN + FP

3. Results

In order to obtain the performance impact from different quality image formats and different sizes of training data in datasets, the experiments were conducted. The first experiment was conducted by training all three possible CT-scanned image file formats by three different CNNs, which are VGG-16, ResNet-50 and MobileNet-V2. The dataset in the first experiment, called the large dataset, was randomly selected (80%) from 2212 CT-scanned images and transformed into 1769 CT-scanned images in BMP, JPG and PNG formats. These 1769 CT-scanned images contained nodules and were labeled as the positive class images. The rest of images in CT-scanned images were used for evaluating the trained networks. To balance the positive images in the large dataset, the negative CT-scanned images were also randomly selected from other CT-scanned images which did not contain nodules. There were 3538 total images in the large dataset that were used to train the networks. In this experiment, due to limitation of the hospital’s computer resources, a batch size was set to 16 per 1 iteration, which means that 16 images were feedforwarded into the networks and the error from each image was obtained and summarized before using this error to find the gradient and train the networks via back propagated algorithm to reduce the loss value. To cover all the images in the dataset, one epoch requires 222 iterations. The term “epoch” means training the neural network with all the training data for one cycle. Therefore, if the networks are trained for 400 epochs, it requires 88,800 iterations, and if the networks are trained for 2000 epochs, it requires 444,000 iterations.
In the first experiment, each model, VGG-16, ResNet-50 and MobileNet-V2, had been trained with three different image file types for 400 epochs. Thus, there were nine trained networks, which were named as L-VGG-JPG, L-VGG-PNG, L-VGG-BMP, L-ResNet-JPG, L-ResNet-PNG, L-ResNet-BMP, L-MobileNet-JPG, L-MobileNet-PNG and L-MobileNet-BMP. These names will be used as the reference in this section. The L is used to explain that the networks had been trained with the large dataset. Figure 5 shows the loss value when training the VGG-16. The loss value graphs of the ResNet-50 and MobileNet show similar behavior to VGG-16, with differences in a training slope and steady-state loss fluctuation value. Loss of the models when trained by different image formats was different at the beginning, however, loss decreased to zero when the number of epochs increased. Loss of VGG-16 (a) was the smallest among the three selected models. By considering loss, the performance of the models when evaluated by the dataset can be obtained. If the loss value is near to zero, there is less error.
The one big concern for a radiologist or oncologist who trains the model in his hospital is when he should stop the training. In order to answer this question, the second experiment was conducted. In the second experiment, the CNN models were trained with more epochs in order to obtain the effect of epochs to loss values. Since VGG-16 models have the smallest loss value from the first experiment, only VGG-16 networks were continuously trained up to 2000 epochs. Then, the training graphs were recorded and divided into four training phases. In phase 1, loss values between 300 and 400 epochs were considered, as shown in Figure 6a. Loss values between 1000 and 1100 epochs, 1500 and 1600 epochs and 1900 and 2000 epochs were considered for phase 2, 3 and 4, respectively, as shown in Figure 6b–d. In each phase, the trendlines of losses vs. epochs were calculated and the slopes of the trendlines were collected. Table 3 shows the slope of all training phases.
From Table 3, the slope of each training phase is very close to each other, since the slopes are very small. This second experiment discovered that the loss values converted to zero and the training slopes converged to 0 for all training phases. Thus, it was not necessary to keep training over 400 epochs. Then, the result explains that if the radiologist trains the network and its loss is small enough, he can stop the training process with confidence.
Because osteosarcoma is a very rare bone cancer, it is possible that CT-scanned images in the dataset for training the CNN model may be very small. Therefore, a third experiment was conducted to obtain the impact from the small training dataset to the models’ performances. This experiment aims to simulate the situation for a small hospital that has a very small number of CT-scanned images of patients. To obtain the effect of a small dataset, therefore, 10% of the 2212 positive class images, or 220 images, were randomly picked for model training. The positive 177 images (80% of 220 images) were also randomly selected to be the positive CT-scanned class images for training, combined with 177 negative CT-scanned class images that did not contain nodules, to create the small training dataset. There were 354 total images in the small dataset that were used to train the VGG-16, ResNet-50 and MobileNet-V2. Then, the CNN models were trained for 400 epochs in the same scenario as the first experiment, and there were nine total trained models output from this experiment. The training graphs of the small dataset were displayed and compared with the training graphs of the large dataset. To see this phenomenon, the training graphs were expanded from 300 to 400 epochs, as shown in Figure 7. The fluctuation of VGG-16 model loss is very small when trained by small and large datasets. In the case of ResNet-50 and MobileNet-V2, loss values fluctuated within a larger range if the models were trained by the small dataset. Loss of ResNet-50 fluctuated about 0.23 when trained by the small dataset and 0.07 when trained by the large dataset. Loss of MobileNet-V2 fluctuated about 0.17 when trained by the small dataset and 0.07 when trained by the large dataset. According to the result, VGG-16 is the most robust CNN model when trained by the large and small dataset.
The first, second and third times of experiments show that the losses were approaching zero for all the models for all conditions. However, they are not reflective of real performance when the trained models are used in real situations where the input images are unseen. Therefore, the fourth experiment was conducted to obtain the model’s performance when validated by unseen CT-scanned images. The unseen positive CT-scanned images are the remaining 443 images of each image file type (JPG, PNG and BMP) from the first experiment, which were not included in the large dataset. The unseen 443 negative CT-scanned images were randomly selected from the CT-scanned images that were not included in any dataset and did not have nodules in the images. Then, all 400 epoch-trained models (nine models trained by the small dataset and nine models trained by the large dataset) were evaluated by 3threedifferent unseen image file types, yielding 54 cases in total. The evaluated results and performance scores for the small dataset trained models are shown in Table 4 and the evaluated results and performance scores for the large dataset trained models are shown in Table 5. The experimental result showed that there is a small effect caused by different unseen file formats to the model’s performance. The results shown concluded that the trained models can achieve the same performance when used in the real scenario where the input image formats are different, such as in the case of the transferred patients.
In the case of the small dataset, VGG-16 models have the best F1-Score when evaluated by unseen images, while the accuracy of all models is close to each other. Moreover, VGG-16 model can achieve the same scores and results (TP, TN, FP and FN) regardless of any trained or tested image file formats. Therefore, if there is a limited training CT-scanned image, the VGG-16 model is the best CNN model choice for detecting nodules, since it is the most robust model with the highest F1-Score. In the case of the large dataset, the accuracy of all models decreased, while the F1-Score increased for some models. In the case of ResNet-50, the models that were trained by JPG file format had a higher F1-Score than models that were trained by BMP and PNG formats, respectively. Although MobileNet-V2 models have similar results as the ResNet-50, the models that were trained by BMP format had the highest F1-Score, followed by models that were trained by PNG and JPG formats, respectively. The VGG-16 model was still robust for any trained or tested image formats in the case of the large dataset in consideration of the precision and recall values, even though it did not have the best F1-Score.
To verify the results in the case of training with the large dataset, the 2000 epochs trained models from the fourth experiment were considered. The evaluated results and performance scores for the large dataset 2000 epochs trained models are shown in Table 6. The accuracy and F1-Score of some models decreased and some models increased.
The uncertainty remains for ResNet-50 and MobileNet-V2 in the 2000 epochs training case. For ResNet-50, the model that was trained by JPG format had the highest F1-Score, followed by models that were trained by PNG and BMP formats, respectively. Even though the F1-Score of the ResNet-50 model that was trained by JPG format slightly increased when epochs were increased from 400 to 2000, the F1-Score of the ResNet-50 models that were trained by PNG and BMP formats decreased. The situation was more confused for MobileNet-V2 cases. The MobileNet-V2 model that was trained by BMP format had the highest F1-Score when the model was trained for 400 epochs, but the MobileNet-V2 model that was trained by PNG format had the highest F1-Score when the model was trained for 2000 epochs. The models that were trained by JPG format had a higher F1-Score than the model that was trained by JPG format when epochs were increased to 2000. For VGG-16, all scores remained the same even if the epochs were increased. From this result, the VGG-16 model is still the most robust CNN model for detecting nodules in a CT-scanned image file. It is not easy to find the number of epochs required for the ResNet-50 and MobileNet-V2 models to have the best F1-Score. In addition, the image file type formats affected the score of the ResNet-50 and MobileNet-V2 inconclusively.
All experiments show low F1-Scores and accuracy in all models, regardless of any training or testing image formats, due to small nodule size in the CT-scanned image. However, the experiments discovered the robustness of each model according to different image formats, which addresses the issue of the transferred patients between hospitals in Thailand. In varieties of common image formats that can be sent to Lerdsin Hospital, the experimental results show a guideline for a radiologist to select the VGG-16 model in order to obtain the consistency, accuracy and F1-Scores output regardless of image format and dataset size. Other models show fluctuation and unpredictable F1-Scores and accuracy when these models are trained and tested with different image formats and a different number of epochs. The experiment also provides a guideline for a training where, if the model loss is low enough, then, it is not necessary to keep training the model. Finally, the radiologist can be comfortable using the VGG-16 model for any common image file formats that are already on hand or in the future.

4. Conclusions and Discussion

This research aims to analyze the effectiveness of learning systems from common image file types to detect osteosarcoma based on CNNs’ models. There are three popular CT-scanned CNN models investigated in this research, which are VGG-16, ResNet-50 and MobileNet-V2 models. Osteosarcoma bone cancer is a rare type of cancer, thus, it is hard to detect even when advanced computer aided technology, such as machine learning or CNNs, is applied [36]. There are limited cases and resources for successful machine learning. Furthermore, there is an additional problem in Thailand. According to the technical and policy issues in the case of transferred patients in Lerdsin Hospital, all patient records and image files must be written in CD-ROM media for safety purposes. The patient’s DICOM files could not be transferred from the source hospital to the destination hospital because they could not be put onto CD-ROM, therefore, the DICOM files must be transformed into other common CT-scanned image file formats [37,38]. The three common CT-scanned image file formats that are sent to Lerdsin Hospital are BMP, JPG and PNG formats [15,39]. This research searched for the answer to the question of the impact from different file formats on the CNN models while training and on performance validation. This conclusion can leverage the telemedicine operation in Thailand. Hospitals in remote areas can use the online detection system by using any kind of CT-scanned image format, yet achieve the same result. The findings indicated that there are uncertain and inconclusive results for ResNet-50 and MobileNet-V2 when the models were trained by different CT-scanned image formats. Also, the F1-Score of the same model was changed when the epochs were increased. This result implied that ResNet-50 and MobileNet-V2 are not robust for osteosarcoma bone cancer metastasized to lung detection when trained by different training image formats. They are dependent on both training image format and number of epochs, which lead to unclear conclusion about how many training epochs are required in order to obtain the best F1-Score and which training image format should be used. However, in the case of the VGG-16 model, the results from both models trained by large and small datasets were consistent. The different CT-scanned image formats did not affect performance scores of the trained models in any case. VGG-16 is the most robust CNN model for osteosarcoma bone cancer metastasized to lung detection. Due to the experimental results, the file formats have little impact on the overall performance scores. Thus, the criteria for selecting image formats must be the image size and quality which could be most practically used. The average CT-scanned image sizes of the JPG, PNG and BMP formats in the datasets are 177.22, 271.10 and 1029.92 Kilobytes, respectively.
Therefore, the PNG image format is preferred for both training and validating models in medical applications, because the PNG format is a patent-free and lossless compression image format that has been proven as the best quality medical image for radiologic application [15,39] with a small storage size [37] that is capable for a small hospital.
F1-Score decreased when the training epochs increased. This phenomenon from this research also has been investigated and discussed for further understanding. According to the positive class images, as shown in Figure 8, the area of the nodule in the patient lung is tiny when compared with the whole image area. This situation makes it difficult even for experienced radiologists to detect the positive nodules in the image. Due to this fact, to increase in the performance of the models for detecting the osteosarcoma nodules in the lung, and the more sophisticated machine learning algorithms or CNN models that can identify the region of interest that has the most probability to contain the nodules, must be further investigated, such as a Region-based Convolutional Neural Network (RCNN) or Single Shot Detection (SSD) framework.

Author Contributions

Conceptualization, C.L. and S.M.; methodology, C.L.; software, C.L. and N.N.; validation, C.L., S.M., N.N. and P.K.; formal analysis, P.K. and M.U.; investigation, S.M. and P.K.; resources, P.K. and M.U.; data curation, C.L. and P.K.; writing—original draft preparation, C.L.; writing—review and editing, C.L. and S.M.; visualization, C.L.; supervision, S.M.; project administration, S.M.; funding acquisition, S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was performed with the financial support from the Office of Research Administration, Ph.D. scholarship from Thammasat University, 1/2019.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, the Belmont Report, the CIOMS Guideline International Conference on Harmonization in Good Clinical Practice, the 45CFR 46.101(b), and approved by the Human Research Ethics Committee of Thammasat University (project code 020/2564 on 2 June 2021) and approved by Lerdsin Hospital Ethics Committee (project code LH62056 on 27 September 2021).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

The raw data supporting the conclusions of this manuscript can be made available on request by the authors to any qualified researcher.

Acknowledgments

We would like to thank orthopaedic and oncologist Piya Kiatisevi, and the radiologist, Meenut Uscharapong, from the Lerdsin Hospital in Thailand for their consulting support.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siegel, R.L. Cancer Statistics 2020. CA Cancer J. Clin. 2020, 70, 7–30. [Google Scholar] [CrossRef] [PubMed]
  2. Cancer.Net Doctor-Approved Patient Information from ASCO Publications, Bone Cancer (Sarcoma of Bone): Statistics, Approved by the Cancer.Net Editorial Board, 01/2021. Available online: https://www.cancer.net/cancer-types/bone-cancer-sarcoma-bone/statistics (accessed on 2 October 2021).
  3. Martin, J.W.; Squire, J.; Zielenska, M. The Genetics of Osteosarcoma. Sarcoma 2012, 2012, 627254. [Google Scholar] [CrossRef] [Green Version]
  4. Matsubara, E.; Mori, T.; Koga, T.; Shibata, H.; Ikeda, K.; Shiraishi, K.; Suzuki, M. Metastasectomy of Pulmonary Metastases from Osteosarcoma: Prognostic Factors and Indication for Repeat Metastasectomy. J. Respir. Med. 2015, 2015, 570314. [Google Scholar] [CrossRef] [Green Version]
  5. Su, W.T.; Chewning, J.; Abramson, S.; Rosen, N.; Gholizadeh, M.; Healey, J.; Meyers, P.; La Quaglia, M.P. Surgical management and outcome of osteosarcoma patients with unilateral pulmonary metastases. J. Pediatr. Surg. 2004, 39, 418–423. [Google Scholar] [CrossRef] [PubMed]
  6. Bramwell, V. Metastatic Osteosarcoma: A Review of Current Issues in Systemic Treatment. Sarcoma 1997, 1, 123–130. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Liang, X.; Yu, J.; Liao, J.; Chen, Z. Convolutional Neural Network for Breast and Thyroid Nodules Diagnosis in Ultrasound Imaging. BioMed Res. Int. 2020, 2020, 1763803. [Google Scholar] [CrossRef] [PubMed]
  8. Lin, Z.; Ye, H.; Zhan, B.; Huang, X. An Efficient Network for Surface Defect Detection. Appl. Sci. 2020, 10, 6085. [Google Scholar] [CrossRef]
  9. Narin, A.; Kaya, C.; Pamuk, Z. Automatic detection of coronavirus disease (COVID-19) using X-ray images and deep convolutional neural networks. Pattern Anal. Appl. 2021, 24, 1207–1220. [Google Scholar] [CrossRef]
  10. Sethy, P.K.; Behera, S.K.; Ratha, P.K.; Biswas, P. Detection of coronavirus Disease (COVID-19) based on Deep Features and Support Vector Machine. Int. J. Math. Eng. Manag. Sci. 2020, 5, 643–651. [Google Scholar] [CrossRef]
  11. Zhang, J.; Xie, Y.; Pang, G.; Liao, Z.; Verjans, J.; Li, W.; Sun, Z.; He, J.; Li, Y.; Shen, C.; et al. Viral Pneumonia Screening on Chest X-rays Using Confidence-Aware Anomaly Detection. IEEE Trans. Med. Imaging 2021, 40, 879–890. [Google Scholar] [CrossRef]
  12. Hemdan, E.E.-D.; Shouman, M.A.; Karar, M.E. COVIDX-Net: A Framework of Deep Learning Classifiers to Diagnose COVID-19 in X-ray Images. Cornell University. Available online: https://arxiv.org/abs/2003.11055 (accessed on 18 October 2021).
  13. Sakshica, D.; Gupta, K. Various Raster and Vector Image File Formats. Int. J. Adv. Res. Comput. Commun. Eng. 2015, 4, 268–271. [Google Scholar] [CrossRef]
  14. Gonzalez, R.C.; Woods, R.E. Digital Image Processing, 3rd ed.; Pearson Education: Upper Saddle River, NJ, USA, 2008; pp. 526–538. [Google Scholar]
  15. Tan, L.K. Image file formats. Biomed. Imaging Interv. J. 2006, 2, e6. [Google Scholar] [CrossRef] [PubMed]
  16. Naseer, A.; Yasir, T.; Azhar, A.; Shakeel, T.; Zafar, K. Computer-Aided Brain Tumor Diagnosis: Performance Evaluation of Deep Learner CNN Using Augmented Brain MRI. Int. J. Biomed. Imaging 2021, 2021, 1–11. [Google Scholar] [CrossRef] [PubMed]
  17. Signoroni, A.; Savardi, M.; Baronio, A.; Benini, S. Deep Learning Meets Hyperspectral Image Analysis: A Multidisciplinary Review. J. Imaging 2019, 5, 52. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  19. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef] [Green Version]
  20. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L. MobileNetV2: Inverted residuals and linear bottlenecks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar] [CrossRef] [Green Version]
  21. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. Available online: https://arxiv.org/pdf/1704.04861.pdf (accessed on 24 October 2021).
  22. Elgendi, M.; Fletcher, R.; Howard, N.; Menon, C.; Ward, R. The Evaluation of Deep Neural Networks and X-Ray as a Practical Alternative for Diagnosis and Management of COVID-19. medRxiv 2020. [Google Scholar] [CrossRef]
  23. Zulkifley, M.A.; Abdani, S.R.; Zulkifley, N.H. Automated Bone Age Assessment with Image Registration Using Hand X-ray Images. Appl. Sci. 2020, 10, 7233. [Google Scholar] [CrossRef]
  24. Peng, J.; Kang, S.; Ning, Z.; Deng, H.; Shen, J.; Xu, Y.; Zhang, J.; Zhao, W.; Li, X.; Gong, W.; et al. Residual convolutional neural network for predicting response of transarterial chemoembolization in hepatocellular carcinoma from CT imaging. Eur. Radiol. 2020, 30, 413–424. [Google Scholar] [CrossRef] [Green Version]
  25. Tsochatzidis, L.; Costaridou, L.; Pratikakis, I. Deep Learning for Breast Cancer Diagnosis from Mammograms—A Comparative Study. J. Imaging 2019, 5, 37. [Google Scholar] [CrossRef] [Green Version]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  27. ElGhany, S.A.; Ibraheem, M.R.; Alruwaili, M.; Elmogy, M. Diagnosis of Various Skin Cancer Lesions Based on Fine-Tuned ResNet50 Deep Network. Comput. Mater. Contin. 2021, 68, 117–135. [Google Scholar] [CrossRef]
  28. Google AI Blog The latest from Google Research, Mark Sandler, Andrew Howard, MobileNetV2: The Next Generation of On-Device Computer Vision Networks. 2018. Available online: https://ai.googleblog.com/2018/04/mobilenetv2-next-generation-of-on.html#1 (accessed on 24 October 2021).
  29. Šimundić, A.-M. Bias in research. Biochem. Med. 2013, 23, 12–15. [Google Scholar] [CrossRef]
  30. Ho, Y.; Wookey, S. The Real-World-Weight Cross-Entropy Loss Function: Modeling the Costs of Mislabeling. IEEE Access 2020, 8, 4806–4813. [Google Scholar] [CrossRef]
  31. Dufourq, E.; Bassett, B.A. Automated Problem Identification: Regression vs. Classification via Evolutionary Deep Networks. In Proceedings of the South African Institute of Computer Scientists and Information Technologists, Thaba Nchu, South Africa, 26–28 September 2017; pp. 1–9. [Google Scholar]
  32. Zhang, Z.; Sabuncu, M.R. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels. In Proceedings of the 32nd Conference on Neural Information Processing Systems, NeurIPS, Montreal, QC, Canada, 3–8 December 2018; pp. 1–11. [Google Scholar]
  33. Sasaki, Y. The Truth of the F-Measure. 2007. Available online: https://www.toyota-ti.ac.jp/Lab/Denshi/COIN/people/yutaka.sasaki/F-measure-YS-26Oct07.pdf (accessed on 25 October 2021).
  34. Chen, H.-C.; Sunardi; Liau, B.-Y.; Lin, C.-Y.; Akbari, V.B.H.; Lung, C.-W.; Jan, Y.-K. Estimation of Various Walking Intensities Based on Wearable Plantar Pressure Sensors Using Artificial Neural Networks. Sensors 2021, 21, 6513. [Google Scholar] [CrossRef]
  35. Serj, M.F.; Lavi, B.; Hoff, G.; Valls, D.P. A Deep Convolutional Neural Network for Lung Cancer Diagnostic. arXiv 2018, arXiv:1804.08170. [Google Scholar]
  36. Sriramakrishnan, P.; Kalaiselvi, T.; Padmapriya, S.T.; Shanthi, N.; Ramkumar, S.; Kalaichelvi, N. An Medical Image File Formats and Digital Image Conversion. Int. J. Eng. Adv. Technol. 2019, 9, 74–78. [Google Scholar]
  37. Ujgare, N.S.; Baviskar, S.P. Conversion of DICOM Image in to JPEG, BMP and PNG Image Format. Int. J. Comput. Appl. 2013, 62, 22–26. [Google Scholar] [CrossRef]
  38. Oladiran, O.; Gichoya, J.; Purkayastha, S. Conversion of JPG Image into DICOM Image Format with One Click Tagging. In Proceedings of the Lecture Notes in Computer Science; Springer: Singapore, 2017; pp. 61–70. [Google Scholar]
  39. Wiggins, R.H.; Davidson, H.C.; Harnsberger, H.R.; Lauman, J.R.; Goede, P.A. Image File Formats: Past, Present, and Future1. RadioGraphics 2001, 21, 789–798. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Structure of transfer learning.
Figure 1. Structure of transfer learning.
Jimaging 08 00002 g001
Figure 2. VGG-16 Architecture of this research.
Figure 2. VGG-16 Architecture of this research.
Jimaging 08 00002 g002
Figure 3. ResNet-50 Architecture of this research.
Figure 3. ResNet-50 Architecture of this research.
Jimaging 08 00002 g003
Figure 4. MobileNet-V2 building block architecture of this research.
Figure 4. MobileNet-V2 building block architecture of this research.
Jimaging 08 00002 g004
Figure 5. The loss of trained VGG-16 models.
Figure 5. The loss of trained VGG-16 models.
Jimaging 08 00002 g005
Figure 6. Loss graph of L-VGG-JPG, L-VGG-PNG and L-VGG-BMP phase I-IV (ad), respectively.
Figure 6. Loss graph of L-VGG-JPG, L-VGG-PNG and L-VGG-BMP phase I-IV (ad), respectively.
Jimaging 08 00002 g006
Figure 7. Training graph between 300 to 400 epochs of 18 CNN models trained by the small (left) and large (right) dataset.
Figure 7. Training graph between 300 to 400 epochs of 18 CNN models trained by the small (left) and large (right) dataset.
Jimaging 08 00002 g007
Figure 8. Example of nodule areas compared with image areas.
Figure 8. Example of nodule areas compared with image areas.
Jimaging 08 00002 g008
Table 1. Osteosarcoma nodules that have metastasized from bone to lung datasets for training.
Table 1. Osteosarcoma nodules that have metastasized from bone to lung datasets for training.
CodesDetailsPositive (No. of Image)Negative (No. of Image)Total (No. of Image)
Small DatasetB1_VGG16JPGVGG16 trained on JPG177177354
B2_VGG16PNGVGG16 trained on PNG177177354
B3_VGG16BMPVGG16 trained on BMP177177354
B4_ResNetJPGResNet50 trained on JPG177177354
B5_ResNetPNGResNet50 trained on PNG177177354
B6_ResNetBMPResNet50 trained on BMP177177354
B7_MobileNetJPGMobileNetV2 trained on JPG177177354
B8_MobileNetPNGMobileNetV2 trained on PNG177177354
B9_MobileNetBMPMobileNetV2 trained on BMP177177354
Large DatasetB10_VGG16JPGVGG16 trained on JPG176917693565
B11_VGG16PNGVGG16 trained on PNG176917693565
B12_VGG16BMPVGG16 trained on BMP176917693565
B13_ResNetJPGResNet50 trained on JPG176917693565
B14_ResNetPNGResNet50 trained on PNG176917693565
B15_ResNetBMPResNet50 trained on BMP176917693565
B16_MobileNetJPGMobileNetV2 trained on JPG176917693565
B17_MobileNetPNGMobileNetV2 trained on PNG176917693565
B18_MobileNetBMPMobileNetV2 trained on BMP176917693565
Total 17,51417,51435,055
Table 2. Reference standard.
Table 2. Reference standard.
Disease PresentDisease AbsentTotal
Index Test PositiveTrue Positive (TP)False Positive (FP)TP + FP
Index Test NegativeFalse Negative (FN)True Negative (TN)FN + TN
TotalTP + FNFP + TN
Table 3. Slopes of trendlines in four training phases.
Table 3. Slopes of trendlines in four training phases.
PeriodTrendline Slopes
JPGPNGBMP
300–400 5 × 10 8 8 × 10 8 7 × 10 8
1000–1100 7 × 10 9 1 × 10 8 1 × 10 8
1500–1600 8 × 10 10 4 × 10 8 2 × 10 10
1900–2000 5 × 10 9 8 × 10 9 3 × 10 9
Table 4. Experimental results of small dataset trained 400 epochs models when evaluated by common file formats.
Table 4. Experimental results of small dataset trained 400 epochs models when evaluated by common file formats.
FormatFormatTPTNFPFNAccuracyPrecisionRecallF1 Score
TrainedTested
VGG-16BMPBMP162355882810.5830.6480.3660.467
JPG161357862820.5840.6520.3630.467
PNG162355882810.5830.6480.3660.467
JPGBMP179349942640.5960.6560.4040.500
JPG174348952690.5890.6470.3930.489
PNG179349942640.5960.6560.4040.500
PNGBMP161352912820.5790.6390.3630.463
JPG154354892890.5730.6340.3480.449
PNG161352912820.5790.6390.3630.463
ResNet-50BMPBMP99394493440.5560.6690.2230.335
JPG100398453430.5620.6900.2250.340
PNG99394493440.5560.6690.2230.335
JPGBMP83408353600.5540.7030.1870.296
JPG81409343620.5530.7040.1830.290
PNG83408353600.5540.7030.1870.296
PNGBMP87385583560.5330.6000.1960.296
JPG83388553600.5320.6010.1870.286
PNG87385583560.5330.6000.1960.296
MobileNet-V2BMPBMP86418253570.5690.7750.1940.310
JPG83419243600.5660.7760.1870.301
PNG86418253570.5690.7750.1940.310
JPGBMP93428153500.5880.8610.2100.337
JPG84425183590.5740.8230.1900.308
PNG93428153500.5880.8610.2100.337
PNGBMP88423203550.5770.8150.1990.319
JPG95424193480.5860.8330.2140.341
PNG88423203550.5770.8150.1990.319
Table 5. Experimental results of large dataset trained 400 epochs models when evaluated by common file formats.
Table 5. Experimental results of large dataset trained 400 epochs models when evaluated by common file formats.
FormatFormatTPTNFPFNAccuracyPrecisionRecallF1 Score
TrainedTested
VGG-16BMPBMP1172701733260.4370.4030.2640.319
JPG1182691743250.4370.4040.2660.321
PNG1172701733260.4370.4030.2640.319
JPGBMP1162751683270.4410.4080.2620.319
JPG1162741693270.4400.4070.2620.319
PNG1162751683270.4410.4080.2620.319
PNGBMP1152721713280.4370.4020.2590.315
JPG1152731703280.4380.4030.2590.315
PNG1152721713280.4370.4020.2590.315
ResNet-50BMPBMP1601902532830.3950.3870.3610.374
JPG1541932502890.3920.3810.3480.364
PNG1601902532830.3950.3870.3610.374
JPGBMP1711842592720.4010.3980.3860.392
JPG1611932502820.3990.3920.3630.377
PNG1711842592720.4010.3980.3860.392
PNGBMP1402501933030.4400.4200.3160.361
JPG1302212223130.3960.3690.2930.327
PNG1402501933030.4400.4200.3160.361
MobileNet-V2BMPBMP1762012422670.4250.4210.3970.409
JPG1871982452560.4340.4330.4220.427
PNG1762012422670.4250.4210.3970.409
JPGBMP1302232203130.3980.3710.2930.328
JPG1272272163160.3990.3700.2870.323
PNG1302232203130.3980.3710.2930.328
PNGBMP1402501933030.4400.4200.3160.361
JPG1522441992910.4470.4330.3430.383
PNG1402501933030.4400.4200.3160.361
Table 6. Experimental results of large dataset trained 2000 epochs models when evaluated by common file formats.
Table 6. Experimental results of large dataset trained 2000 epochs models when evaluated by common file formats.
FormatFormatTPTNFPFNAccuracyPrecisionRecallF1 Score
TrainedTested
VGG-16BMPBMP1112921513320.4550.4240.2500.315
JPG1132881553300.4530.4220.2550.318
PNG1112921513320.4550.4240.2500.315
JPGBMP1152671763280.4310.3950.2590.313
JPG1142731703290.4370.4010.2570.314
PNG1152671763280.4310.3950.2590.313
PNGBMP1122641793310.4240.3850.2530.305
JPG1172651783260.4310.3970.2640.317
PNG1122641793310.4240.3850.2530.305
ResNet-50BMPBMP692891543740.4040.3090.1560.207
JPG643001433790.4110.3090.1440.197
PNG692891543740.4040.3090.1560.207
JPGBMP1801842592630.4110.4100.4060.408
JPG1791842592640.4100.4090.4040.406
PNG1801842592630.4110.4100.4060.408
PNGBMP1062451983370.3960.3490.2390.283
JPG1012461973420.3920.3390.2280.272
PNG1062451983370.3960.3490.2390.283
MobileNet-V2BMPBMP1402092343030.3940.3740.3160.343
JPG1312232203120.3990.3730.2960.330
PNG1402092343030.3940.3740.3160.343
JPGBMP1642252182790.4390.4290.3700.397
JPG1642242192790.4380.4280.3700.397
PNG1642252182790.4390.4290.3700.397
PNGBMP2232012422200.4780.4790.5030.491
JPG2252052382180.4850.4860.5080.497
PNG2232012422200.4780.4790.5030.491
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Loraksa, C.; Mongkolsomlit, S.; Nimsuk, N.; Uscharapong, M.; Kiatisevi, P. Effectiveness of Learning Systems from Common Image File Types to Detect Osteosarcoma Based on Convolutional Neural Networks (CNNs) Models. J. Imaging 2022, 8, 2. https://doi.org/10.3390/jimaging8010002

AMA Style

Loraksa C, Mongkolsomlit S, Nimsuk N, Uscharapong M, Kiatisevi P. Effectiveness of Learning Systems from Common Image File Types to Detect Osteosarcoma Based on Convolutional Neural Networks (CNNs) Models. Journal of Imaging. 2022; 8(1):2. https://doi.org/10.3390/jimaging8010002

Chicago/Turabian Style

Loraksa, Chanunya, Sirima Mongkolsomlit, Nitikarn Nimsuk, Meenut Uscharapong, and Piya Kiatisevi. 2022. "Effectiveness of Learning Systems from Common Image File Types to Detect Osteosarcoma Based on Convolutional Neural Networks (CNNs) Models" Journal of Imaging 8, no. 1: 2. https://doi.org/10.3390/jimaging8010002

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop