Next Article in Journal
Sustained Functioning Impairments and Oxidative Stress with Neurobehavioral Dysfunction Associated with Oral Nicotine Exposure in the Brain of a Murine Model of Ehrlich Ascites Carcinoma: Modifying the Antioxidant Role of Chlorella vulgaris
Previous Article in Journal
Breeding More Crops in Less Time: A Perspective on Speed Breeding
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Synthesis of Microscopic Cell Images Obtained from Bone Marrow Aspirate Smears through Generative Adversarial Networks

1
Department of Computer Engineering, Jeju National University, Jeju 63243, Korea
2
Department of Computer Engineering, Jeju National University, Institute of Information Science & Technology, Jeju 63243, Korea
3
Department of Laboratory Medicine, EONE Laboratories, Incheon 22014, Korea
4
Department of Mechatronics Engineering, Jeju National University, Jeju 63243, Korea
*
Authors to whom correspondence should be addressed.
Biology 2022, 11(2), 276; https://doi.org/10.3390/biology11020276
Submission received: 14 January 2022 / Revised: 26 January 2022 / Accepted: 1 February 2022 / Published: 10 February 2022
(This article belongs to the Section Bioinformatics)

Abstract

:

Simple Summary

This paper proposes a hybrid generative adversarial networks model—WGAN-GP-AC—to generate synthetic microscopic cell images. We generate the synthetic data for the cell types containing fewer data to obtain a balanced dataset. A balanced dataset would help enhance the classification accuracy of each cell type and help with an easy and quick diagnosis that is critical for leukemia patients. In this work, we combine images from three datasets to form a single concrete dataset with variations of multiple microscopic cell images. We provide experimental results that prove the correlation between the original and our synthetically generated data. We also deliver classification results to showcase that the generated synthetic data can be used for real-life experiments and the advancement of the medical domain.

Abstract

Every year approximately 1.24 million people are diagnosed with blood cancer. While the rate increases each year, the availability of data for each kind of blood cancer remains scarce. It is essential to produce enough data for each blood cell type obtained from bone marrow aspirate smears to diagnose rare types of cancer. Generating data would help easy and quick diagnosis, which are the most critical factors in cancer. Generative adversarial networks (GAN) are the latest emerging framework for generating synthetic images and time-series data. This paper takes microscopic cell images, preprocesses them, and uses a hybrid GAN architecture to generate synthetic images of the cell types containing fewer data. We prepared a single dataset with expert intervention by combining images from three different sources. The final dataset consists of 12 cell types and has 33,177 microscopic cell images. We use the discriminator architecture of auxiliary classifier GAN (AC-GAN) and combine it with the Wasserstein GAN with gradient penalty model (WGAN-GP). We name our model as WGAN-GP-AC. The discriminator in our proposed model works to identify real and generated images and classify every image with a cell type. We provide experimental results demonstrating that our proposed model performs better than existing individual and hybrid GAN models in generating microscopic cell images. We use the generated synthetic data with classification models, and the results prove that the classification rate increases significantly. Classification models achieved 0.95 precision and 0.96 recall value for synthetic data, which is higher than the original, augmented, or combined datasets.

1. Introduction

Microscopic images are considered to be an essential and gold standard in the determination and diagnosis of multiple diseases [1]. Bone marrow is a spongy tissue found in the center of the bone containing immature cells called stem cells that produce blood cells. The main task of bone marrow is to produce red blood cells, white blood cells, and platelets [2]. Biopsy of bone marrow includes the collection of bone marrow samples as well as examining the cells, their structure, and counts under the microscope [3]. Bone marrow aspiration is a collection of fluid-containing cells that can be evaluated or examined under a microscope. The slides of stained smears are examined microscopically by specialists, and the cells are evaluated based on the number, appearance, type, etc. Abnormal cellular components indicate the presence of blood diseases. For example, excess granulocytes such as myeloblasts refer to the presence of acute myeloid leukemia [4].
The representation and classification of individual cells have been performed in various histopathology works such as lung cancer, breast cancer, brain cancer, colon cancer, acute leukemia, and many more. Not only classification, but also medical image analysis include the automatic detection of tumors, localization of tumors or cancerous cells, and the early prediction of deadly diseases. Generative adversarial network (GAN) [5] is the emerging framework and has gathered much attention in the medical image analysis domain. Deep learning has been proven to be a powerful tool in modern medical diagnosis and histopathology image analysis [6,7]. GAN has the potential to transform random noise variables into visually realistic images by learning the original data distribution. Initially, the training of Vanilla GAN was unstable. Wasserstein GAN (WGAN) [8], and Wasserstein GAN with gradient penalty (WGAN-GP) [9] has considerably improved the training process of GAN. Conditional GAN (CGAN) was introduced later, where the GAN architecture could be conditioned with special cases. Until this time, the utility of the GAN architecture has increased enormously. Medical image analysis requires abundant data for enhancing the accuracy of machine learning models since, in real-time, there is no room for incorrect results in the medical domain. But on the other hand, it is important to keep patient confidentiality intact. Therefore, there is a need to generate synthetic data that are realistic and maintain the original data distribution.
Hüseyin et al. proposed a framework to automatically detect and classify white blood cells using regional convolutional neural networks [10]. The authors classified five types of white blood cells. The dataset was manually labeled, and variations of R-CNN [11], fast R-CNN [12], and faster R-CNN [13] were used to classify the dataset. Xie et al. proposed a methodology that learns the representation of features and assignment of clusters simultaneously through deep neural networks and calls it deep embedded clustering that is applicable on images [14]. Zhang et al. proposed a probabilistic-based hashing technique for multiple cues of cell-level analysis [15]. In recent work, Ching-Wei Wang et al. introduced a fully automatic bone marrow whole slide image analysis based on deep learning for cell detection and classification [16].
Li Ma et al. proposed a combination of DCGAN [17], and ResNet for blood cell image classification [18]. The authors introduced a new loss function that has improved the discriminative architecture of the GAN model. Shuaizhen et al. developed a weighted feature transfer GAN to synthesize unpaired multi-model medical images accurately [19]. The adversarial loss is combined with the feature map to make meaningful local features of the medical images. Gozes et al. generated synthetic patches of mitosis for enhancing classification of cell images using GAN [20]. In another work, Halicek et al. implemented a conditional GAN for the synthetic generation of hyperspectral cell im-ages of breast cancer obtained from digital histology [21]. DermGAN incorporates the pix2pix architecture to generate synthetic data for clinical skin images [22]. The author also tested the model with a skin condition classifier for classifying malignant conditions. The modified generator GAN MG-GAN was introduced by Poonam et al. to augment medical data for the improvement of cancer classification [23].
In this paper, we propose a hybrid GAN architecture that inherits the supremacy of both WGAN-GP and auxiliary classifier GAN (AC-GAN) [24]. The main contribution of this paper is as follows:
  • We have prepared a new dataset that consists of microscopic cell images obtained from bone marrow aspirate smears collected from three different data sources. The dataset has been prepared with the help of experts from the relevant field;
  • We present a GAN model WGAN-GP-AC that uses the WGAN-GP model combined with the architecture of AC-GAN to generate synthetic microscopic cell images obtained from bone marrow aspirate smears;
  • We use the synthetic data for classification purposes and provide results comparing the performance of classification models using synthetic and original data.

2. Dataset

We collected datasets from three sources and combined them to form a complete dataset for this work. Our main source of data was EONE Laboratories [25] in South Korea. EONE Laboratories provided microscopic images of 17 cell types collected from bone marrow aspirate smears. The dataset contained a total of 12,756 images of individual cells of patients with and without blood diseases. The first dataset contained the following cell types: basophil, eosinophil, neutrophil band, neutrophil segmented, lymphocyte atypical, lymphocyte typical, immature granulocytes, monocytes, erythroblasts, platelets, myelocyte, myeloblast, smudge cells, metamyelocyte, promyelocyte bilobed, promyelocyte and monoblast. Our second data source was from Mendeley data [26] which contained microscopic peripheral blood cell images. The second dataset contained 10,122 individual cell images, which was obtained using the CellaVision DM96 analyzer in the hospital clinic of Barcelona, Core Laboratory. The cell types in this dataset are basophil, eosinophil, neutrophil, lymphocytes, immature granulocytes, monocytes, erythroblasts and platelets. Each image was of size 360 × 360 pixels in JPG format and was annotated by experts in the field. These images were acquired from individuals without any infection or oncologic and hematologic disease. The individuals were also free from any drug consumption for any treatment. Our third data source was the cancer imaging archive [27]. This dataset was prepared by Munich University Hospital and contained 10,300 single cell images labeled by experts and procured from bone marrow aspirate smears. The data were collected from 100 patients diagnosed with acute myeloid leukemia and from 100 patients who had blood disease but were not malignant. The third dataset contained 15 cell types: basophil, eosinophil, erythroblast, smudge cells, lymphocyte atypical, lymphocyte typical, metamyelocyte, monoblast, monocyte, myelocyte, myeloblast, neutrophil band, neutrophil segmented, promyelocyte bilobed, and promyelocyte.
Figure 1 shows sample examples of how each cell types look in microscopic images. In total, the dataset was combined to form 19 cell types.

3. Methodology

In this section, we describe the total implementation process followed to generate synthetic microscopic cell images acquired from bone marrow aspirate smears. As explained in the dataset section, we collected images from three data sources. The first part of the methodology is to collect data and combine the images that belong to the same class. We took the help of experts from EONE Laboratories first to combine the images belonging to the same group. The total number of cell types before preprocessing was eighteen. The cell types were basophil, eosinophil, erythroblast, immature granulocytes, lymphocyte, lymphocyte atypical, lymphocyte typical, metamyelocyte, monoblast, monocyte, myeloblast, myelocyte, neutrophil, neutrophil segmented, neutrophil band, platelet, promyelocyte, promyelocyte bilobed, and smudge cells. Counts per cell type are as mentioned in Table 1.
The overall workflow is presented in Figure 2, where we combine the dataset from three sources and form a dataset with microscopic cell images belonging to 18 cell types. The next phase is data preprocessing which has been done with the help of experts in the medical field. After data preprocessing, we form the final dataset containing twelve classes or cell types.
The dataset is processed through our proposed GAN architecture. For the GAN architecture, we combine the loss function of WGAN-GP and discriminator architecture of auxiliary classifier GAN to form WGAN-GP-AC. The proposed GAN models generate synthetic microscopic individual cell images that machine learning algorithms can further use for classification, detection, and so forth. The generated images are evaluated based on their quality, classification accuracy, and similarity with the original images.

3.1. Data Preprocessing

In the data preprocessing section, redundant data or duplicate images were removed and images were filtered by experts from EONE Laboratories. Firstly, images from three sources belonging to similar cell types were combined. Then, redundant and incorrect data were eliminated through expert reviews. After elimination, images from cell types belonging to the same cell family were merged to form a single group or cell type. For example, lymphocyte, lymphocyte atypical, and lymphocyte typical were combined to form a single cell type, that is, lymphocyte.
Cell type neutrophil was formed by combining images belonging to neutrophil, neutrophil band, and neutrophil segmented. Promyelocyte was formed by merging images of promyelocyte and promyelocyte bilobed. Immature granulocyte contained images from immature granulocyte, myelocyte, and metamyelocyte. After performing all the preprocessing techniques, the total number of cell types was 12 with the following count as mentioned in Table 2.
We have used the stain normalization process so that there would be no bias during the network training. For the classification of each cell type, we have divided the datatset into 70% training, 30% testing and from the training dataset we have used 20% for validation.

3.2. WGAN-GP-AC

Generative adversarial networks (GAN) were introduced by Ian Goodfellow et al. [5] in the year 2014. GAN was proposed as a framework of unsupervised generative models comprising a generator and a discriminator network. GAN is widely used in various fields for different applications.
The generator network G in the GAN architecture is trained to generate synthetic samples, whereas the task of the discriminator D is to discriminate between original and generated samples. This is known as the minmax game where the generator tries to fool the discriminator by producing realistic samples and the discriminator tries to correctly identify real from fake. The generator takes in a random noise variable that retains the data distribution P d over original data distribution x. The generated samples are then passed to the discriminator along with the original samples. The generator tries to minimize l o g ( 1 D ( G ( z ) ) ) with the minmax function defined as in (1):
min G max D f ( D , G ) = E x P d ( x ) [ l o g D ( x ) ] + E z P d ( z ) [ l o g ( 1 D ( G ( z ) ) ) . ]
Generally, the loss function of Vanilla GAN is used to train GAN, which makes the training process harder and the convergence rate slower. The main problem of training with the loss function of Vanilla GAN is that it leads to mode collapse problems. It becomes difficult to assess whether the generator is still in training, has collapsed, or is still converging. Therefore, in this work, we follow the work of Wasserstein GAN with gradient penalty (WGAN-GP) [3]. The Wasserstein GAN [2] solves the problems faced by Vanilla GAN by optimizing the Earth Mover distance, also known as the Wasserstein-1 distance, instead of the Jensen-Shanon divergence (JSD) that is used in the Vanilla GAN. WGAN builds a powerful discriminator that would generate a significant gradient even if the generator performs poorly. The Wasserstein-1 distance can be defined as in (2) with D being the k-Lipschitz function.
min G max D E x P d ( x ) [ D ( x ) ] + E z P z [ D ( G ( z ) ) . ]
WGAN requires critic weight-clipping in a compact space [ c , c ] where the critic is considered the optimal discriminator. WGAN-GP is an improved version of WGAN proposed by Gulrajani et al. that replaces the weight clipping mechanism required in WGAN and complies with the condition of 1-Lipschitz by introducing the gradient penalty. The objective of WGAN-GP is a combination of the original critic loss and the gradient penalty as shown in (3) and (4):
O r i g i n a l C r i t i c L o s s = E x ˜ P d [ D ( x ˜ ) ] E x P r [ D ( G ( x ) ) ]
G r a d i e n t P e n a l t y = σ E x ^ P x ^ [ ( x ^ D ( x ^ ) 2 1 ) 2 ] ,
where x ^ P x ^ defines random samples and P d and P r are data distribution and generator distribution. σ has been selected as 10 according to our data and experiments In our proposed work, the generator receives input noise vector and category or class labels as inputs, which are passed through the dense and activation layers. The images and the labels are then reshaped. After reshaping, it is concatenated and passed through residual upsampling blocks. After which, we perform batch normalization. The synthetic data are generated through Conv2D and tanh activation as shown in Figure 3.
For the discriminator, we use the concept of the auxiliary classifier GAN, which is different from the conditional GAN in the sense that the discriminator in the ACGAN does not receive class information. The task of the discriminator is to output the probability distribution of the image source (real or generated) and to output the probability distribution over the class labels for the particular image that is done by the auxiliary classifier in the discriminator architecture. Therefore, the objective function for the auxiliary classifier GAN is the combination of S L and C L , where S L is the correct source log-likelihood and C L is the log-likelihood of the correct cell type. S L and C L are defined as in (5) and (6):
S L = E [ log P ( S = r e a l | x r e a l ) ] + E [ l o g P ( S = s y n t h e t i c | x s y n t h e t i c ) ]
C L = E [ log P ( C = c | x r e a l ) ] + E [ l o g P ( C = c | x s y n t h e t i c ) ] .
Figure 4 describes the architecture of the discriminator of WGAN-GP-AC. As we can see from Figure 4, the discriminator is provided only with the generated images, but not with the class labels. The images pass through the 2D convolutional layer and residual downsampling blocks. The discriminator tries to maximize the probability of correctly classifying real and synthetic images and correctly predicting the cell types or class labels.
This way, the GAN model generates more realistic images that can be used for real-time classification models or any other machine learning algorithms. Figure 5 and Figure 6 present the residual block architecture for upsampling and downsampling.

4. Experiments and Results

This section defines the evaluation metrics we used to evaluate our model’s performance in generating synthetic microscopic cell images. The evaluation is done in three stages. We first measure how combining the auxiliary classifier’s concept with WGAN-GP performs compared to other existing GAN models. We also compute the training accuracy and loss of the proposed model and see the quality of the generated images. We used a learning rate of 0.001, batch size 64 and patch size as 128 × 128. We provide sample examples of generated images through the proposed WGAN-GP-AC model. We present the error rates of different architectures of GAN on our prepared dataset. Lastly, we use original and synthetic data separately to measure how classification models work. We compare the performance of pre-trained models on original and synthetic data.
The GAN architectures that we use to compare our modelâs performance are auxiliary classifier GAN (AC-GAN) [24], Wasserstein GAN (WGAN) [8], WGAN with gradient penalty (WGAN-GP) [9], information maximizing GAN (InfoGAN) [28], WGAN-GP-Info [29], deep convolutional GAN (DCGAN) [17] and conditional GAN (CGAN) [30].
In Table 3, we present a quantitative comparison of various models on our dataset. The evaluation metrics which we chose are inception score (IS), learned perceptual image patch similarity (LPIPS), recall, precision, F1 score, and Fréchet inception distance (FID). Through the inception score, we compute how realistic the generated images are. The formulation for the inception score is shown as in (7).
e ( E x [ K L D ( p ( y | x ) p ( y ) ) ] ) ,
where K L D is the Kullback–Leibler divergence measuring the difference between the marginal distribution p ( y ) and probability distribution of image × denoted by p ( y | x ) . The LPIPS metric is used to measure the variance in the synthetic samples generated by WGAN-GP-AC. Whereas FID computed the distance of the feature vector between real and synthetic images. It compares the distribution of the synthetic images along with the distribution of the real images used during training the generator. The lower the FID value, the higher the quality of the generated image. The following Equation (8) can be used to compute the FID between real images R and synthetic images S:
F I D = μ R μ S 2 + T R ( Σ R + Σ S 2 ( Σ R Σ S ) 1 / 2 .
Given the real and synthetic distribution, precision measures the quality of generated samples and indicates how accurately the auxiliary classifier is predicting the classes, whereas recall measures the quantity. The higher the value for precision and recall, the better the model’s performance. F1, on the other hand, is the harmonic mean between the precision and the recall and contributes to the measurement of the model’s accuracy on a particular dataset. The experiment result shows that our proposed model WGAN-GP-AC performs better than the other mentioned existing models and improves the quality of multiclass image generation as compared to WGAN-GP and AC-GAN individually.
Figure 7 and Figure 8 show the training and validation accuracy and loss. As can be seen from the figures, our proposed WGAN-GP-AC model produces a consistent performance after 25 epochs. Our model is trained for 50 epochs with 265 iterations per epoch. As can be seen, the model requires less computational overhead to perform significantly better. It achieves a training accuracy of 97.54% and a validation accuracy of 97.32%. Training loss for our model was 0.0692, with a validation loss of 0.0917.
Table 4 presents the comparison of error rates (l1 and l2), peak signal-to-noise ratio (PSNR), and structural similarity index metric (SSIM). The loss functions l1 and l2 are important evaluation metrics for measuring the error rates.
The least absolute deviation is measured by l1 and can be computed to minimize error, which is defined as the sum of all the absolute differences between original and synthetic data as shown in (9):
l 1 = Σ i = 1 n | y r e a l y s y n t h e t i c |
l 2 = Σ i = 1 n ( y r e a l y s y n t h e t i c ) 2
l2, as defined in (10), is another loss function used in GAN to measure error, that is, the squared differences between real and synthetic data. We measure both l1 and l2 errors to test our model with outliers, if any. l1 and l2 are the measure for reconstruction error between the synthetic and the real images. PSNR and SSIM are image quality measures. PSNR computes the peak signal-to-noise ratio between the real and synthetic images. The higher the PSNR, the better is the quality of the synthetic image. Mean squared error (MSE) is the cumulative squared error between the generated and the real image, whereas PSNR computes the peak error. PSNR can be calculated from Equations (11) and (12) where M and N are total rows and columns in the images and R is the maximum possible pixel value of the image.Results show that our proposed model generates less error as compared to other models. WGAN-GP-AC also produces a higher PSNR value and a high structural similarity index, which indicates that the synthetic image generated by our proposed model is of greater quality than other existing models. A crucial aspect of generating synthetic data is to evaluate whether machine learning models can use it for real-life experiments such as classification.
MSE = M , N r e a l ( m , n ) s y n t h e t i c ( m , n ) 2 M * N
PSNR = 10 log 10 R 2 M S E
SSIM uses three quantities, that is, luminance (L), contrast (C), and structure (S) to measure the corresponding pixels and their neighbors in real and synthetic images. L, C and S can be defined as in Equations (13)–(15), where μ and σ denotes mean and standard deviation and C 1 , C 2 and C 3 are constants included for numerical stability. SSIM can be defined as in (16).
L ( r e a l , s y n t h e t i c ) = 2 μ r e a l μ s y n t h e t i c + C 1 μ r e a l 2 + μ s y n t h e t i c 2 + C 1
C ( r e a l , s y n t h e t i c ) = 2 σ r e a l σ s y n t h e t i c + C 2 σ r e a l 2 + σ s y n t h e t i c 2 + C 2
S ( r e a l , s y n t h e t i c ) = σ r e a l s y n t h e t i c + C 3 σ r e a l σ s y n t h e t i c + C 3
SSIM ( r e a l , s y n t h e t i c ) = L ( r e a l , s y n t h e t i c ) * C ( r e a l , s y n t h e t i c ) * S ( r e a l , s y n t h e t i c ) .
In Table 5 and Table 6, we have presented the performance of transfer learning models such as InceptionV3 [31], ResNet [32], VGG16 [33], CNN [34], Xception [35] and VGG19 [36] to measure how accurately the classifiers can classify the cell types after training it with original and synthetic data separately. We have used precision and recall as evaluation metrics. For precision, recall and F1, we have used the computational method as mentioned in [37,38]. The results show that the accuracy of the models is enhanced when synthetic data generated by WGAN-GP-AC are used to train the model.
We have also compared the performance of classification models using original, synthetic, augmented data and a combination of original and synthetic data. For augmentation, we have used three different combinations of augmentation techniques. Augmentation-1 uses scaling, rotation, and color augmentation techniques; Augmentation-2 uses translation, contrast, and scaling methods; and Augmentation-3 uses saturation, scaling, and rotation. We have kept the total number of images for each cases same so that there is a fair comparison.
We have first trained the classification models with original data and also tested the models with original data. The training and testing ratios were 70% and 30%. For validation, we used 20% of the training data. We then followed the same process to get the performance of the classification models with synthetic data, augmented data and mix of original and synthetic data. As can be seen from all the results, the images generated by our proposed WGAN-GP-AC model are of better quality and it performs better, quantitatively, than other GAN models.
In Figure 9 and Figure 10, we have shown each cell type’s accuracy, specificity, and sensitivity for the InceptionV3 classification model using the original and synthetic datasets separately. We present examples of the generated synthetic microscopic images for each cell type in Figure 11.

5. Conclusions and Discussion

In this work, we propose a hybrid GAN architecture that implements the concept of auxiliary classifier GAN in Wasserstein GAN with gradient penalty. We use the loss function of WGAN-GP but implement the discriminator as in the AC-GAN. The discriminators’ task in this work is not only to identify real or fake but also to assign or classify classes, that is, cell types of each image. This architecture helps generate synthetic microscopic bone marrow aspirate smears cell images that can be used for multiclass classification in real life. It enhances the classification results by oversampling the minority classes (i.e., cell types with fewer images) and balancing the dataset.
We first take microscopic images of 19 cell types obtained from bone marrow aspirate smears, which are then preprocessed. In our work, we collaborated with experts to filter data, rearrange images to proper classes if they were marked with wrong cell types, we merged cell types belonging to the same cell family. After preprocessing, the cell types were reduced to 12. Every image of each cell type is processed through our proposed WGAN-GP-AC model. The generator produces synthetic data, and the discriminator evaluates the images as real or fake and tries to classify the cell type for every image. The generated images are evaluated through various evaluation metrics. Our GAN model obtains training and validation accuracy of 97.54% and 97.32%. The quantitative result shows that our model generates less error, has more structural similarity with the original data, and produces better quality images for every cell type as compared to other GAN models. We have included the results for different augmentation techniques and the dataset prepared through a combination of original and synthetic data. We used the same number of images for synthetic, augmented, and a combination of the original and synthetic datasets to evaluate the classification models. The result shows that the classification models perform better for the synthetic dataset generated by our proposed model. We provide accuracy, specificity, and sensitivity scores for each cell type in this work. We provide the results for the classification of each cell type using the original and synthetic datasets separately. We also experiment our data with different classifiers, which indicates that the accuracy of classification models increases while using synthetic data generated from WGAN-GP-AC. In the future, we would like to implement a balancing mechanism that would oversample minority classes or cell types with less images through GAN and undersample majority classes through image similarity measures so that the classification accuracy could enhance their performance. We also plan to generate images with multiple cell types in a single image so that they can be used for disease diagnosis, early detection, and other medical reasons.

Author Contributions

Conceptualization, D.H.; Formal analysis, D.H.; Formal analysis, W.J.K.; Data curation, W.J.K.; Validation, W.J.K.; Supervision, W.J.K.; Funding acquisition, C.-U.K.; Project administration, C.-U.K.; Methodology, D.H.; Writing—review and editing, D.H.; Writing—original draft, D.H.; Visualization, D.H.; Investigation, Y.-C.B.; Resources, Y.-C.B.; Project administration, Y.-C.B.; Supervision, Y.-C.B. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the 2020 scientific promotion program funded by Jeju National University.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest regarding the design of this study, analyses and writing of this manuscript.

References

  1. Gurcan, M.N.; Boucheron, L.E.; Can, A.; Madabhushi, A.; Rajpoot, N.M.; Yener, B. Histopathological image analysis: A review. IEEE Rev. Biomed. Eng. 2009, 2, 147–171. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Cavallo, C.; Boffa, A.; Andriolo, L.; Silva, S.; Grigolo, B.; Zaffagnini, S.; Filardo, G. Bone marrow concentrate injections for the treatment of osteoarthritis: Evidence from preclinical findings to the clinical application. Int. Orthop. 2021, 45, 525–538. [Google Scholar] [CrossRef]
  3. Tomasian, A.; Jennings, J.W. Bone marrow aspiration and biopsy: Techniques and practice implications. Skelet. Radiol. 2021, 51, 81–88. [Google Scholar] [CrossRef]
  4. Hasserjian, R.P. Controversies in the recent (2016) World Health Organization classification of acute myeloid leukemia. Best Pract. Res. Clin. Haematol. 2021, 34, 101249. [Google Scholar] [CrossRef]
  5. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  6. Baxi, V.; Edwards, R.; Montalto, M.; Saha, S. Digital pathology and artificial intelligence in translational medicine and clinical practice. Mod. Pathol. 2021, 35, 23–32. [Google Scholar] [CrossRef]
  7. Lee, K.; Lockhart, J.H.; Xie, M.; Chaudhary, R.; Slebos, R.J.; Flores, E.R.; Chung, C.H.; Tan, A.C. Deep Learning of Histopathology Images at the Single Cell Level. Front. Artif. Intell. 2021, 4, 137. [Google Scholar] [CrossRef] [PubMed]
  8. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  9. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved training of wasserstein gans. arXiv 2017, arXiv:1704.00028. [Google Scholar]
  10. Kutlu, H.; Avci, E.; Özyurt, F. White blood cells detection and classification based on regional convolutional neural networks. Med. Hypotheses 2020, 135, 109472. [Google Scholar] [CrossRef]
  11. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  12. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Washington, DC, USA, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  13. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [Green Version]
  14. Xie, J.; Girshick, R.; Farhadi, A. Unsupervised deep embedding for clustering analysis. In Proceedings of the International Conference on Machine Learning, New York, NY, USA, 19–24 June 2016; pp. 478–487. [Google Scholar]
  15. Zhang, X.; Su, H.; Yang, L.; Zhang, S. Weighted hashing with multiple cues for cell-level analysis of histopathological images. In Proceedings of the International Conference on Information Processing in Medical Imaging, Isle of Skye, UK, 28 June–3 July 2015; pp. 303–314. [Google Scholar]
  16. Wang, C.W.; Huang, S.C.; Lee, Y.C.; Shen, Y.J.; Meng, S.I.; Gaol, J.L. Deep learning for bone marrow cell detection and classification on whole-slide images. Med. Image Anal. 2022, 75, 102270. [Google Scholar] [CrossRef] [PubMed]
  17. Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  18. Ma, L.; Shuai, R.; Ran, X.; Liu, W.; Ye, C. Combining DC-GAN with ResNet for blood cell image classification. Med. Biol. Eng. Comput. 2020, 58, 1251–1264. [Google Scholar] [CrossRef] [PubMed]
  19. Yao, S.; Tan, J.; Chen, Y.; Gu, Y. A weighted feature transfer gan for medical image synthesis. Mach. Vis. Appl. 2021, 32, 1–11. [Google Scholar] [CrossRef]
  20. Gozes, G.; Shkolyar, A.; Gefen, A.; Benayahu, D.; Greenspan, H. Harnessing generative adversarial networks to generate synthetic mitosis images for classification of cell images. In Proceedings of the Medical Imaging 2021, Digital Pathology, Online, 15–19 February 2021; International Society for Optics and Photonics: Washington, DC, USA, 2021; Volume 11603, p. 1160309. [Google Scholar]
  21. Halicek, M.; Ortega, S.; Fabelo, H.; Lopez, C.; Lejeune, M.; Callico, G.M.; Fei, B. Conditional generative adversarial network for synthesizing hyperspectral images of breast cancer cells from digitized histology. In Proceedings of the Medical Imaging 2020, Digital Pathology, Houston, TX, USA, 15–20 February 2020; International Society for Optics and Photonics: Washington, DC, USA, 2020; Volume 11320, p. 113200U. [Google Scholar]
  22. Ghorbani, A.; Natarajan, V.; Coz, D.; Liu, Y. Dermgan: Synthetic generation of clinical skin images with pathology. In Proceedings of the Machine Learning for Health Workshop, Vancouver, BC, Canada, 13 December 2019; pp. 155–170. [Google Scholar]
  23. Chaudhari, P.; Agrawal, H.; Kotecha, K. Data augmentation using MG-GAN for improved cancer classification on gene expression data. Soft Comput. 2020, 24, 11381–11391. [Google Scholar] [CrossRef]
  24. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier gans. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 2642–2651. [Google Scholar]
  25. Eonelab.co.kr, Molecular Diagnostic Research Center. 1983. Available online: https://www.eonelab.co.kr/global/en/main/main.asp (accessed on 15 October 2021).
  26. Acevedo, A.; Merino, A.; Alférez, S.; Molina, Á.; Boldú, L.; Rodellar, J. A dataset of microscopic peripheral blood cell images for development of automatic recognition systems. Data Brief 2020, 30, 105474. [Google Scholar] [CrossRef]
  27. Matek, C.; Schwarz, S.; Marr, C.; Spiekermann, K. A single-cell morphological dataset of leukocytes from AML patients and non-malignant controls (AML-Cytomorphology_LMU). The Cancer Imaging Archive (TCIA) [Internet]. 2019. Available online: https://wiki.cancerimagingarchive.net/pages/viewpage.action (accessed on 29 October 2019).
  28. Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. Infogan: Interpretable representation learning by information maximizing generative adversarial nets. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016; pp. 2180–2188. [Google Scholar]
  29. Hu, B.; Tang, Y.; Eric, I.; Chang, C.; Fan, Y.; Lai, M.; Xu, Y. Unsupervised learning for cell-level visual representation in histopathology images with generative adversarial networks. IEEE J. Biomed. Health Inform. 2018, 23, 1316–1328. [Google Scholar] [CrossRef] [Green Version]
  30. Mirza, M.; Osindero, S. Conditional generative adversarial nets. arXiv 2014, arXiv:1411.1784. [Google Scholar]
  31. Xia, X.; Xu, C.; Nan, B. Inception-v3 for flower classification. In Proceedings of the 2017 2nd International Conference on Image, Vision and Computing (ICIVC), Chengdu, China, 2–4 June 2017; pp. 783–787. [Google Scholar]
  32. Liang, J. Image classification based on RESNET. J. Phys. Conf. Ser. 2020, 1634, 012110. [Google Scholar] [CrossRef]
  33. Kaur, T.; Gandhi, T.K. Automated brain image classification based on VGG-16 and transfer learning. In Proceedings of the 2019 International Conference on Information Technology (ICIT), Bhubaneswar, India, 19–21 December 2019; pp. 94–98. [Google Scholar]
  34. Sharma, N.; Jain, V.; Mishra, A. An analysis of convolutional neural networks for image classification. Procedia Comput. Sci. 2018, 132, 377–384. [Google Scholar] [CrossRef]
  35. Wu, X.; Liu, R.; Yang, H.; Chen, Z. An Xception Based Convolutional Neural Network for Scene Image Classification with Transfer Learning. In Proceedings of the 2020 2nd International Conference on Information Technology and Computer Application (ITCA), Guangzhou, China, 18–20 December 2020; pp. 262–267. [Google Scholar]
  36. Bansal, M.; Kumar, M.; Sachdeva, M.; Mittal, A. Transfer learning for image classification using VGG19: Caltech-101 image data set. J. Ambient Intell. Hum. Comput. 2021, 12, 1–12. [Google Scholar] [CrossRef] [PubMed]
  37. Lucic, M.; Kurach, K.; Michalski, M.; Gelly, S.; Bousquet, O. Are gans created equal? a large-scale study. arXiv 2017, arXiv:1711.10337. [Google Scholar]
  38. Kynkäänniemi, T.; Karras, T.; Laine, S.; Lehtinen, J.; Aila, T. Improved precision and recall metric for assessing generative models. arXiv 2019, arXiv:1904.06991. [Google Scholar]
Figure 1. Sample example of types of microscopic cell images in the dataset.
Figure 1. Sample example of types of microscopic cell images in the dataset.
Biology 11 00276 g001
Figure 2. Overview of the proposed methodology to generate synthetic microscopic cell images obtained from bone marrow aspirate smears through generative adversarial networks.
Figure 2. Overview of the proposed methodology to generate synthetic microscopic cell images obtained from bone marrow aspirate smears through generative adversarial networks.
Biology 11 00276 g002
Figure 3. Generator architecture for the proposed GAN model.
Figure 3. Generator architecture for the proposed GAN model.
Biology 11 00276 g003
Figure 4. Discriminator architecture for the proposed GAN model.
Figure 4. Discriminator architecture for the proposed GAN model.
Biology 11 00276 g004
Figure 5. Illustration of residual block (upsampling) in the proposed architecture.
Figure 5. Illustration of residual block (upsampling) in the proposed architecture.
Biology 11 00276 g005
Figure 6. Illustration of residual block (downsampling) in the proposed architecture.
Figure 6. Illustration of residual block (downsampling) in the proposed architecture.
Biology 11 00276 g006
Figure 7. Training and validation accuracy of WGAN-GP-AC.
Figure 7. Training and validation accuracy of WGAN-GP-AC.
Biology 11 00276 g007
Figure 8. Training and validation loss of WGAN-GP-AC.
Figure 8. Training and validation loss of WGAN-GP-AC.
Biology 11 00276 g008
Figure 9. Accuracy, Specificity and Sensitivity of Individual Cell Types for Original Dataset.
Figure 9. Accuracy, Specificity and Sensitivity of Individual Cell Types for Original Dataset.
Biology 11 00276 g009
Figure 10. Accuracy, Specificity and Sensitivity of Individual Cell Types for Synthetic Dataset.
Figure 10. Accuracy, Specificity and Sensitivity of Individual Cell Types for Synthetic Dataset.
Biology 11 00276 g010
Figure 11. Sample images generated for each cell type by WGAN-GP-AC.
Figure 11. Sample images generated for each cell type by WGAN-GP-AC.
Biology 11 00276 g011
Table 1. Number of images per cell type before preprocessing.
Table 1. Number of images per cell type before preprocessing.
Cell TypeNumber of ImagesDataset-1Dataset-2Dataset-3
Basophil1224570420234
Eosinophil3538106113561121
Erythroblast1547540500507
Immature Grannulocytes288112661615N/A
Lymphocyte1213N/A1213N/A
Lymphocyte Atypical74N/A3
Lymphocyte Typical38181790N/A2028
Metamyelocyte138N/A5
Monoblast2614N/A12
Monocyte25839121013658
Myeloblast31041246N/A1858
Myelocyte3922N/A17
Neutrophil3316N/A3316N/A
Neutrophil Band8242N/A40
Neutrophil Segmented73463588N/A3758
Platelet23391650689N/A
Promyelocyte6926N/A43
Promyelocyte Bilobed1810N/A8
Smudge Cells157N/A8
Table 2. Number of images per cell type after preprocessing.
Table 2. Number of images per cell type after preprocessing.
Cell TypeNumber of Images
Basophil1224
Eosinophil3538
Erythroblast1547
Immature Grannulocytes2933
Lymphocyte5038
Monoblast26
Monocyte2583
Myeloblast3104
Neutrophil10,743
Platelet2339
Promyelocyte87
Smudge Cells15
Table 3. Quantitative comparison on our dataset.
Table 3. Quantitative comparison on our dataset.
ModelsISFIDLPIPSPrecisionRecallF1
AC-GAN8.34 ± 0.8976.30.3494.3794.0194.13
WGAN9.67 ± 0.2572.30.3194.5894.5793.06
WGAN-GP10.06 ± 0.0371.10.2996.7295.3895.11
InfoGAN9.12 ± 0.3773.90.3294.0194.8394.92
WGAN-GP-Info9.94 ± 0.7173.10.3394.4995.0394.02
DCGAN9.89 ± 0.2873.40.3195.6694.9195.50
CGAN9.01 ± 0.7775.20.3493.0193.4892.99
WGAN-GP-AC12.36 ± 0.4167.20.2596.8396.0996.32
Table 4. l1 error, l2 error, PSNR and SSIM comparison of different GAN models on our dataset.
Table 4. l1 error, l2 error, PSNR and SSIM comparison of different GAN models on our dataset.
Modelsl1 Errorl2 ErrorPSNRSSIM
AC-GAN13.9%6.3%30.730.8762
WGAN12.8%5.1%32.610.9135
WGAN-GP12.6%5.4%31.420.9172
InfoGAN11.9%5.7%31.670.9288
WGAN-GP-Info12.3%5.2%31.890.9061
DCGAN12.7%6.1%32.770.9258
CGAN14.3%6.7%32.330.9378
WGAN-GP-AC9.8%4.2%36.710.9616
Table 5. Precision and recall value for different classification models using original and synthetic datasets.
Table 5. Precision and recall value for different classification models using original and synthetic datasets.
Classification
Models
Original Data
Precision
RecallSynthetic Data
Precision
Recall
InceptionV30.930.920.950.96
ResNet0.870.890.90.91
VGG160.930.90.940.93
CNN0.860.880.890.91
Xception0.880.890.920.92
VGG190.910.910.940.96
Table 6. Precision and recall value for different classification models using augmentation methods and combination of original and synthetic data.
Table 6. Precision and recall value for different classification models using augmentation methods and combination of original and synthetic data.
ClassificationAugmention-1Augmentation-2Augmentation-3Original + Synthetic
ModelsPrecisionRecallPrecisionRecallPrecisionRecallPrecisionRecall
InceptionV30.940.930.930.920.940.940.930.94
ResNet0.880.870.890.880.860.850.890.90
VGG160.910.920.900.890.900.900.920.90
CNN0.870.880.860.850.850.840.870.89
Xception0.890.870.870.860.900.890.890.90
VGG190.900.880.890.880.900.910.920.93
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hazra, D.; Byun, Y.-C.; Kim, W.J.; Kang, C.-U. Synthesis of Microscopic Cell Images Obtained from Bone Marrow Aspirate Smears through Generative Adversarial Networks. Biology 2022, 11, 276. https://doi.org/10.3390/biology11020276

AMA Style

Hazra D, Byun Y-C, Kim WJ, Kang C-U. Synthesis of Microscopic Cell Images Obtained from Bone Marrow Aspirate Smears through Generative Adversarial Networks. Biology. 2022; 11(2):276. https://doi.org/10.3390/biology11020276

Chicago/Turabian Style

Hazra, Debapriya, Yung-Cheol Byun, Woo Jin Kim, and Chul-Ung Kang. 2022. "Synthesis of Microscopic Cell Images Obtained from Bone Marrow Aspirate Smears through Generative Adversarial Networks" Biology 11, no. 2: 276. https://doi.org/10.3390/biology11020276

APA Style

Hazra, D., Byun, Y. -C., Kim, W. J., & Kang, C. -U. (2022). Synthesis of Microscopic Cell Images Obtained from Bone Marrow Aspirate Smears through Generative Adversarial Networks. Biology, 11(2), 276. https://doi.org/10.3390/biology11020276

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop