Next Article in Journal
Development of End-to-End Artificial Intelligence Models for Surgical Planning in Transforaminal Lumbar Interbody Fusion
Previous Article in Journal
Full-Field Strain Measurements of the Muscle-Tendon Junction Using X-ray Computed Tomography and Digital Volume Correlation
Previous Article in Special Issue
Prediction of Gait Kinematics and Kinetics: A Systematic Review of EMG and EEG Signal Use and Their Contribution to Prediction Accuracy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Volumetric Imitation Generative Adversarial Networks for Anatomical Human Body Modeling

Department of Electrical and Computer Engineering, Inha University, Incheon 22212, Republic of Korea
*
Author to whom correspondence should be addressed.
Bioengineering 2024, 11(2), 163; https://doi.org/10.3390/bioengineering11020163
Submission received: 9 January 2024 / Revised: 2 February 2024 / Accepted: 6 February 2024 / Published: 7 February 2024
(This article belongs to the Special Issue Intelligent Computer-Aided Designs for Biomedical Applications)

Abstract

:
Volumetric representation is a technique used to express 3D objects in various fields, such as medical applications. On the other hand, tomography images for reconstructing volumetric data have limited utilization because they contain personal information. Existing GAN-based medical image generation techniques can produce virtual tomographic images for volume reconstruction while preserving the patient’s privacy. Nevertheless, these images often do not consider vertical correlations between the adjacent slices, leading to erroneous results in 3D reconstruction. Furthermore, while volume generation techniques have been introduced, they often focus on surface modeling, making it challenging to represent the internal anatomical features accurately. This paper proposes volumetric imitation GAN (VI-GAN), which imitates a human anatomical model to generate volumetric data. The primary goal of this model is to capture the attributes and 3D structure, including the external shape, internal slices, and the relationship between the vertical slices of the human anatomical model. The proposed network consists of a generator for feature extraction and up-sampling based on a 3D U-Net and ResNet structure and a 3D-convolution-based LFFB (local feature fusion block). In addition, a discriminator utilizes 3D convolution to evaluate the authenticity of the generated volume compared to the ground truth. VI-GAN also devises reconstruction loss, including feature and similarity losses, to converge the generated volumetric data into a human anatomical model. In this experiment, the CT data of 234 people were used to assess the reliability of the results. When using volume evaluation metrics to measure similarity, VI-GAN generated a volume that realistically represented the human anatomical model compared to existing volume generation methods.

Graphical Abstract

1. Introduction

Volumetric representation [1,2] is a popular technique to express 3D objects, such as surface modeling [3,4]. Volumetric data are generated mainly by reconstructing from tomographic images, such as computed tomography (CT) and magnetic resonance imaging (MRI) [5,6]. On the other hand, it is challenging to adopt tomographic images for purposes other than medical, such as diagnosis and surgery, because they contain personal information. GAN-based medical image generation techniques [7,8] can produce anatomically meaningful virtual tomographic images applicable to volume reconstruction. On the other hand, the image generation process does not consider the relationship between the adjacent slices of a volume because these methods account for 2D correlations exclusively. This can lead to erroneous results when reconstructing volumes from generated images, particularly when maintaining the 3D structural coherence between adjacent slices. While techniques have been proposed to generate 3D volumes [9,10], these approaches have been limited to generating surface representations and have failed to capture the internal characteristics. Consequently, a volume generation method is needed to reflect the entire 3D human anatomical model, including its internal portion.
This paper introduces the volumetric imitation GAN (VI-GAN), a novel approach that aims to imitate human anatomical models to generate volumetric data. The primary goal of this approach is to generate 3D models that faithfully capture the attributes and 3D structure (external shape, internal slice, and the relationship between vertical slices) within the human anatomical model. The proposed network comprises two main components: a generator to obtain the volumetric data and a discriminator to evaluate the authenticity between the generated volume and the ground truth. The generator performs feature extraction and up-sampling to produce a volume based on the 3D U-Net [11] and ResNet [12] structures. Moreover, the initial feature extraction process from the input image set uses a 3D-convolution-based LFFB (local feature fusion block) [13] to incorporate features at various scales. On the other hand, the discriminator uses 3D convolution to extract the authenticity of the generated volume and ground truth. Using the proposed network structure makes it possible to account for vertical correlations in the volume generation process compared to existing 2D image generation techniques. Moreover, it provides a more realistic representation compared to previous methods focused solely on surface generation because it can faithfully intimate the internal features of the anatomical model. The volume comprises 3D data with higher dimensions than the image, so converging in a specific shape is difficult. Therefore, if the basic distance loss alone is applied to generate the volume, it barely converges in the form of a human anatomical model [14]. Thus, the proposed method devises a reconstruction loss so VI-GAN can generate the volume converging to the human anatomical model. Reconstruction loss includes feature loss and similarity loss. Feature loss is calculated using an overlapping region between the generated volume and the ground truth. Similarity loss is also calculated as an internal similarity based on a structural similarity index map (SSIM) [15].
The spine data from the Digital Korean dataset [16] provided by the Korean Institute of Science and Technology Information (KISTI) and the liver data from CT volumes with multiple organ segmentations (CT-ORG) [17] were applied during the experiment to validate the proposed technique. Evaluation metrics, such as F1-score, dice coefficient, peak signal-to-noise ratio (PSNR), and universal image quality index (UQI), were used to measure the resemblance between the generated volume and the human anatomical model. The VI-GAN outperformed existing methods by producing volumes closely representing the human anatomical model.
The volumetric data generated by the VI-GAN included the external shape and internal structure of the human anatomical model. Therefore, it can be used in various fields, such as diagnosis [18,19] and surgical simulation [20,21]. VI-GAN can produce a 3D human anatomical model that can enhance training efficiency and immersion for medical professionals. Moreover, virtual tomographic images can be produced by decomposing the volumetric data generated by the VI-GAN. Compared to existing medical image generation methods, these images have fewer errors in the relationship between neighboring slices. The tomographic images produced by the VI-GAN can enhance the capabilities of medical professionals to distinguish diseases effectively during the diagnostic process.
The contributions of this paper are as follows. (1) This paper proposes a VI-GAN to create 3D volumetric data that capture the attributes and 3D structure of a human anatomical model (external shape, internal slice, and vertical slice relationships). (2) This paper introduces reconstruction loss that encompasses feature loss and similarity loss to enhance the convergence rate of VI-GAN in volume generation. The feature loss measures the overlapping regions, while the similarity loss quantifies the resemblance between the generated volume and ground truth.

2. Related Works

Several studies have proposed generating the surface of a human anatomical model in the volume format from tomographic images. Balashova et al. [9] proposed a method to reconstruct the surface of the liver using a single X-ray image. They used mask data and images in the training process to generate liver data closer to the ground truth. Henzler et al. [22] generated the surface of animal bones using multiple X-ray images. Their study combined the volumes generated from images of multiple viewpoints. It allowed the production of a high-quality volume with a complete reconstruction of the parts that could be obscured easily from a single viewpoint. Kasten et al. [10] devised a network to reconstruct knee bones from bi-planar X-ray images. Their study synthesized the volumes produced by duplicating axial, coronal, and sagittal images in the z -axis direction in the training process. On the other hand, the generated data could not preserve the characteristics of the entire portion of that model because these methods generated only the surface of the human anatomical model. Furthermore, many of these studies have been proposed in the form of CNNs. The proposed technique requires generating virtual volume data. Hence, applying a GAN specialized for generation is necessary.
Several studies have proposed producing 2D medical data using GANs. These studies have focused mainly on the synthesis and reconstruction of images. Synthesis techniques include changing the style and modality [23,24] and adding characteristics, such as nodules and tumors [25]. Reconstruction techniques cover the super-resolution [26] process. Among these techniques, medical image generation [7,27] can produce the virtual tomographic images required for volume reconstruction. Chuquicusma et al. [28] and Frid-Adar et al. [29] proposed techniques for generating images representing lung modules and liver lesions using a deep convolution GAN (DCGAN). Beer et al. [8] devised a method for generating tomographic images through the progressive growing of GAN (PGGAN) [30] to express skin lesions realistically. These methods generated tomographic images to improve the classification and segmentation performance during the training process. Nevertheless, these medical image generation techniques considered only the 2D correlation within the image. Therefore, erroneous results can be obtained when reconstructing volumes from these generated images because the vertical correlation with the adjacent slices is not considered. Hence, the volume generation technique using a GAN should be proposed to prevent such erroneous results.
Some studies adopted the GAN structure for volume generation by expanding various image generation techniques into 3D space. Wu et al. [31] proposed 3D-GAN and 3D variational autoencoder GAN (3D-VAE-GAN) structures to produce the surface of an object as a volume using a single 2D image. Smith et al. [32] applied the Wasserstein distance to 3D-GAN structures, which improved the volume quality. Volume-based GANs are applied in the medical field, such as classification [33], segmentation [34], denoising [35], and detection [36]. Vox2Vox [37] is one of the volume-based GAN techniques used in the medical field for segmenting brain tumors. Nevertheless, few studies have applied GAN structures in volume generation, particularly for human anatomical models. Thus, it is essential to devise a method for generating volume data similar to a human anatomical model using a GAN.

3. Methods

3.1. Training Process

Figure 1 presents the overall network structure and components of the VI-GAN. The VI-GAN aims to generate volumetric data similar to the human anatomical model. This generation process is achieved using the volume generator G , which incorporates a 3D-convolution-based LFFB (local feature fusion block) [13], denoted as the 3D LFFB. The discriminator D is used to assess the authenticity of the generated volume and the ground truth. In addition, the VI-GAN devises reconstruction loss, including distance, feature, and similarity losses, to converge the generated volumes as a human anatomical model.
The proposed method should reconstruct all the voxels in the generated volume similar to the corresponding voxels in the ground truth volume. The difference between any two corresponding values should be close to zero. Equation (1) represents the training objective for a volumetric dataset. V represents the ground truth volume given during training, and V ^ denotes the corresponding generated volume. V [ m ] and V ^ [ m ] are the m th voxel of the ground truth volume V and generated volume V ^ , respectively. l , w , and h represent x , y , and z -axis resolutions of the volume, respectively.
m V m V ^ m 0 0 m l × w × h , m ( V m , V ^ m R l × w × h )
Volume generator G produces a volume using the input image set I G . The structure of G is represented in Figure 1b using Equation (2). X k represents the intermediate result in the k th layer. The input image set is reconstructed into a volume via each f layer. In the overall volume generation process, the network structures based on 3D U-Net [11] and ResNet [12] are categorized into the encoding, refining, and decoding parts. The encoding, refining, and decoding parts comprise f k down , f k mid , and f k up layers, respectively. The final layer is formed as the f layer; k is the index of those layers. θ k G is the learning parameter of the k th layer in the generator. Before passing to the f layers, 3D LFFB is performed to extract the essential features of the image set. Figure 1d presents the network structure of 3D LFFB.
f k down is the convolution layer that extracts the features from the input through down-sampling. f k mid is the convolution layer that refines features using the ResNet architecture. f k up is the deconvolution layer that reconstructs a volume from features by up-sampling. The final layer f is a convolution layer that generates a volume V ^ in which all voxel values are normalized from 0.0 to 1.0. The kernel sizes of all layers are 4 × 4 × 4 , and the stride sets were assigned for f k down , f k up , and f as two and f k mid as one. n G indicates the total number of layers of the generator; d , r , and u are the layer indices of the encoding, refining, and decoding parts, respectively.
X 0 = I G , X d = f d 1 down ( X d 1 ; θ d 1 G ) , X r = f r 1 mid ( X r 1 ; θ r 1 G ) , X u = f u 1 up ( X u 1 ; θ u 1 G ) V ^ = f ( X n G 1 ; θ n G 1 G ) , 0 < d < r < u < n G
Volume discriminator D distinguishes the authenticity between the generated and ground truth volume. The structure of D is represented in Figure 1c with Equation (3). I D is the input volume. Y k represents the intermediate result in the k th layer. The probability p indicating the authenticity of the input volume is calculated through each g layer. θ k D is the learning parameter of the k th layer in the discriminator. g k is a convolution layer that extracts the features from I D by down-sampling. The final layer, g , is a fully connected layer to generate a probability normalized from 0.0 to 1.0. The kernel sizes and stride sets of all layers are 4 × 4 × 4 and 2, respectively. n D represents the total number of layers in the discriminator.
Y 0 = I D , Y k = g k 1 ( Y k 1 ; θ k 1 D ) , p = g ( Y n D 1 ; θ n D 1 D ) , 0 < k < n D

3.2. Loss Function

The loss function of the generator and discriminator was designed, as shown in Equation (4). L G and L D are the generator and discriminator loss, respectively, of the ground truth volumes V , generated volumes V ^ , and the input image set of the generator I G . L recon is the reconstruction loss, and α is a constant weight assigned to L recon .
L G = E x I G , v ^ V ^ D ( x , v ^ ) 1 2 2 + α L recon L D = E x I G , v V , v ^ V ^ [ D ( x , v ) 1 2 2 + D ( x , v ^ ) 2 2 ]
The reconstruction loss L recon calculates the discrepancy between the generated and the ground truth volumes. Reconstruction loss consists of the distance, feature, and similarity loss. Among these losses, the distance loss L dist is defined using Equation (5). The L1 loss is used for the distance loss.
L dist = E v V , v ^ V ^ v v ^ 1
The voxel values in the generated volume indicate the predicted density of the corresponding area of the human anatomical model. The ground truth volume contains many undefined parts outside the human anatomical model, represented by low-value voxels. Therefore, many voxels in the ground truth volume have low values. The proposed method should reconstruct the entire part of the human anatomical model. Nevertheless, when only the distance loss is used in the regularization term, all voxels are averaged to minimize the distance loss. This process reduces the value of the high-value voxels [14]. The generated volume barely converges with the target model because the shape and characteristics of the human anatomical model are composed mainly of a high-value area. This paper proposes a reconstruction loss that consists of distance loss with feature loss and similarity loss to solve this problem.
Feature loss represents how many high-value voxels overlap between two volumes. Feature loss can be used to emphasize and preferentially reconstruct high-value voxels during the volume-reconstruction process. Feature loss L f is expressed as Equation (6). This loss uses N r number of thresholds; t s is the s th threshold; m is the voxel index in the volume; l , w , and h are the x , y , and z -axis resolutions of the volume, respectively. I ( · ) is the indicator function.
L f = 1 E v V , v ^ V ^ s N r m l × w × h I ( v [ m ] t s ) I ( v ^ [ m ] t s ) m l × w × h I ( I ( v [ m ] t s ) + I ( v ^ [ m ] t s ) ) 0 t s 1
Feature loss can be used to generate the characteristics of a human anatomical model, composed mainly of drastic changes in density. Such characteristics are often represented primarily by a high value. The high-value area within the human anatomical model can be fully reconstructed in the output volume by assigning feature loss during the training process.
Similarity loss L sim reflects the difference in image quality of each internal image slice between the generated and ground truth volumes, which is defined in Equation (7). The SSIM, SSIM ( · , · ) [15], measures the difference in image quality. Similarity loss computes the difference in volume quality calculated based on the SSIM between the internal image slices in the generated and ground truth volume for all z-values. S ( · , k ) is the selector that extracts the k th slice in the volume, and h is the resolution in the z -axis direction of the volume. Using the similarity loss, the internal voxel values can be reconstructed like those of the ground truth volume.
L sim = 1 E v V , v ^ V ^ { 1 h k = 0 h S S I M ( S ( v , k ) , S ( v ^ , k ) ) }
The final reconstruction loss is represented as Equation (8).
L recon = L dist + L f + L sim

3.3. Experimental Setting

This study used CT images of the spine (fifth lumbar vertebra and right hip bone) from the Digital Korean dataset provided by the KISTI [16] and the liver from CT-ORG [17]. The CT data for 94 people for the spine and 140 people for the liver were used. Among them, the CT data of the following were applied: 70 % for training, 20 % for testing, and 10 % for validation. A total of 3117 slices for the fifth lumbar vertebra (average of 33 slices per person), 4579 slices for the right hip bone (average of 49 slices per person), and 19,314 slices for the liver (average of 483 slices per person) were used to generate volumetric data for training. NVIDIA GeForce RTX 3090 Ti with 24,268 MB GPU memory was applied for training. The Adam optimizer [38] was used with a learning rate of 2 × 10 4 . The dropout rates of the middle-ware network blocks in Figure 1b were set to 0.2. The constant α was set to 33.0 when implementing Equation (4).

4. Results

The generated volumes were evaluated using a confusion matrix [39]. Each voxel in the generated volume was classified as positive if it had a high value and negative if it had a low value. In addition, each voxel in the ground truth volume was also categorized as true if it had a high value and false if it had a low value. The states of the voxels were judged by a comparison with the threshold value, whether each voxel value was high or low. Equation (9) expresses the TP (true positive), FP (false positive), and FN (false negative) used for the evaluation metric. t is the threshold; v is the ground truth volume; v ^ is the corresponding generated volume; l , w , and h are the x , y , and z -axis resolutions of the volume, respectively. The thresholds were used to evaluate how well the voxels that formulate a shape and internal characteristics of a human anatomical model were reconstructed. TP, FP, and FN were used to calculate the precision, recall, F1-score, and Dice coefficient [40].
TP = 1 l × w × h m = 0 l × w × h I ( v [ m ] > t ) I ( v ^ [ m ] > t ) FP = 1 l × w × h m = 0 l × w × h I ( v [ m ] t ) I ( v ^ [ m ] > t ) FN = 1 l × w × h m = 0 l × w × h I ( v [ m ] t ) I ( v ^ [ m ] t )
Figure 2 compares the generated volumes between the proposed VI-GAN and existing methods. The quality of the generated volume was evaluated using three criteria: volumetric shape, thresholding result, and internal image slice. The volumetric shape describes how accurately the generated volume represents the external shape of the ground truth volume. The thresholding result describes the internal structure that represents how much high-value voxels overlap between generated and ground truth volume. The internal image slice expresses how similar the internal area is between the generated and ground truth volume. The value of each voxel is a floating point between 0.0 and 1.0. The positions with a lower or higher voxel value are blue or red, respectively. The thresholds were determined for thresholding results by analyzing the voxel value distributions within the ground truth volumes. Specifically, thresholds of 0.4, 0.3, and 0.28 were applied to the fifth lumbar vertebra, right hip bone, and liver, respectively. For the fifth lumber vertebra and right hip bone, the Q3 (third quantile, 75% of data points) [41] values were adopted as thresholds to emphasize the rigid areas within the skeletal system. For the liver, the Q1 (first quantile, 25% of data points) values were used as thresholds to visualize the soft tissue density. For the internal image slice, the slice position is the center of the volume. The slices correspond to the xy plane exactly, the xy plane leaning at 45 , and the yz plane in the cases of the fifth lumbar vertebra, right hip bone, and liver, respectively. All slices in the CT volume were used as input data in Pix2Vox. The real CT volume was applied as input for the end-to-end CNN instead of the synthesized volume.
In the Pix2Vox and end-to-end CNN model, the volumetric shape had an ambiguous form that cannot express the human anatomical model in detail. In addition, the high-value voxels were scattered and could not typically converge to the shape of the human anatomical model. The generated volumes of the Vox2Vox model converged more to the human anatomical model than those of the Pix2Vox and end-to-end CNN models. The volumetric shapes and internal slices depicted as the form of the liver and bone represent this convergence. In the case of the fifth lumber vertebra and right hip bone, however, the high-value voxels in some areas of the volume were not fully reconstructed, which could not make the shape of the human anatomical model clear. In the case of the liver, the volumetric shape did not resemble the ground truth volume, and the organ shape did not appear in the center, which is the correct position for the internal slice. The VI-GAN generated a volumetric shape, high-value voxels, and internal parts that were more similar to the ground truth volume than the other methods. In conclusion, in a qualitative comparison, the VI-GAN generated volume data that are more similar to the human anatomical model than other existing methods.
Table 1 lists the quality measurement of the generated volume between the proposed VI-GAN and other existing methods. The quality was calculated using evaluation factors to measure the disparity between the generated and ground truth volume. The intersection over union (IoU), F1-score (F1), and Dice coefficient (DC) [43] represent the rate of overlapping voxels between thresholded results. The threshold values for calculating the IoU, F1, and DC metrics were 2.5, 1.9, and 2.8, respectively, corresponding to the fifth lumbar vertebra, right hip bone, and liver. These values represent the Q1 (25% of data points) value of the voxel distribution. The L1 error (L1) describes the voxel-wise difference between volumes. The peak signal-to-noise ratio (PSNR) [15], universal quality index (UQI) [44], visual saliency-induced index (VSI) [45], and structural similarity index map (SSIM) showed the similarity between the generated and ground truth volumes calculated using all slices.
A comparison of the VI-GAN with Pix2Vox and end-to-end CNN revealed that the VI-GAN had the highest result for all evaluation factors in all cases (the fifth lumber vertebra, right hip bone, and liver). As a result, the proposed method produces a volume that closely resembles the ground truth, displaying a higher rate of overlap among high-value voxels compared to Pix2Vox and end-to-end CNN. A comparison of the VI-GAN with Vox2Vox using the IoU, F1-score, and Dice coefficient showed that the VI-GAN produced better results in all cases than Vox2Vox. When comparing the metrics of L1, PSNR, UQI, VSI, and SSIM, it is difficult to definitively conclude whether the VI-GAN or Vox2Vox exhibited superior performance on similarity. Furthermore, the disparities in results were relatively minor in most cases. Based on these findings, the volumes produced by Vox2Vox and the VI-GAN showed similar degrees of resemblance to the ground truth. In summary, as listed in Table 1, the VI-GAN effectively generated a volume by accurately capturing high-value voxels compared to other methods while preserving the similarity to the human anatomical model.
Figure 3 presents the assessment results using the intersection over union, F1-score, and Dice coefficient across various threshold values. The objective of this experiment was to evaluate the accuracy in reconstructing high-value voxels, which holds significance in accurately representing the essential information within the volume. This information encompasses the overall appearance of soft tissue or rigid structures within the skeletal system. Furthermore, achieving precise reconstructions of high-value voxels is challenging because of the limited occurrence of such areas in the ground truth volume. The thresholds were selected within the range of Q1 (25% of data points) to Q3 (75% of data points), corresponding to 25%, 37.5%, 50%, 62.5%, and 75% within the voxel distribution. In this experiment, the following thresholds were applied: 0.2500, 0.2875, 0.3250, 0.3625, and 0.4000 for the fifth lumbar vertebra; 0.1900, 0.2175, 0.2450, 0.2725, and 0.3000 for the right hip bone; and 0.2800, 0.3250, 0.3700, 0.4150, and 0.4600 for the liver. Figure 3 shows a consistent trend across all methods in most cases; the results of the evaluation factor tended to decrease as the threshold value increased. This trend shows the difficulty of reconstructing high-value voxels within the volume accurately. Furthermore, the proposed VI-GAN consistently achieved superior results in most cases. Consequently, the VI-GAN demonstrated higher performance in reconstructing high-value voxels than the existing methods.

5. Conclusions

This paper proposed a VI-GAN for generating a volumetric model to describe the human anatomical model using a GAN-based volume generator and discriminator with 3D LFFB. This paper also proposed reconstruction loss, including feature loss and similarity loss, to reconstruct high-value areas and describe the essential characteristics of the model accurately. The experimental result showed that the generated volume of the VI-GAN represents the shape, high-value areas, and internal part better than the other existing methods evaluated. Furthermore, the experimental results with varying threshold values showed that the VI-GAN accurately generates high-value areas compared to existing methods. Based on these results, VI-GANs may enhance the availability of medical data and improve the training efficiency of medical professionals by generating high-quality volumetric data.

Author Contributions

Conceptualization, B.-S.S.; methodology, J.K., Y.L. and B.-S.S.; software, J.K.; validation, J.K. and Y.L.; formal analysis, J.K., Y.L. and B.-S.S.; investigation, J.K. and Y.L.; resources, J.K. and B.-S.S.; data curation, J.K.; writing—original draft preparation, J.K.; writing—review and editing, J.K., Y.L., and B.-S.S.; visualization, J.K.; supervision, Y.L. and B.-S.S.; project administration, B.-S.S.; funding acquisition, B.-S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Research Foundation of Korea (NRF) grants funded by the Korean government (No. NRF-2022R1A2B5B01001553 and No. NRF-2022R1A4A1033549) and Institute of Information & communications Technology Planning & Evaluation (IITP) grant funded by the Korean government (MSIT) (No. RS-2022-00155915, Artificial Intelligence Convergence Innovation Human Resources Development (Inha University)).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The spine CT images from the Digital Korean dataset are not accessible to the public. Please contact the Korea Institute of Science & Technology Information (KISTI) to obtain the data. The liver CT images from CT-ORG are publicly accessible and are available at https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=61080890 (accessed on 30 January 2024).

Acknowledgments

The authors gratefully acknowledge the human data support provided by the Korea Institute of Science & Technology Information (KISTI), which produced these data with the Catholic University of Medicine.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, R.; Huang, T.; Zhang, X.; Liao, H. 4: Interactive Volume Rendering Method Using Dynamic Ray Casting for Autostereoscopic Display. In SID Symposium Digest of Technical Papers; Wiley Online Library: Hoboken, NJ, USA, 2021; Volume 52, pp. 26–29. [Google Scholar]
  2. Fang, C.; An, J.; Bruno, A.; Cai, X.; Fan, J.; Fujimoto, J.; Golfieri, R.; Hao, X.; Jiang, H.; Jiao, L.R.; et al. Consensus recommendations of three-dimensional visualization for diagnosis and management of liver diseases. Hepatol. Int. 2020, 14, 437–453. [Google Scholar] [CrossRef]
  3. Nakao, M.; Nakamura, M.; Mizowaki, T.; Matsuda, T. Statistical deformation reconstruction using multi-organ shape features for pancreatic cancer localization. Med. Image Anal. 2021, 67, 101829. [Google Scholar] [CrossRef]
  4. Kavur, A.E.; Gezer, N.S.; Barış, M.; Aslan, S.; Conze, P.H.; Groza, V.; Pham, D.D.; Chatterjee, S.; Ernst, P.; Özkan, S.; et al. CHAOS challenge-combined (CT-MR) healthy abdominal organ segmentation. Med. Image Anal. 2021, 69, 101950. [Google Scholar] [CrossRef]
  5. Zhao, J.; Zhang, Y.; He, X.; Xie, P. Covid-ct-Dataset: A ct Scan Dataset about COVID-19. arXiv 2020, arXiv:2003.13865. [Google Scholar]
  6. Wisse, L.E.; Chételat, G.; Daugherty, A.M.; de Flores, R.; la Joie, R.; Mueller, S.G.; Stark, C.E.; Wang, L.; Yushkevich, P.A.; Berron, D.; et al. Hippocampal subfield volumetry from structural isotropic 1 mm3 MRI scans: A note of caution. Hum. Brain Mapp. 2021, 42, 539–550. [Google Scholar] [CrossRef]
  7. Costa, P.; Galdran, A.; Meyer, M.I.; Niemeijer, M.; Abràmoff, M.; Mendonça, A.M.; Campilho, A. End-to-end adversarial retinal image synthesis. IEEE Trans. Med. Imaging 2017, 37, 781–791. [Google Scholar] [CrossRef]
  8. Beers, A.; Brown, J.; Chang, K.; Campbell, J.P.; Ostmo, S.; Chiang, M.F.; Kalpathy-Cramer, J. High-resolution medical image synthesis using progressively grown generative adversarial networks. arXiv 2018, arXiv:1805.03144. [Google Scholar]
  9. Balashova, E.; Wang, J.; Singh, V.; Georgescu, B.; Teixeira, B.; Kapoor, A. 3D Organ Shape Reconstruction from Topogram Images. In Proceedings of the International Conference on Information Processing in Medical Imaging, Hong Kong, China, 2–7 June 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 347–359. [Google Scholar]
  10. Kasten, Y.; Doktofsky, D.; Kovler, I. End-to-end convolutional neural network for 3D reconstruction of knee bones from bi-planar X-ray images. In Proceedings of the International Workshop on Machine Learning for Medical Image Reconstruction, Lima, Peru, 8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 123–133. [Google Scholar]
  11. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Athens, Greece, 17–21 October 2016; Springer: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar]
  12. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  13. Jiang, M.; Zhi, M.; Wei, L.; Yang, X.; Zhang, J.; Li, Y.; Wang, P.; Huang, J.; Yang, G. FA-GAN: Fused attentive generative adversarial networks for MRI image super-resolution. Comput. Med. Imaging Graph. 2021, 92, 101969. [Google Scholar] [CrossRef] [PubMed]
  14. Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier gans. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 2642–2651. [Google Scholar]
  15. Setiadi, D.R.I.M. PSNR vs. SSIM: Imperceptibility quality assessment for image steganography. Multimed. Tools Appl. 2021, 80, 8423–8444. [Google Scholar] [CrossRef]
  16. Lee, S.H.; Lee, S.B. Production and usage of Korean human information in KISTI. J. Korea Contents Assoc. 2010, 10, 416–421. [Google Scholar] [CrossRef]
  17. Rister, B.; Yi, D.; Shivakumar, K.; Nobashi, T.; Rubin, D.L. CT-ORG, a new dataset for multiple organ segmentation in computed tomography. Sci. Data 2020, 7, 381. [Google Scholar] [CrossRef]
  18. Dai, W.C.; Zhang, H.W.; Yu, J.; Xu, H.J.; Chen, H.; Luo, S.P.; Zhang, H.; Liang, L.H.; Wu, X.L.; Lei, Y.; et al. CT imaging and differential diagnosis of COVID-19. Can. Assoc. Radiol. J. 2020, 71, 195–200. [Google Scholar] [CrossRef] [PubMed]
  19. Byl, J.; Nelliadi, S.S.; Samuel, B.; Vettukattil, J. True 3D Viewer facilitates accurate diagnosis of lung infarction. Vasc. Dis. Manag. 2021, 18, E267–E268. [Google Scholar]
  20. Shi, W.; Liu, P.X.; Zheng, M. Cutting procedures with improved visual effects and haptic interaction for surgical simulation systems. Comput. Methods Programs Biomed. 2020, 184, 105270. [Google Scholar] [CrossRef] [PubMed]
  21. Munawar, A.; Li, Z.; Nagururu, N.; Trakimas, D.; Kazanzides, P.; Taylor, R.H.; Creighton, F.X. Fully Immersive Virtual Reality for Skull-base Surgery: Surgical Training and Beyond. arXiv 2023, arXiv:2302.13878. [Google Scholar] [CrossRef] [PubMed]
  22. Henzler, P.; Rasche, V.; Ropinski, T.; Ritschel, T. Single-image Tomography: 3D Volumes from 2D Cranial X-Rays. In Computer Graphics Forum; Wiley Online Library: Hoboken, NJ, USA, 2018; Volume 37, pp. 377–388. [Google Scholar]
  23. Xu, Y.; Li, Y.; Shin, B.S. Medical image processing with contextual style transfer. Hum.-Centric Comput. Inf. Sci. 2020, 10, 46. [Google Scholar] [CrossRef]
  24. Qiao, Z.; Qian, Z.; Tang, H.; Gong, G.; Yin, Y.; Huang, C.; Fan, W. CorGAN: Context aware Recurrent Generative Adversarial Network for Medical Image Generation. In Proceedings of the 2020 IEEE International Conference on Bioinformatics and Biomedicine (BIBM), Seoul, Republic of Korea, 16–19 December 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 1100–1103. [Google Scholar]
  25. Wang, Q.; Zhang, X.; Zhang, W.; Gao, M.; Huang, S.; Wang, J.; Zhang, J.; Yang, D.; Liu, C. Realistic lung nodule synthesis with multi-target co-guided adversarial mechanism. IEEE Trans. Med. Imaging 2021, 40, 2343–2353. [Google Scholar] [CrossRef] [PubMed]
  26. Masutani, E.M.; Bahrami, N.; Hsiao, A. Deep learning single-frame and multiframe super-resolution for cardiac MRI. Radiology 2020, 295, 552–561. [Google Scholar] [CrossRef] [PubMed]
  27. Kitchen, A.; Seah, J. Deep generative adversarial neural networks for realistic prostate lesion MRI synthesis. arXiv 2017, arXiv:1708.00129. [Google Scholar]
  28. Chuquicusma, M.J.; Hussein, S.; Burt, J.; Bagci, U. How to fool radiologists with generative adversarial networks? A visual turing test for lung cancer diagnosis. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–7 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 240–244. [Google Scholar]
  29. Frid-Adar, M.; Diamant, I.; Klang, E.; Amitai, M.; Goldberger, J.; Greenspan, H. GAN-based synthetic medical image augmentation for increased CNN performance in liver lesion classification. Neurocomputing 2018, 321, 321–331. [Google Scholar] [CrossRef]
  30. Karras, T.; Aila, T.; Laine, S.; Lehtinen, J. Progressive growing of gans for improved quality, stability, and variation. arXiv 2017, arXiv:1710.10196. [Google Scholar]
  31. Wu, J.; Zhang, C.; Xue, T.; Freeman, B.; Tenenbaum, J. Learning a probabilistic latent space of object shapes via 3D generative-adversarial modeling. Adv. Neural Inf. Process. Syst. 2016, 29, 2–9. [Google Scholar]
  32. Smith, E.J.; Meger, D. Improved adversarial systems for 3D object generation and reconstruction. In Proceedings of the Conference on Robot Learning, PMLR, Mountain View, CA, USA, 13–15 November 2017; pp. 87–96. [Google Scholar]
  33. Kruthika, K.; Maheshappa, H.; Initiative, A.D.N. CBIR system using Capsule Networks and 3D CNN for Alzheimer’s disease diagnosis. Inform. Med. Unlocked 2019, 14, 59–68. [Google Scholar] [CrossRef]
  34. Xu, C.; Xu, L.; Ohorodnyk, P.; Roth, M.; Chen, B.; Li, S. Contrast agent-free synthesis and segmentation of ischemic heart disease images using progressive sequential causal GANs. Med. Image Anal. 2020, 62, 101668. [Google Scholar] [CrossRef]
  35. Ran, M.; Hu, J.; Chen, Y.; Chen, H.; Sun, H.; Zhou, J.; Zhang, Y. Denoising of 3D magnetic resonance images using a residual encoder—Decoder Wasserstein generative adversarial network. Med. Image Anal. 2019, 55, 165–180. [Google Scholar] [CrossRef] [PubMed]
  36. Han, C.; Rundo, L.; Murao, K.; Noguchi, T.; Shimahara, Y.; Milacski, Z.Á.; Koshino, S.; Sala, E.; Nakayama, H.; Satoh, S. MADGAN: Unsupervised medical anomaly detection GAN using multiple adjacent brain MRI slice reconstruction. BMC Bioinform. 2021, 22, 31. [Google Scholar] [CrossRef]
  37. Cirillo, M.D.; Abramian, D.; Eklund, A. Vox2Vox: 3D-GAN for brain tumour segmentation. In Proceedings of the International MICCAI Brainlesion Workshop, Lima, Peru, 4–8 October 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 274–284. [Google Scholar]
  38. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  39. Visa, S.; Ramsay, B.; Ralescu, A.L.; Van Der Knaap, E. Confusion matrix-based feature selection. Maics 2011, 710, 120–127. [Google Scholar]
  40. Taha, A.A.; Hanbury, A. Metrics for evaluating 3D medical image segmentation: Analysis, selection, and tool. BMC Med. Imaging 2015, 15, 29. [Google Scholar] [CrossRef]
  41. Roschger, P.; Gupta, H.; Berzlanovich, A.; Ittner, G.; Dempster, D.; Fratzl, P.; Cosman, F.; Parisien, M.; Lindsay, R.; Nieves, J.; et al. Constant mineralization density distribution in cancellous human bone. Bone 2003, 32, 316–323. [Google Scholar] [CrossRef]
  42. Xie, H.; Yao, H.; Sun, X.; Zhou, S.; Zhang, S. Pix2vox: Context-aware 3d reconstruction from single and multi-view images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2690–2698. [Google Scholar]
  43. Bertels, J.; Eelbode, T.; Berman, M.; Vandermeulen, D.; Maes, F.; Bisschops, R.; Blaschko, M.B. Optimizing the dice score and jaccard index for medical image segmentation: Theory and practice. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Shenzhen, China, 13–17 October 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 92–100. [Google Scholar]
  44. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  45. Zhang, L.; Shen, Y.; Li, H. VSI: A visual saliency-induced index for perceptual image quality assessment. IEEE Trans. Image Process. 2014, 23, 4270–4281. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Overall proposed framework and components of VI-GAN. The blue cubes represent features extracted by the generator, while the green cubes depict features extracted by the discriminator. (a) Overview of the proposed network structure for generating a volume. (b,c) Detailed structure of the volume generator and discriminator. Three dots means repeating the preceding refining and concatenation process in the same manner. (d) 3D local feature fusion block (3D LFFB) that initially extracts the features from an image set before the volume generator.
Figure 1. Overall proposed framework and components of VI-GAN. The blue cubes represent features extracted by the generator, while the green cubes depict features extracted by the discriminator. (a) Overview of the proposed network structure for generating a volume. (b,c) Detailed structure of the volume generator and discriminator. Three dots means repeating the preceding refining and concatenation process in the same manner. (d) 3D local feature fusion block (3D LFFB) that initially extracts the features from an image set before the volume generator.
Bioengineering 11 00163 g001
Figure 2. Qualitative comparison of the generated volumes in the proposed VI-GAN and existing method. The (a) fifth lumber vertebra, (b) right hip bone, and (c) liver volume data are represented. The images consist of a volumetric shape (top row), thresholding result (center row), and internal image slice (bottom row). The volumes generated by the Pix2Vox [42] (first column), end-to-end CNN [10] (second column), Vox2Vox [37] (third column), and VI-GAN (fourth column) models were compared.
Figure 2. Qualitative comparison of the generated volumes in the proposed VI-GAN and existing method. The (a) fifth lumber vertebra, (b) right hip bone, and (c) liver volume data are represented. The images consist of a volumetric shape (top row), thresholding result (center row), and internal image slice (bottom row). The volumes generated by the Pix2Vox [42] (first column), end-to-end CNN [10] (second column), Vox2Vox [37] (third column), and VI-GAN (fourth column) models were compared.
Bioengineering 11 00163 g002
Figure 3. Quality comparison between the volumes of the proposed VI-GAN and existing methods across a range of thresholds.
Figure 3. Quality comparison between the volumes of the proposed VI-GAN and existing methods across a range of thresholds.
Bioengineering 11 00163 g003
Table 1. Quality comparison between the volumes of proposed VI-GAN and existing methods using the evaluation factors. The best results are represented in bold.
Table 1. Quality comparison between the volumes of proposed VI-GAN and existing methods using the evaluation factors. The best results are represented in bold.
MethodFifth Lumber VertebraRight Hip BoneLiver
IoUDCF1L1IoUDCF1L1IoUDCF1L1
Pix2Vox0.5010.6700.6620.1210.3810.5530.5510.1310.2640.4880.4020.139
End-to-end CNN0.5240.6810.6860.1210.3720.5440.5410.1140.2310.4840.3460.155
Vox2Vox0.5960.7430.7420.0960.4570.6280.6260.0750.3930.6120.5550.053
VI-GAN0.7120.8320.8310.1020.8650.9290.9270.0750.5520.7410.6850.056
MethodFifth Lumber VertebraRight Hip BoneLiver
PSNRUQIVSISSIMPSNRUQIVSISSIMPSNRUQIVSISSIM
Pix2Vox16.2260.8390.8270.32915.7640.7290.8290.27915.4860.8840.8310.284
End-to-end CNN16.5730.8490.8360.24416.6760.8170.8720.61314.8040.8390.7760.409
Vox2Vox18.2490.8760.8750.48324.6150.8170.8720.61320.9420.9000.8800.652
VI-GAN18.1890.8830.8740.50524.5180.8160.8650.60026.6610.8860.8740.669
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, J.; Li, Y.; Shin, B.-S. Volumetric Imitation Generative Adversarial Networks for Anatomical Human Body Modeling. Bioengineering 2024, 11, 163. https://doi.org/10.3390/bioengineering11020163

AMA Style

Kim J, Li Y, Shin B-S. Volumetric Imitation Generative Adversarial Networks for Anatomical Human Body Modeling. Bioengineering. 2024; 11(2):163. https://doi.org/10.3390/bioengineering11020163

Chicago/Turabian Style

Kim, Jion, Yan Li, and Byeong-Seok Shin. 2024. "Volumetric Imitation Generative Adversarial Networks for Anatomical Human Body Modeling" Bioengineering 11, no. 2: 163. https://doi.org/10.3390/bioengineering11020163

APA Style

Kim, J., Li, Y., & Shin, B. -S. (2024). Volumetric Imitation Generative Adversarial Networks for Anatomical Human Body Modeling. Bioengineering, 11(2), 163. https://doi.org/10.3390/bioengineering11020163

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop