You are currently viewing a new version of our website. To view the old version click .
Applied Sciences
  • Article
  • Open Access

3 March 2023

A Novel Implicit Neural Representation for Volume Data

and
National Center for Computer Animation, Faculty of Media and Communication, Bournemouth University, Bournemouth BH12 5BB, UK
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.

Abstract

The storage of medical images is one of the challenges in the medical imaging field. There are variable works that use implicit neural representation (INR) to compress volumetric medical images. However, there is room to improve the compression rate for volumetric medical images. Most of the INR techniques need a huge amount of GPU memory and a long training time for high-quality medical volume rendering. In this paper, we present a novel implicit neural representation to compress volume data using our proposed architecture, that is, the Lanczos downsampling scheme, SIREN deep network, and SRDenseNet high-resolution scheme. Our architecture can effectively reduce training time, and gain a high compression rate while retaining the final rendering quality. Moreover, it can save GPU memory in comparison with the existing works. The experiments show that the quality of reconstructed images and training speed using our architecture is higher than current works which use the SIREN only. Besides, the GPU memory cost is evidently decreased.

1. Introduction

Contemporary microscopy technology is widely applied to biology, neuroscience, and medical imaging fields, which bring about large-scale multidimensional datasets and require terabytes or petabytes of data storage. Besides, recent developments in radiological hardware have increased the capabilities of medical imaging to produce high-resolution and 3D images that require high storage capacity. It is desired to manipulate multidimensional datasets for visualization and predictions from multi-channel image intensity values in a 2D or 3D scan. Such datasets pose significant challenges related to storage, manipulation, and rendering. In this paper, we suggest a novel implicit neural representation to compress high-resolution medical volume data using the implicit neural representation (INR) network with high speed and quality in comparison with current works.
There are some works using INR with periodic activation functions (SIREN) [1] to compress the medical images [2,3]. In order to compress the images, these techniques train a neural network using the voxel coordinates. The resulting trained networks represent the individual image stacks. A distinct advantage is that the size of the trained deep network is always less than that of the image stack. However, they also suffer some considerable issues. They usually depend on a huge GPU memory, specifically for high-resolution volume data. Thus, it is impossible for clinicians to access medical volume data based on usual workstations and laptops. On the other hand, in order to deal with high-resolution (HR) image stacks, it is natural to use large-scale neural networks with multiple layers, which results in a long training time and a low rate of compression. Obviously, the main challenge is to both keep a high rate of compression to decrease deep network size and keep the quality of reconstructed images.
To tackle this challenge, we present an architecture combining the SIREN to compress the volume data effectively and a super-resolution module to keep the quality of the reconstruction. Our architecture includes a downsampling module to decrease the resolution of high-resolution (HR) and low-resolution (LR) images, which can fit the downsampled volume data in usual GPUs with low graphical memory. As a result, the SIREN’s layers can be reduced. Training can be accelerated accordingly. The following super-resolution (SR) reconstruction network can recover images up to the quality of the original HR.
In summary, our contributions include:
  • Our architecture consists of the Lanczos downsampling scheme, SIREN deep network, and SRDenseNet upsampling scheme, which increase the speed of training and decrease the demand for GPU memory in comparison with existing INR-based compression techniques;
  • Our architecture can reach both a high compression rate and high quality of the final volume data rendering.
The following part of this paper is organized as follows. Section 2 presents the related work. Section 3 illustrates the methodology. Section 4 shows the experiments and results from the analysis. Section 5 draws our conclusions and presents future works.

3. Methodology

In the following section, we present our architecture and explain its stages in detail.

3.1. Our Architecture

We aim at the medical volume data compression technique to work on usual workstations and laptops instead of the overdependency on high-end GPUs. In order to increase the compression rate and the quality of the reconstruction, we present an architecture that can be applied in many scenarios, as shown in Figure 1. Our architecture consists of three modules, where the first module can downsample the original high-resolution images using Lanczos to decrease the volume data size. The second module can work on the LR volume data for the implicit neural representation. The third module can take the LR volume data back to the original size using SRDenseNet. As a result, the original volume data can be represented by the neural network to achieve compression purposes. The first module may result in information loss. The third module can recover it. There is no guarantee that it is lossless. However, the experiments show our architecture can reach high PSNR.
Figure 1. Our suggested architecture using INR (in this work, SIREN) to compress high-resolution (HR) medical images.

3.2. Lanczos Resampling

The Lanczos resampling technique [54] is an interpolation technique which is used to resample (downsample/upsample) the volumetric medical images. Here, we use Lanczos resampling to downsample our HR images. In the Lanczos resampling technique, each selected sample is replaced with a translated copy of the Lanczos kernel, which is a limited Sinc function. Equation (1) shows the Lanczos kernel formula.
L ( x ) = s i n c ( π x ) s i n c ( π x / a ) i f | x | < a 0 o t h e r w i s e .

3.3. Sinusoidal Representation Networks (SIREN)

We apply the SIREN deep network to medical image stacks. It may be regarded as a function that accepts the 3D coordinate of a voxel x = ( x , y , z ) and outputs the corresponding intensity value σ . In this case, the neural network becomes a continuous implicit function F Θ : x σ and can represent a continuous scene of medical image stacks. The training is to optimize its weights Θ for accurate representation. Recently, INRs have been one of the considerable areas of research focus. Most of the methods adopt a MLP structure using ReLU [9] as the activation function. ReLU-based MLPs are suffering from the lack of capacity to represent fine details. ReLU networks are linear and their second derivative is zero. As a result, it is impossible to model high-order signals. In the various versions of SIREN, the Sine function is used as the activation function to model fine details [1]. We applied a small version to training, and the resulting deep network was so small. As a new representation of the original volume data, it effectively compresses the volume data and saves training time.

3.4. SRDenseNet

In order to obtain the original resolution image, we use SRDenseNet [55] as a super-resolution module, which is a deep network. We selected SRDenseNet because of its generalization. We note that it can work well on a large class of medical images. For example, we use a small training image set from a large class and then perform it on unseen images of the same class. There are three types of SRDenseNets, SRDenseNet-H, SRDenseNet-HL, and SRDenseNet-All. SRDenseNet-H uses high-level features to reconstruct HR images. SRDenseNet-HL uses a combination of low-level and high-level features to reconstruct HR images. SRDenseNet-All, which is used in our implementation, uses a combination of all levels of features with densely skipped connections to reconstruct HR images. SRDenseNet-All is constituted of four stages, low-level features, high-level features, deconvolution layers, and a reconstruction layer.
In the low-level feature stage, the network receives a low-resolution (LR) image as input. A convolution layer learns low-level features. There are eight dense blocks that learn high-level features. Each dense block is constituted of eight convolution layers and each layer produces 16 feature maps. Thus, each block produces 128 feature maps. In the bottleneck layer, the number of feature maps is decreased to a compact model. Then, image resolution is up-scaled from 128 × 128 to 512 × 512 using a deconvolution layer. Finally, with another convolutional layer with a 3 × 3 kernel, the output channels are reduced to a single channel. Figure 2 shows the SRDenseNet-All structure which is used in this paper. In this work, we use the pre-trained model of SRDenseNet in [55]. We trained the SRDenseNet network with LR and HR pairs of slices of our volumetric medical images dataset. Then a pre-trained model of our training is used in our architecture. According to this, SRDenseNet in our architecture can reconstruct HR images of downsampled images of our dataset.
Figure 2. SRDensenet—all architecture with eight blocks.

3.5. Peak Signal-to-Noise Ratio

To compare the quality of the reconstructed image with the original one, we use the Peak Signal-to-Noise Ratio (PSNR) measurement technique. It is given for grey-scale images by the equation:
P S N R = 10 . l o g 10 I 2 M S E
where MSE is the mean square error which is given by the equation below:
M S E = 1 N . M i = 1 N j = 1 M ( y i j y ¯ i j )
I is the maximum intensity value of the image, which is 255 in this work. N is the number of rows and M is the number of columns. y is the original image and y ¯ is the reconstructed image.

4. Results and Discussion

In this section, we compare the numerical results of the applications of INR-based compression techniques with and without our architecture in terms of reconstruction quality, training speed, and GPU adaptivity. In this work, we examine an INR-based method, SIREN with two, three, and four MLP layers. Table 1, Table 2 and Table 3 show the size of the MLPs with two, three, and four layers, respectively, in detail.
Table 1. The estimation of SIREN network size with two layers.
Table 2. The estimation of SIREN network size with three layers.
Table 3. The estimation of SIREN network size with four layers.

4.1. Dataset

In this paper, we are concerned with both 2D and 3D images. We use a volumetric dataset, containing Human CT scan slices obtained from the Visible Human project dataset (https://www.nlm.nih.gov, (accessed on 1 April 2022)). The database constitutes 463 DICOM axial CT scan slices of a human head with a resolution of 512 × 512 pixels and each pixel is made up of 12 bits of grey tone. The thickness of each slice is 0.5 mm.

4.2. Using SIREN with Our Architecture

Based on our architecture, we use the LANCZOS resampling technique to downsample our high-resolution volumetric medical images (512 × 512 × 463) to (128 × 128 × 115). This technique decreases the number of voxels from 121,372,672 to just 1,884,160. For simplicity, we show the results on a selected 2D slice, but it should be considered that it can be expanded to 3D data. Figure 3 shows a selected slice that is downsampled from 512 × 512 × 1 to 128 × 128 × 1 using LANCZOS, which decreases the number of voxels from 262,144 to just 16,384.
Figure 3. Left column shows the high-resolution slices with a size of 512 × 512 and the right is the low-resolution slices with a size of 128 × 128. Low-resolution slices were obtained by applying Lanczos resampling on the high-resolution slices.
After the downsampling stage, our architecture uses SIREN with 2, 3, and 4 layers and 128 neurons for each layer. Figure 4, Figure 5 and Figure 6 show the results, respectively. The Adam was selected as the optimizer with a learning rate of 0.0015 and batch size of 2 16 . MSE was used as the loss function of the MLP. In order to reconstruct the HR slice, we used SRDenseNet and applied it to reconstructed LR images. We also trained it on this kind of CT data in advance so that the trained SRDenseNe is suitable for the targeted class. The final results of using SIREN with our architecture are shown in Figure 7, Figure 8 and Figure 9, respectively. Table 4 compares the results of the SIREN application with 2, 3 and 4 layers, using our architecture.
Figure 4. From left, the first image shows the original down-sampled image. The second image illustrates the reconstructed counterpart of the down-sampled image using SIREN. The next image shows the plot of the PSNR while training the network (best PSNR: 45.38) and the last image shows the plot of the loss values during the SIREN training (loss: 1.4805471 × 10 5 ).
Figure 5. From the left, the first image shows the original down-sampled image. The second image illustrates the reconstructed counterpart of the down-sampled image using SIREN. The next image shows the plot of the PSNR while training the network (best PSNR: 48.29) and the last image shows the plot of the loss values during the SIREN training (loss: 1.0805471 × 10 5 ).
Figure 6. From the left, the first image shows the original down-sampled image. The second image illustrates the reconstructed counterpart of the down-sampled image using a four-layer SIREN. The next image shows the plot of the PSNR during training of the network (best PSNR: 70.63) and the last image shows the plot of the loss values during the four-layer SIREN training (loss: 8.6481705 × 10 8 ).
Figure 7. Results of the whole procedure of our architecture. From the left, the first image shows the original high-resolution (HR) image. The second image shows the original low resolution (LR). The third image shows the reconstructed low resolution (LR) and the last shows the reconstructed high resolution (HR). The best PSNR of reconstructed HR in comparison with the original HR is 34.670.
Figure 8. Results of the whole procedure of our architecture. From the left, the first image shows the original high-resolution (HR) image. The second shows the original low resolution (LR). The third image shows the reconstructed low resolution (LR) and the last shows the reconstructed high resolution (HR). The best PSNR of reconstructed HR in comparison with original HR is 34.865.
Figure 9. Results of the whole procedure of our architecture. From the left, the first image shows the original high-resolution (HR) image. The second shows the original low resolution (LR). The third image shows the reconstructed low resolution (LR) and the last shows the reconstructed high resolution (HR). The best PSNR of reconstructed HR in comparison with the original HR is 35.140.
Table 4. Shows comparison results of the implementation of SIREN with our architecture in terms of quality, speed, GPU memory allocation, and compression rate with 2, 3, and 4 layers of SIREN.

4.3. Using SIREN without Our Architecture [2]

In this section, we show the results of SIREN implementation without our architecture which was implemented in the Ref. [2]. Figure 10, Figure 11 and Figure 12 show the results of high-resolution (HR) reconstruction of 2, 3 and 4 layers of SIREN without our architecture, respectively. Table 5 illustrates the results of using SIREN without our architecture with 2, 3, and 4 layers.
Figure 10. Shows the reconstruction of HR images using a two-layer SIREN without our architecture. The best PSNR is 25.26.
Figure 11. Shows the reconstruction of HR images using a three-layer SIREN without our architecture. The best PSNR is 28.80.
Figure 12. Shows the reconstruction of HR images using a four-layer SIREN without our architecture. The best PSNR is 30.68.
Table 5. Shows comparison results of the implementation of SIREN without our architecture in terms of quality, speed, GPU memory allocation, and compression rate with 2, 3, and 4 layers with SIREN.

4.4. Comparison with Existing Methods

In this section, we compare the performance of SIREN with our architecture and SIREN without our architecture which is used in the Ref. [2] with 2, 3 and 4 layers and 128 neurons for each layer. From the general perspective, both methods use SIREN to compress volumetric medical images. In the Ref. [2] they applied SIREN without any downsampling and upsampling techniques which led to high GPU memory consumption and low training speed. In Figure 13 there are three bar charts, showing that SIREN with our architecture outperforms SIREN without our architecture in the Ref. [2] in terms of quality, speed, and GPU memory consumption with 2, 3, and 4 MLP layers. As can be seen, there are considerable gaps between our technique and SIREN in the Ref. [2] in terms of training speed and GPU memory consumption. The PSNR gap is decreased by increasing the number of layers, but our architecture still outperforms the SIREN in the Ref. [2]. Table 6 compares the results of using SIREN to compress volumetric medical images in the Ref. [2] with using SIREN with our architecture in three factors of PSNR, training speed and GPU memory consumption. Figure 14 compares the final results of using SIREN in the Ref. [2] and using SIREN with our architecture with 2, 3 and 4 layers. Table 7 shows the comparison of the results of Shen’s [52], Mishra’s [53], SIREN [2] and ours in terms of PSNR.
Figure 13. Bar charts compare the results of using SIREN without our architecture [2] and with our architecture. As can be seen, SIREN implementation with our architecture outperforms using SIREN without our architecture [2] in terms of quality, speed, and GPU memory allocation with 2, 3, and 4 MLP layers.
Table 6. Display and comparison of the final results that were obtained using SIREN in the Ref. [2] and using SIREN with our architecture. As it can be seen, using SIREN with our architecture outperforms using SIREN in the Ref. [2] in three terms of quality, speed, and GPU memory allocation with 2, 3 and 4 MLP layers.
Figure 14. Comparison of the quality of reconstructed HR using SIREN in the Ref. [2] and using SIREN with our architecture with 2, 3 and 4 MLP layers.
Table 7. Comparison of results of our technique with SIREN [2], Mishra’s [53] and Shen’s [52] in terms of PSNR.

5. Conclusions

Our architecture significantly outperforms other INR-based techniques without our architecture, which are regular in medical image compression. The experiments show that our proposed architecture is a novel implicit neural representation of medical volume data, reaching both a high compression rate and high quality of reconstruction. The architecture is simple and re-configurable. It may replace volume data and be regarded as a new representation form. Moreover, the three modules can be replaced by others according to different applications. For example, manual annotation may work on low-resolution volume to save time, while rendering may work on high-resolution volume for accuracy.
The quality of our reconstructed high-resolution images with a small version of SIREN is considerably higher than direct SIREN with the same size. We also note that the compression rate depends on the deep network structure. To reach a high compression rate and low loss for volume data, the SR module is crucial. Moreover, the generalization of the SR module is another concern. These are worth our effort in the future.

Author Contributions

Conceptualization, A.S. and H.Y.; methodology, A.S.; software, A.S.; validation, A.S. and H.Y.; formal analysis, A.S.; investigation, A.S. and H.Y.; resources, A.S. and H.Y.; data curation, A.S. and H.Y.; writing—original draft preparation, A.S.; writing—review and editing, A.S. and H.Y.; visualization, A.S.; supervision, H.Y.; project administration, H.Y.; funding acquisition, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
INRImplicit Neural Representation
MLPMulti-Layer Perceptron
CTComputed Tomography
SIRENSinusoidal representation network

References

  1. Sitzmann, V.; Martel, J.N.P.; Bergman, A.W.; Lindell, D.B.; Wetzstein, G. Implicit Neural Representations with Periodic Activation Functions. arXiv 2020, arXiv:2006.09661. [Google Scholar] [CrossRef]
  2. Mancini, M.; Jones, D.K.; Palombo, M. Lossy compression of multidimensional medical images using sinusoidal activation networks: An evaluation study. arXiv 2022, arXiv:2208.01602. [Google Scholar] [CrossRef]
  3. Strümpler, Y.; Postels, J.; Yang, R.; van Gool, L.; Tombari, F. Implicit Neural Representations for Image Compression. arXiv 2021, arXiv:2112.04267. [Google Scholar] [CrossRef]
  4. Wu, J.; Zhang, C.; Zhang, X.; Zhang, Z.; Freeman, W.T.; Tenenbaum, J.B. Learning Shape Priors for Single-View 3D Completion and Reconstruction. arXiv 2018, arXiv:1809.05068. [Google Scholar] [CrossRef]
  5. Khan, M.O.; Fang, Y. Implicit Neural Representations for Medical Imaging Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022; Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S., Eds.; Springer Nature Switzerland: Cham, Switzerland, 2022; pp. 433–443. [Google Scholar]
  6. Park, J.J.; Florence, P.; Straub, J.; Newcombe, R.; Lovegrove, S. DeepSDF: Learning Continuous Signed Distance Functions for Shape Representation. arXiv 2019, arXiv:1901.05103. [Google Scholar] [CrossRef]
  7. Tang, D.; Singh, S.; Chou, P.A.; Haene, C.; Dou, M.; Fanello, S.; Taylor, J.; Davidson, P.; Guleryuz, O.G.; Zhang, Y.; et al. Deep Implicit Volume Compression. arXiv 2020, arXiv:2005.08877. [Google Scholar] [CrossRef]
  8. Nagoor, O.H.; Whittle, J.; Deng, J.; Mora, B.; Jones, M.W. MedZip: 3D Medical Images Lossless Compressor Using Recurrent Neural Network (LSTM). In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; pp. 2874–2881. [Google Scholar] [CrossRef]
  9. Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis. arXiv 2020, arXiv:2003.08934. [Google Scholar] [CrossRef]
  10. Mescheder, L.; Oechsle, M.; Niemeyer, M.; Nowozin, S.; Geiger, A. Occupancy Networks: Learning 3D Reconstruction in Function Space. arXiv 2018, arXiv:1812.03828. [Google Scholar] [CrossRef]
  11. Dupont, E.; Goliński, A.; Alizadeh, M.; Teh, Y.W.; Doucet, A. COIN: COmpression with Implicit Neural representations. arXiv 2021, arXiv:2103.03123. [Google Scholar] [CrossRef]
  12. Skorokhodov, I.; Ignatyev, S.; Elhoseiny, M. Adversarial Generation of Continuous Images. arXiv 2020, arXiv:2011.12026. [Google Scholar] [CrossRef]
  13. Shen, L.; Pauly, J.; Xing, L. NeRP: Implicit Neural Representation Learning with Prior Embedding for Sparsely Sampled Image Reconstruction. arXiv 2021, arXiv:2108.10991. [Google Scholar] [CrossRef] [PubMed]
  14. Eslami, S.M.A.; Rezende, D.J.; Besse, F.; Viola, F.; Morcos, A.S.; Garnelo, M.; Ruderman, A.; Rusu, A.A.; Danihelka, I.; Gregor, K.; et al. Neural scene representation and rendering. Science 2018, 360, 1204–1210. [Google Scholar] [CrossRef] [PubMed]
  15. Tancik, M.; Srinivasan, P.P.; Mildenhall, B.; Fridovich-Keil, S.; Raghavan, N.; Singhal, U.; Ramamoorthi, R.; Barron, J.T.; Ng, R. Fourier Features Let Networks Learn High Frequency Functions in Low Dimensional Domains. In Proceedings of the 34th International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 6–12 December 2020; NIPS’20; Curran Associates Inc.: Red Hook, NY, USA, 2022. [Google Scholar]
  16. Dar, S.U.H.; Özbey, M.; Çatlı, A.B.; Çukur, T. A Transfer-Learning Approach for Accelerated MRI Using Deep Neural Networks. Magn. Reson. Med. 2020, 84, 663–685. [Google Scholar] [CrossRef]
  17. Qin, C.; Schlemper, J.; Caballero, J.; Price, A.N.; Hajnal, J.V.; Rueckert, D. Convolutional Recurrent Neural Networks for Dynamic MR Image Reconstruction. IEEE Trans. Med Imaging 2019, 38, 280–290. [Google Scholar] [CrossRef]
  18. Wu, Y.; Ma, Y.; Liu, J.; Du, J.; Xing, L. Self-attention convolutional neural network for improved MR image reconstruction. Inf. Sci. 2019, 490, 317–328. [Google Scholar] [CrossRef]
  19. Liu, J.; Zhang, Y.; Zhao, Q.; Lv, T.; Wu, W.; Cai, N.; Quan, G.; Yang, W.; Chen, Y.; Luo, L.; et al. Deep iterative reconstruction estimation (DIRE): Approximate iterative reconstruction estimation for low dose CT imaging. Phys. Med. Biol. 2019, 64, 135007. [Google Scholar] [CrossRef] [PubMed]
  20. Wang, J.; Liang, J.; Cheng, J.; Guo, Y.; Zeng, L. Deep learning based image reconstruction algorithm for limited-angle translational computed tomography. PLoS ONE 2020, 15, e0226963. [Google Scholar] [CrossRef]
  21. Xie, H.; Shan, H.; Wang, G. Deep Encoder-Decoder Adversarial Reconstruction (DEAR) Network for 3D CT from Few-View Data. Bioengineering 2019, 6, 111. [Google Scholar] [CrossRef]
  22. Xie, S.; Zheng, X.; Chen, Y.; Xie, L.; Liu, J.; Zhang, Y.; Yan, J.; Zhu, H.; Hu, Y. Artifact removal using improved GoogLeNet for sparse-view CT reconstruction. Sci. Rep. 2018, 8, 6700. [Google Scholar] [CrossRef]
  23. Xu, L.; Zeng, X.; Huang, Z.; Li, W.; Zhang, H. Low-dose chest X-ray image super-resolution using generative adversarial nets with spectral normalization. Biomed. Signal Process. Control 2020, 55, 101600. [Google Scholar] [CrossRef]
  24. Khodajou-Chokami, H.; Hosseini, S.A.; Ay, M.R. A deep learning method for high-quality ultra-fast CT image reconstruction from sparsely sampled projections. Nucl. Instruments Methods Phys. Res. Sect. Accel. Spectrometers Detect. Assoc. Equip. 2022, 1029, 166428. [Google Scholar] [CrossRef]
  25. Wang, T.; He, M.; Shen, K.; Liu, W.; Tian, C. Learned regularization for image reconstruction in sparse-view photoacoustic tomography. Biomed. Opt. Express 2022, 13, 5721–5737. [Google Scholar] [CrossRef] [PubMed]
  26. Gong, K.; Catana, C.; Qi, J.; Li, Q. Direct Reconstruction of Linear Parametric Images From Dynamic PET Using Nonlocal Deep Image Prior. IEEE Trans. Med Imaging 2022, 41, 680–689. [Google Scholar] [CrossRef]
  27. Çiçek, Ö.; Abdulkadir, A.; Lienkamp, S.S.; Brox, T.; Ronneberger, O. 3D U-Net: Learning dense volumetric segmentation from sparse annotation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention; Springer: Berlin/Heidelberg, Germany, 2016; pp. 424–432. [Google Scholar]
  28. Häggström, I.; Schmidtlein, C.R.; Campanella, G.; Fuchs, T.J. DeepPET: A deep encoder–decoder network for directly solving the PET image reconstruction inverse problem. Med Image Anal. 2019, 54, 253–262. [Google Scholar] [CrossRef] [PubMed]
  29. Kandarpa, V.S.S.; Bousse, A.; Benoit, D.; Visvikis, D. DUG-RECON: A Framework for Direct Image Reconstruction Using Convolutional Generative Networks. IEEE Trans. Radiat. Plasma Med Sci. 2021, 5, 44–53. [Google Scholar] [CrossRef]
  30. Yokota, T.; Kawai, K.; Sakata, M.; Kimura, Y.; Hontani, H. Dynamic PET Image Reconstruction Using Nonnegative Matrix Factorization Incorporated With Deep Image Prior. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  31. Zhang, Y.; Hu, D.; Hao, S.; Liu, J.; Quan, G.; Zhang, Y.; Ji, X.; Chen, Y. DREAM-Net: Deep Residual Error Iterative Minimization Network for Sparse-View CT Reconstruction. IEEE J. Biomed. Health Inform. 2023, 27, 480–491. [Google Scholar] [CrossRef]
  32. Xie, H.; Thorn, S.; Liu, Y.H.; Lee, S.; Liu, Z.; Wang, G.; Sinusas, A.J.; Liu, C. Deep-Learning-Based Few-Angle Cardiac SPECT Reconstruction Using Transformer. IEEE Trans. Radiat. Plasma Med Sci. 2023, 7, 33–40. [Google Scholar] [CrossRef]
  33. Hu, D.; Zhang, Y.; Liu, J.; Luo, S.; Chen, Y. DIOR: Deep Iterative Optimization-Based Residual-Learning for Limited-Angle CT Reconstruction. IEEE Trans. Med Imaging 2022, 41, 1778–1790. [Google Scholar] [CrossRef]
  34. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef]
  35. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 1646–1654. [Google Scholar]
  36. Wang, X.; Yi, J.; Guo, J.; Song, Y.; Lyu, J.; Xu, J.; Yan, W.; Zhao, J.; Cai, Q.; Min, H. A Review of Image Super-Resolution Approaches Based on Deep Learning and Applications in Remote Sensing. Remote Sens. 2022, 14. [Google Scholar] [CrossRef]
  37. Arefin, M.R.; Michalski, V.; St-Charles, P.L.; Kalaitzis, A.; Kim, S.; Kahou, S.E.; Bengio, Y. Multi-image super-resolution for remote sensing using deep recurrent networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 206–207. [Google Scholar]
  38. Yu, Y.; Li, X.; Liu, F. E-DBPN: Enhanced deep back-projection networks for remote sensing scene image superresolution. IEEE Trans. Geosci. Remote Sens. 2020, 58, 5503–5515. [Google Scholar] [CrossRef]
  39. Chen, S.; Han, Z.; Dai, E.; Jia, X.; Liu, Z.; Xing, L.; Zou, X.; Xu, C.; Liu, J.; Tian, Q. Unsupervised image super-resolution with an indirect supervised path. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 468–469. [Google Scholar]
  40. Soh, J.W.; Cho, S.; Cho, N.I. Meta-Transfer Learning for Zero-Shot Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  41. Gao, L.; Sun, H.M.; Cui, Z.; Du, Y.; Sun, H.; Jia, R. Super-resolution reconstruction of single remote sensing images based on residual channel attention. J. Appl. Remote Sens. 2021, 15, 016513. [Google Scholar] [CrossRef]
  42. Chen, X.; Wang, X.; Lu, Y.; Li, W.; Wang, Z.; Huang, Z. RBPNET: An asymptotic Residual Back-Projection Network for super-resolution of very low-resolution face image. Neurocomputing 2020, 376, 119–127. [Google Scholar] [CrossRef]
  43. Chen, C.; Gong, D.; Wang, H.; Li, Z.; Wong, K.Y.K. Learning Spatial Attention for Face Super-Resolution. IEEE Trans. Image Process. 2021, 30, 1219–1231. [Google Scholar] [CrossRef]
  44. Wang, J.; Wu, Y.; Wang, L.; Wang, L.; Alfarraj, O.; Tolba, A. Lightweight Feedback Convolution Neural Network for Remote Sensing Images Super-Resolution. IEEE Access 2021, 9, 15992–16003. [Google Scholar] [CrossRef]
  45. Xu, Y.; Luo, W.; Hu, A.; Xie, Z.; Xie, X.; Tao, L. TE-SAGAN: An Improved Generative Adversarial Network for Remote Sensing Super-Resolution Images. Remote Sens. 2022, 14, 2425. [Google Scholar] [CrossRef]
  46. Li, X.; Zhang, D.; Liang, Z.; Ouyang, D.; Shao, J. Fused Recurrent Network Via Channel Attention For Remote Sensing Satellite Image Super-Resolution. In Proceedings of the 2020 IEEE International Conference on Multimedia and Expo (ICME), London, UK, 6–10 June 2020; pp. 1–6. [Google Scholar] [CrossRef]
  47. Klajnšek, G.; Žalik, B. Progressive lossless compression of volumetric data using small memory load. Comput. Med. Imaging Graph. 2005, 29, 305–312. [Google Scholar] [CrossRef]
  48. Senapati, R.K.; Prasad, P.M.K.; Swain, G.; Shankar, T.N. Volumetric medical image compression using 3D listless embedded block partitioning. SpringerPlus 2016, 5, 2100. [Google Scholar] [CrossRef]
  49. Guthe, S.; Goesele, M. GPU-based lossless volume data compression. In Proceedings of the 2016 3DTV-Conference: The True Vision—Capture, Transmission and Display of 3D Video (3DTV-CON), Hamburg, Germany, 4–6 July 2016; pp. 1–4. [Google Scholar] [CrossRef]
  50. Nguyen, K.G.; Saupe, D. Rapid High Quality Compression of Volume Data for Visualization. Comput. Graph. Forum 2001, 20, 49–57. [Google Scholar] [CrossRef]
  51. Dai, Q.; Song, Y.; Xin, Y. Random-Accessible Volume Data Compression with Regression Function. In Proceedings of the 2015 14th International Conference on Computer-Aided Design and Computer Graphics (CAD/Graphics), Xi’an, China, 26–28 August 2015; pp. 137–142. [Google Scholar] [CrossRef]
  52. Shen, H.; David Pan, W.; Dong, Y.; Alim, M. Lossless compression of curated erythrocyte images using deep autoencoders for malaria infection diagnosis. In Proceedings of the 2016 Picture Coding Symposium (PCS), Nuremberg, Germany, 4–7 December 2016; pp. 1–5. [Google Scholar] [CrossRef]
  53. Mishra, D.; Singh, S.K.; Singh, R.K. Lossy Medical Image Compression using Residual Learning-based Dual Autoencoder Model. arXiv 2021, arXiv:2108.10579. [Google Scholar] [CrossRef]
  54. Moraes, T.; Amorim, P.; Silva, J.V.D.; Pedrini, H. Medical image interpolation based on 3D Lanczos filtering. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2020, 8, 294–300. [Google Scholar] [CrossRef]
  55. Tong, T.; Li, G.; Liu, X.; Gao, Q. Image Super-Resolution Using Dense Skip Connections. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4809–4817. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.