Next Article in Journal
Measurement of Trunk Movement during Sit-to-Stand Motion Using Laser Range Finders: A Preliminary Study
Next Article in Special Issue
Subjective Quality Assessment of V-PCC-Compressed Dynamic Point Clouds Degraded by Packet Losses
Previous Article in Journal
A Bayesian Network Approach to Explainable Reinforcement Learning with Distal Information
Previous Article in Special Issue
Compression for Bayer CFA Images: Review and Performance Comparison
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Light Field Image Super-Resolution Using Deep Residual Networks on Lenslet Images

1
School of Information and Communication Engineering, College of Electrical and Computer Engineering, Chungbuk National University, Cheongju 28644, Republic of Korea
2
Electrical Engineering Department, Faculty of Engineering, Assiut University, Assiut 71515, Egypt
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(4), 2018; https://doi.org/10.3390/s23042018
Submission received: 27 December 2022 / Revised: 6 February 2023 / Accepted: 7 February 2023 / Published: 10 February 2023

Abstract

:
Due to its widespread usage in many applications, numerous deep learning algorithms have been proposed to overcome Light Field’s trade-off (LF). The sensor’s low resolution limits angular and spatial resolution, which causes this trade-off. The proposed method should be able to model the non-local properties of the 4D LF data fully to mitigate this problem. Therefore, this paper proposes a different approach to increase spatial and angular information interaction for LF image super-resolution (SR). We achieved this by processing the LF Sub-Aperture Images (SAI) independently to extract the spatial information and the LF Macro-Pixel Image (MPI) to extract the angular information. The MPI or Lenslet LF image is characterized by its ability to integrate more complementary information between different viewpoints (SAIs). In particular, we extract initial features and then process MAI and SAIs alternately to incorporate angular and spatial information. Finally, the interacted features are added to the initial extracted features to reconstruct the final output. We trained the proposed network to minimize the sum of absolute errors between low-resolution (LR) input and high-resolution (HR) output images. Experimental results prove the high performance of our proposed method over the state-of-the-art methods on LFSR for small baseline LF images.

1. Introduction

Light field (LF) captures the intensity and direction of light rays reflected by objects in three-dimensional surroundings. Unlike conventional imaging, which captures the 2D projection of light rays, LF imaging collects data with many dimensions [1]. This abundance of visual information in LF pictures, in addition to their immersive description of the real world, may help several image processing and computer vision tasks, such as depth estimation [2,3], de-occlusion [4,5], salient object detection [6,7], and image post-refocus [8].
Nonetheless, obtaining LF data using plenoptic cameras, such as Raytrix [9], compromises spatial and angular resolutions. Due to a restricted sensor resolution, a plenoptic camera must reduce the spatial resolution of each view to collect more images at a higher angular sampling rate or conversely. Improving LF images’ resolutions is vital, as low-resolution images diminish the performance of low-frequency vision applications. In this paper, we study the LF super-resolution (LFSR) problem.
LFSR methods anticipate subpixel information using the difference between adjacent views [10,11,12,13]. Where these adjacent view images are closely connected in LF, sub-pixel information in each view image may be calculated by leveraging this cross-view correlation, allowing for its super-resolution (SR) reconstruction.
Several deep learning methods with diverse network topologies [14,15,16,17,18,19,20,21,22,23] were recently proposed to accomplish LFSR with enormous LF datasets [24,25,26,27,28]. These methods provide various learning-based SR methods using cross-view correlation through convolutional neural networks (CNN) and transformer-based networks. Although LFSR performance has been steadily improved via careful network design, most present LFSR algorithms underutilize the rich angular information, resulting in performance deterioration, particularly in complicated scenarios. For example, in [19,23], they only considered the spatial and angular information to model the non-local properties of the 4D LF. These methods have been improved upon in [21], by extracting horizontal and vertical epipolar information and spatial and angular information to improve the quality of the resulting images. We argue that Lenslet LF images can provide epipolar and angular information more compactly, allowing the network to model the relationship better and provide more pleasing results [29].
Therefore, this paper proposes a learning-based approach to obtain spatial SR using Lenslet LF images. The Lenslet image is formed by mapping the 4D images into one image using a Periodic Shuffling Operator (PS). We first extract features from input LF images independently. Then, we map the extracted 4D features into Lenslet 2D features. These features are then processed using a sequence of residual groups (RGs) to enhance the spatial resolution and restore some missing details. However, using Lenslet LF images has a blurring effect on the images, which increases with the LF images’ disparity value. Therefore, we process the Lenslet image and sub-aperture images alternately to incorporate angular and spatial information and mitigate the blur effect caused by processing the Lenslet image only. Finally, the final features are aggregated to reconstruct the output image. The quality of the super-resolved images is inversely proportional to the maximum disparity of the scene. In the case of a large disparity, the result is modest and increases inversely with the disparity. In the case of LF with a small disparity, our model comes first in LFSR quality compared with other SR methods.
We conducted several experiments to show our model’s performance in LF super-resolution. The main contributions of our paper are as follows:
  • We propose a different paradigm to increase the spatial-angular interaction by processing the Lenslet image and sub-aperture images to incorporate more information for LFSR.
  • We propose a CNN-based network to work for LFSR using Lenslet Images with superior performance over the state-of-the-art methods in the case of small-baseline LFSR.
  • The remainder of the paper is structured as follows: Section 2 briefly examines the related work. In Section 3, we present our technique for LFSR. Section 4 introduces the conducted experiments to compare our work with the state-of-the-art and discusses the meaning of the obtained results. Finally, Section 5 brings this paper to a close and presents future work to improve the proposed work.

2. Related Work

The objective of LF spatial SR, also known as LF image SR, is to produce high-resolution (HR) LF pictures from low-resolution (LR) inputs. Applying independent single image SR (SISR) algorithms to each sub-aperture picture is a basic way of achieving LF spatial SR (SAI). However, straight SISR for LF spatial SR cannot yield adequate results due to the absence of correlation between distinct viewpoints. Therefore, state-of-the-art methods try to fully model the non-local properties of the 4D LF data by using the information inside every single view (i.e., spatial information) and between multiple views (i.e., angular information) to perform well in LFSR. Since using CNNs in the field of LFSR, their utilization has grown exponentially, and the reconstruction performance has improved continually. Zhang et al. [14] presented a residual network for LFSR. The images in four directions are first stacked and given to separate branches to extract sub-pixel correlations. Then the information from these branches is combined for the final reconstruction. Because just a few side views may be employed, the performance of side views will suffer when compared to the performance of the center view, resulting in undesired inconsistencies in the reconstructed LF pictures. The performance of their later work [15] was increased considerably by using 3D convolutions applied to view image stacks of diverse angle orientations. Jin et al. [16] utilized an all-to-one technique for LFSR and used structural consistency regularization to retain the parallax structure. Yeung et al. [17] proposed to shuffle LF spatial-angular features alternately at a single forward pass using separable convolutions. Wang et al. [18] used deformable convolution on LF images to overcome the disparity problem for LFSR. Wang et al. [19] presented an interactive network (LF-InterNet). In particular, spatial and angular features are extracted and repeatedly interact to extract complementary information step by step. Then, each view image is super-resolved by fusing the interacting features. A network with two parallel branches was suggested by Liu et al. [20]. The top one collects global interview data. The bottom one separately projects each view to deep representations and then models the correlations between all intra-view characteristics using a multi-view context block. Wang et al. [21] designed a disentangling approach by dividing LF into several subspaces. They extracted features using three feature extractors (spatial, angular, horizontal, and vertical epipolar information). The network’s convolution layers only need to analyze information in a single subspace, facilitating LF representation learning. Different from CNN and inspired by recent achievements in Transformers [30], Wang et al. [22] developed a detail-preserving Transformer (DPT) to recover the features of light field (LF) pictures by using gradient maps of light field to direct sequence learning. However, the frameworks of these techniques are all-inclusive models whose supplementary information is not effectively employed for performance enhancement. Later, Liang et al. [23] suggested a Transformer-based LF image SR network in which a spatial Transformer and an angular Transformer were built to simulate, respectively, long-range spatial interdependence and angular correlation.

3. Our Approach

3.1. Problem Formulation

In our approach, the LF is formulated as a 2D array of Sub-Aperture Images (SAI), as shown in Figure 1c, and given by L   R u   ×   v   ×   h   ×   w , with (h, w) and (u, v) spatial and angular resolutions. Therefore, given a low-resolution LF input as L LR   R u   ×   v   ×   h   ×   w , we aim to reconstruct its high-resolution counterpart L HR   R u   ×   v   ×   α h   ×   α w , by enhancing the spatial resolution, where α represents the super-resolution factor. Following recent approaches [14,15,16,17,18,19,20,21,22,23], we set = 2 , 4 and assume that SAIs are distributed in a square array, i.e., u = v = A, where A represents the vertical or horizontal angular resolution. Before feeding L LR to the network, we up-sample the input LF epipolar plane images (EPIs) to the desired output size with α utilizing the Bicubic interpolation. Finally, we arrange the input from the 4D representation into the 3D representation L LR   R uv   ×   h   ×   w .

3.2. Features Extractors

We aim to extract spatial information along with horizontal epipolar, vertical epipolar, and angular information to model the non-local properties of the 4D LF fully. We utilize a convolution filter with a kernel of size 3 × 3 as a spatial feature extractor, as shown in black in Figure 2. This filter is applied to SAI separately. To extract horizontal epipolar, vertical epipolar, and angular information, we utilize another convolution filter with a kernel of size 3 × 3 as a Lenslet feature extractor, as shown in yellow in Figure 2. However, we apply this filter to the Lenslet image. The idea behind operating on the Lenslet image is that rows of the Lenslet image represent horizontal epipolar lines, columns of the Lenslet image represent vertical epipolar lines, and pixels in the Lenslet image represent the angular information. Therefore, one of the key benefits of dealing with a Lenslet image is extracting a large amount of information with a single convolution filter.

3.3. Network Overview

In our method, we process the Lenslet LF similar to the method proposed in [29] for LF angular super-resolution. At the same time, the proposed architecture is designed similarly to the deep residual channel attention networks [31]. Figure 3 depicts the overall design of our network. The proposed network consists of three cascaded stages: initial feature extraction, convolutional neural network (CNN)-based super-resolution network, and final image reconstruction, as shown in Figure 3a. The first and last stages of the network consist of a single 3 × 3 convolution layer, and the middle consists of a long skip connection with cascaded residual angular and residual spatial groups (RG). The angular and spatial groups share the same structure, as shown in Figure 3b. The input LF is processed differently by different network components. For example, the initial feature extractor and spatial groups process each 2D view image VI LR   R h   ×   w independently. While angular groups and the final reconstruction stage process the 2D Lenslet image L LR   R uh   ×   vw .
The mapping between the 3D LF   R uv   ×   h   ×   w and the 2D Lenslet LF   R uh   ×   vw is done using a periodic shuffling operator (PS) [32,33], as shown in Figure 4. In Figure 3a, the three arrows before each block indicate that this block processes each view image independently, while other blocks process the Lenslet image. Initial features are extracted in the first stage of the network to be fed to the main part of the network for processing, where features are extracted from each view image independently and then rearranged to the Lenslet image using a PS. The main part of the network consists of a long skip connection with cascaded angular and spatial RGs. Each RG consists of three residual blocks (RB) in our implementation. Each RB has three cascaded convolution layers with a ReLU in between with a skip connection.

3.4. Loss Function and Training Details

We trained our network only on the luminance component while we upsample the chrominance components using the Bicubic interpolation. We trained our network to learn a mapping from the Low-Resolution (LR) LF image LF LR to the High-Resolution (HR) LF image LF HR . We can write the problem can as follows:
LF HR = f ( LF LR ,   θ )
where f ( . ) is the function mapping from the LR image to the HR image, and θ is the network parameters to be learned while training.
We trained the proposed network to reduce the L1 distance. L1 loss is defined as follows, given a training set with N pairs of LR input and HR output images:
L 1 ( θ ) = 1 N i = 1 N | LF HR i f ( LF LR i ) | 1
Following recent approaches [14,15,16,17,18,19,20,21,22,23], we used 144 LF images for training and 23 for testing from publicly available synthetic and real-world datasets [24,25,26,27,28], as shown in Table 1. The original angular resolution of these datasets is 9 × 9, while we used the middle 5 × 5 views for training and testing. These datasets are divided into three categories based on their disparity value. For example, small-disparity LF includes EPFL and INRIA, medium-disparity LF includes HCInew and HCIold, and large-disparity LF includes STFgantry. The HCInew and HCIold are Synthetic, while other datasets are real-world datasets captured by the Lytro Illum camera.
We extract patches of size 32 × 32 with a stride of one from input and ground-truth images to prepare the training dataset. Our model was trained by ADAM optimizer [34] with β1 = 0.9, β2 = 0.999, and ǫ = 10−8. The initial learning rate is set to 2 × 10−4 and then decreases exponentially by 0.1 every 80 epochs. Our model was trained for 85 epochs in Tensorflow [35] with NVIDIA TITAN RTX GPU.

4. Experiments and Discussion

4.1. Comparison with the State-of-the-Art Methods

To demonstrate the performance of our model in the LFSR task, we compare it with state-of-the-art single image SR (SISR) methods, including VDSR [36], EDSR [37], RCAN [31], and LFSR methods, including resLF [14], MEG-Net [15], LF-ATO [16], LFSSR [17], LF-InterNet [19], LF-DFnet [18], LF-IINet [20], DPT [22], LFT [23], DistgSSR [21] with all models retrained with the same datasets.

4.1.1. Quantitative Comparison

Average PSNR and SSIM are used for the luminance images over all the output views to measure the super-resolution quality. We present numerical results in terms of (PSNR/SSIM) in Table 2 and for 2× and 4× LFSR, respectively. The best results are shown in red, and the second-best results are in blue. The quality of the super-resolved images is inversely proportional to the maximum disparity of the scene. In the case of a large disparity, the result is modest and increases inversely with the disparity. In the case of LF with a small disparity, our model comes first compared with other SR methods and achieves competitive PSNR and SSIM. For example, 0.96 dB and 0.23 dB are higher than the LFT method [23] on the EPFL dataset for 2× and 4×, respectively. Moreover, 1.85 dB and 0.88 dB are higher than the LFT method [23] on the INRIA dataset for 2× and 4×, respectively.

4.1.2. Qualitative Comparison

We compare our results for 2× LFSR visually with one of the state-of-the-art methods [23], as shown in Figure 5. However, it isn’t easy to differentiate between output images as they are very similar, so we urge the reader to check Table 2 and Table 3 for more details. We attribute the significant improvement in the case of small-disparity LF to Lenslet images, which permits the network to accurately understand and simulate the linking between different views of the same scene, recover more texture information, and improve the quality.

4.1.3. Model Efficiency

We compare our proposed model to several competitive methods in terms of the number of parameters, average PSNR, and average SSIM scores. As presented in Table 4, our model achieves the highest PSNR score and second-highest SSIM score for 2× LFSR. For 4× LFSR, our model achieves the second-highest PSNR score and second SSIM score. Except for LFT [23], a transformer-based model, our model is considered the best regarding the number of parameters and average PSNR and SSIM.

4.2. Ablation Study

In this subsection, we validate the effectiveness of two parameters on the final results, including our approach to using Lenslet extractors along with spatial extractors and the patch size used to train the network.

4.2.1. Feature Extractors

As shown in Table 5, we examine three alternatives to highlight the impact of the model’s various components. First, we train the proposed model utilizing the spatial feature extractor only and then utilizing the Lenslet feature extractor only. Finally, we train the proposed model using spatial and Lenslet feature extractors. It is clear from the results that the spatial extractor or Lenslet extractor alone cannot super-resolve LF images with high quality. It is also shown that even though the Lenslet extractor cannot super-resolve high-quality LF images, it may give better results when compared to the spatial extractor alone as it can extract angular information along with the epipolar information.
However, when comparing the model’s efficiency trained using only Lenslet extractor on the small-disparity LF, including EPFL and INRIA, this model performs better than the state-of-the-art methods, which validates our point that Lenslet images can provide a huge amount of information regarding the relationship between different view images.

4.2.2. Patch Size

As shown in Table 6 and Table 7, we examine the effect of patch size on the model using three different sizes 16 × 16, 32 × 32, and 64 × 64. When training for 2× LFSR, the 16 × 16 shows a slightly better performance than the 32 × 32. However, when training for 4× LFSR, the 32 × 32 achieves the best performance. The patch size affects the model trained for 4× LFSR more, especially with the degradation of the input image more than the 2× model.

5. Conclusions and Future Work

In this paper, we proposed a residual convolutional network for LFSR. To effectively explore the non-local property of 4D LF, we adopted the Lenslet LF representation. The Lenslet representation is compact and can provide abundant information. The Lenslet rows represent horizontal epipolar lines, columns represent vertical epipolar lines, and pixels in the Lenslet image represent angular information. Therefore, we alternately process the Lenslet image and view images to incorporate angular and spatial information. We used five datasets for training and testing, and our proposed model achieves the highest average PSNR and the second-highest average PSNR on 2× and 4× LFSR, respectively. In addition, experimental results show that the spatial or Lenslet extractor alone cannot super-resolve LF pictures well. The Lenslet extractor can extract epipolar and angular information; therefore, even though it cannot super-resolve high-quality LF pictures, it may produce better results than the spatial extractor alone. Furthermore, using three different patch sizes, 16 × 16, 32 × 32, and 64 × 64 to train the model, we found that 32 × 32 achieves the best performance on 2× and 4× LFSR.
However, the Lenslet image can provide much information for adjacent views at the same location, but it fails to provide the same amount of information for scenes with large disparity. Therefore, our model shows high performance with images with small disparity and poor performance with images with high disparity because performance is inversely proportional to the maximum disparity in the scene. In the future, we plan to:
  • Shear LF images into different disparity levels; after shearing, the disparity value will become smaller, then our network can extract better features, as proposed in [38].
  • Use a parallax-attention module (PAM) as a final stage, where PAM was designed to capture a global correspondence in stereo images super-resolution [39].
  • Adopt a transformer-based architecture, where transformers can understand the local and global features that benefit images with large disparities, such as the vision [30] and swing transformers [40].

Author Contributions

Conceptualization, A.S. and H.I.; methodology, A.S. and H.I.; software, A.S.; formal analysis, A.S.; investigation, H.-S.K.; resources, H.-S.K.; data curation, A.S.; writing—original draft preparation, A.S.; writing—review and editing, A.S. and H.-S.K.; validation, A.S., H.-S.K.; visualization, H.-S.K.; supervision, H.-S.K.; project administration, H.-S.K.; funding acquisition, H.-S.K. All authors have read and agreed to the published version of the manuscript.

Funding

The Basic Science Research Program partially supported this research through the National Research Foundation of Korea (NRF), funded by the Ministry of Education under Grant 2020R1I1A3A04037680, and by the Research Projects of “Development of automatic screening and hybrid detection system for hazardous material detecting in port container” funded by the Ministry of Oceans and Fisheries.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used in this paper are public datasets. We also provide the proposed method’s test and evaluation codes at: https://github.com/ahmeddiefy/LFSR_Lenslet, created and accessed on 19 December 2022.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wu, G.; Masia, B.; Jarabo, A.; Zhang, Y.; Wang, L.; Dai, Q.; Chai, T.; Liu, Y. Light field image processing: An overview. IEEE J. Sel. Top. Signal Process. 2017, 11, 926–954. [Google Scholar] [CrossRef]
  2. Wang, Y.; Wang, L.; Liang, Z.; Yang, J.; An, W.; Guo, Y. Occlusion-Aware Cost Constructor for Light Field Depth Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 19809–19818. [Google Scholar]
  3. Gao, M.; Deng, H.; Xiang, S.; Wu, J.; He, Z. EPI Light Field Depth Estimation Based on a Directional Relationship Model and Multiviewpoint Attention Mechanism. Sensors 2022, 22, 6291. [Google Scholar] [CrossRef] [PubMed]
  4. Li, Y.; Yang, W.; Xu, Z.; Chen, Z.; Shi, Z.; Zhang, Y.; Huang, L. Mask4D: 4D convolution network for light field occlusion removal. In Proceedings of the ICASSP 2021—2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 2480–2484. [Google Scholar]
  5. Wang, Y.; Wu, T.; Yang, J.; Wang, L.; An, W.; Guo, Y. DeOccNet: Learning to see through foreground occlusions in light fields. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 118–127. [Google Scholar]
  6. Zhang, M.; Ji, W.; Piao, Y.; Li, J.; Zhang, Y.; Xu, S.; Lu, H. LFNet: Light field fusion network for salient object detection. IEEE Trans. Image Process. 2020, 29, 6276–6287. [Google Scholar] [CrossRef] [PubMed]
  7. Zhang, Y.; Chen, G.; Chen, Q.; Sun, Y.; Xia, Y.; Deforges, O.; Hamidouche, W.; Zhang, L. Learning synergistic attention for light field salient object detection. arXiv 2021, arXiv:2104.13916. [Google Scholar]
  8. Jayaweera, S.S.; Edussooriya, C.U.; Wijenayake, C.; Agathoklis, P.; Bruton, L.T. Multi-volumetric refocusing of light fields. IEEE Signal Process. Lett. 2020, 28, 31–35. [Google Scholar] [CrossRef]
  9. RayTrix. 3D Light Field Camera Technology. Available online: http://www.raytrix.de/ (accessed on 30 November 2022).
  10. Bishop, T.E.; Favaro, P. The light field camera: Extended depth of field, aliasing, and superresolution. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 34, 972–986. [Google Scholar] [CrossRef] [PubMed]
  11. Mitra, K.; Veeraraghavan, A. Light field denoising, light field superresolution and stereo camera based refocussing using a GMM light field patch prior. In Proceedings of the 2012 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Providence, RI, USA, 16–21 June 2012; pp. 22–28. [Google Scholar]
  12. Wanner, S.; Goldluecke, B. Variational light field analysis for disparity estimation and super-resolution. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 36, 606–619. [Google Scholar] [CrossRef] [PubMed]
  13. Rossi, M.; Frossard, P. Graph-based light field super-resolution. In Proceedings of the 2017 IEEE 19th International Workshop on Multimedia Signal Processing (MMSP), Luton, UK, 16–18 October 2017; pp. 1–6. [Google Scholar]
  14. Zhang, S.; Lin, Y.; Sheng, H. Residual networks for light field image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 11046–11055. [Google Scholar]
  15. Zhang, S.; Chang, S.; Lin, Y. End-to-end light field spatial super-resolution network using multiple epipolar geometry. IEEE Trans. Image Process. 2021, 30, 5956–5968. [Google Scholar] [CrossRef] [PubMed]
  16. Jin, J.; Hou, J.; Chen, J.; Kwong, S. Light field spatial super-resolution via deep combinatorial geometry embedding and structural consistency regularization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2260–2269. [Google Scholar]
  17. Yeung, H.W.F.; Hou, J.; Chen, X.; Chen, J.; Chen, Z.; Chung, Y.Y. Light field spatial super-resolution using deep efficient spatial-angular separable convolution. IEEE Trans. Image Process. 2018, 28, 2319–2330. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, Y.; Yang, J.; Wang, L.; Ying, X.; Wu, T.; An, W.; Guo, Y. Light field image super-resolution using deformable convolution. IEEE Trans. Image Process. 2020, 30, 1057–1071. [Google Scholar] [CrossRef] [PubMed]
  19. Wang, Y.; Wang, L.; Yang, J.; An, W.; Yu, J.; Guo, Y. Spatial-angular interaction for light field image super-resolution. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 November 2020; pp. 290–308. [Google Scholar]
  20. Liu, G.; Yue, H.; Wu, J.; Yang, J. Intra-inter view interaction network for light field image super-resolution. IEEE Trans. Multimedia 2021, 25, 256–266. [Google Scholar] [CrossRef]
  21. Wang, Y.; Wang, L.; Wu, G.; Yang, J.; An, W.; Yu, J.; Guo, Y. Disentangling light fields for super-resolution and disparity estimation. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 425–443. [Google Scholar] [CrossRef] [PubMed]
  22. Wang, S.; Zhou, T.; Lu, Y.; Di, H. Detail preserving transformer for light field image super-resolution. Proc. Conf. AAAI Artif. Intell. 2022, 36, 2522–2530. [Google Scholar] [CrossRef]
  23. Liang, Z.; Wang, Y.; Wang, L.; Yang, J.; Zhou, S. Light field image super-resolution with transformers. IEEE Signal Process. Lett. 2022, 29, 563–567. [Google Scholar] [CrossRef]
  24. Honauer, K.; Johannsen, O.; Kondermann, D.; Goldluecke, B. A dataset and evaluation methodology for depth estimation on 4D light fields. In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; pp. 19–34. [Google Scholar]
  25. Rerabek, M.; Ebrahimi, T. New light field image dataset. In Proceedings of the 8th International Conference on Quality of Multimedia Experience (QoMEX), Lisbon, Portugal, 6–8 June 2016. [Google Scholar]
  26. Wanner, S.; Meister, S.; Goldluecke, B. Datasets and benchmarks for densely sampled 4D light fields. In Vision Modeling and Visualization; The Eurographics Association: Munich, Germany, 2013; pp. 225–226. [Google Scholar]
  27. Le Pendu, M.; Jiang, X.; Guillemot, C. Light field inpainting propagation via low rank matrix completion. IEEE Trans. Image Process. 2018, 27, 1981–1993. [Google Scholar] [CrossRef] [PubMed]
  28. Vaish, V.; Adams, A. The (new) stanford light field archive. Comput. Graph. Lab. Stanf. Univ. 2008, 6, 3. [Google Scholar]
  29. Salem, A.; Ibrahem, H.; Kang, H.-S. Light Field Reconstruction Using Residual Networks on Raw Images. Sensors 2022, 22, 1956. [Google Scholar] [CrossRef] [PubMed]
  30. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S. An image is worth 16 × 16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  31. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 10–13 September 2018; pp. 286–301. [Google Scholar]
  32. Shi, W.; Caballero, J.; Huszár, F.; Totz, J.; Aitken, A.P.; Bishop, R.; Rueckert, D.; Wang, Z. Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1874–1883. [Google Scholar]
  33. Wang, L.; Guo, Y.; Lin, Z.; Deng, X.; An, W. Learning for video super-resolution through HR optical flow estimation. In Proceedings of the Asian Conference on Computer Vision, Perth, Australia, 2–6 December 2018; pp. 514–529. [Google Scholar]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Abadi, M.; Barham, P.; Chen, J.; Chen, Z.; Davis, A.; Dean, J.; Devin, M.; Ghemawat, S.; Irving, G.; Isard, M. Tensorflow: A system for large-scale machine learning. In Proceedings of the 12th {USENIX} Symposium on Operating Systems Design and Implementation ({OSDI} 16), Savannah, GA, USA, 2–4 November 2016; pp. 265–283. [Google Scholar]
  36. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  37. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  38. Wu, G.; Liu, Y.; Dai, Q.; Chai, T. Learning sheared EPI structure for light field reconstruction. IEEE Trans. Image Process. 2019, 28, 3261–3273. [Google Scholar] [CrossRef] [PubMed]
  39. Wang, L.; Wang, Y.; Liang, Z.; Lin, Z.; Yang, J.; An, W.; Guo, Y. Learning parallax attention for stereo image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 12250–12259. [Google Scholar]
  40. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
Figure 1. 4D vs. Lenslet LF representation. (a) 4D representation: 3 × 3 LF images where each image has four pixels, and each image is represented by a different color, (b) The mapping from the 4D LF of size (u, v, h, w) into the Lenslet image of size (u × h, v × w) using the shuffling operator, (c) a real example of LF scene represented by 3 × 3 images, (d) the Lenslet LF image of the 3 × 3 images in (c).
Figure 1. 4D vs. Lenslet LF representation. (a) 4D representation: 3 × 3 LF images where each image has four pixels, and each image is represented by a different color, (b) The mapping from the 4D LF of size (u, v, h, w) into the Lenslet image of size (u × h, v × w) using the shuffling operator, (c) a real example of LF scene represented by 3 × 3 images, (d) the Lenslet LF image of the 3 × 3 images in (c).
Sensors 23 02018 g001
Figure 2. An illustration of the spatial and Lenslet feature extractors. Here, we have 3 × 3 LF images with nine pixels in each view image and a different color. The spatial feature extractor is shown in black color, while the Lenslet extractor is shown in yellow. We utilize a convolution filter for both extractors with a kernel of size 3 × 3 and a stride of 1. The spatial extractor is applied to each view image independently, as shown in (a), while the Lenslet extractor is applied to the Lenslet image, as shown in (b).
Figure 2. An illustration of the spatial and Lenslet feature extractors. Here, we have 3 × 3 LF images with nine pixels in each view image and a different color. The spatial feature extractor is shown in black color, while the Lenslet extractor is shown in yellow. We utilize a convolution filter for both extractors with a kernel of size 3 × 3 and a stride of 1. The spatial extractor is applied to each view image independently, as shown in (a), while the Lenslet extractor is applied to the Lenslet image, as shown in (b).
Sensors 23 02018 g002
Figure 3. (a) The overall design of our network for LFSR. (b) The residual group (RG) design is used as an angular and spatial group in (a). (c) Residual block (RB) design.
Figure 3. (a) The overall design of our network for LFSR. (b) The residual group (RG) design is used as an angular and spatial group in (a). (c) Residual block (RB) design.
Sensors 23 02018 g003
Figure 4. Periodic shuffling (PS) operation (Depth-to-Space and Space-to-Depth). (a) A sequence of view images with size (H, W, D). (b) Lenslet Image with size (3H, 3W, D/9) contains all the view images in a different order. The mapping from (a) to (b) is called depth-to-space, while from (b) to (a) is called space-to-depth.
Figure 4. Periodic shuffling (PS) operation (Depth-to-Space and Space-to-Depth). (a) A sequence of view images with size (H, W, D). (b) Lenslet Image with size (3H, 3W, D/9) contains all the view images in a different order. The mapping from (a) to (b) is called depth-to-space, while from (b) to (a) is called space-to-depth.
Sensors 23 02018 g004
Figure 5. Visual comparison for different 2× LFSR methods. We chose one scene from each dataset for comparison. In addition, a close-up of image portions in red boxes is provided for Bicubic, LFT [23], our method, and ground truth, respectively.
Figure 5. Visual comparison for different 2× LFSR methods. We chose one scene from each dataset for comparison. In addition, a close-up of image portions in red boxes is provided for Bicubic, LFT [23], our method, and ground truth, respectively.
Sensors 23 02018 g005
Table 1. LF datasets used for training and testing.
Table 1. LF datasets used for training and testing.
DatasetTrainingTestingDisparityData Type
HCInew [24]204[−4, 4]Synthetic
HCIold [26]102[−3, 3]Synthetic
EPFL [25]7010[−1, 1]Real-world
INRIA [27]355[−1, 1]Real-world
STFgantry [28]92[−7, 7]Real-world
Table 2. Numerical comparison for different 2× LFSR methods. The best results are in red, and the second-best in blue.
Table 2. Numerical comparison for different 2× LFSR methods. The best results are in red, and the second-best in blue.
DatasetEPFLHCInewHCIoldINRIASTFgantryAverage
Bicubic29.74/0.94131.89/0.93937.69/0.97931.33/0.95931.06/0.95432.34/0.954
VDSR [36]32.50/0.96034.37/0.95640.61/0.98734.43/0.97435.54/0.97935.49/0.971
EDSR [37]33.09/0.96334.83/0.95941.01/0.98834.97/0.97736.29/0.98236.04/0.974
RCAN [31]33.16/0.96434.98/0.96041.05/0.98835.01/0.97736.33/0.98336.11/0.974
resLF [14]33.62/0.97136.69/0.97443.42/0.99335.39/0.98138.36/0.99037.50/0.982
LFSSR [17]33.68/0.97436.81/0.97543.81/0.99435.28/0.98337.95/0.99037.51/0.983
MEG-Net [15]34.30/0.97737.42/0.97844.08/0.99436.09/0.98538.77/0.99138.13/0.985
LF-ATO [16]34.27/0.97637.24/0.97744.20/0.99436.15/0.98439.64/0.99338.30/0.985
LF-InterNet [19]34.14/0.97237.28/0.97744.45/0.99535.80/0.98538.72/0.99238.08/0.984
LF-DFnet [18]34.44/0.97737.44/0.97944.23/0.99436.36/0.98439.61/0.99338.42/0.985
LF-IINet [20]34.68/0.97737.74/0.97944.84/0.99536.57/0.98539.86/0.99438.74/0.986
DPT [22]34.48/0.97637.35/0.97744.31/0.99436.40/0.98439.52/0.99338.41/0.984
LFT [23]34.80/0.97837.84/0.97944.52/0.99536.59/0.98640.51/0.99438.85/0.986
DistgSSR [21]34.80/0.97937.95/0.98044.92/0.99536.58/0.98640.37/0.99438.92/0.987
Ours35.76/0.97937.49/0.97944.50/0.99438.44/0.98639.16/0.99339.05/0.986
Table 3. Numerical comparison for different 4× LFSR methods. The best results are in red, and the second-best in blue.
Table 3. Numerical comparison for different 4× LFSR methods. The best results are in red, and the second-best in blue.
DatasetEPFLHCInewHCIoldINRIASTFgantryAverage
Bicubic25.14/0.83327.61/0.85332.42/0.93126.82/0.88625.93/0.84727.58/0.870
VDSR [36]27.25/0.87829.31/0.88334.81/0.95229.19/0.92128.51/0.90129.81/0.907
EDSR [37]27.84/0.88629.60/0.88735.18/0.95429.66/0.92628.70/0.90830.20/0.912
RCAN [31]27.88/0.88629.63/0.88835.20/0.95429.76/0.92728.90/0.91130.27/0.913
resLF [14]28.27/0.90430.73/0.91136.71/0.96830.34/0.94130.19/0.93731.25/0.932
LFSSR [17]28.27/0.90830.72/0.91236.70/0.96930.31/0.94530.15/0.93931.23/0.935
MEG-Net [15]28.74/0.91631.10/0.91837.28/0.97230.66/0.94930.77/0.94531.71/0.940
LF-ATO [16]28.52/0.91230.88/0.91437.00/0.97030.71/0.94930.61/0.94331.54/0.938
LF-InterNet [19]28.67/0.91430.98/0.91737.11/0.97230.64/0.94930.53/0.94331.59/0.939
LF-DFnet [18]28.77/0.91731.23/0.92037.32/0.97230.83/0.95031.15/0.94931.86/0.942
LF-IINet [20]29.11/0.92031.36/0.92137.62/0.97431.08/0.95231.21/0.95032.08/0.943
DPT [22]28.93/0.91731.19/0.91937.39/0.97230.96/0.95031.14/0.94931.92/0.941
LFT [23]29.25/0.92131.46/0.92237.63/0.97431.20/0.95231.86/0.95532.28/0.945
DistgSSR [21]28.98/0.91931.38/0.92237.55/0.97330.99/0.95231.63/0.95332.11/0.944
Ours29.48/0.92231.01/0.92137.17/0.97132.08/0.95330.83/0.95132.26/0.944
Table 4. Comparison of the number of model parameters and average PSNR and SSIM for 2× and 4× LFSR. The best results are in red, and the second-best in blue.
Table 4. Comparison of the number of model parameters and average PSNR and SSIM for 2× and 4× LFSR. The best results are in red, and the second-best in blue.
Dataset
#Param.PSNRSSIM#Param.PSNRSSIM
EDSR [37]38.6 M36.040.97438.9 M30.200.912
RCAN [31]15.3 M36.110.97415.4 M30.270.913
resLF [14]6.35 M37.500.9826.79 M31.250.932
LFSSR [17]0.81 M37.510.9831.61 M31.230.935
MEG-Net [15]1.69 M38.130.9851.77 M31.710.940
LF-ATO [16]1.51 M38.300.9851.66 M31.540.938
LF-InterNet [19]4.80 M38.080.9845.23 M31.590.939
LF-DFnet [18]3.94 M38.420.9853.99 M31.860.942
LF-IINet [20]4.84 M38.740.9864.89 M32.080.943
DPT [22]3.73 M38.410.9843.78 M31.920.941
LFT [23]1.11 M38.850.9861.16 M32.280.945
DistgSSR [21]3.53 M38.920.9873.58 M32.110.944
Ours3.21 M39.050.9863.21 M32.260.944
Table 5. Numerical comparison for variants of our proposed network using different feature extractors for 2× LFSR. The best results are in red, and the second-best in blue.
Table 5. Numerical comparison for variants of our proposed network using different feature extractors for 2× LFSR. The best results are in red, and the second-best in blue.
DatasetEPFLHCInewHCIoldINRIASTFgantryAverage
Spatial33.35/0.96434.55/0.96040.68/0.98735.58/0.97735.99/0.98336.03/0.974
Lenslet35.06/0.97636.40/0.97543.35/0.99337.63/0.98536.95/0.98937.88/0.984
Both35.76/0.97937.49/0.97944.50/0.99438.44/0.98639.16/0.99339.05/0.986
Table 6. Numerical comparison for variants of our proposed network trained with different patch sizes for 2× LFSR. The best results are in red, and the second-best in blue.
Table 6. Numerical comparison for variants of our proposed network trained with different patch sizes for 2× LFSR. The best results are in red, and the second-best in blue.
DatasetEPFLHCInewHCIoldINRIASTFgantryAverage
16 × 1635.78/0.98037.46/0.97944.34/0.99438.47/0.98739.36/0.99339.08/0.987
32 × 3235.76/0.97937.49/0.97944.50/0.99438.44/0.98639.16/0.99339.05/0.986
64 × 6435.46/0.97737.17/0.97744.14/0.99438.11/0.98538.98/0.993 38.77/0.985
Table 7. Numerical comparison for variants of our proposed network trained with different patch sizes for 4× LFSR. The best results are in red, and the second-best in blue.
Table 7. Numerical comparison for variants of our proposed network trained with different patch sizes for 4× LFSR. The best results are in red, and the second-best in blue.
DatasetEPFLHCInewHCIoldINRIASTFgantryAverage
16 × 1629.33/0.91830.63/0.91236.63/0.96731.44/0.94930.21/0.94031.65/0.937
32 × 3229.48/0.92231.01/0.92137.17/0.97132.08/0.95330.83/0.95132.26/0.944
64 × 6429.34/0.91830.80/0.91936.77/0.96931.84/0.95130.51/0.94831.85/0.941
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Salem, A.; Ibrahem, H.; Kang, H.-S. Light Field Image Super-Resolution Using Deep Residual Networks on Lenslet Images. Sensors 2023, 23, 2018. https://doi.org/10.3390/s23042018

AMA Style

Salem A, Ibrahem H, Kang H-S. Light Field Image Super-Resolution Using Deep Residual Networks on Lenslet Images. Sensors. 2023; 23(4):2018. https://doi.org/10.3390/s23042018

Chicago/Turabian Style

Salem, Ahmed, Hatem Ibrahem, and Hyun-Soo Kang. 2023. "Light Field Image Super-Resolution Using Deep Residual Networks on Lenslet Images" Sensors 23, no. 4: 2018. https://doi.org/10.3390/s23042018

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop