Next Article in Journal
A Finite-Time Sliding-Mode Controller Based on the Disturbance Observer and Neural Network for Hysteretic Systems with Application in Piezoelectric Actuators
Previous Article in Journal
A Fast Algorithm for Intra-Frame Versatile Video Coding Based on Edge Features
Previous Article in Special Issue
3D Reconstruction with Single-Shot Structured Light RGB Line Pattern
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Comprehensive High-Quality Three-Dimensional Display System Based on a Simplified Light-Field Image Acquisition Method and a Full-Connected Deep Neural Network

1
School of Information and Communication Engineering, Chungbuk National University, Chungbuk 28644, Republic of Korea
2
Department of Electronics Engineering, Incheon National University, Incheon 22012, Republic of Korea
*
Authors to whom correspondence should be addressed.
Sensors 2023, 23(14), 6245; https://doi.org/10.3390/s23146245
Submission received: 24 March 2023 / Revised: 4 July 2023 / Accepted: 5 July 2023 / Published: 8 July 2023
(This article belongs to the Special Issue Sensors for 3D Cameras System)

Abstract

:
We propose a high-quality, three-dimensional display system based on a simplified light field image acquisition method, and a custom-trained full-connected deep neural network is proposed. The ultimate goal of the proposed system is to acquire and reconstruct the light field images with possibly the most elevated quality from the real-world objects in a general environment. A simplified light field image acquisition method acquires the three-dimensional information of natural objects in a simple way, with high-resolution/high-quality like multicamera-based methods. We trained a full-connected deep neural network model to output desired viewpoints of the object with the same quality. The custom-trained instant neural graphics primitives model with hash encoding output the overall desired viewpoints of the object within the acquired viewing angle in the same quality, based on the input perspectives, according to the pixel density of a display device and lens array specifications within the significantly short processing time. Finally, the elemental image array was rendered through the pixel re-arrangement from the entire viewpoints to visualize the entire field-of-view and re-constructed as a high-quality three-dimensional visualization on the integral imaging display. The system was implemented successfully, and the displayed visualizations and corresponding evaluated results confirmed that the proposed system offers a simple and effective way to acquire light field images from real objects with high-resolution and present high-quality three-dimensional visualization on the integral imaging display system.

1. Introduction

Three-dimensional (3D) light field image acquisition techniques encode all 3D information, such as parallax, depth, shading, and the texture of real or virtual objects, into two-dimensional (2D) elemental images, using incoherent illumination [1,2]. These methods tend to be reliable, effective, and capable of supporting many types of 3D displays. The key to success is the acquisition of high-quality light field images. There are two main types of light field image acquisition techniques: single- and multicamera-based methods [3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22]. Single-camera methods can be divided into depth camera and lens array types. The depth camera consists of two sensors, i.e., infrared rays and color sensors, such that the 3D information of the real object is detected simultaneously [3,4,5,6]. The lens array-based single-camera method has traditionally been used for light field imaging as it employs a simple configuration in which the lens array is located between the camera lens and the sensor [7,8,9,10,11,12,13,14]. Here, each elemental lens of the lens array provides different viewpoints of the object; the full 3D information of the object is encoded in the disparity between the captured elemental images. By changing the specifications of the elemental lenses, the depth range of image acquisition, viewing angle, and resolution can be controlled. This method is a simple and efficient representative method capable of acquiring light field images. It was successful in reconstructing fine-quality point cloud models with data obtained from real-world objects through lens array-based light field cameras manufactured by Lytro Inc. [12]. Also, a virtual 3D model-free approach for recovering 3D facial geometry based on a single light field camera and a convolutional neural network using Epipolar plane images was introduced [13]. Thus, it was confirmed that such a lens array-based single-camera method is appropriate for light field imaging techniques. However, the resolution of each elemental image tends to be low as it is affected by optical aberrations of the lenses, resulting in poor image quality. Even though resolution and quality improvement methods based on convolutional neural networks have been applied, they are still insufficient for comfortable observation [13,15,16]. Multicamera-based methods, also known as camera array methods, acquire light field images with high resolution and high quality via multiple interconnected digital cameras [17,18,19,20,21,22]. The main configuration can be one-dimensional, i.e., horizontal or vertical parallax only, or 2D, such that it captures full-parallax images with each camera, providing different viewpoints of the object, similar to a lens array. Light field images can also be captured through parallel or toe-in methods, according to the application. Multicamera-type methods are used to obtain high-resolution and high-quality images of real objects from various perspectives; however, the structures tend to be bulky as they must accommodate a certain number of cameras, and the interconnections between cameras are complex. As such, additional processes are required to correct lens distortion and brightness in each camera. Therefore, it is important to implement the high-quality light field image acquisition method with a simple structure.
Recently, several methods of generating high-quality 3D models from given perspectives based on convolutional neural network models have been introduced [15,16,23,24,25,26,27,28]. For example, some methods estimate the depth of information from the perspectives and generate 3D models for the object [23,24], and other methods represent the view synthesis of 2D images as 3D visualizations [25,26,27,28]. Even though these methods can provide much higher quality 3D visualizations than before, it is difficult to reconstruct high-quality natural-view 3D images that meet our demands through these methods.
In this paper, we propose a comprehensive high-quality 3D display system based on a simplified light field image acquisition method, including a simple structure and high-resolution/high-quality image capturing, as with camera array methods, and a custom-trained full-connected deep neural network model. The main goal of the proposed system is to display the final reconstructed 3D image as high-quality as possible; the high-quality data are obtained from the acquisition stage and high-quality view synthesis is regenerated through a custom-trained full-connected deep neural network within a general environment. First, a moving camera array captures the high-resolution perspectives automatically with the toe-in configuration according to the inputted parameters; then, the brightness distribution is corrected with a histogram-matching process based on the average histogram. Since the proposed system requires the elemental image array (EIA) according to the feature of integral imaging display, a custom-trained full-connected deep neural network model is used to generate the overall viewpoints of the object, based on the corrected perspectives. Finally, the EIA is generated through the pixel re-arrangement method from the view synthesis. Optical experiments were performed to verify that the proposed method can serve as a simple and effective way to display natural-view high-quality light field images of real objects.

2. Proposed Method

Figure 1 shows a schematic configuration of the proposed high-quality 3D display system based on a simplified light field image acquisition method. The proposed system consists of main three sections: acquisition for high-resolution perspectives, generation of the view synthesis, and EIA generation and 3D display.

2.1. Simplified Light Field Image Acquisition Method

Figure 2a shows the simplified light field image acquisition method consisting of a minimized camera array and a multifunctional motor. Here, the camera array consists of Nc high-resolution digital cameras installed and fixed along the vertical axis. When the camera array acquires the light field images with the parallel method, all cameras focus on the object to capture the image. Note that the camera array can capture the perspectives via parallel or toe-in configurations, and a toe-in method is suitable for the proposed system. Here, the central camera focuses on the object as-is, and the other cameras rotate up or down at a given angle based on the central camera’s focusing position. A multifunctional smart motor is mounted on the stage such that the camera array can be shifted horizontally along the stage.
Figure 2b shows a flowchart of the automatic capturing process. The integrated controller supervises the overall process of capturing light field images from the real object. The computer is interconnected with the smart motor controller and camera shutter controller simultaneously and synchronizes the operations of the connected devices. Additionally, to interconnect the cameras and control their shutters simultaneously, we incorporated Raspberry Pi with the switch circuit board. The cameras are connected to the switch circuit board by universal serial bus ports, the switch circuit board is connected to the Raspberry Pi, and the Raspberry Pi is connected to the computer via a Wi-Fi connection using PuTTY, which is an open-source terminal emulator, serial console, and network file transfer application. In the integrated controller, the total number of perspectives (Np), a step of the X-motor responsible for shifting the interval of the camera array horizontally (Xi), a step of the Y-motor controlling the interval of rotation of the camera array (θi), and the initial position of the camera array corresponding to the position of the first perspectives for the X-motor (P0) and Y-motor (R0) are inputted by the user, where X-motor and Y-motor indicate the shifting and rotating units, respectively. Note that the integrated controller does not consider the number of cameras Nc. All the user must do is input the main parameters and execute the system operation and the system starts to capture the perspectives automatically. Here, the camera array is shifted and rotated to the P0 and R0 positions immediately after the main parameters are inputted, and the integrated controller sends the command to execute a Python script on Raspberry Pi to control the camera shutters; the first series of perspectives are captured and saved as the image files. Then, in each step of Xi and θi, the camera array captures the perspectives. When Np-th perspectives are completely captured, the smart motor and translation stage is stopped, and the camera shutters are deactivated; this disables the Raspberry Pi connection. Here, the cameras with only minimized numbers, Nc, are used (instead of Nc × Np cameras), but the desired Nc × Np perspective can be obtained; thus, the proposed system can be considered economically less burdensome. Note that Nc × Np perspectives are captured and stored immediately in the computer as image files, without using any additional data storage.
Before the generation of view synthesis, the color and brightness distribution of all images should be corrected in an identical manner as a pre-processing. This is because the location of the ambient light source and the shape of the object affect the brightness distribution of the images captured by each camera, i.e., the color of each image is slightly different. In the proposed system, after all the perspectives are captured, the average histogram value of the central perspectives captured by each camera is calculated. The histograms of the other perspectives are matched to the calculated average histogram, as shown in Figure 3. This process facilitates uniformity among the perspectives such that they are smoother and brighter and thus more similar to the reference histogram than the original images. Given that each perspective contains information from different viewpoints, the 3D light field display can be provided with high-quality 3D data in the proposed system.

2.2. Train for an Instant-NGP Model with Hash Encoding

Since the number of cameras is minimized, the 3D image cannot be displayed properly because the number of perspectives should be Np × Np, according to the 3D display feature. We considered an integral imaging 3D display including a high-resolution display device as the final output of the proposed system whereby intermediate-view image generation methods are utilized to fill the missing viewpoints; however, there is a disadvantage in that the generated image quality is too low, and it becomes worse when the process is iterated twice or more [1]. Therefore, in the proposed system, in order to display a high-quality 3D image properly, an instant neural graphics primitive (instant-NGP) model, a lightweight simple full-connected deep neural network [28], was trained to extract the high-quality view synthesis of the object in order to match Np × Np perspectives. Compared to the existing neural radiance fields (NeRF) method [26,27], the basic configuration and operation of the model are similar, but it has the advantage of being able to extract higher quality perspectives in a shorter processing time because the instant-NGP model creates a small neural network by performing the multiresolution hash encoding with the parallelized architecture of graphics processing units (GPUs). Note that the conventional NeRF method uses positional encoding to represent a high-frequency function and samples from the high-frequency function using a hierarchical sampling procedure to form a multilayer perceptron. However, the high-frequency function is computationally complicated, and frequency collisions often occur; thus, using a double hierarchical sampling procedure is very slow for multilayer perceptron training.
Figure 4 shows the illustration of the view synthesis generation of the object through the custom-trained instant-NGP model. The general procedure of the model is generating a high-quality 3D scene including all viewpoints of the object from the input perspectives and rendering the high-quality 2D perspectives. As mentioned before, an instant-NGP uses a small network through multiresolution hash encoding with GPU parallelized architecture. In the hash encoding, first, a spaced grid of input perspectives is created from wide to fine at levels from 21 up to 24 (L). The number of L defines how many varieties of grids are created where l indicates the index of level. The sizes of the widest and finest grids are determined as Smin and Smax respectively, and each grid has a different value of corner indices. From Figure 4, the corner indices of the widest and finest grids are illustrated as (cw1, cw2, cw3, cw4) and (cf1, cf2, cf3, cf4), and the average values of corner indices in each grid cannot be equal to other grids’ values. This process gives an opportunity to eliminate hash collisions and the basis of more accurate calculations. Then, the corner indices that are the corresponding hashed integer coordinates are stored at the corresponding hash tables (lookup tables), and they are linearly interpolated according to the relative position in the input 3D scene. Note that a linear interpolation method is a curve fitting using linear polynomials to construct new viewpoints within the range of a discrete set of overall viewing zones, and linear interpolation-based intermediate-view image generation methods are widely used in light field imaging techniques [1,2]. Finally, the full-connected deep neural network extracts Np × Np perspectives according to the camera poses, which are determined by the view parameters of the input perspectives.
Unlike the NeRF, the instant-NGP model uses an Adam optimizer in the training, which is implemented by the tiny-cuda-nn framework [28]. According to this framework, we trained the instant-NGP model to extract the desired view synthesis from the input perspectives through the following processing steps, as shown in Figure 5. Here, first, the model determines the camera poses of input perspectives according to the pre-defined view parameters. From the camera poses of input perspectives, the grids are generated with L levels, and hashes are encoded with linear interpolation. Then, the multilayer perceptron proceeds according to the encoded hashes and defines the camera poses of Np × Np cameras. Simply put, the multilayer perceptron is a deep neural network that expects the 3D volumetric scene of the original object, and a lot of iterations are performed in order to obtain a good-quality multilayer perceptron. Note that the camera pose is a ray tracing calculation in the virtual environment that is performed for each image. Since the instant-NGP uses a volumetric rendering method, an imaginary 3D volume of the object is depicted via the COLMAP structure from the motion package, derived from the camera poses of input images in the virtual space. When the Np × Np camera pose was created, we set the central row as 0° vertically; the central camera of the central row was set to look straight to an imaginary 3D volume, and the interval between each neighboring camera in the same row was set as the θi interval. Also, the interval between each neighboring row was set as the same value. Finally, in the multilayer perceptron, we needed to check whether the top, bottom, and center rows of the Np × Np camera poses matched the input camera poses and whether the related perspectives were rendered. Since the instant-NGP model can be performed 100 times faster than the conventional NeRF model and is trained with the designed scheme, it can obtain a higher-quality view synthesis with less loss of quality in the proposed system. Note that the training parameters should consider the EIA specifications of the next stage.

2.3. Pixel Re-Arrangement Method-Based EIA Generation

Finally, the EIA is generated from the rendered perspectives and displayed on the display unit. Here, the EIA is generated by the pixel re-arrangement method to visualize the entire encoded field-of-view for which each pixel of EIA is re-arranged from the newly rendered perspectives [29,30]. First, the main parameters including the number of virtual cameras, the distance between neighboring cameras, and the distance between the camera and image planes are optimized. Note that the field-of-view of the virtual camera array is set as the acquired viewing angle of the object. Then, each elemental image with the (i,j) index, EIij, can be calculated as follows:
E I i j s , t = P s t i , j ; s = 1 , , N p ,   t = 1 , , N p ,   i = 1 , , m ,   j = 1 , , n ,
where Pst is the set-of perspective images, s and t denote the position of the Np × Np virtual cameras along horizontal and vertical axes, respectively, and m and n are the numbers of pixels of each virtual perspective in horizontal and vertical directions, respectively. Here, the pixel number that belongs to the single elemental lens is the most important parameter in the EIA generation. Accordingly, the entire EIA is generated and displayed on the high-resolution display device and directly reconstructed as a natural-view 3D image when the lens array is mounted in front of the display device. This kind of EIA generation keeps the overall acquired field-of-view of the object on the 3D display.

3. Experiment and Results

The overall stages and related processes of the proposed system were designed and analyzed, and the following experiments and measurements were conducted by the authors to confirm the characteristics and superiority of the proposed system.

3.1. Simplified Light Field Image Acquisition Performance

Figure 6 shows the prototype of the proposed light field image acquisition system. The experimental setup consisted of a computer with an Intel Core i7-12700KF (3.60 GHz) central processing unit, 32 GB of memory, an NVIDIA GeForce RTX 2070 GPU, and a semiconductor-based storage device with 512 GB capacity; the camera array comprised three Sony α6000 high-resolution digital cameras (Nc = 3; resolution: 6000 × 4000 pixels) and a multifunctional smart motor with a Suruga Seiki DS102 controller on a translation stage 600 mm in length. The minimum shifting distance of the X-motor was 0.01 mm (the maximum distance depended on the length of the translation stage, which was 600 mm in the experiment) and the minimum rotating angle of the Y-motor was 0.004° (the maximum was unlimited, i.e., 360°). With this setup, it was possible to capture a sufficiently large number of perspectives within the length range of the translation stage. To interconnect the cameras and control the shutters simultaneously, Raspberry Pi3 was incorporated into the switch circuit board via general-purpose input/output pin connections.
Figure 7a shows the user interface of the integrated controller. The controller was built on the basis of LabVIEW software and synchronized with the operation of the smart motor and camera shutters. When the “Capture” button was pressed after inputting the main parameters, the motor shifted and rotated the camera array accordingly, and the camera array captured the perspectives. Here, the Raspberry Pi3 board controlled the camera shutter through a Python script to capture all perspectives simultaneously. An integrated controller applied a delay time to reduce the effect of vibration on the camera array during image acquisition; in our experiments, the delay time was set to 1 s. Figure 7b shows the brightness distribution difference between the originally captured and corrected perspectives. In the calculation of average brightness, only the background was captured, and we could verify that the brightness was increased from the bottom to the top camera. However, after the histogram matching process, the brightness distribution of all images became identical.
In the experiment, the camera array was shifted horizontally by 597.8 mm and the horizontal viewing angle of the acquired light field image was 20.544°. To match the vertical and horizontal viewing angles, the distance between the neighboring cameras was set to approximately 180 mm such that the vertical viewing angle of the acquired light field image was set as approximately 20.4° with the toe-in capturing method. The distance between the camera and the object was approximately 1000 mm and the number of perspectives along the horizontal axis was Np = 11, where the interval between neighboring perspectives was 59.78 mm and 2.0544°. Note that Np is not limited in the proposed system, given that a total of 3 × 60,000 perspectives can be captured using the translation stage with a 600 mm length. Figure 8 shows an example of 3 × 11 perspectives for a real object, “Belle”, with a height of approximately 1 m. Cam1, Cam2, and Cam3 indicate the camera numbers from bottom to top; 1, 6, and 11 are the horizontal indices.
The resolution of each perspective was 6000 × 4000 pixels, i.e., the same as the camera resolution, where the total imageset (3 × 11 perspectives) size was approximately 136.6 MB in a computer that did not require additional data storage. Therefore, the simplified light field image acquisition system can acquire light field images with a much improved resolution and a larger viewing angle compared to conventional lens array-based commercial light field cameras or depth cameras. As shown in Table 1, only three cameras were utilized to obtain 3 × 11 perspectives, giving the same task to 33 cameras of the camera array; concerning the advantages of the system, it can be observed that the cost performance was good.

3.2. Instant-NGP Model Training and Generation for View Synthesis

Figure 9 shows the user interface of the training process for an instant-NGP model and generation for view synthesis. Both procedures were performed on the Python script. In the training, the iteration of the deep neural network was set as 2122 times within the 30 s and the detected loss of training was 0.00195. In the view synthesis generation, by using the custom-trained instant-NGP model, 11 × 11 camera poses were determined within approximately 4.04 s (30 fps) from the input 3 × 11 perspectives and 11 × 11 perspectives were rendered with a resolution of 4000 × 4000 pixels within approximately 20 min. Since each element lens of the 3D display unit was shaped as a square, the resolution of each rendered perspective was set to 4000 × 4000 pixels, where the overall size in the computer was approximately 339 MB. Unlike previous similar methods [12], the view synthesis generation process becomes simpler by filling in the rest of the viewpoints in consideration of the parameters used to generate the 3D model from the input data, rather than regenerating the 3D model. Note that the output of a custom-trained instant-NGP model is not limited. It can render more images, for example, 50 × 50 or 500 × 500 and more; it only depends on the number of pixels in each elemental image.
Figure 10 shows an example of the newly rendered 11 × 11 perspectives. The number after lines presents the indices of perspectives in each row. Note that the conventional NeRF model requires approximately 1–2 days to output fine-quality images [26]; therefore, a custom-trained instant-NGP model has the significant advantage of being able to produce sufficiently high-quality perspectives in a very short time by performing multiresolution hash encoding.
For the quality and accuracy measurements of the rendered perspectives, we compared the input and output perspectives with the same viewpoint index using peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) methods. Here, PSNR and SSIM values were 27.33 dB and 0.88, respectively, where the corresponding input central perspective (Cam2-6 of Figure 8) was utilized as a reference, as shown in Figure 11. Therefore, it could be observed that the loss of output images of the custom-trained instant-NGP model was significantly less, and both images were almost identical.

3.3. EIA Generation and 3D Display

From the rendered set of different viewpoints, the EIA was generated for the display unit through the pixel re-arrangement process. Here, the field-of-view was given at 20.544° (horizontal)/20.4° (vertical), which was identical to the acquired fields-of-view. An integral imaging display consisted of a 15″ display device with 4K resolution (3840 × 2160 px) and a lens array with 345 × 194 lenses (1 mm pitch and 3.3 focal lengths), as shown in Figure 12a. The EIA with 194 × 194 elemental images was generated where the overall resolution was 2160 × 2160 px and each elemental image consisted of 11 × 11 px. In the experiment, two EIAs were generated and reconstructed for the entire object and for the upper part of the object for more detailed observation, as shown in Figure 12b.
Figure 13 shows optically reconstructed 3D images on the integral imaging display from two EIAs. Here, we verified that the proposed system successfully encoded 3D information about an object, and the virtual 3D images were reconstructed clearly, approximately 20 mm backward from the display device, with perfect depth cues. The viewing angle of the 3D display was approximately 13° according to the lens array’s specifications; however, approximately 20° of fields-of-views of the object were visualized within the 3D display’s viewing angle. Note that the proposed system was not designed only for human faces, but it was confirmed that facial details could be expressed well in the proposed system, like with existing 3D face recovery methods [13].
Although the quality of the reconstructed 3D images was good, we evaluated the quality through quantitative evaluation methods in order to try to prove the quality clearly. Here, the non-reference-type naturalness image quality evaluator (NIQE) [32] and perception-based image quality evaluator (PIQE) [33] methods were used to measure the quality of the displayed images. Note that the reconstructed 3D images were captured manually. Therefore, the PSNR-like reference-type evaluation methods could not be utilized by using the acquired or rendered perspectives as a reference; it was appropriate to separately evaluate the corresponding image itself. Table 2 shows the evaluated NIQE and PIQE scores for the central viewpoints of the reconstructed 3D images that were presented in Figure 13a,b. For the PIQE, the images were evaluated through the default model; for the NIQE, the scores were measured through the custom models that were trained for each image in order evaluate more clearly. These results prove that even though the 3D image was reconstructed through a lens array, various optical issues such as lens aberrations could occur, but the proposed system displayed 3D images of significantly high quality. Accordingly, it confirmed that the proposed system offers a simple and effective way to acquire high-quality and high-resolution light field images of real objects and display high-quality 3D visualizations within the ordinary environment. Note that the values of both methods are inversely correlated to the perceptual quality of images; therefore, a low value indicates better quality [32,33].

4. Conclusions

A high-quality light field 3D display system based on a simplified light field image acquisition method and a custom-trained deep neural network was proposed and implemented. The existing research related to commercial light field cameras, such as 3D model reconstruction or 3D face recovery methods, has been presented successfully, but the purpose of the proposed system is slightly different from these methods. The ultimate goal of the proposed system is to design/implement a high-quality 3D display system, including a high-quality and high-resolution 3D data acquisition method from real-world objects within an ordinary environment and an integral imaging 3D display unit based on a high-resolution display device and large-sized lens array. The proposed system consists of main three stages, and we designed and implemented each stage and combination. The simplified light field image acquisition method acquired high-resolution and high-quality 3D information of a real object (3 × 11 images) through only three high-resolution cameras with a dual-function smart motor that could originally be acquired through a camera array consisting of numerous (33) cameras. In order to match the viewpoints of the real object and the configuration of the 3D display unit, a full-connected instant-NGP deep neural network with a multiresolution hash encoding model was utilized that could render the view synthesis of the object with higher quality within a much shorter processing time compared with the other methods such as NeRF and COLMAP networks. In the experiment, an instant-NGP model was trained to extract 11 × 11 perspectives from acquired images with the same quality. Finally, EIAs were generated from the rendered perspectives through a pixel re-arrangement process and displayed/reconstructed as high-quality 3D images. The experimental results confirm that the proposed system provides a simple and effective way to acquire high-resolution and high-quality light field images of real objects and provide high-quality natural-view 3D images with accurate 3D information. In particular, it was certified that high-quality light field image acquisition/display and image processing can be performed at a relatively high speed in an ordinary environment, without any special equipment and at a low cost. Future research should focus on the fast computation of the proposed system.

Supplementary Materials

The following supporting information can be downloaded at https://www.mdpi.com/article/10.3390/s23146245/s1: Video S1: The performance of a simplified light image acquisition system; Video S2: The 3D image reconstruction for entire object; Video S3: The 3D image reconstruction for the upper part of the object.

Author Contributions

Conceptualization, M.-U.E. and K.-C.K.; methodology, M.-U.E.; software, M.-U.E., T.A. and A.K.; validation, M.-U.E., T.A., A.K., K.-C.K. and N.K.; formal analysis, M.-U.E., T.A., A.K. and K.-C.K.; investigation, M.-U.E.; resources, M.-U.E.; data curation, M.-U.E., T.A. and A.K.; writing—original draft preparation, M.-U.E.; writing—review and editing, M.-U.E., O.-S.N. and N.K.; visualization, M.-U.E.; supervision, S.-H.J., K.-C.K. and N.K.; project administration, K.-C.K. and N.K.; funding acquisition, K.-C.K. and N.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the MSIT (Ministry of Science and ICT), Korea, under the ITRC (Information Technology Research Center) support program (IITP-2023-2020-0-01846) supervised by the IITP (Institute for Information & Communications Technology Planning & Evaluation) and supported by the National Research Foundation of Korea (NRF) grant funded by the Korea Government (nos. NRF-2018R1D1A3B07044041 and NRF-2020R1A2C1101258).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to express their deep gratitude to KAPS Co., Ltd. for producing and supporting the large-sized lens array for the 3D display unit.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Martínez-Corral, M.; Javidi, B. Fundamentals of 3D imaging and displays: A tutorial on integral imaging, light-field, and plenoptic systems. Adv. Opt. Photon. 2020, 10, 512–566. [Google Scholar]
  2. Javidi, B.; Carnicer, A.; Arai, J.; Fujii, T.; Hua, H.; Liao, H.; Martínez-Corral, M.; Pla, F.; Stern, A.; Waller, L.; et al. Roadmap on 3D integral imaging: Sensing, processing, and display. Opt. Express 2020, 28, 32266–32293. [Google Scholar]
  3. Hong, S.; Dorado, A.; Saavedra, G.; Barreiro, J.C.; Martinez-Corral, M. Three-dimensional integral-imaging display from calibrated and depth-hole filtered Kinect information. J. Disp. Technol. 2016, 12, 1301–1308. [Google Scholar]
  4. Xiong, Z.; Zhang, Y.; Wu, F.; Zeng, W. Computational depth sensing: Toward high-performance commodity depth cameras. IEEE Signal Process. Mag. 2017, 34, 55–68. [Google Scholar]
  5. Erdenebat, M.-U.; Kim, B.-J.; Piao, Y.-L.; Park, S.-Y.; Kwon, K.-C.; Piao, M.-L.; Yoo, K.-H.; Kim, N. Three-dimensional image acquisition and reconstruction system on a mobile device based on computer-generated integral imaging. Appl. Opt. 2017, 56, 7796–7802. [Google Scholar]
  6. Erdenebat, M.-U.; Piao, Y.-L.; Kwon, K.-C.; Lee, M.H.; Kwon, K.H.; Kim, M.Y.; Kim, N. Advanced visualization using image super-resolution method for three-dimensional mobile system. Opt. Commun. 2021, 480, 126494. [Google Scholar]
  7. Ng, R.; Levoy, M.; Bredif, M.; Duval, G.; Horowitz, M.; Hanrahan, P. Light field photography with a hand-held plenoptic camera. Stanf. Univ. Comp. Sci. Tech. Rep. 2005, 2005, 02. [Google Scholar]
  8. Lee, S.-K.; Hong, S.-I.; Kim, Y.-S.; Lim, H.-G.; Jo, N.-Y.; Park, J.-H. Hologram synthesis of three-dimensional real objects using portable integral imaging camera. Opt. Express 2013, 21, 23662–23670. [Google Scholar]
  9. Jeon, H.-G.; Park, J.; Choe, G.; Park, J.; Bok, Y.; Tai, Y.-W.; Kweon, I.S. Accurate depth map estimation from a lenslet light field camera. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  10. Kwon, H.; Kizu, Y.; Kizaki, Y.; Ito, M.; Kobayashi, M.; Ueno, R.; Suzuki, K.; Funaki, H. A gradient index liquid crystal microlens array for light-field camera applications. IEEE Photon. Technol. Lett. 2015, 27, 836–839. [Google Scholar]
  11. Jeong, Y.; Kim, J.; Yeom, J.; Lee, C.-K.; Lee, B. Real-time depth controllable integral imaging pickup and reconstruction method with a light field camera. Appl. Opt. 2015, 54, 10333–10341. [Google Scholar]
  12. Perra, C.; Murgia, F.; Giusto, D. An analysis of 3D point cloud reconstruction from light field images. In Proceedings of the 2016 Sixth International Conference on Image Processing Theory, Tools and Applications (IPTA), Oulu, Finland, 12–15 December 2016. [Google Scholar]
  13. Feng, M.; Gilani, S.Z.; Wang, Y.; Mian, A. 3D face reconstruction from light field images: A model-free approach. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  14. Zhu, S.; Lai, A.; Eaton, K.; Jin, P.; Gao, L. On the fundamental comparison between unfocused and focused light field cameras. Appl. Opt. 2018, 57, A1–A11. [Google Scholar]
  15. Kwon, K.-C.; Kwon, K.H.; Erdenebat, M.-U.; Piao, Y.-L.; Lim, Y.-T.; Kim, M.Y.; Kim, N. Resolution-enhancement for an integral imaging microscopy using deep learning. IEEE Photon. J. 2019, 11, 6900512. [Google Scholar]
  16. Kwon, K.H.; Erdenebat, M.-U.; Kim, N.; Khuderchuluun, A.; Imtiaz, S.M.; Kim, M.Y.; Kwon, K.-C. High-quality 3D visualization system for light-field microscopy with fine-scale shape measurement through accurate 3D surface data. Sensors 2023, 23, 2173. [Google Scholar]
  17. Taguchi, Y.; Koike, T.; Takahashi, K.; Naemure, T. TransCAIP: A live 3D TV system using a camera array and an integral photography display with interactive control of viewing parameters. IEEE Trans. Vis. Comput. Graph. 2009, 15, 841–852. [Google Scholar]
  18. Lin, X.; Wu, J.; Zheng, G.; Dai, Q. Camera array based light field microscopy. Biomed. Opt. Express 2015, 15, 3179–3189. [Google Scholar]
  19. Chen, G.; Wang, H.; Liu, M.; Liao, H. Hybrid camera array based calibration for computer-generated integral photography display. J. Opt. Soc. Am. 2015, 35, 1567–1574. [Google Scholar]
  20. Xu, Y.; Maeno, K.; Nagahara, H.; Taniguchi, R.-i. Camera array calibration for light field acquisition. Front. Comput. Sci. 2015, 9, 691–702. [Google Scholar]
  21. Xing, Y.; Xiong, Z.-L.; Zhao, M.; Wang, Q.-H. Real-time integral imaging pickup system using camera array. In Proceedings of the SPIE Photonics West 2018, San Francisco, CA, USA, 19 February 2018. [Google Scholar]
  22. Huang, Y.; Yan, Z.; Jiang, X.; Jing, T.; Chen, S.; Lin, M.; Zhang, J.; Yan, X. Performance enhanced elemental array generation for integral image display using pixel fusion. Front. Phys. 2021, 9, 639117. [Google Scholar]
  23. Shin, C.; Jeon, H.-G.; Yoon, Y.; Kweon, I.S.; Kim, S.J. EPINET: A fully-convolutional neural network using epipolar geometry for depth from light field images. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  24. Kwon, K.-C.; Kwon, K.H.; Erdenebat, M.-U.; Piao, Y.-L.; Lim, Y.-T.; Zhao, Y.; Kim, M.Y.; Kim, N. Advanced three-dimensional visualization system for an integral imaging microscope using a fully convolutional depth estimation network. IEEE Photon. J. 2020, 12, 3900714. [Google Scholar]
  25. Schonberger, J.L.; Frahm, J.-M. Structure-from-motion revisited. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  26. Mildenhall, B.; Srinivasan, P.P.; Tancik, M.; Barron, J.T.; Ramamoorthi, R.; Ng, R. NeRF: Representing scenes as neural radiance fields for view synthesis. Commun. ACM 2022, 65, 99–106. [Google Scholar]
  27. Hu, T.; Liu, S.; Chen, Y.; Shen, T.; Jia, J. EfficientNeRF efficient neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  28. Müller, T.; Evans, A.; Schied, C.; Keller, A. Instant neural graphics primitives with a multiresolution hash encoding. ACM Trans. Graph. 2022, 41, 102. [Google Scholar]
  29. Park, J.; Stoykova, E.; Kang, H.; Hong, S.; Lee, S.; Jung, K. Numerical reconstruction of full parallax holographic stereograms. 3D Res. 2012, 3, 6. [Google Scholar]
  30. Khuderchuluun, A.; Erdenebat, M.-U.; Dashdavaa, E.; Lee, M.H.; Jeon, S.-H.; Kim, N. Simplified digital content generation based on an inverse-directed propagation algorithm for holographic stereogram printing. Appl. Opt. 2021, 60, 4235–4244. [Google Scholar]
  31. Vieira, A.; Duarte, H.; Perra, C.; Tavora, L.; Assuncao, P. Data formats for high efficiency coding of Lytro-Illum light fields. In Proceedings of the 2015 International Conference on Image Processing Theory, Tools and Applications (IPTA), Orleans, France, 10–13 November 2015. [Google Scholar]
  32. Mittal, A.; Soundararajan, R.; Bovik, A.C. Making a completely blind image quality analyzer. IEEE Sig. Process. Lett. 2013, 60, 209–212. [Google Scholar]
  33. Venkatanath, N.; Praneeth, D.; Chandrasekhar, B.M.; Channappayya, S.S.; Medasani, S.S. Blind image quality evaluation using perception based features. In Proceedings of the IEEE 21st National Conference on Communications, Montréal, QC, Canada, 14–23 June 2021. [Google Scholar]
Figure 1. Overall schematic configuration of the proposed high-quality 3D integral imaging display system based on the simplified light field image acquisition method.
Figure 1. Overall schematic configuration of the proposed high-quality 3D integral imaging display system based on the simplified light field image acquisition method.
Sensors 23 06245 g001
Figure 2. (a) The structure of the simplified light field image acquisition system and (b) a flowchart of the main procedure of the integrated controller.
Figure 2. (a) The structure of the simplified light field image acquisition system and (b) a flowchart of the main procedure of the integrated controller.
Sensors 23 06245 g002
Figure 3. Histogram matching for the perspectives captured by different cameras using a calculated average histogram.
Figure 3. Histogram matching for the perspectives captured by different cameras using a calculated average histogram.
Sensors 23 06245 g003
Figure 4. Basic configuration for the rendering of Np × Np perspectives from input Nc × Np-corrected perspectives through the custom-trained instant-NGP model.
Figure 4. Basic configuration for the rendering of Np × Np perspectives from input Nc × Np-corrected perspectives through the custom-trained instant-NGP model.
Sensors 23 06245 g004
Figure 5. The training scheme for the instant-NGP model to extract Np × Np perspectives.
Figure 5. The training scheme for the instant-NGP model to extract Np × Np perspectives.
Sensors 23 06245 g005
Figure 6. Prototype of the simplified light field image acquisition system of the proposed system (Video S1: The performance of a simplified light image acquisition system).
Figure 6. Prototype of the simplified light field image acquisition system of the proposed system (Video S1: The performance of a simplified light image acquisition system).
Sensors 23 06245 g006
Figure 7. (a) User interface of the computer-based integrated controller running on LabVIEW software and (b) the brightness distribution for each camera before (top) and after (bottom) the histogram matching process.
Figure 7. (a) User interface of the computer-based integrated controller running on LabVIEW software and (b) the brightness distribution for each camera before (top) and after (bottom) the histogram matching process.
Sensors 23 06245 g007
Figure 8. The 1st, 6th, and 11th captured perspectives of a real object, “Belle”, for each camera after brightness distribution correction.
Figure 8. The 1st, 6th, and 11th captured perspectives of a real object, “Belle”, for each camera after brightness distribution correction.
Sensors 23 06245 g008
Figure 9. The user interface of (a) the training for an instant-NGP model and (b) the rendering of 11 × 11 perspectives according to the calculation of camera poses.
Figure 9. The user interface of (a) the training for an instant-NGP model and (b) the rendering of 11 × 11 perspectives according to the calculation of camera poses.
Sensors 23 06245 g009
Figure 10. The 1st (left-most), 6th (central), and 11th (right-most) perspectives of the 1st (top), 6th (central), and 11th (bottom) rows of 11 × 11 perspectives.
Figure 10. The 1st (left-most), 6th (central), and 11th (right-most) perspectives of the 1st (top), 6th (central), and 11th (bottom) rows of 11 × 11 perspectives.
Sensors 23 06245 g010
Figure 11. The comparison between the central (6th) perspectives of (a) captured/corrected and (b) newly rendered 11 × 11 images.
Figure 11. The comparison between the central (6th) perspectives of (a) captured/corrected and (b) newly rendered 11 × 11 images.
Sensors 23 06245 g011
Figure 12. (a) A prototype of the integral imaging 3D display and (b) the EIAs for 3D display generated from the entire object (left) and upper part (right).
Figure 12. (a) A prototype of the integral imaging 3D display and (b) the EIAs for 3D display generated from the entire object (left) and upper part (right).
Sensors 23 06245 g012
Figure 13. Reconstructed 3D image of the object from various viewpoints: center (0°), left (−6.5°), right (6.5°), top (6.5°), and bottom (−6.5°) views (a) for the entire object (Video S2: The 3D image reconstruction for entire object) and (b) for the upper part of the object (Video S3: The 3D image reconstruction for the upper part of the object) in order to present more detailed observation.
Figure 13. Reconstructed 3D image of the object from various viewpoints: center (0°), left (−6.5°), right (6.5°), top (6.5°), and bottom (−6.5°) views (a) for the entire object (Video S2: The 3D image reconstruction for entire object) and (b) for the upper part of the object (Video S3: The 3D image reconstruction for the upper part of the object) in order to present more detailed observation.
Sensors 23 06245 g013
Table 1. Comparison for 3 × 11 perspective acquisition based on a commercial light field camera, the camera array, and the proposed system.
Table 1. Comparison for 3 × 11 perspective acquisition based on a commercial light field camera, the camera array, and the proposed system.
CharacteristicsCommercial Light Field Camera [31]Camera Array
(3 × 11 Cameras)
Proposed System
(3 Cameras)
Resolution (single perspective)625 × 434 px6000 × 4000 px6000 × 4000 px
Data quality (for proposed system)LowHighHigh
Data quantityCommonCommonCommon
Data storageNACan be requiredNA
Viewing angleNarrow
(fixed)
Wide
(can be controllable)
Wide
(can be controllable)
Depth-of-fieldCommon
(fixed)
Wide
(can be controllable)
Wide
(can be controllable)
Table 2. Measured NIQE and PIQE scores for the central viewpoints of reconstructed 3D images.
Table 2. Measured NIQE and PIQE scores for the central viewpoints of reconstructed 3D images.
Evaluation MethodCentral Viewpoint (0°) of
3D Image of Figure 13a
Central Viewpoint (0°) of
3D Image of Figure 13b
NIQE3.313.19
PIQE20.0618.57
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Erdenebat, M.-U.; Amgalan, T.; Khuderchuluun, A.; Nam, O.-S.; Jeon, S.-H.; Kwon, K.-C.; Kim, N. Comprehensive High-Quality Three-Dimensional Display System Based on a Simplified Light-Field Image Acquisition Method and a Full-Connected Deep Neural Network. Sensors 2023, 23, 6245. https://doi.org/10.3390/s23146245

AMA Style

Erdenebat M-U, Amgalan T, Khuderchuluun A, Nam O-S, Jeon S-H, Kwon K-C, Kim N. Comprehensive High-Quality Three-Dimensional Display System Based on a Simplified Light-Field Image Acquisition Method and a Full-Connected Deep Neural Network. Sensors. 2023; 23(14):6245. https://doi.org/10.3390/s23146245

Chicago/Turabian Style

Erdenebat, Munkh-Uchral, Tuvshinjargal Amgalan, Anar Khuderchuluun, Oh-Seung Nam, Seok-Hee Jeon, Ki-Chul Kwon, and Nam Kim. 2023. "Comprehensive High-Quality Three-Dimensional Display System Based on a Simplified Light-Field Image Acquisition Method and a Full-Connected Deep Neural Network" Sensors 23, no. 14: 6245. https://doi.org/10.3390/s23146245

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop