Next Article in Journal
An Angle Precision Evaluation Method of Rotary Laser Scanning Measurement Systems with a High-Precision Turntable
Previous Article in Journal
Look-Up-Table-Based Direct-Detection-Faster-Than-Nyquist-Algorithm-Enabled IM/DD Transmission with Severe Bandwidth Limitation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Reconstructing Depth Images for Time-of-Flight Cameras Based on Second-Order Correlation Functions

1
National Space Science Center, Chinese Academy of Sciences, Beijing 100190, China
2
Northeastern University, Shenyang 110819, China
3
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Photonics 2023, 10(11), 1223; https://doi.org/10.3390/photonics10111223
Submission received: 27 September 2023 / Revised: 21 October 2023 / Accepted: 30 October 2023 / Published: 31 October 2023

Abstract

:
Depth cameras are closely related to our daily lives and have been widely used in fields such as machine vision, autonomous driving, and virtual reality. Despite their diverse applications, depth cameras still encounter challenges like multi-path interference and mixed pixels. Compared to traditional sensors, depth cameras have lower resolution and a lower signal-to-noise ratio. Moreover, when used in environments with scattering media, object information scatters multiple times, making it difficult for time-of-flight (ToF) cameras to obtain effective object data. To tackle these issues, we propose a solution that combines ToF cameras with second-order correlation transform theory. In this article, we explore the utilization of ToF camera depth information within a computational correlated imaging system under ambient light conditions. We integrate compressed sensing and non-training neural networks with ToF technology to reconstruct depth images from a series of measurements at a low sampling rate. The research indicates that by leveraging the depth data collected by the camera, we can recover negative depth images. We analyzed and addressed the reasons behind the generation of negative depth images. Additionally, under undersampling conditions, the use of reconstruction algorithms results in a higher peak signal-to-noise ratio compared to images obtained from the original camera. The results demonstrate that the introduced second-order correlation transformation can effectively reduce noise originating from the ToF camera itself and direct ambient light, thereby enabling the use of ToF cameras in complex environments such as scattering media.

1. Introduction

With the development of technology, 3D imaging has been applied in many fields, and depth images have become one of the most important and innovative areas in the field of image sensing science and engineering in recent decades [1]. In applications such as object recognition and remote sensing, depth maps are important tools for capturing the spatial position and motion of detected objects. In the past few decades, 3D imaging technology has mainly been divided into three categories, structured light 3D imaging [2], binocular vision 3D imaging [3,4,5,6] and time-of-flight (ToF) 3D imaging [7,8,9]. The basic principle of structured light 3D imaging is optical triangulation, which reconstructs depth maps by geometric distortions caused by the shape of the object’s surface. Binocular vision 3D imaging uses two eyes (or two images) to capture the shape and position of objects in the scene, calculating 3D information including depth and size by detecting feature points of the objects and the disparity between the two images. Compared to structured light 3D imaging and binocular vision 3D imaging, ToF technology is a 3D imaging technology that uses the time difference of light beam reflection on the object’s surface to measure distance. ToF imaging technology is largely dominated by ToF cameras, as they have small sizes, good resolution, robustness in ambient light conditions, low power consumption, and fast processing capabilities. ToF cameras have advantages in distance measurement compared to traditional cameras because they can measure without being limited by lighting conditions. ToF cameras typically operate within a range of tens of meters to capture high-resolution 3D images at video rates. ToF cameras use modulated infrared light sources to illuminate the scene and resolve the phase shift distance measured by the sensor. However, the quality of depth maps obtained by ToF cameras is easily affected by ambient light and limitations of the camera itself, and phase shift reconstruction is sensitive to noise. In the past few decades, pattern modulation and bucket detection techniques have been proven effective in 3D imaging. Howland et al. achieved single-pixel 3D imaging of the region of interest using distance gating technology and obtained a complete depth map through full-range scanning [10]. Kirmani et al. used a time-correlated single-photon counting system instead of distance gating to obtain depth maps [11]. Sun Mingjie et al. simplified Kirmani’s structure and determined the depth map by sampling the time-varying intensity measured by a high-speed photodiode [9]. Sun Baoqing et al. used a spatially separated single-pixel detector to measure different shadow images and reconstructed 3D images through multi-view stereo vision [12]. Computational ToF imaging has been proven to be a new method for solving many problems in ToF imaging and promoting new applications, such as transient imaging [13,14,15], non-line-of-sight imaging [16,17,18], light transport analysis [19], lensless imaging [20], distance and velocity synchronized imaging, etc. [21]. These methods are mostly based on the waveform design ideas of time-of-flight sensors and temporary modulation codes of light sources. However, though great progress has been made for ToF cameras, the low pixel resolution and signal-to-noise ratio (SNR) in these applications are still bottlenecks that have not been broken through so far [22]. More than a decade ago, the theoretical concept of Computational Ghost Imaging (CGI) was introduced [23,24,25]. It involves the capture of single-pixel bucket-detector measurements on a detection plane, with subsequent data processing to retrieve the intensity distribution of a reference beam. The fluctuation correlation between the signals acquired by the bucket detectors and the reference signal is utilized to reconstruct the image of the target object. This can be achieved even in challenging environments, such as atmospheric turbulence [26] or scattering media [27], where traditional imaging methods may prove ineffective. Subsequently, the success of Compressed Sensing (CS) algorithms [28,29,30,31] in CGI has been observed. CS algorithms are well-suited for the reconstruction of sparse targets. Given the sparsity and noise present in the depth images acquired by ToF sensors, we have opted to employ CS algorithms for processing these images, as they offer robust denoising capabilities and high resolution at low sampling rates. While CS has improved the performance of CGI through its image reconstruction algorithms, its application is constrained by strong sparsity assumptions and limitations in the reconstruction process [32,33,34]. The emergence of Deep Learning (DL) [35,36,37,38,39] presents an opportunity to relax sparsity constraints by recovering images at ultra-low sampling rates (SR) using trained data and untrained strategies [40,41,42,43]. Concurrently, ToF-based novel computational 3D imaging techniques show great promise across various imaging domains [44,45,46,47].
In this paper, a time-modulated randomized spatial illumination pattern and a ToF sensor are used to acquire depth maps. The correlation image sensor does not obtain the phase of the measurement pattern directly, but by obtaining the correlation between the received and reference signals. In order to reduce the noise, a second-order correlation transform is introduced to reconstruct the depth map by combining the ToF principle with CGI, CS, and untrained neural networks. It is found that comparing the depth maps from the original ToF camera, the scheme based on compressed sensing and untrained neural network can obtain a higher peak signal-to-noise ratio (PSNR). Specifically, the TVAL3 algorithm and untrained neural networks yield superior reconstruction images characterized by enhanced image quality, heightened SNR, and increased contrast at a 12.5% sampling rate. In addition, based on the ToF camera, we also conducted experiments with the ToF camera through scattering media, in which the CGI-based CS and DL methods can clearly recover the images, while the ordinary ToF method cannot. We believe that the ToF camera can be used in similar applications where it can be applied to quite complex environments such as scattering media or underwater.

2. Basic Theories and Principles

2.1. Ranging Principle of ToF Cameras

We start from the continuous-wave (CW) modulation of the illuminating source to introduce the ranging principle of the ToF camera as an example. Generally, a ToF camera consists of two parts, i.e., the illuminating source and the detection system. In use, a scene is illuminated by the modulated light source and the reflected beam that carry the information of the scene observed by the sensor in the detection system, where the phase shift between the light source and the reflected light can be measured and then directly converted into the distance between the scene and the optical source according to the ranging formula Equation (1).
The ranging principle of the CW modulated ToF is shown in Figure 1. For simplicity, we take a square wave modulation of the illuminating optical source as an example to illustrate the basic theory of the ToF. The phase delay φ between the modulated emitting and the reflected square wave signals is assumed to be as just shown on the top of Figure 1, which is measured by using the well-known four-step phase shift method [48,49,50]. Whereas, if φ is measured, the distance from the target object to the ToF camera can be calculated according to the following formula [51,52].
d = c 2 φ 2 π f ,
where c is the speed of light in a vacuum and f is the signal frequency. It is shown in Figure 1 that the four-phase control signals C 1 C 4 have π 2 phase delays from each other, which determine the electric charge values Q 1 Q 4 that are the amount of electric charge for the control signals. Generally, the φ can be estimated by using Q 1 Q 4 like
φ = tan 1 Q 4 Q 3 Q 1 Q 2 .
However, Q 1 Q 4 are not able to be directly measured, and the modulated signal is not always a square wave. For a generic output signal c ( t ) of a ToF detector, it can be expressed by the cross-correlation of the emitted signal g ( t ) and the received signal s ( t ) as
c ( τ ) = s ( t ) g ( t ) = lim T 1 T T 2 T 2 s ( t ) g ( t + τ ) d t ,
where
s ( t ) = 1 + a cos ( ω t φ ) ,
and
g ( t ) = cos ( ω t ) ,
Here, a is the attenuated amplitude of the reflected light signal acquired by the sensor and ∗ indicates convolution.
After the calculation, we can know
c ( τ ) = lim T 1 T T 2 T 2 [ 1 + a cos ( ω t φ ) ] · [ cos ( ω t + ω τ ) ] d t , = a 2 · cos ( φ + ω τ ) ,
where c ( τ ) represents the ToF sensor’s output during period T. It actually amounts to the total number of photons received in a period by the certain pixel of the ToF sensor, which is in direct proportion to the electric charge values accumulated on this pixel that are mentioned in Equation (2). Therefore, c ( τ 1 ) c ( τ 4 ) can usually be substituted for Q 1 Q 4 in Equation (2), where τ 1 τ 4 are selected as 0, π 2 , π and 3 π 2 , respectively. The phase shift φ , i.e., Equation (2), is rewritten accordingly as
φ = tan 1 c ( τ 4 ) c ( τ 3 ) c ( τ 1 ) c ( τ 2 ) ,
while the amplitude a is also computed by the following equation:
a = 1 2 [ c ( τ 4 ) c ( τ 3 ) ] 2 + [ c ( τ 1 ) c ( τ 2 ) ] 2 .
Eventually, it is obvious that the distance d can be estimated pixel by pixel via substituting φ in Equation (7) into the Equation (1).

2.2. CGI via a ToF Camera

A classical passive CGI scheme is shown in Figure 2, where a target object illuminated by a light beam is imaged on a DMD and then one of the reflected beams from the DMD is captured by a single pixel bucket detector through a collecting lens. As an indirect imaging method, the image can be retrieved by the correlation computation between the modulated matrix and the actually captured single-pixel optical signals that carry the information of the target object, which can be expressed by the normalized intensity correlation as
g ( 2 ) ( x , y ) = ( 1 / N ) i = 1 N S i ϕ i ( x , y ) ( 1 / N ) i = 1 N S i ( 1 / N ) i = 1 N ϕ i ( x , y ) ,
Here, S i and ϕ i ( x , y ) are the ith ( i = 1 , 2 , , N ) single-pixel signals and the ith modulation speckles, respectively. x and y ( x , y = 1 , 2 , , M ) are the row and column pixel coordinates of each modulated basis ( M × M ) of the DMD.
It is well-known that a ToF sensor can export both the intensity and depth data that carry the information of the target object simultaneously, where both of them come from the counting number of photons received by the ToF sensor in a certain period. Differently, the intensity maps can be acquired directly as a universal detector does while the depth maps are exported by counting the number of photons of different phase signals through a complex mathematical operations mentioned in Equation (1). However, they are essentially determined by the statistics of photon numbers. Therefore, both of them can be available in a CGI system when the ToF camera is used as a bucket detector. That is to say, if the bucket signal S i is replaced by D i , the Equation (9) still works. Therefore, it would be expressed by
g d ( 2 ) ( x , y ) = ( 1 / N ) i = 1 N D i ϕ i ( x , y ) ( 1 / N ) i = 1 N D i ( 1 / N ) i = 1 N ϕ i ( x , y ) ,
where D i = x , y = 1 P 1 , P 2 d i ( x , y ) , which is the ith ( i = 1 , 2 , , N ) single-pixel signal of the ToF depth maps d i ( x , y ) . The size of a ToF sensor is assumed to be P 1 × P 2 pixels and its row and column pixel coordinates are x ( x = 1 , 2 , , P 1 ) and y ( y = 1 , 2 , , P 2 ), respectively.
Although both the intensity and depth maps can be directly extracted from the raw data exported by the ToF camera, they have different map formats due to their different data acquisition modes. For the intensity maps, the 8-bit gray scale format ranging from 0 to 255 is one of the most-used picture formats. The value in each pixel of each map is positively correlated with the photon number received in this pixel of the ToF sensor. As for the depth maps, the value in each pixel is the distance between a certain point on the surface of the target object and the ToF sensor, which ranges from 1 to 5 m with a step of 0.25 m for the ToF camera we used in the experiments. The values that are less than 1 or larger than 5 would be automatically set to be 1 or 5 by the camera. In addition, the values of those pixels without receiving any photon are also set to be 5. Therefore, the depth maps are generally so-called negative images, where the background has bigger pixel values than those of the image of the target in a depth map. It will be seen in the following that the images reconstructed by the depth maps are also negative ones regardless of the CGI and CS algorithms. For convenience, we call a reconstructed image by utilizing the intensity or depth maps of a ToF sensor an intensity or depth image.

2.3. Image Reconstruction Algorithm Based on the U-Net Framework

With the development of DL theory, it has been widely used to solve various inverse problems in computational imaging. However, traditional data-driven DL methods require a large amount of training data to optimize network parameters, and the training time is as long as several hours or even days, which hinders its practical application. In recent years, untrained DL methods that combine DIP theory with physical imaging processes have attracted a lot of attention in computational imaging. The DIP theory can be combined with the physical process of single pixel imaging to form a network framework that can be iterated continuously [53,54,55,56].
In Figure 3, the U-net framework requires an image of the same size as the modulated speckle as the neural network input, and the U-net network will output the estimated image based on the current weight proportion in the network. In the physical process of traditional single pixel imaging, after ignoring the influence of background noise, the obtained bucket signal can be simply described by the following formula:
y = H O T ,
where y is the light intensity value of the barrel signal obtained by the single pixel detector, H is the 2D matrix of the modulation matrix speckles, and O T can be regarded as the 2D reflected light (transmitted light) intensity of the object.
In this neural network framework, the corresponding predicted bucket signals can be obtained based on the predicted images in the network. However, since the output images in the early iterations are of low quality, there are apparent differences between the predicted bucket signals and the real bucket signals. These differences can be regarded as a minimization problem. We take the mean square error (MSE) between the predicted bucket signal and the real bucket signal as the loss function in the neural network. The weights in the neural network are continuously updated by several iterations to minimize the loss function so that the network can output the predicted picture that approximates the real object image. Therefore, in the framework of this neural network, the reconstruction formula of the target object can be represented by the following function:
θ * = arg min θ H θ O i n p u t y t 2 + ξ T θ O i n p u t .
Therefore, this neural network framework is different from the traditional neural network methods [57,58] for it does not require training data sets in the process of reconstructing images, and it only needs the bucket signals obtained by the ToF camera and the estimated bucket signal output by the network.

3. Experimental Scheme and Analysis of Results

To verify the feasibility of the proposed scheme, a passive CGI experiment with the aid of a ToF camera was performed based on the experimental setup in Figure 2. This setup was similar to the traditional CGI, as depicted in Figure 4, but uses a 320 × 240 pixels ToF camera (OPT8241, Texas Instruments, Dallas, TX, USA) instead of a SPD. Additionally, a total reflection prism (TRP) was inserted to adjust optical paths. In Figure 4, a printed binary object illuminated by an optical beam with a wavelength of 850 nm from an infrared light source is imaged by an imaging lens with a focal length of 50 mm on the surface of the optical unit of a DMD (DLP LightCrafter 4500, Texas Instruments), where the TRP is inserted in the optical path to be suitable for the adjustment of the detection system. One of the reflected beams by the DMD, which carries the information of encoded patterns, was captured by the ToF camera that was synchronized with the light source of the camera at the time of delivery. Both the intensity and depth data exported by the camera are obtained by being computed inside the camera.

3.1. Recovering Depth Images Using Different Methods

This section provides the depth image recovery using our scheme of Figure 4 in indoor conditions with ambient light, where a binary object of a Chinese character of the sky is used as the reconstruction target shown in Figure 5a. For comparisons, a traditional ToF depth map is first supplied in Figure 5b, which has an obvious noisy background and lower contrast compared with that of Figure 5a. In CGI experiments, the bucket signals are acquired by summing the intensity of all the pixels of the area of interest of the ToF sensor regardless of what algorithms we refer to, where they are synchronized with the Hadamard bases that are projected on the DMD. In the framework, three well-used image reconstruction algorithms including CGI, basis pursuit (BP) and total variation augmented Lagrangian Alternating Direction Algorithm (TVAL3) are used. It is shown in Figure 5c that the images by CGI are reconstructed at the low SR of 6.25%, 12.5%, 18.75%, 25%, 31.25% and 37.5%, respectively. It is obvious that the image quality increases with the increase in the SR as expected. Therefore, it is believed empirically that the object images could also be successfully recovered by CS-based reconstruction algorithms of BP and TVAL3, which are shown in Figure 5d,e. Unfortunately, the images by BP do not have a better reconstruction quality compared with CGI because of much less SR. However, intuitively, the images reconstructed by TVAL3 appear to have better quality compared to those reconstructed by CGI at the same SR.
In accordance with the principles of ToF cameras, we can ascertain a linear relationship between phase and depth. Consequently, we conduct histogram analysis of the phase map obtained from the camera, aiming to analyze the sources of noise and the reasons for negative depth values. Through our research, we have determined that the signals collected by the detection array primarily consist of three components: photons reflected back from the target object, environmental noise, and photons resulting from multi-path reflections. Figure 6a represents the original depth image obtained using a ToF camera. To more directly illustrate the reasons for the negative aspects of the image, we use the grayscale representation in Figure 6b to depict the original depth map. We configured the camera with an initial frequency of 60 Hz, and based on the camera’s internal parameters, the maximum measurable distance is limited to 5 m. In regions where the sensor receives photons reflected from multiple paths, and conversely, in areas where the sensor does not detect any photons, the camera defaults to measuring the phase values as the maximum possible. This is exemplified in Figure 6a, where the region with a depth value Z of 4.978 is highlighted. In contrast, the region with a depth value of 2.006 corresponds to photons reflected back from objects. It is evident that the values in the regions where photons are reflected back from the object are smaller than in other areas. This also explains why the resulting image displays a negative depth image.
In order to quantitatively evaluate the image quality, the P S N R in decibel (dB) is introduced as
P S N R = 10 · log 10 M A X 2 M S E ,
where M A X = 2 k 1 , which is the maximum of the processed image determined by the image bit depth k and the MSE that is used to measure the difference between the reconstructed image O ˜ and the ground truth image O is given by
M S E = 1 m n i = 0 m 1 j = 0 n 1 ( O ˜ ( i , j ) O ( i , j ) ) 2 .
Here, i , j i = 0 , 1 , 2 , , m 1 ; j = 0 , 1 , 2 , , n 1 is the pixel coordinates of the image.
The relationship between the PSNRs of reconstructed images using different reconstruction algorithms and the SR is compared, which is shown in Figure 7. It is obvious that the PSNR by each reconstructed algorithm all increases with the increase in the SR while the image by TVAL3 has the highest PSNR at each given SR. For example, when the sampling rate is 12.5%, the PSNR values of three algorithms are 8.23 dB, 7.64 dB and 8.52 dB, respectively. Compared to the original depth map with a PSNR value of 5.70 dB, the image quality has improved by 1.44 times, 1.34 times, and 1.49 times, respectively. When the sampling rate is 25%, the PSNR values of the three algorithms are 9.44 dB, 9.76 dB and 9.82 dB, respectively. Compared to the original depth map, the image quality has improved by 1.66 times, 1.71 times, and 1.72 times, respectively. Therefore, by combining the CGI-based scheme and the ToF camera, the noise that most impacts the image quality due to ambient light and the defects in the detector could be suppressed to achieve higher-quality depth images.
In Figure 6, we have identified certain issues in the images, including multipath interference and noise. Since the effective information of the target object is within a range of less than 3 m, we have incorporated a low-pass filter into the later stages of the image processing. This filter helps by nullifying any information exceeding the 3 m threshold. Within this framework, we have opted for four commonly employed image reconstruction algorithms for the purpose of image recovery, namely CGI, BP, TVAL3, and an untrained DL network. It is shown in Figure 8b that the images are reconstructed at the low SR of 6.25%, 12.5%, 18.75%, 25%, 31.25 and 37.5%. It is obvious that the image quality increases with the increase in the SR as expected. Therefore, it is believed empirically that the object images could also be successfully recovered by CS-based reconstruction algorithms of BP and TVAL3, which are shown in Figure 8c,d. Unfortunately, the images by BP do not have a better reconstruction quality compared with CGI because of much less SR. However, intuitively, TVAL3 can reconstruct almost as good even better images as CGI at the same SR. Fortunately, the untrained DL-based method can recover the target images with much better SNR compared with the other three methods, which is depicted in Figure 8e. Note that the batch size, learning rate and SR are set to be 1, 0.9 and 0.25, respectively. The best-reconstruction image is obtained while the iteration number of DL is chosen to be 300.
The relationship between the PSNRs of reconstructed images using different reconstruction algorithms and the SR is compared, which is shown in Figure 9. It is obvious that the PSNR by each reconstructed algorithm increases with the increase in the SR while the image by DL has the highest PSNR at each given SR. The PSNR by DL is 11.42 dB at even 6.25%, which is much higher than those by BP, CGI and TVAL3. However, in comparison to the original ToF images with a PSNR of 5.70 dB, the image quality was notably enhanced when utilizing the framework introduced in this paper at a sampling rate of 6.25 for four distinct reconstruction algorithms. The improvements in image quality were as follows: 1.35 times, 1.08 times, 1.47 times, and 1.75 times, respectively. When the sampling rate reached 25%, the PSNR of CGI-reconstructed images increased to 9.44 dB, translating to a 1.66-fold enhancement in image quality over the original ToF images. BP-reconstructed images achieved a PSNR of 9.76 dB, equating to a 1.71-fold improvement in image quality compared to the original ToF images. TVAL3-reconstructed images obtained a PSNR of 9.82 dB, indicating a 1.72-fold increase in image quality relative to the original ToF images. DL-reconstructed images exhibited a remarkable PSNR of 11.39 dB, signifying a twofold improvement in image quality compared to the original ToF images. However, only the images by DL can be clearly recovered with high quality at not only the much lower SR of 6.25%, but also other higher SRs, while those by other algorithms are either almost blurred, or still have a lot of noisy background. Thus, it could be deduced that the optimization effect of the DL algorithm in depth image reconstruction is more pronounced as the other algorithms’ performances in depth image reconstruction are not so well as those in intensity image reconstruction while the DL algorithm performs even better in depth image reconstruction. These findings demonstrate that the DL algorithm produces superior image quality in the reconstruction of detailed maps. It has to be admitted that the filtering in DL image reconstruction also plays a great role in suppressing noise, some of which is inevitable for a ToF camera thanks to the multi-path interference and the mixed pixels and so on could be eliminated.

3.2. Recovering Depth Images Using Different Methods through Scattering Media

Due to the influence of scattering media, the image quality of ToF cameras deteriorates when imaging through such media. However, the CGI possesses strong anti-interference capabilities and can effectively resist the impact of atmospheric turbulence and scattering media on the image quality. To test whether the proposed scheme can image an object through a scattering medium, we inserted a 0.5 mm thickness transparent plastic sheet into the optical path between the ToF camera and the TRP in Figure 4 as the scattering medium, where the sheet is covered with a layer of dust particles of various sizes to scatter the light beam illuminated on it. In this scenario, the data collected by the ToF camera can be categorized into information from objects affected by scattering media and information from objects unaffected by scattering media. The imaging of the target object information in the ToF camera can be represented as
D ( x ) = α D ( x ) + D s ( x ) ,
In this context, α ( 0 < α < 1 ) represents the transmittance ratio of the scattering medium, while α D ( x ) and D s ( x ) , respectively, represent the distribution of information regarding reflected light and scattered light from the object. The other experimental conditions and the setting of the ToF camera are not changed and the experimental results through the scattering medium are shown in Figure 10, while the corresponding PSNR values are displayed in Figure 11. In Figure 10a, the ToF depth map is almost completely fuzzy and indistinguishable, which means that the traditional ToF camera does not work in this situation. However, by using the ToF camera as a detector in the scheme, which could preserve more high-frequency information of the ToF maps that determine how sharp the edges of the maps are, the depth images can be successfully reconstructed by CGI, BP and TVAL3 at the SRs of 6.25%, 12.5%, 18.75%, 25%, 31.25% and 37.5%, which are shown in Figure 10b–d, respectively. By comparison with the reconstruction images in Figure 7 and Figure 11, it is found that the PSNR of each recovered image by each algorithms at each given SR becomes a little bit worse, at least in most cases we have measured. For example, when the sampling rate is 12.5%, the PSNR values of three algorithms are 5.09 dB, 4.91 dB and 6.17 dB. When the sampling rate is 25%, the PSNR values of the three algorithms are 6.16 dB, 6.05 dB and 6.81 dB. Although image quality recovery in a scattering medium is lower than in standard environments, our proposed solution still facilitates the use of ToF cameras in scattering media, offering a fresh approach to employing ToF cameras in complex environments.
Similarly, when imaging through scattering media using ToF cameras, we incorporate a low-pass filter in our post-processing image treatment. We still employ four image reconstruction algorithms for image restoration, namely CGI, BP, TVAL3, and untrained DL network. It is shown in Figure 12b–e that the images are reconstructed at the low SR of 6.25%, 12.5%, 18.75%, 25%, 31.25 and 37.5%. As expected, when using a low-pass filter, the quality of Image 12 is superior to Image 10, and the image quality increases with the increase in SR. Notably, when reconstructing images through scattering media at a low SR of 6.25% in Figure 13, the PSNRs of depth maps reconstructed by CGI, BP, TVAL3, and DL are 7.03, 7.46, 7.52 and 9.01 dB, respectively. When reconstructing images through scattering media at SR of 25%, the PSNRs of depth images reconstructed by CGI, BP, TVAL3, and DL are 9.59, 9.84, 10.21 and 11.73 dB, respectively. Correspondingly, the PSNRs of the reconstructed depth images are 4.76, 5.95, and 9.00 dB, respectively. These results suggest that DL is particularly effective in reconstructing high-quality images through scattering media at low SRs. DL needs to input an image into the network to be the original image used for iteration, and the number of the iteration and the SR that need to obtain the clear recovered image will decrease when the image is blurry. But this explanation still needs more experimental verification. Therefore, by combing the CGI-based scheme and the ToF camera, the interference caused by the scattering media could be suppressed to realize the imaging of the target object successfully.

4. Conclusions

In conclusion, we have effectively demonstrated innovative applications of ToF cameras as bucket-detectors in both typical environments and scattering media. The introduction of the second-order correlation transformation efficiently mitigates noise originating from the ToF camera itself and direct ambient light. Our approach capitalizes on ToF camera depth information, utilizing various algorithms, including CGI, CS, and DL, to reconstruct images of target objects. Within our proposed framework, we address the common challenges of low resolution and a low signal-to-noise ratio encountered in depth cameras. Research findings indicate that both CGI and CS algorithms are capable of reconstructing negative depth images, and we have meticulously analyzed and mitigated the factors causing negative depth images. Particularly at low sampling rates, CGI and CS yield higher-quality image recovery compared to the original depth images. Remarkably, untrained deep learning networks exhibit substantial advantages in super-low subsampling, achieving a sampling rate of 6.25%, well below the Nyquist limit for image recovery. Furthermore, we have successfully facilitated the deployment of ToF cameras in scattering media, with the proof-of-concept demonstration underscoring the potential of ToF cameras in challenging scenarios such as haze, rain, snow, and underwater environments. Our research provides another avenue for the application of ToF cameras and opens up the possibility of integrating ToF cameras into other imaging systems.

Author Contributions

Conceptualization, T.-L.W.; methodology, T.-L.W. and L.A.; validation, T.-L.W. and L.A.; writing—original draft preparation, T.-L.W. and L.A.; writing—review and editing, Z.-B.S.; data curation, T.-L.W., L.A. and J.Z.; supervision, Z.-B.S.; funding acquisition, Z.-B.S. All authors have read and agreed to the published version of the manuscript.

Funding

This study was funded by National key research and development program (2016YFE0131500); Scientific Instrument Developing Project of the Chinese Academy of Sciences, Grant No.YJKYYQ20190008.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Research data from this study will be made available upon request by contacting the authors.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Szeliski, R. Computer Vision: Algorithms and Applications; Springer Nature: Berlin/Heidelberg, Germany, 2022. [Google Scholar]
  2. Zhang, S.; Huang, P.S. Novel method for structured light system calibration. Opt. Eng. 2006, 45, 083601. [Google Scholar] [CrossRef]
  3. Soltanlou, K.; Latifi, H. Three-dimensional imaging through scattering media using a single pixel detector. Appl. Opt. 2019, 58, 7716–7726. [Google Scholar] [CrossRef]
  4. Geng, J. Structured-light 3D surface imaging: A tutorial. Adv. Opt. Photonics 2011, 3, 128–160. [Google Scholar] [CrossRef]
  5. Salvi, J.; Fernandez, S.; Pribanic, T.; Llado, X. A state of the art in structured light patterns for surface profilometry. Pattern Recognit. 2010, 43, 2666–2680. [Google Scholar] [CrossRef]
  6. Zuo, C.; Huang, L.; Zhang, M.; Chen, Q.; Asundi, A. Temporal phase unwrapping algorithms for fringe projection profilometry: A comparative review. Opt. Lasers Eng. 2016, 85, 84–103. [Google Scholar] [CrossRef]
  7. Bhandari, A.; Raskar, R. Signal processing for time-of-flight imaging sensors: An introduction to inverse problems in computational 3-D imaging. IEEE Signal Process. Mag. 2016, 33, 45–58. [Google Scholar] [CrossRef]
  8. Park, J.; Kim, H.; Tai, Y.W.; Brown, M.S.; Kweon, I. High quality depth map upsampling for 3D-TOF cameras. In Proceedings of the 2011 International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 1623–1630. [Google Scholar] [CrossRef]
  9. Sun, M.J.; Edgar, M.P.; Gibson, G.M.; Sun, B.; Radwell, N.; Lamb, R.; Padgett, M.J. Single-pixel three-dimensional imaging with time-based depth resolution. Nat. Commun. 2016, 7, 12010. [Google Scholar] [CrossRef]
  10. Howland, G.A.; Lum, D.J.; Ware, M.R.; Howell, J.C. Photon counting compressive depth mapping. Opt. Express 2013, 21, 23822–23837. [Google Scholar] [CrossRef] [PubMed]
  11. Kirmani, A.; Colaço, A.; Wong, F.N.; Goyal, V.K. Exploiting sparsity in time-of-flight range acquisition using a single time-resolved sensor. Opt. Express 2011, 19, 21485–21507. [Google Scholar] [CrossRef] [PubMed]
  12. Sun, B.; Edgar, M.P.; Bowman, R.; Vittert, L.E.; Welsh, S.; Bowman, A.; Padgett, M.J. 3D computational imaging with single-pixel detectors. Science 2013, 340, 844–847. [Google Scholar] [CrossRef]
  13. Velten, A.; Wu, D.; Jarabo, A.; Masia, B.; Barsi, C.; Joshi, C.; Lawson, E.; Bawendi, M.; Gutierrez, D.; Raskar, R. Femto-photography: Capturing and visualizing the propagation of light. ACM Trans. Graph. (ToG) 2013, 32, 1–8. [Google Scholar] [CrossRef]
  14. Heide, F.; Hullin, M.B.; Gregson, J.; Heidrich, W. Low-budget transient imaging using photonic mixer devices. ACM Trans. Graph. (ToG) 2013, 32, 1–10. [Google Scholar] [CrossRef]
  15. Peters, C.; Klein, J.; Hullin, M.B.; Klein, R. Solving trigonometric moment problems for fast transient imaging. ACM Trans. Graph. (ToG) 2015, 34, 1–11. [Google Scholar] [CrossRef]
  16. Kirmani, A.; Hutchison, T.; Davis, J.; Raskar, R. Looking around the corner using transient imaging. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 159–166. [Google Scholar] [CrossRef]
  17. Velten, A.; Willwacher, T.; Gupta, O.; Veeraraghavan, A.; Bawendi, M.G.; Raskar, R. Recovering three-dimensional shape around a corner using ultrafast time-of-flight imaging. Nat. Commun. 2012, 3, 745. [Google Scholar] [CrossRef]
  18. Heide, F.; Xiao, L.; Heidrich, W.; Hullin, M.B. Diffuse mirrors: 3D reconstruction from diffuse indirect illumination using inexpensive time-of-flight sensors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3222–3229. [Google Scholar] [CrossRef]
  19. O’Toole, M.; Heide, F.; Xiao, L.; Hullin, M.B.; Heidrich, W.; Kutulakos, K.N. Temporal frequency probing for 5D transient analysis of global light transport. ACM Trans. Graph. (ToG) 2014, 33, 1–11. [Google Scholar] [CrossRef]
  20. Wu, D.; Wetzstein, G.; Barsi, C.; Willwacher, T.; O’Toole, M.; Naik, N.; Dai, Q.; Kutulakos, K.; Raskar, R. Frequency analysis of transient light transport with applications in bare sensor imaging. In Proceedings of the Computer Vision—ECCV 2012: 12th European Conference on Computer Vision, Florence, Italy, 7–13 October 2012; Proceedings, Part I 12. Springer: Berlin/Heidelberg, Germany, 2012; pp. 542–555. [Google Scholar] [CrossRef]
  21. Heide, F.; Heidrich, W.; Hullin, M.; Wetzstein, G. Doppler time-of-flight imaging. ACM Trans. Graph. (ToG) 2015, 34, 1–11. [Google Scholar] [CrossRef]
  22. Heide, F.; Xiao, L.; Kolb, A.; Hullin, M.B.; Heidrich, W. Imaging in scattering media using correlation image sensors and sparse convolutional coding. Opt. Express 2014, 22, 26338–26350. [Google Scholar] [CrossRef] [PubMed]
  23. Shapiro, J.H. Computational ghost imaging. Phys. Rev. A 2008, 78, 061802. [Google Scholar] [CrossRef]
  24. Pittman, T.B.; Shih, Y.; Strekalov, D.; Sergienko, A.V. Optical imaging by means of two-photon quantum entanglement. Phys. Rev. A 1995, 52, R3429. [Google Scholar] [CrossRef] [PubMed]
  25. Ferri, F.; Magatti, D.; Gatti, A.; Bache, M.; Brambilla, E.; Lugiato, L.A. High-resolution ghost image and ghost diffraction experiments with thermal light. Phys. Rev. Lett. 2005, 94, 183602. [Google Scholar] [CrossRef]
  26. Cheng, J. Ghost imaging through turbulent atmosphere. Opt. Express 2009, 17, 7916–7921. [Google Scholar] [CrossRef]
  27. Wu, Y.; Yang, Z.; Tang, Z. Experimental study on anti-disturbance ability of underwater ghost imaging. Laser Optoelectron. Prog. 2021, 58, 611002. [Google Scholar] [CrossRef]
  28. Takhar, D.; Laska, J.N.; Wakin, M.B.; Duarte, M.F.; Baron, D.; Sarvotham, S.; Kelly, K.F.; Baraniuk, R.G. A new compressive imaging camera architecture using optical-domain compression. Proc. Comput. Imaging IV 2006, 6065, 43–52. [Google Scholar] [CrossRef]
  29. Donoho, D.L. Compressed sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  30. Duarte, M.F.; Davenport, M.A.; Takhar, D.; Laska, J.N.; Sun, T.; Kelly, K.F.; Baraniuk, R.G. Single-pixel imaging via compressive sampling. IEEE Signal Process. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef]
  31. Katz, O.; Bromberg, Y.; Silberberg, Y. Compressive ghost imaging. Appl. Phys. Lett. 2009, 95, 131110. [Google Scholar] [CrossRef]
  32. Katkovnik, V.; Astola, J. Compressive sensing computational ghost imaging. JOSA A 2012, 29, 1556–1567. [Google Scholar] [CrossRef]
  33. Yu, W.K.; Li, M.F.; Yao, X.R.; Liu, X.F.; Wu, L.A.; Zhai, G.J. Adaptive compressive ghost imaging based on wavelet trees and sparse representation. Opt. Express 2014, 22, 7133–7144. [Google Scholar] [CrossRef]
  34. Chen, Z.; Shi, J.; Zeng, G. Object authentication based on compressive ghost imaging. Appl. Opt. 2016, 55, 8644–8650. [Google Scholar] [CrossRef] [PubMed]
  35. Song, X.; Dai, Y.; Qin, X. Deep depth super-resolution: Learning depth super-resolution using deep convolutional neural network. In Proceedings of the Computer Vision—ACCV 2016: 13th Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; Revised Selected Papers, Part IV 13. Springer: Berlin/Heidelberg, Germany, 2017; pp. 360–376. [Google Scholar] [CrossRef]
  36. Hinton, G.E.; Osindero, S.; Teh, Y.W. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef] [PubMed]
  37. Ranzato, M.; Boureau, Y.L.; Cun, Y. Sparse feature learning for deep belief networks. In NeurIPS Proceedings: Advances in Neural Information Processing Systems 20 (NIPS 2007); Curran Associates: San Jose, CA, USA, 2007. [Google Scholar]
  38. Tao, Q.; Li, L.; Huang, X.; Xi, X.; Wang, S.; Suykens, J.A. Piecewise linear neural networks and deep learning. Nat. Rev. Methods Prim. 2022, 2, 42. [Google Scholar] [CrossRef]
  39. Lyu, M.; Wang, W.; Wang, H.; Wang, H.; Li, G.; Chen, N.; Situ, G. Deep-learning-based ghost imaging. Sci. Rep. 2017, 7, 17865. [Google Scholar] [CrossRef]
  40. He, Y.; Wang, G.; Dong, G.; Zhu, S.; Chen, H.; Zhang, A.; Xu, Z. Ghost imaging based on deep learning. Sci. Rep. 2018, 8, 6469. [Google Scholar] [CrossRef]
  41. Shimobaba, T.; Endo, Y.; Nishitsuji, T.; Takahashi, T.; Nagahama, Y.; Hasegawa, S.; Sano, M.; Hirayama, R.; Kakue, T.; Shiraki, A.; et al. Computational ghost imaging using deep learning. Opt. Commun. 2018, 413, 147–151. [Google Scholar] [CrossRef]
  42. Barbastathis, G.; Ozcan, A.; Situ, G. On the use of deep learning for computational imaging. Optica 2019, 6, 921–943. [Google Scholar] [CrossRef]
  43. Nah, S.; Hyun Kim, T.; Mu Lee, K. Deep multi-scale convolutional neural network for dynamic scene deblurring. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3883–3891. [Google Scholar] [CrossRef]
  44. Wang, C.; Mei, X.; Pan, L.; Wang, P.; Li, W.; Gao, X.; Bo, Z.; Chen, M.; Gong, W.; Han, S. Airborne near infrared three-dimensional ghost imaging lidar via sparsity constraint. Remote Sens. 2018, 10, 732. [Google Scholar] [CrossRef]
  45. Mei, X.; Wang, C.; Pan, L.; Wang, P.; Gong, W.; Han, S. Experimental demonstration of vehicle-borne near infrared three-dimensional ghost imaging LiDAR. In Proceedings of the 2019 Conference on Lasers and Electro-Optics (CLEO), San Jose, CA, USA, 5–10 May 2019; pp. 1–2. [Google Scholar] [CrossRef]
  46. Li, Z.P.; Huang, X.; Jiang, P.Y.; Hong, Y.; Yu, C.; Cao, Y.; Zhang, J.; Xu, F.; Pan, J.W. Super-resolution single-photon imaging at 8.2 kilometers. Opt. Express 2020, 28, 4076–4087. [Google Scholar] [CrossRef]
  47. Li, Z.P.; Ye, J.T.; Huang, X.; Jiang, P.Y.; Cao, Y.; Hong, Y.; Yu, C.; Zhang, J.; Zhang, Q.; Peng, C.Z.; et al. Single-photon imaging over 200 km. Optica 2021, 8, 344–349. [Google Scholar] [CrossRef]
  48. Foix, S.; Alenya, G.; Torras, C. Lock-in time-of-flight (ToF) cameras: A survey. IEEE Sens. J. 2011, 11, 1917–1926. [Google Scholar] [CrossRef]
  49. Lange, R.; Seitz, P. Solid-state time-of-flight range camera. IEEE J. Quantum Electron. 2001, 37, 390–397. [Google Scholar] [CrossRef]
  50. Kolb, A.; Barth, E.; Koch, R. ToF-sensors: New dimensions for realism and interactivity. In Proceedings of the 2008 IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, Anchorage, AK, USA, 23–28 June 2008; pp. 1–6. [Google Scholar] [CrossRef]
  51. Oggier, T.; Lehmann, M.; Kaufmann, R.; Schweizer, M.; Richter, M.; Metzler, P.; Lang, G.; Lustenberger, F.; Blanc, N. An all-solid-state optical range camera for 3D real-time imaging with sub-centimeter depth resolution (SwissRanger). Proc. Opt. Des. Eng. 2004, 5249, 534–545. [Google Scholar] [CrossRef]
  52. Li, L. Time-of-Flight Camera—An Introduction; Texas Instruments, Technical White Paper; Texas Instruments: Dallas, TX, USA, 2014; p. SLOA190B. [Google Scholar]
  53. Liu, S.; Meng, X.; Yin, Y.; Wu, H.; Jiang, W. Computational ghost imaging based on an untrained neural network. Opt. Lasers Eng. 2021, 147, 106744. [Google Scholar] [CrossRef]
  54. Wang, F.; Wang, C.; Chen, M.; Gong, W.; Zhang, Y.; Han, S.; Situ, G. Far-field super-resolution ghost imaging with a deep neural network constraint. Light. Sci. Appl. 2022, 11, 1. [Google Scholar] [CrossRef]
  55. Lin, J.; Yan, Q.; Lu, S.; Zheng, Y.; Sun, S.; Wei, Z. A Compressed Reconstruction Network Combining Deep Image Prior and Autoencoding Priors for Single-Pixel Imaging. Photonics 2022, 9, 343. [Google Scholar] [CrossRef]
  56. Wang, C.H.; Li, H.Z.; Bie, S.H.; Lv, R.B.; Chen, X.H. Single-Pixel Hyperspectral Imaging via an Untrained Convolutional Neural Network. Photonics 2023, 10, 224. [Google Scholar] [CrossRef]
  57. Li, F.; Zhao, M.; Tian, Z.; Willomitzer, F.; Cossairt, O. Compressive ghost imaging through scattering media with deep learning. Opt. Express 2020, 28, 17395–17408. [Google Scholar] [CrossRef]
  58. Rizvi, S.; Cao, J.; Hao, Q. Deep learning based projector defocus compensation in single-pixel imaging. Opt. Express 2020, 28, 25134–25148. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The ranging principle of the CW modulated ToF.
Figure 1. The ranging principle of the CW modulated ToF.
Photonics 10 01223 g001
Figure 2. The schematic diagrams of CGI. The size of the object is 32 × 32 pixels, the focal length of the converging lens is 50mm, the detector has a resolution of 1280 × 960 pixels, and the dashed line connecting to the computer represents the computer simultaneously controlling the detector and DMD.
Figure 2. The schematic diagrams of CGI. The size of the object is 32 × 32 pixels, the focal length of the converging lens is 50mm, the detector has a resolution of 1280 × 960 pixels, and the dashed line connecting to the computer represents the computer simultaneously controlling the detector and DMD.
Photonics 10 01223 g002
Figure 3. Schematic diagram of the image reconstruction using a neural network.
Figure 3. Schematic diagram of the image reconstruction using a neural network.
Photonics 10 01223 g003
Figure 4. The schematic diagrams of CGI based on a ToF camera.
Figure 4. The schematic diagrams of CGI based on a ToF camera.
Photonics 10 01223 g004
Figure 5. Experimental results of imaging reconstruction using depth maps at different SRs. (a) target object, (b) ToF image, (ce) the recovered images by CGI, BP and TVAL3. The SRs from left to right is 6.25%, 12.5%, 18.75%, 25%, 31.25% and 37.5%.
Figure 5. Experimental results of imaging reconstruction using depth maps at different SRs. (a) target object, (b) ToF image, (ce) the recovered images by CGI, BP and TVAL3. The SRs from left to right is 6.25%, 12.5%, 18.75%, 25%, 31.25% and 37.5%.
Photonics 10 01223 g005
Figure 6. The original images exported by the ToF camera. (a) ToF original depth image, and (b) the original depth image in grayscale.
Figure 6. The original images exported by the ToF camera. (a) ToF original depth image, and (b) the original depth image in grayscale.
Photonics 10 01223 g006
Figure 7. Plots of the PSNRs of the reconstructed depth images versus the SRs by different algorithms. The black, red and blue lines denote the PSNRs by CGI, BP and TVAL3.
Figure 7. Plots of the PSNRs of the reconstructed depth images versus the SRs by different algorithms. The black, red and blue lines denote the PSNRs by CGI, BP and TVAL3.
Photonics 10 01223 g007
Figure 8. Experimental results of imaging reconstruction at different SRs. (a) ToF image, (be) the recovered images by CGI, BP, TVAL3 and DL. The SRs from left to right are 6.25%, 12.5%, 18.75%, 25%, 31.25 and 37.5%.
Figure 8. Experimental results of imaging reconstruction at different SRs. (a) ToF image, (be) the recovered images by CGI, BP, TVAL3 and DL. The SRs from left to right are 6.25%, 12.5%, 18.75%, 25%, 31.25 and 37.5%.
Photonics 10 01223 g008
Figure 9. Plots comparing the PSNR and SRs for depth image reconstruction using different algorithms after applying low-pass filtering.
Figure 9. Plots comparing the PSNR and SRs for depth image reconstruction using different algorithms after applying low-pass filtering.
Photonics 10 01223 g009
Figure 10. Experimental results of reconstruction using the depth maps through the scattering media at different SRs. (a) ToF image, (bd) the recovered images by CGI, BP and TVAL3. The SRs from left to right are 6.25%, 12.5%, 18.75%, 25%, 31.25% and 37.5%.
Figure 10. Experimental results of reconstruction using the depth maps through the scattering media at different SRs. (a) ToF image, (bd) the recovered images by CGI, BP and TVAL3. The SRs from left to right are 6.25%, 12.5%, 18.75%, 25%, 31.25% and 37.5%.
Photonics 10 01223 g010
Figure 11. Plots comparing the PSNR and SRs for the reconstruction of depth images through scattering media using different algorithms.
Figure 11. Plots comparing the PSNR and SRs for the reconstruction of depth images through scattering media using different algorithms.
Photonics 10 01223 g011
Figure 12. Experimental results of reconstruction using the depth maps through the scattering media at different SRs. (a) ToF image, (bd) the recovered images by CGI, BP and TVAL3. The SRs from left to right are 6.25%, 12.5%, 18.75%, 25%, 31.25% and 37.5%.
Figure 12. Experimental results of reconstruction using the depth maps through the scattering media at different SRs. (a) ToF image, (bd) the recovered images by CGI, BP and TVAL3. The SRs from left to right are 6.25%, 12.5%, 18.75%, 25%, 31.25% and 37.5%.
Photonics 10 01223 g012
Figure 13. Plots comparing PSNR and SRs for the reconstruction of depth images through scattering media using different algorithms after applying low-pass filtering.
Figure 13. Plots comparing PSNR and SRs for the reconstruction of depth images through scattering media using different algorithms after applying low-pass filtering.
Photonics 10 01223 g013
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, T.-L.; Ao, L.; Zheng, J.; Sun, Z.-B. Reconstructing Depth Images for Time-of-Flight Cameras Based on Second-Order Correlation Functions. Photonics 2023, 10, 1223. https://doi.org/10.3390/photonics10111223

AMA Style

Wang T-L, Ao L, Zheng J, Sun Z-B. Reconstructing Depth Images for Time-of-Flight Cameras Based on Second-Order Correlation Functions. Photonics. 2023; 10(11):1223. https://doi.org/10.3390/photonics10111223

Chicago/Turabian Style

Wang, Tian-Long, Lin Ao, Jie Zheng, and Zhi-Bin Sun. 2023. "Reconstructing Depth Images for Time-of-Flight Cameras Based on Second-Order Correlation Functions" Photonics 10, no. 11: 1223. https://doi.org/10.3390/photonics10111223

APA Style

Wang, T. -L., Ao, L., Zheng, J., & Sun, Z. -B. (2023). Reconstructing Depth Images for Time-of-Flight Cameras Based on Second-Order Correlation Functions. Photonics, 10(11), 1223. https://doi.org/10.3390/photonics10111223

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop