Next Article in Journal
Performance Analysis of the Four-Wave Mixing Effect with Various Tapered Fiber Waist Diameters
Next Article in Special Issue
Phase Imaging through Scattering Media Using Incoherent Light Source
Previous Article in Journal
Evaluation of Renal Ischemia–Reperfusion Injury Using Optical Coherence Tomography Based on Fractal Dimension
Previous Article in Special Issue
Recognition of Orbital Angular Momentum of Vortex Beams Based on Convolutional Neural Network and Multi-Objective Classifier
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

TSDSR: Temporal–Spatial Domain Denoise Super-Resolution Photon-Efficient 3D Reconstruction by Deep Learning

1
School of Computer Science and Technology, Zhejiang Sci-Tech University, Hangzhou 310018, China
2
Key Laboratory of Optical Field Manipulation of Zhejiang Province, Zhejiang Sci-Tech University, Hangzhou 310018, China
3
Institute of Land Aviation, Beijing 101121, China
*
Authors to whom correspondence should be addressed.
Photonics 2023, 10(7), 744; https://doi.org/10.3390/photonics10070744
Submission received: 25 April 2023 / Revised: 20 June 2023 / Accepted: 27 June 2023 / Published: 28 June 2023
(This article belongs to the Special Issue Nonlinear Optics and Hyperspectral Polarization Imaging)

Abstract

:
The combination of a single-photon avalanche diode detector with a high-sensitivity and photon-efficient reconstruction algorithm can realize the reconstruction of target range image from weak light signal conditions. The limited spatial resolution of the detector and the substantial background noise remain significant challenges in the actual detection process, hindering the accuracy of 3D reconstruction techniques. To address this challenge, this paper proposes a denoising super-resolution reconstruction network based on generative adversarial network (GAN) design. Soft thresholding is incorporated into the deep architecture as a nonlinear transformation layer to effectively filter out noise. Moreover, the Unet-based discriminator is introduced to complete the high-precision detail reconstruction. The experimental results show that the proposed network can achieve high-quality super-resolution range imaging. This approach has the potential to enhance the accuracy and quality of long-range imaging in weak light signal conditions, with broad applications in fields such as robotics, autonomous vehicles, and biomedical imaging.

1. Introduction

Active optical imaging has lower environmental requirements and is more versatile than passive imaging detection. However, optical detection typically requires a large number of photons to reduce background noise. This can be challenging in remote sensing [1], non-visual imaging [2], and other applications [3], where the amount of signal photons is limited by light flux and integration time. Traditional systems based on photomultiplier tubes struggle to accurately reconstruct scene information under these conditions. single-photon avalanche diode (SPAD) is an avalanche photodiode operating in a Geiger mode, which achieves detection sensitivity at the single-photon level by utilizing the avalanche breakdown phenomenon. SPAD-based laser radar systems are highly sensitive and can detect very weak echo signals [4]. The SPAD detector records the photon arrival time, and a 3D reconstruction algorithm calculates the target range information based on the photon count histogram.
During the measurement process, dark counts can introduce additional noise and uncertainties, impacting the accuracy of imaging. Moreover, sparse echo signal scenes are often accompanied by strong background noise, such as sunlight, fog, rain, and snow. To improve detection accuracy in various scenarios, it is necessary to reduce the impact of background noise on the system using a reconstruction algorithm. Several methods have been proposed to reconstruct high-noise SPAD detection data [5,6,7,8,9,10]. Moreover, in recent years, deep learning-based methods have advanced photon-efficient reconstruction imaging even further [11,12,13,14,15].
For large-scale SPAD arrays, the packaging technology poses difficulties, leading to a mutual restriction between temporal and spatial resolution. To achieve better range accuracy and reduce noise from crosstalk between pixels, the spatial resolution of the SPAD array is often sacrificed, resulting in a lack of detail in reconstructed images. However, there is currently a lack of methods that utilize deep learning to improve the resolution of photon-efficient range imaging. Single-image super-resolution (SISR) is a widely used computer vision technique in astronomical and biomedical imaging. Neural networks, particularly GANs, offer a wide range of uses in SISR and have expanded quickly [16,17,18,19,20]. The introduction of SISR technology into photon-efficient imaging can further enhance reconstruction image resolution [21,22,23,24,25,26].
In this paper, a photon-efficient super-resolution convolutional neural network is proposed. The network uses the Non-Local Sparse Attention (NLSA) module [15] to extract features and addresses the high noise typically present in SPAD detector data by directly solving 3D data and introducing adaptive soft thresholding for denoising [27]. However, soft thresholding may filter out weak echo signals, so a more reasonable loss function is needed to balance noise removal and signal preservation. To achieve this, the proposed method uses a Unet-based discriminator to output both local and global discriminant information [28]. The discriminator guides the network to generate more details using pixel-level discrimination information, and the total hybrid loss function includes Mean Absolute Error (MAE), Structural Similarity (SSIM), and adversarial loss from the Unet-based discriminators. Simulation experiments conducted in different noise environments and scenarios demonstrate the proposed method’s excellent reconstruction results through qualitative and quantitative analysis and comparison with previously reported algorithms. Ablation experiments explore the influence of each module on network performance. The proposed network’s 3D super-resolution reconstruction performance is further validated through experiments on a pulsed lidar system based on an SPAD array in the real world.

2. Related Work

2.1. Reconstruction Photon-Efficient Imaging

Recently, there has been a significant amount of work focused on improving the quality of reconstructed images by suppressing background noise. This has been achieved through various techniques such as leveraging the distribution characteristics of signal photons or integrating multiple sensors. Shin et al. [5,6] suppressed Poisson noise on a pixel-by-pixel basis by establishing a SPAD trigger probability model. Rapp et al. [7] improved the great likelihood estimation method by incorporating an adaptive super-resolution algorithm that leverages the signal photon distribution law. In challenging conditions, Halimi et al. [8] utilized a layered Bayesian algorithm combined with multi-spectral single-photon lidar to estimate range image. Chen et al. [9] proposed a deep domain adversarial adaptation technique that addresses the domain shift issue in photon-efficient imaging by making advantage of a potent network structure.
With the development of deep learning technique, neural network has become a new direction to solve the problem of efficient photon imaging with its powerful feature extraction ability [11]. Peng et al. [12] proposed an end-to-end reconfigurable network with a denoising module as the main architecture. Non-local attention module was introduced to extract the temporal-spatial domain long-range correlation feature of SPAD detection data. Zang et al. [13] proposed a lightweight neural network that can be implemented on an embedded hardware platform. Unet was used as the main body of the network to integrate multidimensional space-time spatial characteristics. Zhao et al. [14] constrained the network with gradient regularization function and introduced the ADAG module into the Unet network model to recover more accurate edge details. Yang et al. [15] designed a convolutional encoder to directly reconstruct the incoming 3D data into range and reflectivity images. NLSA module was used to extract non-local features with low computational cost.

2.2. Single-Image Super-Resolution

Super-Resolution Convolutional Neural Network (SRCNN) [16] first applies deep convolutional neural networks to SISR. Over the years, researchers have made significant advancements in optimizing the network architecture by incorporating various techniques such as stacking convolutional layers, adding jump connections and integrating attention modules [17,18,19,20]. For instance, Zhang et al. [17] incorporated the channel attention module into extremely deep residual network, which improved reconstruction accuracy significantly. Wang et al. [20] explored the use of sparsity in SISR by incorporating a sparse mask that identifies important and unimportant regions in images, leading to further enhancements in the reconstruction process.
GANs have been used in SISR to improve the perceived quality of the output images. Ledig et al. [21] proposed the SRGAN network, which uses adversarial losses and content losses to strengthen the generator. Enhanced Super-Resolution Generative Adversarial Networks (ESRGAN) [22] refine the adversarial training process by introducing perceived loss and a Residual-in-Residual Dense Block (RRDB). Ma et al. [23] introduced a gradient guidance via an additional branch in the network to alleviate the problem of structural distortion and inconsistency. Li et al. [24] addressed potentially undesired image artifacts using a regional awareness adversarial learning strategy. Liang et al. [25] proposed an LDL framework to regularize adversarial training by clearly distinguishing between visual artifacts and realistic details. These advancements have significantly improved the performance of GAN-based SISR methods, making GANs a popular choice in the field.

3. Method

3.1. Forward Model

Figure 1 illustrates a typical pulsed lidar system based on photon-efficient imaging technology. The system comprises a pulsed laser that emits a pulsed laser beam s(t) at a fixed period T. A SPAD is used to collect the arrival time of the echo signal. In extremely weak echo scenes, the amount of light that can enter the detector during each repetition period is very small, typically less than one photon per pixel. Under these conditions, the impact of pulse pile-up on the detection result is negligible. For each pixel q of the SPAD array during a time period Δt, the expected number Φq of photons that can be detected is
Φ q = 0 Δ t [ σ α q s ( t 2 z q c ) + B ] d t ,
where σ contains the quantum efficiency of SPAD and distance attenuation, c is the speed of light, αq represents the reflectivity of the target area corresponding to the pixel cell, zq is the distance between the target area and the detector, and the response caused by environmental noise and dark count is denoted as B.
To avoid producing distance aliasing during detection, it is essential to ensure that T > 2zmax/c, where zmax represents the farthest distance of the target. During each illumination pulse repetition cycle, detections by the SPAD are treated as separate events. The probability density function of the avalanche multiplication event of pixel q in a full detection process can be written as follows:
f q ( t , z q ) = σ α q s ( t 2 z q / c ) + B σ α q T 0 T 1 s ( t 2 z q / c ) d t + B ( T 1 T 0 ) ,
where T0 and T1 represent the start time and end time of detection.
The SPAD detector generates point cloud data M in the low light flux scenario by accumulating many detection results. M contains the response of the detector to the echo signal photon and the background noise photon, which can be reconstructed to obtain the range image. By employing the maximum likelihood estimation (MLE) method [29], the range information on q can be expressed as follows:
z q M L = arg max z q [ 0 , c T ) i = 1 k q log [ s ( t ( i ) 2 z q / c ) ] ,
where kq is the total number of photons that the detector responds to on q. The proposed method is detailed in the next section.

3.2. Network Architecture

This research proposes the Temporal–Spatial domain Denoise Super-Resolution (TSDSR) neural network architecture based on GAN which comprises a generator and a discriminator in order to produce greater quality super-resolution range reconstruction. To reconstruct an image using a neural network, the estimation process can be expressed as follows:
{ Y r a n g e H R } = G θ = θ * ( M ) ,
where G(.) represents the generator neural network, θ is the parameters of the neural network, θ* is the parameters after training, and Y r a n g e H R is the High Resolution (HR) range image of the target scene.

3.2.1. Generator

The structure of the Residual Channel-wise Soft Thresholding (RCST) unit is illustrated in Figure 2. The main body is composed of residual units, and adaptive soft thresholding is added. Equation (2) indicates that in the presence of an echo signal, photon detection events have a higher probability density, and there is a greater chance of obtaining a high value in the statistical detection results. The reactions caused by background noise and dark counts are asymmetric and randomly dispersed. In photon reconstruction imaging tasks with low Signal-to-Background Ratio (SBR), it is essential to eliminate the impact of background noise and dark count.
To achieve denoising, the noise information is transformed into a feature close to zero through convolution, which is then further transformed into zero through soft thresholding. The soft thresholding function can be expressed as follows:
y = { x τ 0 x + τ x > τ τ x τ x < τ .
The selection of threshold τ is critical to ensuring optimum denoising performance in RCST. RCST combines soft thresholding with residual units. The feature information is transformed into a C-dimensional vector by absolute value global mean pooling after Rectified Linear Unit (ReLU) activation and convolution processing (AGAP). The vector value is then scaled to the range (0,1) with a sigmoid after two layers of fully connected (FC) processing. The output is multiplied element-wise with the vector output from AGAP processing to obtain a vector of size 1 × 1 × C, where each value represents a separate threshold for the corresponding channel. These thresholds are all positive and can be learned continuously during back propagation. RCST units avoid discarding too much useful information in a single filtering step by employing shortcut connections. By reusing RCST units in generators, noise in features is gradually reduced.
The NRC module integrates multiple functions, including denoising, feature extraction and data compression. With NRC as the main component of the network, the spatial resolution of the data is expanded by the upsampling module when the data are compressed to 64 channels. Finally, the feature data are multiplied by two 3 × 3 filters to produce an HR range image of the target scene.

3.2.2. Discriminator

In the GAN, the discriminator obtains discriminant information according to the input data distribution and uses the difference between the real image and the generated image discriminant information as the loss guidance generator. Improving the accuracy of the information expressed by the discriminator can prompt the generator to produce more realistic output. However, as a classification network, most discriminators cannot extract different features from the input data simultaneously. When trying to learn global semantics and details at the same time, these discriminators often lose their ability to express themselves effectively. Therefore, global and local feature extraction can be regarded as different tasks. Some studies use multiple independent discriminators to perceive the global and local information of images separately. The network proposed in this paper chooses U-net as the main body of the discriminator and obtains the global and local feature information from it. As shown in Figure 3, the DU structure is divided into two parts: the encoder performs the classification task to extract the global image features, while the decoder performs the semantic segmentation task to output the discrimination information pixel by pixel.
The encoder consists of multiple convolutions, activation functions, and downsampling operations to map a 128 × 128 × 1 input to a 4 × 4 × 1024 size feature. Global discriminant information is obtained through global summation pooling (GSP) and FC layers. The decoder uses a similar structure to gradually up-sample the encoder output features to the size of the discriminator input to complete the semantic segmentation task. During this process, the features from each layer of the encoder are merged with the decoder output features of the same resolution. By fusing shallow and deep information in this way, details can be better recovered. The global discriminant information is based on the similarity of deep image features. However, in super-resolution tasks, more attention needs to be paid to the image details. DU fuses encoder and decoder information to better guide the generator in narrowing the detail gap between real and fake samples.

3.2.3. Loss Function

The super-resolution mission is essentially an ill-posed problem. The use of mixed loss function can avoid imposing rigid constraints on Low Resolution (LR) to HR and reduce the difficulty of network training. MAE and adversarial loss are used as hybrid loss functions to guide generators in training. The total loss of the generator can be expressed as
L G = M A E λ 1 E Y [ λ 2 log D e n U ( G ( M ) ) + λ 3 log D d e U ( G ( M ) ) ] ,
where D e n U ( ) represents the encoder output of the discriminator, D d e U ( ) represents the discriminant matrix output by the decoder of the discriminator, n is the total number of pixels, and λ is the hyperparameter. In the process of network training in this paper, λ1 = 0.005, λ2 = 0.3 and λ3 = 0.7. This hyperparameter setting encourages the generator to pay more attention to local details. The loss of the discriminator consists of L D e n U calculated from the output of the encoder and L D d e U calculated from the output of the decoder:
L D e n U = E R [ log D e n U ( R ) ] E Y [ 1 log D e n U ( G ( M ) ) ] ,
L D d e U = E R [ log D d e U ( R ) ] E Y [ 1 log D d e U ( G ( M ) ) ] ,
where R is HR ground truth. The total discriminator loss LD is
L D = L D e n U + L D d e U .

4. Experiments

4.1. Dataset and Training Detail

Firstly, the histogram of photon detection probability is established by combining ground truth images of range and reflectivity in the NYU v2 [30] dataset according to Equation (2) Then, the histogram is sampled using a non-homogeneous Poisson process to obtain SPAD measurements. The spatial resolution of simulation data used in training is 64 × 64, with time divided into 1024 intervals, each interval being 80 ps apart. The measurements included 9 different noise scenes, with an average of 2, 10 and 50 background photons per pixel and an average of 2, 5 and 10 signal photons per pixel. A total of 38,624 and 2372 measurements were generated for training and validation, respectively. The deep learning neural network code was implemented using PyTorch. ADAM optimizer was used for iteration in the training process, with a learning rate of 2 × 10−4 and each epoch decaying to 0.95 of the previous one. Network training requires approximately 20 epochs, with a batch size of 4 and approximately 193 k iterations.

4.2. Numerical Simulation

A total of 72 sets of 128 × 128 × 1024 testing data were generated in the simulation experiments, which included 8 scenarios and 9 different noise levels (SBR = 10:2; 10:10; 10:50; 5:2; 5:10; 5:50; 2:2; 2:10; 2:50) provided by the Middlebury [31] dataset. The proposed method was compared with Peng’s [12] method, Zhao’s [14] method, and NLSA-Encoder [15]. In the simulation experiment, the proposed method achieved end-to-end reconstruction and super-resolution. In contrast, the compared methods first reconstructed the test data to obtain a 128 × 128 LR range image and then performed super-resolution using other methods.
The result of range image reconstruction in low SBR (2:50) Art scene is shown in Figure 4. The first row shows the super-resolution using the Bicubic method, while the second row shows the super-resolution using the ESRGAN neural network method trained on the range reconstruction results. In the numerical simulation experiments, the super-resolution results obtained using bicubic interpolation appear relatively blurry from a human visual perception standpoint.
In order to better compare the various reconstruction methods, this paper shows more results with ESRGAN in Figure 5. The Peak Signal-to-Noise Ratio (PSNR) was calculated using the ground truth range image and the HR range reconstruction result. Peng’s method outperformed in lower SBR environments due to its use of total variation (TV) as part of the loss. However, TV constraints obscure local details, resulting in less detail in the panes of the refactoring results in the Laundry scene compared to the other methods. Zhao’s method uses gradient regularization to penalize the neural network to obtain more accurate edges. In the high-noise environment, more singular points appear in Zhao reconstruction results. This situation shows that this method enhances the expression of some noise as well as the boundary information. The NLSA-Encoder method yields more balanced reconstruction results but lacks some details, such as the panes in the Laundry scene, the small square in the middle of the Moebius scene, and the lower-right corner of the Dolls scene. The proposed method provides better denoising performance and produces more detailed reconstruction results.
To compare the reconstruction performance of various methods, four quantitative indicators were selected. The following is the formula for each quantitative index.
Root Mean Square Error (RMSE) is represented by the following formula:
R M S E ( R , Y ) = 1 n 1 n ( R q - Y q ) 2 ,
accuracy is represented as follows:
A c c u r a c y ( C ) = 1 n 1 n C q ,
PSNR is formulated as
P S N R ( R , Y ) = 10 log 10 [ z max 2 / 1 n 1 n ( R q - Y q ) 2 ] ,
and the Universal Image Quality Index (UIQI) [32] can be expressed as follows:
U I Q I ( R , Y ) = 4 σ R Y μ R μ Y ( σ R 2 + σ Y 2 ) ( μ R 2 + μ Y 2 ) ,
where σ is variance, μ is mean value, n is total number of pixels. The value at any position of the confidence matrix C determined by the threshold δ can be represented as follows:
C q ( R , Y ) = { 0 max ( R q / Y q , Y q / R q ) > δ 1 max ( R q / Y q , Y q / R q ) < δ ,
Table 1 lists the average values of the indicators for each method in each noise environment. In terms of quantitative indexes, Peng’s method has poor overall performance. Zhao’s method performs well in environments with high signal photon numbers but is susceptible to noise. NLSA-Encoder is more prominent in low-SBR environments. The average performance of the proposed network is the best. This method performs best or second-best in all kinds of noise environments. The results show that TSDSR has the ability to accurately super-resolution reconstruct the range image.

4.3. Real-World Experiments

To validate the performance of the proposed method, a SPAD-based pulsed laser system was constructed as shown in Figure 1 to collect real-world experimental data. The fiber pulse laser emitted pulses with a pulse width of 1 ns, a peak power of 500 mW (near-field) or 1.5 mJ (far-field), and a wavelength of 1064 nm at a repetition rate of 20 KHz. The laser beam was diffused by an external lens, resulting in a beam divergence angle of 25 mrad (near-field) or 15 mrad (far-field). The laser transmitter triggered the SPAD synchronously with the pulse to detect the flight time of the echo photon. The SPAD model used was the GD5551 InGaAs, with a trigger exposure time of 4096 ns, time resolution of 1 ns, and spatial resolution of 64 × 64.
The target scene photos and reconstruction results of the real experiment data are illustrated in Figure 6. In the reconstruction process, MLE utilized 20,000 frames of data to reconstruct 64 × 64 LR range images as a reference, while the other methods used only 400 frames to reconstruct HR range images. Peng’s method failed to filter out the noise, whereas Zhao’s method restored more edge information, but the left-hand reconstruction of the middle doll was less effective. Although the NLSA-Encoder did not have much noise, it struggled to accurately capture details in certain weak echo regions, such as the edges of the dolls and flagpoles. The NLSA-Encoder-reconstructed rabbit doll on the left is smaller than the result of MLE method.
TSDSR showed the highest quality for the reconstructed data in the table part of the image. The shape of the right side of the table and the signboard on the rooftop were well restored. These details were difficult to observe in the LR reconstructed images of the MLE method, indicating the importance of super-resolution. The qualitative comparison of the reconstruction results of real-world experiments proved that the proposed method outperformed other methods in super-resolution range reconstruction.

5. Discussion

Photon-efficient imaging is limited by significant noise and photon detection hardware. In this paper, an end-to-end network is proposed for super-resolution range reconstruction from photon-efficient measurements. To thoroughly examine the contribution of each module to the network’s performance, three network architectures were compared: one without adaptive soft thresholding (A w/o AST), one without DU (B w/o DU), and one containing all components (C Proposed).
The ablation experiment used the same test data as the simulation experiment. Figure 7 shows the reconstruction results under low SBR (2:50) Art and Laundry scenes. The comparison highlights that the network’s reconstruction results without using adaptive soft thresholding introduce a lot of noise. While the adaptive soft thresholding module’s denoising capabilities are evident, it also leads to the problem of detail loss. Among the panes in the Laundry scene, the A network, which does not use adaptive soft thresholding, displays the most detailed results. During adaptive soft thresholding, some useful information may be accidentally filtered out. Choosing an appropriate threshold can reduce the likelihood of signal photon filtering, resulting in a higher-quality reconstructed image. In both test scenarios, the proposed network demonstrated outstanding performance in terms of image reconstruction quality and PSNR. This is because the DU penalty helps the network achieve a more reasonable threshold, which better balances the trade-off between detail recovery and denoising performance.
Compared to existing convolutional neural network approaches, the employed technique involves learnable soft thresholding to effectively filter out the substantial amount of background noise present in the original data captured by the SPAD detector. Furthermore, a Unet-based discriminator is introduced to guide the network in generating high-fidelity, high-resolution range images. This discriminator serves as both a classifier and a segmenter, enhancing the preservation of the signal while minimizing potential information loss due to soft thresholding. In numerical simulation experiments, the network effectively preserves fine details. However, in real-world experiments, there is an occurrence of unexpected filtering of some echo signals. For instance, in the window part of the far-field experiment (Figure 6), the resulting image appears excessively blurred, potentially due to the negative impact of the introduced soft thresholding. Future work can focus on further optimizing the thresholding process or selecting more suitable loss functions to enhance the network’s reconstruction performance.

6. Conclusions

To address the problem of poor single photon imaging quality caused by the limited spatial resolution of the SPAD array, a GAN-based neural network called TSDSR is utilized for super-resolution reconstruction. This method uses a residual structure and an NLSA module for feature extraction. The discriminator employs the Unet architecture, with the encoder performing classification by image and the decoder by pixel. This architectural improvement results in a stronger discriminator, which is encouraged to maintain a more powerful data representation, making it more difficult for the generator to deceive the discriminator and thereby enhancing the quality of generated samples. Soft threshold processing is added to the residual module to lessen the negative impact of background noise on reconstruction quality. The network’s high reconstruction performance is validated through numerical simulations of various scenarios, and ablation experiments are conducted to examine the influence of each module on network performance. Furthermore, the proposed method has been evaluated in real-world tests. Compared to other existing deep learning methods, the proposed network can reconstruct HR range images of the target with higher quality.

Author Contributions

Conceptualization, Z.T. and X.Y.; methodology, Z.T.; software, Z.T.; validation, Z.T., X.J. and X.Y.; formal analysis, J.H.; investigation, L.X.; resources, X.Y.; data curation, L.W.; writing—original draft preparation, Z.T.; writing—review and editing, X.Y. and X.J.; visualization, X.J.; supervision, B.Z.; project administration, X.Y.; funding acquisition B.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not Applicable.

Informed Consent Statement

Not Applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, Z.-P.; Ye, J.-T.; Huang, X.; Jiang, P.-Y.; Cao, Y.; Hong, Y.; Yu, C.; Zhang, J.; Zhang, Q.; Peng, C.-Z.; et al. Single-Photon Imaging over 200 km. Optica 2021, 8, 344–349. [Google Scholar] [CrossRef]
  2. Yang, J.; Li, J.; He, S.; Wang, L.V. Angular-Spectrum Modeling of Focusing Light inside Scattering Media by Optical Phase Conjugation. Optica 2019, 6, 250–256. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Maccarone, A.; McCarthy, A.; Ren, X.; Warburton, R.E.; Wallace, A.M.; Moffat, J.; Petillot, Y.; Buller, G.S. Underwater Depth Imaging Using Time-Correlated Single-Photon Counting. Opt. Express 2015, 23, 33911–33926. [Google Scholar] [CrossRef] [PubMed]
  4. Richardson, J.A.; Grant, L.A.; Henderson, R.K. Low Dark Count Single-Photon Avalanche Diode Structure Compatible With Standard Nanometer Scale CMOS Technology. IEEE Photonics Technol. Lett. 2009, 21, 1020–1022. [Google Scholar] [CrossRef] [Green Version]
  5. Shin, D.; Kirmani, A.; Goyal, V.K.; Shapiro, J.H. Photon-Efficient Computational 3-D and Reflectivity Imaging With Single-Photon Detectors. IEEE Trans. Comput. Imaging 2015, 1, 112–125. [Google Scholar] [CrossRef] [Green Version]
  6. Shin, D.; Xu, F.; Venkatraman, D.; Lussana, R.; Villa, F.; Zappa, F.; Goyal, V.K.; Wong, F.N.C.; Shapiro, J.H. Photon-Efficient Imaging with a Single-Photon Camera. Nat. Commun. 2016, 7, 12046. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Rapp, J.; Goyal, V.K. A Few Photons Among Many: Unmixing Signal and Noise for Photon-Efficient Active Imaging. IEEE Trans. Comput. Imaging 2017, 3, 445–459. [Google Scholar] [CrossRef]
  8. Halimi, A.; Maccarone, A.; Lamb, R.A.; Buller, G.S.; McLaughlin, S. Robust and Guided Bayesian Reconstruction of Single-Photon 3D Lidar Data: Application to Multispectral and Underwater Imaging. IEEE Trans. Comput. Imaging 2021, 7, 961–974. [Google Scholar] [CrossRef]
  9. Chen, S.; Halimi, A.; Ren, X.; McCarthy, A.; Su, X.; McLaughlin, S.; Buller, G.S. Learning Non-Local Spatial Correlations To Restore Sparse 3D Single-Photon Data. IEEE Trans. Image Process. 2020, 29, 3119–3131. [Google Scholar] [CrossRef] [PubMed]
  10. Chen, Y.; Yao, G.; Liu, Y.; Su, H.; Hu, X.; Pan, Y. Deep Domain Adversarial Adaptation for Photon-Efficient Imaging. Phys. Rev. Appl. 2022, 18, 54048. [Google Scholar] [CrossRef]
  11. Lindell, D.B.; O’Toole, M.; Wetzstein, G. Single-Photon 3D Imaging with Deep Sensor Fusion. ACM Trans. Graph. 2018, 37, 111–113. [Google Scholar] [CrossRef]
  12. Peng, J.; Xiong, Z.; Huang, X.; Li, Z.-P.; Liu, D.; Xu, F. Photon-Efficient 3d Imaging with a Non-Local Neural Network. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Volume 16, pp. 225–241. [Google Scholar]
  13. Zang, Z.; Xiao, D.; Li, D.D.-U. Non-Fusion Time-Resolved Depth Image Reconstruction Using a Highly Efficient Neural Network Architecture. Opt. Express 2021, 29, 19278–19291. [Google Scholar] [CrossRef] [PubMed]
  14. Zhao, X.; Jiang, X.; Han, A.; Mao, T.; He, W.; Chen, Q. Photon-Efficient 3D Reconstruction Employing a Edge Enhancement Method. Opt. Express 2022, 30, 1555–1569. [Google Scholar] [CrossRef] [PubMed]
  15. Yang, X.; Tong, Z.; Jiang, P.; Xu, L.; Wu, L.; Hu, J.; Yang, C.; Zhang, W.; Zhang, Y.; Zhang, J. Deep-Learning Based Photon-Efficient 3D and Reflectivity Imaging with a 64 64 Single-Photon Avalanche Detector Array. Opt. Express 2022, 30, 32948–32964. [Google Scholar] [CrossRef] [PubMed]
  16. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a Deep Convolutional Network for Image Super-Resolution. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Volume 13, pp. 184–199. [Google Scholar]
  17. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image Super-Resolution Using Very Deep Residual Channel Attention Networks. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  18. Zhang, Y.; Tian, Y.; Kong, Y.; Zhong, B.; Fu, Y. Residual Dense Network for Image Super-Resolution. In Proceedings of the IEEE conference on computer vision and pattern recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 2472–2481. [Google Scholar]
  19. Hussein, S.A.; Tirer, T.; Giryes, R. Correction Filter for Single Image Super-Resolution: Robustifying off-the-Shelf Deep Super-Resolvers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 1428–1437. [Google Scholar]
  20. Wang, L.; Dong, X.; Wang, Y.; Ying, X.; Lin, Z.; An, W.; Guo, Y. Exploring Sparsity in Image Super-Resolution for Efficient Inference. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, Nashville, TN, USA, 20–25 June 2021; pp. 4917–4926. [Google Scholar]
  21. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-Realistic Single Image Super-Resolution Using a Generative Adversarial Network. In Proceedings of the IEEE conference on computer vision and pattern recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  22. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. Esrgan: Enhanced Super-Resolution Generative Adversarial Networks. In Proceedings of the European conference on computer vision (ECCV) workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  23. Ma, C.; Rao, Y.; Cheng, Y.; Chen, C.; Lu, J.; Zhou, J. Structure-Preserving Super Resolution with Gradient Guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 7769–7778. [Google Scholar]
  24. Li, W.; Zhou, K.; Qi, L.; Lu, L.; Lu, J. Best-Buddy Gans for Highly Detailed Image Super-Resolution. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtual, 22 February–1 March 2022; Volume 36, pp. 1412–1420. [Google Scholar]
  25. Liang, J.; Zeng, H.; Zhang, L. Details or Artifacts: A Locally Discriminative Learning Approach to Realistic Image Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5657–5666. [Google Scholar]
  26. Mei, Y.; Fan, Y.; Zhou, Y. Image Super-Resolution with Non-Local Sparse Attention. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 3517–3526. [Google Scholar]
  27. Zhao, M.; Zhong, S.; Fu, X.; Tang, B.; Pecht, M. Deep Residual Shrinkage Networks for Fault Diagnosis. IEEE Trans. Ind. Informatics 2019, 16, 4681–4690. [Google Scholar] [CrossRef]
  28. Schonfeld, E.; Schiele, B.; Khoreva, A. A U-Net Based Discriminator for Generative Adversarial Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 8207–8216. [Google Scholar]
  29. Houwink, Q.; Kalisvaart, D.; Hung, S.-T.; Cnossen, J.; Fan, D.; Mos, P.; Ülkü, A.C.; Bruschini, C.; Charbon, E.; Smith, C.S. Theoretical Minimum Uncertainty of Single-Molecule Localizations Using a Single-Photon Avalanche Diode Array. Opt. Express 2021, 29, 39920–39929. [Google Scholar] [CrossRef] [PubMed]
  30. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor Segmentation and Support Inference from Rgbd Images. ECCV 2012, 7576, 746–760. [Google Scholar]
  31. Scharstein, D.; Pal, C. Learning Conditional Random Fields for Stereo. In Proceedings of the 2007 IEEE Conference on Computer Vision and Pattern Recognition, Minneapolis, MN, USA, 17–22 June 2007; pp. 1–8. [Google Scholar]
  32. Wang, Z.; Bovik, A.C. A Universal Image Quality Index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
Figure 1. Pulsed LiDAR imaging system based on SPAD.
Figure 1. Pulsed LiDAR imaging system based on SPAD.
Photonics 10 00744 g001
Figure 2. (a) Generator general architecture. (b) Non-local sparse attention Residual Channel-wise soft thresholding (NRC) module. (c) RCST unit.
Figure 2. (a) Generator general architecture. (b) Non-local sparse attention Residual Channel-wise soft thresholding (NRC) module. (c) RCST unit.
Photonics 10 00744 g002
Figure 3. U-Net-based discriminator. Brighter colors in the decoder output correspond to lower confidence (discriminator considers fake).
Figure 3. U-Net-based discriminator. Brighter colors in the decoder output correspond to lower confidence (discriminator considers fake).
Photonics 10 00744 g003
Figure 4. Comparison of different methods for HR range image reconstruction in Art scene. The image resolution is 256 × 256 and the noise environment is 2:50. The “Target” represents the ground truth range image provided in the dataset.
Figure 4. Comparison of different methods for HR range image reconstruction in Art scene. The image resolution is 256 × 256 and the noise environment is 2:50. The “Target” represents the ground truth range image provided in the dataset.
Photonics 10 00744 g004
Figure 5. Three scenes comparing different methods of HR range reconstruction image. The image resolution is 256 × 256. The noise environment is 5:10 (Laundry), 2:10 (Moebius) and 5:50 (Dolls). The “Target” represents the ground truth range image provided in the dataset. The content enclosed in the square highlights significant differences observed among the methods.
Figure 5. Three scenes comparing different methods of HR range reconstruction image. The image resolution is 256 × 256. The noise environment is 5:10 (Laundry), 2:10 (Moebius) and 5:50 (Dolls). The “Target” represents the ground truth range image provided in the dataset. The content enclosed in the square highlights significant differences observed among the methods.
Photonics 10 00744 g005
Figure 6. Real-world experiment target scenes and reconstruction results. The resolution of the range image by MLE method is 64×64. The resolution of the other images is 128×128. The last column of images was reconstructed by TSDSR. The content enclosed in the square highlights significant differences observed among the methods. (a) Near-field experimental target scene. (b) Far-field experimental target scene. (c) Reconstruction results of near-field experimental detection data by various methods.
Figure 6. Real-world experiment target scenes and reconstruction results. The resolution of the range image by MLE method is 64×64. The resolution of the other images is 128×128. The last column of images was reconstructed by TSDSR. The content enclosed in the square highlights significant differences observed among the methods. (a) Near-field experimental target scene. (b) Far-field experimental target scene. (c) Reconstruction results of near-field experimental detection data by various methods.
Photonics 10 00744 g006
Figure 7. Art and Laundry scenes comparing different methods of HR range reconstruction image. The image resolution is 256 × 256 and the noise environment is 2:50. The “Target” represents the ground truth range image provided in the dataset. The content enclosed in the square highlights significant differences observed among the methods.
Figure 7. Art and Laundry scenes comparing different methods of HR range reconstruction image. The image resolution is 256 × 256 and the noise environment is 2:50. The “Target” represents the ground truth range image provided in the dataset. The content enclosed in the square highlights significant differences observed among the methods.
Photonics 10 00744 g007
Table 1. Under 9 different SBRs conditions, the average performance of various reconstruction methods was quantitatively evaluated across 8 different test scenes. A lower value of RMSE (Root Mean Squared Error) indicates a closer resemblance of the reconstructed image to the ground truth image. Higher values of accuracy, PSNR, and UIQI indicate a closer resemblance of the reconstructed image to the ground truth image. Optimal performance is represented by bold characters. The abbreviation “NSE” refers to the NLSA-Encoder method.
Table 1. Under 9 different SBRs conditions, the average performance of various reconstruction methods was quantitatively evaluated across 8 different test scenes. A lower value of RMSE (Root Mean Squared Error) indicates a closer resemblance of the reconstructed image to the ground truth image. Higher values of accuracy, PSNR, and UIQI indicate a closer resemblance of the reconstructed image to the ground truth image. Optimal performance is represented by bold characters. The abbreviation “NSE” refers to the NLSA-Encoder method.
SBRRMSEAccuracy (δ = 1.03)
PengZhaoNSEProposedPengZhaoNSEProposed
10:20.02050.01740.02110.01890.97630.97860.97980.9788
10:100.02040.01770.02110.01910.97590.97850.97960.9789
10:500.02070.01980.02130.01940.97590.97790.97910.9784
5:20.02180.01890.02140.02010.97430.97700.97810.9783
5:100.02200.02120.02170.02050.97400.97650.97700.9783
5:500.02330.02780.02290.02170.97230.96980.97470.9771
2:20.02450.02320.02340.02240.96940.97200.97390.9756
2:100.02620.03550.02480.02410.96600.96720.97000.9728
2:500.03180.03640.03080.03040.95340.96240.95380.9592
AVG0.02350.02420.02320.02180.97080.97330.97400.9753
SBRPSNRUIQI
PengZhaoNSEProposedPengZhaoNSEProposed
10:260.72962.18861.00961.1920.97910.98250.98110.9823
10:1060.69162.11960.83761.1940.97880.98350.98130.9831
10:5060.31658.08260.64460.7570.97400.98030.98040.9809
5:259.87460.15960.26560.9030.97580.97990.97920.9813
5:1059.78057.82459.97360.6610.97520.97810.97900.9805
5:5059.00457.57659.33659.9260.97170.97050.97600.9786
2:258.21958.75358.93359.6160.96920.97270.97520.9761
2:1057.46555.24958.06258.3880.96490.95630.96960.9717
2:5055.09354.57455.41555.2640.94920.95080.95230.9524
AVG59.01958.50359.38659.7670.97090.97270.97490.9763
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tong, Z.; Jiang, X.; Hu, J.; Xu, L.; Wu, L.; Yang, X.; Zou, B. TSDSR: Temporal–Spatial Domain Denoise Super-Resolution Photon-Efficient 3D Reconstruction by Deep Learning. Photonics 2023, 10, 744. https://doi.org/10.3390/photonics10070744

AMA Style

Tong Z, Jiang X, Hu J, Xu L, Wu L, Yang X, Zou B. TSDSR: Temporal–Spatial Domain Denoise Super-Resolution Photon-Efficient 3D Reconstruction by Deep Learning. Photonics. 2023; 10(7):744. https://doi.org/10.3390/photonics10070744

Chicago/Turabian Style

Tong, Ziyi, Xinding Jiang, Jiemin Hu, Lu Xu, Long Wu, Xu Yang, and Bo Zou. 2023. "TSDSR: Temporal–Spatial Domain Denoise Super-Resolution Photon-Efficient 3D Reconstruction by Deep Learning" Photonics 10, no. 7: 744. https://doi.org/10.3390/photonics10070744

APA Style

Tong, Z., Jiang, X., Hu, J., Xu, L., Wu, L., Yang, X., & Zou, B. (2023). TSDSR: Temporal–Spatial Domain Denoise Super-Resolution Photon-Efficient 3D Reconstruction by Deep Learning. Photonics, 10(7), 744. https://doi.org/10.3390/photonics10070744

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop