Next Article in Journal
VEP Score of a Left Eye Had Predictive Values for Achieving NEDA-3 over Ten Years in Patients with Multiple Sclerosis
Next Article in Special Issue
Virtual Light Sensing Technology for Fast Calculation of Daylight Autonomy Metrics
Previous Article in Journal
Supervised Object-Specific Distance Estimation from Monocular Images for Autonomous Driving
Previous Article in Special Issue
Anisotropic SpiralNet for 3D Shape Completion and Denoising
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

Adaptive High-Resolution Imaging Method Based on Compressive Sensing

1
School of Mechanical Engineering, Shijiazhuang Tiedao University, Shijiazhuang 050000, China
2
School of Engineering, Hong Kong University, Hong Kong, China
3
School of Artificial Intelligence and Big Data, Hebei Polytechnic Institute, Shijiazhuang 050000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(22), 8848; https://doi.org/10.3390/s22228848
Submission received: 29 September 2022 / Revised: 30 October 2022 / Accepted: 12 November 2022 / Published: 16 November 2022
(This article belongs to the Collection Computational Imaging and Sensing)

Abstract

:
Compressive sensing (CS) is a signal sampling theory that originated about 16 years ago. It replaces expensive and complex receiving devices with well-designed signal recovery algorithms, thus simplifying the imaging system. Based on the application of CS theory, a single-pixel camera with an array-detection imaging system is established for high-pixel detection. Each detector of the detector array is coupled with a bundle of fibers formed by fusion of four bundles of fibers of different lengths, so that the target area corresponding to one detector is split into four groups of target information arriving at different times. By comparing the total amount of information received by the detector with the threshold set in advance, it can be determined whether the four groups of information are calculated separately. The simulation results show that this new system can not only reduce the number of measurements required to reconstruct high quality images but can also handle situations wherever the target may appear in the field of view without necessitating an increase in the number of detectors.

1. Introduction

As a type of signal sampling theory, CS theory overcomes the shortcomings of traditional Nyquist theorem sampling, namely a large amount of data and a large waste of storage space and time. Researchers proposed that if the original signal is sparse or compressible, a precise reconstruction can be recovered from a small number of random measurements by solving linear equations [1,2,3,4]. Since the non-adaptive linear measurement results of a small number of compressible signals or images contain enough information for reconstruction processing, CS theory can be used to directly obtain the compressed signal representation without first sampling the signal [4,5,6,7]. CS theory transfers the burden of sampling to data processing, it shifts the focus of the imaging system from the traditional design of expensive receiver hardware to the novel design of signal recovery algorithms [5,7]. At present, CS technology has been widely used in three areas—wireless communication, array signal processing, and imaging systems—and plays a major role in fault diagnosis and signal recognition [8,9]. A single-pixel camera is an important application of compressed sensing theory. It uses a digital micromirror device (DMD) to perform optical calculations on the linear projection of images on the pseudo-random binary pattern [10]. Single-pixel camera technology has been widely used in spectral imaging, radar imaging, and medical imaging [11,12]. Similarly, it also supports a variety of applications for three-dimensional (3-D) imaging, such as multispectral imaging [13], depth estimation of imaging scenes [14], and video rate improvement [15].
In previous research [16], the sparsity problem was identified, which has a great impact on array detection imaging systems based on CS theory; consequently, a method using irregularly arranged fibers was proposed to solve this problem. However, this method should be applied under a certain condition, i.e., the target should always be in the middle of the field of view. The reconstruction of an image using fibers with fewer data is a hot research topic; one method is proposed by Yu et al. [17]. However, our method is based on CS theory rather than machine learning algorithm, and uses fewer datasets. Another related method is to conjunct the photon-counting imaging system with a fiber optic taper to extend the field of view of the images [18]. Therefore, in this article, taking advantage of the flexibility of the fiber array arrangement, the threshold judgment method is used to determine the position of the target, which facilitates the rapid completion of the pixels imaging process without increasing the number of detectors.

2. Theory

2.1. CS and Single-Pixel Camera

The CS theory enables researchers to stably reconstruct the image with fewer measurement results than the number of reconstructed pixels [7]. It comprises three main parts: sparse representation of signal, measurement matrix and reconstruction algorithm.
Figure 1 shows the principle of CS. If the N-dimensional signal x is K-sparse on the Ψ transform domain, it will be processed in the Φ domain. The measurement matrix y can be obtained after M linear measurements by:
y = Φ Ψ α = Θ α
where Θ is called the effective observation matrix. An accurate or approximate solution can be obtained by the optimal norm using a small number of measured values y , so that x = Ψ α has a higher probability of being reconstructed accurately. The process of signal reconstruction faces a typical complexity non-deterministic polynomial (NP) problem. According to CS theory, if the condition of restricted isometry property (RIP) is met [3], it can be converted to the L 1 norm to reconstruct x, and can be written as:
α * = arg min α α 1 s . t .   y Θ α 2 ε
where ε is the maximum noise energy limit during reconstruction.
The single-pixel camera is an optical computer that sequentially measures the inner products y [ m ] = x , ϕ m between an N-pixel sampled version x of the incident light-field from the scene under view and a set of two-dimensional (2-D) test functions { ϕ m } [9]. The structure of the single-pixel CS camera is shown in Figure 2.
The light-field is focused by biconvex Lens 1 onto a DMD consisting of an array of N small mirrors. Each mirror corresponds to a particular pixel in x , and ϕ m can independently rotate in the direction controlled by the codes. The reflected light is then collected by biconvex Lens 2 and focused onto a single photon detector that integrates the product x [ n ] ϕ m [ n ] to compute the measurement y [ m ] = x , ϕ m as its output voltage and digitized by the A/D converter. Values of ϕ m can be obtained by dithering the mirrors back and forth during the photodiode integration time [7].

2.2. Problem

Baraniuk et al. proposed that if the relationship between the number of random measurements M and the K-sparse vectors satisfies
M O ( K log ( N / K ) )
these compressible vectors can be exactly reconstructed and approximated stably with high probability [19]. That is to say, at least M measurements are required to reconstruct a K-sparse signal with high quality. Among them, sparsity can be simply understood as the number of non-zero elements in the signal information, and values close to zero are sometimes approximated to zero. Large coefficients generally contain signal-related information, while sparse signals are small coefficients (such as background) that contain little information.
For example, a single-pixel CS camera is used to image a target and reconstruct an image of 256 × 256 pixels (65,536 pixels), in which background information occupies more than 70%, and a high-quality reconstructed image that meets the recognition needs may require approximately 1700 random measurements. This reflects the advantage of CS theory; that is, compared with 256 × 256 detector array imaging, it not only reduces the detection cost but also shortens the detection time compared with the traditional single-pixel detection. However, in actual detection, more and more complex scenes or fine targets need to be captured, sampled, and processed faster in 3D imaging systems, with lower power consumption and larger detection pixels. Therefore, a compromise solution was proposed in previous studies: an array detector with a small number of pixels and a DMD array are used to image the target. Each detector can be regarded as a single-pixel camera, using compressed sensing theory to reconstruct the corresponding sub-image block to achieve the purpose of reducing the reconstruction time through parallel measurement and calculation. In the array detection imaging system, assuming that a 4 × 4 pixel array detector is used to detect the target to reconstruct a high-quality image of 256 × 256 pixels, the sub-imaging block corresponding to each detector is 64 × 64 pixels (4096 pixels). This method of image cutting results in several sub-imaging blocks that may be full of target information (close to 0-sparse), and the number of measurements calculated by Equation (3) is close to 4096 pixels, which removes the advantage of the CS technology in the array imaging system. Simultaneously, there may be several sub-imaging blocks that are almost full of background information, resulting in redundant measurement work. In previous studies, assuming that the target is always in the middle of the field of view and the background information is distributed around, it is proposed that the number of pixels of the image block corresponding to the target and the background information can be redistributed using an optical fiber arrangement, thereby shortening the high-quality reconstruction time for low-sparse sub-image blocks, and the redundant calculations of high-sparse sub-image blocks reconstruction are reduced [16]. However, the target may appear anywhere in the actual field of view, as shown in Figure 3. In this case, the previous optical fiber arrangement method cannot be used.

2.3. Proposed System

For this situation where the target position is not fixed, the structure of an adaptive imaging system is proposed, as shown in Figure 4 and Figure 5.
The pulsed beam is emitted by the laser and reflected from the target. The reflected signal reaches the DMD array board after passing through lens 1, and then is focused on the optical fiber array by lens 2. The DMD array board is composed of DMD boards arranged in 6 × 6, and each DMD board is composed of small mirrors arranged in 100 × 100 (that is, the final reconstructed target image is 600 × 600 pixels). At the input of the optical fiber array, 36 groups of optical fiber bundles are closely arranged in a 6 × 6 arrangement. Each group of optical fiber bundles consists of four optical fibers, and the lengths of these four optical fibers follow an arithmetic sequence. The tolerance ∆L can ensure that the time length of the signals returning from the same sub-region of the target do not overlap. The output of the four optical fibers is fused into one optical fiber and connected to a detector in the detector array, and the 36 detectors in the detector array are also arranged in a 6 × 6 arrangement. It is equivalent to each detector receiving the signal reflected by its corresponding one DMD board. The computer first judges the amount of signal received according to a preset threshold. If the total signal amount of one 100 × 100 pixels sub-image area exceeds the threshold within a certain period, the signal of the four time periods of this sub-image area is respectively measured and processed in parallel using CS theory. The reconstructed four 50 × 50 pixel images are finally spliced into a 100 × 100 pixel reconstructed image. On the contrary, when the total amount of signal is less than the threshold, the information of this sub-image area does not need to be processed separately, but can be directly used to reconstruct a 100 × 100 pixel image. Finally, all the reconstructed images are spliced together in the correct position to obtain a reconstructed image of 600 × 600 pixels.
It can be concluded from the previous analysis that when more information of the target is processed, longer measurement time is required for high-quality reconstruction. Therefore, the threshold is used to determine the location of the target, the sub-image blocks with target information greater than the threshold are further divided into four parts, and the measurement time is reduced through parallel calculation to achieve the purpose of shortening the overall detection and imaging time.

3. Comparative Experiment

In order to better demonstrate the advantages of the proposed system, a simulation system is built, and a series of simulation experiments are performed. Simulation parameters mainly include laser radar (LADAR) ranging equation, fiber loss, detector conversion efficiency, etc. The LADAR range equation describes the relevant parameters of the radar system, target characteristics, receiving parameters, and atmospheric transmission [20,21]. In the following simulation experiments, it is assumed: that the laser transmission in the atmosphere conforms to the geometrical optics principle; the atmosphere is uniform and isotropic; the target object belongs to the Lambert radiator; and the laser energy distribution on the surface of the target object is uniform [18]. The equation to compute signal power at the detector P det that incorporates these efficiency terms is expressed by:
P det = τ 0 τ a D R 2 ρ t ( d A ) P t R 2 θ R ( θ t R ) 2
where P t is the laser transmitted power, θ t is the laser transmitter beam diameter and angular divergence, R is the range between target and detector, d A is the effective target area, ρ t is the target surface reflectivity, θ R is the target surface angular dispersion, τ 0 is the transmission of the optics, τ a is the atmospheric transmission, and D R is the area of the circular receiver aperture with diameter [18].
The parameters mentioned above are set to appropriate values, as shown in Table 1.
Figure 6a is the original array detection imaging method. For the 6 × 6 array-detection, each detector corresponds to a detection area of 100 × 100 pixels. The 36 sub-image blocks are reconstructed separately and then spliced together to obtain a 600 × 600 pixel reconstruction image.
Figure 6b is the method to obtain the reconstructed image based on the threshold. It can be seen that in the original 100 × 100 pixel block, six received target signals are higher than the threshold, and they are divided into four 50 × 50 pixel image blocks and reconstructed respectively. The remaining 100 × 100 pixel image blocks do not need to be subdivided, and all the signals received by their corresponding detectors are directly used to reconstruct the target image of the corresponding 100 × 100 pixel area. Finally, these 54 image blocks are spliced together to obtain a reconstructed image of 600 × 600 pixels.
It should be noted that the value of the length difference Δ L of the fiber bundle is the key to whether the reflected signal of the target area with low sparsity can be subdivided correctly. Δ L should ensures that the information of the four sub-imaging blocks does not interfere with each other. If Δ L is too small, the information in different areas will overlap and the reconstruction will fail. If Δ L is too large, it will waste fiber cost and increase detection time. Therefore, Δ L must satisfy:
Δ L > c × ( Δ S / c )
where c is the speed of light, and Δ S is the depth of the target. It can be obtained that the value of Δ S should be greater than Δ S . Information for the next sub-imaging block will not reach the detector until the detector receives all the information of the previous sub-imaging block.
The quality of the reconstructed image is characterized by the difference between the reconstructed image and the original image. The higher the number of measurements, the smaller the obtained variance value, and the better the quality of the reconstructed image. Through multiple simulation and recognition, we set a variance value, which corresponds to a high-quality reconstructed image that meets the requirements of recognition. We also set the threshold of segmented pixel detection for the detector of the improved system. Figure 7a shows the intensity image of the target. Three groups of simulation experiments were organized, ten times for each group. The original and proposed system models were used to reconstruct 600 × 600 pixel image, and to calculate the average number of measurements required to obtain high-quality reconstruction results set in advance, as shown in Figure 7b.
In a parallel measurement system, the total measurement time depends on the image block with the maximum number of measurements required for reconstruction. After the reconstruction effect is set, more measurements are required, which means that the corresponding image block is more important for the overall target recognition, and may even affect the overall recognition result. Therefore, the time consumption of the two systems can be compared by comparing the peak value of the relationship curve in Figure 7b. It can be concluded that, compared with the original method, the peak value corresponding to the proposed method is significantly reduced. The extra time cost caused by the fiber array and threshold judgment only accounts for a small proportion, and becomes less as the demand for imaging pixels increases.

4. Threshold Selection

After verifying the advantages of the proposed system, it is considered whether the measurement time can be further shortened by adjusting the threshold. The target is placed in two positions, and three thresholds are used to reconstruct the same level of high-quality images. Figure 8a–c correspond to the first position of the target, and Figure 8d–f correspond to the second position of the target. The thresholds used in Figure 8a,d are lower than those in Figure 8b,e, and higher than those in Figure 8c,f. It can be seen from the blue and yellow blocks in Figure 8 that different thresholds result in different target area subdivision schemes. Due to the further subdivision of image blocks, the abscissa of Figure 9 corresponds to the number of sub-image blocks in Figure 8a–f, as 54, 45, 60, 51, 42, and 54, respectively.
By comparing the peaks of the curves in Figure 9a,b, it can be obtained that the number of measurements required to reconstruct a high-quality image with a larger threshold (Threshold 2) is much higher than the others. The difference between the number of measurements required for Threshold 1 and Threshold 3 is relatively small, between which the number of measurements of Threshold 1 is slightly more. However, the lower the threshold selected, the more image blocks are subdivided.

5. Conclusions

The single-pixel camera is able to efficiently and scalably handle high-dimensional data sets from hyperspectral imaging. Target detection and recognition have increasingly higher requirements for imaging pixels. Using a single-pixel camera to build an array-detection imaging system can greatly reduce hardware burden; however, the time cost of reconstructing images will also increase. Therefore, an improved adaptive method is proposed, which uses a flexible optical fiber array structure to connect to the detector array. The computer performs a threshold judgment on the total signals received by each detector in the detector array, and further divides the sub-image blocks that exceed the threshold (low-sparse); then, images are reconstructed separately, so the aim of shortening the reconstruction time can be achieved by parallel measurement and calculation. The comparative experiment proves the effect of the system for shortening the measurement time, and another simulation result provides suggestions for the selection of the threshold.

Author Contributions

Conceptualization, Z.W. and Y.G.; methodology, Z.W. and Y.G.; software, Z.W.; validation, X.D. and J.C.; formal analysis, Z.W.; investigation, Y.G.; resources, J.C.; data curation, Z.W.; writing—original draft preparation, Z.W.; writing—review and editing, Y.G.; visualization, J.C.; supervision, J.C.; project administration, Z.W.; funding acquisition, J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Donoho, D.L. Compressed sensing. IEEE Trans. Inform. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  2. Candès, E.J.; Tao, T. Near optimal signal recovery from random projections: Universal encoding strategies? IEEE Trans. Inform. Theory 2006, 52, 5406–5425. [Google Scholar] [CrossRef] [Green Version]
  3. Candès, E.J. Compressive sampling. Proc. Int. Cong. Math. 2006, 3, 1433–1452. [Google Scholar]
  4. Baraniuk, R.G. Compressive sensing. IEEE Signal Proc. Mag. 2007, 24, 118–121. [Google Scholar] [CrossRef]
  5. Candès, E.J.; Wakin, M.B. An introduction to compressive sampling. IEEE Signal Proc. Mag. 2008, 25, 21–30. [Google Scholar] [CrossRef]
  6. Willett, R.M.; Marcia, R.F.; Nichols, J.M. Compressed sensing for practical optical imaging systems: A tutorial. Opt. Eng. 2011, 50, 07260. [Google Scholar]
  7. MDuarte, F.; Davenport, M.A.; Takbar, D.; Laska, J.N. Single-pixel imaging via compressive sampling. IEEE Signal Proc. Mag. 2008, 25, 83–91. [Google Scholar] [CrossRef] [Green Version]
  8. Lahbib, N.D.; Cherif, M.; Hizem, M.; Bouallegue, R. Channel Estimation for TDD Uplink Massive MIMO Systems Via Compressed Sensing. In Proceedings of the IEEE International Wireless Communications and Mobile Computing Conference, Tangier, Morocco, 24–28 June 2019; pp. 1680–1684. [Google Scholar]
  9. Ma, Y.F.; Jia, X.S.; Bai, H.J.; Wang, G.L.; Liu, G.Z.; Guo, C.M. A new fault diagnosis method using deep belief network and compressive sensing. J. Vibro Eng. 2020, 22, 83–97. [Google Scholar] [CrossRef] [Green Version]
  10. Takhar, D.; Laska, J.N.; Wakin, M.B.; Duarte, M.F.; Baron, D.; Sarvotham, S.; Kelly, K.F.; Baraniuk, R.G. A new compressive imaging camera architecture using optical-domain compression. Proc. Comput. Imaging IV 2006, 6065, 43–52. [Google Scholar]
  11. Lee, S.W.; Jung, Y.H.; Lee, M.J.; Lee, W.Y. Compressive Sensing-Based SAR Image Reconstruction from Sparse Radar Sensor Data Acquisition in Automotive FMCW Radar System. Sensors 2021, 21, 7283. [Google Scholar] [CrossRef] [PubMed]
  12. Graff, C.G.; Sidky, E.Y. Compressive Sensing in Medical Imaging. Appl. Opt. 2015, 54, C23–C44. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Ke, J.; Alieva, T.; Oktem, F.S.; Silveira, P.E.; Wetzstein, G.; Willomitzer, F. Computational Optical Sensing and Imaging 2021: Introduction to the feature issue. Appl. Opt. 2022, 61, COSI1–COSI4. [Google Scholar] [CrossRef]
  14. Ye, J.T.; Huang, X.; Li, Z.P.; Xu, F. Compressed sensing for active non-line-of-sight imaging. Opt. Exp. 2021, 29, 1749–1763. [Google Scholar] [CrossRef]
  15. Kaiguo, X.; Zhisong, P.; Pengqiang, M. Video Compressive sensing reconstruction using unfolded LSTM. Sensors 2022, 22, 7172. [Google Scholar] [CrossRef] [PubMed]
  16. Cao, J.Y.; Han, S.K.; Liu, F.; Zhai, Y.; Xia, W.Z. Research on the high pixels ladar imaging system based on compressive sensing. Opt. Eng. 2019, 58, 013103. [Google Scholar] [CrossRef]
  17. Yu, Z.; Ju, Z.; Zhang, X.; Meng, Z.; Yin, F.; Xu, K. High-speed multimode fiber imaging system based on conditional generative adversarial network. Chin. Opt. Lett. 2021, 19, 081101. [Google Scholar] [CrossRef]
  18. Zheng, T.X.; Shen, G.Y.; Li, Z.H.; Wu, E.; Chen, X.L.; Wu, G. Single-photon imaging system with a fiber optic taper. Optoelectron. Lett. 2018, 14, 267–270. [Google Scholar] [CrossRef]
  19. Baraniuk, R.; Steeghs, P. Compressive Radar Imaging. In Proceedings of the 2007 IEEE Radar Conference, Waltham, MA, USA, 17–20 April 2007; pp. 128–133. [Google Scholar]
  20. Richard, D.R.; Stephen, C.C. Direct-Detection LADAR Systems; SPIE Press: Bellingham, WA, USA, 2009; pp. 15–25. [Google Scholar]
  21. Xia, W.Z.; Han, S.K.; Ullah, N.; Cao, J.Y.; Wang, L.; Cao, J.; Cheng, Y.; Yu, H.Y. Design and modeling of three-dimensional laser imaging system based on streak tube. Appl. Opt. 2017, 56, 487–497. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Linear measurement of CS [4].
Figure 1. Linear measurement of CS [4].
Sensors 22 08848 g001
Figure 2. Schematic diagram of single-pixel camera.
Figure 2. Schematic diagram of single-pixel camera.
Sensors 22 08848 g002
Figure 3. Possible target positions within the field of view.
Figure 3. Possible target positions within the field of view.
Sensors 22 08848 g003
Figure 4. Schematic diagram of the proposed imaging system.
Figure 4. Schematic diagram of the proposed imaging system.
Sensors 22 08848 g004
Figure 5. Connection structure diagram of fiber array and detector array.
Figure 5. Connection structure diagram of fiber array and detector array.
Sensors 22 08848 g005
Figure 6. Comparison of original (a) and proposed (b) methods.
Figure 6. Comparison of original (a) and proposed (b) methods.
Sensors 22 08848 g006
Figure 7. The intensity image of the target and the result of the comparison experiment.
Figure 7. The intensity image of the target and the result of the comparison experiment.
Sensors 22 08848 g007
Figure 8. CS measurement scheme diagram with two target positions and three thresholds.
Figure 8. CS measurement scheme diagram with two target positions and three thresholds.
Sensors 22 08848 g008
Figure 9. Comparison curve of the number of measurements.
Figure 9. Comparison curve of the number of measurements.
Sensors 22 08848 g009
Table 1. System simulation parameters.
Table 1. System simulation parameters.
ParameterValue
laser power (J)1
fiber diameter (μm)125
atmospheric transmission1
laser dispersion angular0.012
detector dark current (A)10−9
CCD pixels800 × 800
optics transmission1
detector quantum efficiency0.75
width of laser pulse (ns)6
background power (w/m2)100
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, Z.; Gao, Y.; Duan, X.; Cao, J. Adaptive High-Resolution Imaging Method Based on Compressive Sensing. Sensors 2022, 22, 8848. https://doi.org/10.3390/s22228848

AMA Style

Wang Z, Gao Y, Duan X, Cao J. Adaptive High-Resolution Imaging Method Based on Compressive Sensing. Sensors. 2022; 22(22):8848. https://doi.org/10.3390/s22228848

Chicago/Turabian Style

Wang, Zijiao, Yufeng Gao, Xiusheng Duan, and Jingya Cao. 2022. "Adaptive High-Resolution Imaging Method Based on Compressive Sensing" Sensors 22, no. 22: 8848. https://doi.org/10.3390/s22228848

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop