Next Article in Journal
Enhancing Contrast of Spatial Details in X-ray Phase-Contrast Imaging through Modified Fourier Filtering
Next Article in Special Issue
Research on Ground Object Echo Simulation of Avian Lidar
Previous Article in Journal
Tuning the Photonic Spectrum of Superlattice Structures with Magnetic Fields: An Anisotropic Perspective
Previous Article in Special Issue
Comparative Analysis of the Effect of Cutting Piezoelectric Ceramics on Optically Oriented Compensation Capability
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Super-Resolution Imaging Enhancement through a 2D Scanning Galvanometer: Algorithm Formulation and Application in Aerial Optoelectronic Systems

1
Changchun University of Science and Technology, Changchun 130000, China
2
Key Laboratory of Airborne Optical Imaging and Measurement, Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences, Changchun 130033, China
3
University of Chinese Academy of Sciences, Beijing 100039, China
*
Author to whom correspondence should be addressed.
Photonics 2023, 10(11), 1203; https://doi.org/10.3390/photonics10111203
Submission received: 13 September 2023 / Revised: 20 October 2023 / Accepted: 24 October 2023 / Published: 27 October 2023
(This article belongs to the Special Issue Advances in Photoelectric Tracking Systems)

Abstract

:
As the fields of aviation and aerospace optics continue to evolve, there is an increasing demand for enhanced detection capabilities in equipment. Nonetheless, in applications where both optical and mechanical constraints are stringent, the continuous expansion of optical aperture and focal length is impractical. Given the existing technological landscape, employing super-resolution algorithms to enhance the imaging capability of optical systems is both practical and highly relevant. This study capitalizes on using a 2D scanning galvanometer in optical systems to acquire micro-displacement information. Initially, an imaging model for optical systems equipped with a 2D scanning galvanometer was established, and the displacement vectors for both forward and sweep image motions were defined. On this foundation, we incorporated micro-displacement information that can induce high-frequency aliasing. Subsequently, the motion paths of the galvanometer were planned and modeled. To align image sequences with micro-displacement correlations, the Lucas–Kanade (L-K) optical flow method was employed with multi-layer pyramid iteration. Then, super-resolution reconstruction was performed using kernel regression techniques. Ultimately, we tested the algorithm on an aeronautical optoelectronic pod to evaluate its impact on optical resolution and imaging quality. Compared with the original images, the 16-frame image demonstrated a 39% improvement in optical resolution under laboratory conditions. Moreover, the algorithm exhibited satisfactory performance under both nighttime and daytime conditions, as well as during aerial tests.

1. Introduction

As the aviation and aerospace photoelectric payloads’ imaging distance extends and the resolution increases, challenges arise. Under constraints involving volume, weight, and other factors, the aperture and focal length of the payload’s optical system cannot be increased infinitely. In electronics, high-resolution images correspond to a higher pixel count. The detector must arrange a denser pixel distribution per unit area, imposing very stringent demands on the design and manufacture of the detector, and the transmission capability of the data link [1]. Research focusing on obtaining higher-resolution images at the current level of optoelectronic technology has intensified [2,3,4,5,6].
In the field of aerospace and aviation, super-resolution imaging typically involves the photoelectric load acquiring low-resolution image information of a single frame or a sequence of multiple frames for a scene. Algorithms are then employed to reconstruct high-resolution images. For image detail accuracy, multi-frame super-resolution algorithms have found widespread use [7,8,9].
Multi-frame super-resolution imaging encompasses two vital steps. The first involves the procurement of low-resolution images containing high-frequency aliasing information. Primary technical paths include micro-scanning and sub-pixel stitching. The former employs a micro-motion mechanism, moving the optical-system-formed image on the detector at a 1/N pixel pitch, to capture several low-resolution images containing high-frequency aliasing information.
A rapid micro-scanning imaging device was developed by Fortin and Chevrette [10]. Comprising three main sections—the micro-scanning head, the controller, and the power amplifier—the system enables four operation modes. These modes are fixed position, 2 × 2, 3 × 3, and 4 × 4 micro-scan. In the case of a given staring array imager, the resolution is doubled.
Utilization of a piezoelectric nanopositioner for micro-displacements was demonstrated by Wiltse and Miller [11]. This approach provides substantial enhancements in minimum resolvable temperature/minimum resolvable contrast curves, allowing resolution beyond the sensor’s native Nyquist frequency. Increases in target range discrimination of approximately 18% are attributed to four-point micro-scanning.
An infrared micro-scanning lens was designed by Zhang et al. [12] for constructing an infrared micro-scanning optical system. The entire field of view (FOV)’s displacement uniformity and ideal image quality on the image plane are ensured by this system. Optimal results were achieved with 2 × 2 magnification.
However, the micro-scanning method, requiring a micro-motion mechanism for aliased image acquisition, poses high demands on volume, power consumption, processing, and displacement accuracy. Therefore, its application range may be subject to some limitations.
Sub-pixel stitching technology, another approach, employs sub-pixel-level detector stitching to enhance the imaging system’s resolution. Initially applied on the French SPOT-5 satellite [13], this technique enabled an imaging resolution of up to 2.5 m. The method entails placing two linear array detectors in dislocation (0.5-pixel displacement) on the optical system’s image plane. With the optical system’s resolution designed as 5 m, an increase to 2.5 m is achieved. The Swiss Leica Company and the German Aerospace Center’s joint development of the ADS40 aerial surveying and mapping camera also utilized detector splicing to achieve super-resolution in the imaging system [14].
Tang et al. [15] introduced a mosaic technique that is grounded in the geometric structure of TDI CCD sensors and the object space projection plane. The sensors are projected into the object space indirectly, along and across the target trajectory’s direction. This method has been applied in the daily processing systems of ZY-1 02C and ZY-3 satellites, yielding positive results. However, the application of sub-pixel stitching is primarily limited to line detector systems.
The second step in the process involves a high-resolution image reconstruction algorithm that uses a multi-frame image sequence. Typically, multi-frame reconstruction algorithms are categorized into the frequency domain method [16,17], non-uniform interpolation method [18,19], iterative back-projection method [20], convex set projection method [21,22], and probability analysis method [23,24]. Although these methods can achieve enhanced super-resolution results, high-performance algorithms often require significant computing resources. Consequently, they are unsuitable for aerospace and aviation platforms with high real-time demands and limited computing capabilities.
In this research, a 2D scanning galvanometer is planned to be utilized for aliasing information acquisition, along with a non-radial kernel function to merge multi-frame images and achieve super-resolution images. Initially, the mathematical relationship among the swing angle of the two-dimensional mirror, the image point’s displacement on the image plane, and the exposure time is defined. Then, constraints are given for the super-resolution imaging of the photoelectric system with two-dimensional swinging. A super-resolution algorithm processing flow, based on prior displacement information, is then established. Following this, low-resolution image information is registered, fused, and reconstructed, and the multi-frame sequence depth’s influence on the imaging effect is examined. Finally, the algorithm is tested in the laboratory, daytime, nighttime, and real-application scenarios. Experimental evidence indicates that the algorithm enhances image resolution and performs well in the aerial camera when applied.

2. Acquisition of Micro-Displacement

The degradation in aerial camera imaging quality is mainly attributed to image motion. To address this issue, commonly used methods encompass optical, mechanical, electronic, and image compensation techniques. Yet, most optical systems presently employ a 2D scanning galvanometer for image motion compensation.
In this study, the 2D scanning galvanometer is leveraged to carry out image motion compensation, imposing sub-pixel micro-displacement on the sensor to capture multiple frames of low-resolution image sequences with aliasing information. First, an imaging model of the optical system incorporating a 2D scanning galvanometer is created, and the mathematical relationships among the 2D scanning galvanometer oscillating vector, target moving vector, and micro-displacement vector are defined. Subsequently, the 2D scanning galvanometer’s usage in the system is detailed, and its trajectory model is established.

2.1. The Imaging System Model

The active changes in the aerial camera’s optical axis are primarily associated with two factors: alterations in aircraft attitude and the active swinging of the camera movement mechanism. The aircraft’s ground-target relative moving speed is translated into the image-plane coordinate system with the aircraft track coordinate system, body coordinate system, fixed mirror, optical system, and two-dimensional pendulum mirror. Based on the light transmission relationship, a mathematical model is established between the target moving speed and the moving speed of the point on the image plane. The fundamental optical structure of an aerial camera is illustrated in Figure 1.
The transformation in the homogeneous coordinate change involves multiple coordinate systems, namely, the ground coordinate system, the fixed-mirror coordinate system, the 2D scanning galvanometer coordinate system, and the image-plane coordinate system. The track angle of attack, sideslip angle, and roll angle are represented by α, β, and γ, respectively. Likewise, the camera’s sweep angle, depression angle, and push-broom angle are designated by φ, ϕ , and η, respectively. The transformation of the speed of the image point in the image-plane coordinate system is accomplished through a sequence from the track coordinate system, via the aircraft coordinate system, to the image-plane coordinate system.
In Figure 1, the camera’s sweep angle is assumed to be θ, with the focal length being represented by f; the flight height, by H; and the track speed, by ν. The coordinate system’s center point is defined as the intersection point between the central axis of the lens and the fixed mirror. Under the track coordinate system, the sideslip angle’s impact on the image motion vector on the image plane is formulated as
ν β = T 1 K T 2 cos β sin β 0 sin β cos β 0 0 0 1 ν 0 0 .
where T1 and T2 are the vector forms of the reflection law and their expressions is
T 1 = I 2 C 1 C 1 T ,
T 2 = I 2 C 2 C 2 T .
where C1 and C2 are identified as the normal vectors of the 2D scanning galvanometer and the mirror, respectively; their expressions are as follows:
C 1 = 1 0 0 0 cos ϕ sin ϕ 0 sin ϕ cos ϕ sin π 4   0   cos π 4 T ,
C 2 = cos φ 0 sin φ 0 1 0 sin φ 0 cos φ 1 0 0 0 cos η sin η 0 sin η cos η 1 0 0 0 cos ϕ sin ϕ 0 sin ϕ cos ϕ cos π 4   0   sin π 4 T .
In Equations (2) and (3), the 4 × 4 identity matrix is represented by I, and the refraction transformation matrix of the lens in Equation (1) is denoted by K, as follows:
K = f H 1 0 0 0 1 0 0 0 1 .
Similarly, the influence of the flight path angle of attack (α) on the image motion vector is as follows:
ν α = T 1 K T 2 cos α 0 sin α 0 1 0 sin α 0 cos α ν 0 0 .
Following the sequence of sideslip angle, attack angle, and roll angle, the image motion vector ( ν s p e e d ) caused by the target on the detector target surface can be derived as follows:
ν s p e e d = T 1 K T 2 1 0 0 0 cos γ sin γ 0 sin γ cos γ cos α 0 sin α 0 1 0 sin α 0 cos α cos β sin β 0 sin β cos β 0 0 0 1 ν 0 0
An assumption is made that the camera rotates at constant speed around the x-axis, with the rotation speed being ω. The swing speed on the image plane is then formulated as follows:
v roll = f ω .
The vector can be written as
ν = [ 0   v roll   0 ] T .
This leads to the synthesis of the vector and the image motion velocity vector of the entire system as follows:
ν s y s t e m = ν s p e e d + ν .
On this foundation, the aliasing required for super-resolution is attainable by adding the micro-displacement vector, and the final vector needed for super-resolution is
ν s r = ν s p e e d + ν + ν m d .
where ν m d represents the superimposed micro-displacement vector in single-step compensation. Its value might be one-half pixel, one-third pixel, or other step sizes to gather aliasing information. This can be chosen based on the required sequence depth and moving path parameters.

2.2. Two-Dimensional Scanning Galvanometer Movement Mode

In the staring imaging mode, the two-dimensional oscillating mirror is operated multiple times within a restricted range to obtain aliasing information. Using the image sequence depth of 4 as an example, Figure 2 illustrates the change in the optical-axis pointing angle brought by the 2D scanning galvanometer during the imaging process.
An assumption is further made that the aircraft flies at a constant speed. Figure 3 illustrates sequential moments (t1, t2, t3, and t4) where the center of the boresight points to specific locations (points 1, 2, 3, and 4). The angle of the optical-axis pointing in the four imaging changes is minimal. In the cross-sectional view of the direction, the displacement of the center of the visual axis between moments t1 and t4, and between t2 and t3 is Δ x . Similarly, the displacement in the side view of the direction between moments t1 and t2, and between t3 and t4 is Δ y .
These changes in pointing ultimately influence the position of the image point on the image plane. During the imaging process with a sequence depth of 4, the trajectory of the image point on the detector’s image plane is depicted in Figure 4.
From moments t1 to t4 of the staring imaging process, to capture the aliasing information of the ground target, the swing mirror not only compensates for the image motion caused by velocity and swing but also adds forward and backward micro-displacements at t1 and t3. Additionally, micro-displacements are added in the swing direction at t2 and t4. The red area represents the position on the image plane corresponding to the center of the boresight. As illustrated in Figure 4, the camera acquires the first image at the initial position. When forming the second image, the image point corresponding to the center of the visual axis is shifted to the right by 1/2 pixel from its previous position due to the micro-displacement added by the swing mirror. The process then proceeds to acquire the third image, shifting 1/2 pixel to the rear side, and moving the fourth image to the front by 1/2 pixel in a similar fashion.
In summary, the pendulum mirror’s swing angle must take into account two effects: compensation for image motion and superimposed micro-displacement to capture aliasing information. The swing path of the pendulum mirror and the micro-displacement step size can vary. The micro-displacement step size might be planned directly according to 1/3 or 1/4 pixel, but the depth of the imaging sequence changes in response. The longer the imaging sequence, the greater the compensation required for image motion by the pendulum mirror. The design of the step size and path must align with the specific system, aiming to gather sufficient aliasing information.

3. Super-Resolution Algorithm Process and Principle

In this chapter, the processing flow of the super-resolution algorithm is presented. Given the optical–mechanical structure characteristics of the aerial visible-light camera, aliased images are achieved with the regular swing of the two-dimensional pendulum mirror. The sequences are subsequently registered, and an anisotropic kernel function is employed for super-resolution reconstruction. This method is similar to the work carried out by Wronski et al. [22], where the data processing includes alignment and anisotropic regression kernel construction. However, in this paper, the micro-displacement relationship between sequence images is defined, greatly enhancing the algorithm’s computing speed and robustness.
Initially, the original image sequence that has been acquired is registered according to the reference frame. The micro-displacement vector matrix for each frame in the image sequence relative to the base frame is ascertained. The exact positional relationship of the images within the known sequence enables the exclusion of mis-registered regions and objects, thereby enhancing registration accuracy.
Subsequently, the local gradient information of the image is calculated. The shape of the kernel function is then adjusted based on the local gradient tensor matrix. The anisotropic kernel function alters the shape in accordance with changes at the image’s edge, with different weights being allocated to various positions to preserve the details of the reconstructed image.
Finally, these images are weighted and accumulated to produce a super-resolution image. The complete process is visually represented in Figure 5.

3.1. Image Sequence Alignment Based on Controllable Displacement

The problem of image registration forms the foundation for ensuring the accuracy of image super-resolution. In the image processing flow described in this paper, alignment must first be performed on the image obtained through micro-displacement before it can be input into the image algorithm sequence. Generally, the first image in the sequence is selected as the reference frame, and subsequent images are aligned with this frame. This principle is depicted in Figure 6.
In the alignment of an image sequence with an aliasing relationship to the same reference frame, the camera samples the same position in space N times. Due to the presence of micro-displacement, different sampling positions on the detector’s target surface are achieved. Consequently, the color information of the three RGB spectral segments at the same spatial position can be obtained, allowing for the direct generation of color images [25].
The characteristics of aerial imaging include a relatively long imaging distance and minimal sudden or rapid target position changes within the FOV. These characteristics align with the fundamental assumption of the L-K optical flow method [26]. To attain sub-pixel-level registration accuracy, the pyramid method is utilized using multiple iterations. The control over registration accuracy is mainly executed by selecting pyramid layers and the search range of the registration module. Considering the moving distance of the image plane caused by single-step micro-displacement, there is no need to search the registration area extensively. A constrained search range significantly enhances both registration accuracy and computational efficiency. If the calculated displacement vector of the current frame is less than or equal to the current single-step micro-displacement distance, the data are retained. Otherwise, the pixels in that area are discarded, enhancing registration robustness and minimizing the effects of misregistration on image quality, and it is as follows:
S x _ F i n a l S y _ F i n a l = S x S y i f   S x Δ x | | S y Δ y d i s c a r d e l s e .
where Δ x and Δ y symbolize the displacement triggered by the single-step movement of the micro-motion mechanism on the detector’s target surface.

3.2. Super-Resolution Reconstruction Based on Anisotropic Gaussian Kernel

Kernel regression is an approach for fitting a nonlinear model, fundamentally employing the kernel function as a weight function to create a regression model. Unlike the traditional isotropic kernel function, in image applications, there is a desire for the kernel shape to correlate with the image’s edge information. The kernel function’s shape at various positions is adjusted in accordance with the current point’s intensity information, the surrounding point image intensity information, and other factors. This ensures that the edge information in the image after the regression operation is more pronounced, effectively enhancing image authenticity and the signal-to-noise ratio, as well as realizing anisotropy in the edge area [27,28]. The adaptive kernel regression method consists of two basic steps, where the first step is an initial estimate of the image gradients, which is made using some kind of gradient estimator. The second step is to use this estimate to measure the main directions of local gradients in the image. In the second step, this orientation information is used to adaptively “steer” the local kernel, resulting in elongated, elliptical contours spread along the directions of the local edge structure. A steering matrix is defined as follows:
H i s t e e r = h μ i C i 1 2 .
This matrix is a symmetric covariance matrix of local pixel gray-level differences. By appropriately choosing C i and h , the algorithm can control the direction of the kernel function and the denoising effect. The local edge structure is related to the gradient covariance. Assuming a Gaussian kernel for the kernel function, the mathematical form of the steering matrix is as follows:
K H i s t e e r ( x i x ) = det ( C i ) 2 π h 2 μ i 2 exp ( x i x ) T C i ( x i x ) 2 h 2 μ i 2 .
where Ci is the local covariance matrix of a kernel function, which can be expressed by Equation (16) and is obtained with the singular-value decomposition (SVD) of the covariance matrix based on the local gradient.
C i = γ i U θ i Λ i U θ i T ,
where U θ i denotes the function of the rotation matrix, Λ i represents the function of the stretching matrix, and γ i is the radius amplitude of the ellipse; their definitions are shown in Equations (17) and (18). By using these three factors, a kernel function constructed from local gradients can control the weight of the kernel function at the image’s edge. It provides greater weight to the tangential direction of the edge, thereby enhancing the clarity of the target contour information.
U θ i T = cos θ i sin θ i sin θ i cos θ i ,
Λ i = σ i 0 0 σ i 1 .

4. Experimental Verification

4.1. Experimental Device Construction

The algorithm delineated in this paper is implemented in an aerial imaging system, including the processing flow and the super-resolution algorithm. Initially, the algorithm is laboratory-tested, where the two-dimensional swing mirror only performs swing compensation and micro-displacement superposition. Subsequently, in-flight tests are conducted. Speed compensation, swing compensation, and micro-displacement superposition are added to the motion of the 2D scanning galvanometer, validating the algorithm’s effectiveness in real-world applications.

4.2. Comparison with Demosaicing Techniques

In this paper, the implemented algorithm exploits the high-frequency aliasing introduced by multi-frame sequence image information to restore the color of the Bayer format image, instead of using interpolation. Several classical interpolation methods are selected for comparison, including VNG [29], FlexISP [30], and ADMM [31] algorithms. Real scenes serve as input images, and the comparative results are illustrated in Figure 7.

4.3. Resolution Test

The algorithm is evaluated using resolution test targets. The test device is shown in Figure 8.
Using a collimator with a focal length of 1.6 m, the width of the resolution target ranges from 0.04 mm to 0.02 mm from the widest to the narrowest, and the width difference between two adjacent groups of stripes is 0.002 mm. Its dimensions are shown in Figure 9.
The micro-displacement required for super-resolution is derived from the 2D scanning galvanometer in the optical system, with constant exposure time and gain. Under laboratory conditions, the resolution target is photographed, with the original image being depicted in Figure 10.
It can be analyzed from Figure 10 that the resolution of the original image is 13.8 lp/mm and the angular resolution is 0.000045 rad. Different numbers of image sequences are employed to assess the super-resolution algorithm, and the results are presented in Figure 11.
Through visual observation, we find the smallest target that can be resolved in each sequence depth result. It can be seen from Table 1 that when the image sequence depth is changed, the values of optical resolution and angular resolution change accordingly.
The tests demonstrate that the multi-frame super-resolution algorithm significantly reduces noise in the image, noticeably enhancing the resolution of details.

4.4. Spatial Resolution Improvement Test

4.4.1. Daytime Outdoor Scene Imaging Test

To further evaluate the algorithm, both daytime and nighttime scenes are tested. In the daytime scene, a multi-frame sequence of 8 or 12 images is used, with integration time and gain maintained. The imaging scene is portrayed in Figure 12.
The comparative effects between the bicubic interpolation image and the super-resolution image are displayed in Figure 13, which is the detail in the red box in Figure 12. The super-resolution image exhibits superior edge details and texture restoration capabilities.

4.4.2. Nighttime Outdoor Scene Imaging Test

In the night scene, the image sequence retains identical imaging conditions, ensuring consistent exposure time and gain during the process. A total of 12 frames are utilized, and the imaging scene is captured in Figure 14.
A comparison of the interpolated original image with the super-resolution image under nighttime imaging conditions reveals superior noise control in the super-resolution image. The restoration of target details in the super-resolution image is significantly more effective than in the original. Details that were previously submerged in noise within the original image were well restored, as shown in Figure 15.

4.4.3. Influence of Accumulation Cycle on Performance

To examine the impact of image sequence length on the super-resolution effect, tests were conducted on both sunlight and night scenes. A selection of 4 to 16 images was used for each test. Figure 16 illustrates the PSNR and SNR of the super-resolution images, highlighting that the number of images in the sequence positively affects these parameters during nighttime. In a static scene, the inclusion of more image sequences yields improved super-resolution effects.
Contrarily, in daytime scenarios, increasing the number of image frames does not continuously enhance the PSNR and SNR of the super-resolution images in Figure 17. An upward trend becomes stagnant or even declines after the number of image frames reaches eight. This stagnation is attributed to the increasing complexity of image registration and the rising probability of misregistration in image details with the addition of image sequences. As a result, the image evaluation factor may remain static or decrease. Therefore, a reasonable control of sequence depth from 8 to 12 frames is suggested based on the test results.

4.4.4. Results of Flight Experiments

The algorithm was deployed in a photoelectric load and tested during a flight. The pendulum mirror’s movements were set to eight, with a step size of half a pixel, following a circular rectangular path; the speed-to-height ratio of the system was 44.5; the pixel size was 5.5 microns; and the focal length of the lens was 180 mm. Figure 18 depicts the experimental results.
The test shows that the algorithm can be applied in the optoelectronic pod, and a careful inspection of local details shows that the algorithm has good efficacy in enhancing image resolution.

5. Conclusions

A super-resolution algorithm has been formulated for aeronautical optoelectronic systems equipped with a 2D scanning galvanometer. Initially, within a standard aerial optical system with a two-dimensional swing mirror, a mathematical model was established between the target displacement and image point on the image plane. The swing path was designed, and the swing vector was analyzed.
Subsequently, based on aerial ground imaging characteristics, a super-resolution algorithm was designed, employing a multi-layer pyramid optical flow method to register sequential images with known micro-displacement, thereby substantially improving registration accuracy and robustness.
An anisotropic kernel function was generated using the gradient tensor matrix, and this kernel function was employed to perform super-resolution reconstruction of the image. This approach significantly preserves image edge details and augments system resolution.
An aeronautical optoelectronic load incorporating the algorithm underwent tests under both daytime and nighttime conditions. The results demonstrate the algorithm’s ability to markedly suppress noise and enhance image resolution in night scenes. In a resolution test, 16 frames in our algorithm increased the resolution of the optical system by 39%.
Furthermore, an increase in the sequence’s number of images can elevate the algorithm’s performance, but beyond a certain limit, gains in SNR and PSNR slow down due to factors such as misregistration and compensation of nonlinear errors caused by increased image motion. Data indicate that a sequence depth of 8–12 images should generally be selected for optimal results.

Author Contributions

Conceptualization T.M.; methodology, T.M.; software, T.M., C.L. and Y.H.; validation, Y.X.; formal analysis, T.M. and Y.L.; investigation T.M. and F.Y.; resources, Y.X. and L.M.; writing—original draft preparation, T.M. and Y.H.; writing—review and editing, T.M. and Y.H.; visualization, T.M. and L.M.; supervision, Y.L. and H.S.; project administration, Y.L., Y.X. and H.S.; All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

This study did not create new data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Gouveia, L.C.P.; Choubey, B.J.S.r. Advances on CMOS image sensors. Proc. IEEE 2016, 36, 231–239. [Google Scholar] [CrossRef]
  2. Jacobsen, K.J. High resolution satellite imaging systems-an overview. Photogramm. Fernerkund. Geoinf. 2005, 2005, 487. [Google Scholar]
  3. Moser, G.; Serpico, S.B.; Benediktsson, J.A.J. Land-cover mapping by Markov modeling of spatial–contextual information in very-high-resolution remote sensing images. Proc. IEEE 2012, 101, 631–651. [Google Scholar] [CrossRef]
  4. Yue, L.; Shen, H.; Li, J.; Yuan, Q.; Zhang, H.; Zhang, L.J. Image super-resolution: The techniques, applications, and future. Signal Process. 2016, 128, 389–408. [Google Scholar] [CrossRef]
  5. Zhu, Y.; Chen, X.; Yuan, W.; Chu, Z.; Wong, K.-y.; Lei, D.; Yu, Y.J. A waveguide metasurface based quasi-far-field transverse-electric superlens. Opto-Electron. Adv. 2021, 4, 210013. [Google Scholar] [CrossRef]
  6. Xiong, Z.; Melzer, J.E.; Garan, J.; McLeod, E.J. Optimized sensing of sparse and small targets using lens-free holographic microscopy. Opt. Express 2018, 26, 25676–25692. [Google Scholar] [CrossRef]
  7. Dorr, F.J. Satellite image multi-frame super resolution using 3D wide-activation neural networks. Remote. Sens. 2020, 12, 3812. [Google Scholar] [CrossRef]
  8. Xu, J.; Liang, Y.; Liu, J.; Huang, Z.J. Multi-frame super-resolution of Gaofen-4 remote sensing images. Sensors 2017, 17, 2142. [Google Scholar] [CrossRef]
  9. Bishara, W.; Su, T.-W.; Coskun, A.F.; Ozcan, A.J. Lensfree on-chip microscopy over a wide field-of-view using pixel super-resolution. Opt. Express 2010, 18, 11181–11191. [Google Scholar] [CrossRef]
  10. Fortin, J.; Chevrette, P. Realization of a Fast Microscanning Device for Infrared Focal Plane Arrays; SPIE: Bellingham, WA, USA, 1996; Volume 2743. [Google Scholar]
  11. Wiltse, J.; Miller, J. Imagery improvements in staring infrared imagers by employing subpixel microscan. J. Optical. Eng. 2005, 44, 056401. [Google Scholar] [CrossRef]
  12. Zhang, X.; Huang, W.; Xu, M.; Jia, S.; Xu, X.; Li, F.; Zheng, Y.J. Super-resolution imaging for infrared micro-scanning optical system. Opt. Express 2019, 27, 7719–7737. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, C.; Everitt, J.H.; Murden, D. Evaluating high resolution SPOT 5 satellite imagery for crop identification. Comput. Electron. Agric. 2011, 75, 347–354. [Google Scholar] [CrossRef]
  14. Reulke, R.; Becker, S.; Haala, N.; Tempelmann, U.J. Determination and improvement of spatial resolution of the CCD-line-scanner system ADS40. ISPRS J. Photogramm. Remote. Sens. 2006, 60, 81–90. [Google Scholar] [CrossRef]
  15. Tang, X.; Hu, F.; Wang, M.; Pan, J.; Jin, S.; Lu, G. Inner FoV Stitching of Spaceborne TDI CCD Images Based on Sensor Geometry and Projection Plane in Object Space. Remote. Sens. 2014, 6, 6386–6406. [Google Scholar] [CrossRef]
  16. Datsenko, D.; Elad, M.J.M. Example-based single document image super-resolution: A global MAP approach with outlier rejection. Syst. Signal Process. 2007, 18, 103–121. [Google Scholar] [CrossRef]
  17. Kim, K.I.; Kwon, Y. Example-based learning for single-image super-resolution. In Proceedings of the Joint Pattern Recognition Symposium, Munich, Germany, 10–13 June 2008; pp. 456–465. [Google Scholar]
  18. Keren, D.; Peleg, S.; Brada, R. Image sequence enhancement using sub-pixel displacements. In Proceedings of the Computer Vision and Pattern Recognition, Computer Society Conference, Ann Arbor, MI, USA, 5–9 June 1988. [Google Scholar]
  19. Alam, M.S.; Bognar, J.G.; Hardie, R.C.; Yasuda, B.J.J. Infrared image registration and high-resolution reconstruction using multiple translationally shifted aliased video frames. IEEE Trans. Instrum. Meas. 2002, 49, 915–923. [Google Scholar] [CrossRef]
  20. Irani, M.; Peleg, S.J.I. Super resolution from image Sequences. In Proceedings of the 10th International Conference on Pattern Recognition, Atlantic City, NJ, USA, 16–21 June 1990. [Google Scholar]
  21. Xie, W.; Zhang, F.; Chen, H.; Qin, Q. Blind Super-Resolution Image Reconstruction Based on POCS Model. In Proceedings of the International Conference on Measuring Technology & Mechatronics Automation, Zhangjiajie, China, 11–12 April 2009. [Google Scholar]
  22. Stark, H.; Oskoui, P. High-resolution image recovery from image-plane arrays, using convex projections. JOSA A 1989, 6, 1715. [Google Scholar] [CrossRef]
  23. Tom, B.C.; Katsaggelos, A.K. Reconstruction of A High-Resolution Image by Simultaneous Registration, Restoration, and Interpolation Of Low-Resolution Images. In Proceedings of the Image Processing, International Conference, Washington, DC, USA, 23–26 October 1995. [Google Scholar]
  24. Schultz, R.R.; Stevenson, R.L.J. Extraction of high-resolution frames from video sequences. IEEE Trans. Image Process 1996, 5, 996–1011. [Google Scholar] [CrossRef] [PubMed]
  25. Wronski, B.; Garcia-Dorado, I.; Ernst, M.; Kelly, D.; Krainin, M.; Liang, C.-K.; Levoy, M.; Milanfar, P.J. Handheld multi-frame super-resolution. ACM Trans. Graph. 2019, 38, 1–18. [Google Scholar] [CrossRef]
  26. Lucas, B.D.; Kanade, T. An iterative image registration technique with an application to stereo vision. In Proceedings of the IJCAI’81: 7th International Joint Conference on Artificial Intelligence, Vancouver, BC, Canada, 24–28 August 1981; pp. 674–679. [Google Scholar]
  27. Takeda, H.; Farsiu, S.; Milanfar, P.J. Kernel regression for image processing and reconstruction. IEEE Trans. Image Process 2007, 16, 349–366. [Google Scholar] [CrossRef]
  28. Takeda, H.; Farsiu, S.; Milanfar, P. Robust kernel regression for restoration and reconstruction of images from sparse noisy data. In Proceedings of the 2006 International Conference on Image Processing, Atlanta, Georgia, 8–11 October 2006; pp. 1257–1260. [Google Scholar]
  29. Chang, E.; Cheung, S.; Pan, D.Y. Color filter array recovery using a threshold-based variable number of gradients. In Proceedings of the Sensors, Cameras, and Applications for Digital Photography, San Jose, CA, USA, 27–28 January 1999; pp. 36–43. [Google Scholar]
  30. Heide, F.; Steinberger, M.; Tsai, Y.-T.; Rouf, M.; Pająk, D.; Reddy, D.; Gallo, O.; Liu, J.; Heidrich, W.; Egiazarian, K.J. Flexisp: A flexible camera image processing framework. ACM Trans. Graph. 2014, 33, 1–13. [Google Scholar] [CrossRef]
  31. Tan, H.; Zeng, X.; Lai, S.; Liu, Y.; Zhang, M. Joint demosaicing and denoising of noisy bayer images with ADMM. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 2951–2955. [Google Scholar]
Figure 1. The base optical structure of an aerial camera.
Figure 1. The base optical structure of an aerial camera.
Photonics 10 01203 g001
Figure 2. Diagram of boresight pointing at different moments of imaging.
Figure 2. Diagram of boresight pointing at different moments of imaging.
Photonics 10 01203 g002
Figure 3. Diagram of micro-displacement and angle. (a) The change in the boresight in the cross-sectional view of flight direction; (b) the change in the boresight in the side view of flight direction.
Figure 3. Diagram of micro-displacement and angle. (a) The change in the boresight in the cross-sectional view of flight direction; (b) the change in the boresight in the side view of flight direction.
Photonics 10 01203 g003
Figure 4. Schematic diagram of the motion of the image point on the image sensor.
Figure 4. Schematic diagram of the motion of the image point on the image sensor.
Photonics 10 01203 g004
Figure 5. Overview of super-resolution algorithm.
Figure 5. Overview of super-resolution algorithm.
Photonics 10 01203 g005
Figure 6. Sub-pixel displacements from 2D scanning galvanometer.
Figure 6. Sub-pixel displacements from 2D scanning galvanometer.
Photonics 10 01203 g006
Figure 7. Comparison with demosaicing techniques. (a) Bayer; (b) VNG; (c) FlexISP; (d) ADMM; (e) ours.
Figure 7. Comparison with demosaicing techniques. (a) Bayer; (b) VNG; (c) FlexISP; (d) ADMM; (e) ours.
Photonics 10 01203 g007
Figure 8. Optical resolution test device.
Figure 8. Optical resolution test device.
Photonics 10 01203 g008
Figure 9. Resolution test target dimension drawing.
Figure 9. Resolution test target dimension drawing.
Photonics 10 01203 g009
Figure 10. Raw image of the resolution test target.
Figure 10. Raw image of the resolution test target.
Photonics 10 01203 g010
Figure 11. Experimental results of the algorithm on multi-frame resolution target images. (a) The result of 4 frames; (b) the result of 8 frames; (c) the result of 12 frames; (d) the result of 16 frames.
Figure 11. Experimental results of the algorithm on multi-frame resolution target images. (a) The result of 4 frames; (b) the result of 8 frames; (c) the result of 12 frames; (d) the result of 16 frames.
Photonics 10 01203 g011
Figure 12. Imaging scenes under daytime conditions.
Figure 12. Imaging scenes under daytime conditions.
Photonics 10 01203 g012
Figure 13. Imaging scenes under daytime conditions. (a) Bicubic interpolation image; (b) 8-frame super-resolution image; (c) 12-frame super-resolution image.
Figure 13. Imaging scenes under daytime conditions. (a) Bicubic interpolation image; (b) 8-frame super-resolution image; (c) 12-frame super-resolution image.
Photonics 10 01203 g013
Figure 14. Imaging scenes under night conditions.
Figure 14. Imaging scenes under night conditions.
Photonics 10 01203 g014
Figure 15. Comparison of night imaging details. (a) Bicubic interpolation image; (b) 8-frame super-resolution image; (c) 12-frame super-resolution image.
Figure 15. Comparison of night imaging details. (a) Bicubic interpolation image; (b) 8-frame super-resolution image; (c) 12-frame super-resolution image.
Photonics 10 01203 g015
Figure 16. The relationship between the number of frames, and PSNR and SNR in the night scene.
Figure 16. The relationship between the number of frames, and PSNR and SNR in the night scene.
Photonics 10 01203 g016
Figure 17. The relationship between the number of frames, and PSNR and SNR in the daytime scene.
Figure 17. The relationship between the number of frames, and PSNR and SNR in the daytime scene.
Photonics 10 01203 g017
Figure 18. Flight test image. (a) Bicubic interpolation image and details; (b) 8-frame super-resolution image and details.
Figure 18. Flight test image. (a) Bicubic interpolation image and details; (b) 8-frame super-resolution image and details.
Photonics 10 01203 g018
Table 1. Resolution statistical table.
Table 1. Resolution statistical table.
Number of ImagesOptical
Resolution
Angular
Resolution
Resolution Improvement Compared with RAW
414.7 lp/mm0.0000425 rad6%
815.6 lp/mm0.0000400 rad13%
1216.7 lp/mm0.0000375 rad21%
1619.2 lp/mm0.0000325 rad39%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, T.; Liang, C.; Han, Y.; Yuan, F.; Meng, L.; Xu, Y.; Shen, H.; Liu, Y. Super-Resolution Imaging Enhancement through a 2D Scanning Galvanometer: Algorithm Formulation and Application in Aerial Optoelectronic Systems. Photonics 2023, 10, 1203. https://doi.org/10.3390/photonics10111203

AMA Style

Ma T, Liang C, Han Y, Yuan F, Meng L, Xu Y, Shen H, Liu Y. Super-Resolution Imaging Enhancement through a 2D Scanning Galvanometer: Algorithm Formulation and Application in Aerial Optoelectronic Systems. Photonics. 2023; 10(11):1203. https://doi.org/10.3390/photonics10111203

Chicago/Turabian Style

Ma, Tianxiang, Chao Liang, Yuting Han, Fang Yuan, Lingtong Meng, Yongsen Xu, Honghai Shen, and Yunqing Liu. 2023. "Super-Resolution Imaging Enhancement through a 2D Scanning Galvanometer: Algorithm Formulation and Application in Aerial Optoelectronic Systems" Photonics 10, no. 11: 1203. https://doi.org/10.3390/photonics10111203

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop