Next Article in Journal
Learning Soft Mask Based Feature Fusion with Channel and Spatial Attention for Robust Visual Object Tracking
Next Article in Special Issue
Demodulation Method for Loran-C at Low SNR Based on Envelope Correlation–Phase Detection
Previous Article in Journal
Laryngeal Lesion Classification Based on Vascular Patterns in Contact Endoscopy and Narrow Band Imaging: Manual Versus Automatic Approach
Previous Article in Special Issue
Implicit and Explicit Regularization for Optical Flow Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

High-Efficiency Microsatellite-Using Super-Resolution Algorithm Based on the Multi-Modality Super-CMOS Sensor

1
Key Laboratory of 3D Information Acquisition and Application, MOE, Capital Normal University, Beijing 100048, China
2
Engineering Research Centre of Spatial Information Technology, MOE, Capital Normal University, Beijing 100048, China
3
The Institute of Remote Sensing, Beijing 100011, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(14), 4019; https://doi.org/10.3390/s20144019
Submission received: 2 June 2020 / Revised: 17 June 2020 / Accepted: 15 July 2020 / Published: 20 July 2020
(This article belongs to the Special Issue Data, Signal and Image Processing and Applications in Sensors)

Abstract

:
To realize the application of super-resolution technology from theory to practice, and to improve microsatellite spatial resolution, we propose a special super-resolution algorithm based on the multi-modality super-CMOS sensor which can adapt to the limited operation capacity of microsatellite computers. First, we designed an oblique sampling mode with the sensor rotated at an angle of 26.56 ( arctan 1 2 ) to obtain high overlap ratio images with sub-pixel displacement. Secondly, the proposed super-resolution algorithm was applied to reconstruct the final high-resolution image. Because the satellite equipped with this sensor is scheduled to be launched this year, we also designed the simulation mode of conventional sampling and the oblique sampling of the sensor to obtain the comparison and experimental data. Lastly, we evaluated the super-resolution quality of images, the effectiveness, the practicality, and the efficiency of the algorithm. The results of the experiments showed that the satellite-using super-resolution algorithm combined with multi-modality super-CMOS sensor oblique-mode sampling can increase the spatial resolution of an image by about 2 times. The algorithm is simple and highly efficient, and can realize the super-resolution reconstruction of two remote-sensing images within 0.713 s, which has good performance on the microsatellite.

1. Introduction

Since the 20th century, issues such as resources, environment, disasters, and national security have become the main problems in the development of human society. As one of the important means to solve these problems, space-to-earth observation technology has received much more attention from all countries, which has promoted the rapid development of aerospace and remote-sensing technology. At present, refined remote sensing has become the main direction of the development of the current remote-sensing field.
Generally speaking, there are mainly two ways to improve satellite image resolution. One is known as the “hardware way”. Since 1970, CCD (Charge-coupled Device) and CMOS (Complementary Metal Oxide Semiconductor) image sensors have been widely used in satellite remote sensing. The most direct method is to reduce the size of the photosensitive unit by improving the sensor fabrication process to increase the number of pixels in the image. Increasing camera focal length is another “hardware way”. It can reduce the instantaneous field of view while ground sampling, but it results in more processing difficulty and higher manufacturing cost of optical parts. Due to the limitation of technology level and image data transmission rate, the resolution of images obtained by high-resolution CCD camera is very limited [1]. Considering these factors, the technology of image super-resolution has been proposed, which is called the “software way”. To achieve the effects of improving image resolution, highlight feature details, and improve image quality, the super-resolution technology increases the number of image pixels by a variety of ways such as information fusion, prior constraint, and so on [2]. By dealing with images, it can break the limitations of current imaging devices. However, this technology has limited by application scenarios, and the improvement effect is susceptible to the redundancy of complementary information, which restricts the wide application of super-resolution technology in the field of remote sensing.
Therefore, we present a method of combining hardware equipment with software algorithm to improve the resolution of microsatellite imaging. The authors designed a hardware equipment which is known as multi-modality super-CMOS sensor, or MS-CMOS for short. Its sampling mode can effectively increase the sampling density in both the satellite flight direction and the perpendicular direction and provide the best data for the satellite-using super-resolution algorithm. For the software algorithm, which is based on the sub-pixel super-resolution theories, we proposed a special super-resolution algorithm using MS-CMOS data which can adapt to the limited operation capacity of microsatellite computers.
This paper is organized as follows: some related works are mentioned in Section 2. The multi-modality super-CMOS sensor is introduced in Section 3. The data-simulation method is proposed in Section 4. The super-resolution algorithm based on the oblique sampling model is proposed in Section 5. The experiments and discussion are shown in Section 6. Section 7 shows the conclusion.

2. Techniques for Improving the Resolution of Remote-Sensing Images

2.1. Satellite–Ground Combined Super-Resolution Technology

Super-resolution (SR) is the process of obtaining high-resolution (HR) images from one or more low-resolution (LR) images. However, there is not a general SR method or technology that can solve all the SR problems. The SR algorithm and parameters should be selected according to different application scenarios and conditions. Because of the limitation of system design and the complexity of the imaging environment, the spatial resolution of the image is restricted. Therefore, scholars have proposed a theory which is to combine the hardware and software, satellite, and ground to reduce the cost of human resources and material resources, and effectively improve image resolution. This method can realize multi-image fusion and improve the resolution by designing the special arrangement of multi-sensor and bypassing the image registration and correction.
French Centre National d’Etudes Spatiales (CNES) proposed a CCD arrangement method to increase the spatial resolution of SPOT5 satellite sensor. The SPOT5 satellite SR mode, SUPERMODE sampling and HIPERMODE sampling were proposed. The two sampling methods used the customized two rows CCD device with 0.5 pixel staggered in the linear array direction and 3.5 pixel staggered in the satellite flight direction [3] (pp. 318–320). HIPERMODE sampling method realized the fusion of two 5 m resolution images to obtain one 2.5~3 m HR image. The concept of STAGGERED ARRAY was proposed by the Hot Spot Recognition Sensors (HSRS) which was developed by Deutsches Zentrum für Luft und Raumfahrt (DLR) [4]. The core of the HSRS sensor was a staggered 512 × 2 double infrared ray array. The dislocation arrangement was specific: stagger 0.5 pixels in the array direction and stagger 2 whole pixels in the vertical array direction to obtain multiple images with pixel offset. After processing, the spatial resolution of the camera can be improved. The Leica ADS40 line scanner system consisted of two staggered CCD linear arrays. The ADS40 sampled at twice the frequency in the flight direction to obtain multiple images with half-pixel offset [5] (p. 565). In addition, a method of recovering HR images from interlaced lines was proposed, which can improve the spatial resolution of the camera [6]. Israel had developed a series of EROS (Earth Remote Observation System) satellites. The outstanding feature of this satellite was that it had the ability of single orbit imaging. This function allowed repeated sampling of the same area by tilting the camera. The resolution of EROS-B1 satellite can reach 0.85 m, and the over-sampling mode can reach a resolution of 0.5 m [7]. Nigerian satellite 2 was a small satellite with a spatial resolution of 2.5 m. The satellite design allowed intentional yaw, which reduced the width of the pixel’s ground track to achieve image SR [8]. SkySat series satellites used the method of surface array push-frame combined with ground processing, by the sequence image registration and SR algorithm processing on the ground, the image resolution was improved from 1.3 m to 0.9 m, and sub-meter imaging was realized [9,10]. Moreover, a satellite platform was designed to capture images with overlapping parts continuously to solve the problem of poor imaging quality caused by the limited payload of Nanosatellite [11]. They used the normalized cross-correlation method to determine the sub-pixel transformation between images, laying a foundation for CubeSats to achieve SR.
In China, Zhou, C. P. [12] (pp. 1–27), [13] proposed the satellite oblique-mode sampling method, which combined satellite hardware with ground-based image processing software to improve the spatial resolution. By tilting the linear array detector 45° and adjusting the integration time, Zhou, F. [14] improved the resolution of the sampled image by about 1.64 times compared with that before rotation.

2.2. Software-Hardware Combined Sub-Pixel Super-Resolution Algorithm

The sub-pixel imaging is a method to obtain multiple LR images with mutual dislocation information of the same target first, and then extract the redundant information in the shifted image to improve the resolution, which is consistent with the core of two technologies mentioned above. The proposed SR algorithm is based on the similar idea, it is necessary to discuss the sub-pixel SR algorithm in this section.
Since the last century, there had existed resolution improvement methods which used displacement images. Aizawa [15] (pp. 289–292) proposed a method to improve the resolution of a single CCD sensor by software synthesis after displacement imaging using CCD. Hardie [16] used micro-scanning to reduce the interference noise generated by image sub-sampling, to realize the synthesis of multiple digital images and improve the resolution.
In the remote-sensing imaging system, An [17] used adjacent layers of optical fibers arranged into hexagonal patterns to obtain multiple images of sub-pixel dislocation, designed the interpolation function of image fusion, and improved the spatial resolution. Wu [18] designed a rotary micro-scanner, which can effectively improve the spatial resolution of infrared images and improve the function value of the system. Sun and Yu [19] believed that sub-pixel scanning via mobile sensors would be the key to SR reconstruction. They proposed a sur-pixel scanning method with low processing difficulty, which verified its advantages in engineering. Li [20] designed a registration method based on the sub-pixel displacement estimation to artificially improve the computational efficiency of image SR reconstruction, and verified the effectiveness of the method through experiments. Zhao [21] used a dynamic single-pixel camera to obtain sequential images with sub-pixel displacement and reconstructed the final HR image using wiener filter. Compared with the traditional single-pixel imaging system, this method had better spatial resolution and SNR (Signal Noise Ratio). Zhang [22] had designed a controllable sub-pixel micro-scanning infrared optical system which can achieve any step size of sampling, with the purpose of obtaining sub-pixel LR images, realize the de-noising and SR reconstruction of infrared images at the same time. Wang [23] used Risley prisms to design a low cost and high-efficiency method that can improve the image resolution of the imaging system when the resolution is limited by the pixel size. They established the mathematical model of SR reconstruction for single-camera sub-pixel imaging.

2.3. Remote-Sensing Image Post-Processing Technology

Remote-sensing image post-processing technology refers to the use of multiple LR images of the same scene in multiple time phases, combined with image registration and image SR algorithms, to achieve the SR reconstruction of remote-sensing image. Because the remote-sensing data has multi-phase, multi-angle, and multi-band complementary information, it can greatly meet the needs of SR reconstruction for complementary information. The traditional SR reconstruction of remote-sensing image mainly uses multi-temporal image data, in fact, the idea of SR reconstruction originated from processing multi-temporal Landsat remote-sensing image. By applying the theory of using single or multiple satellites to post-process and synthesize multiple images of the same target to improve image resolution, China’s Qian Xuesen Laboratory of Space Technology had broken through the key techniques of image registration and reconstruction of remote-sensing images SR processing, and carried out an experiment to improve the resolution of GaoFen-4 images [24,25].
For remote-sensing satellites, the theoretical research and academic discussion of image SR algorithm do not have the basis of engineering temporarily, which is not enough to achieve the purpose of improving the spatial resolution. There are many problems in the hardware design of satellite imaging sensor to improve the spatial resolution, such as the difficulty of realization, the long period of research and development, and the high cost of manufacture. Many image SR technologies, including deep learning, have a short application time in the field of remote sensing. The common idea is to conduct data fusion for LR images, which is prone to generate false information and requires the support of high-performance computer equipment, the processing capacity of satellite computer cannot meet the requirements. The combination of hardware design and software algorithm has become a new idea to improve the resolution of remote-sensing image for its low cost and convenience.

3. Multi-Modality Super-CMOS Sensor

3.1. Structure and Imaging Model

In 1969, a CCD Image Sensor was developed at the Alcatel-Lucent Bell Labs. The CCD Image Sensor has many advantages such as small volume, high resolution, high sensitivity, and so on. The research of the sensor in this paper choses CMOS device which has the same origin with CCD electric coupling device, but the price is lower.
We name this special sensor as multi-modality super-CMOS sensor, MS-CMOS for short. Its hardware structure is similar to a planar array CCD sensor, as shown in Figure 1. The imaging and output of MS-CMOS are similar to a linear array CCD sensor. Each row that forms the MS-CMOS is separately controlled, images simultaneously, and outputs independently.

3.2. Imaging Model and Sampling Model

Unlike the traditional push-broom or push-frame imaging along the satellite flight direction, MS-CMOS can achieve different types of imaging by controlling parameters such as CMOS row number, rotation angle, and so on. The MS-CMOS imaging model can be represented by the function (1):
I m = F ( N , M , α , T N , t i )
In this function:
I m : MS-CMOS imaging model,
N: number of line CMOS sensors number(vertical),
M: pixel number of line CMOS sensor(lateral),
α : CMOS sensor rotation angle,
T N : sampling time of each line CMOS sensor,
t i : integration time.
When N = 1, MS-CMOS can realize linear array CMOS sensor function.
When I 1 = I 2 = … = I N , MS-CMOS can realize planar array CMOS sensor function.
when α 0 , the sensor implements oblique-mode sampling, as shown in Figure 2.

4. MS-CMOS Imaging Simulation

The microsatellite with MS-CMOS sensor is scheduled to be launched this year. In the course of this research, the real MS-CMOS remote-sensing images cannot be obtained at the moment, the simulation function is given for the rotation angle sampling model of the sensor, which makes the simulation imaging process be realized by computer programming, simplifies the simulation process, and obtains the experimental data.
Since each row that forms the MS-CMOS is separately controlled, images simultaneously, and outputs independently, we design an imaging simulation mode of a single-line-array CMOS sensor.
The design of imaging simulation mode is based on the theory of image down-sampling. Down-sampling is an operation that reduces image resolution by using image interpolation algorithm to reduce image size. For an image which size is M · N , the resolution is reduced by S . times, and the image size is reduced by S . times. The flow chart of the imaging simulation mode is shown in Figure 3.
During the simulation process, the input HR images are sampled as LR images, we define a d-times down-sampling of the image, i.e., all the pixels in the d-d window are merged into a single pixel for output.
The linear array MS-CMOS sampling simulation mode is shown in function (2):
I = F ( α , A x , A y , d , m , p )
In this function:
I : MS-CMOS imaging model,
α: CMOS sensor rotation Angle,
A x : abscissa of simulation starting point A ( x , y )
A y : ordinate of simulation starting point A ( x , y )
d: down-sampling times,
m : pixel number of a linear array sensor,
p: sampling frequency of a linear array sensor,
When α = 0, it becomes the conventional simulation mode.
We use tan α = a b stead of the angle α in the input, so the input values of the mode include m , p , d , a , b and the size of the input original HR image r × c . All the other variables in the following functions are process variables. Points A , B , C , D . form a down-sampling window in clockwise order, as shown in Figure 4.
Building the mode of the jth window in the ith imaging, which is shown in function (3):
{ m 2 = ( k × a ) 2 + ( k × b ) 2 A x i j = ( [ k × a ] + 1 ) + ( i 1 ) × d p A y i j = 1 d 2 = ( n × a ) 2 + ( n × b ) 2 B x i j = A x i j n × a B y i j = A y i j + n × b C x i j = B x i j + n × b C y i j = B y i j + n × a D x i j = A x i j + n × b D y i j = A y i j + n × a
The loop function (4) is used in the simulation of the imaging down-sampling window:
{ A x i ( j + 1 ) = B x i j A y i ( j + 1 ) = B y i j B x i ( j + 1 ) = A x i ( j + 1 ) n × a B y i ( j + 1 ) = A y i ( j + 1 ) + n × b C x i ( j + 1 ) = B x i ( j + 1 ) + n × b C y i ( j + 1 ) = B y i ( j + 1 ) + n × a D x i ( j + 1 ) = C x i j D y i ( j + 1 ) = C y i j
The simulation data can be obtained by traversing the whole original HR image.

5. Satellite-Using Super-Resolution Algorithm

Compared with the phenomenon that the effective resolution improvement ratio of the image processed by the existing SR methods is much smaller than the magnification of the image, the sequence images obtained by MS-CMOS in oblique-mode sampling contain high overlap and fixed dislocation information in the same area, which have more non-redundant information than that obtained by traditional optical satellites. First, the oblique mode can improve the resolution of the image in the satellite flight direction by increasing the sampling frequency. Secondly, sampling with any rotation angle realizes periodic cycle imaging, after the nth imaging, the image element of the (n + 1)th becomes superposition with that of the first imaging. The periodic overlap relations are used to reconstruct the SR image without registration, so that the resolution can be improved in the perpendicular direction. In this way, the SR process can be simplified, and more image information can be recovered.
The satellite-using SR algorithm can meet the requirements of the limited processing capacity and the limited available resources of the on-board computer, realize the SR reconstruction of multi-image and improve the satellite spatial resolution.

5.1. Parameters Correlation Analysis

According to the above mode, if we set: n is the minimum number of used MS-CMOS, α is the rotation angle, f is the sampling frequency, d is the spacing distance of sampling, S . is the times of resolution improvement (theoretical value), the relationship between sampling frequency and sampling spacing distance is shown in formula (5):
f = 1 d .
When the MS-CMOS is tilting an angle of α, the relationship between rotation angle and the minimum number of used MS-CMOS is shown in formula (6):
tan α = 1 n
At this point, the sensor sampling frequency and the sampling spacing distance meet the function (7) and (8):
f = 1 + n 2
d = 1 f = 1 1 + n 2
During the imaging process, we use n rows of MS-CMOS to get n images. The SR processing of these images can increase the resolution of the images equally in both the satellite flight direction and the perpendicular direction, the increase time is:
S = 1 + n 2
For example, when tan α = 1, i.e., α = 45 , at this time n = 1 , the satellite uses 1 row of MS-CMOS to get 1 image, the sampling frequency is 2 times of the original, the image resolution can be improved by 2 times (theoretical value).
When tan α = 1 2 , i.e., α = 26.56 , at this time n = 2 , the satellite uses 2 rows of MS-CMOS to get 2 images, the sampling frequency is 5 times of the original, the image resolution can be improved by 5 times (theoretical value).

5.2. Image Super-Resolution Reconstruction

Theoretically, MS-CMOS is capable of imaging at various angles. According to the requirements of the microsatellite design, this section discusses the algorithm of 26.56 ( arctan 1 2 ) rotation angle, and takes 26.56 ( arctan 1 2 ) as an example in the follow-up experiments.
According to the earlier discussion, when α = 26.56 , the microsatellite uses 2 rows of MS-CMOS to get 2 images, the sampling frequency is 5 times of the original. The sampling mode is shown in Figure 5:
In Figure 5, 1 and 2 represent MS-CMOS row numbers, T 1 T 5 represent the imaging time. For a complete remote-sensing image obtained by splicing N strip images at N times, there exists pixel misplacement between rows in the image. The overlap rule is shown in formula (10):
{ ( T M x T ( M + 1 ) x , T M y T ( M + 1 ) y ) = ( 1 4 , 1 2 ) ( T M x T ( M + 2 ) x , T M y T ( M + 2 ) y ) = ( 1 2 , 1 )
Secondly, the overlapping relationship between CMOS1 and CMOS2 is analyzed. Comparing the image obtained by CMOS2 at T 3 time with that obtained by CMOS1 at T 1 time, there is a 1/2 pixel overlap region of 0.5 pixel displacement in x’ direction and 0 pixel displacement in y’ direction, as shown in the red box of Figure 5. The pattern of dislocation and overlap between the images is shown in formula (11):
{ CMOS 1 , T 1 x , T 1 y = CMOS 2 , 1 2 T 3 x , T 3 y CMOS 1 , T 2 x , T 2 y = CMOS 2 , 1 2 T 4 x , T 4 y CMOS 1 , T 3 x , T 3 y = CMOS 2 , 1 2 T 5 x , T 5 y CMOS 1 , T ( N 2 ) x , T ( N 2 ) y = CMOS 2 , 1 2 T N x , T N y
According to the two kinds of overlapping patterns, the SR reconstruction result is recorded as H, the CMOS1 imaging is L 1 , and the CMOS2 imaging is L 2 . The reconstruction model is shown in formula (12):
H ( i , j ) = ( L 1 ( i + 1 , j + 1 ) + L 2 ( i , j ) ) / 2
H ( i , j ) ,   L 1 ( i , j ) and L 2 ( i , j ) : The gray value of the pixel in ( i , j ) position.

5.3. Image Deformation Correction

Images sampled in oblique mode are distorted due to the image storage and display mechanism of the computer. Oblique sampling makes the real data change from the traditional rectangle to the parallelogram with jagged edges, and the images in the computer are automatically arranged into a rectangle display, which results in the deformation of the images.
Deformation correction is the inverse transformation of rotation angle, which means the coordinate transformation of each pixel in the image. The super-resolution image is recorded as SR, H has the same meaning as above, the inverse transformation mode is shown in formula (13) and (14):
S R   = F ( α , H )
Specifically:
{ S R x = α × H x S R y = H y
In this function:
S R x ,   S R y : abscissa and ordinate of pixel ( i , j ) in SR image,
α : MS-CMOS rotate angle,
H x , H y : abscissa and ordinate of pixel ( i , j ) in SR reconstruction result.

6. Experiments and Discussion

6.1. Simulation Experiments

There are two types of simulation data. The first kind of data is the experimental data of satellite-using SR algorithm. We use the simulation algorithm to conduct the down-sampling operation with the resolution of the original HR image decreased by 10 times. Specifically, we simulate a 2-rows MS-CMOS with an angle of 26.56 ( arctan 1 2 ) , 5 times of the frequency sampling. The second kind of data is used as the comparison data of the experimental results of the algorithm, which simulates the conventional sampling of microsatellite, we sample the original HR image with a resolution reduction of 10 times.
The original HR images selected for the simulation include one SkySat-2 image (seashore remote-sensing image with a resolution of 0.8 m), two Worldview-4 images (farmland and city remote-sensing images with a resolution of 0.31 m) and one target image (with a resolution of 2 m). These images are rich in texture details, which can provide a good reference for the following experiments. Figure 6 shows the original data, Figure 7 and Figure 8 show the experimental data, Figure 9 shows the comparison data. The areas selected by the red box in each type of the images are the same, those local images will be enlarged in Section 6.3 and Section 6.4 to analyze the details.

6.2. Algorithm Experiments

We experiment the 26.56 ( arctan 1 2 ) algorithm by using MATLAB (Matrix & Laboratory software) R2016a. The experimental content includes two parts: image SR reconstruction and image deformation correction. Figure 10 shows the seashore SR image, Figure 11 shows the farmland SR image, Figure 12 shows the city SR image, Figure 13 shows the target SR image. The areas selected by the red box in each type of the images are the same, those local images will be enlarged in Section 6.3 and Section 6.4 to analyze the details.

6.3. Evaluation of Super-Resolution Image Quality

In this section, we discuss the enlarged red box selected areas of the simulation data and the experimental results. Figure 14 shows the local magnification images of the seashore image in Figure 10, Figure 9, Figure 6, and Figure 7 in turn. Figure 15 shows the local magnification images of the farmland image in Figure 11, Figure 9, Figure 6, and Figure 7 in turn. Figure 16 shows the local magnification images of the city image in Figure 12, Figure 9, Figure 6, and Figure 7 in turn.
In the experiment, we select areas with rich texture and distinct features to enlarge, and analyze the ability of this algorithm in restoring the edge and other details. Figure 14 shows the chosen area of the seashore image which is an offshore structure in an Egyptian port, there is a distinct contour feature in the upper left corner of Figure 14d. Figure 14b lost the contour information in the upper left corner. However, Figure 14a, processed by the algorithm, successfully recovers part of the contour information, which is more similar to Figure 14d the original image. Figure 15 shows the chosen area of the farmland image which is the crops and houses in Xinjiang, China. The area contains an orderly array of crops and buildings, the crops in this area have distinct texture features, as shown in Figure 15d. In Figure 15b, the image information is indistinguishable, only fuzzy boundaries can be seen. However, Figure 15a, processed by the algorithm, not only contains the boundary, but also has parts of the crops texture information and roof features, which is more similar to Figure 15d the original image. Figure 16 shows the chosen area of the city image which is buildings in Sydney. The area contains distinct contour and edge features, as shown in Figure 16d. In Figure 16b, the edges are indistinguishable, and the feature information is lost. However, in Figure 16a, processed by the algorithm, buildings can be well distinguished, and the image retains some contour and edge information, which is more similar to Figure 16d the original image.
In addition, comparing the image (b) and the image (c) in Figure 14, Figure 15 and Figure 16, the images obtained by oblique-mode sampling are richer in edge details than those obtained by conventional mode sampling, which indicates that the special imaging design of MS-CMOS has improved the image resolution to a certain extent.

6.4. Evaluation of Algorithm Effectiveness

The effectiveness evaluation of the algorithm is to evaluate the ability of improving the image resolution. In this paper, the improvement of target image resolution is calculated to realize evaluation.
The target image usually consists of pairs of lines of different thickness, with different line pairs representing different resolution sizes. We evaluate the improvement times of image resolution by determining the corresponding value of the limit resolvable line pairs of the target image (minimum resolvable contras).
Let x represent the minimum value of the three lateral lines in the target image which can be resolved to the limit, let y represent the minimum value of the three vertical lines in the target image which can be resolved to the limit, the calculation formula of target image resolution is:
T a r g e t   R e s o l u t i o n = x 2 + y 2 2
We enlarge the red box of the target image in Section 6.1 and Section 6.2, calculate the resolution of each target image.
Figure 17 shows the selected area of target image which is the minimum resolvable line pairs. The image resolution of the original target image (Figure 17d) is 2 m. From the image (b) in Figure 17, the minimum resolvable value of line pairs is 19 m, according to the formula (15), the target resolution represented in Figure 17b is 19.00 m. From the image (c) in Figure 17, the minimum resolvable value of lateral line pairs is 13 m, the minimum resolvable value of vertical line pairs is 15 m, according to the formula (15), the target resolution represented in Figure 17c is 14.03 m. It confirms the conclusion of the previous section on the subjective evaluation of other experimental results, the resolution of MS-CMOS oblique-mode sampling data has been improved compared with the conventional sampling of sensors.
From the image (a) in Figure 17, the minimum resolvable value of line pairs is 9 m, according to the formula (15), the target resolution represented in Figure 17 image (a) is 9.00 m. Compared with the conventional sampling image, the resolution of SR image is increased by 19.00 9.00 2.11 times through the processing method of MS-COMS combined with the algorithm. The result is close to the theoretical value of 5 times, which proves that the proposed algorithm is effective.

6.5. Evaluation of Algorithm Practicality

Since the rotation of the satellite sensor cannot be accurately transferred to the theoretical design value of 26.56 , this section discusses the influence of rotation error on image resolution improvement. We rotate the original image of the target every 10 (rotate 180 clockwise and 180 counterclockwise respectively) to obtain 36 new target images, in which the anticlockwise rotation is represented by a negative value, repeat the experiment process and calculate the resolution. We calculate the lateral line pairs, the vertical line pairs and the image resolution improvement of the target image after clockwise rotation and counterclockwise rotation respectively, and the results are shown in the tables.
In Table 1, when the data is rotated clockwise, the improvement of image resolution is more influenced by the limit resolvable scale of the vertical line pairs. The minimum value of the image resolution improvement appears at 140 , which is 1.51 times. The complementary angle of 26.56 is the fourth quadrant angle, which is the clockwise 153.44 . The data around this angle, which are 140 ~ 160 images, the times of vertical resolution improvement of these data all show local minimum values. This means that when the angle between the actual orbit and the predetermined orbit appears clockwise due to the attitude adjustment or yaw of the satellite, when the algorithm is actually operated on the microsatellite, the spatial resolution can be improved at least 1.51 times.
In Table 2, when the data is rotated counterclockwise, the improvement of image resolution is more influenced by the limit resolvable scale of lateral line pairs. The minimum value of the image resolution improvement appears at 140 , which is 1.41 times. The redundancy angle of 26.56 is the third quadrant angle, which is the counterclockwise 116.56 ( 116.56 ). The data around this angle, which are 120 ~ 150 images, the times of lateral resolution improvement of these data all show local minimum values. This means that when the angle between the actual orbit and the predetermined orbit appears counterclockwise due to the attitude adjustment or yaw of the satellite, when the algorithm is actually operated on the microsatellite, the spatial resolution can be improved at least 1.41 times.
Based on the above analysis, it can be seen that there are differences in the resolution improvement times of the images which are used for the algorithm experiment after being rotated. When the satellite rotates clockwise, the image resolution of the target image increases by 1.78 times on average, with a minimum increase of 1.51 times; when the satellite rotates counterclockwise, the image resolution of the target image increases by 1.80 times on average, with a minimum increase of 1.41 times.
Therefore, the algorithm satisfies the preset technical index (the spatial resolution is increased by 2 1.41 times) when the project was designed, the algorithm has on-board practicality.

6.6. Evaluation of Algorithm Efficiency

The spaceborne data acquisition and processing system is an embedded system which is suitable for real-time and multi-task processing. Its main functions include collecting, framing, and transmitting the satellite telemetry information, executing the indirect remote control command, measuring, and controlling the satellite attitude, correcting the orbit autonomous fault, diagnosis and processing of satellite, and satellite missions and resources management. The hardware part mainly includes a FPGA (Field Programmable Gate Array) chip, a spaceborne computer, storage equipment, and so on. The software part includes pre-embedded development image processing programs, image compression programs, storage file systems, and so on. Some indicators of the satellite-borne data acquisition and processing system are as follows:
  • Microsatellite storage space: 1 TB
  • Data transmission capacity: 450 Mbps
  • Pixel size of MS-CMOS sensor: 7.4 µm × 7.4 µm
  • Pixel number of MS-CMOS sensor: 5144 × 3800
  • FPGA type: XILINX UltraScale
  • FPGA computing power: 3.15 Tflops (0.7 × 4.5 Tflops (theoretical maximum value))
  • HP (Hewlett-Packard) desktop computing power: 211.2 Gflops
The ground running environment of the algorithm is a HP desktop and MATLAB R2016a software. The algorithm comes from the theory of image interpolation. Besides the function of image reading, display, and storage, the core code of the program uses the basic image pixel operation and geometric transformation operation in MATLAB, and the method is simple. The experimental data obtained by simulation is smaller than the real remote-sensing images. We enlarge the experimental data to approximate the size of the microsatellite images for evaluation, the results are shown in Table 3. The program runs with less cache data, less memory, high efficiency, and short time. The time used includes the length of the image display function. When the algorithm is transplanted to the microsatellite computer, the image display function will not be needed. The actual running time will be further shortened. Because the FPGA chip has more computing power than the HP desktop, the proposed algorithm can adapt to the limited computer resource environment of the satellite, the algorithm has a certain satellite-using efficiency.

7. Conclusions

In this paper, a simple and efficient microsatellite-using SR algorithm is proposed for multi-modality super-CMOS sensor. First, the spatial resolution is improved from the hardware design by using the oblique sampling mode of the MS-CMOS sampling at several times of the conventional sampling frequency. Secondly, based on the displacement relationship between the oblique pattern sequences, the image SR reconstruction model and distortion correction is established, and several LR images are synthesized into one HR image to improve the spatial resolution. It is proved that the algorithm can effectively improve the satellite resolution, meet the preset engineering indexes, and have the feasibility of operating on the satellite.
Due to the limited resources on the microsatellite, complex algorithms cannot be realized. With the improvement of microsatellite technology, image preprocessing and other operations will be completed on the satellite, so that the image will get better results. When the data is transmitted to the ground, the resolution can be further improved, and the image information can be enriched by combining with other SR algorithms.

Author Contributions

Conceptualization, K.Z., C.Y. and C.Z.; Data curation, K.Z. and R.Z.; Formal analysis, K.Z. and C.Y.; Funding acquisition, C.Y., X.L., C.Z. and R.Z.; Investigation, K.Z., C.Z. and R.Z.; Methodology, K.Z., C.Y. and C.Z.; Resources, X.L. and R.Z.; Software, K.Z. and C.Y.; Supervision, X.L. and C.Z.; Validation, K.Z. and C.Z.; Visualization, K.Z.; Writing—original draft, K.Z.; Writing—review & editing, K.Z., C.Y., X.L., C.Z. and R.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the interdisciplinary research fund of Capital Normal University “Theory of micro-nano satellite multi-mode imaging and hardware acceleration of in-orbit image processing”.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, X.L. Super-Resolution Reconstruction Algorithm of Image Based on Neighborhood Learning and Sparse Atomic Clustering Dictionary. Ph.D. Thesis, Southeast University, Nanjing, China, 2016. [Google Scholar]
  2. Shen, H.F.; Ai, T.H.; Liu, Y.L.; Wang, Y. Development and application of super resolution image reconstruction technique. Meas. Control Technol. 2009, 6, 5–11. [Google Scholar]
  3. Dhérété, P.; Rougé, B. Image de-blurring and application to SPOT5 THR satellite imaging. In Proceedings of the 2003 IEEE International Geoscience and Remote Sensing Symposium, Toulouse, France, 21–25 July 2003. [Google Scholar]
  4. Skrbek, W.; Lorenz, E. HSRS: An infrared sensor for hot spot detection. SPIE’S Int. Symp. Opt. Sci. Eng. Instrum. 1998, 3437, 167–175. [Google Scholar]
  5. Sandau, R.; Braunecker, B.; Driescher, H.; Eckardt, A.; Hilbert, S.; Hutton, J.; Kirchhofer, W.; Lithopoulos, E.; Reulke, R.; Wicki, S. Design Principles of the LH Systems ADS40 Airborne Digital Sensor; ISPRS: Amsterdam, The Netherlands, 2000; p. 565. [Google Scholar]
  6. Reulke, R.; Becker, S.; Haala, N.; Tempelmann, U. Determination and improvement of spatial resolution of the CCD-line-scanner system ADS40. ISPRS J. Photogramm. Remote Sens. 2006, 60, 81–90. [Google Scholar] [CrossRef]
  7. Wang, L. An introduction to Isreali commercial high-resolution remote sensing satellite program. Sci. Surv. Map. 2002, 27, 52–54. [Google Scholar]
  8. Cawthorne, A.; Purll, D.; Eves, S. Very high resolution imaging using small satellites. In Proceedings of the 6th Responsive Space Conference, Los Angeles, CA, USA, 28 April–1 May 2008. [Google Scholar]
  9. Murthy, K.; Shearn, M.; Smiley, B.D.; Chau, A.H.; Levine, J.; Robinson, D. SkySat-1: Very high-resolution imagery from a small satellite. SPIE Remote Sens. 2014, 9241, 92411E. [Google Scholar]
  10. D’Angelo, P.; Mattyus, G.; Reinartz, P. Skybox image and video product evaluation. Int. J. Image Data Fusion 2015, 7, 1–16. [Google Scholar] [CrossRef] [Green Version]
  11. Lüdemann, J.; Barnard, A.; Malan, D.F. Sub-pixel image registration on an embedded Nanosatellite Platform. Acta Astronaut. 2019, 161, 293–303. [Google Scholar] [CrossRef]
  12. Zhou, C.P.; Yao, H.J.; Ma, J.; Chen, X.W. The study of theoretical method for improving the spatial resolution of satellite images with CCD cameras. Int. Symp. Remote Sens. 2000, 16, 543–547. [Google Scholar]
  13. Zhou, C.P.; Tian, Y.; Ji, T.K.; Wu, S.L.; Zhang, F.P.; Xu, D.; Shi, C.Y.; Yao, Y.H. The study of method for improving the spatial resolution of satellite images with CCD cameras. J. Remote Sens. 2002, 179–182. [Google Scholar] [CrossRef]
  14. Zhou, F.; Wang, H.Y.; Ma, W.P.; Liu, Z.J.; Zhou, C.P.; Chen, Q.; Ou Yang, X.L.; Jiang, L.; Xia, D.S. Study on a new method called tilting mode sampling for EO sensor. Spacecr. Recovery Remote Sens. 2005, 26, 47–51. [Google Scholar]
  15. Aizawa, K.; Komatsu, T.; Saito, T. A scheme for acquiring very high resolution images using multiple cameras. In Proceedings of the ICASSP-92: 1992 IEEE International Conference on Acoustics, Speech, and Signal Processing, San Francisco, CA, USA, 23–26 March 1992. [Google Scholar]
  16. Gillette, J.C.; Stadtmiller, T.M.; Hardie, R.C. Aliasing reduction in staring infrared imagers utilizing subpixel techniques. Opt. Eng. 1995, 34, 3130–3137. [Google Scholar] [CrossRef]
  17. An, B.W.; Xue, B.B.; Pan, S.D.; Chen, G.L. Sub-pixel processing for super-resolution scanning imaging system with fiber bundle coupling. Chin. Opt. Lett. 2011, 9, 081001. [Google Scholar]
  18. Wu, X.-S.; Deng, F.-Y.; Chen, M.; Cai, Y. Development of rotary infrared micro-scanner. J. Infrared Millim. Terahertz Waves 2012, 30, 263–267. [Google Scholar] [CrossRef]
  19. Sun, M.; Yu, K. A sur-pixel scan method for super-resolution reconstruction. Optik 2013, 124, 6905–6909. [Google Scholar] [CrossRef]
  20. Li, X. An improved two-stage image registration algorithm for super-resolution. IEEJ Trans. Electr. Electron. Eng. 2014, 9, 415–420. [Google Scholar] [CrossRef]
  21. Zhao, Y.; Chen, Q.; Sui, X.; Gao, H. Super Resolution Imaging Based on a Dynamic Single Pixel Camera. IEEE Photonics J. 2017, 9, 1–11. [Google Scholar] [CrossRef]
  22. Zhang, X.F.; Huang, W.; Xu, M.F.; Jia, S.Q.; Xu, X.R.; Li, F.B.; Zheng, Y.D. Super-resolution imaging for infrared micro-scanning optical system. Opt. Express 2019, 27, 7719–7737. [Google Scholar] [CrossRef] [PubMed]
  23. Wang, Z.; Cao, J.; Hao, Q.; Zhang, F.; Cheng, Y.; Kong, X. Super-resolution imaging and field of view extension using a single camera with Risley prisms. Rev. Sci. Instrum. 2019, 90, 033701. [Google Scholar] [CrossRef] [PubMed]
  24. Li, F.; Xin, L.; Guo, Y.; Gao, J.; Jia, X. A Framework of Mixed Sparse Representations for Remote Sensing Images. IEEE Trans. Geosci. Remote. Sens. 2016, 55, 1–12. [Google Scholar] [CrossRef]
  25. Li, F.; Xin, L.; Guo, Y.; Gao, D.; Kong, X.; Jia, X. Super-Resolution for GaoFen-4 Remote Sensing Images. IEEE Geosci. Remote. Sens. Lett. 2018, 15, 28–32. [Google Scholar] [CrossRef]
Figure 1. The diagram of MS-CMOS structure.
Figure 1. The diagram of MS-CMOS structure.
Sensors 20 04019 g001
Figure 2. The diagram of oblique-mode sampling.
Figure 2. The diagram of oblique-mode sampling.
Sensors 20 04019 g002
Figure 3. The flow chart of the imaging simulation mode.
Figure 3. The flow chart of the imaging simulation mode.
Sensors 20 04019 g003
Figure 4. The diagram of four points.
Figure 4. The diagram of four points.
Sensors 20 04019 g004
Figure 5. The diagram of 2 rows 26.56 ° ( arctan 1 2 ) oblique sampling mode.
Figure 5. The diagram of 2 rows 26.56 ° ( arctan 1 2 ) oblique sampling mode.
Sensors 20 04019 g005
Figure 6. The original high-resolution images: (a) seashore remote-sensing image, the chosen area is an offshore structure in an Egyptian port; (b) farmland remote-sensing image, the selected area is the crops and houses in Xinjiang, China; (c) city remote-sensing image, the selected area is buildings in Sydney; (d) target image, the selected area is minimum resolvable line pairs.
Figure 6. The original high-resolution images: (a) seashore remote-sensing image, the chosen area is an offshore structure in an Egyptian port; (b) farmland remote-sensing image, the selected area is the crops and houses in Xinjiang, China; (c) city remote-sensing image, the selected area is buildings in Sydney; (d) target image, the selected area is minimum resolvable line pairs.
Sensors 20 04019 g006
Figure 7. CMOS1 oblique-mode sampling simulation images: (a) seashore remote-sensing image, the chosen area is an offshore structure in an Egyptian port; (b) farmland remote-sensing image, the selected area is the crops and houses in Xinjiang, China; (c) city remote-sensing image, the selected area is buildings in Sydney; (d) target image, the selected area is minimum resolvable line pairs.
Figure 7. CMOS1 oblique-mode sampling simulation images: (a) seashore remote-sensing image, the chosen area is an offshore structure in an Egyptian port; (b) farmland remote-sensing image, the selected area is the crops and houses in Xinjiang, China; (c) city remote-sensing image, the selected area is buildings in Sydney; (d) target image, the selected area is minimum resolvable line pairs.
Sensors 20 04019 g007
Figure 8. CMOS2 oblique-mode sampling simulation images: (a) seashore remote-sensing image; (b) farmland remote-sensing image; (c) city remote-sensing image; (d) target image.
Figure 8. CMOS2 oblique-mode sampling simulation images: (a) seashore remote-sensing image; (b) farmland remote-sensing image; (c) city remote-sensing image; (d) target image.
Sensors 20 04019 g008
Figure 9. MS-CMOS conventional mode sampling simulation images: (a) seashore remote-sensing image, the chosen area is an offshore structure in an Egyptian port; (b) farmland remote-sensing image, the selected area is the crops and houses in Xinjiang, China; (c) city remote-sensing image, the selected area is buildings in Sydney; (d) target image, the selected area is minimum resolvable line pairs.
Figure 9. MS-CMOS conventional mode sampling simulation images: (a) seashore remote-sensing image, the chosen area is an offshore structure in an Egyptian port; (b) farmland remote-sensing image, the selected area is the crops and houses in Xinjiang, China; (c) city remote-sensing image, the selected area is buildings in Sydney; (d) target image, the selected area is minimum resolvable line pairs.
Sensors 20 04019 g009
Figure 10. The seashore remote-sensing image experimental results: super-resolution images, the chosen area is an offshore structure in an Egyptian port.
Figure 10. The seashore remote-sensing image experimental results: super-resolution images, the chosen area is an offshore structure in an Egyptian port.
Sensors 20 04019 g010
Figure 11. The farmland remote-sensing image experimental results: super-resolution images, the selected area is the crops and houses in Xinjiang, China.
Figure 11. The farmland remote-sensing image experimental results: super-resolution images, the selected area is the crops and houses in Xinjiang, China.
Sensors 20 04019 g011
Figure 12. The city remote-sensing image experimental results: super-resolution images, the selected area is buildings in Sydney.
Figure 12. The city remote-sensing image experimental results: super-resolution images, the selected area is buildings in Sydney.
Sensors 20 04019 g012
Figure 13. The target remote-sensing image experimental results: super-resolution images, the selected area is minimum resolvable line pairs.
Figure 13. The target remote-sensing image experimental results: super-resolution images, the selected area is minimum resolvable line pairs.
Sensors 20 04019 g013
Figure 14. The local magnification of the seashore remote-sensing image: (a) super-resolution image, (b) conventional mode simulation image, (c) CMOS1 oblique-mode simulation images, (d) original high-resolution image.
Figure 14. The local magnification of the seashore remote-sensing image: (a) super-resolution image, (b) conventional mode simulation image, (c) CMOS1 oblique-mode simulation images, (d) original high-resolution image.
Sensors 20 04019 g014
Figure 15. The local magnification of the farmland remote-sensing image: (a) super-resolution image, (b) conventional mode simulation image, (c) CMOS1 oblique-mode simulation images, (d) original high-resolution image.
Figure 15. The local magnification of the farmland remote-sensing image: (a) super-resolution image, (b) conventional mode simulation image, (c) CMOS1 oblique-mode simulation images, (d) original high-resolution image.
Sensors 20 04019 g015
Figure 16. The local magnification of the city remote-sensing image: (a) super-resolution image, (b) conventional mode simulation image, (c) CMOS1 oblique-mode simulation images, (d) original high-resolution image.
Figure 16. The local magnification of the city remote-sensing image: (a) super-resolution image, (b) conventional mode simulation image, (c) CMOS1 oblique-mode simulation images, (d) original high-resolution image.
Sensors 20 04019 g016
Figure 17. The local magnification of the target remote-sensing image: (a) super-resolution image, (b) conventional mode simulation image, (c) CMOS1 oblique-mode simulation images, (d) original high-resolution image.
Figure 17. The local magnification of the target remote-sensing image: (a) super-resolution image, (b) conventional mode simulation image, (c) CMOS1 oblique-mode simulation images, (d) original high-resolution image.
Sensors 20 04019 g017
Table 1. Comparison of clockwise rotation resolution improvement times.
Table 1. Comparison of clockwise rotation resolution improvement times.
AngleTimes of Lateral Resolution ImprovementTimes of Vertical Resolution ImprovementTimes of Image Resolution Improvement
2.112.112.11
10°1.901.821.86
20°1.671.821.74
30°1.742.111.89
40°1.501.701.59
50°1.502.251.77
60°1.462.111.70
70°1.292.381.62
80°1.732.381.98
90°1.902.382.10
100°1.902.001.95
110°2.001.891.95
120°2.131.451.72
130°2.251.501.77
140°
150°
160°
1.78
2.25
2.00
1.33 1.38 1.38 1.51
1.67
1.61
170°1.581.501.54
180°2.111.631.84
Average1.831.851.78
Table 2. Comparison of counterclockwise rotation resolution improvement times.
Table 2. Comparison of counterclockwise rotation resolution improvement times.
AngleTimes of Lateral Resolution ImprovementTimes of Vertical Resolution ImprovementTimes of Image Resolution Improvement
2.112.112.11
−10°2.111.821.94
−20°2.501.752.01
−30°2.381.381.71
−40°2.001.311.57
−50°2.111.361.61
−60°2.111.551.79
−70°2.252.382.31
−80°2.001.901.95
−90°1.801.901.85
−100°1.732.001.84
−110°1.582.251.81
−120°
−130°
−140°
−150°
1.38 1.21 1.21 1.29 2.25
1.80
1.78
1.80
1.67
1.44
1.41
1.48
−160°1.902.001.95
−170°1.901.801.85
−180°2.111.641.84
Average1.881.831.80
Table 3. Algorithm running time.
Table 3. Algorithm running time.
NameImage SizeTime (s)
Seashore image4142 × 5661 (8.03 M)0.852
Farmland image2628 × 3240 (3.28 M)0.679
City image2928 × 3762 (6.20 M)0.748
Target image1024 × 1024 (1.00 M)0.584
Average4.63 M0.713

Share and Cite

MDPI and ACS Style

Zhang, K.; Yang, C.; Li, X.; Zhou, C.; Zhong, R. High-Efficiency Microsatellite-Using Super-Resolution Algorithm Based on the Multi-Modality Super-CMOS Sensor. Sensors 2020, 20, 4019. https://doi.org/10.3390/s20144019

AMA Style

Zhang K, Yang C, Li X, Zhou C, Zhong R. High-Efficiency Microsatellite-Using Super-Resolution Algorithm Based on the Multi-Modality Super-CMOS Sensor. Sensors. 2020; 20(14):4019. https://doi.org/10.3390/s20144019

Chicago/Turabian Style

Zhang, Ke, Cankun Yang, Xiaojuan Li, Chunping Zhou, and Ruofei Zhong. 2020. "High-Efficiency Microsatellite-Using Super-Resolution Algorithm Based on the Multi-Modality Super-CMOS Sensor" Sensors 20, no. 14: 4019. https://doi.org/10.3390/s20144019

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop