Next Article in Journal
Satellite Measured Ionospheric Magnetic Field Variations over Natural Hazards Sites
Previous Article in Journal
Effect of Lunar Complex Illumination on In Situ Measurements Obtained Using Visible and Near-Infrared Imaging Spectrometer of Chang’E-4
Previous Article in Special Issue
Ex Post Analysis of Water Supply Demand in an Agricultural Basin by Multi-Source Data Integration
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Fusion of China ZY-1 02D Hyperspectral Data and Multispectral Data: Which Methods Should Be Used?

1
College of Resource Environment and Tourism, Capital Normal University, Beijing 100048, China
2
College of Geospatial Information Science and Technology, Capital Normal University, Beijing 100048, China
3
Key Laboratory of 3D Information Acquisition and Application, Capital Normal University, Beijing 100048, China
4
Logistics Support Department, Capital Normal University, Beijing 100048, China
*
Author to whom correspondence should be addressed.
Co-first author.
Remote Sens. 2021, 13(12), 2354; https://doi.org/10.3390/rs13122354
Submission received: 12 May 2021 / Revised: 14 June 2021 / Accepted: 14 June 2021 / Published: 16 June 2021

Abstract

:
ZY-1 02D is China’s first civil hyperspectral (HS) operational satellite, developed independently and successfully launched in 2019. It can collect HS data with a spatial resolution of 30 m, 166 spectral bands, a spectral range of 400~2500 nm, and a swath width of 60 km. Its competitive advantages over other on-orbit or planned satellites are its high spectral resolution and large swath width. Unfortunately, the relatively low spatial resolution may limit its applications. As a result, fusing ZY-1 02D HS data with high-spatial-resolution multispectral (MS) data is required to improve spatial resolution while maintaining spectral fidelity. This paper conducted a comprehensive evaluation study on the fusion of ZY-1 02D HS data with ZY-1 02D MS data (10-m spatial resolution), based on visual interpretation and quantitative metrics. Datasets from Hebei, China, were used in this experiment, and the performances of six common data fusion methods, namely Gram-Schmidt (GS), High Pass Filter (HPF), Nearest-Neighbor Diffusion (NND), Modified Intensity-Hue-Saturation (IHS), Wavelet Transform (Wavelet), and Color Normalized Sharping (Brovey), were compared. The experimental results show that: (1) HPF and GS methods are better suited for the fusion of ZY-1 02D HS Data and MS Data, (2) IHS and Brovey methods can well improve the spatial resolution of ZY-1 02D HS data but introduce spectral distortion, and (3) Wavelet and NND results have high spectral fidelity but poor spatial detail representation. The findings of this study could serve as a good reference for the practical application of ZY-1 02D HS data fusion.

Graphical Abstract

1. Introduction

The ZY-1 02D Satellite, also known as a 5-m optical satellite, is the first operational civil hyperspectral (HS) satellite, independently developed and successfully operated by China as the China–Brazil Earth Resources Satellite. It was launched on 12 September 2019, and one of the three main payloads is an advanced HS imager developed by the Shanghai Institute of Technical Physics (SITP), Chinese Academy of Sciences. The ZY-1 02D HS imager has 166 spectral bands ranging from 400 nm to 2500 nm, a spatial resolution of 30 m, and a swath width of 60 km, allowing it to provide detailed spectral information about ground features. Compared with other on-orbit or planned satellites (e.g., Environmental Mapping and Analysis Program, Precursore Iperspettrale della Missione Applicativa, and DLR Earth Sensing Imaging Spectrometer), the spectral resolution and the swath width of the ZY-1 02D are more advantageous. Unfortunately, due to the unavoidable trade-offs between spectral resolution, spatial resolution and signal-to-noise ratio [1,2], the ZY-1 02D HS data cannot directly make full use of its advantages in some specific applications [3,4,5,6,7]. In this case, a simple and feasible solution that improves spatial resolution while preserving spectral fidelity is required. Nowadays, data fusion has become an important data processing method for achieving the aforementioned goal, as it can fuse HS data with MS data to combine the benefits of both [8]. Given the availability of ZY-1 02D MS data and the benefits of imaging under the same conditions as HS data, this paper focuses on fusing ZY-1 02D HS data with ZY-1 02D MS data to fully excavate and effectively use its internal information [9,10].
In recent years, with the development of high-precision quantitative remote sensing applications, a large amount of remote sensing data with both high spatial resolution and high spectral resolution has become urgently needed [11,12,13,14]. To address this issue, domestic and international researchers have conducted a lot of research in the hopes of improving spatial resolution and enriching information while preserving the spectral fidelity of the original image [15]. Data fusion technology can synthesize the effective information of various image data, eliminate the redundancy and contradictions between multiple sources of information, and produce a composite image with improved interpretability [16,17]. So far, several mature image fusion methods have been developed, including the Gram-Schmidt (GS) transform [18], Intensity-Hue-Saturation (IHS) [19], High-Pass Filter (HPF) [20], Nearest-Neighbor Diffusion (NND) [21], Wavelet Transform (Wavelet) [22], Principal Component Analysis (PCA) [23], Color Normalized Spectral Sharpening (CNSS) [24] and Color Normalized Sharping (Brovey) [25], etc. To evaluate the quality of the fused result, statistical indicators (mean, standard deviation, and mean gradient, etc.) or ground object classification accuracy, and target extraction accuracy are typically used [26,27,28,29]. Among the numerous studies, Huang et al. [30] compared the performance of HPF, IHS and super-resolution Bayesian (Pansharp) in the image fusion application of Mapping Satellite-1 (TH-1) and found that HPF is the best. Sun et al. [31] made a performance comparison of five different fusion methods to find a suitable method for GaoFen-2 (GF-2), and the results show that the fused images transformed by HCS and GS have good performance in both visual interpretation analysis and ground object classification. Du et al. [32] used four methods including Pansharp, GS and Wavelet to carry out data fusion on GaoFen-1 (GF-1), finding that the GS method could effectively improve the spatial resolution and enrich the texture information. Huang et al. [33] fused the ZiYuan-3 (ZY-3) satellite data with a variety of commonly used image fusion methods and evaluated the fusion results from both qualitative and quantitative aspects. Obviously, it can be seen that these fusion methods have good performance in high-spatial-resolution satellite data fusion. In addition, there have also been some studies considering the use of multi-source data fusion to explore the best fusion method for different data, such as Ren et al. [34], who conducted a comprehensive evaluation study on the fusion results of GF-5 HS data with three MS data (namely GF-1, GF-2 and Sentinel-2A), and the results showed that LANARAS, Adaptive Gram-Schmidt (GSA), and modulation transfer function (MTF)-generalized Laplacian pyramid (GLP) methods were more suitable for fusing GF-5 with GF-1 data, while MTF-GLP and GSA methods were recommended for fusing GF-5 with GF-2 data, and GSA and smoothing filtered-based intensity modulation (SFIM) could be used to fuse GF-5 with S2A data. Ghimire et al. [35] also created an optimal image fusion and quality evaluation strategy for various satellite image data (GF-1, GF-2, GF-4, Landsat-8 OLI, and MODIS).
The preceding studies point us in the right direction for data fusion. It can be found that different fusion methods have different fusion performance when applied to remote sensing data. This disparity can be attributed to the characteristics of remote sensing data as well as the fusion method. As a result, in practice, it is critical to select the appropriate fusion method according to the image characteristics and application purpose [36]. Given the short launch time of the ZY-1 02D satellite, there are currently few related research and results regarding image fusion of ZY-1 02D data. Because the fusion methods suitable for existing satellite data are not necessarily suitable for the ZY-1 02D, it is critical to investigate its suitable fusion methods, and this is of great significance for future application and research using ZY-1 02D data. Furthermore, when compared to multi-source remote sensing data, ZY-1 02D provides HS and MS data under the same imaging conditions, reducing the uncertainty caused by input data. The fusion results of ZY-1 02D will have more application value. For these reasons, we will look into image fusion on ZY-1 02D HS and ZY-1 02D MS data. There are numerous fusion methods available for us to use in this research [37,38,39,40,41,42], but many of them are not user-friendly in terms of operability, computing resource requirements, and professional requirements. Taking into account the needs of users in practical applications, the focus of this research is to identify suitable methods for ZY-1 02D data among the existing well-known methods.
The objective of this paper is to find appropriate methods for fusing ZY-1 02D HS and ZY-1 02D MS data from among some well-known methods. To meet the main objective, we test six common image fusion methods (i.e., GS, HPF, IHS, Wavelet, NND, and Brovey) and use a comprehensive evaluation framework to evaluate their performance, including aspects of their visual interpretation and quantitative metrics. This study will serve as a reference for the choice of fusion methods for the ZY-1 02D data, thereby further broadening its application in a variety of fields.

2. Materials and Methods

2.1. ZY-1 02D Data

At 11:26 a.m. on 12 September 2019, the ZY-1 02D satellite was launched into the planned orbit by a Long March-4B carrier rocket from the Taiyuan Satellite Launch Center in China’s Shanxi Province. It is currently in a solar synchronous orbit 778 km above the earth, with a five-year expected lifespan. The two sensors on board enable it to effectively acquire 8-band MS data with a width of 115 km and 166-band HS data with a width of 60 km. Specifically, the spatial resolution of the MS data is 10 m, the HS data is 30 m, and the spectral resolution of the HS reaches 10 nm and 20 nm in the visible-near infrared (VIS-NIR) and short-wave infrared (SWI) ranges. The main parameters of the ZY-1 02D HS and MS sensors are shown in Table 1. Figure 1 shows the spectral response function of MS sensor.
For the experiment, the ZY-1 02D HS image and MS image covering Anxin County, Baoding City, Hebei Province, China were used (Figure 2). They were collected on 7 October 2020. There are several common surface types in the area, mainly artificial buildings, farmland, vegetation, and water.

2.2. Data Preprocessing

The preprocessing of the data was divided into two sections: (1) MS data preprocessing, and (2) HS data preprocessing. MS data preprocessing included image mosaicking, image clipping and image registration. First, the two MS images were mosaicked into one image, and then the part that overlapped with the HS image was clipped out (the size of the clipped image was 7670 × 8291 pixels). Then, the automatic registration tool in ENVI (i.e., Image Registration Workflow) was used to register it with the HS image (RMS Error: 0.31), and the processed MS image was used as input data for the fusion experiment. HS data preprocessing included radiometric calibration, atmospheric correction, and band extraction. Radiation calibration was used to covert the gray value (digital number, DN) into top of atmosphere (TOA) reflectance. FLAASH model was used for atmospheric correction. The atmospheric model was set to mid-latitude summer, and the aerosol model parameters were set to city. Since the spectral range of the HS image (396~2501 nm) was wider than that of the MS (452–1047 nm), a subset of the data spanning the range 452~1047 nm was extracted from the HS data, and the bands were grouped to match specific MS bands (Table 2), and the processed HS image was used as the input data of the fusion experiment. All of the preceding procedures were carried out using the ENVI 5.3 software.

2.3. Fusion Method

In this experiment, six commonly used fusion methods were compared, namely GS, HPF, NND, IHS, Wavelet, and Brovey. The methodological flowchart is shown in Figure 3. In the fusion experiment, 8 MS bands and their corresponding 75 HS bands (as shown in Table 2) were input into different algorithms, respectively.

2.3.1. Gram-Schmidt (GS)

GS transform can remove redundant information by converting an HS image to orthogonal space. The transformed components are orthogonal in the orthogonal space, and the degree of information retention varies little between them. Compared with PCA transform, this method avoids the problem of information over-concentration [43,44]. Its advantage is that the process is simple, there are no restrictions on the number of input bands, and the spectral information of the original low-spatial-resolution image can be well preserved. During the experiment, the HS data was input as a low-resolution image, and the MS data were input as a high-resolution image (the following methods also used the same setting). The resampling method was set to nearest neighbor. This method was implemented in ENVI 5.3.

2.3.2. High-Pass Filter (HPF)

The HPF fusion method extracts structural detail information from the high-spatial-resolution image using a high-pass filter operator and then superimposes the detail information on the low-spatial-resolution image to achieve a combination of the two [45]. The advantages of this method include a simple algorithm, a small amount of calculation, and no limit on the number of input bands. The keral size and weighting factor were set to 5 × 5 and the minimum value, respectively, while the other parameters remained unchanged. This method was implemented in ERDAS IMAGINE 2014.

2.3.3. Nearest-Neighbor Diffusion (NND)

NND was proposed by the Rochester Institute of Technology (RIT) in the United States, which uses the Nearest-Neighbor Diffusion pan-sharpening algorithm for fusion [21]. The principle is to first perform down-sampling processing on the high-spatial-resolution image to make the spatial resolution consistent with the low-spatial-resolution image data; then, the spectral band contribution vector is calculated via linear regression, obtaining the nearest super-pixel difference factor of each pixel in the original high-spatial-resolution image; and finally, the linear mixed model is used to obtain the fused image. It is characterized by fast operation speed and high spectral fidelity. This method was implemented in ENVI 5.3.

2.3.4. Modified Intensity-Hue-Saturation (IHS)

IHS is a color representation system that employs intensity, hue, and saturation. When using this method for image fusion, it is primarily divided into the following three steps: (1) Resampling the low-spatial-resolution image to match the spatial resolution of the high-spatial-resolution image, then converting it from RGB space to IHS space; (2) histogram matching the high-resolution image and the I component of the low-resolution image, then replacing the I component of the low-resolution image with the new luminance component; (3) inverse transformation of the above result and its restoration to RGB space. This method has been widely used because of its high spatial detail enhancement capabilities. Compared with the traditional IHS, modified IHS overcomes the limitation of three input bands by fusing multi-band data via multiple iterations. In the B03 band, for an example, a combination of 13-14-15 and 16-17-18 is used for two iterations of fusion to produce a fused result of 13-18. The resampling method was set to nearest neighbor, and the ratio celling was set to 2.0. Because there is no ZY-1 02D in the sensor options, the band information (including center wavelength, wavelength, etc.) was customized according to the provided data file. This method was implemented in ERDAS IMAGINE 2014.

2.3.5. Wavelet Transform (Wavelet)

Wavelet transform is a spatial signal decomposition and reconstruction fusion technology. Its basic principle is to perform wavelet forward transformation on a low-spatial-resolution image and a high-resolution image to obtain high-frequency information from the high-resolution image and low-frequency information from the low-resolution image, respectively, and then generate a fused image using inverse wavelet transformation [46]. The spectral transform and resampling method were set to single band and nearest neighbor, respectively. This method was implemented in ERDAS IMAGINE 2014.

2.3.6. Color Normalized Sharping (Brovey)

The Brovey fusion method first normalizes the high-spectral-resolution data, before multiplying it by the high-spatial-resolution image to obtain the fusion result. Each band in the RGB image is multiplied by the ratio of the high-resolution data to the RGB data, and the RGB image is then resampled to the high-resolution pixel size. The resampling method was set to nearest neighbor. This method was implemented in ERDAS IMAGINE 2014.

2.4. Quality Evaluation Methods

We evaluate the quality of fusion results using two criteria: qualitative evaluation (i.e., visual interpretation) and quantitative evaluation (i.e., statistical metrics).

2.4.1. Qualitative Evaluation

Visual interpretation is the method used in qualitative evaluation, and it refers to the observer’s subjective evaluation of the fusion result with respect to both the overall effect and the local effect via visual perception. Qualitative evaluation has become an indispensable part of the quality evaluation of remote sensing fusion images due to its quick and simple advantages.

2.4.2. Quantitative Evaluation

The use of various remote sensing image statistical metrics to evaluate the quality of the fusion results is referred to as quantitative evaluation. The advantages and disadvantages of various fusion methods can be discovered through statistics and analysis of the aforementioned metrics. In this experiment, five statistical metrics (i.e., mean, standard deviation, entropy, mean gradient, and correlation coefficient) were selected to quantitatively evaluate the fused results from the four aspects of brightness, clarity, information content, and spectral fidelity (Figure 4). The calculations of the five metrics are performed as follows.
The gray mean value is primarily used to describe the average brightness of the image. When the gray mean value of the fused image is close to that of the original multispectral image, it indicates that the fusion effect is good. It is defined as
Mean = 1 M × N i = 1 M j = 1 N I ( i , j )
where M and N are the total number of rows and columns of the image, i and j are the pixel positions, I ( i , j ) indicates the gray value of the pixel located in the i-th row and j-th column of the image.
The standard deviation is frequently used to describe the uniformity of image grayscale. The larger the standard deviation, the more dispersed the grayscale distribution of the image and the higher the image contrast. It is defined as
Std = 1 M × N i = 0 M 1 j = 0 N 1 ( I ( i , j ) I ) 2
where I represents the gray mean value of the image.
The entropy is an important indicator for measuring the richness of image information because it reflects the average information content in the image. It is defined as
  Entropy = P ( x i ) log ( 2 , p ( x i ) )
where x i represents the random variable, and P ( x i ) is the output probability function.
The mean gradient refers to the obvious difference in the gray scale near the border or both sides of the shadow line of the image, indicating that the gray scale change rate is high, and the magnitude of this change rate can be used to express the clarity of the image. It can sensitively reflect the rate at which the image expresses the contrast of small details and characterize the relative clarity and texture of the image. The larger the mean gradient, the clearer the image is. It is defined as
G = 1 ( M 1 ) ( N 1 ) i = 1 M j = 1 N ( ( Z ( x i , y j ) x i ) 2 + ( Z ( x i , y j ) y j ) 2 ) 2
where Z ( x i , y j ) x i represents the gradient in the horizontal direction, and Z ( x i , y j ) y j represents the gradient in the vertical direction.
The correlation coefficient indicates how similar the images were before and after fusion. A high correlation coefficient indicates that the fused image is close to the original image and has good spectral fidelity. It is defined as
C = i = 1 M j = 1 N [ R ( i , j ) F ˉ R ] [ F ( i , j ) F ˉ ] i = 1 M j = 1 N [ R ( i , j ) F ˉ R ] 2 i = 1 M j = 1 N [ F ( i , j ) F ˉ ] 2
In addition to the statistical metrics listed above, the spectral curve of typical ground objects can also be used to evaluate the quality of the fusion results [47]. In this experiment, the spectral curves of typical ground objects (vegetation, water, and artificial building) are extracted, and a comparison is drawn between the original HS image and the fused image, respectively, to quantify the benefits and drawbacks of the fusion methods.

3. Results

3.1. Qualitative Evaluation

Figure 5 shows false-color images of the original HS image, the six fused results (R: 954 nm; G: 765 nm; B: 482 nm), and the original MS image (R: B09; G: B08; B: B02). There are some color differences between different fused images and the original HS image from the perspective of the entire image, but they all retain the main spectral characteristics of the original HS image. The colors of the HPF and Wavelet images are the closest to the original HS image, and the tone of the two is lighter than that of the original HS image, with almost no difference between the HPF and the Wavelet images. When compared to the original HS image, the color of the NND and GS images is more orange, and the contrast between adjacent objects in the two is not as clear, indicating that the NND and GS fusion methods perform worse than the above two methods. The IHS image and the original HS image have a distinct spectral difference, which is reflected in the darker tone of the IHS image as well as the spectral distortion phenomenon. Nevertheless, IHS images possess clarity. The color deviation between the Brovey image and the original HS image is the greatest, indicating that the Brovey image has significant spectral distortion.
To further evaluate the visual interpretation effect of the fused images, especially the enhancement effect of the spatial details, Figure 6 and Figure 7 show the detail of several typical ground objects, namely artificial buildings, farmland, vegetation and water.
When the edge and texture differences between the original HS image and fused image are compared, it can be found that, with the exception of the Wavelet image, the clarity of the other five fused images is higher than that of the original HS image, indicating that the above five fusion methods are able to improve the spatial resolution of the original HS image, thereby improving the accuracy and reliability of the visual interpretation. The IHS image has the best clarity and good spatial sharpening effect. The boundaries between building, road, and farmland are the clearest, and the outlines of vegetation and artificial fences in the water are the most visible, indicating that the IHS fusion method improves the spatial details of the original image the most. Unfortunately, the IHS image contains some spectral distortions, resulting in significant color differences between the IHS image and the original HS image. HPF and GS fusion methods are second only to IHS in terms of spatial detail enhancement. Specifically, the contours of aquatic vegetation are more visible in the HPF fused image, and the details of some buildings and farmland features are blurred. The detailed spatial information of the Brovey fused image is slightly lost, which is reflected in the fact that the artificial fence edge in the water is difficult to identify, and its spectral distortion is more pronounced. The spatial resolutions of the NND and Wavelet fused images are low, and the visual interpretation effect is not optimal. Between them, the Wavelet image has the closest color match to the original HS image, but the effect of its spatial detail representation is poor, and there are some obvious distortions and unclear texture features in the buildings.

3.2. Quantitative Evaluation

3.2.1. Statistical Metrics

Considering a large number of HS bands, eight HS bands are quantitatively accounted for this experiment, corresponding to the center wavelength positions of the eight MS bands. Table 3 shows the mean, standard deviation, entropy, mean gradient, and correlation coefficient of the original HS image and six fused images.
By comparing the statistical metric values of the six fused images in Table 3 with the original HS image, the following five results can be obtained.
(1)
The HPF and GS fused images (gray mean value of the HPF image: 101~160; GS: 96~154) had high brightness similarity with the original HS image (100~155); the NND image (103~147) and Wavelet image (99~166) were second to them; the IHS image (73~131) had some spectral distortion when compared to the original HS image; the mean value of the Brovey image (29~49) was much lower than that of the original HS image, the mean value between the two had the largest deviation, and the spectral distortion of the Brovey image was the most significant.
(2)
The standard deviations of HPF, IHS and GS fused images were relatively large. Among them, the HPF image had the largest standard deviation in the visible light (standard deviation of the HPF fused image: 23~27), and the IHS image had the biggest standard deviations in the near-infrared and red edge bands (IHS image: 22~24), and the standard deviations of the three fused images were higher than the original HS image, greatly improving the information richness of the original HS image; the standard deviations of the NND and Wavelet images were slightly lower than those of the above three fused images. The standard deviation of the Wavelet image (16~21) was slightly higher than that of the NND (12~22); the standard deviation of the Brovey image (6~8) was much lower than that of the original image, indicating that the gray level of the fused image after Brovey transformation was not sufficiently dispersed and the tone tended to be single.
(3)
Except for the Brovey image, the entropy of the other five fused images was improved when compared to the original HS image, and there was no significant difference in the entropy of different fusion images (5–7). Among them, the entropy of the Wavelet image was the highest, followed by the HPF and GS images. The information content of the three was nearly identical, with the exception of the IHS, which was slightly lower; the information content of Brovey fusion images was lower than that of the original HS images.
(4)
The mean gradient of the fused image was higher than that of the original HS image, indicating that the six fusion methods were able to improve the original image’s ability to represent spatial details. The mean gradients of HPF and IHS fused images were larger, and their spatial detail information enhancement effects were the best, as shown in Table 3. Between them, the mean gradient of HPF image was more visible in the visible light band (mean gradient of HPF fusion image: 47~61). The IHS fusion image outperformed the other two fusion methods in the red edge and near-infrared bands (55~57); the mean gradient of GS and Brovey fused images were slightly lower than the above two fused images. Between them, the mean gradient of Brovey image (43~57) was slightly higher than GS image (41~58); the mean gradient of NND and Wavelet fused was the lowest, indicating that the spatial resolution of these two fused images was low.
(5)
Similar to the visual interpretation result, the correlation coefficients of Wavelet, HPF and NND were relatively large, indicating that they have excellent spectral fidelity performance. Except for the red band, the correlation coefficients of Wavelet and HPF fused images were all greater than 0.9; the correlation coefficient between the IHS fusion image and the original image was second only to HPF and Wavelet, but it performed poorly in the blue and red bands; and the Brovey fused image had the lowest correlation coefficient, indicating that there was a large spectral difference between it and the original HS image.
To summarize, the HPF image not only preserved the spectral characteristics of the original HS image to a large extent, but also improved the spatial resolution of the HS data, making it the best choice for ZY-1 02D HS data enhancement; although Wavelet and NND fused images had a large amount of information and a high spectral fidelity, the spatial detail representation effect of the two was not satisfactory; the IHS fused image was clear and rich in texture information, but there was obvious spectral distortion; the performance of the GS fusion image was at a medium level; the Brovey transform had a mediocre performance in spectral fidelity and spatial detail representation, and had a serious spectral distortion problem, making it far inferior to the above five methods.

3.2.2. The Spectral Curves of Typical Ground Objects

To more intuitively evaluate the spectral fidelity of the different fused images, this experiment takes typical objects (i.e., vegetation, artificial building, and water) as examples to compare the differences between the six fused results and the original HS image. To ensure that the results are representative, the spectral curves of the three objects are averaged over three homogeneous areas with 5 × 5 pixels. As shown in Figure 8, the left column is the three spectral curves of six fused images and HS image. In addition, the spectral differences between the six fused images and the original HS image are also shown (the right column of Figure 8), which is convenient for identifying the differences in the spectral fidelity of the different fusion methods.
By comparing the spectral curves of different fused images and the original HS image, we can obtain similar results to those presented above. Overall, the HPF, Wavelet, NND and GS fused images possessed high spectral fidelity; while the IHS and Brovey fused images exhibited significant spectral distortion. The spectral curves of the fused images have roughly the same trend as the original HS image and fluctuate around the original HS spectral curve with slight variations. In the spectral curves of vegetation, the difference between the Wavelet and HPF fused image and the original HS image was the smallest, indicating that the two fusion methods had the best spectral fidelity. Between them, the Wavelet fused image was more prominent in the visible light band, while the HPF fused image was more prominent in the visible light and near-infrared (NIR) bands; the spectral fidelity of the NND and GS fused images were second, with a lager difference between their spectral curves. In the spectral curves of the building, the Wavelet fused image had a higher degree of overlap and a smaller difference with the original HS image; the spectral curves of HPF and GS fused images basically fitted the original HS image after about 510 nm; and the performance of NND was quite different in the NIR band (approximately 765 nm). The curve in the second half of the spectral curve was very similar to that of the original HS image, whereas the curve in the first half was quite different from the gray value of the original HS image. This result is consistent with the color deviation of the building in the visual interpretation effect. In the spectral curves of water, the NND and the original HS image were highly overlapped; the difference between Wavelet and GS was slightly larger; the change trend of the HPF fused image was the same as the original image, but compared to the above fused methods, there was a larger difference between HPF and the original HS, which is slightly different from the high correlation coefficient obtained for the HPF method in the metric statistical results. This may have been caused by the small amount of water in the entire image. Similar to the visual interpretation results, the IHS and Brovey fused images were significantly different from the original HS image with respect to the spectral curves of the three ground objects. There were obvious spectral distortions in the two fused images, particularly the Brovey image, which had the lowest spectral fidelity.
In conclusion, the HPF, Wavelet, NND, and GS fusion methods performed well, and different fusion methods performed differently when confronted with different ground objects. HPF and Wavelet had the best spectral fidelity in the vegetation and building areas, respectively, while NND had the best spectral fidelity in the water area. It can be found that the HPF, Wavelet, NND and GS fusion methods have good performance, and different fusion methods have different performance when facing different ground objects. Specifically, HPF and Wavelet have the best spectral fidelity in vegetation and building areas, and NND has the best spectral fidelity in the water area.

4. Discussion

4.1. Performance of Fusion Methods

In this paper, six well-known fusion methods were successfully used to enhance the spatial resolution and the information of ZY-1 02D data. Compared to the original HS data, the six fused images were clearer and easier to visually interpret. Furthermore, there were also some noticeable differences among the six fused results. HPF, in particular, demonstrated excellent performance with respect to both spectral fidelity and spatial resolution enhancement. The reason for this could be that the HPF injects the spatial details of the MS data into the HS data via a high-pass filter and has a low-pass filter to maintain the spectral separation of HS data. This result was also discovered in the fusion experiment of Mapping Satellite-1 by Huang et al. [30]. The difference is that the research object of this study is HS data with high spectral resolution, which are better suited to narrow-band spectroscopy research. Aside from HPF, Wavelet and NND exhibited high robustness in terms of spectral fidelity. Between them, the spatial resolution of the Wavelet fused image was slightly lower, and the boundary between different features was not significant, but it was always clearer than the original HS data. As concluded by Sun et al. [31] and Du et al. [32], GS performs well when applied to remote sensing data. On the contrary, the spectral fidelity performances of IHS and Brovey were slightly inferior, and there is still room for improvement in spatial resolution [48,49]. The spectral distortion of IHS is mainly caused by the forced direct replacement of the I component [50], whereas the spectral distortion of Brovey was affected by the simple multiplication of HS and MS [25]. It is worth noting that these two methods are simple in terms of theory, simple to implement, and quick to calculate, and so can be used for some low-demand applications.

4.2. Selection of Quantitative Metrics

The quality evaluation of the fused results in remote sensing data fusion research is too complicated to establish a unified standard. Until now, it has been performed using a combination of visual interpretation and quantitative metrics [51]. As a result, we used the same method to evaluate the quality of six fused images. The five quantitative metrics selected for this paper adhered to three principles: (1) they were able to evaluate the quality of fused images with respect to different aspects (typically spatial resolution enhancement capability and spectral fidelity); (2) they were able to distinguish the performance differences of the six fusion methods; and (3) they were simple and easy to calculate. The comparison results also demonstrate the applicability of the five metrics. In addition to the common metrics listed above, many researchers have attempted to supplement the quality evaluation system by improving or proposing new metrics [52,53,54]. There is still a long way to go before we establish a unified evaluation standard. In practical applications, appropriate quantitative metrics should be selected based on geographic conditions, application requirements, and data sources.

4.3. Limitations

The study was carried out on a single date in a single study area with four typical surface types; it would be preferable if different study areas with more surface types in different periods were used [55]. Because the spectral range of the HS image is wider than that of the MS, this study only performed fusion processing on the HS data in the 452~1047 nm range. In subsequent research, we will fully exploit the advantages of the wide spectral range to investigate suitable fusion methods for shortwave of infrared band, as well as extending the practical application of ZY-1 02D data. In reality, the primary purpose of data fusion is to prepare for remote sensing image applications such as image classification [56], change detection [57], and so on. As a result, more research into the application of fused images is required in the future to obtain better results and conclusions.

5. Conclusions

Six fusion methods, GS, HPF, IHS, NND, Wavelet, and Brovey, were used in this paper to realize the fusion of ZY-1 02D HS data and MS data. The six fusion results were compared and analyzed using a combination of qualitative and quantitative evaluation methods, and the following conclusions were drawn: (1) Considering the three aspects of visual effect, spectral fidelity, and spatial detail expression, the HPF method was the most suitable for the fusion of ZY-1 02D HS data and MS data. In comparison to the original HS image, the HPF fused image maintained its spectral characteristics while improving its spatial resolution, enriching its information, and providing the best fusion performance. (2) Different fusion methods perform differently for different datasets. In practice, appropriate data fusion methods should be selected according to the data type and specific needs. Six commonly used fusion methods were used in this study for fusion processing of ZY-1 02D satellite HS and MS images, providing a significant reference for future ZY-1 02D data processing and application-related research.

Author Contributions

H.L., D.Q. and L.D. designed and developed the research idea. H.L., D.Q. and S.W. conducted the field data collection. H.L., D.Q., Y.L. and S.W. processed all remaining data. H.L. and D.Q. performed the data analysis and wrote the manuscript. H.L., D.Q., Y.L., S.W. and L.D. contributed to result and data interpretation, discussion, and revision of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the landscape plant maintenance monitoring and intelligent diagnosis model development of Capital Normal University (no. 21220030003).

Acknowledgments

The authors are very thankful for Zou Hanyue, Fan Tianxing and Chen Yong for their valuable support.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Zhang, L.; Shen, H. Progress and future of remote sensing data fusion. J. Remote Sens. 2016, 20, 1050–1061. [Google Scholar]
  2. Xie, Q.; Zhou, M.; Zhao, Q.; Meng, D.; Zuo, W.; Xu, Z. Multispectral and Hyperspectral Image Fusion by MS/HS Fusion Net. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 1585–1594. [Google Scholar]
  3. Ma, Y.; Zhang, Y.; Mei, X.; Dai, X.; Ma, J. Multifeature-Based Discriminative Label Consistent K-SVD for Hyperspectral Image Classification. IEEE J. Stars. 2019, 12, 4995–5008. [Google Scholar] [CrossRef]
  4. Qu, Y.; Qi, H.; Ayhan, B.; Kwan, C.; Kidd, R. DOES multispectral/hyperspectral pansharpening improve the performance of anomaly detection? In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Fort Worth, TX, USA, 23–28 July 2017; pp. 6130–6133. [Google Scholar]
  5. Li, W.; Wu, G.; Zhang, F.; Du, Q. Hyperspectral Image Classification Using Deep Pixel-Pair Features. IEEE Trans. Geosci. Remote Sens. 2017, 55, 844–853. [Google Scholar] [CrossRef]
  6. Ferraris, V.; Dobigeon, N.; Qi, W.; Chabert, M. Robust Fusion of Multi-Band Images with Different Spatial and Spectral Resolutions for Change Detection. IEEE Trans. Comput. Imaging 2017, 3, 175–186. [Google Scholar] [CrossRef] [Green Version]
  7. Gómez-Chova, L.; Tuia, D.; Moser, G.; Camps-Valls, G. Multimodal classification of remote sensing images: A review and future directions. Proc. IEEE 2015, 103, 1560–1584. [Google Scholar] [CrossRef]
  8. Fauvel, M.; Tarabalka, Y.; Benediktsson, J.A.; Chanussot, J.; Tilton, J.C. Advances in Spectral-Spatial Classification of Hyperspectral Images. Proc. IEEE 2013, 101, 652–675. [Google Scholar] [CrossRef] [Green Version]
  9. Ghassemian, H. A review of remote sensing image fusion methods. Inf. Fusion 2016, 32, 75–89. [Google Scholar] [CrossRef]
  10. Pandit, V.R.; Bhiwani, R.J. Image Fusion in Remote Sensing Applications: A Review. Int. J. Comput. Appl. 2015, 120, 22–32. [Google Scholar]
  11. Huang, B.; Zhao, Y. Research Status and Prospect of Spatiotemporal Fusion of Multi-source Satellite Remote Sensing Imagery. Acta Geod. Cartogr. Sin. 2017, 46, 1492–1499. [Google Scholar]
  12. Wang, H.; Peng, J.; Wu, W. Remote Sensing Image Fusion Using Wavelet Packet Transform. J. Image Graph. 2002, 9, 68–73. [Google Scholar]
  13. Zhou, F.; Hang, R.; Liu, Q.; Yuan, X. Pyramid Fully Convolutional Network for Hyperspectral and Multispectral Image Fusion. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2019, 12, 1–10. [Google Scholar] [CrossRef]
  14. Yokoya, N.; Grohnfeldt, C.; Chanussot, J. Hyperspectral and Multispectral Data Fusion: A comparative review of the recent literature. IEEE Geosci. Remote Sens. Mag. 2017, 5, 29–56. [Google Scholar] [CrossRef]
  15. Shao, Y.; Zhu, C.; Zhang, X.; Shen, Q. Comparison of diffirent fusion methods and their performance evaluation to high spatial resolution remote sensing data of GF. Bulletin Surv. Mapp. 2019, 5, 5–10. [Google Scholar]
  16. Wang, G.; Li, Y.; Zeng, Y.; Jin, L. Comparison and analysis is of pixel level image fusion algorithms application to ALOS data. Sci. Surv. Mapp. 2008, 33, 121–124. [Google Scholar]
  17. Kaczynski, R.; Donnay, J.P.; Muller, F. Satellite image maps of Warsaw in the scale 1:25,000. Earsel Adv. Remote Sens. 1995, 4, 100–103. [Google Scholar]
  18. Aiazzi, B.; Baronti, S.; Selva, M. Improving Component Substitution Pansharpening Through Multivariate Regression of MS +Pan Data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  19. Tu, T.; Huang, P.S.; Hung, C.; Chang, C. A fast intensity-hue-saturation fusion technique with spectral adjustment for IKONOS imagery. IEEE Geosci. Remote Sens. 2004, 1, 309–312. [Google Scholar] [CrossRef]
  20. Metwalli, M.R.; Nasr, A.H.; Allah, O.S.F.; El-Rabaie, S. Image fusion based on principal component analysis and high-pass filter. In Proceedings of the 2009 International Conference on Computer Engineering & Systems, Cairo, Egypt, 14–16 December 2009; pp. 63–70. [Google Scholar]
  21. Sun, W.; Chen, B.; Messinger, D.W. Nearest-neighbor diffusion-based pan-sharpening algorithm for spectral images. Opt. Eng. 2013, 53, 13107. [Google Scholar] [CrossRef] [Green Version]
  22. Zhou, J.; Civco, D.L.; Silander, J.A. A wavelet transform method to merge Landsat TM and SPOT panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  23. Shah, V.P.; Younan, N.H.; King, R.L. An Efficient Pan-Sharpening Method via a Combined Adaptive PCA Approach and Contourlets. IEEE Trans. Geosci. Remote 2008, 46, 1323–1335. [Google Scholar] [CrossRef]
  24. Klonus, S.; Ehlers, M. Image Fusion Using the Ehlers Spectral Characteristics Preserving Algorithm. GISci. Remote Sens. 2007, 44, 93–116. [Google Scholar] [CrossRef]
  25. Tu, T.; Lee, Y.C.; Chang, C.P.; Huang, P.S. Adjustable intensity-hue-saturation and Brovey transform fusion technique for IKONOS/QuickBird imagery. Opt. Eng. 2005, 44, 116201. [Google Scholar] [CrossRef] [Green Version]
  26. Zeng, Y.; Huang, W.; Liu, M.; Zhang, H.; Zou, B. Fusion of satellite images in urban area: Assessing the quality of resulting images. In Proceedings of the 2010 18th International Conference on Geoinformatics, Beijing, China, 18–20 June 2010; pp. 1–4. [Google Scholar]
  27. Teo, T.; Fu, Y. Spatiotemporal Fusion of Formosat-2 and Landsat-8 Satellite Images: A Comparison of “Super Resolution-Then-Blend” and “Blend-Then-Super Resolution” Approaches. Remote Sens. 2021, 13, 606. [Google Scholar] [CrossRef]
  28. Amarsaikhan, D.; Saandar, M.; Ganzorig, M.; Blotevogel, H.H.; Egshiglen, E.; Gantuyal, R.; Nergui, B.; Enkhjargal, D. Comparison of multisource image fusion methods and land cover classification. Int. J. Remote Sens. 2012, 33, 2532–2550. [Google Scholar] [CrossRef]
  29. Karathanassi, V.; Kolokousis, P.; Ioannidou, S. A comparison study on fusion methods using evaluation indicators. Int. J. Remote Sens. 2007, 28, 2309–2341. [Google Scholar] [CrossRef]
  30. Huang, H.; Feng, Y.; Zhang, M.; Li, M. Research on Fusion of Mapping Satellite-1 Imagery and Its Evaluation. Bulletin Surv. Mapp. 2013, 430, 6–9. [Google Scholar]
  31. Sun, P.; Dong, Y.; Chen, W.; Ma, J.; Zou, Y.; Wang, J.; Chen, H. Research on fusion of GF-2 imagery and quality evaluation. Remote Sens. Land Resour. 2016, 28, 108–113. [Google Scholar]
  32. Du, Y.; Zhang, X.; Mao, Z.; Chen, J. Performances of conventional fusion methods evaluated for inland water body observation using GF-1 image. Acta Oceanol. Sin. 2019, 38, 172–179. [Google Scholar] [CrossRef]
  33. Huang, X.; Wen, D.; Xie, J.; Zhang, L. Quality Assessment of Panchromatic and Multispectral Image Fusion for the ZY-3 Satellite: From an Information Extraction Perspective. IEEE Geosci. Remote Sens. 2014, 11, 753–757. [Google Scholar] [CrossRef]
  34. Ren, K.; Sun, W.; Meng, X.; Yang, G.; Du, Q. Fusing China GF-5 Hyperspectral Data with GF-1, GF-2 and Sentinel-2A Multispectral Data: Which Methods Should Be Used? Remote Sens. 2020, 12, 882. [Google Scholar] [CrossRef] [Green Version]
  35. Ghimire, P.; Lei, D.; Juan, N. Effect of Image Fusion on Vegetation Index Quality—A Comparative Study from Gaofen-1, Gaofen-2, Gaofen-4, Landsat-8 OLI and MODIS Imagery. Remote Sens. 2020, 12, 1550. [Google Scholar] [CrossRef]
  36. Li, L.; She, M.; Luo, H. Comparison on fusion algorithms of ZY-3 panchromatic and multi-spectral images. Trans. Chin. Soc. Agric. Eng. 2014, 30, 157–165. [Google Scholar]
  37. Liu, Z.; Zheng, Y.; Han, X. Unsupervised Multispectral and Hyperspectral Image Fusion with Deep Spatial and Spectral Priors; Springer International Publishing: Cham, Switzerland, 2021; Volume 12628, pp. 31–45. [Google Scholar]
  38. Dian, R.; Li, S.; Kang, X. Regularizing Hyperspectral and Multispectral Image Fusion by CNN Denoiser. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 1124–1135. [Google Scholar] [CrossRef]
  39. Shao, Z.; Cai, J. Remote Sensing Image Fusion With Deep Convolutional Neural Network. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 1656–1669. [Google Scholar] [CrossRef]
  40. Kwan, C.; Choi, J.H.; Chan, S.; Jin, Z.; Budavari, B. Resolution enhancement for hyperspectral images: A super-resolution and fusion approach. In Proceedings of the ICASSP 2017–2017 IEEE International Conference on Acoustics, Speech and Signal, New Orleans, LA, USA, 5–9 March 2017; pp. 6180–6184. [Google Scholar]
  41. Loncan, L.; de Almeida, L.B.; Bioucas-Dias, J.M.; Briottet, X.; Chanussot, J.; Dobigeon, N.; Fabre, S.; Liao, W.; Licciardi, G.A.; Simoes, M.; et al. Hyperspectral Pansharpening: A Review. IEEE Geosci. Remote Sens. Mag. 2015, 3, 27–46. [Google Scholar] [CrossRef] [Green Version]
  42. Lillo Saavedra, M.; Gonzalo, C. Spectral or spatial quality for fused satellite imagery? A trade-off solution using the waveletà trous algorithm. Int. J. Remote Sens. 2007, 27, 1453–1464. [Google Scholar] [CrossRef]
  43. Jelének, J.; Kopačková, V.; Koucká, L.; Mišurec, J. Testing a Modified PCA-Based Sharpening Approach for Image Fusion. Remote Sens. 2016, 8, 794. [Google Scholar] [CrossRef] [Green Version]
  44. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6,011,875, 4 January 2000. [Google Scholar]
  45. Schowengerdt, R.A. Reconstruction of multispatial, multispectral image data using spatial frequency content. Photogramm. Eng. Rem. Sens. 1980, 46, 1325–1334. [Google Scholar]
  46. Nunez, J.; Otazu, X.; Fors, O.; Prades, A.; Pala, V.; Arbiol, R. Multiresolution-based image fusion with additive wavelet decomposition. IEEE Trans. Geosci. Remote 1999, 37, 1204–1211. [Google Scholar] [CrossRef] [Green Version]
  47. Chen, Y.; Sun, K.; Bai, T.; Peng, Y. Quality assessment on image fusion methods for GF-2 data. Sci. Surv. Mapp. 2017, 42, 35–40. [Google Scholar]
  48. Choi, M. A new intensity-hue-saturation fusion approach to image fusion with a tradeoff parameter. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1672–1682. [Google Scholar] [CrossRef] [Green Version]
  49. Zhou, H.; Shuang, W.U.; Mao, D.; Wang, A.; Bingyang, Y.U. Improved Brovey method for multi-sensor image fusion. J. Remote Sens. 2012, 16, 343–360. [Google Scholar]
  50. El-Mezouar, M.C.; Taleb, N.; Kpalma, K.; Ronsin, J. An IHS-Based Fusion for Color Distortion Reduction and Vegetation Enhancement in IKONOS Imagery. IEEE Trans. Geosci. Remote Sens. 2011, 49, 1590–1602. [Google Scholar] [CrossRef]
  51. Pei, W.; Wang, G.; Yu, X. Performance evaluation of different references based image fusion quality metrics for quality assessment of remote sensing Image fusion. Geosci. Remote Sens. Symp. 2012, 2280–2283. [Google Scholar] [CrossRef]
  52. Sengupta, A.; Seal, A.; Panigrahy, C.; Krejcar, O.; Yazidi, A. Edge Information Based Image Fusion Metrics Using Fractional Order Differentiation and Sigmoidal Functions. IEEE Access 2020, 8, 88385–88398. [Google Scholar] [CrossRef]
  53. Rodriguez-Esparragon, D.; Marcello, J.; Eugenio, F.; Garcia-Pedrero, A.; Gonzalo-Martin, C. Object-based quality evaluation procedure for fused remote sensing imagery. Neurocomputing 2017, 255, 40–51. [Google Scholar] [CrossRef]
  54. Rodríguez-Esparragón, D. Evaluation of the performance of spatial assessments of pansharpened images. Geosci. Remote Sens. Symp. 2014, 1619–1622. [Google Scholar] [CrossRef]
  55. Lu, H.; Fan, T.; Ghimire, P.; Deng, L. Experimental Evaluation and Consistency Comparison of UAV Multispectral Minisensors. Remote Sens. 2020, 12, 2542. [Google Scholar] [CrossRef]
  56. Seal, A.; Garcia-Pedrero, A.; Bhattacharjee, D.; Nasipuri, M.; Gonzalo-Martin, C. Multi-scale RoIs selection for classifying multi-spectral images. Multidimens. Syst. Sign. Process 2020, 31, 745–769. [Google Scholar] [CrossRef]
  57. Wang, B.; Jaewan, C.; Seokeun, C.; Soungki, L.; Wu, P.; Yan, G. Image Fusion-Based Land Cover Change Detection Using Multi-Temporal High-Resolution Satellite Images. Remote Sens. 2017, 9, 804. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Spectral response function of multispectral sensor.
Figure 1. Spectral response function of multispectral sensor.
Remotesensing 13 02354 g001
Figure 2. Location of study area. (a) hyperspectral image; (b) multispectral image.
Figure 2. Location of study area. (a) hyperspectral image; (b) multispectral image.
Remotesensing 13 02354 g002
Figure 3. Methodological flowchart of the research. GS: Gram-Schmidt; HPF: High-Pass Filter; NND: Nearest-Neighbor Diffusion; IHS: Intensity-Hue-Saturation; Wavelet: Wavelet Transform; Brovey: Color Normalized Sharping.
Figure 3. Methodological flowchart of the research. GS: Gram-Schmidt; HPF: High-Pass Filter; NND: Nearest-Neighbor Diffusion; IHS: Intensity-Hue-Saturation; Wavelet: Wavelet Transform; Brovey: Color Normalized Sharping.
Remotesensing 13 02354 g003
Figure 4. Structure of the fused image quality evaluation system.
Figure 4. Structure of the fused image quality evaluation system.
Remotesensing 13 02354 g004
Figure 5. Experimental results of ZY-1 02D hyperspectral and multispectral data fusion. (a) ZY-1 02D hyperspectral image; (b) Brovey image; (c) GS image; (d) HPF image; (e) ZY-1 02D multispectral image; (f) IHS image; (g) NND image; (h) Wavelet image.
Figure 5. Experimental results of ZY-1 02D hyperspectral and multispectral data fusion. (a) ZY-1 02D hyperspectral image; (b) Brovey image; (c) GS image; (d) HPF image; (e) ZY-1 02D multispectral image; (f) IHS image; (g) NND image; (h) Wavelet image.
Remotesensing 13 02354 g005
Figure 6. The details of farmland and artificial buildings. (a) ZY-1 02D original hyperspectral image; (b) GS image; (c) HPF image; (d) IHS image; (e) NND image; (f) Wavelet image; (g) Brovey image.
Figure 6. The details of farmland and artificial buildings. (a) ZY-1 02D original hyperspectral image; (b) GS image; (c) HPF image; (d) IHS image; (e) NND image; (f) Wavelet image; (g) Brovey image.
Remotesensing 13 02354 g006
Figure 7. The details of vegetation and water.(a) ZY-1 02D original hyperspectral image; (b) GS image; (c) HPF image; (d) IHS image; (e) NND image; (f) Wavelet image; (g) Brovey image.
Figure 7. The details of vegetation and water.(a) ZY-1 02D original hyperspectral image; (b) GS image; (c) HPF image; (d) IHS image; (e) NND image; (f) Wavelet image; (g) Brovey image.
Remotesensing 13 02354 g007
Figure 8. Spectral curves of typical objects. Top: vegetation spectral curves of six fused images and HS image (a), difference between fused images and HS image in vegetation spectral curve (d); Middle: artificial building spectral curves of six fused images and HS image (b), difference between fused images and HS image in building spectral curve (e); Bottom: water spectral curves of six fused images and HS image (c), difference between fused images and HS image in water spectral curve (f).
Figure 8. Spectral curves of typical objects. Top: vegetation spectral curves of six fused images and HS image (a), difference between fused images and HS image in vegetation spectral curve (d); Middle: artificial building spectral curves of six fused images and HS image (b), difference between fused images and HS image in building spectral curve (e); Bottom: water spectral curves of six fused images and HS image (c), difference between fused images and HS image in water spectral curve (f).
Remotesensing 13 02354 g008
Table 1. Parameters of ZY-1 02D hyperspectral and multispectral sensors.
Table 1. Parameters of ZY-1 02D hyperspectral and multispectral sensors.
SensorsBandsSpectral Range/nmSpatial Resolution/mSpectral Resolution/nmSwath Width/km
MSB02452~52110 115
B03522~607
B04635~694
B05776~895
B06416~452
B07591~633
B08708~752
B09871~1047
HSVIS-NIR396~1039301060
SWI1056~2501 20
Table 2. Correspondence between the eight MS bands and seventy-five HS bands.
Table 2. Correspondence between the eight MS bands and seventy-five HS bands.
Spectral Range/nmBands of MS (HS)
452~521B02 (4~12)
522~607B03 (13~18)
635~694B04 (22~33)
776~895B05 (40~55)
416~452B06 (1~3)
591~633B07 (19~21)
708~752B08 (34~39)
871~1047B09 (56~75)
Table 3. Quantitative evaluation statistics metrics.
Table 3. Quantitative evaluation statistics metrics.
Number of Bands (Wavelength)Fusion MethodMeanStandard DeviationEntropyMean GradientCorrelation Coefficient
B02 (48 nm)Original HS100.364812.76726.040426.88511.0000
GS96.933819.67236.124147.80240.9027
HPF103.557524.39266.089654.46260.9448
IHS73.185824.11575.270849.49110.8613
NND103.820114.21276.018329.04720.9237
Wavelet102.710518.27656.111238.48770.9711
Brovey29.68656.37604.972546.15800.7841
B03 (56 nm)Original HS109.753614.07126.079230.42751.0000
GS108.766621.0176.060654.88790.8979
HPF113.848125.30816.130958.95710.9479
IHS104.023322.50636.032652.43680.8986
NND111.232915.77976.045832.91320.9501
Wavelet112.767819.17296.132241.96920.9682
Brovey33.63756.78815.03751.80420.7923
B04 (66 nm)Original HS105.648815.09966.101732.10221.0000
GS103.723822.55026.128355.08250.8966
HPF108.942726.68346.128460.85550.9426
IHS104.123923.56096.040454.74680.8995
NND108.638216.49246.050233.90250.9411
Wavelet111.492120.26556.150244.34680.9664
Brovey32.27697.27675.026853.66610.7904
B05 (83 nm)Original HS148.183410.95445.764924.21741.0000
GS147.757917.86055.922656.45830.9281
HPF148.472119.78516.025849.29870.9095
IHS129.568523.51025.891156.81610.9252
NND138.335514.07125.879730.26960.9403
Wavelet156.021516.01695.854834.22370.9813
Brovey48.55665.99844.916247.46230.8193
B06 (43 nm)Original HS98.283111.91646.062625.35951.0000
GS104.128415.32136.105541.89930.8887
HPF101.606224.55616.071555.17040.9341
IHS79.584222.25625.683944.54850.8834
NND115.303919.28736.047445.64130.8802
Wavelet99.297317.69686.099837.46420.9696
Brovey33.48795.62965.000143.93590.7811
B07 (61 nm)Original HS107.934114.51226.096731.15871.0000
GS105.926422.25586.128354.59360.8972
HPF110.596126.21066.130960.14160.9442
IHS104.679222.36296.043355.70090.8985
NND109.816315.96876.052733.11660.9502
Wavelet112.510720.22156.150844.29710.9688
Brovey32.22557.26895.027953.63640.8894
B08 (76 nm)Original HS147.684411.23185.766124.99811.0000
GS147.212517.77535.925257.08810.9289
HPF139.32721.36676.016153.93280.8806
IHS116.049322.17316.022955.15460.6981
NND127.733421.62025.938156.69210.8184
Wavelet134.625520.46866.077645.90710.8207
Brovey48.96516.08444.902856.18310.8106
B09 (95 nm)Original HS154.83279.95055.740122.14831.0000
GS153.136918.26315.879351.55270.9529
HPF159.398118.01646.025847.08320.9775
IHS130.148122.28895.891155.18310.9226
NND146.238112.78015.858327.79780.9461
Wavelet159.349915.88875.854836.20660.9832
Brovey48.01086.18524.899546.14930.8071
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lu, H.; Qiao, D.; Li, Y.; Wu, S.; Deng, L. Fusion of China ZY-1 02D Hyperspectral Data and Multispectral Data: Which Methods Should Be Used? Remote Sens. 2021, 13, 2354. https://doi.org/10.3390/rs13122354

AMA Style

Lu H, Qiao D, Li Y, Wu S, Deng L. Fusion of China ZY-1 02D Hyperspectral Data and Multispectral Data: Which Methods Should Be Used? Remote Sensing. 2021; 13(12):2354. https://doi.org/10.3390/rs13122354

Chicago/Turabian Style

Lu, Han, Danyu Qiao, Yongxin Li, Shuang Wu, and Lei Deng. 2021. "Fusion of China ZY-1 02D Hyperspectral Data and Multispectral Data: Which Methods Should Be Used?" Remote Sensing 13, no. 12: 2354. https://doi.org/10.3390/rs13122354

APA Style

Lu, H., Qiao, D., Li, Y., Wu, S., & Deng, L. (2021). Fusion of China ZY-1 02D Hyperspectral Data and Multispectral Data: Which Methods Should Be Used? Remote Sensing, 13(12), 2354. https://doi.org/10.3390/rs13122354

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop