Next Article in Journal
A Fuzzy-Based Approach for Sensing, Coding and Transmission Configuration of Visual Sensors in Smart City Applications
Previous Article in Journal
Depth Errors Analysis and Correction for Time-of-Flight (ToF) Cameras
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessment of Pansharpening Methods Applied to WorldView-2 Imagery Fusion

Key Laborary of Digital Earth Sciences, Institute of Remote Sensing and Digital Earth, Chinese Academy of Sciences, Beijing 100094, China
*
Authors to whom correspondence should be addressed.
Sensors 2017, 17(1), 89; https://doi.org/10.3390/s17010089
Submission received: 11 October 2016 / Revised: 30 November 2016 / Accepted: 26 December 2016 / Published: 5 January 2017
(This article belongs to the Section Remote Sensors)

Abstract

:
Since WorldView-2 (WV-2) images are widely used in various fields, there is a high demand for the use of high-quality pansharpened WV-2 images for different application purposes. With respect to the novelty of the WV-2 multispectral (MS) and panchromatic (PAN) bands, the performances of eight state-of-art pan-sharpening methods for WV-2 imagery including six datasets from three WV-2 scenes were assessed in this study using both quality indices and information indices, along with visual inspection. The normalized difference vegetation index, normalized difference water index, and morphological building index, which are widely used in applications related to land cover classification, the extraction of vegetation areas, buildings, and water bodies, were employed in this work to evaluate the performance of different pansharpening methods in terms of information presentation ability. The experimental results show that the Haze- and Ratio-based, adaptive Gram-Schmidt, Generalized Laplacian pyramids (GLP) methods using enhanced spectral distortion minimal model and enhanced context-based decision model methods are good choices for producing fused WV-2 images used for image interpretation and the extraction of urban buildings. The two GLP-based methods are better choices than the other methods, if the fused images will be used for applications related to vegetation and water-bodies.

1. Introduction

The WorldView-2 (WV-2) satellite, launched in October 2009, offers eight multispectral (MS) bands of 1.84-m spatial resolution and a panchromatic (PAN) band of 0.46 m spatial resolution [1]. The MS bands cover the spectrum from 400 nm to 1050 nm, and include four conventional visible and near-infrared MS bands: blue (B, 450–510 nm), green (G, 510–580 nm), red (R, 630–690 nm), and near-IR1 (NIR1, 770–895 nm); and four new bands: coastal (C, 400–450 nm), yellow (Y, 585–625 nm), red edge (RE, 705–745 nm), and near-IR2 (NIR2, 860–1040 nm). The PAN band has a spectral response range of 450–800 nm, which covers shorter NIR spectral range than some common PAN bands of 450–900 nm. The WV-2 images have been widely used in various fields, e.g., geological structure interpretation [1], Antarctic land cover mapping [2], bamboo patch mapping [3], high density biomass estimation for wetland vegetation [4], mapping natural vegetation on a coastal site [5], predicting forest structural parameters [6], and especially for the detection of urban objects. Since numerous applications need high-spatial-resolution (HSR) MS images, it is highly desirable to fuse the eight MS bands and the PAN band to produce HSR MS imagery for better monitoring the Earth’s surface.
Numerous pansharpening methods have been proposed in the last decades to produce spatially enhanced MS images by fusing the MS and PAN images. These methods are divided into two categories: the component substitution (CS) family and multi-resolution analysis (MRA) family. The CS approaches focus on the substitution of a component that is obtained by a spectral transformation of the MS bands with the PAN image. The representative CS methods are the intensity-hue-saturation [7,8], principal component analysis [9], and Gram-Schmidt spectral sharpening (GS) [10,11] methods. The CS methods are easy to implement, and the generated fused MS images yield high spatial quality. However, the CS methods suffer from spectral distortions since the local dissimilarities between the PAN and MS channels, which are caused by different spectral response ranges, are not considered by them. The MRA-based techniques rely on the injection of the spatial details that are obtained through a multi-resolution decomposition of the PAN image into the up-sampled MS bands. Multi-resolution decomposition methods, such as “à trous” wavelet transform [12,13], undecimated or decimated Wavelet transform [14,15,16], Laplacian pyramids [17], Contourlet [18,19,20], and Curvelet [21], are often employed to extract spatial details of the PAN image. Although the MRA-based methods better preserve spectral information of the original MS images than the CS methods, they may cause spatial distortions, such as ringing or aliasing effects, originating shifts or blurred contours and textures [22]. Numerous hybrid schemes combining CS and MRA-based methods are developed to maximize spatial improvement and minimize spectral distortions [23,24,25,26]. In addition, several new pansharpening methods were proposed for the fusion of WV-2 imagery, i.e., the Hyperspherical Color Sharpening (HCS) [27] method and the improved Non-subsampled Contourlet Transform (NSCT) method [28]. These methods were proved to be better than early CS methods, such as GS, PCA.
Several studies have performed comparisons and analyses of some widely used state-of-the-art pansharpening methods, using test images covering different regions from several sensors. Previous studies showed that a pansharpening method may give different performances for test images from different sensors [29,30]. A noticeable point for the WV-2 is that the spectral ranges of the PAN band overlap limited party of the spectral ranges of the C, NIR1, and NIR2 bands. This will result in relative low correlation coefficients between these bands and the PAN bands, which may lead to spectral distortions of the fused version of these bands [31]. Regarding the wide use of the fused WV-2 images, it is urgent to evaluate the performances of different state-of-the-art pansharpening methods applied to WV-2 imagery. Some of the previous comparisons also used test images recorded by WV-2 [30,32,33,34,35] and other sensors. In these works, the early pansharpening methods, such as GS, PANSHARP, Ehlers, modified intensity-hue-saturation (M-IHS), high pass filter (HPF), principal component analysis (PCA), and wavelet-PCA (W-PCA) methods were assessed regarding quality indexes and visual inspection, usually using one or two test images covering urban areas. However, a fusion product providing the best performance in terms of quality indexes and visual inspection may be the best choice for applications such as image interpretation, but it may be not the best choice for applications related to classification and objects identification, i.e., the extraction of buildings, vegetation, and water-bodies [29,36,37]. Consequently, it is important to evaluate the widely used state-of-the-art pansharpening methods from the point of applications, such as land cover classification and object extraction. The purpose of this study was to assess the performances of the existing state-of-the-art pansharpening methods applied to WV-2 imagery, using information indices related to land cover classification and information extraction, as well as quality indexes and visual inspection. Several test images, presenting typical image scenes covering urban, suburban, and rural regions, are employed in the experiments. In addition, the newly proposed HCS, and NSCT methods, which are rarely included in previous comparisons, will be included in this work.
In this study, eight state-of-the-art algorithms, most of which have been demonstrated to outperform some other methods were assessed using both quality indices and information indices, along with visual inspection. The selected algorithms include four methods belonging to the CS family and four methods belonging to the MRA family. The four CS methods including Gram-Schmidt (GS) [10], adaptive GS (GSA) [38], Haze- and Ratio-based (HR) [39], and HCS [27] were compared. The four MRA methods include undecimated “à trous” wavelet transform (ATWT) using additive injection model [40,41], Generalized Laplacian pyramids (GLP) using spectral distortion minimal model (SDM) and context-based decision model (CBD) [42,43], and the improved NSCT method introduced in [28]. Traditional image quality indices couple with visual inspection were adopted to assess the quality of the fused images. Four comprehensive indices, including Dimensionless Global Relative Error of Synthesis (ERGAS) [44], Spectral Angle Mapper (SAM) [45], Q2n [46,47], and spatial correlation coefficient (SCC) [48] were employed to measure the spectral distortion between the fused and the original MS bands. Regarding the application purpose of the high-resolution fused images, which includes land cover classification of urban or suburban areas, bamboo and forest mapping, and so on, some widely used indexes, derived from the fusion products, were assessed to evaluate the information presentation ability of the fusion products. The employed indexes include morphological building index (MBI) [49], normalized difference vegetation index (NDVI), and normalized difference water index (NDWI). The information presentation of a fusion product was assessed using the correlation coefficient (CC) between an index derived from the fusion product and the same index derived from the corresponding original MS image. A higher CC value implies a better information preservation ability of the fusion product.
This paper is organized as follows: the eight selected pansharpening methods are introduced in Section 2, as well as the quality indexes; the experimental results with visual and quantitative comparisons with other outstanding fusion methods are presented in Section 3. Discussions are presented in Section 4, whereas the conclusions are summarized in Section 5.

2. Methodology

2.1. Algorithms

The algorithms used for the comparisons are introduced in the following subsections. MS and P represent the original low-resolution MS image and high-resolution PAN image, respectively. M S ˜ and M S ^ represent the up-sampled MS and the fused MS images, respectively. A general formulation of CS fusion is given by:
M S ^ i = M S ˜ i + g i ( P I L ) , i = 1 , , N
in which the subscript i indicates the ith spectral band; g i is the injection gains of the ith band, while the intensity image I L is defined by Equation (2):
I L =   i = 1 N w i MS i ˜
where wi is the weight of the ith MS band, and N is the number of MS bands.
Similarity, a general formulation for MRA-based methods can be given by Equation (3):
M S ^ i = M S ˜ i + g i ( P P L ) , i = 1 , , N
where PL is the low-frequency component of the PAN band. PL can be obtained by different approaches, such as low-pass filter, Laplacian pyramid and wavelet decomposition.

2.1.1. GS and GSA

GS is a representative method of the CS family, the fusion process of which is described by Equation (1), with the injection gains g i given by Equation (4):
g i =   cov ( M S i , I L ) var ( I L )
where cov (X, Y) is the covariance between the two images X and Y, and var (X) is the variance of X.
Several versions of GS can be achieved by changing the method for generating IL. One way to obtain IL is simply averaging the MS components (i.e., using Equation (2) with setting w i = 1 for all i = 1 , , N ). This modality is referred as GS or GS mode 1. Another way is using the low-pass version of P as IL. This modality is referred as GS mode 2.
An enhanced version, called adaptive GS (GSA), is proposed by assigning IL as a weighted average of the original MS bands, as Equation (2). The weights wi in Equation (2) are calculated with the minimum mean square error (MSE) solution of Equation (5):
P 0 =   i = 1 N w i M S i + ε
where P 0 is the degraded version of P, with the same pixel sizes of the original MS bands. P 0 is generated by low-pass filtering of P, followed by decimation. Both the GS and GSA methods are included in the experiments in this study.

2.1.2. HR

The HR method is based on the assumption that the ratio of a HSR MS band to a low-spatial-resolution (LSR) MS band is equal to the ratio of a HSR PAN image to an assumed LSR PAN image a fusion method that considers haze [50,51,52]. The fused ith MS band M S ^ i can be calculated by using Equation (6) according to HR fusion method:
M S ^ i = ( M S ˜ i H i ) P H p P L H p + H i , i = 1 , , N
where P L is a low-pass version of P; Hi and Hp denote the haze values in the ith MS band and the PAN band, respectively. The values of Hi and Hp can be determined using the minimum grey level values in M S i and P according to an image-based dark-object subtraction method [50,51,52].

2.1.3. HCS

HCS is a pansharpening method designed for WV-2 imagery based on transforming the MS bands into hyperspherical color space. The HCS method offers two modes, including the native mode and the smart mode. The process of the native mode is described as follows:
(a)
The squares of the multispectral intensity (I2) and the PAN (P2) are calculated using Equations (7) and (8), respectively:
I 2 = i = 1 N M S ˜ i 2
P 2 = ( P ) 2
(b)
Calculate the mean (uP) and standard deviation (σp) of P2, as well as the mean (uI) and standard deviation (σP) of I2.
(c)
The P2 is adjusted to the mean and standard deviation of I2, using Equation (9):
P 2 =   σ I σ P ( P 2 u P + σ P ) + u I σ I
(d)
The square root of the adjusted P2 is assigned to Iadj (i.e., I a d j = P 2 ), Iadj is used in the reverse transform from HCS color space back to the original color space, using Equation (10):
M S ˜ i = { I a d j c o s φ i , i = 1 I a d j s i n φ 1 s i n φ 2 s i n φ i 1 c o s φ i , i = 2 , , N 1 I a d j s i n φ 1 s i n φ 2 s i n φ i 1 , i = N ,
in which φ i is defined using Equation (11):
φ i = { tan 1 ( k = 2 N M S ˜ k 2 M S ˜ i ) , i = 1 tan 1 ( k = i + 1 N M S ˜ k 2 M S ˜ i ) , i = 2 , , N 2 tan 1 ( M S ˜ i + 1 M S ˜ i ) , i = N 1
For the smart case, similarity to P2, a PS2 is calculated by using the low-pass version of the PAN image (PL), i.e., P S 2 = ( P L ) 2 . The PL is generated by an average filtering with size of 7 × 7. Both the means and standard deviations of PS2 and P2 are adjusted to those of I2. Then, the adjusted intensity Iadj is assigned by using Equation (12):
I a d j = ( P 2 P S 2 ) I 2
Finally, the fused image can be generated by using Equations (10) and (11). The HCS method using the smart mode is considered in the experiment in this study.

2.1.4. ATWT

The ATWT method is a MRA fusion approach that extracting spatial details using “à trous” wavelet transform. The fusion scheme utilizing the additive injection model can be formulated as:
M S ^ i = M S ˜ i + P P L , i = 1 , , N
where PL is the low frequency component of P and is generated by the “à trous” wavelet.
The “à trous” wavelet is a kind of non-orthogonal wavelet that is different from orthogonal and biorthogonal. It is a redundant transform, since decimation is not implemented during the process of wavelet transform while the orthogonal or biorthogonal wavelet can be carried out using either decimation or undecimation mode.
The whole process of the ATWT fusion can be divided into two steps [40]:
(1)
Use the à trous wavelet transform to decompose the PAN image to n wavelet planes. Usually, n = 2 or 3.
(2)
Add the wavelet planes (i.e., spatial details) of the decomposed PAN images to each of the spectral bands of the MS image to produce fused MS bands.

2.1.5. GLP

The fusion process of GLP can also be formulated as Equation (3). For the GLP method, the low-frequency component of the PAN band PL is generated by up-sampling the down-sampled version of the original PAN band P. The down-sampling of P is implemented using a low-pass reduction filter that matches the MTF of the band, whereas the up-sampling of P is carried out by using an ideal expansion low-pass filter.
For the GLP fusion, different detail injection models are designed to obtain the injection coefficients gi. The most used models are the spectral distortion minimizing (SDM) model and the context-based decision (CBD) model.
For the SDM model, the injection coefficients gi can be obtained using Equation (14):
g i ( m , n ) = β ( m , n ) M S ˜ i ( m , n ) P L ( m , n ) , i = 1 , , N
where β ( m , n ) is equal to 1 for all pixels in the original SDM model [42], whereas β ( m , n ) is defined as the ratio between average local standard deviations of resampled MS bands and local standard deviation of PL , for the enhanced SDM (ESDM) model proposed by Aiazzi [43]:
β ( m , n ) 1 N k = 1 N v a r [ M S ˜ k ] ( m , n ) v a r [ P L ] ( m , n )
In the CBD model, the space-varying coefficients gi is defined as Equation (16):
g i ( m , n ) = { min { σ M S ˜ i ( m , n ) σ P L ( m , n ) , c } i f   ρ i ( m , n ) θ i , 0 i f   ρ i ( m , n ) < θ i ,
in which ρ i ( m , n ) is the local correlation coefficient between M S ˜ i and PL calculated on a square sliding window of size L × L centered on pixel (m, n).
The CBD model is uniquely defined by the set of thresholds 0 < θ i 1 , for i = 1 , , N , generally different for each band, and by the window size L depending on the spatial resolutions and scale ratio of the images be merged, as well as the landscape characteristics (typically, 7 L 11 to avoid loss of local sensitivity with L > 11 and statistical instability with L < 7). The thresholds may be related to the spectral content of the Pan image, e.g., θ i = 1 ρ i , where ρ i is the global correlation coefficient between the kth band and the Pan image spatially degraded to the same resolution. A clipping constant c was introduced to avoid numerical instabilities (empirically, 2 c 3 ).
For the Enhanced CBD (ECBD) model proposed by Aiazzi [43], the coefficients g i is calculated as Equation (17):
g i ( m , n ) = min { σ M S ˜ i ( m , n ) σ P L ( m , n ) ρ i ( m , n ) ρ i ¯ , c }
where ρ i ¯ is the global correlation coefficient between M S ˜ i and P L . Both the GLP methods using the ESDM and the ECBD models are considered in the experiment.

2.1.6. NSCT

The contourlet transform (CT) is proved to be a better approach than the wavelet for pan-sharpening. CT is implemented by a multiscale decomposition using the Laplacian pyramid followed by a local directional transform using the directional filter bank (DFB).
The NSCT is a shift-invariant version of CT and has excellent multilevel and multi-direction properties. NSCT is built upon the non-subsampled pyramid filter banks (NSPFBs) and the non-subsampled directional filter banks (NSDFBs). The NSPFB employed by NSCT is a 2-D two-channel non-subsampled filter bank, whereas the NSDFB employed by NSCT is a shift-invariant version of the critically sampled DFB in CT. The details of the NSCT can be attended in [28,53]. An improved version of the standard NSCT-based method is introduced in [28]. The process of the improved NSCT method with mode 2 (NSCT_M2) in [28] is descripted as follows.
(a)
Each original MS band M S i , is decomposed using 1-level NSCT to get one coarse level, C 0 M S i , and one fine level, C 1 M S i ;
(b)
The PAN band is decomposed using3-level NSCT into one coarse level, C 0 P A N , and three fine levels, which are denoted as C 1 P A N , C 2 P A N , and C 3 P A N , respectively.
(c)
The coefficients of each MS band, C 0 M S and C 1 M S , are up-sampled to the scale of the PAN band using the bi-linear interpolation algorithm.
(d)
The coarse level of the fused ith MS band, C 0 F i , is the up-sampled coarse level of the ith MS band C 0 M S i , whereas the fine levels 2 ( C 2 F i ) and 3 ( C 3 F i ) of the fused ith MS band are the fine levels 2 ( C 2 P A N ) and 3 ( C 3 P A N ) of the PAN band.
(e)
The fused fine level 1, C 1 F i , is obtain by fusing the coefficients of the same level obtained from both the ith MS band and the PAN band. For each pixel (x, y), the coefficients of the fused fine level 1, C 1 F i ( x , y ) , is determined according to Equation (18):
C 1 F i ( x , y ) = { C 1 M S i ( x , y ) , L E M S i ( x , y ) > L E PAN ( x , y ) C 1 P A N ( x , y ) , o t h e r w i s e
where L E M S i ( x , y ) and LEPAN(x, y) are the local energy of pixel (x, y) for the ith MS band and the PAN band, respectively, calculated within a (2M + 1) × (2P + 1) window using the formula shown in Equation (19):
L E ( x , y ) = i = M M j = P P ( C 1 ( x + i , y + j ) ) 2
The inverse NSCT is applied to the fused coefficients to provide the fused ith MS band.
This improved version was demonstrated to provide pansharpened images with a good spectral quality.

2.2. Quality Indexes

Quality assessment of fusion products can be performed using two approaches. The first approach considers fusing images at a spatial resolution lower than the original resolution and uses the original MS image as a reference to assess the quality of the fused images. Several indexes have been proposed for evaluating the spatial and spectral distortions of the fused image with respect to an available reference image. The widely used spectral quality indexes include Root Mean Square Error (RMSE), Relative Average Spectral Error (RASE) [54], ERGAS, SAM, Universal Image Quality Index (UIQI) [55], Q4 [46], Q2n, and Peak Signal-to-Noise Ratio (PSNR) [56], whereas the widely used spatial indexes include SCC and Structural SIMilarity (SSIM) [57]. The second approach uses quality indexes that do not require a reference image but operate on relationships among the original images and the fusion products. This approach has the advantage of validating the products at the original scale, thus avoiding any hypothesis on the behavior at different scales. However, appropriate indexes requiring no reference should be exploited to assess the quality of the fusion product. The Quality with no reference index (QNR) was one of the mostly used indexes [58]. It is composed by the product of two separate values that quantify spectral and spatial distortions, respectively. However, QNR is proved to be lower reliability than the indexes belonging to the first approach [30], since it can be affected by slightly mismatches among the original image bands. The acquisition modality of WV-2 can led to s mall temporal misalignments among the MS bands, since the eight MS bands of are arranged in two arrays of four bands each. Consequently, the QNR index is not used in this study. Four quality indexes belong to the first approach, including ERGAS, SAM, Q2n, and SCC, are employed to assess the quality of fused images at data level. These indexes are chosen due to they are widely used in literatures related to fusion of remote sensing imagery.

2.2.1. ERGAS

The global index ERGAS is an improved version of the index named Relative Average Spectral Error (RASE), which is defined based on Root Mean Square Error (RMSE). The formula of ERGAS [44] is defined as follows:
ERGAS = 100 R 1 N k = 1 N ( R M S E ( M S ^ k , M S k ) u ( M S k ) ) 2
where u ( M S k ) is the mean of the kth band of the reference image; R is the spatial resolution ratio between the MS and PAN bands; RMSE is defined as Equation (21):
R M S E ( M S ^ k , M S k ) = E [ ( M S ^ k M S k ) 2 ]
The optimal value for ERGAS is 0, since it is defined as a weighted sum of RMSE values.

2.2.2. SAM

SAM expresses the spectral similarity between a fused image and a reference image with the average spectral angle of all pixels involved [45]. Let two spectral vectors V = { V 1 , V 2 , , V N } and V ^ = { V ^ 1 , V ^ 2 ,   , V ^ N } present the reference spectral pixel and the fused spectral pixel, respectively, their spectral angle SAM is defined as in Equation (22):
SAM = arccos ( V , V ^ | V | | V ^ | )
where X , Y stands for the inner-product of the two vectors X and Y, and |X| stands for the modulus of a vector X.
The smaller the spectral angle, the higher the similarity between the two vectors. Since the angle is independent of the magnitudes of the two vectors, the index SAM is not affected by solar illumination factors.

2.2.3. Q2n

Q2n is a generalization of Q index for monoband images [47] and an extension of Q4 [46]. Q2n is derived from the theory of hypercomplex numbers, particularly of 2n-ones [59,60]. For a 2n-on hypercomplex random variable z (in boldface) is written in the following form:
z = z 0 + i 1 z 1 + + i 2 n 1 z 2 n 1
where z 0 , z 1 , , z 2 n 1 are real numbers, and i 0 , i 1 , , i 2 n 1 are hypercomplex unit vectors, the conjugate z * is given by:
z * = z 0 i 1 z 1 i 2 n 1 z 2 n 1
and the modulus | z | is defined by Equation (25):
| z | = z 0 2 + z 1 2 + + z 2 n 1 2
Give two 2n-on hypercomplex random variables z and v, the hypercomplex covariance between z and v is defined as Equation (26):
c o v ( z , v ) E [ ( z z ¯ ) ( v v ¯ ) ] = E [ z v * ] z ¯ v ¯ * = σ z , v
The hypercomplex CC between the two 2n-on random variables are defined as the normalized covariance:
q ( z , v ) σ z , v σ z σ v
in which σ z and σ v are the square roots of the variances of z and v, and are obtained by Equations (28) and (29), respectively:
σ z = E [ | z | 2 ] | z ¯ | 2
σ v = E [ | v | 2 ] | v ¯ | 2
The index Q2n can be computed from Equation (30):
Q 2 n M × M = σ z , v σ z σ v 2 z ¯ v ¯ ( z ¯ ) 2 + ( v ¯ ) 2 2 σ z σ v σ z 2 + σ v 2
where M is the size of the local window used to calculate Q 2 n .
Finally, Q2n is obtained by averaging the magnitudes of all Q 2 n M × M over the whole image, according to Equation (31):
Q 2 n = E [ | Q 2 n M × M | ]
According to [46], the value of M is suggested to be 32.

2.2.4. SCC

To assess the spatial quality of the fusion products, the spatial details presented in the fused images will be compared with those presented in the reference image by calculating the correlation coefficient between the spatial details extracted from the two images. Similar to the procedure proposed by Otazu et al. [48], the spatial information presented in the two images to be compared is extracted by using a Laplacian filter, then, the correlations between the two filtered images are calculated band by band. However, an overall correlation coefficient of the two edge images with eight bands is calculated in this study. A high SCC value indicates that many of the spatial details of the reference image are presented in the fused image.

2.3. Information Indexes

In order to assess the ability of information extraction of the fusion products employed in specific remote sensing applications, we assessed the quality of fused images by the use of a series of information indices, which are proved to be useful in land cover classification and information extraction in previous studies. Three indices are employed in this study: MBI, NDVI, and NDWI. The accuracy of an information index derived from a fusion product is assessed using the CC between the information index and the same information index derived from the corresponding reference MS image. A higher CC value implies a better information preservation ability of the fusion product in terms of the information index. Henceforth, the CC values calculated for MBI, NDVI, and NDWI are denoted as CMBI, CNDVI, and CNDWI, respectively. The employed three information indices are introduced in this subsection.

2.3.1. NDVI

Based on the principle that vegetation has a strong reflectance in the near-infrared (NIR) channel but a strong absorption in the red (Red) channel, the NDVI is defined as Equation (32):
N D V I = N I R R e d N I R + R e d

2.3.2. NDWI

The NDWI is defined using the spectral value of the Green band (G) and the NIR band as Equation (33):
N D W I = G N I R G + N I R

2.3.3. MBI

The MBI is proposed by Huang et al. [49], aiming to represent spectral and spatial characteristics of buildings using a series of morphological operators. The MBI is calculated according to the following steps:
(a)
A brightness image b is generated by setting the value of each pixel p to be the maximum digital number of the visible bands. Only the visible channels are considered due to they have the most significant contributions to the spectral property of buildings.
(b)
The directional white top-hat (WTH) reconstruction is employed to highlight bright structures that have a size equal to or smaller than the size of the structure element (SE), and meanwhile suppresses other dark structures in the image. WTH with linear SE is defined as Equation (34):
W T H ( s , θ ) = b γ S E ( s , θ )
in which γ S E ( s , θ ) is an opening by reconstruction operator using a linear SE with a size of s and a direction of θ .
(c)
The difference morphological profiles (DMP) of white top-hat transforms are employed to model building structures in a multi-scale manner:
{ D M P W T H = { D M P W T H ( s , θ ) : s m i n s s m a x , θ D } D M P W T H ( s , θ ) = | W T H ( s + Δ s , θ ) W T H ( s , θ ) |
(d)
Finally, MBI is defined based on DMP using Equation (36):
M B I = s , d i r D M P W T H ( s , θ ) N D × N s
where ND and NS are the directionality and the scale of the DMPs, respectively. The definition of the MBI is based on the fact that building structures have high local contrast and, hence, have larger feature values in most of the directions of the profiles. Accordingly, buildings will have large MBI values.

3. Experimental Results

3.1. Datasets

Six datasets clipped from three WV-2 scenes were considered in this study. One scene covers Beijing City, China, whereas the other two scenes cover Pingdingshan City, Hebei Province, China. The Bejing scene was acquired on 21 September 2013, with an off-nadir angle of 13.7°. One of the Hebei scenes was acquired on 29 April 2014, with an average off-nadir angle of 12.5°, whereas the other was obtained on 21 August 2014, with an average off-nadir angle of 8.6°. The locations of the three scenes are shown in Figure 1. The red rectangles in the figure indicate the locations of the six datasets. Two datasets covering urban areas were selected from the Beijing scene, whereas the other four datasets were obtained from the other two WV-2 scenes. Two of them cover suburb areas, whereas the other two cover rural areas. For all the datasets, the radiometric resolution is 16 bits, whereas the spatial resolution ratio R is 4. Each dataset has a size of 256 × 256 pixels at MS scale. The two urban images are henceforth referred as I1 and I2, respectively, whereas the two suburban images are referred as I3 and I4, respectively. The two rural images are referred as I5 and I6, respectively. The typical image objects shown in the six images are listed in Table 1; the six images are shown in Figure 2.
To evaluate the quality of the fusion products with respect to the original MS images, the degraded datasets were produced by reducing the original MS and PAN images to a spatial resolution of 8 m and 2 m, respectively. All the fusion experiments were performed on the degraded datasets. The fused 2-m images were compared with the 2-m true MS images to assess their quality. The degraded images were generated using averaging in this study [61]. The spatial resolution ratio R in this study is 4 for all the tested images.
According to the image objects included in the six test images, different information indices were chosen for each of the images (Table 1). The fusion products of the two urban images were assessed with respect to CMBI and CNDVI, due to no water bodies can be observed from these two images. The fusion products of the two suburban images were assessed using all the three indices, whereas the fusion products of the two rural images were assessed using only CNDVI and CNDWI.

3.2. Fusing Using the Selected Algorithms

All the fusion algorithms were implemented in MATLAB version 2014b. For all the selected eight fusion methods, the up-sampled MS images were produced using the bi-cubical interpolation approach. For the HR method, the low-spatial-resolution version of PAN image (PL) was obtained by averaging the PAN pixels in an R × R window, followed by the bi-cubically up-sampled to the resolution of the original PAN image. The GLP methods using the ESDM model and the ECBD model were used in the fusion experiment; the two methods are referred as GLP_ESDM and GLP_ECBD, respectively, in this study. For the HR method, the haze values for the MS and PAN bands were determined using the values of the pixel that offering the lowest value in the PAN image.

3.3. Quality Indexes

3.3.1. Assessment for the Two Urban Images

The image quality indices of the fusion products of the two urban images are shown in Table 2. The fusion products of these two images are partly shown in Figure 3 and Figure 4, respectively, in order to facilitate the observation of the details of the fused images. The images in each figure in this work are stretched by using an identical histogram obtained from the corresponding reference MS images.
The HR method offers the highest Q8 and SCC values and the lowest ERGAS and SAM values for the two urban images, indicating that the HR method gives the best performances in both spectral and spatial quality indices. The NSCT_M2 methods yields the lowest Q8 and SCC values, the highest ERGAS and SAM values, indicating the poorest performance in spectral preservation. The GSA method provides Q8 values slight lower than the highest values provided by the HR method, followed by the GS, GLP_ESDM, GLP_ECBD, ATWT, and HCS methods. The GLP_ESDM performs the best among the MRA methods in terms of both spectral and spatial quality indices. The HCS method yields the poorest performance among the CS methods, in terms of all the four quality indices for the two urban images.
Visual comparisons for the fusion products of the two urban images (Figure 3 and Figure 4) show that the fusion products generated by the GS, HCS, and NSCT_M2 methods show significant spectral distortions in shadow covered regions; this is especially obvious for the fusion products of I2. In addition, the fused image generated by the GLP_ECBD method for I2 also show significant spectral distortions in shadow covered regions, which is consistent with the poor performance of the GLP_ECBD method in terms of quality indexes. Although the fused images generated by the ATWT and GLP_ECBD methods show more sharpened boundaries between different objects than the other fusion products, they seem to be over sharpened and show spectral distortions. This is very obvious for the fusion products of I2. The two HCS-fused images are more blurred than other fusion products. It also can be observed that the fusion products generated by GSA, HR, and GLP_ESDM methods yield lower spectral distortions and more sharpened boundaries between different objects than other fusion products. In addition, the HR-fused images provide more texture details in vegetation covered areas.
According to the quality indexes and visual inspection, the HR, GSA, and GLP_ESDM methods give better performances than the other methods, whereas the NSCT_M2 and HCS methods offer the poorest performances, for the two urban images.

3.3.2. Assessment for the Two Suburban Images

The image quality indices of the fusion products of the two suburban images are shown in Table 3. Similarly, parts of the fusion products of the two images are shown in Figure 5 and Figure 6, respectively.
The GSA method offers best performances in terms of ERGAS, SAM, and Q8, whereas the GS method provides the highest SCC values, for the two suburban images. The NSCT_M2 method gives the poorest performances in terms of ERGAS, SAM, and Q8. Among the MRA methods, the GLP_ECBD method offers the highest Q8 value, whereas the ATWT method provides the highest SCC values. Although the HR method provides Q8 values slightly lower than the highest value provided by the GSA method, the former offers the lowest SCC value among the CS methods. The HR method also provides the highest SAM values among the CS methods. Although the HR method gives the best performances for the two urban images, it offers relative poor performances for the two suburban images.
Visual comparisons of the fusion products of the two suburban images show that obvious spectral distortions can be found from the water-body and shadow covered regions in the fused images generated by GS, HCS, ATWT, and NSCT_M2 methods. In addition, the fusion products of the HCS and NSCT_M2 methods are significant more blurred than other products, due to few spatial details are injected into the fused images. The fusion products generated by HR, GLP_ESDM, and GLP_ECBD methods are more sharpened and yield lower spectral distortions than other fusion products. Although the GSA method gives the best performance in terms of Q8, the corresponding fusion products seem to be more blurred than the fusion products generated by the HR, GLP_ESDM, GLP_ECBD methods. Although the GS method offers the highest SCC values, the two GS-fused images show significant spectral distortions. The two HR-fused images also show very blurred boundaries between vegetation and non-vegetation objects, which may contribute to the relative low SCC values and high SAM values provided by this method. According to the quality indexes and visual inspection, the GLP_ECBD method gives better performances than other methods for the two suburban images.

3.3.3. Assessment for the Two Rural Images

The image quality indices of the fusion products of the two rural images are shown in Table 4, whereas the sub-images of fusion products of the two images are shown in Figure 7 and Figure 8, respectively.
The eight fusion methods give different performances for I5 and I6, which may due to the fact that the land cover types of the two scenes are obviously different. For I5, the HR and GLP_ESDM methods give the highest Q8 and SCC values; the GS and NSCT_M2 methods provide the lowest Q8 values; the GSA method yields the lowest SCC values. However, for I6, the HR, GLP_ECBD, and ATWT methods offer the highest Q8 values, whereas the ATWT, NSCT_M2, and GLP_ECBD methods provide the highest SCC values. The HCS and GS methods provide the lowest Q8 and SCC values, respectively, for I6. The GLP_ESDM gives a slightly better performance than the GLP_ECBD method for I5. In contrast, the former offers a poorer performance than the latter for I6.
Obvious spectral distortions can be observed from the fused images generated by the GS, GSA, and HCS methods for I5, especially for the water-body covered regions and rooftop regions. The fusion products generated by the GS, GSA, and HCS methods for I5 are also more blurred than other fusion products. Conversely, the fused image generated by GLP_ECBD for I5 seems to be too sharpened, due to over injection of the spatial details. Consequently, the HR and GLP_ESDM methods give the best performances for I5, with respect to both the quality indexes and visual inspection.
The eight fusion products for I6 show no significant differences. Among the eight fusion products for I6, the product generated by the NSCT_M2 method is the most blurred; whereas pixels corresponding to vegetation in the product generated by the GLP_ESDM method seem to be too light colored. The HR-fused image shows slightly more details than other products, especially in vegetation covered regions. With respect to the quality indexes and visual inspection, the HR, GLP_ECBD, and ATWT methods yield the best performances for I6.

3.4. Information Preservation

For each of the fusion products, a CC between an information index derived from the fused image and the same index derived from the corresponding reference MS image was calculated. The CC values for MBI (CMBI), NDVI (CNDVI) and NDWI (CNDWI) for fusion products of the six test datasets are listed in Table 5.

3.4.1. CMBI

The fused products generated from the two urban (I1 and I2) and tow suburban (I3 and I4) images were assessed in terms of CMBI. All the fusion products of the two urban images offer CMBI values that are significant higher than those of the up-sampled MS images (EXPs), indicating that all the fusion products show obvious improvements in terms of CMBI. This is due to the fact that spatial details extracted from the PAN bands were injected to produce these images. Generally, the performances of the eight methods in terms of CMBI are consistent with those in terms of Q8 and SCC. The HR, GSA, and GS methods offer the highest CMBI values, whereas the NSCT_M2 and HCS methods offer the lowest CMBI values, for both the urban and suburban test images. The CS methods give slightly better performances than the MRA methods in terms of CMBI, for the four test images. This is consistent with the performances of the eight methods in terms of quality indexes, for the four test images. This is also consistent with the conclusion of previous studies, which imply that the fusion products generated by the CS methods offer good visual and geometrical impression [31].
Among the MRA methods, the GLP_ESDM method offers the best performance for the two urban images, whereas the GLP_ECBD method performs the best for the two suburban images, in terms of CMBI. This is also consistent with the performances of the two methods in terms of quality indexes. According to the assessment results with respect to the quality indexes and visual inspection, the HR, GSA, and GLP_ESDM methods outperform other methods for the two urban images, whereas the GLP_ECBD method gives the best performances for the two suburban images. Actually, although the four method give different performances vary different image scenes, they outperform the other methods for all the four images, in terms of CMBI, as well as the quality indexes and visual inspection. Consequently, the HR, GSA, GLP_ESDM, and GLP_ECBD methods may be the best choice for producing fused urban/suburban WV-2 images used for image interpretation and buildings extraction.

3.4.2. CNDVI and CNDWI

The fusion products of all the six images were assessed using CNDVI, whereas only the fusion products for the two suburban images and the two rural images were assessed using CNDWI. We discuss the two indexes together because both of the two indexes measure the differences between inter-band relationships of the fused image and those of the reference MS image. In addition, both of them are related to the NIR bands.
It can be observed that some of the fusion products offer CNDVI or CNDWI values that are lower than those of provided by the corresponding up-sampled MS images (EXPs), and the highest CNDVI values are just slightly higher than those of the EXPs. This indicates that the fusion products show limited improvements of in terms of CNDVI and CNDWI, which may be caused by the spectral distortions of the fused NIR bands.
The HCS method offers relative high CNDVI and CNDWI values for all the test images. This may due to the fact that the inter-band relationships of up-sampled LSR MS bands are preserved in the corresponding HCS-fused images. Since the HCS-fused images show significant spectral distortions in terms of both quality indexes and visual inspection, it is regarded as offering poor performance for specific applications. Hence, we do not discuss about the performance of this method henceforth in this section.
For the two urban images, the GLP_ESDM, HR, and GS methods offer the highest CNDVI and CNDWI values. With respect to the fact the GS-fused images show obvious spectral distortions and only the GLP_ESDM method offers higher CNDVI and CNDWI values than those of the corresponding EXPs, the GLP_ESDM method is a better choice for producing fusion products that will be used in applications related to urban vegetation and water-bodies.
For I3 and I4, the GS, ATWT, and GLP_ECBD methods provide the highest CNDVI and CNDWI values, whereas the HR and NSCT_M2 methods offer the lowest CNDVI and CNDWI values. With respect to the fact that the GS and ATWT methods give poor performances in terms of quality indexes and visual inspection, the GLP_ECBD method give the best performances for the two suburban images in terms of NDVI and NDWI information preservation, as well as quality indexes and visual inspection.
For I5, the GLP_ESDM, ATWT, and HR methods offer higher CNDVI and CNDWI values than other methods. With respect to the fact that the ATWT method gives poor performances in terms of quality indexes and visual inspection and only the GLP_ESDM method offers higher CNDVI and CNDWI values than those of the corresponding EXPs, the GLP_ESDM method is the better choice for rural WV-2 images with similar image objects with I5, in terms of NDVI and NDWI information preservation, as well as quality indexes and visual inspection.
For I6, the GLP_ECBD, ATWT, and GLP_ESDM methods provide higher CNDVI and CNDWI values than the other methods. With respect to the fact that the HR, GLP_ECBD, and ATWT methods outperform the other methods in terms of quality indexes and visual inspection only the GLP_ECBD method offers higher CNDVI and CNDWI values than those of the corresponding EXPs, the GLP_ECBD method is the best choice for rural WV-2 images with similar image objects with I6, in terms of NDVI and NDWI information preservation, as well as quality indexes and visual inspection.

4. Discussion

Generally, the comparisons of different pansharpening methods are performed by assessing fusion products using spectral and spatial quality indexes, as well as visual inspection. However, a good performance in terms of quality indexes and visual inspection does not always result in a good choice for different application purposes. The NDVI, NDWI, and MBI index, which are widely used in applications related land cover classification, the extraction of vegetation area, buildings, and water bodies, were employed in this study to evaluate the performances of the selected pansharpening methods in terms of the information presentation ability. In this study, the performances of eight selected state-of-art pan-sharpening methods were assessed using information indices (NDVI, NDWI and MBI), along with current image quality indices (ERGAS, SAM, Q2n and SCC) and visual inspection, with six datasets from two WV-2 scenes.

4.1. General Performances of the Selected Pansharpening Methods

Generally, the HR, GSA, GLP_ESDM, and GLP_ECBD methods give better performances than the other methods, whereas the NSCT and HCS methods offer the poorest performances, for most of the test images, in terms of quality indexes and visual inspection. The four methods also give slightly different performances for images including different image objects. For example, the HR, GSA, GLP_ESDM methods give the best performances for the two urban images, whereas the GLP_ECBD provides the best performances for the two rural images. However, the fusion products of the four methods offer good visual quality for most images. Consequently, the HR, GSA, GLP_ESDM, and GLP_ECBD methods are good choices if the fused WV-2 images will be used for image interpretation.
The results of the assessments using the three information indices show that the rank of the selected eight fusion methods in terms of CMBI is a little similar with those in terms of Q8 and SCC. This may indicate that the assessment using only the quality indexes and visual inspection is sufficient for selecting a best fusion method for producing fused urban WV-2 images used for image interpretation and applications related to urban buildings. The order of eight methods for in terms of CNDVI is similar with that in terms of CNDWI. This is due to the fact that both CNDVI and CNDWI measure the differences between the inter-band relationships of a fused image and those of the corresponding reference MS image. In contrast, the orders of the eight methods in terms of CNDVI and CNDWI are significant different from those in terms of Q8 and SCC. This indicates that a fusion method offering the best performance for a certain image in terms of quality indexes and visual inspection does not always provide the highest CNDVI and CNDWI values. Generally, the GLP_ESDM method outperforms the other methods for I1, I2 and I5, whereas the GLP_ECBD method provides the best performances for I3, I4 and I6, in terms of CNDVI and CNDWI, as well as quality indexes and visual inspection. This indicates that the GLP_ESDM is the best choice for images with similar objects with I1, I2 and I5, whereas the GLP_ECBD is the best choice for images with similar objects with I3, I4 and I6, for producing fusion products used for applications related to vegetation or water-bodies. In addition, the fusion products show limited improvements in terms of CNDVI and CNDWI. This indicates that it is hard for the fusion products to preserve the NDVI and NDWI information obtained from the corresponding up-sampled MS images. Consequently, it is necessary to evaluate fusion products using information indices (i.e., NDVI and NDWI) if fused WV-2 images will be used for applications related to vegetation and water-bodies.

4.2. Effects for Different Spectral Ranges between the PAN and MS Bands

A noticeable point for the WV-2 is that the spectral range of the PAN band covers limited portion of the spectral ranges of the C, NIR1 and NIR2 bands. This results in relative low correlation coefficients between these bands and the PAN band. It is interesting to see the performances of the selected pansharpening methods on the two NIR bands and the C band of WV-2. In order to assess the spectral distortion of each fused band, the CC value between each fused band and the corresponding reference band was calculated for each fusion product. The CC values for the fusion products of I1, I3, I5 and I6 are shown in Table 6. The CC values of I2 and I4 are not presented because they are similar with those of I1 and I3, respectively.
It can be seen from Table 6 that the CC values of the two NIR bands are significantly lower than those of the other bands for all the fusion products, indicating that the fused NIR bands show more spectral distortions than the other bands. This is caused by the relative low correlation coefficients between the two NIR bands and the PAN band. This is also revealed by previous studies, the higher the correlation between the PAN band and each MS band, the better the success of fusion [31]. Generally, the four CS methods offer higher CC values for the two NIR bands than the four MRA methods for most of the test images. This is consistent with the result of visual inspection of these fusion products. However, the two GLP-based methods, which provide good performances in terms of NDVI and NDWI information preservation, offer relative low CC values for the fused NIR1 band, due to the low CC between the PAN and the NIR1 band. This proves again that it is necessary to evaluate fusion products using information indices (i.e., NDVI and NDWI) if fused WV-2 images will be used for applications related to vegetation and water-bodies.

4.3. How to Extend the Selected Pansharpening Methods to Other HSR Satellite Images

As introduced in the previous sections, the HR, GSA, GLP_ESDM, and GLP_ECBD methods are good choices for producing fused WV-2 images used for image interpretation and applications related to urban buildings. The two GLP-based methods outperform other methods for generating fused WV-2 images used for applications related to vegetation and water-bodies. It is interesting for the readers that whether these methods give similar performances to the sensors having a similar PAN spectral range with WV-2, such as GeoEye-1, and WorldView-3/4.
Actually, the selected pansharpening methods can be categorized into two groups, according to the approaches employed to generate the synthetic PAN band P L , which mainly contains the low-frequency component of the original PAN band. For the first group, P L is generated by applying filters to the original PAN band, or by up-sampling the degraded version of the original PAN band. In contrast, for the second group, the intensity image I L , which can be seen as another approach for generating the synthetic PAN band P L , is generated using the weighted combination of the LSR MS bands. The methods belong to the first group include HR, ATWT, NSCT and the two GLP-based methods, whereas the methods belong to the second group include GS, GSA and HCS methods. For the first group, the low-frequency component of P L has relative low correlations with the C, NIR1 and NIR2 bands, but has relative high correlations with the other spectral bands. This result in the fact that the details of the PAN band have relative high correlations with the B, G, Y, R and RE bands, but relative low correlations with the C, NIR1 and NIR2 bands. This may result in the fact that a large amount of spatial details are injected into the B, G, Y, R and RE bands, but only a small amount of the spatial details are injected into the C, NIR1 and NIR2 bands, especial for the case the injection gains are determined considering the relationship between each MS band and the PAN band. For the second group, the low-frequency component of the intensity image is related or partly related to the C, NIR1 and NIR2 bands. This may result in the fact that the low-frequency component of the C, NIR1 and NIR2 bands may be injected into the B, G, Y, R and RE bands, and hence may lead to spectral distortions of these bands. An exception occurs for GSA, since the intensity image I L employed the GSA method have low CC with the C, NIR1 and NIR2 bands, due to the weights wi obtained using Equation (5) are very low for these bands.
According to the introduction about the algorithms of the selected methods, different injection gains g i are employed by these methods. The GS and GSA methods use a band-dependent model considering the relationship between each MS band and the PAN band. The GSA method outperform the GS method due to the intensity image I L employed the former have low CC values with the C, NIR1 and NIR2 bands. It can be seen from Table 6 that the CC values for the B, G, Y, R, and RE bands of the GSA-fused images are significantly higher than those of the GS-fused image. The HR method uses the SDM model, which is also band-dependent. The ATWT method employs a simple additive injection model with weights for each band equal to 1, whereas the two GLP-based methods use the ESDM and ECBD models, respectively. Among these models, only the ESDM and ECBD models consider the local dissimilarity between the MS and PAN bands. According to the experimental results, the two GLP-based methods give good performances in terms of NDVI and NDWI information preservation. This may due to the fact that only the ESDM and ECBD models consider the local dissimilarity between the MS and PAN bands. It is also demonstrated by previous studies that local dissimilarity between the MS and PAN bands should be considered by pansharpening methods to reduce spectral distortions.
As a result of the above analyses about the algorithms of the selected pansharpening methods, we can obtain the following conclusions. Firstly, for the spectral bands with relative high correlations with the PAN band, the synthetized PAN band should be obtained using the original PAN band and the injection gains should considering the relationship between each MS band and the PAN band. Secondly, for the spectral bands with relative low correlations with the PAN band, further experiments should be designed to evaluate which approach is better for generating the synthetized PAN band. However, there is no doubting that local dissimilarity between the MS and PAN bands should be considered for the fusion of these bands, i.e., the NIR band, especially for the case that the fused images will be used in applications related to vegetation and water-bodies.
According to the analysis, we can conclude that the GSA, HR, GLP_ESDM, and GLP_ECBD methods can also provide good performances for similar sensors, such as GeoEye-1, WorldView-3, WorldView-4, for the cases that the fusion products will be used in image interpretation or urban buildings. Actually, it is proved by previous studies that the performances of these newly proposed methods are sensor independent [30]. However, for the case that the fusion products will be used in applications related to vegetation or water-bodies, the GLP-ESDM and GLP_ECBD methods or other fusion methods consider local dissimilarity between the MS and PAN bands are better choices.

5. Conclusions

The performances of eight state-of-art pan-sharpening methods for WV-2 imagery were assessed using information indices (NDVI, NDWI, and MBI), along with current image quality indices (ERGAS, SAM, Q2n, and SCC) and visual inspection, with six WV-2 datasets. The main findings and conclusions derived from our analyses are as follows:
(1)
Generally, the HR, GSA, GLP_ESDM and GLP_ECBD methods give better performances than the other methods, whereas the NSCT and HCS methods offer the poorest performances, for most of the test images, in terms of quality indexes and visual inspection. Some of the fusion products generated by the GS and ATWT methods show significant spectral distortions. In addition, the performances of the eight methods in terms of CMBI are consistent with those in terms of Q8 and SCC. Consequently, the HR, GSA, GLP_ESDM, and GLP_ECBD methods are good choices if the fused WV-2 images will be used for image interpretation and applications related to urban buildings. The four methods can also provide good performances for other WV-2 image scenes, for producing fused images used for image interpretation.
(2)
The order of the pansharpening methods in terms of CNDVI is consistent with that in terms of CNDWI. This is because both of the two indexes measure the differences between inter-band relationships of the fused image and those of the reference MS image, and both of them are related to the quality of the fused NIR1 bands. The GLP_ESDM method offers higher CNDVI and CNDWI values for I1, I2 and I5, whereas the GLP_ECBD method provides higher CNDVI and CNDWI values for I3, I4 and I6, as well as good performances in terms of quality indexes and visual inspection. Consequently, the GLP_ESDM and GLP_ECBD methods are better than other methods, if the fused WV-2 images will be used for applications related to vegetation and water-bodies. However, for this case, it is better to select a best method by comparing the indexes CNDVI and CNDWI, as well as quality indexes and visual inspection, since the GLP_ESDM and GLP_ECBD methods may give different performances for images with different land cover objects.
(3)
According to the experimental results of this work and the analyses the algorithms of the selected pansharpening methods, we can offer two suggestions for the fusion of images obtained by sensors similar with WV-2, such as Geoeye-1 and Worldview-3/4. Firstly, for the spectral bands with relative high correlations with the PAN band, the synthetized PAN band should be obtained using the original PAN band and the injection gains should considering the relationship between each MS band and the PAN band. The HR, GSA, GLP_ESDM, and GLP_ECBD method also can offer good performances for scenes obtained by GeoEye-1 and Worldview-3/4, for producing fused images used for interpretation and applications related to urban buildings. Secondly, for the spectral bands with relative low correlations with the PAN band, local dissimilarity between the MS and PAN bands should be considered for the fusion of these bands, i.e., the NIR band, especially for the case that the fused images will be used in applications related to vegetation and water-bodies.

Acknowledgments

The authors wish to acknowledge the anonymous reviewers for providing helpful suggestions that greatly improved the manuscript. This research was supported by the One Hundred Person Project of Chinese Academy of Sciences (grant No. Y34005101A), Youth Foundation of Director of Institution of Remote Sensing and Digital Earth, Chinese Academy of Sciences (grant No. Y6SJ1100CX), the Key Research Program of Chinese Academy of Sciences (Grant No. ZDRW-ZS-2016-6-1-3), the National Science and Technology Support Program of China (grant No. 2015BAB05B05-02), and the “Light of West China” Program of Chinese Academy of Sciences (grant No. 2015-XBQN-A-07).

Author Contributions

Hui Li drafted the manuscript and was responsible for the research design, experiments and the analyses. Linhai Jing provided technical guide and reviewed the manuscript. Yunwei Tang processed the WV-2 data.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhang, R.; Zeng, M.; Chen, J. Study on geological structural interpretation based on Worldview-2 remote sensing image and its implementation. Procedia Environ. Sci. 2011, 10, 653–659. [Google Scholar]
  2. Jawak, S.D.; Luis, A.J. A spectral index ratio-based Antarctic land-cover mapping using hyperspatial 8-band Worldview-2 imagery. Polar Sci. 2013, 7, 18–38. [Google Scholar] [CrossRef]
  3. Ghosh, A.; Joshi, P.K. A comparison of selected classification algorithms for mapping bamboo patches in lower gangetic plains using very high resolution Worldview-2 imagery. Int. J. Appl. Earth Obs. Geoinf. 2014, 26, 298–311. [Google Scholar] [CrossRef]
  4. Mutanga, O.; Adam, E.; Cho, M.A. High density biomass estimation for wetland vegetation using Worldview-2 imagery and random forest regression algorithm. Int. J. Appl. Earth Obs. Geoinf. 2012, 18, 399–406. [Google Scholar] [CrossRef]
  5. Rapinel, S.; Clement, B.; Magnanon, S.; Sellin, V.; Hubert-Moy, L. Identification and mapping of natural vegetation on a coastal site using a Worldview-2 satellite image. J. Environ. Manag. 2014, 144, 236–246. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  6. Ozdemir, I.; Karnieli, A. Predicting forest structural parameters using the image texture derived from Worldview-2 multispectral imagery in a dryland forest, Israel. Int. J. Appl. Earth Obs. Geoinf. 2011, 13, 701–710. [Google Scholar] [CrossRef]
  7. Carper, W.J.; Lillesand, T.M.; Kiefer, R.W. The use of intensity-hue-saturation transformations for merging SPOT panchromatic and multispectral image data. Photogramm. Eng. Remote Sens. 1990, 56, 459–467. [Google Scholar]
  8. Zhou, X.; Liu, J.; Liu, S.; Cao, L.; Zhou, Q.; Huang, H. A GIHS-based spectral preservation fusion method for remote sensing images using edge restored spectral modulation. ISPRS J. Photogramm. Remote Sens. 2014, 88, 16–27. [Google Scholar] [CrossRef]
  9. Welch, R.; Ehlers, M. Merging multiresolution SPOT HRV and Landsat TM data. Photogramm. Eng. Remote Sens. 1987, 53, 301–303. [Google Scholar]
  10. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6,011,875, 4 January 2000. [Google Scholar]
  11. Aiazzi, B.; Baronti, S.; Selva, M.; Alparone, L. MS + Pan image fusion by an enhanced Gram-Schmidt spectral sharpening. In New Developments and Challenges in Remote Sensing; Bochenek, Z., Ed.; Millpress: Rotterdam, The Netherlands, 2007; pp. 113–120. [Google Scholar]
  12. Shensa, M.J. The discrete wavelet transform—Wedding the à trous and Mallat algorithms. IEEE Trans. Sign. Process. 1992, 40, 2464–2482. [Google Scholar] [CrossRef]
  13. Teggi, S.; Cecchi, R.; Serafini, F. TM and IRS-1C-PAN data fusion using multiresolution decomposition methods based on the ‘a tròus’ algorithm. Int. J. Remote Sens. 2003, 24, 1287–1301. [Google Scholar] [CrossRef]
  14. Mallat, S.G. Multifrequency channel decompositions of images and wavelet models. IEEE Trans. Acoust. Speech Sign. Process. 1989, 37, 2091–2110. [Google Scholar] [CrossRef]
  15. Yocky, D.A. Multiresolution wavelet decomposition image merger of Landsat Thematic Mapper and SPOT panchromatic data. Photogramm. Eng. Remote Sens. 1996, 62, 1067–1074. [Google Scholar]
  16. Amolins, K.; Zhang, Y.; Dare, P. Wavelet based image fusion techniques—An introduction, review and comparison. ISPRS J. Photogramm. Remote Sens. 2007, 62, 249–263. [Google Scholar] [CrossRef]
  17. Aiazzi, B.; Alparone, L.; Baronti, S.; Lotti, F. Lossless image compression by quantization feedback in a content-driven enhanced Laplacian pyramid. IEEE Trans. Image Process. 1997, 6, 831–843. [Google Scholar] [CrossRef] [PubMed]
  18. Yang, X.H.; Jiao, L.C. Fusion algorithm for remote sensing images based on nonsubsampled Contourlet transform. Acta Autom. Sin. 2008, 34, 274–281. [Google Scholar] [CrossRef]
  19. Saeedi, J.; Faez, K. A new pan-sharpening method using multiobjective particle swarm optimization and the shiftable contourlet transform. ISPRS J. Photogramm. Remote Sens. 2011, 66, 365–381. [Google Scholar] [CrossRef]
  20. Dong, W.; Li, X.E.; Lin, X.; Li, Z. A bidimensional empirical mode decomposition method for fusion of multispectral and panchromatic remote sensing images. Remote Sens. 2014, 6, 8446–8467. [Google Scholar] [CrossRef]
  21. Dong, L.; Yang, Q.; Wu, H.; Xiao, H.; Xu, M. High quality multi-spectral and panchromatic image fusion technologies based on Curvelet transform. Neurocomputing 2015, 159, 268–274. [Google Scholar] [CrossRef]
  22. Alparone, L.; Wald, L.; Chanussot, J.; Thomas, C.; Gamba, P.; Bruce, L.M. Comparison of pan-sharpening algorithms: Outcome of the 2006 GRS-S data-fusion contest. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3012–3021. [Google Scholar] [CrossRef] [Green Version]
  23. Gonzalez-Audicana, M.; Saleta, J.L.; Catalan, R.G.; Garcia, R. Fusion of multispectral and panchromatic images using improved IHS and PCA mergers based on wavelet decomposition. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1291–1299. [Google Scholar] [CrossRef]
  24. Guo, Q.; Liu, S. Performance analysis of multi-spectral and panchromatic image fusion techniques based on two wavelet discrete approaches. Optik Int. J. Light Electron Opt. 2011, 122, 811–819. [Google Scholar] [CrossRef]
  25. Chen, F.; Qin, F.; Peng, G.; Chen, S. Fusion of remote sensing images using improved ICA mergers based on wavelet decomposition. Procedia Eng. 2012, 29, 2938–2943. [Google Scholar] [CrossRef]
  26. Liu, Y.; Liu, S.P.; Wang, Z.F. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  27. Padwick, C.; Deskevich, M.; Pacifici, F.; Smallwood, S. Worldview-2 pan-sharping. In Proceedings of the 2010 Conference of American Society for Photogrammetry and Remote Sensing, San Diego, CA, USA, 21 May 2010.
  28. El-Mezouar, M.C.; Kpalma, K.; Taleb, N.; Ronsin, J. A pan-sharpening based on the non-subsampled Contourlet transform: Application to Worldview-2 imagery. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 1806–1815. [Google Scholar] [CrossRef]
  29. Du, Q.; Younan, N.H.; King, R.; Shah, V.P. On the performance evaluation of pan-sharpening techniques. IEEE Geosci. Remote Sens. Lett. 2007, 4, 518–522. [Google Scholar] [CrossRef]
  30. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A critical comparison among pansharpening algorithms. IEEE Trans. Geosci. Remote Sens. 2015, 33, 2565–2586. [Google Scholar] [CrossRef]
  31. Thomas, C.; Ranchin, T.; Wald, L.; Chanussot, J. Synthesis of multispectral images to high spatial resolution: A critical review of fusion methods based on remote sensing physics. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1301–1312. [Google Scholar] [CrossRef] [Green Version]
  32. Zhang, Y.; Mishra, R.K. A review and comparison of commercially available pan-sharpening. In Proceedings of the International Geoscience and Remote Sensing Symposium (IGARSS), Munich, Germany, 22–27 July 2012.
  33. Yuhendra; Alimuddin, I.; Sumantyo, J.T.S.; Kuze, H. Assessment of pan-sharpening methods applied to image fusion of remotely sensed multi-band data. Int. J. Appl. Earth Obs. Geoinf. 2012, 18, 165–175. [Google Scholar] [CrossRef]
  34. Nikolakopoulos, K. Quality assessment of ten fusion techniques applied on Worldview-2. Eur. J. Remote Sens. 2015, 48, 141–167. [Google Scholar] [CrossRef]
  35. Maglione, P.; Parente, C.; Vallario, A. Pan-sharpening Worldview-2: IHS, Brovey and Zhang methods in comparison. Int. J. Eng. Technol. 2016, 8, 673–679. [Google Scholar]
  36. Huang, X.; Wen, D.; Xie, J.; Zhang, L. Quality assessment of panchromatic and multispectral image fusion for the ZY-3 satellite: From an information extraction perspective. IEEE Geosci. Remote Sens. Lett. 2014, 11, 753–757. [Google Scholar] [CrossRef]
  37. Ghassemian, H. A review of remote sensing image fusion methods. Inf. Fusion 2016, 32, 75–89. [Google Scholar] [CrossRef]
  38. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS + Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  39. Jing, L.; Cheng, Q. Two improvement schemes of PAN modulation fusion methods for spectral distortion minimization. Int. J. Remote Sens. 2009, 30, 2119–2131. [Google Scholar] [CrossRef]
  40. Nunez, J.; Otazu, X.; Fors, O.; Prades, A.; Pala, V.; Arbiol, R. Multiresolution-based image fusion with additive Wavelet decomposition. IEEE Trans. Geosci. Remote Sens. 1999, 37, 1204–1211. [Google Scholar] [CrossRef] [Green Version]
  41. Vivone, G.; Restaino, R.; Dalla Mura, M.; Licciardi, G.; Chanussot, J. Contrast and error-based fusion schemes for multispectral image pansharpening. IEEE Geosci. Remote Sens. Lett. 2014, 11, 930–934. [Google Scholar] [CrossRef]
  42. Garzelli, A.; Nencini, F. Interband structure modeling for Pan-sharpening of very high-resolution multispectral images. Inf. Fusion 2005, 6, 213–224. [Google Scholar] [CrossRef]
  43. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. MTF-tailored multiscale fusion of high-resolution MS and PAN imagery. Photogramm. Eng. Remote Sens. 2006, 72, 591–596. [Google Scholar] [CrossRef]
  44. Ranchin, T.; Aiazzi, B.; Alparone, L.; Baronti, S.; Wald, L. Image fusion—The ARSIS concept and some successful implementation schemes. ISPRS J. Photogramm. Remote Sens. 2003, 58, 4–18. [Google Scholar] [CrossRef]
  45. Yuhas, R.; Goetz, A.; Boardman, J. Discrimination among Semi-Arid Landscape Endmembers Using the Spectral Angle Mapper (sam) Algorithm. In Summaries of the Third Annual JPL Airborne Geoscience Workshop; Jet Propulsion Laboratory: Pasadena, CA, USA, 1992; pp. 147–149. [Google Scholar]
  46. Alparone, L.; Baronti, S.; Garzelli, A.; Nencini, F. A global quality measurement of pan-sharpened multispectral imagery. IEEE Geosci. Remote Sens. Lett. 2004, 1, 313–317. [Google Scholar] [CrossRef]
  47. Garzelli, A.; Nencini, F. Hypercomplex quality assessment of multi/hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2009, 6, 662–665. [Google Scholar] [CrossRef]
  48. Otazu, X.; Gonzalez-Audicana, M.; Fors, O.; Nunez, J. Introduction of sensor spectral response into image fusion methods: Application to wavelet-based methods. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef] [Green Version]
  49. Huang, X.; Zhang, L. Morphological building/shadow index for building extraction from high-resolution imagery over urban areas. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2012, 5, 161–172. [Google Scholar] [CrossRef]
  50. Chavez, P.S., Jr. An improved dark-object subtraction technique for atmospheric scattering correction of multispectral data. Remote Sens. Environ. 1988, 24, 459–479. [Google Scholar] [CrossRef]
  51. Chavez, P.S. Image-based atmospheric corrections—Revisited and improved. Photogramm. Eng. Remote Sens. 1996, 62, 1025–1036. [Google Scholar]
  52. Moran, M.S.; Jackson, R.D.; Slater, P.N.; Teillet, P.M. Evaluation of simplified procedures for retrieval of land surface reflectance factors from satellite sensor output. Remote Sens. Environ. 1992, 41, 169–184. [Google Scholar] [CrossRef]
  53. Zhang, Q.; Guo, B.L. Multifocus image fusion using the nonsubsampled contourlet transform. Sign. Process. 2009, 89, 1334–1346. [Google Scholar] [CrossRef]
  54. Wald, L. Quality of high resolution synthesised images: Is there a simple criterion? In Proceedings of the International Conference on Fusion Earth Data, Sophia Antipolis, France, 26–28 January 2000.
  55. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Signal Process. Lett. 2002, 9, 81–84. [Google Scholar] [CrossRef]
  56. Liu, J.; Wang, X.; Chen, M.; Liu, S.G.; Shao, Z.F.; Zhou, X.R.; Liu, P. Illumination and contrast balancing for remote sensing images. Remote Sens. 2014, 6, 1102–1123. [Google Scholar] [CrossRef]
  57. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Geosci. Remote Sens. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  58. Alparone, L.; Alazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef]
  59. Smith, W.D. Quaternions, Octonions, and Now, 16-ons and 2n-ons, New Kinds of Numbers, 2004. Available online: http://www.scorevoting.net/WarrenSmithPages/homepage/nce2.pdf (accessed on 30 September 2014).
  60. Ebbinghaus, H.B. Numbers; Springer: New York, NY, USA, 1991. [Google Scholar]
  61. Jing, L.; Cheng, Q.; Guo, H.; Lin, Q. Image misalignment caused by decimation in image fusion evaluation. Int. J. Remote Sens. 2012, 33, 4967–4981. [Google Scholar] [CrossRef]
Figure 1. The locations of the three WV-2 scenes employed in this work. The locations of the six test images used in the experiments are marked by the rectangles in red.
Figure 1. The locations of the three WV-2 scenes employed in this work. The locations of the six test images used in the experiments are marked by the rectangles in red.
Sensors 17 00089 g001
Figure 2. The selected six dataset used in the experiments. (a) I1; (b) I2; (c) I3; (d) I4; (e) I5; (f) I6.
Figure 2. The selected six dataset used in the experiments. (a) I1; (b) I2; (c) I3; (d) I4; (e) I5; (f) I6.
Sensors 17 00089 g002
Figure 3. The degraded and 2-m fused images of I1, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Figure 3. The degraded and 2-m fused images of I1, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Sensors 17 00089 g003
Figure 4. The degraded and 2-m fused images of I2, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Figure 4. The degraded and 2-m fused images of I2, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Sensors 17 00089 g004
Figure 5. The original and 2-m fused images of I3, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Figure 5. The original and 2-m fused images of I3, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Sensors 17 00089 g005
Figure 6. The original and 2-m fused images of I4, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Figure 6. The original and 2-m fused images of I4, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Sensors 17 00089 g006
Figure 7. The original and 2-m fused images of I5, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Figure 7. The original and 2-m fused images of I5, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Sensors 17 00089 g007
Figure 8. The original and 2-m fused images of I6, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Figure 8. The original and 2-m fused images of I6, shown in band 5-7-2 composition. (a) 8-m MS; (b) 2-m PAN; (c) True 2-m MS; (d) GS; (e) GSA; (f) HR; (g) HCS; (h) ATWT; (i) GLP_ESDM; (j) GLP_ECBD; and (k) NSCT_M2.
Sensors 17 00089 g008
Table 1. The description of the selected six datasets.
Table 1. The description of the selected six datasets.
ImageLocationTypeObjectsInformation Indices
I1BeijingUrbanHigh buildings, squares, roads, vegetation, shadowsMBI, NDVI
I2BeijingUrbanModerate buildings, squares, roads, vegetation, shadowsMBI, NDVI
I3PingdingshanSuburbanLow buildings, squares, roads, vegetation, shadows, water bodiesMBI, NDVI, NDWI
I4PingdingshanSuburbanLow buildings, squares, roads, vegetation, shadows, water bodiesMBI, NDVI, NDWI
I5PingdingshanRuralBuilding, roads, farms, water bodiesNDVI, NDWI
I6PingdingshanRuralVegetation, water bodies, bare soilsNDVI, NDWI
Table 2. Quality indices of the fused images generated by the selected algorithms for the two urban images.
Table 2. Quality indices of the fused images generated by the selected algorithms for the two urban images.
ImageMethodERGASSAMQ8SCC
I1GS1.642.190.9570.911
GSA1.281.920.9740.908
HR1.261.760.9770.911
HCS1.952.580.8970.881
ATWT2.092.720.8790.863
GLP_ESDM1.562.050.9560.898
GLP_ECBD1.932.360.9560.888
NSCT_M22.223.660.8490.861
EXP1.642.190.8570.582
I2GS2.703.790.9090.844
GSA2.223.450.9460.839
HR2.203.140.9510.849
HCS2.873.990.8500.809
ATWT2.764.020.8610.799
GLP_ESDM2.513.470.9160.824
GLP_ECBD3.144.260.9050.785
NSCT_M23.045.030.8310.795
EXP3.843.990.7970.499
Table 3. Quality indices of the fused images generated by the selected algorithms for the two suburban images.
Table 3. Quality indices of the fused images generated by the selected algorithms for the two suburban images.
ImageMethodERGASSAMQ8SCC
I3GS1.311.900.9080.885
GSA1.031.760.9420.883
HR1.382.230.9270.849
HCS1.281.980.8810.871
ATWT1.211.930.9070.872
GLP_ESDM1.391.910.9020.835
GLP_ECBD1.331.910.9210.856
NSCT_M21.392.410.8870.869
EXP1.621.980.8370.796
I4GS1.742.850.8710.841
GSA1.432.740.9190.831
HR1.843.250.8930.779
HCS1.743.010.8500.817
ATWT1.662.980.8680.821
GLP_ESDM1.822.820.8760.796
GLP_ECBD1.792.880.8920.805
NSCT_M21.883.640.8430.818
EXP2.213.010.7590.669
Table 4. Quality indices of the fused images generated by the selected algorithms for the two rural images.
Table 4. Quality indices of the fused images generated by the selected algorithms for the two rural images.
ImageMethodERGASSAMQ8SCC
I5GS1.662.410.8220.838
GSA1.562.660.8870.799
HR1.322.060.9140.882
HCS1.612.230.8370.861
ATWT1.592.370.8530.857
GLP_ESDM1.491.990.8750.874
GLP_ECBD1.902.450.8740.845
NSCT-M21.712.830.8290.855
EXP2.302.230.7480.665
I6GS1.882.950.7620.756
GSA1.161.750.8570.785
HR1.081.600.8730.812
HCS1.051.510.7350.848
ATWT0.971.490.8580.868
GLP_ESDM1.121.520.8570.809
GLP_ECBD0.991.440.8680.854
NSCT-M21.101.770.8200.865
EXP1.091.510.8180.849
Table 5. The CC values for the information indices for the fused images generated by the selected eight algorithms.
Table 5. The CC values for the information indices for the fused images generated by the selected eight algorithms.
MethodCMBICNDVICNDWI
I1I2I3I4I1I2I3I4I5I6I3I4I5I6
GS0.9730.9720.9690.9340.9220.8770.9290.8910.9120.9460.9150.8580.8760.950
GSA0.9750.9790.9810.9590.9150.8550.9240.8790.8840.9660.9080.8530.8300.977
HR0.9760.9780.9770.9500.9230.8850.8990.8160.9270.9690.8760.7800.8950.981
HCS0.9620.9490.9230.8480.9260.8830.9170.8800.9400.9740.9110.8590.9170.986
ATWT0.9650.9560.9340.8820.9110.8490.9280.8900.9280.9750.9120.8510.9150.986
GLP_ESDM0.9650.9600.9450.8990.9280.8870.9150.8710.9420.9720.9010.8410.9220.984
GLP_ECBD0.9640.9570.9610.9260.8920.8110.9250.8760.9060.9770.9090.8510.8600.988
NSCT_M20.9650.9480.9050.8420.8600.7780.8980.8620.8960.9680.8730.8000.8710.982
EXP0.8250.7980.8290.7830.9260.8830.9170.8800.9400.9740.9110.8590.9170.986
Table 6. The CC values for the fused bands for I1, I3, I5 and I6.
Table 6. The CC values for the fused bands for I1, I3, I5 and I6.
ImageMethodCC
CBGYRRENIR1NRI2Avg
I1GS0.9850.9850.9880.9920.9900.9900.9780.9780.986
GSA0.9870.9870.9910.9950.9930.9920.9780.9770.988
HR0.9900.9900.9930.9960.9940.9930.9770.9770.989
HCS0.8190.9450.9820.9890.9870.9840.9700.9690.956
ATWT0.8490.9310.9850.9910.9860.9870.9740.9710.959
GLP_ESDM0.9590.9790.9860.9900.9900.9880.9750.9730.980
GLP_ECBD0.9820.9820.9860.9890.9860.9850.9670.9590.980
NSCT_M20.8050.9060.9850.9930.9870.9880.9710.9650.950
EXP0.9520.9480.9440.9420.9390.9280.9170.9160.936
I3GS0.9580.9570.9620.9670.9620.9520.8990.9030.945
GSA0.9770.9770.9860.9900.9840.9770.9060.9060.963
HR0.9760.9750.9830.9860.9800.9690.8300.8230.940
HCS0.8360.9460.9770.9790.9690.9680.8920.8930.933
ATWT0.9180.9320.9800.9820.9750.9710.9000.8980.945
GLP_ESDM0.9190.9640.9780.9850.9800.9660.8420.8440.935
GLP_ECBD0.9620.9630.9750.9820.9750.9650.8760.8610.945
NSCT_M20.8860.9060.9800.9780.9690.9680.8720.8630.928
EXP0.9220.9210.9210.9230.9200.9100.8860.8850.911
I5GS0.9580.9510.9450.9470.9530.9080.8680.8750.926
GSA0.9690.9640.9630.9620.9670.9810.9490.9570.964
HR0.9680.9660.9740.9680.9670.9840.9540.9650.968
HCS0.7390.8920.9540.9660.9670.9840.9620.9730.930
ATWT0.9010.9240.9760.9670.9650.9870.9660.9750.958
GLP_ESDM0.9130.9620.9750.9700.9700.9840.9490.9630.961
GLP_ECBD0.9420.9350.9590.9670.9680.9850.9640.9740.962
NSCT_M20.8580.8880.9700.9600.9570.9810.9570.9680.943
EXP0.9400.9320.9230.9250.9310.9760.9660.9740.946
I6GS0.9570.9590.9670.9670.9660.9480.8020.8000.921
GSA0.9700.9740.9850.9820.9800.9630.7500.7470.919
HR0.9790.9800.9870.9870.9850.9690.8580.8540.950
HCS0.8790.9660.9780.9710.9640.9360.8480.8460.924
ATWT0.8960.9240.9800.9790.9710.9500.8580.8620.928
GLP_ESDM0.9050.9620.9760.9780.9790.9550.8690.8480.934
GLP_ECBD0.9560.9600.9720.9710.9670.9450.7900.8290.924
NSCT_M20.8730.9050.9820.9780.9680.9530.8010.8290.911
EXP0.9040.8990.8930.8990.9000.7880.7470.7480.848

Share and Cite

MDPI and ACS Style

Li, H.; Jing, L.; Tang, Y. Assessment of Pansharpening Methods Applied to WorldView-2 Imagery Fusion. Sensors 2017, 17, 89. https://doi.org/10.3390/s17010089

AMA Style

Li H, Jing L, Tang Y. Assessment of Pansharpening Methods Applied to WorldView-2 Imagery Fusion. Sensors. 2017; 17(1):89. https://doi.org/10.3390/s17010089

Chicago/Turabian Style

Li, Hui, Linhai Jing, and Yunwei Tang. 2017. "Assessment of Pansharpening Methods Applied to WorldView-2 Imagery Fusion" Sensors 17, no. 1: 89. https://doi.org/10.3390/s17010089

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop