Next Article in Journal
A Low-Cost Open Hardware System for Collecting Traffic Data Using Wi-Fi Signal Strength
Previous Article in Journal
Integration of Autonomous Wireless Sensor Networks in Academic School Gardens
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An IHS-Based Pan-Sharpening Method for Spectral Fidelity Improvement Using Ripplet Transform and Compressed Sensing

1
School of Urban Design, Wuhan University, 8 Donghu South Road, Wuhan 430072, China
2
Collaborative Innovation Center of Geospatial Technology, 129 Luoyu Road, Wuhan 430079, China
3
State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan 430079, China
*
Author to whom correspondence should be addressed.
Sensors 2018, 18(11), 3624; https://doi.org/10.3390/s18113624
Submission received: 27 August 2018 / Revised: 14 October 2018 / Accepted: 23 October 2018 / Published: 25 October 2018
(This article belongs to the Section Remote Sensors)

Abstract

:
Pan-sharpening aims at integrating spectral information from a multi-spectral (MS) image and spatial information from a panchromatic (PAN) image in a fused image with both high spectral and spatial resolutions. Numerous pan-sharpening methods are based on intensity-hue-saturation (IHS) transform, which may cause evident spectral distortion. To address this problem, an IHS-based pan-sharpening method using ripplet transform and compressed sensing is proposed. Firstly, the IHS transform is applied to the MS image to separate intensity components. Secondly, discrete ripplet transform (DRT) is implemented on the intensity component and the PAN image to obtain multi-scale sub-images. High-frequency sub-images are fused by a local variance algorithm and, for low-frequency sub-images, compressed sensing is introduced for the reconstruction of the intensity component so as to integrate the local information from both the intensity component and the PAN image. The specific fusion rule is defined by local difference. Finally, the inverse ripplet transform and inverse IHS transform are coupled to generate the pan-sharpened image. The proposed method is compared with five state-of-the-art pan-sharpening methods and also the Gram-Schmidt (GS) method through visual and quantitative analysis of WorldView-2, Pleiades and Triplesat datasets. The experimental results reveal that the proposed method achieves relatively higher spatial resolution and more desirable spectral fidelity.

Graphical Abstract

1. Introduction

Remote sensing (RS) is a general approach for the knowledge extraction of the Earth’s surface structure and content through acquiring and interpreting the spectral characteristics from a great distance [1]. However, the well-known trade-off between spatial resolution and spectral resolution has always precluded the further application of RS products. Pan-sharpening is a desirable solution to settle such dilemma. In fact, most of the Earth observation satellite images, such as IKONOS, QuickBird, and the WorldView family (including WorldView-2/3/4), can only provide a panchromatic (PAN) image with high spatial resolution but low spectral resolution together, with a multi-spectral (MS) image with low spatial resolution but high spectral resolution, respectively [2]. Pan-sharpening, a special case of image fusion, is capable of obtaining an image with both high spatial and high spectral resolutions. In other words, the technique provides a fused image with the same spectral response as the MS image and with the spatial resolution of the PAN image. The pan-sharpened images have been proved to possess great potential for multiple applications, including land use/land cover (LULC) classification, land change detection and environment monitoring [3,4,5,6,7,8].
Generally, pan-sharpening methods can be classified into three main categories: component substitution (CS)-based methods, multi-resolution analysis (MRA)-based methods, and model-based methods. In addition, there are a number of subcategories of pan-sharpening methods. Detailed information regarding these subcategories has been investigated in the references [1,9,10,11]. Specifically, CS-based methods include intensity-hue-saturation (IHS) [12,13,14], the Gram-Schmidt (GS) method [1,15], and principle component analysis (PCA) [16,17]. CS-based methods firstly separate components into other color space by spectral transform (such as IHS, GS, and PCA). Secondly, the component with sufficient spatial details is substituted by the PAN image to enhance spatial resolution. Finally, the pan-sharpened MS images are obtained by performing an inverse spectral transform. The second group (MRA-based methods) usually applies digital filters, covering wavelet transform [18,19], contourlet transform [20], curvelet transform [21], Laplacian pyramids [22] and ripplet transform [23] to obtain the multi-scale representation of the MS images. The MRA-based approaches [24] firstly employ MRA transforms to decompose PAN images into multi-scale components. Then, the equivalent spatial information contained in the high-frequency component is injected into an up-sampled MS image. Detailed illustrations of CS- and MRA-based methods are well documented [24]. In addition, the edge-preserving filters are frequently integrated with MRA transforms [25]. The third category is model-based methods, such as compressed sensing-based methods [26,27] and sparse matrix factorization technique-based methods [28]. The compressed sensing-based methods perform pan-sharpening by reconstructing high-resolution MS images using image patches extracted from both source PAN and MS images. Each category of pan-sharpening method possesses inherent superiorities and limitations. CS-based methods, especially the IHS method, can improve the spatial resolution of MS images significantly while preserving high convenience and efficiency for implementation. However, CS-based methods result in spectral distortion in the substitution of components due to the ignorance of local differences and the spectral mismatch that exists between the PAN image and the corresponding MS image [12,29,30]. Compared to CS-based methods, MRA-based methods achieve superior spectral fidelity but the improvement of spatial resolution is limited. The characteristics of model-based methods will not be discussed as the merits and demerits of such methods vary with the specific model applied. For further knowledge, multiple studies are available for inspection [1,4,9,31,32,33,34,35]. In addition, the spectral characteristics of the PAN image is claimed to have impacts on the quality of pan-sharpened image [36].
The exploration of compressed sensing in improving pan-sharpening performance is well documented [37,38]. Compressed sensing is able to reconstruct the sparse signal (e.g., digital image) [39,40,41] in a linear combination of sparse and low-dimensional projection. Pan-sharpening methods based on compressed sensing are claimed [38,42,43] to perform better than conventional methods. For instance, the method proposed by Yang and Li [43] utilizes a slide window algorithm to obtain image patches. Then, a discrete cosine transform (DCT) dictionary and orthogonal matching pursuit (OMP) algorithm are integrated to fuse the images. Li and Yang [26] conduct pan-sharpening based on compressed sensing by randomly selecting basis atoms in a trained dictionary extracted from the PAN and MS images. Ghahremani and Ghassemian [23] propose the compressed-sensing-injection (CSI) method using a dictionary consisting only of the spatial details extracted from the PAN images by ripplet transform. Furthermore, an iterative method improving on the basis of the CSI method is proposed by Ghahremani and Ghassemian [27] to implement image pan-sharpening with the aim of preserving original spectral information. In such a method, a dictionary extracted only from the PAN images is used to reconstruct the high spatial resolution intensity images. This method is more efficient and effective for the reconstruction of high-resolution MS images compared with other studies, while only using high-resolution intensity and its sparsest coefficient. In addition, Wang et al. [44] construct the dictionary by using several sub-dictionaries composed of image patches. The patches are directly extracted from each spectral channel of an MS image. The dictionary construction takes the correlation between PAN and MS images into consideration to transform pan-sharpening into image restoration.
However, most IHS-based algorithms utilize MRA transforms to improve spectral fidelity without the aid of a compressed sensing technique, which limits the local information preservation of the pan-sharpened image [16,18,45,46]. A few methods utilize MRA tools and compressed sensing, e.g., the method based on wavelet transform and compressed sensing proposed by Cheng et al. [47]. Nevertheless, most existing studies possess intrinsic limitations in the extraction and reconstruction of spatial details using conventional MRA tools. Conventional MRA tools (e.g., discrete wavelet transform and à trous wavelet transform), have limited capability to resolve one-dimensional (1D) singularities or two-dimensional (2D), isotropic singularities. Hence, the pan-sharpened images usually suffer from Gibbs phenomena. Furthermore, curvelet transform is suggested as representing 2D singularities more efficiently. However, the scaling rule to achieve anisotropic directionality has not yet been thoroughly demonstrated.
In recent years, the shift invariant ripplet transform (RT) [48] has been proposed and extensively adopted in feature extraction. RT is able to represent 2D singularities (e.g., edges and textures) more efficiently at multiple scales and in multiple directions along a family of curves. Several studies [23,48,49] demonstrate that RT is superior to wavelet- and curvelet-based MRA transforms in terms of the effectiveness and efficiency of edge and texture extraction and reconstruction. In view of this, an IHS-based pan-sharpening method combining the high spectral fidelity of discrete ripplet transform (DRT) [49] and desirable local information preservation of compressed sensing [42,50] is proposed in this paper. Specifically, IHS transform is firstly performed on MS images to obtain the intensity component. Then, smoothing filter-based intensity modulation (SFIM) [51] is introduced to modulate spatial information contained in the intensity component without altering its spectral characteristic. The ripplet transform [23] is implemented on the intensity component and the PAN images to extract high-frequency (HF) and low-frequency (LF) sub-images. Subsequently, compressed sensing technique is applied to reconstruct the LF sub-images. Various fusion rules are adopted for the HF sub-images and LF sub-images according to their specific characteristics. The local variance algorithm [39] and local difference weighted algorithm are respectively selected as the fusion rules for HF and LF coefficients. Finally, the pan-sharpened image is obtained by inverse ripplet transform and inverse IHS transform. Experiments on WorldView-2, Pleiades and Triplesat datasets corroborate that the proposed method ensures superior spectral fidelity compared to several state-of-the-art approaches.
The reminder of the paper is structured as follows. Section 2 briefly reviews the related works. The proposed method is introduced in Section 3. The experimental results implemented on multiple RS images and discussion are provided in Section 4. Section 5 draws a brief conclusion of this paper.

2. Related Work

The four relevant techniques utilized in pan-sharpening are briefly reviewed here. Section 2.1 discusses the principle of the IHS-based pan-sharpening method and the cause of spectrum distortion. Section 2.2 introduces discrete ripplet transform. Sparse representation of images and the compressed sensing theory are briefly illustrated in Section 2.3.

2.1. IHS Transform and Spectral Distortion

The framework of IHS and IHS-like pan-sharpening methods is firstly presented in [12] and thoroughly illustrated in [52] as:
M F b = M U P b + G b · ( P H M I U P ) , b = 1 , , B
where M F is the pan-sharpened image, M denotes the original MS image, I is the intensity component separated by IHS transform, G represents the vector modulating the gains of spatial details injection, B is the number of bands in the MS image, and b denotes the b th band of the pan-sharpened or MS image. The subscript “ U P ” denotes the up-sampling process. In practice, the up-sampled image is produced by cubic convolution interpolation for operational convenience. It is noteworthy that this process can be implemented more properly by integrating with up-sampling, a low-pass filter and modulate transfer function (MTF) [53,54]. P H M is the histogram-matched PAN image. The up-sampled I component is generated by:
I U P = b = 1 B w b M U P b
where w is the weight vector which can be empirically set as 1 B [13] and can be experimentally chosen to weigh the spectral overlapping between PAN and MS images [24].
According to reference [12], in the inverse IHS transform for obtaining a pan-sharpened image, the input hue ( H ) and saturation ( S ) components remain unchanged, while the input ( I ) component is replaced by P H M , essentially. We denote intensity variation ( δ ) occurs in the IHS-based pan-sharpening method which can be measured as:
δ = b = 1 B M F b M U P b = P H M I U P
Due to the sensor type, local dissimilarities, and spectral mismatch between the PAN and MS images [12,29,30], δ is not equal to zero. Furthermore, the variation of I leads to a distorted S component of the pan-sharpened image. Thus, under the prerequisite of instrumentally and temporally identical pan-sharpening, spectral distortion can be remitted by decreasing δ by means of reconstructing a new intensity component that combines spectral characteristics of I and the PAN image. The proposed method is developed based on this concept. Remarkably, the maximization of image quality score index Q4 [52] is an ideal approach for distortion minimization, which is proved superior in extensive practice.

2.2. Discrete Ripplet Transform (DRT)

Ripplet transform [49] is proposed to address the problem that 2-D singularities along arbitrarily shaped curves cannot be resolved by conventional wavelet transform. RT generalizes the curvelet by introducing two parameters, i.e., support, c , and degree, d . The two parameters empower RT with the anisotropic capability to represent singularities along arbitrarily shaped curves.
The discretized representation of RT (i.e., DRT) is actually based on the discretization of the parameters of ripplets, which is similar to discrete curvelet transform [55]. In reference [48,49], DRT is defined at scale a j = 2 j , orientation θ l = 2 Π c · 2 j ( 1 1 / d ) · l , and position x k j , l = R θ l 1 ( c 2 j k 1 , 2 j / d k 2 ) , where k = [ k 1 , k 2 ] T , j , k 1 , k 2 , l , d , m , n N ,   and   d = n m . In the frequency domain, we can represent the frequency response of the ripplet function as:
ρ j ^ ( r , w ) = c 1 2 a m + n 2 n W ( 2 j · r ) V ( c 1 · 2 j m n n · w l )
where the radial window W and the angular window V satisfy the following admissible conditions:
j = 0 + | W ( 2 j · r ) | 2 = 1
i = + | V ( c 1 · 2 1 d 1 · w l ) | 2 = 1
given c , d and j . The “wedge” corresponding to the ripplet function in the frequency domain is as follows:
H j , l ( r , θ ) = { 2 j | r | 2 2 j , | θ 2 1 d 1 · l | π 2 2 j }
The DRT of an M × N image f ( n 1 , n 2 ) will be in the form of
R j , k , l = n 1 = 0 M 1 n 2 = 0 N 1 f ( n 1 , n 2 ) ρ j , k , l ( n 1 , n 2 ) ¯
The image can be reconstructed through inverse discrete ripplet transform (IDRT):
f ˜ ( n 1 , n 2 ) = j k l R j , k , l ρ j , k , l ( n 1 , n 2 )

2.3. Compressed Sensing and Sparse Representation

The basic assumption of images in sparse representation (SR) is that each original image can be represented as a linear combination of small number of atoms from a specific dictionary [42,56,57]. The main information, including local spatial features and internal structure of original images, can be concentrated in the atoms using SR. With the assumption that an original image patch s has a consistent structure, the sparse representation α of the original image patch using a specific dictionary D can be modeled as:
α ^ = m i n   α 0 s . t . D α s 2 2 = 0
where · d = 0 , 2 denotes the L d -norm, and α 0 is the number of non-zero components in α . Practically, directed by compressed sensing theory [42,50], we decimate a few measurements of y using decimation matrix M as:
y = M s
where M n × k , k is the number of atoms in the dictionary, and n k . Combining with the sparse representation of images and compressed sensing [27], the sparsest α ensures that image patch s can be recovered from y through the optimization process as:
α ^ = m i n   α 0 s . t . y M D α 2 2 ε
where s = D α , and ε is a fairly small constant that denotes the error tolerance. The optimization of Equation (12) is a Non-deterministic Polynomial (NP)-hard problem [26,27]. According to references [26,58], Equation (12) can be converted into the L 1 -norm minimization problem as:
α ^ = m i n   α 1 s . t . y M D α 2 2 ε
The basis pursuit (BP) algorithm [59] is the well-known and most commonly used optimization method to solve Equation (13). The image patches reconstruction using the BP algorithm can be represented in the standard linear formation as:
m i n c T θ s . t . A θ = b ,   θ 0
where θ = ( α + , α ) ,   c = ( 1 ; 1 ) , A = ( M D ;   M D ) ,   b = y , and α + and α are the positive and negative value vectors of α , respectively. α ^ in Equation (12) is obtained by α ^ = α + α . The sparsely reconstructed image is obtained by s ^ = D α ^ .

3. Proposed Method

Based on the concept introduced in Section 2.1, the proposed method focuses on reducing spectral distortion δ by obtaining a reconstructed I component with spatial details and spectral characteristics from the PAN and MS images, respectively. We perform SFIM on the I component and the PAN image at first, and then utilize DRT to obtain HF and LF sub-images of the I component. The local variance algorithm is adopted as the fusion rule for the HF coefficient containing spatial information such as textures and edges. The LF coefficient is reconstructed using a compressed sensing technique and consequently fused according to local differences.

3.1. Spectral Fidelity Improvement of Intensity Component Using Sfim

In the proposed method, SFIM is performed on the PAN images and the I component to generate the new I component ( I S F I M ) with modulated spatial details while spectral diversity is maintained. SFIM has two advantages which encourage its adoption for improving the spectral fidelity of I S F I M . On one hand, the smoothing filter kernel of SFIM leads to a reduction of co-registration accuracy sensitivity. On the other hand, SFIM can be performed on individual color components. The smoothing filter kernel is an average function to obtain a local average on each pixel in a high-resolution PAN image. A smoothing filter kernel with size of s × s can be represented as:
1 s ( 1 1 1 1 )
SFIM can be represented as:
I S F I M ( x , y ) = I ( x , y ) · P ( x , y ) P M e a n ( x , y )
where I ( x , y ) is a pixel of the I component, P ( x , y ) is the corresponding pixel in the PAN image, and P M e a n is the smoothed mean image of the PAN image. The size of the kernel is dependent on the ratio of resolution between the PAN and MS images, which generally equals 4 for most high resolution satellite images [51], e.g., WorldView-2, Pleiades-1A, or Triplesat (Table 1). The kernel size of the smoothing filter is not constant but changes with the experiment data. The optimization of kernel size of SFIM is experimentally discussed in Section 4.3.1.

3.2. Intensity Component Reconstruction Using DRT and Compressed Sensing

DRT and compressed sensing are combined to obtain a reconstructed I component ( I R ) so as to adequately integrate the local spatial details and spectral information of the I component and the PAN image while minimizing the local dissimilarities and spectral mismatch between the two [12,29,30]. Specifically, DRT is advantageous to preserve the spectral features and spatial structures of source images. Compressed sensing is effective in extracting local spatial information by using image pairs [47]. In this study, two-level DRT with c = 1 and d = 4 is applied to obtain the HF and LF sub-images of the I component and the PAN image; note that c = 1 and d = 4 are shown to be sufficiently high for DRT to capture 2-D anisotropic details in high spatial resolution satellite images, such as IKONOS and QuickBird [23].

3.2.1. High-Frequency Sub-Images Fusion

The HF sub-images decomposed by DRT contain sufficient spatial details of original images. The local variation of wavelet energy can represent the abundance of local information contained in the original images. The local variance of wavelet energy is utilized to fuse the high-frequency sub-images of the I component ( H F I ) and high-frequency sub-images of PAN images ( H F P ) without losing spatial details or distorting spectral characteristics [60,61,62]. Following the set-up in [61], we define the local region Q in the image with size 3 × 3 , whose centre is ( x , y ) . For every region Q in H F I or H F P , the principle of HF sub-images fusion rule is defined as:
V ( x , y ) = 1 3 × 3 ( H ( x , y ) H ¯ ) 2
where V is the local variance value of the pixel, H ¯ is the mean value of a specific local region in H F I or H F P , and H ( x , y ) is the central pixel value. The fused high-frequency sub-images ( H F F ) can be obtained as:
H F F ( x , y ) = { H F I ( x , y ) , V H F I ( x , y ) > V H F P ( x , y ) H F P ( x , y ) , V H F I ( x , y ) V H F P ( x , y )
where V H F I ( x , y ) and V H F P ( x , y ) are the local variance values of a specific pixel H F I ( x , y ) or H F P ( x , y ) , respectively.

3.2.2. Low-Frequency Sub-Images Fusion

The low-frequency sub-images of the I component ( L F I ) and the PAN image ( L F P ) decomposed by DRT contain the majority of the energy and spectral information of the I component and the PAN image, which are essential for pan-sharpening. Sparse representation and compressed sensing are introduced to get the representation of low-frequency sub-images efficiently and precisely. Predefined dictionaries are usually adopted to reconstruct images in multiple existing studies of image sparse representation and compressed sensing [43,56,63]. However, spectral characteristics of satellite images vary according to the sensor type, imaging location and imaging time [47]. Therefore, many pan-sharpening methods based on sparse representation and compressed sensing establish the dictionaries through image patches extracted from the original satellite images [23,27,30,43,44,64].
We construct the dictionary D using image patches of L F P sub-images. To combine the local information of L F I and L F P , we divide L F I and L F P into K small and partially overlapped image patches with size l × l and overlapping ratio σ using the patch-by-patch strategy. Following the set-up in the reference [27], all K patches from L F P sub-images are normalized, from which is subtracted the mean of all patches to form the dictionary D . We denote D = { d 1 , d 2 , , d K } is the dictionary to be calculated, and the Gaussian random matrix is adopted as the decimation matrix M . Thus, each patch of L F I (i.e., L F I k , k { 1 ,   , K } ), should be reconstructed into L F I R k according to Equation (13) as:
α L F I k ^ = m i n   α k 1 s . t .   y L F I k M D α k 2 2 ε , k = 1 ,   , K
where y L F I k is column vector lexicographically rearranged from L F I k . According to Section 2.3, the estimated k th patch is generated as L F I R k ^ = D α L F I k ^ via the BP algorithm.
To ensure the fused patch is capable of integrating both the local wavelet energy and structures from L F I and L F P , the local difference weighted algorithm [62,65] is adopted to fuse L F I R k and the corresponding k th L F P patch ( L F P k ). The local difference weighted value calculated in the patch with size l × l can be represented as:
W ( x , y ) ¯ = i = x k x + k j = y k y + k | L F I R k ( i , j ) L F P k ( i , j ) | l × l
where k = l / 2 , L F I R k ( i , j ) and L F P k ( i , j ) denote the pixel values of a specific L F I R patch and corresponding L F P k patch, W ( x , y ) ¯ is the local difference weighted value, and ( x , y ) is the central coordinate of this specific patch. The fusion rule is defined as:
L F F R k ( x , y ) = λ [ a × L F I R k ( x , y ) + b × L F P k ( x , y ) ] + μ
where L F F R k ( x ,   y ) is the k th fused low-frequency patch, and λ is the credibility of the pixel value λ ( 0 , 1 ] . Furthermore, a, b and μ can be defined as:
a = 1 b
b = W ( x , y ) ¯ m a x [ L F I R k ( i , j ) L F P k ( i , j ) , 1 i M , 1 j N ]
  μ = ( 1 λ ) × i = 1 l j = 1 l L F P k ( i , j ) l × l
The fused low-frequency patches L F F not only contain the spectral information in L F I but also preserve the characteristics of L F P , which differs from several pan-sharpening methods based on sparse representation and compressed sensing that have been recognized as excellent [23,26,47]. Subsequently, L F F are obtained by rearranging L F F R into image patches with the same size as L F I and L F P . The procedure of low-frequency sub-images fusion is represented by the flowchart in Figure 1.
The four main steps of low-frequency sub-images fusion can be briefly summarized as: (1) The L F I and L F P sub-images are separately divided into K small and partially overlapped image patches. All the image patches are normalized before the mean of each patch is subtracted. (2) All K patches from L F P are utilized for the dictionary composition. (3) For each patch from L F I , compressed sensing is performed by solving Equation (18) to acquire the reconstructed L F I patch. (4) The L F F patch is fused using a local difference weighted algorithm.

3.3. The Over-All Pan-Sharpening Method

The technical flow of the proposed method is shown in Figure 2. The basic steps of the improved IHS pan-sharpening method based on ripplet transform and compressed sensing are represented as follows:
  • Data pre-processing: The MS and PAN images are co-registered precisely at first. In the meantime, the MS image is up-sampled into the same pixel size and spatial scale as PAN image using cubic convolution. This step ensures the MS and PAN images have the same pixel size and geographic coordinates.
  • IHS transform: The MS image is converted into IHS space to obtain the I component according to Equation (2). The histogram of the I component and of the PAN image are matched.
  • SFIM processing: SFIM is performed on the I component and the PAN image based on Equation (15) to obtain I S F I M . In SFIM, the smoothing filter size should be defined experimentally.
  • DRT decomposition: The new intensity component ( I S F I M ) and the PAN image are decomposed using two-level DRT with c = 1 and d = 4 . The high frequency sub-images ( H F I , H F P ) and low frequency sub-images ( L F I , L F P ) are generated by performing DRT.
  • High-frequency sub-images fusion: H F I and H F P are fused to obtain H F F according to the local variance algorithm in Section 3.2.1.
  • Low-frequency sub-images fusion: Compressed sensing and the local weighted algorithm discussed in Section 3.2.2 is performed on L F I and L F P to obtain L F F .
  • IDRT transform: IDRT is performed on H F F and L F F to obtain the reconstructed intensity I component ( I R ) following Equation (9).
  • Inverse IHS transform: By inverse IHS transform, we transform the reconstructed intensity component ( I R ) and original H component and S component back into multi-band spaces to obtain the pan-sharpened image.

4. Results and Discussion

In this study, the proposed method, together with five state-of-the-art existing methods, is implemented on three datasets (WorldView-2, Pleiades-1A, and TripleSat) each with size 1024 × 1024 for comparison. The specific spatial resolution, spectral response, quantization value, and data source of the three datasets are listed in the Table 1. The datasets cover land objects such as buildings, roads, bridges, farmlands, rivers, shoreline, etc. To guarantee the generality of the experiment, we only fuse three spectral channels in the MS image: red (R), green (G), and blue (B). In fact, the proposed method on the basis of IHS transform has the potential to be extended to fuse four or more spectral bands.
The proposed method is compared with five state-of-the-art pan-sharpening methods, including adaptive IHS (AIHS) [13], wavelet transform and sparse representation (WT-SR) [66], wavelet based IHS (WT-IHS) [19], curvelet transform and independent component analysis (CT-ICA) [67], and ripplet transform based on injected details (CSI) [23]. Moreover, GS sharpening [15] is a well-known method shown to be outstanding on many satellite images for its generality and effectiveness. For fair comparison, the GS method is utilized as a standard benchmark in this study. The corresponding performances are evaluated by six quality indices. All the algorithms and indices used in experiments are implemented in MATLAB 2015b on a macOS 10.13 laptop.
In order to balance the performance and computation time of the proposed method, we empirically define the averaging filter kernel size as 5 × 5 when performing SFIM, and set the image patch size of 7 × 7 with an overlapping ratio of 15% in low-frequency sub-images fusion. In Section 4.3.1, we have a brief discussion on the size of averaging filter kernel selection in SFIM. The effects of the image patch size and overlap ratio on WorldView-2, Pleiades and Triplesat are analyzed in Section 4.3.2. To ensure the fairness of the comparison, the overlapping ratio of the CSI method is set to be 50% as recommended by [27].

4.1. Qualitative Comparison

The source MS and PAN images from the three datasets and corresponding pan-sharpened images produced by seven individual approaches are presented in Figure 3, Figure 4 and Figure 5, respectively. The up-sampled MS images are regarded as ground truth and collaborated with GS sharpened images as benchmarks. For fairer comparisons, the zoomed images for the local area are provided at the bottom corner of each image.
The pan-sharpening results reveal that the GS method possesses ideal universality for the three experimental datasets. The GS method is able to enrich spatial information significantly with a tolerable spectral distortion. Furthermore, in general, the proposed method outperforms all the other methods (including the GS method) on the three datasets. It can be seen in Figure 3, Figure 4 and Figure 5 that the pan-sharpened image of the proposed method has the closest spectral characteristics of MS images and possesses more abundant (or approximately equivalent) spatial information than other pan-sharpened images.
From the perspective of spatial details upgrading, AIHS and WT-IHS methods perform better than WT-SR and CT-ICA. The spatial structures and textures contained in pan-sharpened images output by AIHS and WT-HIS are more visually cognizable. The AIHS method causes more obvious spectral distortion on the three datasets, which is deduced by the highest brightness of pan-sharpened images by AIHS. However, from the perspective of the preservation of spectral information, WT-SR and CT-ICA results reveal no obvious visual spectral diversity degradation. However, the zoomed images suggest these methods (especially WT-SR) perform poorly in improving spatial details. It can also be seen that the CSI method has a remarkable performance on the WorldView-2 and Triplesat datasets.
Since human vision is not sensitive enough to perceive spectral distortion, we calculate the absolute difference values between the fusion results and the original MS image according to Equation (24) as:
D i f f e r e n c e ( x , y ) = 1 B x = 1 M y = 1 N b = 1 B | F ( x , y , b ) R ( x , y , b ) |
where ( x , y ) is the coordinate of a specific pixel, B is the number of bands, M × N is the size of images, and F and R denote the pan-sharpened image and referenced MS image. Figure 6, Figure 7 and Figure 8 are difference images, which denote the absolute values of spectral information difference between each fused image and referenced MS image on every pixel for the three datasets. Blue represents finer differences and red suggests larger differences in spectral information.
Figure 6, Figure 7 and Figure 8 demonstrate that the proposed method is superior to other methods in spectral fidelity improvement. However, when implemented on the Triplesat dataset, the proposed method (Figure 8g) causes a slight spectral distortion. The difference images of the GS method are generally fine for all datasets when compared with other methods. In particular, the difference images reveal that the AIHS and WT-IHS methods lead to more serious spectral degradation, which is in accordance with visual analyses.

4.2. Quantitative Comparison

In quantitative comparison, correlation coefficient (CC) [10], root mean square error (RMSE) [47], the relative global synthesis error (ERGAS) [10], spectral angle mapper (SAM) [10], image quality score index (Q4) [68], and quality with no reference (QNR) [69] are adopted to evaluate the performance of individual pan-sharpening methods. Specifically, the application of CC, RMSE, ERDAS, SAM, Q4, and QNR with D λ can assess the spectral correlation between the original MS image and the pan-sharpened image. The standard deviation (STD) and QNR with D s reflect the richness of spatial details in the MS and pan-sharpened images. The ideal values of CC, Q4 and QNR are 1. The smaller the values for RMSE, ERGAS and SAM, the better. The quantitative results of these different methods on the three datasets are reported in Table 2, Table 3 and Table 4, respectively. For better observation, the optimal value of each quality metric is labeled in bold.
The quantitative comparison results, as represented in Table 2, Table 3 and Table 4, reveal the proposed method to be superior than other methods as it achieves the optimum value for most quality indices. Such a result is in accordance with our previous qualitative comparison. The CSI method achieves comparable values for the WorldView-2 dataset (specifically for STD of the red band and SAM) and the Triplesat dataset (specifically for CC, RMSE, STD of the blue and red bands and D s ) but fails to achieve the same overall (i.e., average) score for the WorldView-2 and Pleiades datasets. The AIHS method has relatively higher STD values on the three datasets, and even achieves the highest value of STD of the red band for the Pleiades dataset. Similar to the WT-IHS method, however, the AIHS method is also not capable of maintaining spectral information, with inferior values for CC, RMSE ERGAS, SAM and Q4. After the proposed method, CSI method performs best on the WorldView-2 and Triplesat datasets.

4.3. Discussion on Independence Factors and Contributions

In this section, we discuss the three independence factors of the proposed method and briefly conclude the contributions. In Section 4.3.1, the influence of smoothing filter kernel size in SFIM is discussed. The dependence of image patch size and overlapping ratio in low-frequency sub-images fusion is investigated in Section 4.3.2. The contributions are provided in Section 4.3.3.

4.3.1. Discussion on Averaging Filer Kernel Size in SFIM

As mentioned in Section 3.1, the definition of averaging filer kernel size in SFIM depends on the scale ratio between PAN and MS images. The scale ratio of WorldView-2, Pleiades and Triplesat is 4 (Table 1), but the kernel size is theoretically odd [51]. Hence, 3 × 3 and 5 × 5 are the two potential optimal kernel sizes of the experiment datasets. Following the guidance in [51], we derive the CC between the pan-sharpened images with the original PAN and MS images for the three datasets. The specific results, as reported in Table 5 (where “PS” represents the pan-sharpened images), reveal 5 × 5 as the optimal kernel size for the study, as the sums of CC between PS with PAN and PS with MS are relatively higher than that of 3 × 3 .
Furthermore, we increase the smoothing kernel size from 7 × 7 towards 13 × 13 to discuss the influence of kernel size in SFIM. It is noteworthy that the CCs of PS and PAN, and PS and MS become gradually identical with the increase in kernel size. Such a trend reveals that the large kernel size will excessively modulate MS images towards PAN images.

4.3.2. Dependence on Image Patch Size and Overlapping Ratio

According to Section 3.2.2, image patch size l × l and the overlapping ratio σ are the only two parameters in LF sub-images fusion. The formation of dictionary D is dependent on image patch size and overlapping ratio. There is a trade-off between the completeness (i.e., number of image patches) and compactness (i.e., size of image patches) of the dictionary [70]. On one hand, a too-large image patch size accelerates computation but weakens the ability of compressed sensing to reconstruct the LF component. On the other hand, a too-small image patch size leads to high coherence of atoms in the dictionary.
We increase image patch size from 5 × 5 to 11 × 11 and the overlapping ratio from 5% to 20%. Experimental results on WorldView-2, Pleiades-1A, and Triplesat data are reported in Table 6, Table 7 and Table 8 (where “Avg” means the average value of the index of all the spectral channels), separately.
According to this statistical evidence, we can draw three conclusions:
  • The increase of the overlapping ratio σ is beneficial to performance but leads to large increases in computation time.
  • For WorldView-2, Pleiades-1A, and Triplesat data, 7 × 7 is the optimal image patch size.
  • In our experiments, σ = 15 % is the most desirable value to balance performance and computation time.

4.3.3. Contributions

In general, comprehensive evaluations, both visually and by objective indices, reveal the proposed method as optimal for datasets from WorldView-2, Pleiades, and Triplesat, when compared with AIHS [13], WT-SR [66], WT-IHS [19], CT-ICA [67], CSI [23] and GS [15]. Notably, the CSI method also achieved comparable results with the exception of the Pleiades dataset. The results indicate that the proposed method achieves better spectral fidelity while yielding greater (or equivalent) spatial information improvement than the other state-of-the-art methods. Furthermore, the results demonstrate that the integration of DRT and compressed sensing can ensure the balance of spatial detail reconstruction and spectral diversity preservation.

5. Conclusions

In this paper, an improved IHS-based pan-sharpening method using ripplet transform and compressed sensing is proposed, which overcomes the spectral distortion that occurs in the original IHS method. Firstly, SFIM is performed on the PAN image and the I component separated from MS image to improve spectral fidelity. The proposed method adopts different fusion rules according to individual characteristics of the HF and LF coefficients obtained by DRT. The fusion rule based on local variance of wavelet energy is utilized for HF sub-images. For LF information, we use a compressed sensing technique to maintain the local structures and spectral information. The LF coefficients of the I component and the PAN image are divided into small patches and a dictionary is constructed using patches of low-frequency components of PAN. The reconstructed low-frequency components of the I component and the PAN image are generated using the BP algorithm, and the local difference weighted algorithm is used to generate new LF components. For the combination of IHS transform, DRT and compressed sensing, the proposed algorithm can produce pan-sharpened images with both high spatial and high spectral resolution efficiently and effectively.
The performance of the proposed method and five state-of-the-art methods is comprehensively evaluated using WorldView-2, Pleiades-1A and Triplesat datasets. The specific evaluation indices include CC, RMSE, STD, ERGAS, SAM, Q4 and QNR. The experimental results demonstrate that the proposed method performs better than other fusion methods and overcomes spectral distortion while improving spatial resolution significantly. In future work, more effort should be dedicated to improving the efficiency of the proposed method, as the BP algorithm is time-consuming and can thus be potentially replaced by other reconstruction algorithms. In addition, the practicability of the algorithm images captured by other remote sensors has yet to be demonstrated.

Author Contributions

C.Y. and R.M. conceived and designed the experiments. C.Y. proposed the methodology and validation, implemented the source code, and composed the manuscript. Q.Z. improved the framework of this study and reviewed the paper. H.L. provided detailed advice in the reviewing and editing of this paper.

Funding

This study is supported by the National Natural Science Foundation of China (No.41331175 and 51378399).

Acknowledgments

The authors wish to acknowledge the anonymous reviewers for providing helpful suggestions that greatly improved the manuscript. In the meantime, the authors would like to thank Dapeng Wu (Dept. of Electrical and Computer Engineering, University of Florida, USA) for helping us in the implementation of Ripplet Transform.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ghassemian, H. A review of remote sensing image fusion methods. Inf. Fusion 2016, 32, 75–89. [Google Scholar] [CrossRef]
  2. Otazu, X.; Gonzalez-Audicana, M.; Fors, O.; Nunez, J. Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef] [Green Version]
  3. Pohl, C.; Van Genderen, J.L. Review article Multisensor image fusion in remote sensing: Concepts, methods and applications. Int. J. Remote Sens. 1998, 19, 823–854. [Google Scholar] [CrossRef] [Green Version]
  4. Simone, G.; Farina, A.; Morabito, F.C.; Serpico, S.B.; Bruzzone, L. Image fusion techniques for remote sensing applications. Inf. Fusion 2002, 3, 3–15. [Google Scholar] [CrossRef] [Green Version]
  5. Melgani, F.; Bazi, Y. Markovian fusion approach to robust unsupervised change detection in remotely sensed imagery. IEEE Geosci. Remote Sens. Lett. 2006, 3, 457–461. [Google Scholar] [CrossRef]
  6. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A. Remote Sensing Image Fusion; CRC Press: Boca Raton, FL, USA, 2015. [Google Scholar]
  7. Kavzoglu, T.; Tonbul, H. An experimental comparison of multi-resolution segmentation, SLIC and K-means clustering for object-based classification of VHR imagery. Int. J. Remote Sens. 2018, 1–17. [Google Scholar] [CrossRef]
  8. Ibarrola-Ulzurrun, E.; Gonzalo-Martin, C.; Marcello-Ruiz, J.; Garcia-Pedrero, A.; Rodriguez-Esparragon, D. Fusion of high resolution multispectral imagery in vulnerable coastal and land ecosystems. Sensors 2017, 17, 228. [Google Scholar] [CrossRef] [PubMed]
  9. Li, S.; Kang, X.; Fang, L.; Hu, J.; Yin, H. Pixel-level image fusion: A survey of the state of the art. Inf. Fusion 2017, 33, 100–112. [Google Scholar] [CrossRef]
  10. Garzelli, A.; Nencini, F.; Alparone, L.; Aiazzi, B.; Baronti, S. Pan-sharpening of multispectral images: A critical review and comparison. In Proceedings of the IGARSS 2004, IEEE International Geoscience and Remote Sensing Symposium, Anchorage, AK, USA, 20–24 September 2004. [Google Scholar]
  11. Thomas, C.; Ranchin, T.; Wald, L.; Chanussot, J. Synthesis of multispectral images to high spatial resolution: A critical review of fusion methods based on remote sensing physics. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1301–1312. [Google Scholar] [CrossRef] [Green Version]
  12. Tu, T.-M.; Su, S.-C.; Shyu, H.-C.; Huang, P.S. A new look at IHS-like image fusion methods. Inf. Fusion 2001, 2, 177–186. [Google Scholar] [CrossRef]
  13. Rahmani, S.; Strait, M.; Merkurjev, D.; Moeller, M.; Wittman, T. An adaptive IHS pan-sharpening method. IEEE Geosci. Remote Sens. Lett. 2010, 7, 746–750. [Google Scholar] [CrossRef]
  14. Guo, Q.; Ehlers, M.; Wang, Q.; Pohl, C.; Hornberg, S.; Li, A. Ehlers pan-sharpening performance enhancement using HCS transform for n-band data sets. Int. J. Remote Sens. 2017, 38, 4974–5002. [Google Scholar] [CrossRef]
  15. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent US6011875A, 19 April 1998. [Google Scholar]
  16. Gonzalez-Audicana, M.; Saleta, J.L.; Catalan, R.G.; Garcia, R. Fusion of multispectral and panchromatic images using improved IHS and PCA mergers based on wavelet decomposition. IEEE Trans. Geosci. Remote Sens. 2004, 42, 1291–1299. [Google Scholar] [CrossRef] [Green Version]
  17. Faragallah, O.S. Enhancing multispectral imagery spatial resolution using optimized adaptive PCA and high-pass modulation. Int. J. Remote Sens. 2018, 1–15. [Google Scholar] [CrossRef]
  18. Nunez, J.; Otazu, X.; Fors, O.; Prades, A.; Pala, V.; Arbiol, R. Multiresolution-based image fusion with additive wavelet decomposition. IEEE Trans. Geosci. Remote Sens. 1999, 37, 1204–1211. [Google Scholar] [CrossRef] [Green Version]
  19. Pradhan, P.S.; King, R.L.; Younan, N.H.; Holcomb, D.W. Estimation of the number of decomposition levels for a wavelet-based multiresolution multisensor image fusion. IEEE Trans. Geosci. Remote Sens. 2006, 44, 3674–3686. [Google Scholar] [CrossRef]
  20. Shah, V.P.; Younan, N.H.; King, R.L. An efficient pan-sharpening method via a combined adaptive pca approach and contourlets. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1323–1335. [Google Scholar] [CrossRef]
  21. Nencini, F.; Garzelli, A.; Baronti, S.; Alparone, L. Remote sensing image fusion using the curvelet transform. Inf. Fusion 2007, 8, 143–156. [Google Scholar] [CrossRef]
  22. Li, H.; Jing, L.; Tang, Y. Assessment of pansharpening methods applied to worldview-2 imagery fusion. Sensors 2017, 17, 89. [Google Scholar] [CrossRef] [PubMed]
  23. Ghahremani, M.; Ghassemian, H. Remote sensing image fusion using ripplet transform and compressed sensing. IEEE Geosci. Remote Sens. Lett. 2015, 12, 502–506. [Google Scholar] [CrossRef]
  24. Aiazzi, B.; Alparone, L.; Baronti, S.; Carlà, R.; Garzelli, A.; Santurri, L. Sensitivity of pansharpening methods to temporal and instrumental changes between multispectral and panchromatic data sets. IEEE Trans. Geosci. Remote Sens. 2017, 55, 308–319. [Google Scholar] [CrossRef]
  25. Meng, X.; Li, J.; Shen, H.; Zhang, L.; Zhang, H. Pansharpening with a guided filter based on three-layer decomposition. Sensors 2016, 16, 1068. [Google Scholar] [CrossRef] [PubMed]
  26. Li, S.; Yang, B. A new pan-sharpening method using a compressed sensing technique. IEEE Trans. Geosci. Remote Sens. 2011, 49, 738–746. [Google Scholar] [CrossRef]
  27. Ghahremani, M.; Ghassemian, H. A compressed-sensing-based pan-sharpening method for spectral distortion reduction. IEEE Trans. Geosci. Remote Sens. 2016, 54, 2194–2206. [Google Scholar] [CrossRef]
  28. Huang, B.; Song, H.; Cui, H.; Peng, J.; Xu, Z. Spatial and spectral image fusion using sparse matrix factorization. IEEE Trans. Geosci. Remote Sens. 2014, 52, 1693–1704. [Google Scholar] [CrossRef]
  29. Zhou, X.; Liu, J.; Liu, S.; Cao, L.; Zhou, Q.; Huang, H. A GIHS-based spectral preservation fusion method for remote sensing images using edge restored spectral modulation. ISPRS-J. Photogramm. Remote Sens. 2014, 88, 16–27. [Google Scholar] [CrossRef]
  30. Imani, M.; Ghassemian, H. Pansharpening optimisation using multiresolution analysis and sparse representation. Int. J. Image Data Fusion. 2017, 8, 270–292. [Google Scholar] [CrossRef]
  31. Li, S.; Yang, B.; Hu, J. Performance comparison of different multi-resolution transforms for image fusion. Inf. Fusion 2011, 12, 74–84. [Google Scholar] [CrossRef]
  32. Rehman, U.N.; Ehsan, S.; Abdullah, M.S.; Akhtar, J.M.; Mandic, P.D.; McDonald-Maier, D.K. Multi-scale pixel-based image fusion using multivariate empirical mode decomposition. Sensors 2015, 15, 10923–10947. [Google Scholar] [CrossRef] [PubMed]
  33. Eghbalian, S.; Ghassemian, H. Multi spectral image fusion by deep convolutional neural network and new spectral loss function. Int. J. Remote Sens. 2018, 39, 3983–4002. [Google Scholar] [CrossRef]
  34. Li, H.; Jing, L.; Tang, Y.; Ding, H. An improved pansharpening method for misaligned panchromatic and multispectral data. Sensors. 2018, 18, 557. [Google Scholar] [CrossRef] [PubMed]
  35. Zhang, L.; Zhang, J. A novel remote-sensing image fusion method based on hybrid visual saliency analysis. Int. J. Remote Sens. 2018, 1–23. [Google Scholar] [CrossRef]
  36. Matsuoka, M.; Tadono, T.; Yoshioka, H. Effects of the spectral properties of a panchromatic image on pan-sharpening simulated using hyperspectral data. Int. J. Image Data Fusion. 2016, 7, 339–359. [Google Scholar] [CrossRef] [Green Version]
  37. Li, Z.; Leung, H. Fusion of multispectral and panchromatic images using a restoration-based method. IEEE Trans. Geosci. Remote Sens. 2009, 47, 1482–1491. [Google Scholar] [CrossRef]
  38. Zhang, Q.; Liu, Y.; Blum, R.S.; Han, J.; Tao, D. Sparse representation based multi-sensor image fusion for multi-focus and multi-modality images: A review. Inf. Fusion 2018, 40, 57–75. [Google Scholar] [CrossRef]
  39. Olshausen, B.A.; Field, D.J. Emergence of simple-cell receptive field properties by learning a sparse code for natural images. Nature 1996, 381, 607. [Google Scholar] [CrossRef] [PubMed]
  40. Candes, E.; Romberg, J.; Tao, T. Stable Signal Recovery from Incomplete and Inaccurate Measurements. Mathematics. Commun. Pure Appl. Math. 2005, 1207–1223. [Google Scholar] [CrossRef]
  41. Lustig, M.; Donoho, D.L.; Santos, J.M.; Pauly, J.M. Compressed Sensing MRI. IEEE Signal Process. Mag. 2008, 25, 72–82. [Google Scholar] [CrossRef]
  42. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  43. Yang, B.; Li, S. Multifocus image fusion and restoration with sparse representation. IEEE Trans. Instrum. Meas. 2010, 59, 884–892. [Google Scholar] [CrossRef]
  44. Wang, J.; Peng, J.; Jiang, X.; Feng, X.; Zhou, J. Remote-sensing image fusion using sparse representation with sub-dictionaries. Int. J. Remote Sens. 2017, 38, 3564–3585. [Google Scholar] [CrossRef]
  45. Zhang, Y.; Hong, G. An IHS and wavelet integrated approach to improve pan-sharpening visual quality of natural colour IKONOS and QuickBird images. Inf. Fusion 2005, 6, 225–234. [Google Scholar] [CrossRef]
  46. Huang, Q.; Zhao, Z.; Gao, Q.; Meng, Y.; Ma, J.; Chen, J. improved fusion method for remote sensing imagery using NSCT and extend IHS. IFAC-PapersOnLine. 2015, 48, 64–69. [Google Scholar] [CrossRef]
  47. Cheng, J.; Liu, H.; Liu, T.; Wang, F.; Li, H. Remote sensing image fusion via wavelet transform and sparse representation. ISPRS-J. Photogramm. Remote Sens. 2015, 104, 158–173. [Google Scholar] [CrossRef]
  48. Xu, J.; Wu, D. Ripplet transform type II transform for feature extraction. IET Image Process. 2012, 6, 374–385. [Google Scholar] [CrossRef]
  49. Xu, J.; Yang, L.; Wu, D. Ripplet: A new transform for image processing. J. Vis. Commun. Image Represent. 2010, 21, 627–639. [Google Scholar] [CrossRef] [Green Version]
  50. Elad, M.; Aharon, M. Image denoising via learned dictionaries and sparse representation. In Proceedings of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR’06), New York, NY, USA, 17–22 June 2006; pp. 895–900. [Google Scholar]
  51. Liu, J.G. Smoothing filter-based intensity modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  52. Garzelli, A.; Nencini, F. Fusion of panchromatic and multispectral images by genetic algorithms. In Proceedings of the 2006 IEEE International Symposium on Geoscience and Remote Sensing, Denver, CO, USA, 31 July–4 August 2006; pp. 3810–3813. [Google Scholar]
  53. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. Context-driven fusion of high spatial and spectral resolution images based on oversampled multiresolution analysis. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2300–2312. [Google Scholar] [CrossRef]
  54. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. MTF-tailored multiscale fusion of high-resolution MS and pan imagery. Photogramm. Eng. Remote Sens. 2015, 72, 591–596. [Google Scholar] [CrossRef]
  55. Starck, J.-L.; Candès, E.J.; Donoho, D.L. The curvelet transform for image denoising. IEEE Trans. Image Process. 2002, 11, 670–684. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  56. Mairal, J.; Elad, M.; Sapiro, G. Sparse representation for color image restoration. IEEE Trans. Image Process. 2008, 17, 53–69. [Google Scholar] [CrossRef] [PubMed]
  57. Liu, Y.; Liu, S.; Wang, Z. A general framework for image fusion based on multi-scale transform and sparse representation. Inf. Fusion 2015, 24, 147–164. [Google Scholar] [CrossRef]
  58. Jiang, Y.; Wang, M. Image fusion with morphological component analysis. Inf. Fusion 2014, 18, 107–118. [Google Scholar] [CrossRef]
  59. Bruckstein, A.M.; Donoho, D.L.; Elad, M. From Sparse Solutions of Systems of Equations to Sparse Modeling of Signals and Images. SIAM Rev. 2009, 51, 34–81. [Google Scholar] [CrossRef] [Green Version]
  60. Woodcock, C.E.; Strahler, A.H. The factor of scale in remote sensing. Remote Sens. Environ. 1987, 21, 311–332. [Google Scholar] [CrossRef]
  61. Chu, H.; Zhu, W. Fusion of IKONOS satellite imagery using IHS transform and local variation. IEEE Geosci. Remote Sens. Lett. 2008, 5, 653–657. [Google Scholar] [CrossRef]
  62. Dong, Z.; Wang, Z.; Liu, D.; Zhang, B.; Zhao, P.; Tang, X.; Jia, M. SPOT5 multi-spectral (MS) and panchromatic (PAN) image fusion using an improved wavelet method based on local algorithm. Comput. Geosci. 2013, 60, 134–141. [Google Scholar] [CrossRef]
  63. Rubinstein, R.; Zibulevsky, M.; Elad, M. Double sparsity: Learning sparse dictionaries for sparse signal approximation. IEEE Trans. Signal Process. 2010, 58, 1553–1564. [Google Scholar] [CrossRef]
  64. Vicinanza, M.R.; Restaino, R.; Vivone, G.; Mura, M.D.; Chanussot, J. A pansharpening method based on the sparse representation of injected details. IEEE Geosci. Remote Sens. Lett. 2015, 12, 180–184. [Google Scholar] [CrossRef]
  65. Cost, S.; Salzberg, S. A weighted nearest neighbor algorithm for learning with symbolic features. Mach. Learn. 1993, 10, 57–78. [Google Scholar] [CrossRef] [Green Version]
  66. Liu, Y.; Wang, Z. A practical pan-sharpening method with wavelet transform and sparse representation. In Proceedings of the 2013 IEEE International Conference on Imaging Systems and Techniques (IST), Beijing, China, 22–23 October 2013; pp. 288–293. [Google Scholar]
  67. Ghahremani, M.; Ghassemian, H. Remote-sensing image fusion based on curvelets and ICA. Int. J. Remote Sens. 2015, 36, 4131–4143. [Google Scholar] [CrossRef]
  68. Alparone, L.; Baronti, S.; Garzelli, A.; Nencini, F. A global quality measurement of pan-sharpened multispectral imagery. IEEE Geosci. Remote Sens. Lett. 2004, 1, 313–317. [Google Scholar] [CrossRef]
  69. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef]
  70. Zhu, X.X.; Bamler, R. A sparse image fusion algorithm with application to pan-sharpening. IEEE Trans. Geosci. Remote Sens. 2013, 51, 2827–2836. [Google Scholar] [CrossRef]
Figure 1. The procedure of low-frequency sub-images fusion.
Figure 1. The procedure of low-frequency sub-images fusion.
Sensors 18 03624 g001
Figure 2. The technical flow of the proposed pan-sharpening method.
Figure 2. The technical flow of the proposed pan-sharpening method.
Sensors 18 03624 g002
Figure 3. WorldView-2 original images and fusion results of different methods. (a) Original multispectral image (MS); (b) Original panchromatic image (PAN); (c) Gram-Schmidt [15] pan-sharpened image (GS); (d) Adaptive IHS [13] pan-sharpened image (AIHS); (e) Wavelet transform and sparse representation method [66] pan-sharpened image (WT-SR); (f) Wavelet-based IHS [19] pan-sharpened image (WT-IHS); (g) Curvelet transform and independent component analysis [67] pan-sharpened image (CT-ICA); (h) Ripplet transform based on injected details [23] pan-sharpened image (CSI); (i) pan-sharpened image by the proposed method (proposed).
Figure 3. WorldView-2 original images and fusion results of different methods. (a) Original multispectral image (MS); (b) Original panchromatic image (PAN); (c) Gram-Schmidt [15] pan-sharpened image (GS); (d) Adaptive IHS [13] pan-sharpened image (AIHS); (e) Wavelet transform and sparse representation method [66] pan-sharpened image (WT-SR); (f) Wavelet-based IHS [19] pan-sharpened image (WT-IHS); (g) Curvelet transform and independent component analysis [67] pan-sharpened image (CT-ICA); (h) Ripplet transform based on injected details [23] pan-sharpened image (CSI); (i) pan-sharpened image by the proposed method (proposed).
Sensors 18 03624 g003
Figure 4. Pleiades original images and fusion results of different methods. (a) MS; (b) PAN; (c) GS; (d) AIHS; (e) WT-SR; (f) WT-IHS; (g) CT-ICA; (h) CSI; (i) proposed.
Figure 4. Pleiades original images and fusion results of different methods. (a) MS; (b) PAN; (c) GS; (d) AIHS; (e) WT-SR; (f) WT-IHS; (g) CT-ICA; (h) CSI; (i) proposed.
Sensors 18 03624 g004
Figure 5. Triplesat original images and fusion results of different methods. (a) MS; (b) PAN; (c) GS; (d) AIHS; (e) WT-SR; (f) WT-IHS; (g) CT-ICA; (h) CSI; (i) proposed.
Figure 5. Triplesat original images and fusion results of different methods. (a) MS; (b) PAN; (c) GS; (d) AIHS; (e) WT-SR; (f) WT-IHS; (g) CT-ICA; (h) CSI; (i) proposed.
Sensors 18 03624 g005
Figure 6. The difference image between the fused images and the referenced MS image for WorldView-2 data. (a) GS; (b) AIHS; (c) WT-SR; (d) WT-IHS; (e) CT-ICA; (f) CSI; (g) proposed.
Figure 6. The difference image between the fused images and the referenced MS image for WorldView-2 data. (a) GS; (b) AIHS; (c) WT-SR; (d) WT-IHS; (e) CT-ICA; (f) CSI; (g) proposed.
Sensors 18 03624 g006
Figure 7. The difference image between the fused images and the referenced MS image for Pleiades data. (a) GS; (b) AIHS; (c) WT-SR; (d) WT-IHS; (e) CT-ICA; (f) CSI; (g) proposed.
Figure 7. The difference image between the fused images and the referenced MS image for Pleiades data. (a) GS; (b) AIHS; (c) WT-SR; (d) WT-IHS; (e) CT-ICA; (f) CSI; (g) proposed.
Sensors 18 03624 g007
Figure 8. The difference image between the fused images and the referenced MS image for Triplesat data. (a) GS; (b) AIHS; (c) WT-SR; (d) WT-IHS; (e) CT-ICA; (f) CSI; (g) proposed.
Figure 8. The difference image between the fused images and the referenced MS image for Triplesat data. (a) GS; (b) AIHS; (c) WT-SR; (d) WT-IHS; (e) CT-ICA; (f) CSI; (g) proposed.
Sensors 18 03624 g008
Table 1. Characteristics of three datasets used in this study.
Table 1. Characteristics of three datasets used in this study.
FeaturesWorldView-2Pleiades-1ATriplesat
PANMSPANMSPANMS
Spatial Resolution0.46 m1.84 m0.5 m2 m0.8 m3.2 m
Spectral Response450–800 nmCostal (C):
400–450 nm
Red (R):
630–690 nm
470–830 nmBlue (B):
430–550 nm
450–650 nmBlue (B):
440–510 nm
Blue (B):
450–510 nm
Red Edge (RE):
705–745 nm
Green (G):
500–620 nm
Green (G):
510–590 nm
Green (G):
510–580 nm
Near IR-1 (NIR-1):
770–895 nm
Red (R):
590–710 nm
Red (R):
600–670 nm
Yellow (Y):
585–625 nm
Near IR-2 (NIR-2):
860–900 nm
Near IR (NIR):
740–940 nm
Near IR (NIR):
760–910 nm
Quantization Value11 Bits12 Bits or 16 Bits10 Bits
Imaging LocationHuang Long Stadium in Hangzhou, ChinaXuanen county government in Enshi, ChinaQingpu district in Shanghai, China
Imaging Time20 December 200913 April 201518 September 2017
Table 2. Evaluation indices of fusion results for WorldView-2 data.
Table 2. Evaluation indices of fusion results for WorldView-2 data.
IndexGSMethod
AIHSWT-SRWT-IHSCT-ICACSIProposed
Correlation coefficient (CC)R0.9450.8960.9120.9100.9380.9410.952
G0.9360.8790.9340.9150.9240.9530.957
B0.9400.9250.9270.9120.9470.9480.949
Average0.9410.9000.9240.9120.9360.9470.953
Root mean square error (RMSE)R15.0216.9615.6716.2414.6815.1714.91
G14.8616.8515.9416.3614.9613.8213.64
B14.0717.2316.0715.9115.2413.9413.65
Average14.6517.0515.8916.1714.9614.3114.07
Standard deviation (STD)R63.6159.6558.3463.2862.3466.2465.27
G65.0961.2357.9767.9557.6864.1665.33
B62.8257.2958.2661.2758.2963.9065.28
Average63.8459.3958.1964.1759.4364.7765.29
The relative global synthesis error (ERGAS)2.8733.6813.4873.3353.0152.9742.762
Spectral angle mapper (SAM)5.0967.6208.6737.2416.2923.2453.271
Image quality score index (Q4)0.8310.5890.5340.5910.6280.8960.913
Quality with no reference (QNR) D s 0.1290.1450.1320.1520.1370.1250.121
D λ 0.1430.1860.1740.1630.1510.1340.132
Average0.8080.6690.7810.7480.7950.8180.854
Table 3. Evaluation indices of fusion results for Pleiades data.
Table 3. Evaluation indices of fusion results for Pleiades data.
IndexGSMethod
AIHSWT-SRWT-IHSCT-ICACSIproposed
CCR0.8330.6750.8060.7130.8670.8210.875
G0.8020.6230.8140.7560.8590.8340.861
B0.7990.6940.7950.6970.8740.8060.868
Average0.8110.6640.8050.7220.8670.8200.868
RMSER10.1413.6810.2912.409.79910.289.014
G9.99613.5510.3712.529.3689.8679.257
B10.1113.2711.1512.709.25710.029.119
Average9.77313.5010.6012.549.47510.059.130
STDR53.0256.5647.9251.3752.5155.6955.24
G53.9752.4646.5852.1753.1154.7255.67
B54.1253.6547.1452.6452.8955.0255.94
Average53.7054.1647.2252.0052.8455.1555.62
ERGAS1.7092.6892.0542.3371.6801.8061.547
SAM1.6432.0811.8541.9221.7251.7341.549
Q40.8780.6570.7620.7150.8650.8900.894
QNR D s 0.1190.2640.1810.2540.2130.1570.119
D λ 0.1020.1180.0970.1540.1050.0940.086
Average0.8100.6750.8340.7600.7980.8250.874
Table 4. Evaluation indices of fusion results for Triplesat data.
Table 4. Evaluation indices of fusion results for Triplesat data.
IndexGSMethod
AIHSWT-SRWT-IHSCT-ICACSIproposed
CCR0.9050.7570.8520.8840.9390.9610.975
G0.9170.7780.8670.9040.9280.9630.982
B0.8930.8020.8400.8980.9220.9750.966
Average0.9050.7790.8530.8950.9300.9660.974
RMSER19.8330.8125.4226.7322.4012.4210.08
G16.7729.9624.3628.3922.3312.189.422
B13.9230.4625.0527.5720.5010.7511.60
Average16.7430.4124.9427.5621.7411.7810.37
STDR59.8060.0647.5657.0151.6662.1763.52
G61.0760.6148.7357.7951.8662.6262.07
B60.5859.5947.3456.5050.1462.3762.17
Average60.4860.0947.8857.1051.1362.3962.59
ERGAS1.4023.9251.8752.6131.5341.3941.360
SAM1.9965.3413.5224.8982.8472.0021.987
Q40.8100.6750.8060.7610.8610.9250.979
QNR D s 0.1770.3290.1680.2140.1750.1370.149
D λ 0.1380.2050.1770.1980.1540.1160.103
Average0.8100.6680.7150.6940.7670.8320.869
Table 5. Impact of smoothing filer kernel size in SFIM on three datasets.
Table 5. Impact of smoothing filer kernel size in SFIM on three datasets.
DatasetCCKernel Size
3 × 35 × 57 × 79 × 911 × 1113 × 13
WorldView-2PS and PAN0.7380.7650.7760.7970.8310.841
PS and MS0.9690.9580.9350.9020.8810.867
Sum1.7071.7231.7211.6991.7121.708
PleiadesPS and PAN0.7150.7310.7420.7570.7720.791
PS and MS0.8790.8680.8470.8330.8250.802
Sum1.5941.5991.5891.5901.5771.593
TriplesatPS and PAN0.7320.7880.8050.8350.8420.846
PS and MS0.9810.9740.9550.9210.9130.902
Sum1.7131.7621.7601.7561.7551.746
Table 6. Impact of image patch size and overlapping ratio on WorldView-2 image.
Table 6. Impact of image patch size and overlapping ratio on WorldView-2 image.
Image Patch SizeOverlapping RatioCC (Avg)RMSE (Avg)STD (Avg)ERGASSAMQ4QNR (Avg)Time (s)
5 × 55%0.88719.0255.213.3304.3150.8040.75140.91
5 × 510%0.90618.2757.933.2054.2980.8270.76656.43
5 × 515%0.92117.4959.663.1194.1020.8400.79482.72
5 × 520%0.93417.3360.573.0123.9260.8560.801107.5
7 × 75%0.94916.2163.012.8873.6140.8910.83561.25
7 × 710%0.95114.9564.672.8243.3590.9070.84785.60
7 × 715%0.95314.0765.292.7623.2710.9130.854119.7
7 × 720%0.95413.6565.332.6373.1530.9180.862162.1
9 × 95%0.92414.7363.292.8353.8250.8820.82672.71
9 × 910%0.93014.6563.372.8293.8140.8970.833105.8
9 × 915%0.93514.5863.452.8203.8020.9030.839156.1
9 × 920%0.93714.4663.582.8133.7960.9090.841182.9
11 × 115%0.91815.3262.092.9643.9670.8730.81186.50
11 × 1110%0.92415.2462.172.9473.9560.8790.817117.1
11 × 1115%0.93115.1362.222.9393.9430.8850.824165.6
11 × 1120%0.93515.0962.362.9323.9370.8910.829196.4
Table 7. Impact of image patch size and overlapping ratio on Pleiades-1A image.
Table 7. Impact of image patch size and overlapping ratio on Pleiades-1A image.
Image Patch SizeOverlapping RatioCC (Avg)RMSE (Avg)STD (Avg)ERGASSAMQ4QNR (Avg)Time (s)
5 × 55%0.78810.26744.3282.3392.3570.7860.78124.26
5 × 510%0.7959.90146.5832.3282.3400.8010.79359.67
5 × 515%0.8139.82548.7722.2242.2140.8150.80580.41
5 × 520%0.8219.65849.9352.1072.0980.8340.82995.93
7 × 75%0.8379.24653.8391.7151.7720.8630.84749.47
7 × 710%0.8549.22454.3481.6061.6180.8790.86180.65
7 × 715%0.8689.13055.6171.5471.5490.8940.874124.3
7 × 720%0.8829.06756.8111.4351.4210.9120.886165.7
9 × 95%0.8119.42551.2671.8091.7950.8510.83550.66
9 × 910%0.8209.31951.9831.7751.6820.8600.84261.42
9 × 915%0.8299.22152.6171.6611.6070.8730.85698.71
9 × 920%0.8369.17853.2091.5871.5340.8890.868174.9
11 × 115%0.8059.72950.6641.9241.8950.8470.82272.83
11 × 1110%0.8149.63351.7091.8731.8620.8510.829103.6
11 × 1115%0.8279.52153.1051.7051.7340.8640.838138.7
11 × 1120%0.8339.41354.0181.5821.6070.8770.845190.1
Table 8. Impact of image patch size and overlapping ratio on Triplesat image.
Table 8. Impact of image patch size and overlapping ratio on Triplesat image.
Image Patch SizeOverlapping RatioCC (Avg)RMSE (Avg)STD (Avg)ERGASSAMQ4QNR (Avg)Time (s)
5 × 55%0.88311.30451.1782.0163.3570.9170.79835.34
5 × 510%0.89811.29552.5911.9123.1020.9210.80362.40
5 × 515%0.90711.18753.8341.8872.9450.9250.811105.4
5 × 520%0.92411.06254.3671.6542.7980.9340.826162.7
7 × 75%0.95110.56360.4981.4562.2230.9550.84784.64
7 × 710%0.96210.44562.0671.3982.1480.9680.853125.3
7 × 715%0.97410.36862.5861.3601.9870.9790.869141.5
7 × 720%0.98910.01763.0541.3491.9050.9820.874188.4
9 × 95%0.94110.66455.3921.4892.4820.9400.84194.51
9 × 910%0.95210.53256.0181.4472.4050.9540.849110.9
9 × 915%0.96610.46756.9811.3942.3490.9630.855145.2
9 × 920%0.97510.39557.7651.2512.2970.9770.863191.6
11 × 115%0.93310.89456.2151.6642.6640.9330.83286.32
11 × 1110%0.94110.75257.5411.5982.6150.9410.839119.7
11 × 1115%0.95710.66358.1651.4352.5330.9560.847158.5
11 × 1120%0.969-10.59759.0851.3572.4980.9680.851200.2

Share and Cite

MDPI and ACS Style

Yang, C.; Zhan, Q.; Liu, H.; Ma, R. An IHS-Based Pan-Sharpening Method for Spectral Fidelity Improvement Using Ripplet Transform and Compressed Sensing. Sensors 2018, 18, 3624. https://doi.org/10.3390/s18113624

AMA Style

Yang C, Zhan Q, Liu H, Ma R. An IHS-Based Pan-Sharpening Method for Spectral Fidelity Improvement Using Ripplet Transform and Compressed Sensing. Sensors. 2018; 18(11):3624. https://doi.org/10.3390/s18113624

Chicago/Turabian Style

Yang, Chen, Qingming Zhan, Huimin Liu, and Ruiqi Ma. 2018. "An IHS-Based Pan-Sharpening Method for Spectral Fidelity Improvement Using Ripplet Transform and Compressed Sensing" Sensors 18, no. 11: 3624. https://doi.org/10.3390/s18113624

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop