Next Article in Journal
ARIES: A Novel Multivariate Intrusion Detection System for Smart Grid
Previous Article in Journal
A Low Power Sigma-Delta Modulator with Hybrid Architecture
Previous Article in Special Issue
Pyramid Inter-Attention for High Dynamic Range Imaging
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Variational Bayesian Pansharpening with Super-Gaussian Sparse Image Priors

1
Departamento de Ciencias de la Computación e Inteligencia Artificial, Universidad de Granada, 18071 Granada, Spain
2
Departamento de Lenguajes y Sistemas Informáticos, Universidad de Granada, 18071 Granada, Spain
3
Department of Electrical Engineering and Computer Science, Northwestern University, Evanston, IL 60208-3118, USA
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(18), 5308; https://doi.org/10.3390/s20185308
Submission received: 21 July 2020 / Revised: 8 September 2020 / Accepted: 11 September 2020 / Published: 16 September 2020
(This article belongs to the Special Issue Digital Imaging with Multispectral Filter Array (MSFA) Sensors)

Abstract

:
Pansharpening is a technique that fuses a low spatial resolution multispectral image and a high spatial resolution panchromatic one to obtain a multispectral image with the spatial resolution of the latter while preserving the spectral information of the multispectral image. In this paper we propose a variational Bayesian methodology for pansharpening. The proposed methodology uses the sensor characteristics to model the observation process and Super-Gaussian sparse image priors on the expected characteristics of the pansharpened image. The pansharpened image, as well as all model and variational parameters, are estimated within the proposed methodology. Using real and synthetic data, the quality of the pansharpened images is assessed both visually and quantitatively and compared with other pansharpening methods. Theoretical and experimental results demonstrate the effectiveness, efficiency, and flexibility of the proposed formulation.

1. Introduction

Remote sensing sensors simultaneously capture a Multispectral (MS) low resolution image along with a single-band high resolution image of the same area, referred to as Panchromatic (PAN) image. However, MS high-resolution images are needed by many applications, such as land use and land cover analyses or change detection. Pansharpening is a technique that fuses the MS and PAN images into an MS high resolution image that has the spatial resolution of the PAN image and the spectral resolution of the MS one.
In this paper we formulate the pansharpening problem following the Bayesian framework. Within this framework, we use the sensor characteristics to model the observation process as a conditional probability distribution. The observation process describes both the MS high resolution image to MS low resolution image relationship and how the PAN image is obtained from the MS high resolution one. This probability distributions provides fidelity to the observed data in the pansharpened image reconstruction process. together with from fidelity to the data, Bayesian methods incorporate prior knowledge on the MS high resolution image in the form of prior probability distributions. Crisp images, such as high resolution MS images, are expected to have Super-Gaussian (SG) statistics, while upsampled images suffer from blur that smooths out sharp gradients, making them more Gaussian in their statistics [1]. Our goal is to integrate the sharp edges of the PAN image into the pansharpened image, leading to less Gaussian statistics which makes SG priors a suitable choice. SG priors have been successfully applied to other image processing tasks, such as compressed sensing [2], blind deconvolution [1,3] and blind color deconvolution [4] and so it is also expected to produce good results in pansharpening. However, the form of the SG prior does not allow us to obtain the posterior distribution in an analytical way, making full Bayesian inference intractable. Hence, in this paper, we use the variational Bayesian inference to estimate the distribution of the pansharpened image as well as the model parameters from the MS low resolution and PAN images.
The rest of the paper is organized as follows: a categorization and short review of related pansharpening methods is presented in Section 2. In Section 3 the pansharpening problem is mathematically formulated. Following the Bayesian modelling and inference, in Section 4 we propose a fully Bayesian method for the estimation of all the problem unknowns and model parameters. In Section 5, the quality of the pansharpened images is assessed both visually and quantitatively and compared with other classic and state-of-the-art pansharpening methods. In Section 6 we discuss the obtained results and finally, Section 7 concludes the paper.

2. Related Work

Early pansharpening techniques, such as in the Brovey method [5], substituted some bands for image visualization or performed simple arithmetic transformations. Other classical methods included the transformation of the MS image and the substitution of one of its components by the high spatial resolution PAN image. Examples of this strategy are Principal Components Analysis PCA substitution [6], Brovey Transform [7] and Intensity-Hue-Saturation (IHS) [8] methods. A review of those early methods, among others, can be found in [9].
Over the past 20 years, numerous methods have been presented and, in an attempt to bring some order to the diversity of approaches, different reviews, comparisons and classifications have been proposed in the literature (see, for instance, [10,11,12,13,14,15,16,17]) each one with different criteria and, therefore, with a different categorization. Nevertheless, in the last years, there seems to be a consensus in three main categories, namely Component Substitution (CS), Multi-Resolution Analysis (MRA) and Variational Optimization (VO) [15,16,17]. Additionally, the increasing number of Deep Learning (DL)-based pansharpening methods proposed in recent years can be regarded as a new category.
The Component Substitution (CS) category includes the most widely used pansharpening methods. CS methods [12] usually upsample the MS image to the size of the PAN image and transform it to another space that separates the spatial and spectral image components. Then, the transformed component containing the spatial information is substituted by the PAN image (possibly, after histogram matching). Finally, the backward transform is applied to obtain the pansharpened image. Examples of these methods include the already mentioned PCA substitution [6], IHS methods [8,18,19], the Gram–Schmidt (GS) methods [20] and Brovey transform [7]. In [21], the transformation is replaced by any weighted average of the MS bands. It is shown that this approach generalizes any CS image fusion method. Determination of the weights has been carried out in different ways. For instance, in [22] the weights are optimally estimated to minimize the mean squared error while in [23] they are set to the correlation coefficient between a single band low resolution image (obtained from the MS image) and each MS band. A local criterion, based on the belonging of a given pixel to a fuzzy cluster, was applied in [24] to estimate weights that are different for each pixel of the image. To obtain a crisper MS high-resolution image, in [25] a Wiener deconvolution of the upsampled MS bands was performed before fusion.
In general, CS-based methods produce spectral distortions due to the different statistics of the PAN image and the transformed component containing the spatial details. To tackle this issue, Multi-Resolution Analysis (MRA) methods decompose the MS and PAN images to different levels, extract spatial details from the decomposed PAN image, and inject them into the finer scales of the MS image. This principle is also known as the ARSIS concept [10]. The High-Pass Filtering (HPF) algorithm in [11,18], can be considered to be the first approach in this category where only two levels are considered. Multi-scale decompositons, such as the wavelet transform (WT) [26,27,28], the Generalized Laplacian Pyramid (GLP) [29,30,31] or the Non-Subsampled Contourlet Transform (NSCT) [32,33,34], were used to bring more precision to the methods. The “a trous” wavelet transform (AWT) was the preferred decomposition technique [26,28] until the publication of [31] showed the advantages of GLP over AWT. This was later corroborated in [14] where a comparison of different methods based on decimated and undecimated WT, AWT, GLP and NSCT concluded that GLP outperforms AWT because it better removes aliasing. MRA category also includes the Smoothing Filter Based Intensity Modulation (SFIM) method [35,36], which first upsamples the MS image to the size of the PAN one and then uses a simplified solar radiation and land surface reflection model to increase its quality, and the Indusion method [37] in which upscaling and fusion steps are carried out together.
Deep Learning (DL) techniques have gained prominence in the past years and several methods have been proposed for pansharpening. As far as we know, the use of Deep Neural Networks (DNN) for pansharpening were first introduced in [38] where a Modified Sparse Denoising Autoencoder  (MSDA) algorithm was proposed. For the same task, a Coupled Sparse Denoising Autoencoder (CSDA) was used in [39]. Convolutional neural networks were introduced in [40] and also used, for instance, in [41]. Instead of facing the difficult task of learning the whole image, residual networks [42,43] learn, from upsampled MS and PAN patches, only the details of the MS high-resolution image that are not already in the upsampled MS image and add them to it to obtain the pansharpened image. To adjust the size of the MS image to the size of the PAN one in a coarse-to-fine manner, two residual networks in cascade were set in the so called Progressive Cascade Deep Residual Network (PCDRN)  [44]. In [45] a multi-scale approach is followed by learning a DNN to upsample each NSCT directional sub-band from the MS and PAN images. In general, the main weaknesses of the DL techniques are the high computational resources needed for training, the need of a huge amount of training data, which, in the case of pansharpening, might not be available, and the poor generalization to satellite images not used during training. The absence of ground-truth MS high-resolution images, needed for training these DL methods, is a problem pointed-out by [46] where a non-supervised generative adversarial network (Pan-GAN) was proposed. The GAN aims to generate pansharpened images that are consistent with the spectral information of the MS image while maintaining the spatial information of the PAN image. However, the generalization of this technique to satellite images different from the ones used for training is not clear. The adaptation of general image fusion methods, like the U2Fusion method in [47], to the pansharpening problem is a promising research area.
From a practical perspective, Variational Optimization (VO)-based methods present advantages both from a theoretical as well as computational points of view [48]. VO-based methods mathematically model the relation between the observed images and the original MS high resolution image, building an energy functional based on some desired properties of the original image. The pansharpened image is obtained as the image that minimizes this energy functional [49]. This mathematical formulation allows to rigorously introduce and process features that are visually important into the energy functional. Variational optimization can be considered as a particular case of the Bayesian approach [50], where the estimated image is obtained by maximizing the posterior probability distribution of the MS high resolution image. Bayesian methods for pansharpening formulate the relations between the observed images and the original MS high resolution image as probability distributions, model the desired properties as prior distributions and use Bayes’ theory to estimate the pansharpened image based on the posterior distribution of the original MS high resolution  image.
Following the seminal P+Xs method [51], the PAN image is usually modelled as a combination of the bands of the original high resolution mutispectral image. However, in [49] this model was generalized by substituting the intensity images by their gradients. Note that while the P+Xs method [51] preserves spectral information, it produces blurring artifacts. To remove blur while preserving spectral similarity, other restrictions are introduced as reasonable assumptions or prior knowledge about the original image such as Laplacian prior [52], total variation [53,54], sparse representations [55], band correlations [56,57], non-local priors [58,59], etc. Spectral information is also preserved by enforcing the pansharpened image to be close to the observed MS one when downsampled to the size of the latter [52,60,61]. A special class of VO-based methods are the super-resolution methods which model pansharpening as the inverse problem of recovering the original high-resolution image by fusing the MS image and the PAN (see [52,62] for a recent review and [63] for a recent work). Deconvolution methods, such as [64], also try to solve the inverse problem but the upsampling of the MS image to the size of the PAN one is performed prior to the pansharpening procedure. Registration and fusion are carried out simultaneously in [65].
Note that the variational Bayesian approach, also followed in this paper, is more general than variational optimization. While VO-based methods aim at obtaining a single estimate of the pansharpened image, the variational Bayesian approach estimates the whole posterior distribution of the pansharpened images and the model parameters, given the observations. When a single image is needed, the mode of the distribution is usually selected, but other solutions can be obtained, for instance, by sampling the estimated distribution. Even more, the proposed approach allows us to simultaneously estimate the model parameters along with the pansharpened image using the same framework.

3. Problem Formulation

Let us denote by y the MS high-resolution image hypothetically captured with an ideal high-resolution sensor with B bands y b , b = 1 , , B , of size p = m × n pixels, that is, y = [ y 1 T , , y B T ] T , where the superscript T denotes the transpose of a vector or matrix. Note that each band of the image is flattened into a column vector containing its pixels in lexicographical order. Unfortunately, this high-resolution image is not available in real applications. Instead, we observe an MS low-resolution image Y = [ Y 1 T , , Y B T ] T with B bands Y b of size P = M × N pixels with M < m , N < n .
The bands in this image are flattened as well to express them as a column vector. The relation between each low-resolution band, Y b , and its corresponding high-resolution one, y b , is defined by
Y b = D H y b + n b = B y b + n b ,
where D is P × p decimation operator, H is a p × p blurring matrix, B = D H , and the capture noise n b is modeled as additive white Gaussian noise with variance β b 1 .
A single band high-resolution PAN image covering a wide range of frequencies is also provided by the sensor. This PAN image x of size p = m × n is modelled as an spectral average of the unknown high-resolution bands y b , as 
x = b = 1 B λ b y b + v ,
where λ b > 0 are known quantities that depend on each particular satellite sensor, and the capture noise v is modeled as additive white Gaussian noise with variance γ 1 .
Once the image formation is formulated, let us use the Bayesian formulation to tackle the problem of recovering y , the MS high resolution image, using the observed Y , its degraded MS low resolution and PAN x .

4. Bayesian Modelling and Inference

We model the distribution of each low resolution image Y b , b = 1 , , B , following the degradation model in Equation (1) as a Gaussian distribution with mean B y b and covariance matrix β b 1 I . Then, the distribution of the observed image Y is modelled by
p ( Y | y , β ) = b = 1 B N ( Y b | B y b , β b 1 I ) ,
with β = { β 1 , , β b } .
Analogously, using the degradation model in Equation (2), the distribution of the PAN image x is given by
p ( x | y , γ ) = N ( x | b = 1 B λ b y b , γ 1 I ) .
The starting point for Bayesian methods is to choose a prior distribution for the unknowns. In this paper, we use SG distributions as priors for the MS high resolution image as
p ( y | α ) = b = 1 B ν = 1 J p ( y b ν | α b ν ) = b = 1 B ν = 1 J i = 1 p Z ( α b ν ) exp [ α b ν ρ ( y b ν ( i ) ) ] ,
with α b ν > 0 and α = { α 11 , α L B } and Z ( α b ν ) is a partition function. In Equation (5), y b ν = F ν y b is a filtered version of the b-th band, y b , where { F ν } ν = 1 J is a set of J high-pass filters, y b ν ( i ) is the i-th pixel value of y b ν , and  ρ ( · ) is a penalty function. The image priors are placed on the filtered image y b ν . It is well-known that the application of high-pass filters to natural images returns sparse coefficients. Most of the coefficients are zero or close to zero while only the edge related coefficients remain large. Sparse priors enjoy SG properties, heavier tails, more peaked and positive excess kurtosis compared to the Gaussian distribution. The distribution mass is located around zero, but large values have a higher probability than in a Gaussian distribution. For  p ( y b ν | α b ν ) in Equation (5) to be SG, ρ ( · ) has to be symmetric around zero and the function ρ ( s ) increasing and concave for s ( 0 , ) . This condition is equivalent to ρ ( s ) / | s | being decreasing on ( 0 , ) , and allows ρ to be represented as
ρ y b ν ( i ) = inf η b ν ( i ) > 0 1 2 η b ν ( i ) y b ν 2 ( i ) ρ * 1 2 η b ν ( i )
ρ y b ν ( i ) L y b ν ( i ) , η b ν ( i ) = 1 2 η b ν ( i ) y b ν 2 ( i ) ρ * 1 2 η b ν ( i ) ,
where inf denotes the infimum, ρ * ( · ) is the concave conjugate of ρ ( · ) and η b ν = { η b ν ( i ) } i = 1 p are a set of positive parameters. The relationship dual to Equation (6) is given by [66]
ρ * 1 2 η b ν ( i ) = inf y b ν ( i ) 1 2 η b ν ( i ) y b ν 2 ( i ) ρ y b ν ( i ) .
To achieve sparsity, the function ρ should suppress most of the coefficients in y b ν and preserve a small number of key features. Table 1 shows some penalty functions, corresponding to SG distributions (see [1]).
From Equations (3)–(5), the joint probability distribution p ( Θ , Y , x ) , with  Θ = { y , β , γ , α } the set of all unknowns, is given by
p ( Θ , Y , x ) = p ( Y | y , β ) p ( x | y , γ ) p ( y | α ) p ( β ) p ( γ ) p ( α ) ,
where flat hyperpriors p ( β ) , p ( γ ) and p ( α ) on the model hyperparameters have been included.
Following the Bayesian paradigm, inference will be based on p ( Θ | Y , x ) . Since this posterior distribution cannot be analytically calculated due to the form of the SG distribution, in this paper we use the mean-field variational Bayesian model [67] to approximate p ( Θ | Y , x ) by the distribution q ( Θ ) of the form q ( Θ ) = θ Θ q ( θ ) , that minimizes the Kullback–Leibler divergence [68] defined as
KL q ( Θ ) | | p ( Θ | Y , x ) = q ( Θ ) log q ( Θ ) p ( Θ , Y , x ) d Θ + log p ( Y ) + log p ( x ) .
The Kullback–Leibler divergence is always non-negative and it is equal to zero if and only if q ( Θ ) = p ( Θ | Y , x ) .
Even with this factorization, the SG prior for y hampers the evaluation of this divergence, but the quadratic bound for ρ in Equation (7) allows us to bound the prior in Equation (5) with a Gaussian form such that
p y b ν ( i ) | α b ν Z ( α b ν ) exp [ α b ν L y b ν ( i ) , η b ν ( i ) ] , η b ν ( i ) > 0 .
We then define the lower bound of the prior M ν ( y ν , η ν | α ν ) = b M ν b ( y b ν , η b ν | α b ν ) where
M ν b ( y b ν , η b ν | α b ν ) = i = 1 p Z ( α b ν ) exp [ α b ν L y b ν ( i ) , η b ν ( i ) ]
and obtain the lower bound of the joint probability distribution
F ( Θ , Y , x , η ) = p ( Y | y , β ) p ( x | y , γ ) ν = 1 J M ν ( y ν , η ν | α ν )
to obtain the inequality log p ( Θ , Y , x ) log F ( Θ , Y , x , η ) .
Utilizing the lower bound F ( Θ , Y , x , η ) for the posterior probability distribution in Equation (10) we minimize KL q ( Θ ) | | F ( Θ , Y , x , η ) instead of KL q ( Θ ) | | p ( Θ | Y , x ) .
As shown in [67], for each unknown θ Θ , the estimated q ( θ ) will have the form
q ( θ ) exp log F ( Θ , Y , x , η ) q ( Θ θ ) ,
where Θ θ represents all the variables in Θ except θ and · q ( Θ θ ) denotes the expected value calculated using the distribution q ( Θ θ ) . When point estimates are required θ ^ = θ q ( θ ) is used.
For variables with a degenerate posterior approximation, that is, for  θ { β , γ , α } , the value where the posterior degenerates is given by [67]
θ ^ = arg max θ log F ( Θ , Y , x , η ) q ( Θ θ ) .
Let us now obtain the analytic expressions for each unknown posterior approximation.

4.1. High Resolution Multispectral Image Update

Using Equation (14) we can show in a straightforward way that the posterior distribution for the high resolution MS image will have the form
q ( y ) = N ( y | y , Σ y ) ,
where the inverse of the covariance matrix is given by
Σ y 1 = diag ( β ) B T B + γ ( λ λ T ) I p × p + ν α 1 ν F ν T diag ( η 1 ν ) F ν 0 p × p 0 p × p 0 p × p α 2 ν F ν T diag ( η 2 ν ) F ν 0 p × p 0 p × p 0 p × p α B ν F ν T diag ( η B ν ) F ν ,
with ⊗ denoting the Kronecker product, diag ( · ) is a diagonal matrix formed from the elements of a vector and the mean is obtained as
Σ y 1 y = diag ( β ) B T Y + γ diag ( λ ) I p × p x T , x T , , x T T .

4.2. Variational Parameters Update

To estimate the value of the variational parameters, η introduced in Equation (7), we need to solve, for each band b { 1 , , B } , filter ν { 1 , , J } , and pixel i { 1 , , p } , the optimization problem
η ^ b ν ( i ) = arg min η b ν ( i ) L y b ν ( i ) , η b ν ( i ) q ( y ) = arg min η b ν ( i ) 1 2 η b ν ( i ) u b ν 2 ( i ) ρ * 1 2 η b ν ( i ) ,
where u b ν ( i ) = y b ν 2 ( i ) . Since
ρ * η ^ b ν ( i ) 2 = min x 1 2 η ^ b ν ( i ) x 2 ρ ( x )
whose minimum is achieved at x = u b ν ( i ) , we have, differentiating the right hand side of (19) with respect to x,
η ^ b ν ( i ) = ρ ( u b ν ( i ) ) ) / u b ν ( i ) .

4.3. Model Parameters Update

The estimates of the noise variance in the degradation models in Equations (3) and (4) are obtained using Equation (15) as
β ^ b 1 = tr ( Y b B y b ) ( Y b B y b ) T q ( Θ ) P , b = 1 , , B ,
γ ^ 1 = tr ( x b = 1 B λ b y b ) ( x b = 1 B λ b y b ) T q ( Θ ) p ,
where tr ( · ) represents the trace of the matrix.
From Equation (14) we obtain the following distribution for the parameter α b ν of the SG prior in Equation (5).
q ( α b ν ) = c o n s t + i = 1 p log Z ( α b ν ) exp α b ν ρ ( y b ν ( i ) ) .
The mode of this distribution can be obtained (see [69]) by solving
Z ( α ^ b ν ) α ^ ν s = tr F ν T F ν y b ν y b ν T p .
The p penalty function shown in Table 1 produces proper priors, for which the partition function can be evaluated, but the log penalty function produces an improper prior. We tackle this problem examining, for  α b ν 1 , the behavior of
Z ( α b ν , K ) 1 = K K exp α b ν ρ ( t ) d t
and keeping in Z ( α b ν ) / α b ν the term that depends on α b ν . This produces for the log prior
Z ( α ^ b ν ) α ^ b ν = ( α ^ b ν 1 ) 1 .

4.4. Calculating the Covariance Matrices

The matrix Σ y in Equation (17) must be explicitly computed to find its trace and also to calculate η ^ b ν ( i ) . However, since its calculation is very intense, we propose the following approximation. We  first approximate diag ( η b ν ) using
diag ( η b ν ) z ( η b ν ) I p × p ,
where z ( η b ν ) is calculated as the mean of the values in η b ν .
We then use the approximation
Σ y 1 Σ y 1 1 0 p × p 0 p × p 0 p × p Σ y 2 1 0 p × p 0 p × p 0 p × p Σ y B 1
with
Σ y b 1 β b B T B + γ λ b 2 I p × p + ν α b ν z ( η b ν ) F ν T F ν = C b , b = 1 , , B .
Finally we have
y b ν 2 ( i ) ( y b ν ( i ) ) 2 + 1 p tr C b 1 F ν T F ν .

4.5. Proposed Algorithm

Based on the previous derivations, we propose the Variational Bayesian SG Pansharpening Algorithm in Algorithm 1. The linear equations problem in Equation (18), used in step 4 of Algorithm 1, has been solved using the Conjugate Gradient approach.
Algorithm 1: Variational Bayesian SG pansharpening.
Require: Observed multispectral image, Y , panchromatic image x , and  λ parameter.
 Set Σ y ( 0 ) = 0 and n = 0 . y b ( 0 ) is obtained by bicubic interpolation of Y b , b = 1 , , B .
while convergence criterion is not met do
  1. Set n = n + 1 .
  2. Obtain β ( n ) , γ ( n ) and α b ν ( n ) from Equations (22), (23) and (25) respectively.
  3. Using y ( n 1 ) and Σ y ( n 1 ) , update the variational parameters η ^ b ν ( n ) , b , ν from Equation (21).
  4. Using β ( n ) , γ ( n ) , α b ν ( n ) , and  η ^ b ν ( n ) , update Σ y 1 ( n ) in Equation (17) and solve Equation (18) for y ( n ) .
end while
 Output the high resolution hyperspectral image y ^ = y ( n ) .

5. Materials and Methods

To test the performance of the proposed methodology on different kind of images, five satellite images were used: three LANDSAT 7-ETM+ [70] images, a SPOT-5 [71] image and a FORMOSAT-2 [72] image. LANDSAT MS images have six bands and a ratio between PAN and MS images p / P = 4 . Figure 1 and Figure 2 show RGB color images formed by the bands B4, B3 and B2 of LANDSAT MS images, and their corresponding PAN images. Figure 1 corresponds to an area from Chesapeake Bay (US) while Figure 2 depicts two areas from Neatherland.SPOT-5 MS images have four bands and two PAN images, with resolution ratios of p / P = 4 and p / P = 16 , are provided. FORMOSAT-2 MS images also have four bands and a ratio between PAN and MS images p / P = 16 . Figure 3a,c show the RGB color images formed from bands B3, B2 and B1 bands of a SPOT-5 image from Roma (IT) and a FORMOSAT-2 MS image from Salon-de-Provence (FR) and Figure 3b,d their corresponding PAN images.
Both the observed Y and x images have been normalized to the range [ 0 , 1 ] before running Algorithm 1. The convergence criterion in the algorithm was y ( n ) y ( n 1 ) 2 / y ( n ) 2 10 6 or 50 iterations were reached, whatever occurs first. The relationship between the MS high resolution image and the panchromatic image in Equation (2) is governed by the parameters λ that need to be set before pansharpening is carried out. If we knew the sensor spectral response characteristics, the values of λ could be estimated from them. For instance, for LANDSAT 7-ETM+, Figure 4 shows the sensor spectral response curves for the MS bands B1-B6, shown in color, and the PAN band shown in black. For this sensor, the PAN band mainly overlaps B2-B4 MS bands, and λ coefficients could be obtained from this overlapping (see [52]). In this paper, however, a more general approach is followed to estimate λ from the observations. First, we define X = D x , a version of the PAN image downsampled to the size of the MS image. Then, since the sensor spectral response is the same in high and low resolution, the parameters λ can be obtained by solving
λ = argmin λ X b = 1 B λ b Y b 2 ,
subject   sto λ b 0 , b , b = 1 B λ b = 1 .
Table 2 shows the λ s associated to the different considered observed images. For the LANDSAT 7-ETM+ images only the first four bands are positive and λ 5 and λ 6 values are 0 since we know that bands B5 and B6 are not covered by the panchromatic sensor. For this process, each band is normalized to the interval [ 0 , 1 ] . Note that due to the normalization, the estimated λ values do not only depend on the sensor spectral response but also on the observed area characteristics. This explains the differences between the obtained λ values for the images in Figure 2a,c. Although those images are from the same area of Netherlands, clouds in Figure 2a modify the estimation of the values of λ .

6. Discussion

Within the variational Bayesian methodology, two methods are proposed in this paper: one using the log penalty function (see Table 1), hence, named log method, and another using the p penalty function, with p = 1 , referred as 1 method. The proposed methods have been compared with the following classic and state-of-the-art pansharpening methods: the Principal Component Analysis (PCA) [6], the Intensity–Hue–Saturation (IHS) transform [19], the Brovey transform (Brovey) [7], the Band-Dependent Spatial-Detail (BDSD) method in [22], the Gram-Schmidt (GS) method in [20], the Gram-Schmidt adaptive (GSA) method in [21], the Partial Replacement Adaptive Component Substitution (PRACS) method in [23], the High Pass Filtering (HPF) algorithm in [18], the Smoothing Filter Based Intensity Modulation (SFIM) method [35,36], the Indusion method in [37], the Additive A Trous Wavelet Transform (ATWT) in [26], the Additive Wavelet Luminance Proportional (AWLP) method in [28], the ATWT Model 2 (ATWT-M2) and ATWT Model 3 (ATWT-M3) methods in [10], the Generalized Laplacian Pyramid (GLP)-based methods in [29], concretely the modulation transfer functions (MTF)-GLP, GLP with High Pass Modulation (MTF-GLP-HPM), and GLP with Context Based Decision (MTF-GLP-CBD) methods, and the pansharpening method using a Total Variation (TV) image model in [53]. We have used the implementation of the methods and measures provided by the Pansharpening Toolbox (https://rscl-grss.org/coderecord.php?id=541) [13]. For those methods not included in the toolbox we have used the code provided by the authors. The code of the proposed methods will be publicly available at https://github.com/vipgugr. We have also included the results of bilinear interpolating the MS image to the size of the PAN, marked as EXP, as a reference. Both quantitative and qualitative comparisons of the different methods have been performed.

6.1. Quantitative Comparison

A common problem in pansharpening is the nonexistence of a MS high resolution ground-truth image to compare with. Hence we performed two kinds of quantitative comparisons. Firstly, the images obtained using the different methods have been compared following Wald’s protocol [73] as follows: the observed MS image, Y , and the PAN image, x , are downsampled by applying the operator D to generate low resolution versions of them. Then, pansharpening is applied to those low resolution images and the obtained estimation of the MS image, y ^ , is quantitatively compared with the observed MS image, Y . Secondly, the different methods have been compared using Quality with No Reference (QNR) measures [13,74]. As previously stated, for the LANDSAT image in Figure 1, the resolution ratio between MS and PAN images is p / P = 4 . Since the SPOT-5 satellite provides two PAN images, two experiments were carried out on the image in Figure 3, one with a decimation ratio of 4 and another with a ratio of 16. For the FORMOSAT-2 image the ratio is p / P = 16 . However, for the sake of completeness, two experiments were also carried out, one assuming a decimation ratio of 4 and another with a ratio of 16.
Both spatial and spectral quality metrics have been used to compare the results obtained using the different methods. Details for the metrics used is shown below:
Spatial measures:
  • Q
    Universal Quality Index (UQI) [75] averaged on all MS bands.
    Range: [-1, 1]
    The higher the the better.
  • Q4, Q8
    Instances of the Q 2 n [76] index taking values. Suitable to measure quality for multiband images having an arbitrary number of spectral bands. Q4 is used for SPOT-5 and FORMOSAT-2 images which have four bands and Q8 for the LANDSAT image with six bands.
    Range: [0, 1]
    The higher the better.
  • Spatial Correlation Coefficient (SCC) [77]
    Measures the correlation coefficient between compared images after the application of a Sobel filter.
    Range: [0, 1]
    The higher the better.
  • QNR spatial distortion ( D S ) [78]
    Measures the spatial distortion between MS bands and PAN image.
    Range: [0, 1]
    The lower the better.
Spectral measures:
  • Spectral Angle Mapper (SAM) [79]
    For spectral fidelity. Measures the mean angle between the corresponding pixels of the compared images in the space defined by considering each spectral band as a coordinate axis
    Range: [0, 180]
    The lower the better.
  • Erreur Relative Globale Adimensionnelle de Synthese (ERGAS) [80]
    Measures spectral consistency between compared images.
    Range: [ 0 , [
    The lower ERGAS value the better consistency, specially for values lower than the number of image bands B.
  • QNR spectral distortion ( D λ ) [78]
    This measure is derived from the differences between the inter-band Q index values computed for HR and LR images.
    Range: [0, 1]
    The lower, the better.
Spatial and spectral measures:
  • Jointly Spectral and Spatial Quality Index (QNR) [78]
    QNR is obtained as the product of (1- D S ) and (1- D λ ).
    Range: [0, 1]
    The higher the better.
Table 3 shows the obtained figures of merit using Wald’s protocol for the LANDSAT image in Figure 1. As it is clear from the table, 1 outperforms all the other methods both in spectral fidelity and the incorporation of spatial details. Note the high SCC value (meaning that the details in the PAN image have been successfully incorporated into the pansharpened image) while also obtaining the lowest spectral distortion as evidenced by the SAM and ERGAS values. The TV method obtains the second best results except for the SAM metric, for this metric, the proposed log method has the second best value. This method also obtains the third best values for ERGAS and SCC measures. GLP based and PRACS methods also obtain high values for the Q, Q8 indices and low value for SAM. However, their ERGAS and SCC performance is worse. Table 4 shows the QNR quantitative results for the LANDSAT image in Figure 1. In this table, the proposed methods achieve competitive results. Log obtains the best D λ value and this method together with 1 obtain second and third QNR scores, respectively. Note that EXP obtained the highest score using QNR since bilinear interpolation of the observed MS low resolution image is used as the MS high resolution estimation to calculate D S and D λ calculations.
Table 5 and Table 6 show the quantitative results using Wald’s protocol for the LANDSAT images in Figure 2a,c, respectively. PRACS outperforms all other methods on the image in Figure 2a (see Table 5) and the proposed 1 and log obtain the first and second best scores on the image in Figure 2c (see Table 6). Table 7 and Table 8 show the obtained QNR figures of merit for those two images. The proposed methods produce good D S , D λ and QNR values for both images, both above 0.9 which supports their good performance. Again the EXP results are the best in all the measures for Table 8 and provides the best D λ , for the image associated to Table 7. The 1 method obtains the best D S for this image and BDSD the highest QNR.
Figure 5 and Figure 6 show a zoomed in region of the RGB color images formed by bands B4, B3, and B2 of MS ground truth images used to apply Wald’s protocol and also the absolute error images for the methods in Table 7 and Table 8. In those images, the darker the intensity the lower the absolute error. Figure 5 and Figure 6 are consistent with the quantitative comparison shown in Table 5 and Table 6, respectively. The best results for the image in Figure 2a were obtained using PRACS, while for the image in Figure 2c the best performing method is 1. Note that brighter areas in Figure 5e,f correspond to the borders of cloudy areas in Figure 2a. We argue that since clouds alter the weights of λ estimated using Equation (30), the boundaries of clouds and land areas in Figure 2a are not well resolved. This explains a worse behavior of the proposed methods in the cloudy areas of this image.
Table 9 and Table 10 show, respectively, the quantitative results using Wald’s protocol for the SPOT-5 and the FORMOSAT-2 images in Figure 3 for the decimation ratios p / P = 4 and p / P = 16 . The proposed log obtains the best figures of merit for the SPOT image in Figure 3a with p / P = 4 except for Q and Q4 metrics. The Q values obtained by log and 1 are slightly lower than those obtained by BDSD. Note that BDSD achieved the third best general figures just below the proposed log and 1 algorithms. With p / P = 16 the proposed log algorithm provides the best results except for Q, Q4 and SAM values, where competitive values are obtained. The proposed log achieves a slightly lower Q value than PRACS and a slightly higher SAM value than Brovey. In general, PRACS is the second best performing method for this image for p / P = 16 . For the FORMOSAT-2 image in Figure 3c, the proposed 1 and log algorithms obtained the best numerical results for a p / P = 4 magnification. Both methods provide similar results, which are better than all the one provided by the competing methods. For a ratio p / P = 16 , there is not a clear winner. The proposed methods are competitive in this image although they do not stand out in any of the measures. Table 11 and Table 12 show, respectively, the QNR quantitative results for the SPOT-5 and the FORMOSAT-2 images in Figure 3 for the decimation ratios p / P = 4 and p / P = 16 . In Table 11, EXP achieves the best D S , D λ and QNR scores. In this table, the proposed methods obtain good scores. The log method obtains the second best D λ and D S values and very high QNR values for both decimation ratios. Results for the FORMOSAT image, shown in Table 12, are very similar although in this case, BDSD obtains the best D S and QNR values for p / P = 4 and ATWT-M3 for p / P = 16 .
Table 13 shows the required CPU time in seconds on a 2.40GHz Intel® Xeon® CPU for the pansharpening of a MS image with 4 bands to a 1024 × 1024 size, for p / P = 4 and p / P = 16 , using the different methods under comparison. Equation (18) has been solved using the Conjugate Gradient method which required, to achieve convergence, less than 30 iterations for the 1 prior and at least 1000 iterations for the log prior. This explains the differences between their required CPU time. Note that the proposed methods automatically estimate the model parameters which increases the running time but makes our methods parameter free.

6.2. Qualitative Comparison

Figure 7 shows a small region of interest of the observed LANDSAT-7 images in Figure 1 and the pansharpening results with p / P = 4 obtained by the proposed methods and the competing ones with the best quantitative performance, that is, PRACS, MTF-GLP-HPM, MTF-GLP-CBD and TV methods. All color images in this figure are RGB images formed from the B4, B3 and B2 Landsat bands. Since we are using full resolution images, there is no ground truth to compare with, so a visual analysis of the resulting images is performed. The improved resolution of all the pansharpening results in Figure 7c–h with respect to the observed MS image in Figure 7a is evident. PRACS, MTF-GLP-HPM and MTF-GLP-CBD images in Figure 7c–e have a lower detail level than TV and the proposed 1 method, see Figure 7f,g, respectively. See, for instance, the staircase effects in some diagonal edges not present in the TV and proposed 1 method results. The PRACS, MTF-GLP-HPM and MTF-GLP-CBD methods produce similar, but lower, spectral quality than the proposed method, which is consistent with the numerical results in Table 3 and discussion presented in Section 6.1. The image obtained using the 1 method, Figure 7g, has colors closer to those of the observed MS image than the TV image, Figure 7f, which is also somewhat noisier. The log method is very good at removing noise in the image (see the sea area) but it tends to remove other fine details too.
Figure 8 shows a region of interest of the observed SPOT-5 images in Figure 3a,b and the pansharpening results with p / P = 16 obtained using the competing methods with the best performance on this image, that is, Brovey, PRACS, ATWT-M3, and TV methods, and the proposed 1 and log methods. All color images in this figure are RGB images formed from the bands B3, B2 and B1 of the SPOT-5 image. The Brovey method (Figure 8c) produces the highest spectral distortion, however, it also recovers more spatial details in the image (see the airport runway and plane). ATWT-M3 (Figure 8e), on the other hand, produces a blurry image. PRACS produces a sharper image, see Figure 8d, but details in the PAN image do not seem to be well-integrated. The TV method (Figure 8f) and the proposed 1 and log methods (Figure 8g,h obtain the most consistent results, with high spatial details and low spectral distortion. However, TV introduces staircase artifacts on diagonal lines that are not noticeable in the 1 and log images. As with the LANDSAT-7 image, the log image in Figure 8h lacks some small details, removed by the method along with noise.

7. Conclusions

A variational Bayesian methodology for the pansharpening problem has been proposed. In this methodology, we model the relation between the MS high resolution image and the PAN image as a linear combination of the MS bands whose weights are estimated from the available data. The observed MS image is modelled as a downsampled version of the original MS image. The expected characteristics of the pansharpened image are incorporated in the form of SG sparse image priors. Two penalty functions corresponding to SG distributions are used, 1 and log. All the unknowns and model parameters have been automatically estimated within the variational Bayesian modelling and inference, and an efficient algorithm has been obtained.
The proposed 1 and log methods have been compared to classic and state-of-the-art methods obtaining very good results both quantitative and qualitatively. In general, they have obtained the best quantitative results for LANDSAT-7 ETM+, SPOT-5 and FORMOSAT-2 images with a resolution ratio of 4 and SPOT-5 with a resolution ratio of 16. Competitive results were also obtained for the FORMOSAT-2 image with a resolution ratio of 16. They stand out in terms of spectral consistency while improving the spatial resolution of pansharpened images. We argue that the superior spectral consistency of SG methods arises from the modelling of the PAN image which selectively incorporates PAN detailed information into the different MS high resolution bands without changing their spectral properties. Qualitatively, SG methods produce results consistent with the observed PAN and MS images and with the numerical results previously described. The log method is better at removing noise in the images, at the cost of removing some fine details.

Author Contributions

Conceptualization, M.V., R.M.; methodology, M.V.; software, F.P.-B., M.V.; validation, F.P.-B.; formal analysis, J.M., M.V.; investigation, F.P.-B., M.V.; resources, J.M., M.V.; data curation, F.P.-B.; writing—original draft preparation, F.P.-B., M.V.; writing—review and editing, A.K.K., J.M. and R.M.; visualization, F.P.-B., J.M. and M.V.; supervision, A.K.K., R.M.; project administration, J.M., R.M.; funding acquisition, J.M., R.M. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Spanish Ministerio de Economía y Competitividad under contract DPI2016-77869-C2-2-R, by the Ministerio de Ciencia e Innovación under contract PID2019-105142RB-C22, and the Visiting Scholar Program at the University of Granada.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
MSmultispectral
PANpancrhomatic
PCAprincipal components analysis
IHSintensity-hue-saturation
CScomponent substitution
MRAmulti-resolution analysis
VOvariational optimization
DLdeep learning
CScomponent substitution
GSGram-Schmidt
HPFhigh-pass filtering
WTwavelet transform
GLPgeneralized Laplacian pyramid
NSCTnon-subsampled contourlet transform
AWT“a trous” wavelet transforms
SFIMsmoothing filter based intensity modulation
DNNdeep neural networks
MSDAmodified sparse denoising autoencoder
CSDAcoupled sparse denoising autoencoder
PCDRNprogressive cascade deep residual network
GANgenerative adversarial network
SGsuper-Gaussian
BroveyBrovey transform
BDSDband-dependent spatial-detail
GSAGram-Schmidt adaptive
PRACSpartial replacement adaptive component substitution
ATWTadditive “a trous” wavelet transform
ATLPadditive wavelet luminance proportional
MTFmodulation transfer functions
HPMhigh pass modulation
CBDcontext based decision
TVtotal variation
UQIuniversal quality index
SCCspatial correlation coefficient
SAMspectral angle mapper
ERGASerreur relative globale adimensionnelle de synthese
QNRquality with no reference

References

  1. Babacan, S.D.; Molina, R.; Do, M.N.; Katsaggelos, A.K. Bayesian blind deconvolution with general sparse image priors. In Proceedings of the European Conference on Computer Vision, Florence, Italy, 7–13 October 2012. [Google Scholar]
  2. Rubio, J.; Vega, M.; Molina, R.; Katsaggelos, A.K. A general sparse image prior combination in Compressed Sensing. In Proceedings of the 21st European Signal Processing Conference (EUSIPCO 2013), Marrakech, Morocco, 9–13 September 2013; pp. 1–5. [Google Scholar]
  3. Zhou, X.; Vega, M.; Zhou, F.; Molina, R.; Katsaggelos, A.K. Fast bayesian blind deconvolution with Huber super Gaussian priors. Digit. Signal Process. 2017, 60, 122–133. [Google Scholar]
  4. Pérez-Bueno, F.; Vega, M.; Naranjo, V.; Molina, R.; Katsaggelos, A. Fully automatic blind color deconvolution of histological images using super Gaussians. In Proceedings of the 28th European Signal Processing Conference, EUSIPCO 2020, Amsterdam, The Netherlands, 18–22 January 2021. [Google Scholar]
  5. Ehlers, M. Multisensor image fusion techniques in remote sensing. ISPRS J. Photogramm. Remote Sens. 1991, 46, 19–30. [Google Scholar]
  6. Chavez, P.S.; Kwarteng, A.Y. Extracting spectral contrast in Landsat thematic mapper image data using selective principal component analysis. Photogramm. Eng. Remote Sens. 1989, 55, 339–348. [Google Scholar]
  7. Gillespie, A.R.; Kahle, A.B.; Walker, R.E. Color enhancement of highly correlated images. II. Channel ratio and “chromaticity” transformation techniques. Remote Sens. Environ. 1987, 22, 343–365. [Google Scholar]
  8. Carper, W.J.; Lillesand, T.M.; Kiefer, R.W. The use of intensity-hue-saturation transformations for merging SPOT panchromatic and multispectral image data. Photogramm. Eng. Remote. Sens. 1990, 56, 459–467. [Google Scholar]
  9. Pohl, C.; Genderen, J.L.V. Multi-sensor image fusion in remote sensing: Concepts, methods, and applications. Int. J. Remote Sens. 1998, 19, 823–854. [Google Scholar]
  10. Ranchln, T.; Wald, L. Fusion of high spatial and spectral resolution images: The ARSIS concept and its implementation. Photogramm. Eng. Remote Sens. 2000, 66, 49–61. [Google Scholar]
  11. Schowengerdt, R.A. Remote Sensing: Models and Methods for Image Processing; Elsevier: Amsterdam, The Netherlands, 2006. [Google Scholar]
  12. Amro, I.; Mateos, J.; Vega, M.; Molina, R.; Katsaggelos, A. A survey of classical methods and new trends in pansharpening of multispectral images. EURASIP J. Adv. Signal Process. 2011, 2011, 79. [Google Scholar]
  13. Vivone, G.; Alparone, L.; Chanussot, J.; Dalla Mura, M.; Garzelli, A.; Licciardi, G.A.; Restaino, R.; Wald, L. A critical comparison among pansharpening algorithms. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2565–2586. [Google Scholar] [CrossRef]
  14. Alparone, L.; Baronti, S.; Aiazzi, B.; Garzelli, A. Spatial Methods for Multispectral Pansharpening: Multiresolution Analysis Demystified. IEEE Trans. Geosci. Remote Sens. 2016, 54, 2563–2576. [Google Scholar]
  15. Duran, J.; Buades, A.; Coll, B.; Sbert, C.; Blanchet, G. A survey of pansharpening methods with a new band-decoupled variational model. ISPRS J. Photogramm. Remote Sens. 2017, 125, 78–105. [Google Scholar] [CrossRef] [Green Version]
  16. Kahraman, S.; Erturk, A. Review and performance comparison of pansharpening algorithms for RASAT images. Istanb. Univ. J. Electr. Electron. Eng. 2018, 18, 109–120. [Google Scholar] [CrossRef] [Green Version]
  17. Meng, X.; Shen, H.; Li, H.; Zhang, L.; Fu, R. Review of the pansharpening methods for remote sensing images based on the idea of meta-analysis: Practical discussion and challenges. Inf. Fusion 2019, 46, 102–113. [Google Scholar] [CrossRef]
  18. Chavez, P.S.; Sides, S.; Anderson, J. Comparison of three different methods to merge multiresolution and multispectral data: Landsat TM and SPOT panchromatic. Photogramm. Eng. Remote. Sens. 1991, 57, 295–303. [Google Scholar]
  19. Tu, T.M.; Su, S.C.; Shyu, H.C.; Huang, P.S. A new look at IHS-like image fusion methods. Inf. Fusion. 2001, 2, 177–186. [Google Scholar] [CrossRef]
  20. Laben, C.A.; Brower, B.V. Process For Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6,011,875, 4 January 2000. [Google Scholar]
  21. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS + Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  22. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE pan sharpening of very high resolution multispectral images. IEEE Trans. Geosci. Remote Sens. 2007, 46, 228–236. [Google Scholar] [CrossRef]
  23. Choi, J.; Yu, K.; Kim, Y. A new adaptive component-substitution-based satellite image fusion by using partial replacement. IEEE Trans. Geosci. Remote Sens. 2010, 49, 295–309. [Google Scholar] [CrossRef]
  24. Shahdoosti, H.R.; Javaheri, N. Pansharpening of clustered MS and Pan images considering mixed pixels. IEEE Geosci. Remote Sens. Lett. 2017, 14, 826–830. [Google Scholar] [CrossRef]
  25. Palsson, F.; Sveinsson, J.R.; Ulfarsson, M.O.; Benediktsson, J.A. MTF-based deblurring using a Wiener filter for CS and MRA pansharpening methods. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 2255–2269. [Google Scholar] [CrossRef]
  26. Nuñez, J.; Otazu, X.; Fors, O.; Prades, A.; Pala, V.; Arbiol, R. Multiresolution-Based Image fusion with additive wavelet decomposition. IEEE Trans. Geosci. Remote. Sens. 1999, 37, 1204–1211. [Google Scholar] [CrossRef] [Green Version]
  27. King, R.; Wang, J. A wavelet based algorithm for pansharpening Landsat 7 imagery. In Proceedings of the IEEE 2001 International Geoscience and Remote Sensing Symposium, Sydney, Australia, 9–13 July 2001; pp. 849–851. [Google Scholar]
  28. Otazu, X.; Gonzalez-Audicana, M.; Fors, O.; Nuñez, J. Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods. IEEE Trans. Geosci. Remote. Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef] [Green Version]
  29. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. MTF-tailored multiscale fusion of high-resolution MS and Pan Imagery. Photogramm. Eng. Remote. Sens. 2006, 72, 591–596. [Google Scholar] [CrossRef]
  30. Aiazzi, B.; Alparone, L.; Baronti, S.; Pippi, I.; Selva, M. Generalised Laplacian pyramid-based fusion of MS + P image data with spectral distortion minimisation. ISPRS Internat. Arch. Photogramm. Remote Sens. 2002, 34, 3–6. [Google Scholar]
  31. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A.; Selva, M. Advantages of Laplacian pyramids over ”à trous” wavelet transforms for pansharpening of multispectral images. In Image and Signal Processing for Remote Sensing XVIII; International Society for Optics and Photonics: Edinburgh, UK, 2012; p. 853704. [Google Scholar]
  32. Shah, V.P.; Younan, N.H.; King, R.L. An efficient pan-sharpening method via a combined adaptive PCA approach and contourlets. IEEE Trans. Geosci. Remote Sens. 2008, 46, 1323–1335. [Google Scholar] [CrossRef]
  33. Amro, I.; Mateos, J.; Vega, M. Parameter estimation in Bayesian super-resolution pansharpening using contourlets. In Proceedings of the IEEE International Conference on Image Processing ICIP 2011, Brussels, Belgium, 11–14 September 2011; pp. 1345–1348. [Google Scholar]
  34. Upla, K.P.; Gajjar, P.P.; Joshi, M.V. Pan-sharpening based on non-subsampled contourlet transform detail extraction. In Proceedings of the 2013 Fourth National Conference on Computer Vision, Pattern Recognition, Image Processing and Graphics (NCVPRIPG), Jodhpur, India, 18–21 December 2013; pp. 1–4. [Google Scholar]
  35. Liu, J. Smoothing filter-based intensity modulation: A spectral preserve image fusion technique for improving spatial details. Int. J. Remote Sens. 2000, 21, 3461–3472. [Google Scholar] [CrossRef]
  36. Wald, L.; Ranchin, T. Liu’Smoothing filter-based intensity modulation: A spectral preserve image fusion technique for improving spatial details’. Int. J. Remote Sens. 2002, 23, 593–597. [Google Scholar] [CrossRef]
  37. Khan, M.M.; Chanussot, J.; Condat, L.; Montanvert, A. Indusion: Fusion of multispectral and panchromatic images using the induction scaling technique. IEEE Geosci. Remote Sens. Lett. 2008, 5, 98–102. [Google Scholar] [CrossRef] [Green Version]
  38. Huang, W.; Xiao, L.; Wei, Z.; Liu, H.; Tang, S. A new pan-sharpening method with deep neural networks. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1037–1041. [Google Scholar] [CrossRef]
  39. Cai, W.; Xu, Y.; Wu, Z.; Liu, H.; Qian, L.; Wei, Z. Pan-sharpening based on multilevel coupled deep network. In Proceedings of the 2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 7046–7049. [Google Scholar]
  40. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by convolutional neural networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef] [Green Version]
  41. Eghbalian, S.; Ghassemian, H. Multi spectral image fusion with deep convolutional network. In Proceedings of the 2018 9th International Symposium on Telecommunications, Tehran, Iran, 17–19 December 2018; pp. 173–177. [Google Scholar]
  42. Wei, Y.; Yuan, Q.; Shen, H.; Zhang, L. Boosting the Accuracy of Multispectral Image Pansharpening by Learning a Deep Residual Network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1795–1799. [Google Scholar] [CrossRef] [Green Version]
  43. Li, N.; Huang, N.; Xiao, L. PAN-Sharpening via residual deep learning. In Proceedings of the 2017 IEEE International Geoscience and Remote Sensing Symposium, Fort Worth, TX, USA, 23–28 July 2017; pp. 5133–5136. [Google Scholar]
  44. Yang, Y.; Tu, W.; Huang, S.; Lu, H. PCDRN: Progressive Cascade Deep Residual Network for Pansharpening. Remote Sens. 2020, 12, 676. [Google Scholar] [CrossRef] [Green Version]
  45. Huang, W.; Fei, X.; Feng, J.; Wang, H.; Liu, Y.; Huang, Y. Pan-sharpening via multi-scale and multiple deep neural networks. Signal Process. Image Commun. 2020, 85, 115850. [Google Scholar] [CrossRef]
  46. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Guo, X.; Jiang, J. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  47. Xu, H.; Ma, J.; Jiang, J.; Guo, X.; Ling, H. U2Fusion: A Unified Unsupervised Image Fusion Network. IEEE Trans. Pattern Anal. Mach. Intell. 2020. [Google Scholar] [CrossRef]
  48. Chan, T.; Shen, J.; Vese, L. Variational PDE models in image processing. Not. AMS. 2003, 50, 14–26. [Google Scholar]
  49. Fang, F.; Li, F.; Shen, C.; Zhang, G. A variational approach for pan-sharpening. IEEE Trans. Image Process. 2013, 22, 2822–2834. [Google Scholar] [CrossRef]
  50. Loncan, L.; de Almeida, L.B.; Bioucas-Dias, J.M.; Briottet, X.; Chanussot, J.; Dobigeon, N.; Fabre, S.; Liao, W.; Licciardi, G.A.; Simões, M.; et al. Hyperspectral pansharpening: A review. IEEE Geosci. Remote Sens. Mag. 2015, 3, 27–46. [Google Scholar] [CrossRef] [Green Version]
  51. Ballester, C.; Caselles, V.; Igual, L.; Verdera, J.; Rougé, B. A variational model for P+XS image fusion. Int. J. Comput. Vis. 2006, 69, 43–58. [Google Scholar] [CrossRef]
  52. Molina, R.; Vega, M.; Mateos, J.; Katsaggelos, A. Variational posterior distribution approximation in Bayesian super resolution reconstruction of multispectral images. Appl. Comput. Harmon. Anal. 2008, 24, 251–267. [Google Scholar] [CrossRef] [Green Version]
  53. Vega, M.; Mateos, J.; Molina, R.; Katsaggelos, A. Super resolution of multispectral images using TV image models. In Proceedings of the 2th International Conference on Knowledge-Based and Intelligent Information & Engineering Systems, Zagreb, Croatia, 3–5 September 2008; pp. 408–415. [Google Scholar]
  54. Palsson, F.; Sveinsson, J.R.; Ulfarsson, M.O. A new pansharpening algorithm based on total variation. IEEE Geosci. Remote Sens. Lett. 2014, 11, 318–322. [Google Scholar] [CrossRef]
  55. Yang, X.; Jian, L.; Yan, B.; Liu, K.; Zhang, L.; Liu, Y. A sparse representation based pansharpening method. Future Gener. Comput. Syst. 2018, 88, 385–399. [Google Scholar] [CrossRef]
  56. Vega, M.; Mateos, J.; Molina, R.; Katsaggelos, A. Super resolution of multispectral images using l1 image models and interband correlations. J. Signal Process. Syst. 2011, 65, 509–523. [Google Scholar] [CrossRef] [Green Version]
  57. Zhang, M.; Li, S.; Yu, F.; Tian, X. Image fusion employing adaptive spectral-spatial gradient sparse regularization in UAV remote sensing. Signal Process. 2020, 170, 107434. [Google Scholar] [CrossRef]
  58. Duran, J.; Buades, A.; Coll, B.; Sbert, C. A nonlocal variational model for pansharpening image fusion. SIAM J. Imaging Sci. 2014, 7, 761–796. [Google Scholar] [CrossRef]
  59. Fu, X.; Lin, Z.; Huang, Y.; Ding, X. A variational pan-sharpening with local gradient constraints. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 10265–10274. [Google Scholar]
  60. Li, W.; Hu, Q.; Zhang, L.; Du, J. Pan-sharpening with a spatial-enhanced variational model. J. Appl. Remote Sens. 2018, 12, 035018. [Google Scholar] [CrossRef]
  61. Chen, Y.; Wang, T.; Fang, F.; Zhang, G. A pan-sharpening method based on the ADMM algorithm. Front. Earth Sci. 2019, 13, 656–667. [Google Scholar] [CrossRef]
  62. Garzelli, A. A review of image fusion algorithms based on the super-resolution paradigm. Remote Sens. 2016, 8, 797. [Google Scholar] [CrossRef] [Green Version]
  63. Tian, X.; Chen, Y.; Yang, C.; Gao, X.; Ma, J. A Variational Pansharpening Method Based on Gradient Sparse Representation. IEEE Signal Process. Lett. 2020, 27, 1180–1184. [Google Scholar] [CrossRef]
  64. Vivone, G.; Simões, M.; Dalla Mura, M.; Restaino, R.; Bioucas-Dias, J.M.; Licciardi, G.A.; Chanussot, J. Pansharpening based on semiblind deconvolution. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1997–2010. [Google Scholar] [CrossRef]
  65. Chen, C.; Li, Y.; Liu, W.; Huang, J. SIRF: Simultaneous Satellite Image Registration and Fusion in a Unified Framework. IEEE Trans. Image Process. 2015, 24, 4213–4224. [Google Scholar] [PubMed] [Green Version]
  66. Rockafellar, R. Convex Analysis; Princeton University Press: Princeton, NJ, USA, 1996. [Google Scholar]
  67. Bishop, C. Mixture models and EM. In Pattern Recognition and Machine Learning; Springer: Berlin/Heidelberg, Germany, 2006; pp. 454–455. [Google Scholar]
  68. Kullback, S. Information Theory and Statistics; Dover Publications: Mineola, NY, USA, 1959. [Google Scholar]
  69. Vega, M.; Molina, R.; Katsaggelos, A. Parameter Estimation in Bayesian Blind Deconvolution with Super Gaussian Image Priors. In Proceedings of the 2014 22nd European Signal Processing Conference (EUSIPCO), Lisbon, Portugal, 1–5 September 2014; pp. 1632–1636. [Google Scholar]
  70. U.S. Geological Survey. Landsat Missions. Available online: https://www.usgs.gov/land-resources/nli/landsat (accessed on 16 September 2020).
  71. Satellite Imaging Corporation. SPOT-5 Satellite Sensor. Available online: https://www.satimagingcorp.com/satellite-sensors/other-satellite-sensors/spot-5/ (accessed on 16 September 2020).
  72. Satellite Imaging Corporation. FORMOSAT-2 Satellite Sensor. Available online: https://www.satimagingcorp.com/satellite-sensors/other-satellite-sensors/formosat-2/ (accessed on 16 September 2020).
  73. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote. Sens. 1997, 63, 691–699. [Google Scholar]
  74. Alparone, L.; Wald, L.; Chanussot, J.; Thomas, C.; Gamba, P.; Bruce, L. Comparison of pansharpening algorithms: Outcome of the 2006 GRS-S data-fusion contest. IEEE Trans. Geosci. Remote. Sens. 2007, 45, 3012–3020. [Google Scholar]
  75. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Sign. Proc. Lett. 2002, 9, 81–84. [Google Scholar]
  76. Garzelli, A.; Nencini, F. Hypercomplex quality assessment of multi/hyperspectral images. IEEE Geosci. Remote Sens. Lett. 2009, 6, 662–665. [Google Scholar]
  77. Pratt, W.K. Correlation Techniques of Image Registration. IEEE Trans. Aerosp. Electron. Syst. 1974, AES-10, 353–358. [Google Scholar]
  78. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar]
  79. Yuhas, R.H.; Goetz, A.F.H.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the Spectral Angle Mapper (SAM) algorithm. In Proceedings of the Summaries 3rd Annual JPL Airborne Geoscience Workshop, Pasadena, CA, USA, 1–5 June 1992; pp. 147–149. [Google Scholar]
  80. Wald, L. Quality of high resolution synthesized images: Is there a simple criterion? Proc. Int. Conf. Fusion Earth Data 2000, 1, 99–105. [Google Scholar]
Figure 1. Observed LANDSAT 7-ETM+ Chesapeake Bay image: (a) 1024 × 1024 multispectral (MS), (b) 2048 × 2048 panchromatic (PAN).
Figure 1. Observed LANDSAT 7-ETM+ Chesapeake Bay image: (a) 1024 × 1024 multispectral (MS), (b) 2048 × 2048 panchromatic (PAN).
Sensors 20 05308 g001
Figure 2. Observed LANDSAT 7-ETM+ Netherland images: (a) 128 × 128 MS, (b) 256 × 256 PAN, (c) 2048 × 2048 MS, (d) 4096 × 4096 PAN.
Figure 2. Observed LANDSAT 7-ETM+ Netherland images: (a) 128 × 128 MS, (b) 256 × 256 PAN, (c) 2048 × 2048 MS, (d) 4096 × 4096 PAN.
Sensors 20 05308 g002
Figure 3. Observed SPOT-5 Roma image: (a) 1024 × 1024 MS, (b) 4096 × 4096 PAN. FORMOSAT-2 Salon-de-Provence image: (c) 1024 × 1024 MS, (d) 4096 × 4096 PAN.
Figure 3. Observed SPOT-5 Roma image: (a) 1024 × 1024 MS, (b) 4096 × 4096 PAN. FORMOSAT-2 Salon-de-Provence image: (c) 1024 × 1024 MS, (d) 4096 × 4096 PAN.
Sensors 20 05308 g003
Figure 4. LANDSAT 7-ETM+ band spectral response normalized to one.
Figure 4. LANDSAT 7-ETM+ band spectral response normalized to one.
Sensors 20 05308 g004
Figure 5. (a) Ground truth 128 × 128 image from Figure 2a. The normalized maximum absolute error minus the absolute error images images for the following methods, (b) Partial Replacement Adaptive Component Substitution (PRACS), (c) modulation transfer functions (MTF)-generalized Laplacian pyramid (GLP)-context based decision (CBD), (d) Total Variation (TV), (e) 1 and (f) log.
Figure 5. (a) Ground truth 128 × 128 image from Figure 2a. The normalized maximum absolute error minus the absolute error images images for the following methods, (b) Partial Replacement Adaptive Component Substitution (PRACS), (c) modulation transfer functions (MTF)-generalized Laplacian pyramid (GLP)-context based decision (CBD), (d) Total Variation (TV), (e) 1 and (f) log.
Sensors 20 05308 g005
Figure 6. (a) Ground truth 256 × 256 image from Figure 2c. The normalized maximum absolute error images for the following methods: (b) PRACS, (c) MTF-GLP-CBD, (d) TV, (e) 1 and (f) log.
Figure 6. (a) Ground truth 256 × 256 image from Figure 2c. The normalized maximum absolute error images for the following methods: (b) PRACS, (c) MTF-GLP-CBD, (d) TV, (e) 1 and (f) log.
Sensors 20 05308 g006
Figure 7. A region of interest of the LANDSAT 7-ETM+ Chesapeake Bay image in Figure 1a. Observed images: (a) 128 × 64 MS, (b) 256 × 128 PAN. 256 × 128 pansharpened images by: (c) PRACS, (d) MTF-GLP-High Pass Modulation (HPM), (e) MTF-GLP-CBD, (f) TV, (g) 1 and (h) log methods.
Figure 7. A region of interest of the LANDSAT 7-ETM+ Chesapeake Bay image in Figure 1a. Observed images: (a) 128 × 64 MS, (b) 256 × 128 PAN. 256 × 128 pansharpened images by: (c) PRACS, (d) MTF-GLP-High Pass Modulation (HPM), (e) MTF-GLP-CBD, (f) TV, (g) 1 and (h) log methods.
Sensors 20 05308 g007
Figure 8. A region of interest of the SPOT-5 Roma image in Figure 3a. Observed images: (a) 128 × 128 MS, (b) 512 × 512 PAN. 512 × 512 pansharpened images by: (c) Brovey, (d) PRACS, (e) Additive A Trous Wavelet Transform (ATWT)-M3, (f) TV, (g) 1 and (h) log methods.
Figure 8. A region of interest of the SPOT-5 Roma image in Figure 3a. Observed images: (a) 128 × 128 MS, (b) 512 × 512 PAN. 512 × 512 pansharpened images by: (c) Brovey, (d) PRACS, (e) Additive A Trous Wavelet Transform (ATWT)-M3, (f) TV, (g) 1 and (h) log methods.
Sensors 20 05308 g008
Table 1. Some possible penalty functions.
Table 1. Some possible penalty functions.
Label ρ ( s ) ρ ( s ) / | s |
p , 0 < p 1 1 p | s | p | s | p 2
log log ( ϵ + | s | ) ( ϵ + | s | ) 1 | s | 1
Table 2. Estimated λ values for the different sensors.
Table 2. Estimated λ values for the different sensors.
SensorImageB1B2B3B4B5B6
LANDSAT 7-ETM+Figure 10.09860.10110.25760.542700
LANDSAT 7-ETM+Figure 2a0.01830.42430.05760.499800
LANDSAT 7-ETM+Figure 2c00.22830.16110.610600
SPOT-5Figure 3a00.29930.68970.0110--
FORMOSAT-2Figure 3c0.03840.556600.4051--
Table 3. Quantitative results on the LANDSAT 7-ETM+ Chesapeake Bay image. Bold values indicate the best score.
Table 3. Quantitative results on the LANDSAT 7-ETM+ Chesapeake Bay image. Bold values indicate the best score.
MethodQ8QSAMERGASSCC
EXP0.83350.83832.02235.11130.8718
PCA0.78300.80742.79375.80860.8569
IHS0.67950.67342.66407.85860.8223
Brovey0.67980.67902.16057.52260.8148
BDSD0.75990.78292.46625.92440.8539
GS0.74470.74013.30528.72580.8040
GSA0.80100.82022.40335.06880.8764
PRACS0.83630.84232.09984.86550.8774
HPF0.81370.82432.22215.12790.8834
SFIM0.80080.81842.18575.02280.8905
Indusion0.80520.81672.29065.54120.8495
ATWT0.81000.82082.34825.36420.8809
AWLP0.82670.83152.25165.38730.8734
ATWT-M20.77370.78022.60045.81330.8317
ATWT-M30.78840.79252.57205.89730.8367
MTF-GLP0.81570.82582.24185.13070.8838
MTF-GLP-HPM0.80370.82062.18785.02360.8918
MTF-GLP-CBD0.82000.82922.24995.10150.8809
TV0.84920.86062.13624.25050.9163
1 0.85950.86941.85184.09540.9220
log0.79510.78561.88394.48190.9007
Table 4. QNR Quantitative results on the LANDSAT 7-ETM+ Chesapeake Bay image. Bold values indicate the best score.
Table 4. QNR Quantitative results on the LANDSAT 7-ETM+ Chesapeake Bay image. Bold values indicate the best score.
D λ D S QNR
EXP0.00960.01660.9740
PCA0.05550.14590.8066
IHS0.10490.28880.6366
Brovey0.09630.22900.6968
BDSD0.10550.14860.7616
GS0.07160.19920.7435
GSA0.06050.11670.8298
PRACS0.01540.08230.9035
HPF0.08570.13080.7947
SFIM0.08340.11460.8116
Indusion0.05950.05250.8911
ATWT0.10440.15000.7612
AWLP0.10450.15170.7596
ATWT-M20.15300.23270.6499
ATWT-M30.12270.19750.7041
MTF-GLP0.09170.13780.7831
MTF-GLP-HPM0.08900.12150.8004
MTF-GLP-CBD0.05850.10810.8397
TV0.04250.09190.8695
l10.03380.05270.9153
log0.00900.03640.9549
Table 5. Quantitative results using Wald’s protocol on the LANDSAT 7-ETM+ Netherland image in Figure 2a. Bold values indicate the best score.
Table 5. Quantitative results using Wald’s protocol on the LANDSAT 7-ETM+ Netherland image in Figure 2a. Bold values indicate the best score.
Q8QSAMERGASSCC
EXP0.47270.88493.04457.72350.8686
PCA0.37590.76173.816512.64200.8205
IHS0.33220.74791.763310.75980.8570
Brovey0.28920.76750.000010.78090.8492
BDSD0.72050.95201.72964.67480.9777
GS0.38600.78333.348611.78500.8323
GSA0.55430.84742.508610.29900.8707
PRACS0.82300.97200.95582.90970.9878
HPF0.64200.90452.06997.43870.9458
SFIM0.59500.90431.88988.17780.9379
Indusion0.41080.84063.69639.65210.8269
ATWT0.55820.87412.68599.65070.9267
AWLP0.47150.87412.205910.10570.9195
ATWT-M20.39430.84363.78798.72890.8606
ATWT-M30.48610.86853.52747.75060.8829
MTF-GLP0.59750.89462.25448.12790.9351
MTF-GLP-HPM0.56590.89262.01339.02210.9272
MTF-GLP-CBD0.60950.90012.17267.82900.9392
TV0.47980.89063.48737.12070.8977
l10.48150.90443.37836.94220.9022
log0.49310.89202.91876.93210.8952
Table 6. Quantitative results using Wald’s protocol on the LANDSAT 7-ETM+ Netherland image in Figure 2c. Bold values indicate the best score.
Table 6. Quantitative results using Wald’s protocol on the LANDSAT 7-ETM+ Netherland image in Figure 2c. Bold values indicate the best score.
Q8QSAMERGASSCC
EXP0.78740.78673.33625.92790.8521
PCA0.61670.48545.707812.81270.5788
IHS0.53430.37784.741712.95260.5918
Brovey0.54650.40633.460513.08760.5960
BDSD0.75750.77333.93486.86100.7856
GS0.58990.43696.115813.92750.5522
GSA0.73230.74444.09207.20780.7535
PRACS0.78220.78293.47876.32440.8037
HPF0.71670.72413.81357.32660.7603
SFIM0.69080.71814.66939.02620.7261
Indusion0.72300.73463.82297.21770.7631
ATWT0.69480.69714.04028.03060.7240
AWLP0.71100.70663.81248.17330.7165
ATWT-M20.63320.62884.52238.55420.6135
ATWT-M30.69930.69674.30287.44200.7201
MTF-GLP0.71160.71723.86327.51340.7457
MTF-GLP-HPM0.68750.71334.67809.11100.7161
MTF-GLP-CBD0.75190.75953.75226.86980.7740
TV0.78430.80653.84025.73510.8519
l10.81180.81963.13375.18310.8853
log0.77500.76823.08105.31150.8818
Table 7. QNR quantitative results on the LANDSAT 7-ETM+ Netherland image in Figure 2a. Bold values indicate the best score.
Table 7. QNR quantitative results on the LANDSAT 7-ETM+ Netherland image in Figure 2a. Bold values indicate the best score.
D λ D S QNR
EXP0.01040.05930.9309
PCA0.24630.39980.4523
IHS0.26320.40350.4394
Brovey0.21820.38730.4790
BDSD0.01590.05050.9344
GS0.23350.40670.4548
GSA0.21390.32400.5314
PRACS0.06650.21060.7369
HPF0.17110.26380.6102
SFIM0.16100.25130.6282
Indusion0.13540.16120.7252
ATWT0.19680.29610.5654
AWLP0.19770.29540.5653
ATWT-M20.17270.31270.5686
ATWT-M30.11870.21430.6924
MTF-GLP0.17960.27910.5915
MTF-GLP-HPM0.16880.26610.6100
MTF-GLP-CBD0.17410.27780.5965
TV0.09120.11270.8064
l10.03420.03860.9286
log0.01570.06270.9225
Table 8. QNR quantitative results on the LANDSAT 7-ETM+ Netherland image in Figure 2c. Bold values indicate the best score.
Table 8. QNR quantitative results on the LANDSAT 7-ETM+ Netherland image in Figure 2c. Bold values indicate the best score.
D λ D S QNR
EXP0.00670.01570.9778
PCA0.18930.50190.4038
IHS0.20400.59730.3205
Brovey0.19790.52270.3829
BDSD0.01340.01070.9761
GS0.19990.52370.3811
GSA0.09860.23370.6907
PRACS0.02550.11760.8599
HPF0.15940.27720.6075
SFIM0.11240.22720.6859
Indusion0.13520.20540.6871
ATWT0.18220.32340.5534
AWLP0.17650.31010.5682
ATWT-M20.19030.34670.5290
ATWT-M30.09170.15810.7647
MTF-GLP0.16560.29330.5897
MTF-GLP-HPM0.11640.24240.6694
MTF-GLP-CBD0.08540.19530.7359
TV0.03810.11860.8478
l10.02280.07040.9084
log0.00730.01990.9730
Table 9. Quantitative results using Wald’s protocol on the SPOT-5 Roma image. Bold values indicate the best score.
Table 9. Quantitative results using Wald’s protocol on the SPOT-5 Roma image. Bold values indicate the best score.
p / P 416
MethodQ4QSAMERGASSCCQ4QSAMERGASSCC
EXP0.87660.88591.70483.78570.86400.73250.74072.50712.84410.6049
PCA0.40670.53605.164612.33460.27880.39270.50915.72086.29110.2443
IHS0.40720.52383.995112.23420.27720.39730.50514.45086.17700.2520
Brovey0.41240.53371.841312.19600.25940.40190.51242.40006.17180.2482
BDSD0.85590.88252.05654.27760.82350.59470.62313.03284.30500.2804
GS0.41020.53645.052312.12720.26340.39850.51245.58496.14710.2405
GSA0.48970.53842.989311.03630.19320.49970.53543.11895.37030.2164
PRACS0.82200.83802.05364.89360.72980.72910.74582.51902.96470.5295
HPF0.74880.76952.06326.33880.62500.58880.61242.83704.59690.2988
SFIM0.77440.78601.96586.06940.64470.60520.62322.67924.44250.3079
Indusion0.74730.78942.16095.97170.67610.53010.59353.56894.96560.3408
ATWT0.69280.71712.23587.64330.51830.58490.60922.89324.72450.3099
AWLP0.70660.72602.17987.62460.50750.59470.61732.85044.69630.3052
ATWT-M20.72290.73432.30846.31450.46780.67230.68222.72913.32770.3965
ATWT-M30.78370.79302.23945.19350.67140.69240.70192.70263.10940.4801
MTF-GLP0.72890.75072.12016.79750.57660.57750.60232.92754.83550.3022
MTF-GLP-HPM0.75530.76752.00056.49230.59720.59280.61112.73884.70660.3073
MTF-GLP-CBD0.77180.78702.01886.04900.62150.60210.62172.77674.56320.3118
TV0.74720.78933.18826.13930.61620.64800.68723.51093.97630.3265
1 0.86170.87832.06884.15570.81960.64090.70173.77933.71170.3792
log0.86360.87621.60533.36730.89230.73230.73952.42282.70720.6262
Table 10. Quantitative results using Wald’s protocol on the FORMOSAT-2 Salon-de-Provence image. Bold values indicate the best score.
Table 10. Quantitative results using Wald’s protocol on the FORMOSAT-2 Salon-de-Provence image. Bold values indicate the best score.
p / P 416
MethodQ4QSAMERGASSCCQ4QSAMERGASSCC
EXP0.86100.86171.81583.72570.84600.69180.69252.50572.69340.5906
PCA0.78580.80892.20985.07110.65950.74150.76682.63732.74850.5870
IHS0.79120.81591.96374.62930.66950.74450.77202.45272.58030.5894
Brovey0.78610.81461.91064.53320.66440.73850.76872.43452.55970.5793
BDSD0.84430.85452.02924.26080.72320.77010.78232.54602.72170.5821
GS0.79790.81952.13754.91930.66190.75060.77532.57612.68600.5888
GSA0.78590.81672.18555.25080.65080.76070.77722.55102.76980.5796
PRACS0.84950.85921.95814.08420.73730.79470.80162.42402.36830.6230
HPF0.84690.85231.97994.22100.77820.78130.78852.46092.53880.5900
SFIM0.84700.85231.98214.27180.77550.78170.78882.46002.56580.5874
Indusion0.83040.83712.00754.43610.74070.75190.76982.49862.69410.5758
ATWT0.83910.84492.03744.56750.76630.79010.79652.46792.53990.6161
AWLP0.84080.84731.88964.35490.76710.79090.79792.36892.45400.6150
ATWT-M20.82770.83262.23074.21640.69750.76720.77042.54992.37090.6325
ATWT-M30.83250.83562.21374.05860.73410.76390.76602.56592.38780.6384
MTF-GLP0.84770.85331.98834.27050.77040.79010.79682.47142.55630.6176
MTF-GLP-HPM0.84760.85321.98894.33130.76760.79040.79712.46662.59370.6158
MTF-GLP-CBD0.84930.85482.00764.25750.77300.78460.79162.52702.63940.6123
TV0.86960.87902.14373.76020.78760.78070.78782.79062.43740.5956
1 0.89460.89741.92003.35260.84570.76910.76253.07262.60410.5503
log0.87060.86831.75973.36860.87510.68890.68482.49752.60090.6050
Table 11. QNR Quantitative results on the SPOT-5 Roma image. Bold values indicate the best score.
Table 11. QNR Quantitative results on the SPOT-5 Roma image. Bold values indicate the best score.
p / P 416
D λ D S QNR D λ D S QNR
EXP0.00410.01500.98090.00010.03120.9687
PCA0.20470.40940.46970.30940.50350.3429
IHS0.23890.41580.44470.35740.51430.3121
Brovey0.18040.37990.50820.28900.47540.3730
BDSD0.01080.09220.89800.03880.03440.9281
GS0.19640.41000.47410.30450.50440.3447
GSA0.21940.34210.51350.32670.42870.3846
PRACS0.03250.15550.81710.06560.21620.7324
HPF0.08510.14050.78640.21490.25560.5844
SFIM0.06610.12560.81670.19490.24160.6107
Indusion0.05800.03700.90720.24580.15870.6345
ATWT0.13100.21190.68490.23980.30370.5293
AWLP0.10300.19500.72210.20150.28140.5738
ATWT-M20.07280.20020.74160.09960.16910.7482
ATWT-M30.01620.03490.94940.04930.03280.9195
MTF-GLP0.10400.15860.75390.25110.30000.5242
MTF-GLP-HPM0.08580.14410.78250.23140.28680.5481
MTF-GLP-CBD0.06570.12720.81540.19220.26810.5912
TV0.33990.18300.53940.18660.15100.6906
l10.02770.03780.93560.19270.15000.6862
log0.00560.02720.96740.04220.03800.9214
Table 12. QNR quantitative results on the FORMOSAT-2 Salon-de-Provence image. Bold values indicate the best score.
Table 12. QNR quantitative results on the FORMOSAT-2 Salon-de-Provence image. Bold values indicate the best score.
p / P 416
D λ D S QNR D λ D S QNR
EXP0.00870.08370.90830.00860.09310.8990
PCA0.11080.21900.69450.15030.30830.5877
IHS0.10830.21270.70200.15080.30270.5921
Brovey0.08590.19640.73450.12150.28150.6312
BDSD0.01790.01500.96730.02640.17750.8008
GS0.09700.20840.71490.13660.29710.6069
GSA0.12540.20950.69140.16180.29630.5899
PRACS0.06210.16560.78260.08780.24130.6921
HPF0.08960.11300.80750.11470.15590.7473
SFIM0.08770.11210.81010.11280.15560.7492
Indusion0.05100.01740.93250.09380.09650.8187
ATWT0.12270.15290.74320.13550.19510.6958
AWLP0.12900.15000.74040.13930.18870.6983
ATWT-M20.11840.15500.74490.09670.12380.7915
ATWT-M30.08710.09510.82610.05860.03880.9049
MTF-GLP0.10040.12770.78470.14370.19960.6854
MTF-GLP-HPM0.09800.12690.78750.14100.19890.6881
MTF-GLP-CBD0.09100.12220.79790.13170.19190.7017
TV0.11140.07870.81860.09510.18110.7410
l10.04250.05140.90830.05940.08470.8609
log0.00940.09170.89980.01740.10110.8832
Table 13. Elapsed CPU time in seconds for the different pansharpening methods on a 1024 × 1024 image and with different p / P ratios.
Table 13. Elapsed CPU time in seconds for the different pansharpening methods on a 1024 × 1024 image and with different p / P ratios.
PpPCAIHSBroveyBDSDGS
512 × 512 1024 × 1024 0.90.040.040.90.4
256 × 256 1024 × 1024 0.30.030.030.80.3
PpGSAPRACSHPFSFIMIndusion
512 × 512 1024 × 1024 12.20.20.50.5
256 × 256 1024 × 1024 1.61.20.20.180.3
PpATWTAWLPATWT-M2ATWT-M3MTF-GLP
512 × 512 1024 × 1024 111410100.9
256 × 256 1024 × 1024 33770.6
PpMTF-HPMMTF-CBDTV1log
512 × 512 1024 × 1024 0.821.5 10 3 8081.1 10 4
256 × 256 1024 × 1024 0.60.61.6 10 3 3 10 3 2.8 10 4

Share and Cite

MDPI and ACS Style

Pérez-Bueno, F.; Vega, M.; Mateos, J.; Molina, R.; Katsaggelos, A.K. Variational Bayesian Pansharpening with Super-Gaussian Sparse Image Priors. Sensors 2020, 20, 5308. https://doi.org/10.3390/s20185308

AMA Style

Pérez-Bueno F, Vega M, Mateos J, Molina R, Katsaggelos AK. Variational Bayesian Pansharpening with Super-Gaussian Sparse Image Priors. Sensors. 2020; 20(18):5308. https://doi.org/10.3390/s20185308

Chicago/Turabian Style

Pérez-Bueno, Fernando, Miguel Vega, Javier Mateos, Rafael Molina, and Aggelos K. Katsaggelos. 2020. "Variational Bayesian Pansharpening with Super-Gaussian Sparse Image Priors" Sensors 20, no. 18: 5308. https://doi.org/10.3390/s20185308

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop