Next Article in Journal
Utilizing Visible Band Vegetation Indices from Unmanned Aerial Vehicle Images for Maize Phenotyping
Next Article in Special Issue
Spatiotemporal Prediction of Conflict Fatality Risk Using Convolutional Neural Networks and Satellite Imagery
Previous Article in Journal
Production of Annual Nighttime Light Based on De-Difference Smoothing Algorithm
Previous Article in Special Issue
A Review of GAN-Based Super-Resolution Reconstruction for Optical Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Unsupervised CNN-Based Pansharpening Framework with Spectral-Spatial Fidelity Balance

by
Matteo Ciotola
1,*,
Giuseppe Guarino
1 and
Giuseppe Scarpa
2
1
Department of Electrical Engineering and Information Technology (DIETI), University Federico II, Via Claudio 21, 80125 Naples, Italy
2
Department of Engineering, University Parthenope, Centro Direzionale ISOLA C4, 80133 Naples, Italy
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(16), 3014; https://doi.org/10.3390/rs16163014
Submission received: 7 June 2024 / Revised: 12 August 2024 / Accepted: 14 August 2024 / Published: 16 August 2024
(This article belongs to the Special Issue Weakly Supervised Deep Learning in Exploiting Remote Sensing Big Data)

Abstract

:
In recent years, deep learning techniques for pansharpening multiresolution images have gained increasing interest. Due to the lack of ground truth data, most deep learning solutions rely on synthetic reduced-resolution data for supervised training. This approach has limitations due to the statistical mismatch between real full-resolution and synthetic reduced-resolution data, which affects the models’ generalization capacity. Consequently, there has been a shift towards unsupervised learning frameworks for pansharpening deep learning-based techniques. Unsupervised schemes require defining sophisticated loss functions with at least two components: one for spectral quality, ensuring consistency between the pansharpened image and the input multispectral component, and another for spatial quality, ensuring consistency between the output and the panchromatic input. Despite promising results, there has been limited investigation into the interaction and balance of these loss terms to ensure stability and accuracy. This work explores how unsupervised spatial and spectral consistency losses can be reliably combined preserving the outocome quality. By examining these interactions, we propose a general rule for balancing the two loss components to enhance the stability and performance of unsupervised pansharpening models. Experiments on three state-of-the-art algorithms using WorldView-3 images demonstrate that methods trained with the proposed framework achieve good performance in terms of visual quality and numerical indexes.

1. Introduction

Due to technological and physical constraints, many Earth observation systems, such as GeoEye, Plaiades or WorldView, acquire a multispectral (MS) image with high spectral resolution and a single-band panchromatic (PAN) image with high spatial resolution. A multi-resolution fusion process, called pansharpening, is then employed to estimate a full resolution MS image from the original PAN and MS components.
Due to the unavailability of the MS image at full resolution as a ground truth reference, pansharpening remains a challenging, ill-posed problem that is far from being resolved. Consequently, several approaches have been proposed over the years, which can be roughly collected into four main categories [1]: Component Substitution (CS) [2,3,4], Multi-Resolution Analysis (MRA) [5,6,7], variational optimization (VO) [8,9], and machine/deep learning (ML/DL) [10,11,12].
Both CS and MRA methods incorporate the PAN component into the resized (by interpolation) MS image, though they employ different techniques. The CS methods execute the injection in a transform domain, such as principal component analysis (PCA) [13], Gram-Schmidt (GS) projection [14], Brovey [15], or generalized Intensity-Hue-Saturation (G-IHS) transform [16], where the “strongest” component is substituted with a suitably equalized version of the PAN image. The MRA methods, on the other hand, work on “detail” (high-frequency) components, requiring multi-resolution decomposition transforms like Wavelets [17,18] or Laplacian pyramids [19,20]. In contrast, VO methods depend on the optimization of suitable acquisition or representation models, such as resolution downgrading models [21], sparse representations [22], total variation [8], and low-rank matrix representations [23]. In recent years, deep learning methodologies have garnered increasing attention for their remarkable efficacy in addressing a multitude of intricate computer vision and image processing challenges, fostering heightened expectations across diverse application domains, including remote sensing [24,25,26,27,28,29,30,31].
The advent of the first pansharpening convolutional neural network (PNN) by Masi et al. [32] marked a significant milestone in this trajectory. After this seminal work, numerous alternative approaches have been proposed [33,34,35,36,37,38,39,40,41,42], among others, in attempts to progressively refine the resolution enhancement capabilities of pansharpening methodologies. Nowadays, deep learning is the most popular approach for pansharpening. In fact, due to the importance it holds at this moment, many pansharpening toolboxes have included algorithms based on deep learning in their comparative analysis [1,43].
A significant obstacle in leveraging deep learning for pansharpening is the lack of ground-truth data, which hinders the use of simple and effective supervised training procedures. Early proposals addressed this issue by resorting to supervised training in the reduced resolution domain, following the pansharpening synthesis assessment protocol proposed by Wald [44]. With this approach, the network is trained using downsampled versions of the PAN and MS components, with the original MS image serving as the ground truth. The underlying assumption is that methods optimized in the reduced resolution space will perform equally well on full-resolution images. Nevertheless, there is an intrinsic limitation in this hypothesis, stemming from a statistical mismatch between synthesized images and real ones due to their resolution gap. In real world images there are, indeed, several fine-grained patterns that completely disappear during the downsampling process needed for traning deep learning methods in a supervised manner. For example, car shapes, vegetation, toproofs or ground textures, such as horizontal traffic signs, parking lines, are clearly visible at PAN scale resolution but already difficult to distinguish in the original MS. Eventually, these patterns can not be learned by the network using synthetic training samples, reflecting a lack of definition on the full resolution images [45,46,47]. Experimental evidence has demonstrated that this assumption does not hold: good performances at low resolution hardly rise to full resolution, and low-quality images are often obtained [46]. This limitation has recently motivated several studies aimed at circumventing the resolution downgrade. They resort to training mechanisms which do not require any ground truth during the training phase.
More specifically, some of these methods exploit adversarial training procedures, which range from the use of two distinct discriminators for separately checking the spectral and spatial consistency of the outcomes [48,49,50], or cycleGANs [51]. While these techniques have demonstrated the capacity to produce fused images of superior quality, they are plagued by challenges such as unstable training, lack of interpretability, and mode collapse, significantly inhibiting the results’ overall quality. Furthermore, comprehending the internal mechanisms and behaviours of Generative Adversarial Networks (GANs) is not easy, thereby complicating the delicate control of injected details.
Some other methods [45,47,52,53,54,55] rely on loss functions that integrate metrics utilized for the no-reference evaluation of pansharpening outputs or are otherwise related to them. This approach allows for the direct adjustment of the delicate balance between spectral fidelity and the injection of structural details during the training process. Regardless of the underlying philosophy (adversarial or metrics-based training procedure), the unsupervised protocol complicates the training process of deep learning-based pansharpening methods. Whereas with the supervised mechanism there is a single loss that controls both the spectral and spatial content, through the ground-truth, in the unsupervised case there is a need to independently compare the two qualities with the input images. Thus, although during the training phase one can experience a convergent trend of the overall loss, one or more terms could stabilize on a non-optimal value or, in the worst case, diverge. This creates the need to separate the spectral quality problem from the spatial one as much as possible. An incorrect balance between these components can lead to less than satisfactory results. For example, in [45,46] the images sometimes lack of brightness. In [53] the results have some colour aberrations, particularly visible on the rooftops of some buildings. In [56], instead, the outcomes show blurred effects. Finally, in [47] some images present oversharpening. This is likely due to a not perfect balance between the two loss components, which exacerbates their interdependence, generating instability and/or degenerate results.
Motivated by these considerations, we propose a new framework for training unsupervised deep learning-based pansharpening methods that avoid instabilities between the two loss functions. We start from the full-resolution training framework proposed in [46], applying a modified version of the loss function proposed in [47]. Carrying out several experiments varying the balance between the two loss components, we evaluate to what extent the combination of the two losses can be reliable without impairing the global quality of the outcomes, also analyzing their impact on the pansharpened products. Then, we define a general criterion for correctly balancing the unsupervised loss terms to prevent unstable behaviour. Experiments with this balanced training framework on three state-of-the-art CNN-based pansharpening models of diverse sizes, performed on images acquired by the WorldView-3 satellite, demonstrate its efficacy and broad applicability.
In the rest of the manuscript, we account for deepening the problem, demonstrating how unsupervised loss functions for deep learning-based pansharpening methods can be expressed as a combination of metrics capable of checking spectral and spatial consistencies (Section 2), we provide experimental results (Section 3) and a related discussion (Section 4), before drawing conclusions (Section 5).

2. Materials and Methods

2.1. Unsupervised Losses for Pansharpening

The invalidity of the scale-invariance assumption has driven researchers towards innovative unsupervised approaches. In the absence of ground truth (GT), the primary challenge becomes defining an appropriate loss function that provides meaningful guidance for network optimization. Unfortunately, it is impractical to use a single loss that effectively controls both spectral and spatial image characteristics. To address this, all unsupervised deep learning-based methods define the loss as a weighted sum (or product) of spectral ( L λ ) and structural ( L S ) terms, balanced by hyperparameters α and β , respectively, along with auxiliary terms to achieve secondary objectives:
L = α L λ + β L S + auxiliary terms
In the supervised case, where a single loss (or multiple highly correlated losses) is used, an incorrect learning rate ( η ) results in total instability of the training mechanism, causing the loss to diverge. In the unsupervised case, two or more ideally independent losses are used (one controlling low frequencies and the other high frequencies) and can be optimized separately. However, this hypothesis is not valid in practice: these losses are somehow linked, and as demonstrated in Section 3 and Section 4, exhibiting antagonistic behavior. A learning rate that is too high can cause not only the divergence of the overall loss but also the failure to optimize one of the components, leading to negative effects such as color variations or total absence of details in extreme cases (as shown in Section 3). Therefore, the hyperparameters α , β are closely tied to the η parameter used during both the training phase, for initial weight production, and the target adaptation phase, for adapting the weights to the target image and producing the pansharpened image. Thus, even though α , β , η values can be found in a three-dimensional space, the degrees of freedom remain two. Therefore, the learning rate will be fixed for all methods mentioned and the experimental analysis in Section 3.
Starting from these assumptions, the main aim of this work is to identify an approach to optimize both components as much as possible, ensuring a trend of all loss curves that is as regular as possible, with a final value that balances spectral and spatial fidelity.
Table 1 lists the main symbols used in the rest of the manuscript, while Table 2 summarizes some of the most commonly used unsupervised spectral and spatial loss functions for deep learning-based pansharpening and their respective balancing parameters. Analyzing Table 2, it is evident that there is no consensus on the loss function to be used. This is due to the novelty of these methods, allowing room for improvement, and the different approaches to various challenges. Compared to well-established supervised deep learning-based pansharpening algorithms, this new category of methods has greater computational complexity [55], sensitivity to local and global misalignment [47], different overlapping wavelengths between PAN and MS components [48], which can cause spatial patterns to appear only in PAN or some MS bands, and competitive behaviors between spectral and spatial components. This last issue has been tackled in different ways by the methods summarized in Table 2, using different loss functions with various balancing hyperparameters.
More in detail, Luo et al. [45] employ a combination of Mean Square Error (MSE) and Structure SIMilarity (SSIM) for both spatial and spectral consistency terms (with unitary α and β ), together with the Quality with No Reference (QNR) metric [58], re-defined as a loss:
L QNR = 1 QNR = 1 1 D λ Q ( M ^ , M ) a · 1 D S Q ( M ^ , M , P ) b
where the spectral distortion index D λ Q is defined as:
D λ Q = 1 B ( B 1 ) i = 1 B j = 1 , j j B Q ( M i , M j ) Q ( M ^ i , M ^ j ) p p ,
and the structural distortion index D S Q is formulated as:
D S Q = 1 B i = 1 B Q ( M ^ i , P ) Q ( M i , P ) q ,
in which Q ( · , · ) represents the Universal Image Quality Index (UIQI) [59]. a and b are combination weights to ensure a correct balance between the spectral and spatial consistency terms. Usually both these weights are set to 1.
Similarly, Zhou et al. [49] employ L QNR within an adversarial training framework, adjusting the loss balance to α = 0.0002 and β = 0.0001 for the adversarial losses.
The adversarial losses also characterize the training framework of PanGAN [48]. The L a d v is combined with an Euclidean distance for both spectral and spatial losses, with α = 1 and β = 5 . To better preserve the high-frequency details, Ma et al. propose computing the spatial loss on the gradients of the images.
This idea, along with the 2 -norm for the spectral loss functions, is maintained in [56]. However, for the spatial consistency check, the 1 -norm is used. These two components are multiplied, respectively, with α = 0.1 and β = 1.0 .
In contrast, [57] utilize a combination of Spectral Angle Mapper (SAM) [60] and MSE, and a combination of UIQI and MSE, for the spectral and structural consistency with weights of α = 0.79 and β = 0.20 , respectively.
Z-PNN [46] and its faster version [55] employ an 1 -norm loss for assessing the spectral consistency and a modified version of the local correlation coefficient ( D ρ ) [61] for the evaluation of the spatial accuracy. In this case, the two loss terms are weighted with α = 1 and β = 0.36 .
Finally, λ -PNN [47] further refines the loss proposed in Z-PNN by modifying its spectral component. In this case, L λ is defined as:
L λ = α 1 · ERGAS M ^ , M + D λ M ^ , M ,
where ERGAS is defined as:
ERGAS = 100 R 1 B b = 1 B RMSE b μ b GT 2 ,
and D λ is the Khan’s consistency spectral index [62]:
D λ = 1 Q 2 n M ^ , M .
Q 2 n [63] is the multiband extension of UIQI. Each pixel of an image with B spectral bands can be represented as a hypercomplex number with one real part and B 1 imaginary parts. Let z and z ^ be the hypercomplex representations of a generic ground truth pixel and its prediction, respectively, then Q 2 n can be expressed as the product of three terms:
Q 2 n = E | σ z , z ^ | σ z σ z ^ · 2 σ z σ z ^ σ z 2 + σ z ^ 2 · 2 μ z μ z ^ | μ z | 2 + | μ z ^ | 2 ,
in which the first factor measures correlation the second assesses contrast changes and the third evaluates the mean bias between z and z ^ .
Another pivotal element contributing to the effectiveness of pansharpening methods, as demonstrated in [46,47,55], is the target-adaptive operating modality. This approach was initially introduced in the context of supervised pansharpening by [36] and subsequently adapted to the unsupervised training framework in [46].
Remote sensing images exhibit substantial variability due to differences in the depicted scenes, sensor characteristics, and weather/light acquisition conditions. Often, the training sets used in pansharpening comprise a limited number of images, typically acquired under similar conditions and with the same sensor. Consequently, models trained under these constraints may perform suboptimally when applied to new, off-training data.
The target-adaptive modality addresses this limitation by unfreezing the network pre-trained weights and performing a few cycles of fine-tuning on the target image. This process utilizes a subset of data extracted from the target image itself, allowing the model to adapt to the specific characteristics of the new data. This protocol not only facilitates the accomplishment of high-quality results but also provides valuable insights into the loss properties, including stability and overall performance in terms of the quality of the results.

2.2. Proposed Framework with Balanced Spectral-Spatial Loss

The aim of this work is first to experimentally demonstrate the contrastive dependence between the spectral and spatial loss components. This antagonistic nature can be mitigated by using pairs of properly selected α , β hyperparameters.
The general rule proposed, and the main contribution of this work, is to preserve as much of the informative content of the spectral bands as possible by using a pair of values that minimize spectral degradation while maximizing detail gain. This can be achieved by analyzing the hyperparameter space to find the balance that results in a convergent trend for both losses, identifying the pair that best preserves the final spectral loss value (compared with the most convergent results obtained by other balances), while effectively minimizing the spatial loss.
As noted in the following sections, these values are not very sensitive to the CNN employed (allowing for a faster search using a scaled network) and are more linked to the sensors’ technology.
To demonstrate the strong generalization capability of this approach, we will apply the same analysis on three different architectures trained under the same conditions, with varying depths (Z-PNN [46], PanNet [11], and BDPN [37]), obtaining very competitive results compared to most state-of-the-art methods.
Z-PNN [46] is an advanced version of the PNN [32], which originally comprises only three convolutional layers. Z-PNN introduces a skip connection, altering the original structure. This network layout results in 104,36 trainable parameters with a number of floating point operations (FLOPs) required for a single inference of 0.29 GFLOPs. The detailed structure of the network is shown in Figure 1.
PanNet [11] is a medium-depth network with ten convolutional layers. This architecture leverages high-frequency information from M and P to maintain separated spectral and spatial information. Specifically, the deep residual network includes a preprocessing convolutional layer to increase the number of features, followed by four residual blocks for enhanced feature extraction, and a single convolutional layer for postprocessing and matching the spectral dimensions. This network design provides 78,504 trainable parameters and 0.32 GFLOPs and is illustrated in Figure 2.
The Bi-Directional Pansharpening Network (BDPN) [37] is a two-stream network for pansharpening. It employs a bidirectional pyramid structure to process M and P separately, in line with the principles of Multi-Resolution Analysis. Multilevel details are extracted from P and injected into the M image to reconstruct M ^ . A flowchart of the BDPN architecture is provided in Figure 3. This network, characterized by numerous multiscale convolutional layers, is highly deep, with 1,484,412 trainable parameters and 3.80 GFLOPs.
The proposed analysis utilizes a modified version of the loss function introduced in [55]. This is a multitask loss as described in Equation (1), devoid of any auxiliary terms. Thus, we first perform pansharpening on real full-resolution samples. Then, we check the consistency with the two input components. To check the spectral consistency with the MS input, we perform a spatial degradation using a Gaussian-shaped low-pass filter that matches the Modulation Transfer Function (MTF) [2] to preserve contrast levels between scales, followed by decimation. To assess spatial consistency, we compare the output with the PAN. Specifically, the spectral loss term L λ comprises two components: one based on a pixel-wise error measure (ERGAS), and the other on first- and second-order statistics ( D λ ). Experimental observations indicate that these two components are somewhat correlated. Indeed, we have experimentally observed that ERGAS not only plays a more significant role in ensuring accuracy but also exhibits greater conflict with the spatial loss term L S . Based on these considerations, and to simplify the subsequent analysis, we decided to omit the second spectral component D λ , resulting in:
L λ = ERGAS M ^ , M .
Conversely, the spatial loss term remains unaltered.
In the proposed validation framework, we will also employ the target-adaptive protocol [46,64]. This approach will enable us to examine the network’s response under the extreme condition of overfitting on the target image. Utilizing this protocol, we will explore the α β hyperparameter space, analyzing the network and loss function both quantitatively, separately evaluating spatial and spectral accuracy, and qualitatively through visual inspection.
We have tested our hypothesis on images acquired by the WorldView-3 satellite. To our knowledge, this satellite is currently one of the most advanced, acquiring MS with 8 bands at 1.24 m resolution and a PAN image at 0.31 m resolution. Due to the high resolution at which this satellite operates and technological constraints, such as different sensor positions and slight time delays during acquisitions, the images provided are a challenging test for modern pansharpening methods. To address the misalignment problem specific to this satellite, we use a coregistration-at-loss paradigm, as proposed in [47]. The combined use of these tools (MTF low-pass filtering, coregistration-at-loss, target adaptation) allows us to obtain results that can be easily generalized to other sensors, making this analysis adaptable to satellites with different architectures and technologies.

3. Results

In this Section, after analyzing an experiment that demonstrate the antagonistic behaviour of the spectral and spatial loss components, giving us the opportunity to empirically define a modality with which set the hyperparameters α and β , we carry out the comparative perforance assessent of the proposed framework, studying visual and numerical results.

3.1. Experimental Setup

To obtain a reliable and generalizable assessment, we conducted experiments on a large and varying dataset. We used several images acquired by the WorldView-3 satellite (see Table 3), with three images exclusively for training and validation. We utilized one image for validating the “cross-scenario” generalization capability of our solution and also for testing. This image, with different weather and sunlight conditions, place, has never been used for training. Finally, the last image was exclusively used for testing. As there is no connection between this last test image and the training process, this experiment provides the most meaningful information about the effectiveness of our proposal. The given dataset volume and partition provide a good proxy for the quality assessment and, partially, also for the generalization capacity. Since our solution is based on a training and target-adaptation procedure, the only pre-processing step of our dataset is patch generation: training and validation are carried out on 512 × 512 crops at PAN scale, while testing is performed on images of 2048 × 2048 .
To evalaute the performances of methods, trained and fine-tuned with our proposed framework, we compared them with several state-of-the-art methods. We rely on two benchmark toolboxes [1,43], which contain algorithms belonging to all four main categories recalled in Section 1: CS, MRA, VO and ML. We have exploited most of the methods provided in this collections, except few VO solutions that suffer from library incompatibilities. In addition to these, we selected five state-of-the-art unsupervised deep learning-based algorithm, retrained on our datasets to ensure a fair comparison. A summary of all comparative methods is provided in Table 4.
It is worth underlining that assessing the performance of pansharpening algorithms is still an open problem due to the lack of ground truth. We consider full-resolution no-reference-based indexes, which separately assess spectral and spatial fidelity. Specifically, we consider the spectral distortion index ( D λ ), proposed by Khan et al. [62], along with the reprojection indexes, R-SAM and R-ERGAS, proposed in [61] for spectral quality assessment, and the correlation distortion index, D ρ , for spatial quality assessment.
In our experiments, we used a NVIDIA DGX Station A100, with an AMD EPYC 7742 64-Core Processor CPU, 512 GB RAM, and four NVIDIA A100-SXM4 GPUs with 40 GB of graphic memory each, produced by NVIDIA Corporation, Santa Clara (CA). USA. To assess the impact of different weightings, we pre-trained the three models using the Adam optimizer with a learning rate of η = 2 · 10 4 for 50 epochs.

3.2. Finding the Correct α / β Balance

We optimized the parameters α , β , beginning with an initial coarse search across a broad hyperparameter space, followed by a more precise refinement within the range α , β 10 3 , 10 1 × 10 2 , 10 2 . Utilizing a logarithmic scale, we sampled 500 configurations to balance the spectral and spatial loss components during target adaptation on the validation set over 200 epochs. For each sampled configuration, we recorded, alongside the final pansharpening results, the values of the loss terms at the end of the tuning iterations. Additionally, we logged the loss values from the pre-trained model, denoted as L λ ( 0 ) and L S ( 0 ) , to serve as a reference for evaluating the benefits of tuning for each component. Figure 4 and Figure 5 present the L λ and L S values achieved during fine-tuning for the sampled configurations ( α , β ) , for Z-PNN (left), PanNet (middle), and BDPN (right), for one of the Adelaide validation images. Similar results have been obtained with different validation images, which are not presented for the sake of brevity.
In this phase, we focused on generating comprehensive scatter plots to visualize the impact of various α , β configurations on spectral and spatial loss components. Square (Figure 4) and circular (Figure 5) markers represent the respective loss components, with marker size corresponding to the loss values. A quantile-based color-coded clustering facilitates easier visual inspection of the scatter plots. In all scatter plots, triangular markers denote unstable configurations, attributed to spectral loss in Figure 4 spatial loss in Figure 5. Unstable configurations are classified as weakly divergent ( L * ( 0 ) < L * < 10 L * ( 0 ) ) and strongly divergent ( L * > 10 L * ( 0 ) ). Two red lines delineate an ideal domain where both losses exhibit convergent behavior.

3.3. Comparative Analysis

By examining the scatter plots, we observe a transition region for all three models where the black squares change to blue ones (i.e., where the spectral loss value is close to q 1 ) in Figure 4 and where the dots change from blue to black, in Figure 5. This convergence basin defines the values of α and β that optimize performance both spectrally and spatially. To maintain symmetry, we select the pair α , β = 0.03 , 0.3 from this region as the optimal value, as it is roughly at the center of our plots. Since we are interested not only in finding the optimal value but also in studying the effects of various loss components on the final results, we have uniformly sampled 9 β values from the horizontal line α = 0.03 , comprising the optimal balance value chosen. These configurations were uniformly chosen for the Z-PNN, PanNet, and BDPN networks to establish a basis for testing proposed solutions. We increased the number of tuning iterations to 500 for these cases to achieve more precise results. With these models, we first analyzed the effects of wrong balance visually inspecting the results.
Sample results from the Adelaide test dataset for these selected ( α , β ) configurations are displayed in Figure 6, Figure 7 and Figure 8. Results from the PairMax dataset, which yielded similar outcomes, are omitted for brevity.
Thus, we also deepen the correlation between the losses and the metrics used for the evaluation. We quantitatively evaluated our sampled weights using established spectral consistency indexes D λ [62], R-ERGAS and R-SAM and the spatial consistency index D ρ [61]. Given that the three spectral indexes provided comparable indications, we present only D λ for brevity. The experiments have been conducted on both the test sets and the values obtained are visually plotted in Figure 9, Figure 10 and Figure 11 for Z-PNN, PanNet and BDPN, respectively. The plot on the left refers to the values obtained on the three Adelaide crops, averaged for the fixed ( α , β ) . while on the right are displayed the averaged values of spectral consistency index D λ and the structural consistency index D ρ obtained on the three PairMax images.
Finally, numerical results are summarized in Table 5, comparing our “Balanced” solutions with other state-of-the-art methods for pansharpening on both testing datasets.

4. Discussion

To evaluate the impact of different weightings on the spectral and spatial loss components, we focus on Figure 4 and Figure 5, which respectively display the L λ and L S values achieved during fine-tuning by the three CNNs for the sampled ( α , β ) configurations.
In this validation phase, we begin by examining the results for Z-PNN, as shown in the scatter plots on the left in both figures. As anticipated, when β is smaller relative to α , the spectral loss decreases (Figure 4, leftmost scatter), and vice versawhen β increases relative to α . These scatter plots clearly illustrate that the two loss terms present an antagonist behaviour between each other: each loss term diverges as the other approaches zero when the balance ( α , β ) is outside a specific range, roughly delineated by the red pair of lines. Moreover, it is possible to identify a basin in which both loss components exhibit convergent behaviour. In this range of values, the antagonistic nature of the losses, balancing the need to preserve low frequencies while injecting more high frequencies, creates an optimal trade-off between colours and details in the final results. In this region, we have selected the parameters α , β = 0.03 , 0.3 , which reflect values close to q 1 for both losses. Similar observations can be extended to the other two architectures, PanNet and BDPN. Notably, deeper architectures demonstrate a broader the convergence basin for the two loss components. This suggests that deeper networks have a greater capacity to balance spectral and spatial losses effectively, with reference to their initial value L * ( 0 ) , across a wider range of ( α , β ) values.
In the testing phase, to understand the effects of this trade-off between spectral and spatial consistency on the outcome, we fine-tuned the three architectures for 500 epochs on the test dataset. Figure 6, Figure 7 and Figure 8 present sample (taken from Adelaide dataset) results obtained by Z-PNN, PanNet, and BDPN, respectively, fine-tuned with selected ( α , β ) balances, ranging from spectrally boosted (high α / β ) to spatially boosted (low α / β ) settings. For all methods, higher α / β ratios produce a faithful colour response (compared to the MS) but result in poor spatial details (compared to the PAN), exhibiting blur and noisy textural patterns. Conversely, lower α / β ratios enhance the injection of PAN details into the pansharpened image, at the cost of spectral degradation (chromatic shift). Finally, a good balance guarantees precise colours and clean details, correctly injected within the scene.
In addition to visual inspection, we quantitatively evaluated how different ( α , β ) configurations impact on the evaluation metrics. The values obtained are visually plotted in Figure 9, Figure 10 and Figure 11 for Z-PNN, PanNet, and BDPN, respectively. This analysis provides interesting insights: focusing on the red lines, indicating D λ , at parity of α , for high β weights D λ diverges as expected. However, for low β values, the spectral consistency metric remains substantially constant. Conversely, moving to the D ρ curve, plotted in blue, this trend becomes specular. The convergence of D λ ( D ρ ) is guaranteed even with not large α / β (low α / β ) ratios, and the network can correctly inject details without any critical loss of colours until a specific limit configuration, the optimal configuration.
This configuration (indicated as “Balanced” version) is used for overall comparison.
Table 5 presents numerical results for both the Adelaide and Munich (PairMax) datasets. The “Balanced” configuration, for all three methods, achieves highly competitive indexes across all metrics. It is worth noting a general difficulty for BDPN in preserving colours compared to the other two CNNs, likely due to its architecture, which includes several layers for feature extraction from the P component but relatively few for processing the information from M. Nonetheless, these results remain competitive compared to many other deep-learning methods and most model-based approaches.

5. Conclusions

In this work, we demonstrate the antagonistic nature between spectral and spatial losses necessary for training deep learning-based pansharpening algorithms. Unsupervised approaches show critical sensitivity to this balance, which can lead to degraded performance in color or detail. Through extensive validation with three state-of-the-art CNNs, we empirically define a general rule for effectively balancing these losses to achieve an optimal trade-off between spectral and spatial quality. This involves identifying the optimal α / β ratio, which constrains spectral loss values while minimizing structural loss. The α and β parameters are chosen based on the transition area between black and blue squares in specific figures.
We also analyze the effects of individual losses and their combinations on final results, both through visual inspection and evaluation metrics. Our balanced approach compares favorably with several state-of-the-art pansharpening methods, yielding promising results.
However, our solution has limitations: it is not suitable for optimizing adversarial losses, which must be contrastive by nature. It also does not address challenges such as moving objects or pansharpening of bands not covered by PAN wavelengths. Future improvements could involve developing a more precise formula to automatically and dynamically set the optimal balance based on image content.

Author Contributions

Conceptualization, M.C. and G.S.; Methodology, M.C. and G.S.; Software, M.C. and G.G.; Validation, M.C. and G.G.; Formal analysis, M.C.; Investigation, M.C.; Resources, M.C.; Data curation, M.C. and G.G.; Writing—original draft, M.C.; Writing—review and editing, M.C., G.G. and G.S.; Visualization, M.C.; Supervision, G.S.; Project administration, G.S.; Funding acquisition, G.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

Adelaide courtesy of Maxar©. Fortaleza, Mexico City, Xian (Maxar©) provided by ESA. Restrictions apply to the availability of these data. Data were obtained from ESA and are available on https://earth.esa.int/eogateway/missions/worldview-3 with the permission of Maxar©. Munich is part of PairMax dataset (https://resources.maxar.com/product-samples/pansharpening-benchmark-dataset).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Vivone, G.; Dalla Mura, M.; Garzelli, A.; Restaino, R.; Scarpa, G.; Ulfarsson, M.O.; Alparone, L.; Chanussot, J. A New Benchmark Based on Recent Advances in Multispectral Pansharpening: Revisiting Pansharpening With Classical and Emerging Pansharpening Methods. IEEE Geosci. Remote Sens. Mag. 2021, 9, 53–81. [Google Scholar] [CrossRef]
  2. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS+Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  3. Lolli, S.; Alparone, L.; Garzelli, A.; Vivone, G. Haze Correction for Contrast-Based Multispectral Pansharpening. IEEE Geosci. Remote Sens. Lett. 2017, 14, 2255–2259. [Google Scholar] [CrossRef]
  4. Vivone, G. Robust Band-Dependent Spatial-Detail Approaches for Panchromatic Sharpening. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6421–6433. [Google Scholar] [CrossRef]
  5. Otazu, X.; González-Audícana, M.; Fors, O.; Núñez, J. Introduction of sensor spectral response into image fusion methods. Application to wavelet-based methods. IEEE Trans. Geosci. Remote Sens. 2005, 43, 2376–2385. [Google Scholar] [CrossRef]
  6. Alparone, L.; Garzelli, A.; Vivone, G. Intersensor Statistical Matching for Pansharpening: Theoretical Issues and Practical Solutions. IEEE Trans. Geosci. Remote Sens. 2017, 55, 4682–4695. [Google Scholar] [CrossRef]
  7. Vivone, G.; Restaino, R.; Chanussot, J. A regression-based high-pass modulation pansharpening approach. IEEE Trans. Geosci. Remote Sens. 2017, 56, 984–996. [Google Scholar] [CrossRef]
  8. Palsson, F.; Sveinsson, J.R.; Ulfarsson, M.O. A New Pansharpening Algorithm Based on Total Variation. IEEE Geosci. Remote Sens. Lett. 2014, 11, 318–322. [Google Scholar] [CrossRef]
  9. Wang, T.; Fang, F.; Li, F.; Zhang, G. High-Quality Bayesian Pansharpening. IEEE Trans. Image Process. 2019, 28, 227–239. [Google Scholar] [CrossRef]
  10. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. CNN-based Pansharpening of Multi-Resolution Remote-Sensing Images. In Proceedings of the Joint Urban Remote Sensing Event 2017, Dubai, United Arab Emirates, 6–8 March 2017. [Google Scholar]
  11. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A Deep Network Architecture for Pan-Sharpening. In Proceedings of the ICCV 2017, Venice, Italy, 22–29 October 2017. [Google Scholar] [CrossRef]
  12. Deng, L.J.; Vivone, G.; Jin, C.; Chanussot, J. Detail injection-based deep convolutional neural networks for pansharpening. IEEE Trans. Geosci. Remote Sens. 2020, 59, 6995–7010. [Google Scholar] [CrossRef]
  13. Chavez, P.S.; Kwarteng, A.W. Extracting spectral contrast in Landsat thematic mapper image data using selective principal component analysis. Photogramm. Eng. Remote Sens. 1989, 55, 339–348. [Google Scholar]
  14. Laben, C.A.; Brower, B.V. Process for Enhancing the Spatial Resolution of Multispectral Imagery Using Pan-Sharpening. U.S. Patent 6011875, 4 January 2000. [Google Scholar]
  15. Gillespie, A.R.; Kahle, A.B.; Walker, R.E. Color enhancement of highly correlated images. II. Channel ratio and “chromaticity” transformation techniques. Remote Sens. Environ. 1987, 22, 343–365. [Google Scholar] [CrossRef]
  16. Tu, T.M.; Huang, P.S.; Hung, C.L.; Chang, C.P. A fast intensity hue-saturation fusion technique with spectral adjustment for IKONOS imagery. IEEE Geosci. Remote Sens. Lett. 2004, 1, 309–312. [Google Scholar] [CrossRef]
  17. Ranchin, T.; Wald, L. Fusion of high spatial and spectral resolution images: The ARSIS concept and its implementation. Photogramm. Eng. Remote Sens. 2000, 66, 49–61. [Google Scholar]
  18. Khan, M.M.; Chanussot, J.; Condat, L.; Montanvert, A. Indusion: Fusion of Multispectral and Panchromatic Images Using the Induction Scaling Technique. IEEE Geosci. Remote Sens. Lett. 2008, 5, 98–102. [Google Scholar] [CrossRef]
  19. Aiazzi, B.; Alparone, L.; Baronti, S.; Garzelli, A. Context-driven fusion of high spatial and spectral resolution images based on oversampled multiresolution analysis. IEEE Trans. Geosci. Remote Sens. 2002, 40, 2300–2312. [Google Scholar] [CrossRef]
  20. Restaino, R.; Mura, M.D.; Vivone, G.; Chanussot, J. Context-Adaptive Pansharpening Based on Image Segmentation. IEEE Trans. Geosci. Remote Sens. 2017, 55, 753–766. [Google Scholar] [CrossRef]
  21. Vivone, G.; Simões, M.; Dalla Mura, M.; Restaino, R.; Bioucas-Dias, J.M.; Licciardi, G.A.; Chanussot, J. Pansharpening Based on Semiblind Deconvolution. IEEE Trans. Geosci. Remote Sens. 2015, 53, 1997–2010. [Google Scholar] [CrossRef]
  22. Vicinanza, M.R.; Restaino, R.; Vivone, G.; Mura, M.D.; Chanussot, J. A Pansharpening Method Based on the Sparse Representation of Injected Details. IEEE Geosci. Remote Sens. Lett. 2015, 12, 180–184. [Google Scholar] [CrossRef]
  23. Palsson, F.; Ulfarsson, M.O.; Sveinsson, J.R. Model-Based Reduced-Rank Pansharpening. IEEE Geosci. Remote Sens. Lett. 2020, 17, 656–660. [Google Scholar] [CrossRef]
  24. Vitale, S.; Ferraioli, G.; Pascazio, V. A New Ratio Image Based CNN Algorithm for SAR Despeckling. In Proceedings of the IGARSS 2019, Yokohama, Japan, 28 July–2 August 2019; pp. 9494–9497. [Google Scholar] [CrossRef]
  25. Fayad, I.; Ienco, D.; Baghdadi, N.; Gaetano, R.; Alvares, C.A.; Stape, J.L.; Ferraço Scolforo, H.; Le Maire, G. A CNN-based approach for the estimation of canopy heights and wood volume from GEDI waveforms. Remote Sens. Environ. 2021, 265, 112652. [Google Scholar] [CrossRef]
  26. Nyborg, J.; Pelletier, C.; Lefèvre, S.; Assent, I. TimeMatch: Unsupervised cross-region adaptation by temporal shift estimation. ISPRS J. Photogramm. Remote Sens. 2022, 188, 301–313. [Google Scholar] [CrossRef]
  27. Vitale, S.; Ferraioli, G.; Pascazio, V. Analysis on the Building of Training Dataset for Deep Learning SAR Despeckling. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4015005. [Google Scholar] [CrossRef]
  28. Haq, M.A. CNN based automated weed detection system using UAV imagery. Comput. Syst. Sci. Eng. 2022, 42, 837–849. [Google Scholar]
  29. Belmouhcine, A.; Burnel, J.C.; Courtrai, L.; Pham, M.T.; Lefèvre, S. Multimodal Object Detection in Remote Sensing. In Proceedings of the IGARSS 2023—2023 IEEE International Geoscience and Remote Sensing Symposium, Pasadena, CA, USA, 16–21 July 2023; pp. 1245–1248. [Google Scholar] [CrossRef]
  30. Haq, M.A.; Hassine, S.B.H.; Malebary, S.J.; Othman, H.A.; Tag-Eldin, E.M. 3D-CNNHSR: A 3-Dimensional Convolutional Neural Network for Hyperspectral Super-Resolution. Comput. Syst. Sci. Eng. 2023, 47, 2689–2705. [Google Scholar]
  31. Guarino, G.; Ciotola, M.; Vivone, G.; Scarpa, G. Band-Wise Hyperspectral Image Pansharpening Using CNN Model Propagation. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5500518. [Google Scholar] [CrossRef]
  32. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by Convolutional Neural Networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef]
  33. Yuan, Q.; Wei, Y.; Meng, X.; Shen, H.; Zhang, L. A Multiscale and Multidepth Convolutional Neural Network for Remote Sensing Imagery Pan-Sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 978–989. [Google Scholar] [CrossRef]
  34. Liu, X.; Wang, Y.; Liu, Q. Psgan: A Generative Adversarial Network for Remote Sensing Image Pan-Sharpening. In Proceedings of the 2018 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 873–877. [Google Scholar]
  35. Shao, Z.; Cai, J. Remote Sensing Image Fusion With Deep Convolutional Neural Network. IEEE J. Sel. Topics Appl. Earth Observ. 2018, 11, 1656–1669. [Google Scholar] [CrossRef]
  36. Scarpa, G.; Vitale, S.; Cozzolino, D. Target-Adaptive CNN-Based Pansharpening. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5443–5457. [Google Scholar] [CrossRef]
  37. Zhang, Y.; Liu, C.; Sun, M.; Ou, Y. Pan-Sharpening Using an Efficient Bidirectional Pyramid Network. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5549–5563. [Google Scholar] [CrossRef]
  38. He, L.; Rao, Y.; Li, J.; Chanussot, J.; Plaza, A.; Zhu, J.; Li, B. Pansharpening via Detail Injection Based Convolutional Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 1188–1204. [Google Scholar] [CrossRef]
  39. Guarino, G.; Ciotola, M.; Vivone, G.; Poggi, G.; Scarpa, G. PCA-CNN Hybrid Approach for Hyperspectral Pansharpening. IEEE Geosci. Remote Sens. Lett. 2023, 20, 5511505. [Google Scholar] [CrossRef]
  40. Gong, M.; Zhang, H.; Xu, H.; Tian, X.; Ma, J. Multipatch Progressive Pansharpening With Knowledge Distillation. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5401115. [Google Scholar] [CrossRef]
  41. Gong, M.; Ma, J.; Xu, H.; Tian, X.; Zhang, X.P. D2TNet: A ConvLSTM Network With Dual-Direction Transfer for Pan-Sharpening. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5409114. [Google Scholar] [CrossRef]
  42. Kumar, D.G.; Joseph, C.; Subbarao, M.V. An Efficient PAN-sharpening of Multispectral Images using Multi-scale Residual CNN with Sparse Representation. In Proceedings of the 2024 International Conference on Integrated Circuits and Communication Systems (ICICACS), Raichur, India, 23–24 February 2024; pp. 1–8. [Google Scholar] [CrossRef]
  43. Deng, L.j.; Vivone, G.; Paoletti, M.E.; Scarpa, G.; He, J.; Zhang, Y.; Chanussot, J.; Plaza, A. Machine Learning in Pansharpening: A benchmark, from shallow to deep networks. IEEE Geosci. Remote Sens. Mag. 2022, 10, 279–315. [Google Scholar] [CrossRef]
  44. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolution: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  45. Luo, S.; Zhou, S.; Feng, Y.; Xie, J. Pansharpening via Unsupervised Convolutional Neural Networks. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4295–4310. [Google Scholar] [CrossRef]
  46. Ciotola, M.; Vitale, S.; Mazza, A.; Poggi, G.; Scarpa, G. Pansharpening by Convolutional Neural Networks in the Full Resolution Framework. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5408717. [Google Scholar] [CrossRef]
  47. Ciotola, M.; Poggi, G.; Scarpa, G. Unsupervised Deep Learning-Based Pansharpening With Jointly Enhanced Spectral and Spatial Fidelity. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5405417. [Google Scholar] [CrossRef]
  48. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Guo, X.; Jiang, J. Pan-GAN: An unsupervised pan-sharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  49. Zhou, H.; Liu, Q.; Wang, Y. PGMAN: An Unsupervised Generative Multiadversarial Network for Pansharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6316–6327. [Google Scholar] [CrossRef]
  50. Liu, X.; Liu, X.; Dai, H.; Kang, X.; Plaza, A.; Zu, W. Mun-GAN: A Multiscale Unsupervised Network for Remote Sensing Image Pansharpening. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5404018. [Google Scholar] [CrossRef]
  51. Zhou, H.; Liu, Q.; Weng, D.; Wang, Y. Unsupervised Cycle-Consistent Generative Adversarial Networks for Pan Sharpening. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5408814. [Google Scholar] [CrossRef]
  52. Qu, Y.; Baghbaderani, R.K.; Qi, H.; Kwan, C. Unsupervised Pansharpening Based on Self-Attention Mechanism. IEEE Trans. Geosci. Remote Sens. 2021, 59, 3192–3208. [Google Scholar] [CrossRef]
  53. Ni, J.; Shao, Z.; Zhang, Z.; Hou, M.; Zhou, J.; Fang, L.; Zhang, Y. LDP-Net: An Unsupervised Pansharpening Network Based on Learnable Degradation Processes. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 5468–5479. [Google Scholar] [CrossRef]
  54. Wang, D.; Zhang, P.; Bai, Y.; Li, Y. MetaPan: Unsupervised Adaptation With Meta-Learning for Multispectral Pansharpening. IEEE Geosci. Remote Sens. Lett. 2022, 19, 5513505. [Google Scholar] [CrossRef]
  55. Ciotola, M.; Scarpa, G. Fast Full-Resolution Target-Adaptive CNN-Based Pansharpening Framework. Remote Sens. 2023, 15, 319. [Google Scholar] [CrossRef]
  56. Uezato, T.; Hong, D.; Yokoya, N.; He, W. Guided deep decoder: Unsupervised image pair fusion. In Proceedings of the European Conference on Computer Vision 2020, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 87–102. [Google Scholar]
  57. Xiong, Z.; Liu, N.; Wang, N.; Sun, Z.; Li, W. Unsupervised Pansharpening Method Using Residual Network With Spatial Texture Attention. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5402112. [Google Scholar] [CrossRef]
  58. Alparone, L.; Aiazzi, B.; Baronti, S.; Garzelli, A.; Nencini, F.; Selva, M. Multispectral and panchromatic data fusion assessment without reference. Photogramm. Eng. Remote Sens. 2008, 74, 193–200. [Google Scholar] [CrossRef]
  59. Wang, Z.; Bovik, A.C. A universal image quality index. IEEE Signal Processing Letters 2002, 9, 81–84. [Google Scholar] [CrossRef]
  60. Kruse, F.A.; Lefkoff, A.; Boardman, J.; Heidebrecht, K.; Shapiro, A.; Barloon, P.; Goetz, A. The spectral image processing system (SIPS)—interactive visualization and analysis of imaging spectrometer data. Remote Sens. Environ. 1993, 44, 145–163. [Google Scholar] [CrossRef]
  61. Scarpa, G.; Ciotola, M. Full-Resolution Quality Assessment for Pansharpening. Remote Sens. 2022, 14, 1808. [Google Scholar] [CrossRef]
  62. Khan, M.M.; Alparone, L.; Chanussot, J. Pansharpening Quality Assessment Using the Modulation Transfer Functions of Instruments. IEEE Trans. Geosci. Remote Sens. 2009, 47, 3880–3891. [Google Scholar] [CrossRef]
  63. Garzelli, A.; Nencini, F. Hypercomplex Quality Assessment of Multi/Hyperspectral Images. IEEE GRSL 2009, 6, 662–665. [Google Scholar] [CrossRef]
  64. Scarpa, G.; Gargiulo, M.; Mazza, A.; Gaetano, R. A CNN-Based Fusion Method for Feature Extraction from Sentinel Data. Remote Sens. 2018, 10, 236. [Google Scholar] [CrossRef]
  65. Vivone, G.; Dalla Mura, M.; Garzelli, A.; Pacifici, F. A Benchmarking Protocol for Pansharpening: Dataset, Preprocessing, and Quality Assessment. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6102–6118. [Google Scholar] [CrossRef]
  66. Garzelli, A.; Nencini, F.; Capobianco, L. Optimal MMSE pan sharpening of very high resolution multispectral images. IEEE Trans. Geosci. Remote Sens. 2008, 46, 228–236. [Google Scholar] [CrossRef]
  67. Garzelli, A. Pansharpening of Multispectral Images Based on Nonlocal Parameter Optimization. IEEE Trans. Geosci. Remote Sens. 2015, 53, 2096–2107. [Google Scholar] [CrossRef]
  68. Choi, J.; Yu, K.; Kim, Y. A New Adaptive Component-Substitution-Based Satellite Image Fusion by Using Partial Replacement. IEEE Trans. Geosci. Remote Sens. 2011, 49, 295–309. [Google Scholar] [CrossRef]
  69. Vivone, G.; Restaino, R.; Chanussot, J. Full Scale Regression-Based Injection Coefficients for Panchromatic Sharpening. IEEE Trans. Image Process. 2018, 27, 3418–3431. [Google Scholar] [CrossRef] [PubMed]
  70. Alparone, L.; Wald, L.; Chanussot, J.; Thomas, C.; Gamba, P.; Bruce, L. Comparison of pansharpening algorithms: Outcome of the 2006 GRS-S Data-Fusion Contest. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3012–3021. [Google Scholar] [CrossRef]
  71. Restaino, R.; Vivone, G.; Dalla Mura, M.; Chanussot, J. Fusion of Multispectral and Panchromatic Images Based on Morphological Operators. IEEE Trans. Image Process. 2016, 25, 2882–2895. [Google Scholar] [CrossRef] [PubMed]
  72. Wei, Y.; Yuan, Q.; Shen, H.; Zhang, L. Boosting the Accuracy of Multispectral Image Pansharpening by Learning a Deep Residual Network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1795–1799. [Google Scholar] [CrossRef]
Figure 1. A-PNN model for Z-PNN [46].
Figure 1. A-PNN model for Z-PNN [46].
Remotesensing 16 03014 g001
Figure 2. PanNet architecture [11].
Figure 2. PanNet architecture [11].
Remotesensing 16 03014 g002
Figure 3. BDPN model [37].
Figure 3. BDPN model [37].
Remotesensing 16 03014 g003
Figure 4. Spectral loss terms after 200 iterations for different α , β configurations.
Figure 4. Spectral loss terms after 200 iterations for different α , β configurations.
Remotesensing 16 03014 g004
Figure 5. Spatial loss terms after 200 iterations for different α , β configurations.
Figure 5. Spatial loss terms after 200 iterations for different α , β configurations.
Remotesensing 16 03014 g005
Figure 6. Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by Z-PNN through target adaptation for 500 epochs with different ( α , β ) settings.
Figure 6. Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by Z-PNN through target adaptation for 500 epochs with different ( α , β ) settings.
Remotesensing 16 03014 g006
Figure 7. Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by PanNet through target adaptation for 500 epochs with different ( α , β ) settings.
Figure 7. Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by PanNet through target adaptation for 500 epochs with different ( α , β ) settings.
Remotesensing 16 03014 g007
Figure 8. Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by BDPN through target adaptation for 500 epochs with different ( α , β ) settings.
Figure 8. Sample results taken from WV3 Adelaide dataset: input MS and PAN followed by related pansharpenings obtained by BDPN through target adaptation for 500 epochs with different ( α , β ) settings.
Remotesensing 16 03014 g008
Figure 9. Full-resolution accuracy indexes for Z-PNN, with α = 0.03 and β 3 · 10 5 , 3 · 10 4 , , 3 · 10 3 .
Figure 9. Full-resolution accuracy indexes for Z-PNN, with α = 0.03 and β 3 · 10 5 , 3 · 10 4 , , 3 · 10 3 .
Remotesensing 16 03014 g009
Figure 10. Full-resolution accuracy indexes for PanNet, with α = 0.03 and β 3 · 10 5 , 3 · 10 4 , , 3 · 10 3 .
Figure 10. Full-resolution accuracy indexes for PanNet, with α = 0.03 and β 3 · 10 5 , 3 · 10 4 , , 3 · 10 3 .
Remotesensing 16 03014 g010
Figure 11. Full-resolution accuracy indexes for BDPN, with α = 0.03 and β 3 · 10 5 , 3 · 10 4 , , 3 · 10 3 .
Figure 11. Full-resolution accuracy indexes for BDPN, with α = 0.03 and β 3 · 10 5 , 3 · 10 4 , , 3 · 10 3 .
Remotesensing 16 03014 g011
Table 1. List of the main symbols and operators used in the paper.
Table 1. List of the main symbols and operators used in the paper.
SymbolDescription
Rresolution ratio
Bnumber of multispectral bands
M, Poriginal multispectral and panchromatic components
Isimulated P obtained by linearly combining M components
M ^ pansharpened image
M ^ ( R × ) downscaled version of M ^
M ˜ ( R × ) upscaled version of M
M ^ lp low-pass filtered versions of M ^
P hp , M ^ hp high-pass filtered versions of P and M ^
P ˜ B × expanded panchromatic component
L λ , L S , L spectral, spatial and total loss
L a d v Binary Cross-Entropy loss
· spatial and spectral average
image gradient operation
ρ local correlation coefficient
ρ σ suitably estimated pixel/band-wise upper-bound for ρ
u ( · ) step function
Table 2. A brief summary of the unsupervised losses currently proposed in the literature for high-resolution pansharpening. PGMAN [49] and SSQ [45] also include L QNR in their respective loss functions. These loss functions reflect different strategies to solve some of the most common issues of unsupervised deep learning-based pansharpening algorithms. The Reader is referred to the original papers for more details.
Table 2. A brief summary of the unsupervised losses currently proposed in the literature for high-resolution pansharpening. PGMAN [49] and SSQ [45] also include L QNR in their respective loss functions. These loss functions reflect different strategies to solve some of the most common issues of unsupervised deep learning-based pansharpening algorithms. The Reader is referred to the original papers for more details.
Loss Name L λ L S α β
L SSQ [45] M ^ lp M ˜ 2 + [ 1 SSIM ( M ^ lp , M ˜ ) ] P I 2 + [ 1 SSIM ( P , I ) ] 1.001.00
L PGMAN [49] L a d v ( M ^ , M ) L a d v ( I , P ) 2.001.00
L PanGAN [48] M ^ M 2 + L a d v ( M ^ , M ) P I 2 + L a d v ( P , I ) 1.005.00
L GDD [56] M ^ M 2 P I 1 0.101.00
L UAPNet [57] 0.001 · M ^ lp M ˜ 2 + SAM M ^ , M 0.045 · M ^ hp P hp 2 + Q M ^ , P ˜ 0.790.20
L Z PNN [46] M ^ M 1 ( 1 ρ σ ) u ( ρ max ( M ˜ , P ) ρ σ ( M ^ , P ) ) 1.000.36
L λ PNN [47] 0.04 · ERGAS M ^ , M + D λ M ^ , M ( 1 ρ σ ) u ( ρ max ( M ˜ , P ) ρ σ ( M ^ , P ) ) 1.253.75
Table 3. WV3 Dataset, with number of crops for training, validation and test. Adelaide courtesy of DigitalGlobe©. Fortaleza, Mexico City, Xian (DigitalGlobe©) provided by ESA. Munich is part of PairMax dataset [65].
Table 3. WV3 Dataset, with number of crops for training, validation and test. Adelaide courtesy of DigitalGlobe©. Fortaleza, Mexico City, Xian (DigitalGlobe©) provided by ESA. Munich is part of PairMax dataset [65].
WorldView-3 (GSD at Nadir: 0.31 m)
DatasetTrainingValidationTest
(PAN Size)(512 × 512)(512 × 512)(2048 × 2048)
Fortaleza328-
Mexico City328-
Xian328-
Adelaide-243
Munich (PairMax)--3
Table 4. Detailed list of all reference methods.
Table 4. Detailed list of all reference methods.
Component Substitution (CS)
BT-H [3], BDSD [66], C-BDSD [67], BDSD-PC [4], GS [14], GSA [2], C-GSA [20], PRACS [68]
Multiresolution Analysis (MRA)
AWLP [5], MTF-GLP [6], MTF-GLP-FS [69], MTF-GLP-HPM [6], MTF-GLP-HPM-H [3],
MTF-GLP-HPM-R [7], MTF-GLP-CBD [70], C-MTF-GLP-CBD [20], MF [71]
Variational Optimization (VO)
FE-HPM [21], SR-D [22], TV [8]
Supervised Deep Learning-based
PNN [32], A-PNN [36], A-PNN-TA [36], BDPN [37], DiCNN [38], DRPNN [72],
FusionNet [12], MSDCNN [33], PanNet [11]
Unsupervised Deep Learning-based
QSS [45], GDD [56], PanGan [48], Z-PNN [46], λ -PNN [47]
Table 5. Average results on WV3-Test. Left: Adelaide. Right: Munich (PairMax). Best, Second, and Third ranked comparative algorithms are in red, blue, and green, while proposed Balanced solutions are in bold.
Table 5. Average results on WV3-Test. Left: Adelaide. Right: Munich (PairMax). Best, Second, and Third ranked comparative algorithms are in red, blue, and green, while proposed Balanced solutions are in bold.
MethodAdelaidePairMax
D λ R-SAMR-ERGAS D ρ D λ R-SAMR-ERGAS D ρ
EXP0.07245.14464.65280.84410.06372.67883.91060.8542
BT-H0.06525.13394.41040.07570.07932.93674.16150.0487
BDSD0.09696.37265.37570.12190.11983.38525.59610.0749
C-BDSD0.11986.66546.16080.18220.13703.90096.73130.1128
BDSD-PC0.08025.74584.84760.08850.10753.32665.44720.0585
GS0.08545.39084.89400.09340.12813.55005.01210.0722
GSA0.06125.50154.42110.07390.07543.80934.41560.0507
C-GSA0.06145.64684.45000.12990.07823.73754.41100.0603
PRACS0.06025.13214.25190.17150.06282.94083.91870.1953
AWLP0.04935.13133.91300.10280.04322.55073.07840.0793
MTF-GLP0.04935.15703.93480.08420.04162.45022.94810.0556
MTF-GLP-FS0.05135.16463.99890.10900.04282.45652.97920.0672
MTF-GLP-HPM0.04925.20243.94510.08910.04792.96813.38310.0605
MTF-GLP-HPM-H0.04885.23523.95500.08130.04252.56822.95280.0492
MTF-GLP-HPM-R0.05105.17823.99880.12090.04232.47422.98610.0711
MTF-GLP-CBD0.05155.16434.00690.11240.04322.45882.99500.0701
C-MTF-GLP-CBD0.05655.15374.18450.20810.04682.60513.25400.1628
MF0.04445.13063.75840.11280.04242.49813.03710.0780
FE-HPM0.05035.17664.03670.11150.04272.54003.10300.0728
SR-D0.05575.34384.28330.30140.03402.32902.77680.1857
TV0.03524.10763.34270.21490.04031.72892.88770.1684
PNN0.10597.29786.59810.46120.41638.02929.11460.4754
A-PNN0.05985.27484.27790.51440.16933.54924.39840.6649
A-PNN-FT0.05585.13004.17300.30870.08452.56423.52050.3352
BDPN0.12295.80255.77120.16240.29446.58287.69260.3107
DiCNN0.12076.41655.97540.39570.21694.93396.07720.4101
DRPNN0.11685.84075.65790.19450.19814.90186.59790.1861
FusionNet0.07305.49704.76550.41330.23824.55295.30200.3402
MSDCNN0.13496.13175.89020.20050.39237.65916.66390.3484
PanNet0.05545.05634.17370.34030.06092.54163.32210.2970
QSS0.05905.27824.27300.28530.08283.19633.56710.2843
PanGan0.658614.633013.76310.71340.381615.056317.58950.1317
GDD0.194911.23207.44590.64400.30449.96708.69310.5867
Z-PNN0.03744.85443.29250.09460.09193.57593.76660.1012
λ -PNN0.02033.76312.53630.05740.03412.36902.66060.0515
Balanced Z-PNN0.01943.47592.37980.10840.04152.18832.51030.1126
Balanced PanNet0.01703.31672.25240.05920.03322.03812.41030.0435
Balanced BDPN0.04125.37973.38330.10250.09343.37603.51320.0884
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ciotola, M.; Guarino, G.; Scarpa, G. An Unsupervised CNN-Based Pansharpening Framework with Spectral-Spatial Fidelity Balance. Remote Sens. 2024, 16, 3014. https://doi.org/10.3390/rs16163014

AMA Style

Ciotola M, Guarino G, Scarpa G. An Unsupervised CNN-Based Pansharpening Framework with Spectral-Spatial Fidelity Balance. Remote Sensing. 2024; 16(16):3014. https://doi.org/10.3390/rs16163014

Chicago/Turabian Style

Ciotola, Matteo, Giuseppe Guarino, and Giuseppe Scarpa. 2024. "An Unsupervised CNN-Based Pansharpening Framework with Spectral-Spatial Fidelity Balance" Remote Sensing 16, no. 16: 3014. https://doi.org/10.3390/rs16163014

APA Style

Ciotola, M., Guarino, G., & Scarpa, G. (2024). An Unsupervised CNN-Based Pansharpening Framework with Spectral-Spatial Fidelity Balance. Remote Sensing, 16(16), 3014. https://doi.org/10.3390/rs16163014

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop