Next Article in Journal
Mapping Irrigated Rice in Brazil Using Sentinel-2 Spectral–Temporal Metrics and Random Forest Algorithm
Previous Article in Journal
BresNet: Applying Residual Learning in Backpropagation Neural Networks to Predict Ground Surface Concentration of Primary Air Pollutants
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Robust Pansharpening: A Large-Scale High-Resolution Multi-Scene Dataset and Novel Approach

1
School of Computer Technology and Application, Qinghai University, Xining 810016, China
2
Intelligent Computing and Application Laboratory of Qinghai Province, Qinghai University, Xining 810016, China
3
School of Computer and Information Technology, Beijing Jiaotong University, Beijing 100044, China
4
Department of Computer Science and Technology, Tsinghua University, Beijing 100084, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Remote Sens. 2024, 16(16), 2899; https://doi.org/10.3390/rs16162899
Submission received: 8 July 2024 / Revised: 5 August 2024 / Accepted: 6 August 2024 / Published: 8 August 2024

Abstract

:
Pansharpening, a pivotal task in remote sensing, involves integrating low-resolution multispectral images with high-resolution panchromatic images to synthesize an image that is both high-resolution and retains multispectral information. These pansharpened images enhance precision in land cover classification, change detection, and environmental monitoring within remote sensing data analysis. While deep learning techniques have shown significant success in pansharpening, existing methods often face limitations in their evaluation, focusing on restricted satellite data sources, single scene types, and low-resolution images. This paper addresses this gap by introducing PanBench, a high-resolution multi-scene dataset containing all mainstream satellites and comprising 5898 pairs of samples. Each pair includes a four-channel (RGB + near-infrared) multispectral image of 256 × 256 pixels and a mono-channel panchromatic image of 1024 × 1024 pixels. To avoid irreversible loss of spectral information and achieve a high-fidelity synthesis, we propose a Cascaded Multiscale Fusion Network (CMFNet) for pansharpening. Multispectral images are progressively upsampled while panchromatic images are downsampled. Corresponding multispectral features and panchromatic features at the same scale are then fused in a cascaded manner to obtain more robust features. Extensive experiments validate the effectiveness of CMFNet.

1. Introduction

With the continuous advancement of remote sensing technology, a substantial amount of Earth observation data can be readily obtained. Due to the technological constraints of remote sensing imaging sensors, remote sensing images are typically provided as low-resolution multispectral (MS) images and high-resolution panchromatic (PAN) images. The task of pansharpening [1] has emerged to obtain images with high spatial and spectral resolutions. Pansharpening aims to reconstruct a high-resolution MS image from a low-resolution MS image guided by a PAN image. The pansharpened images enable better analysis and interpretation of the data, serving as an indispensable preprocessing step for downstream tasks in remote sensing, such as land cover classification [2,3], object recognition [4,5], and change detection [6,7]. This provides high-quality data for downstream remote sensing [8,9,10] applications.
Deep learning methods [11,12,13,14] have been widely applied in recent years and have made significant progress in pansharpening. They usually utilize deep neural network models [15,16] to learn the complex mapping relationship between MS and PAN images through end-to-end training. Existing deep learning-based pansharpening algorithms for remote sensing images can be broadly categorized into three types: (1) pixel-level fusion, (2) feature-level fusion, and (3) pixel–feature-level fusion. Pixel-level fusion splices the up-sampled MS image and PAN image directly on the channel dimension and then inputs them into the neural model for processing; this includes methods such as PNN [17], MSDCNN [18], GPPNN [19], and so on. While pixel-level fusion is easy to implement, it may potentially destroy the spectral information of the MS image, resulting in fused images that are not sufficiently clear in certain scenarios. Feature-level fusion extracts modality-aware features from the panchromatic and multispectral images independently, followed by information fusion in the feature space. Algorithms in this category include PanNet [20], TFNet [21], and FusionNet [22], among others. This kind of method can better retain the band information of MS images. Still, it needs many transformations and calculations and is easily affected by the feature-extraction algorithm. Pixel–feature-level fusion refers to injecting PAN images as reference information into the network in stages based on image super-resolution tasks to guide the entire spatial information reconstruction process; this includes methods such as SFIIN [23] and MIDP [24]. Some of these algorithms introduce an information-driven framework to reduce redundancy and enhance model performance.
Although deep learning-based pansharpening methods for remote sensing images have achieved promising results, numerous challenges and issues remain to be addressed. On the one hand, the current research on pansharpening relies on small-scale patch cropping, limited satellite sources, and a single category of scenes, which can lead to the poor overall generalization capability of the models. On the other hand, in existing methods, multispectral and panchromatic images are typically concatenated for channel splicing, leading to coupled feature-extraction processes [17,18]. Alternatively, downsampling operations may be applied to low-spatial-resolution multispectral images based on a well-performing dual-branch structure, without considering progressive sampling, resulting in irreversible loss of spectral information [20,21]. Therefore, both are undesirable. Additionally, the lack of consideration for cascaded multiscale fusion during the fusion process indirectly increases the difficulty of image detail recovery. Therefore, a unified high-resolution dataset and stable high-performance baseline are urgently needed for pansharpening to better meet the requirements of practical applications.
We have compiled PanBench, an extensive dataset tailored specifically for the task of pansharpening. PanBench aims to address the aforementioned challenges by aggregating remote sensing data from multiple satellites. It boasts support for 10 prominent satellites, encompassing a collection of 5898 meticulously curated pairs of high-resolution samples. Additionally, PanBench features six meticulously annotated land cover classifications. This comprehensive dataset empowers us to tackle the intricate challenges and variability encountered in practical scenarios, thereby fostering robust generalization capabilities. The breadth and depth of PanBench make significant contributions to the advancement of pansharpening techniques and applications in remote sensing. Subsequently, based on this dataset, we introduce the Cascaded Multiscale Fusion Network (CMFNet). To avoid irreversible loss of spectral information, direct downsampling operations are eschewed on low-spatial-resolution multispectral images. Instead, hierarchical image feature extraction is performed separately on the input panchromatic and multispectral images. Dense cascading and top-down cascading additions are incorporated to enhance feature interconnection at the decoding layer. The fusion of multispectral and panchromatic features at varying scales facilitates the acquisition of more resilient features. Hierarchical feature maps are directly and densely connected, leveraging all coarse low-dimensional features to produce satisfactory high-resolution depth outputs with minimal attenuation in the decoding layers. Consequently, the fused features are decoded to reconstruct the pansharpened image.
To sum up, our main contributions are as follows:
  • We construct PanBench, a large-scale, high-resolution, multi-scene dataset containing all mainstream satellites for pansharpening, including six finely annotated classifications of ground feature coverage. Moreover, we can extend this dataset with applications such as super-resolution and colorization tasks.
  • We propose CMFNet, a high-fidelity fusion network for pansharpening, and the experimental results show that the algorithm has good generalization ability and significantly surpasses the currently representative pansharpening methods.
  • The PanBench dataset is not only suitable for pansharpening tasks in the field of remote sensing, but also supports other computer vision tasks, such as image super-resolution and image coloring. This dataset demonstrates strong adaptability and extensibility across different tasks.
Additionally, we implemented a unified toolbox to evaluate our datasets and mainstream pansharpening models at https://github.com/XavierJiezou/Pansharpening (accessed on 1 August 2024), which facilitates the high-quality development of pansharpening.

2. Related Works

2.1. Datasets for Pansharpening

The dataset plays a crucial role in developing and evaluating pansharpening algorithms. However, previous research (Table 1) exhibits issues in the following aspects: (1) The majority of studies employ two to four satellite images to train and validate the effectiveness of pansharpening algorithms. Among them, GaoFen2 (GF2), IKONOS (IN), QuickBird (QB), WorldView2 (WV2), and WorldView3 (WV3) are commonly used [25]. For instance, PanNet [20] utilizes the IN, WV2, and WV3 datasets, while FusionNet [22] employs the GF2, QB, WV2, and WV3 datasets. However, satellite data sources are relatively limited, and the algorithm’s generalization on other satellites needs to be verified. (2) The input multispectral scales are small. For example, PSGAN [26] and RSIF [27] use 64 × 64 MS images as the input, while MIDP [24] and SFIIN [23] adopt 32 × 32 MS images as the input. This is not competent at batch processing tasks with large-resolution remote sensing images of the real world. (3) The datasets used did not specify land cover categories or included only a single category, such as PGCU [28], PanDiff [29], and USSCNet [30]. There is also some work to divide multiple scenes, but the single source of the satellite cannot prove the generalization on other satellites [31,32]. G. Vivone et al. [33] built a dataset called PAirMax, consisting of 14 panchromatic and multispectral image pairs from different satellites on different landscapes, Meng et al. [34] presented a comprehensive baseline dataset consisting of 2270 pairs of HR PAN/LR MS images from six satellites and three thematic surface features. But, none of them made particularly detailed classifications of different land scenarios.
It is well known that satellites providing PAN and MS image data are suitable for pansharpening missions. GaoFen1 is the first satellite in China’s high-resolution Earth observation system, providing 2-meter PAN and 8-meter MS resolution. GaoFen2 provides 0.8 m PAN and 3.2 m MS resolution; GaoFen6 is similar to GaoFen1, with slightly lower resolution, but higher spectral resolution. The GaoFen series of satellites is suitable for land and resource survey and agriculture and environmental monitoring. Landsat7 and Landsat8 in the United States provide 30 m MS and 15 m PAN resolution, respectively, suitable for surface change monitoring and environmental research. IKONOS is the world’s first commercial high-resolution imaging satellite, providing 1-meter PAN and 4-meter MS resolution for urban planning and disaster emergency response. QuickBird offers 0.61 m PAN and 2.44 m MS resolution, widely used in fine geographic information systems and urban planning; WorldView2 offers a PAN resolution of 0.46 meters and 8 MS bands; WorldView3 offers a PAN resolution of 0.31 m and 16 MS bands; WorldView4 offers 0.31 m PAN and 1.24 m MS resolution, and the WorldView series is primarily used for fine-grained GIS and infrastructure monitoring. The pansharpening of images can allow better analysis and interpretation of data, as well as better service to downstream remote sensing applications. This paper constructs PanBench (Figure 1), a large-scale, high-resolution, and multi-scene dataset containing all major satellites used for pansharpening. To ensure the scene’s diversity, according to [37,38], the scene from PanBench is divided into water, urban, ice/snow, crops, vegetation, and barren. It ensures the algorithm can effectively handle different scenarios and produce reliable results.

2.2. Algorithms for Pansharpening

Pansharpening aims to fuse low-resolution MS images with high-resolution PAN images to obtain high-resolution MS images. Traditional methods for fusing PAN and MS images include component substitution [26], multiresolution analysis [34], and variational-optimization (VO)-based methods [39]. Component substitution methods, such as Intensity–Hue–Saturation (HIS), HSV, principal component analysis (PCA), and the Gram–Schmidt (GS) algorithm, efficiently combine high-frequency details from panchromatic images with multispectral images, but often cause spectral distortion. Multiresolution analysis methods, like the wavelet transform, High-Pass Filtering (HPF), the LP transform, MTF_GLP, and MTF_GLP_HPM, improve image quality by fusing decomposed low-frequency information, but are complex and may lose spatial details. VO methods, known for their pansharpening performance, seek optimized functions based on prior constraints or assumptions, yet selecting appropriate constraints remains challenging.
In recent years, deep learning-based [40] pansharpening methods have received significant attention, which rely on large-scale data to learn the nonlinear relationship between ideal fused high-resolution MS images and low-resolution MS and PAN images. In general, deep learning-based pansharpening methods take the original MS as the ground truth based on the Wald protocol [41], conduct network training under reduced resolution, and apply the trained model to the original PAN and MS images directly to obtain full-resolution fusion images. Typical pansharpening methods based on deep learning mainly include two kinds of network structures: single-branch and double-branch. For instance, inspired by the success of residual networks [42], Yang [20] proposed a deep network called PanNet for pansharpening, which is trained in the high-pass domain to preserve spatial information. The up-sampled MS image is added to the output of the final network through residual connections, thereby preserving spectral information. Cai [35] developed a novel convolutional neural network-based deep super-resolution pansharpening algorithm (SRPPNN) that uses multiscale features in MS images to reconstruct spatial information. Distinguishing the single-branch structure above, two-branch structures usually extract features from PAN and low-resolution MS images, respectively, and then fuse them in hidden space to reconstruct high-resolution MS images. For example, Xu [19] proposed a new paradigm combining depth expansion and observation models to develop the model-driven pansharpening network GPPNN. This pansharpening model takes PAN and low-resolution MS images into account. However, the above pansharpening methods did not explicitly perform complementary learning of information between PAN and MS images, further limiting the performance. Zhou [23] proposed a new mutual information-driven pansharpening framework, which can reduce information redundancy and improve model performance. Zhu [28] has developed a new probability-based global cross-modal upsampling method for translation sharpening, taking full advantage of the global information of each pixel in low-resolution MS images and the cross-modal information of guided PAN images. Existing pansharpening methods often suffer from limited generalization due to reliance on small-scale patch data and single scene types, leading to poor performance in diverse scenarios. Additionally, spectral information loss occurs due to the coupling of multispectral and panchromatic images during feature extraction, as seen in methods like PanNet and GPPNN. Furthermore, insufficient multiscale fusion of spectral and spatial information results in unstable image synthesis performance, as highlighted by limitations in methods such as PGCU. Moreover, the existing methods do not consider the cascaded fusion of multiscale spectral information and spatial information, resulting in unstable image synthesis performance. So, in our proposed approach, hierarchical image feature extraction is performed on the input PAN and MS images, respectively, and dense cascade and top-down cascade addition are used to improve the interconnection between the features of the decoding layer. The hierarchical feature maps are directly densely connected, and all coarse low-dimensional features are exploited to generate satisfactory high-resolution depth outputs without much attenuation through the decoding layers.

3. PanBench Dataset

We constructed PanBench, a novel unified evaluation dataset for pansharpening, which supports 10 mainstream satellites, contains 5898 high-resolution sample pairs, and has six manually labeled land cover classifications. This enables us to capture the challenges of complexity and variability encountered in practical generalization applications and contributes to further advancements in this field.

3.1. Multi-Source Satellite

A broader range of data sources is required to develop effective algorithms for pansharpening tasks under various conditions. The PanBench created includes the primary satellite sources currently available for pansharpening, as shown in Figure 1a. Not all satellites support pansharpening, and some satellites, such as Sentinel-2, do not meet pansharpening’s conditions because they do not have sensors mounted that capture the panchromatic band. The commonly used pansharpening in the GaoFen series includes GF1 and GF2. PanBench adds the GF6 dataset. The GF1, GF2, and GF6 datasets contain 625, 783, and 578 image pairs, respectively. The WorldView series also supplements the WV4 dataset based on the commonly used WV2 and WV3. The WV2, WV3, and WV4 datasets contain 578, 567, and 500 image pairs, respectively. The Landsat series adds the LC7 dataset (576 image pairs) to the LC8 dataset (484 image pairs). In addition, there are the frequently used QB (551 image pairs) and IN (656 image pairs) datasets (Figure 1b). Thereby, the model can learn the characteristics and rules of satellites more comprehensively and accurately and improve the robustness of the model. As we all know, each satellite has a different resolution of radiation. Radiometric resolution refers to the sensor’s ability to distinguish the smallest differences in radiance, typically expressed in bits [43]. For example, satellites like Landsat7 have an 8-bit radiometric resolution, allowing 256 different intensity levels, while WorldView3 and WorldView4 have an 11-bit radiometric resolution, allowing 2048 intensity levels. This higher radiometric resolution enables the sensors to capture finer spectral details, thereby improving the accuracy and reliability of the imagery for precise remote sensing applications such as land cover classification, change detection, and environmental monitoring. All images in the PanBench dataset have been radiologically corrected to ensure the consistency of image data across different sensors and imaging conditions.

3.2. Data Processing

Loading the whole large remote sensing image may occupy many computing resources and much time. The data size can be reduced and the efficiency of data processing can be improved by clipping the image. Firstly, we performed a series of preprocessing operations on the collected source data, radiometric calibration [44], atmospheric correction [45], orthometric correction [46], and image alignment, to initially extract and enhance useful information in the images, and the specific data processing flow is shown in Figure 2. Secondly, we segmented the entire large-scale remote sensing image into some image pairs, which were composed of four-channel (RGB + near-infrared) MS images with a 256 × 256 pixel size and single-channel PAN images with a 1024 × 1024 pixel size. Compared with the training datasets used in the current representative literature, such as those utilized in the GPPNN, PanNet, and PGCU, PanBench offers a significantly larger clipping scale. This extensive dataset provides a more comprehensive and diverse training resource, enhancing the generalization capability of pansharpening models. The larger the image size, the more detail, context information, and pixel information can be captured, helping to accurately capture the subtle features and structure in the image. For LC7 and LC8, the spatial resolution of the MS component is half that of PAN’s. In other cases, the MS component’s spatial resolution is a quarter of PAN’s. For satellites like WV-2 and WV-3, which provide eight multispectral (MS) bands, we selected four bands to form the unified dataset. These bands are Red, Green, Blue, and near-infrared, chosen for their broad applicability and significance in remote sensing tasks. Ultimately, PanBench comprises 5898 image pairs.

3.3. Scene Classification

The ultimate objective of pansharpening is to use fused images to provide crucial spatial information for downstream tasks such as urban planning, land management, disaster response, and more. Therefore, it is necessary to verify the accuracy of the pansharpening method in various scenarios. In this study, by labeling the training data of different scenarios, we reclassified the data into six basic land cover types according to DeepGlobe 2018 [37], Cheng [38], and Meng [47]: water, urban, ice/snow, crops, vegetation, and barren. The specific sample counts are shown in Figure 3. The algorithm’s performance in different scenes may differ, so optimizing the model’s parameters and the algorithm design is necessary by verifying and evaluating different scenes.

4. Methodology

4.1. Overall Framework

4.1.1. Problem Definition

The purpose of pansharpening is to fuse a low-resolution MS image x ms R 4 × H 4 × W 4 with a high-resolution PAN image x pan R 1 × H × W to produce a high-resolution color image y R 4 × H × W with multispectral information, where H and W represent the height and width of the image, respectively. Pansharpening’s task flow can be expressed in the following formula:
y = f θ ( x m s , x p a n ) ,
where f θ is a parameterized neural network model.

4.1.2. Overall Pipeline

The fusion of MS images and PAN images is a key task in remote sensing image processing, aiming to make full use of the spectral information of MS images and the high spatial resolution of PAN images to obtain more comprehensive and high-quality images. In past research, some fusion methods tried to improve the practicability and computational efficiency of the algorithm through complex calculation processes, but often introduced some unnecessary information loss in the processing, especially in the downsampling stage of MS images. The fusion framework proposed in this paper is dedicated to solving these problems, and the core motivation is to simplify the whole fusion process. By upsampling the MS images and avoiding other cumbersome transformations, it tries to preserve the original information of the MS images to the greatest extent. Such a processing method helps prevent the information loss introduced in the downsampling stage and improves the practicability of the algorithm. The upsampled MS images and panchromatic images are fused by keeping the same scale. By keeping the same scale, the information integration between the two becomes more intuitive and effective. This helps to avoid the problems introduced by scale differences, making the fused images more accurate and consistent.
In this paper, we propose CMFNet, a high-fidelity fusion network for pansharpening. It primarily consists of three components: the multiscale MS encoder, the multiscale PAN encoder, and the multiscale fusion autoencoder. The overall pipeline of CMFNet is illustrated in Figure 4, which is described as follows:
  • Multiscale MS encoder: The MS image x m s passes through a convolutional layer with a kernel size of 3 × 3 to obtain the features F s R 4 C × H 4 × W 4 . The features are passed through a multiscale MS encoder, obtaining hierarchical image features of three resolutions.
  • Multiscale PAN encoder: Similar to the multiscale MS encoder, the PAN features F p R C × H × W are pre-extracted without changing the image size, and three features corresponding to the MS scale are obtained by the multiscale PAN encoder.
  • Multiscale fusion autoencoder: The fused features F o R C × H × W are obtained from the features of the three corresponding scales obtained from the multiscale MS encoder and the PAN encoder, respectively. Finally, the fused features are output by a convolution layer.
Figure 4. The overall framework of our proposed CMFNet, where F o represent the output of the autoencoder.
Figure 4. The overall framework of our proposed CMFNet, where F o represent the output of the autoencoder.
Remotesensing 16 02899 g004

4.2. Cascaded Multiscale Fusion Network

4.2.1. Multiscale MS Encoder

Ground objects and scenes have different characteristics at different scales due to their various sizes and shapes. Through multiscale feature extraction, the pansharpening network can focus on local details and global structure at the same time, to better capture the subtle changes and context information of ground objects [48,49]. We introduce the multiscale MS encoder module within CMFNet, as shown in Figure 5a. The MS features F s are processed through three blocks [50] and two upsampling operations, yielding three features at distinct resolutions, denoted as S = S i R ( C × 2 3 i ) × H 2 ( 3 i ) × W 2 ( 3 i ) | i [ 1 , 2 , 3 ] , where the upsampling operation denotes the interpolation method.

4.2.2. Multiscale PAN Encoder

PAN images have higher spatial resolution than MS images. Multiscale feature extraction of PAN images can compensate for the lack of spatial perception of MS images and capture a wide range of structures and spatial relationships from PAN images to better guide the pansharpening process and improve the clarity and quality of the pansharpening results. Therefore, we designed the multiscale PAN encoder module in CMFNet, as shown in Figure 5b. For the PAN image feature F p , three features corresponding to the scale of MS are obtained by three blocks, the same as the MS encoder and two downsampling operations, that is P = P i R ( C × 2 i 1 ) × H 2 ( i 1 ) × W 2 ( i 1 ) | i [ 1 , 2 , 3 ] , where the downsampling operation denotes the pooling layer.

4.2.3. Multiscale Fusion Autoencoder

  • Multiscale fusion encoder: The encoder is responsible for gradually downsampling the input image and extracting high-level semantic features [15,53,54]. To be able to make full use of the information of the images of the multiscale MS encoder and multiscale PAN encoder, it is necessary to deeply fuse S and P at the same scale. After the multiscale MS and PAN encoders, we have two feature sets S and P , representing MS and PAN images, respectively. Since high-resolution MS images must have high spatial and spectral resolutions, their features must have both spatial and spectral information. To do this, the two feature sets must be concatenated and added at the same scale. That is, F 1 = S 3 + P 1 , F 2 = S 2 + P 2 , and F 3 = S 1 + P 3 . Then, the block, the same as the multiscale MS encoder, is used to encode the concatenated feature maps into a more compact representation after each addition, and the end of the multiscale fusion encoder (Figure 5c) is the feature set E = E i R ( C × 2 i 1 ) × H 2 ( i 1 ) × W 2 ( i 1 ) | i [ 1 , 2 , 3 ] , which encodes the spatial and spectral information of the two input images:
    E i = Block ( F i ) ,
    E i = Block ( S 4 i + P i + ( Downsample ( E i 1 ) ) .
  • Multiscale fusion decoder: The decoder (Figure 5c) corresponds to the encoder, and the upsampled feature map is fused with the feature map in the corresponding encoder by the feature fusion operation. This can help to recover the detail and texture information of the image [55]. Specifically, we upsampled the features of the set E and superimposed and fused them with the features of the corresponding scale of E . In encoder downsampling, some details and local information may be lost due to the loss of information or resolution degradation caused by downsampling. Therefore, in the process of decoding E , multiscale injection of S and P can obtain the details and local information from the encoder in the decoder, which can effectively connect and fuse the low-level and high-level features, and finally, output the fusion result F o .

5. Experiments and Results

5.1. Implementation Details

5.1.1. Training Settings

The dataset was divided into training, validation, and test sets in a ratio of 8:1:1, which means the training set comprises 4718 image pairs, and both the validation and test sets contain 590 image pairs each. The batch size was set to 16. We used the Adam [56] optimizer with a learning rate of 0.002. If there is no improvement in the performance of the validation set for ten consecutive epochs, the learning rate is reduced to 10%. Cascaded layers in the multiscale feature fusion were set to 3, and the initial number of channels for the CMFNet was 32. We used the MSE as the object function. The training was performed on four NVIDIA GeForce RTX 3090 GPUs (Santa Clara, CA, USA).

5.1.2. Evaluation Metrics

Since the expected high-resolution MS images are not available, we followed Wald’s protocol and used the original MS remote sensing images as the ground truth to evaluate various methods. The preprocessed PAN and MS remote sensing images were subsampled, respectively, and their resolution was reduced to 1/4 of the original as the input data. The evaluation of the pansharpening algorithms for PAN and MS image fusion involves adapting widely used image quality assessment (IQA) metrics. These metrics include the peak signal-to-noise ratio (PSNR), the structural similarity index (SSIM) [57], the spectral angle mapper (SAM) [58], relative dimensionless global error synthesis (ERGAS) [59], the spatial correlation coefficient (SCC) [60], and the mean-squared error (MSE) (× 10 4 ). These metrics measure fidelity, similarity, spectral and spatial distortion, and spatial correlation. By employing these metrics, the performance of the pansharpening algorithms can be objectively evaluated and compared in real-world scenarios. This evaluation framework enables comprehensive analysis and comparison of different methods for pansharpening. D λ , D s , and the QNR are non-reference metrics used for real-world full-resolution scenes. D λ measures the spectral distortion. D s evaluates the spatial distortion. The QNR quantifies the quality of the fused image without a reference image.
To assess the performance of CMFNet in the task of pansharpening, we have selected eight representative state-of-the-art (SOTA) methods from 2016 to 2023 for comparative analysis. These methods include the PNN [17], PanNet [20], MSDCNN [18], TFNet [21], FusionNet [22], GPPNN [19], SRPPNN [35], and PGCU [28]. We conducted comprehensive experiments on PanBench to evaluate the performance of our model. Then, we performed experiments on various classification scenarios to ensure the generalizability of our model across diverse real-world contexts.

5.1.3. Quantitative Comparison

The comparative results of the nine algorithms on the PanBench are presented in Table 2, with the best values highlighted in bold black. Our proposed method achieved the best overall performance compared to other state-of-the-art pansharpening methods, firmly establishing its superiority. Specifically, in terms of the PSNR, our method outperformed the closest competitor, TFNet, by approximately 1.67. In addition to the PSNR, notable improvements can also be observed in other metrics, indicating reduced spectral distortion and preserved spatial textures. We also evaluated our models on full-resolution images without downsampling them. It is important to note that, in this setting, there will be no target images available for training. Given that the main focus of this experiment is the generalization ability of cascaded multiscale models, we directly applied the optimized networks to the original PAN and MS images to generate the desired high-resolution MS images. The results are reported in Table 2, the best results are in boldface. As evidence, the proposed CMFNet performed well on full-resolution images, achieving competitive results. To conduct a more in-depth comparative analysis, we compared the performance of different methods on each satellite, including both full-resolution and downsampled results, as shown in Table 3 and Table 4. It can be seen that, in the downsampling experiments, the proposed method achieved the best performance in every metric for all satellites. In the full-resolution experiments, the method demonstrated competitive performance across all satellites, with CMFNet achieving the best results on the GF1, WV3, and WV4 images. Additionally, we conducted a comprehensive evaluation for each scene category, as shown in Table 5. CMFNet consistently achieved the best metrics in the downsampled evaluation and demonstrated competitive performance in the full-resolution results. It can be seen that the performance of each metric was optimal due to the simple structure of the water. Other scenes exhibited different performances due to variations in texture structure, consistent with objective facts. This further elucidates the generalizability of the PanBench dataset, capable of accommodating challenges encountered in real-world scenarios.

5.1.4. Qualitative Comparison

We also present qualitative comparisons of the visualization results to demonstrate the effectiveness of our method. For the resolution-reduction experiment, the visualization results of 4 of the 10 satellites involved were selected for display. Non-commercial satellites are represented by GF1 and GF2 (as shown in Figure 6), and commercial satellites are represented by QB and IN (as shown in Figure 7). Visualizations of other satellites are also presented when the scene category is displayed. In the scene category, the visualization results of water, urban, ice/snow, crops, vegetation, and barren are highlighted (as shown in Figure 8, Figure 9, Figure 10, Figure 11, Figure 12 and Figure 13). The MAE difference between the output pansharpening results and the ground truth is described, where red indicates a poor generation effect and blue indicates a good generation effect. Compared to other competing methods, our model has smaller spatial and spectral distortions. We note that our proposed approach is closer to the fundamental truth than other comparative approaches. The experimental results prove the effectiveness of this method, which successfully reduces the information redundancy and improves the quality of the pansharpening results. For full-resolution experiments, some typical samples are represented in Figure 14, which clearly shows the appealing results of CMFNet.

5.2. Ablation Studies

5.2.1. Impact of the Multiscale Cascading

By two decoupled MS and PAN encoders, we can obtain three hierarchical multiscale features with dimensions of 64 × 64, 128 × 128, and 256 × 256. The autoencoder seamlessly fuses the corresponding features of the same scale in a cascading manner. As depicted in Table 6, we conducted experiments to investigate the impact of various cascading levels on the fusion outcomes. Notably, the performance of pansharpening exhibits linear growth as the number of cascading layers increases, reaching its optimal state when the number of cascading layers equals three. This compellingly validates the necessity of hierarchical cascading fusion.

5.2.2. Impact of the Cascaded Injection

In the multiscale fusion decoder, we incorporated a cascading strategy for information injection, whereby the spectral information from the MS image and the spatial information from the PAN image are directly injected across the encoder into the decoder. To validate the effectiveness of this strategy, we conducted ablative research, and the experimental results are shown in Table 7. Following the introduction of cascading information injection, the PSNR increased by 0.4164, and the ERGAS decreased by approximately 0.1. The consistent performance improvements on all metrics provide substantial evidence for the effectiveness of this strategy. One possible explanation is that, for the computer vision task of pansharpening, which belongs to low-level visual computations, the cross-connection of feature injection serves as an efficacious means to alleviate the loss of information during the encoding process.

5.2.3. Impact of the Block

Ablation experiments were conducted to assess the impact of different types of blocks on model performance. ResNet [42], NAFNet [51], ConvNeXt [52], and the previously proposed DiffCR block [50] were selected for comparative analysis. As illustrated in Table 8, the experimental results demonstrate that the DiffCR block exhibits superior performance across multiple metrics. However, the performance differences among the various blocks are not substantial, highlighting the overall superiority of our multiscale cascaded network architecture. In CMFNet, the flexibility to interchange any type of block is emphasized; even ResNet blocks yield satisfactory results, while the ConvNeXt, NAFNet, and DiffCR blocks exhibit even more outstanding performance.

5.2.4. Scalability of the PanBench Dataset

It is worth mentioning that the PanBench dataset we constructed is not only suitable for pansharpening tasks in the field of remote sensing, but also supports other general computer vision tasks such as image super-resolution [63,64,65], image colorization [66], and image classification. In addition, we evaluated the performance of CMFNet on image super-resolution and image colorization tasks through ablation experiments, as shown in Table 9. Compared to pansharpening, super-resolution [67] reconstruction only receives PAN images as the input, while colorization only receives MS images as the input. As can be seen from Table 9, colorization is the most challenging task, followed by super-resolution and then pansharpening.

6. Conclusions

This paper presents PanBench, a large-scale, high-resolution, and multi-scene dataset encompassing the prominent satellites commonly used for pansharpening. The dataset is made available in an open-source manner, to facilitate the development of novel pansharpening methods. To achieve high-fidelity synthesis, we propose CMFNet, a new model designed explicitly for panchromatic sharpening. Experimental results on visual restoration and semantic recovery quality demonstrate the effectiveness of the proposed approach, surpassing existing representative pansharpening methods. The experimental findings also indicate the algorithm’s strong generalization capability and impressive performance. We firmly believe that the PanBench dataset will benefit the community, while our evaluations provide valuable directions for future research endeavors.

Author Contributions

Conceptualization, S.W. and X.Z.; methodology, S.W., X.Z., and P.T.; software, S.W. and X.Z.; the investigation, S.W. and X.Z.; writing—original draft preparation, S.W., K.L. and J.X.; writing—review and editing, S.W., T.C. and P.T.; project administration, T.C.; funding acquisition, T.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of Qinghai Province (No. 2024-ZJ-708).

Data Availability Statement

The original contributions presented in the study are included in the article, further inquiries can be directed to the corresponding author.

Acknowledgments

This research was supported by the High-performance computing center of Qinghai University. Thanks to the Data Processing Intersections team of the School of Computer Science for their support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, H.; Ma, J. GTP-PNet: A residual learning network based on gradient transformation prior for pansharpening. ISPRS-J. Photogramm. Remote Sens. 2021, 172, 223–239. [Google Scholar] [CrossRef]
  2. Cao, X.; Yao, J.; Xu, Z.; Meng, D. Hyperspectral Image Classification with Convolutional Neural Network and Active Learning. IEEE Trans. Geosci. Remote Sens. 2020, 58, 4604–4616. [Google Scholar] [CrossRef]
  3. Lv, Z.; Zhang, P.; Sun, W.; Benediktsson, J.A.; Li, J.; Wang, W. Novel Adaptive Region Spectral-Spatial Features for Land Cover Classification with High Spatial Resolution Remotely Sensed Imagery. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5609412. [Google Scholar] [CrossRef]
  4. Liu, J.; Li, S.; Zhou, C.; Cao, X.; Gao, Y.; Wang, B. SRAF-Net: A Scene-Relevant Anchor-Free Object Detection Network in Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5405914. [Google Scholar] [CrossRef]
  5. Li, X.; Wang, Z.; Zhang, B.; Sun, F.; Hu, X. Recognizing object by components with human prior knowledge enhances adversarial robustness of deep neural networks. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 8861–8873. [Google Scholar] [CrossRef] [PubMed]
  6. Asokan, A.; Anitha, J. Change detection techniques for remote sensing applications: A survey. Earth Sci. Inform. 2019, 12, 143–160. [Google Scholar] [CrossRef]
  7. Wu, C.; Du, B.; Zhang, L. Fully convolutional change detection framework with generative adversarial network for unsupervised, weakly supervised and regional supervised change detection. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 9774–9788. [Google Scholar] [CrossRef] [PubMed]
  8. Cao, X.; Fu, X.; Xu, C.; Meng, D. Deep spatial-spectral global reasoning network for hyperspectral image denoising. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5504714. [Google Scholar] [CrossRef]
  9. Jiang, K.; Wang, Z.; Yi, P.; Jiang, J.; Xiao, J.; Yao, Y. Deep distillation recursive network for remote sensing imagery super-resolution. Remote Sens. 2018, 10, 1700. [Google Scholar] [CrossRef]
  10. Wang, Z.; Jiang, K.; Yi, P.; Han, Z.; He, Z. Ultra-dense GAN for satellite imagery super-resolution. Neurocomputing 2020, 398, 328–337. [Google Scholar] [CrossRef]
  11. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image Super-Resolution Using Deep Convolutional Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef]
  12. Girshick, R. Fast r-cnn. In Proceedings of the ICCV, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  13. Shao, Z.; Lu, Z.; Ran, M.; Fang, L.; Zhou, J.; Zhang, Y. Residual Encoder–Decoder Conditional Generative Adversarial Network for Pansharpening. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1573–1577. [Google Scholar] [CrossRef]
  14. Ma, J.; Yu, W.; Chen, C.; Liang, P.; Guo, X.; Jiang, J. Pan-GAN: An unsupervised pansharpening method for remote sensing image fusion. Inf. Fusion 2020, 62, 110–120. [Google Scholar] [CrossRef]
  15. Hu, X.; Li, K.; Zhang, W.; Luo, Y.; Lemercier, J.M.; Gerkmann, T. Speech separation using an asynchronous fully recurrent convolutional neural network. NeuralIPS 2021, 34, 22509–22522. [Google Scholar]
  16. Li, K.; Yang, R.; Hu, X. An efficient encoder-decoder architecture with top-down attention for speech separation. arXiv 2022, arXiv:2209.15200. [Google Scholar]
  17. Masi, G.; Cozzolino, D.; Verdoliva, L.; Scarpa, G. Pansharpening by Convolutional Neural Networks. Remote Sens. 2016, 8, 594. [Google Scholar] [CrossRef]
  18. Yuan, Q.; Wei, Y.; Meng, X.; Shen, H.; Zhang, L. A Multiscale and Multidepth Convolutional Neural Network for Remote Sensing Imagery Pan-Sharpening. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 978–989. [Google Scholar] [CrossRef]
  19. Xu, S.; Zhang, J.; Zhao, Z.; Sun, K.; Liu, J.; Zhang, C. Deep Gradient Projection Networks for Pan-Sharpening. In Proceedings of the CVPR, Nashville, TN, USA, 20–25 June 2021; pp. 1366–1375. [Google Scholar]
  20. Yang, J.; Fu, X.; Hu, Y.; Huang, Y.; Ding, X.; Paisley, J. PanNet: A Deep Network Architecture for Pan-Sharpening. In Proceedings of the ICCV, Venice, Italy, 22–29 October 2017; pp. 5449–5457. [Google Scholar]
  21. Liu, X.; Liu, Q.; Wang, Y. Remote sensing image fusion based on two-stream fusion network. Inf. Fusion 2020, 55, 1–15. [Google Scholar] [CrossRef]
  22. Deng, L.J.; Vivone, G.; Jin, C.; Chanussot, J. Detail Injection-Based Deep Convolutional Neural Networks for Pansharpening. IEEE Trans. Geosci. Remote Sens. 2020, 59, 6995–7010. [Google Scholar] [CrossRef]
  23. Zhou, M.; Huang, J.; Yan, K.; Yu, H.; Fu, X.; Liu, A.; Wei, X.; Zhao, F. Spatial-Frequency Domain Information Integration for Pan-Sharpening. In Proceedings of the ECCV, Tel Aviv, Israel, 23–27 October 2022; Springer: Cham, Switzerland, 2022; pp. 274–291. [Google Scholar]
  24. Zhou, M.; Yan, K.; Huang, J.; Yang, Z.; Fu, X.; Zhao, F. Mutual Information-Driven Pan-Sharpening. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022; pp. 1798–1808. [Google Scholar]
  25. Pohl, C.; Moellmann, J.; Fries, K. Standardizing quality assessment of fused remotely sensed images. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2017, 42, 863–869. [Google Scholar] [CrossRef]
  26. Liu, Q.; Zhou, H.; Xu, Q.; Liu, X.; Wang, Y. PSGAN: A Generative Adversarial Network for Remote Sensing Image Pan-Sharpening. IEEE Trans. Geosci. Remote Sens. 2020, 59, 10227–10242. [Google Scholar] [CrossRef]
  27. Jin, X.; Huang, S.; Jiang, Q.; Lee, S.J.; Wu, L.; Yao, S. Semisupervised Remote Sensing Image Fusion Using Multiscale Conditional Generative Adversarial Network with Siamese Structure. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 7066–7084. [Google Scholar] [CrossRef]
  28. Zhu, Z.; Cao, X.; Zhou, M.; Huang, J.; Meng, D. Probability-Based Global Cross-Modal Upsampling for Pansharpening. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 14039–14048. [Google Scholar]
  29. Meng, Q.; Shi, W.; Li, S.; Zhang, L. PanDiff: A Novel Pansharpening Method Based on Denoising Diffusion Probabilistic Model. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5611317. [Google Scholar] [CrossRef]
  30. Liu, Q.; Meng, X.; Shao, F.; Li, S. Supervised-unsupervised combined deep convolutional neural networks for high-fidelity pansharpening. Inf. Fusion 2023, 89, 292–304. [Google Scholar] [CrossRef]
  31. Zhou, B.; Shao, F.; Meng, X.; Fu, R.; Ho, Y.S. No-Reference Quality Assessment for Pansharpened Images via Opinion-Unaware Learning. IEEE Access 2019, 7, 40388–40401. [Google Scholar] [CrossRef]
  32. Agudelo-Medina, O.A.; Benitez-Restrepo, H.D.; Vivone, G.; Bovik, A. Perceptual quality assessment of pansharpened images. Remote Sens. 2019, 11, 877. [Google Scholar] [CrossRef]
  33. Vivone, G.; Dalla Mura, M.; Garzelli, A.; Pacifici, F. A benchmarking protocol for pansharpening: Dataset, preprocessing, and quality assessment. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 6102–6118. [Google Scholar] [CrossRef]
  34. Meng, X.; Xiong, Y.; Shao, F.; Shen, H.; Sun, W.; Yang, G.; Yuan, Q.; Fu, R.; Zhang, H. A large-scale benchmark data set for evaluating pansharpening performance: Overview and implementation. IEEE Geosci. Remote Sens. Mag. 2020, 9, 18–52. [Google Scholar] [CrossRef]
  35. Cai, J.; Huang, B. Super-Resolution-Guided Progressive Pansharpening Based on a Deep Convolutional Neural Network. IEEE Trans. Geosci. Remote Sens. 2020, 59, 5206–5220. [Google Scholar] [CrossRef]
  36. Gastineau, A.; Aujol, J.F.; Berthoumieu, Y.; Germain, C. Generative Adversarial Network for Pansharpening with Spectral and Spatial Discriminators. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4401611. [Google Scholar] [CrossRef]
  37. Demir, I.; Koperski, K.; Lindenbaum, D.; Pang, G.; Huang, J.; Basu, S.; Hughes, F.; Tuia, D.; Raskar, R. DeepGlobe 2018: A Challenge to Parse the Earth through Satellite Images. In Proceedings of the CVPRW, Salt Lake City, UT, USA, 18–22 June 2018; pp. 172–181. [Google Scholar]
  38. Cheng, G.; Han, J.; Lu, X. Remote Sensing Image Scene Classification: Benchmark and State of the Art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef]
  39. Li, S.; Yang, B. A new pansharpening method using a compressed sensing technique. TGRS 2010, 49, 738–746. [Google Scholar]
  40. Huang, W.; Xiao, L.; Wei, Z.; Liu, H.; Tang, S. A New Pan-Sharpening Method with Deep Neural Networks. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1037–1041. [Google Scholar] [CrossRef]
  41. Wald, L.; Ranchin, T.; Mangolini, M. Fusion of satellite images of different spatial resolutions: Assessing the quality of resulting images. Photogramm. Eng. Remote Sens. 1997, 63, 691–699. [Google Scholar]
  42. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the CVPR, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  43. Verde, N.; Mallinis, G.; Tsakiri-Strati, M.; Georgiadis, C.; Patias, P. Assessment of Radiometric Resolution Impact on Remote Sensing Data Classification Accuracy. Remote Sens. 2018, 10, 1267. [Google Scholar] [CrossRef]
  44. Lin, S.; Gu, J.; Yamazaki, S.; Shum, H.Y. Radiometric calibration from a single image. In Proceedings of the CVPR, Washington, DC, USA, 27 June–2 July 2004; Volume 2, p. II. [Google Scholar]
  45. Liang, S.; Fang, H.; Chen, M. Atmospheric correction of Landsat ETM+ land surface imagery. I. Methods. IEEE Trans. Geosci. Remote Sens. 2001, 39, 2490–2498. [Google Scholar] [CrossRef]
  46. Barzaghi, R.; Carrion, D.; Carroccio, M.; Maseroli, R.; Stefanelli, G.; Venuti, G. Gravity corrections for the updated Italian levelling network. Appl. Geomat. 2023, 15, 773–780. [Google Scholar] [CrossRef]
  47. Meng, Q.; Zhao, M.; Zhang, L.; Shi, W.; Su, C.; Bruzzone, L. Multilayer Feature Fusion Network with Spatial Attention and Gated Mechanism for Remote Sensing Scene Classification. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6510105. [Google Scholar] [CrossRef]
  48. Li, K.; Yang, R.; Sun, F.; Hu, X. IIANet: An Intra-and Inter-Modality Attention Network for Audio-Visual Speech Separation. In Proceedings of the Forty-first International Conference on Machine Learning (ICML), Vienna, Austria, 7–12 July 2024. [Google Scholar]
  49. Li, K.; Luo, Y. Subnetwork-To-Go: Elastic Neural Network with Dynamic Training and Customizable Inference. In Proceedings of the ICASSP, Seoul, Republic of Korea, 22–27 May 2024; pp. 6775–6779. [Google Scholar]
  50. Zou, X.; Li, K.; Xing, J.; Zhang, Y.; Wang, S.; Jin, L.; Tao, P. DiffCR: A Fast Conditional Diffusion Framework for Cloud Removal from Optical Satellite Images. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5612014. [Google Scholar] [CrossRef]
  51. Chen, L.; Chu, X.; Zhang, X.; Sun, J. Simple Baselines for Image Restoration. In Proceedings of the ECCV, Tel Aviv, Israel, 23–27 October 2022; pp. 17–33. [Google Scholar]
  52. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A ConvNet for the 2020s. In Proceedings of the CVPR, Tel Aviv, Israel, 23–27 October 2022; pp. 11976–11986. [Google Scholar]
  53. Li, K.; Chen, G. Spmamba: State-space model is all you need in speech separation. arXiv 2024, arXiv:2404.02063. [Google Scholar]
  54. Li, K.; Xie, F.; Chen, H.; Yuan, K.; Hu, X. An audio-visual speech separation model inspired by cortico-thalamo-cortical circuits. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 10, 76–80. [Google Scholar] [CrossRef] [PubMed]
  55. Li, K.; Luo, Y. On the design and training strategies for RNN-based online neural speech separation systems. In Proceedings of the ICASSP, Rhodes, Greece, 4–9 June 2023; pp. 1–5. [Google Scholar]
  56. Liu, M.; Zhang, W.; Orabona, F.; Yang, T. Adam: A Stochastic Method with Adaptive Variance Reduction. arXiv 2021, arXiv:2011.11985. [Google Scholar]
  57. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  58. Yuhas, R.H.; Goetz, A.F.; Boardman, J.W. Discrimination among semi-arid landscape endmembers using the Spectral Angle Mapper (SAM) algorithm. In Proceedings of the JPL Airborne Geoscience Workshop, Pasadena, CA, USA, 1–5 June 1992; Volume 1, pp. 147–149. [Google Scholar]
  59. Renza, D.; Martinez, E.; Arquero, A. A New Approach to Change Detection in Multispectral Images by Means of ERGAS Index. IEEE Geosci. Remote Sens. Lett. 2012, 10, 76–80. [Google Scholar] [CrossRef]
  60. Zhou, J.; Civco, D.L.; Silander, J.A. A wavelet transform method to merge Landsat TM and SPOT panchromatic data. Int. J. Remote Sens. 1998, 19, 743–757. [Google Scholar] [CrossRef]
  61. Aiazzi, B.; Baronti, S.; Selva, M. Improving component substitution pansharpening through multivariate regression of MS + Pan data. IEEE Trans. Geosci. Remote Sens. 2007, 45, 3230–3239. [Google Scholar] [CrossRef]
  62. Vivone, G.; Restaino, R.; Chanussot, J. Full scale regression-based injection coefficients for panchromatic sharpening. IEEE Trans. Image Process. 2018, 27, 3418–3431. [Google Scholar] [CrossRef]
  63. Huang, J.; Li, K.; Wang, X. Single image super-resolution reconstruction of enhanced loss function with multi-gpu training. In Proceedings of the ISPA/BDCloud/SocialCom/Su, Xiamen, China, 16–18 December 2019; pp. 559–565. [Google Scholar]
  64. Saharia, C.; Ho, J.; Chan, W.; Salimans, T.; Fleet, D.J.; Norouzi, M. Image Super-Resolution via Iterative Refinement. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 4713–4726. [Google Scholar] [CrossRef]
  65. Li, K.; Yang, S.; Dong, R.; Wang, X.; Huang, J. Survey of single image super-resolution reconstruction. IET Image Process. 2020, 14, 2273–2290. [Google Scholar] [CrossRef]
  66. Saharia, C.; Chan, W.; Chang, H.; Lee, C.; Ho, J.; Salimans, T.; Fleet, D.; Norouzi, M. Palette: Image-to-Image Diffusion Models. In Proceedings of the ACM SIGGRAPH, Vancouver, BC, Canada, 7–11 August 2022; pp. 1–10. [Google Scholar]
  67. Zhou, M.; Yan, K.; Pan, J.; Ren, W.; Xie, Q.; Cao, X. Memory-augmented deep unfolding network for guided image super-resolution. Int. J. Comput. Vis. 2023, 131, 215–242. [Google Scholar] [CrossRef]
Figure 1. All mainstream satellites for pansharpening are supported by our PanBench dataset, along with their respective sample quantities.
Figure 1. All mainstream satellites for pansharpening are supported by our PanBench dataset, along with their respective sample quantities.
Remotesensing 16 02899 g001
Figure 2. Data preprocessing flow chart.
Figure 2. Data preprocessing flow chart.
Remotesensing 16 02899 g002
Figure 3. Number of the six scenes of land cover classification.
Figure 3. Number of the six scenes of land cover classification.
Remotesensing 16 02899 g003
Figure 5. The detailed structure of the three components in CMFNet. The upward and downward arrows indicate upsampling and downsampling, respectively, while the plus sign denotes summation. The term “block” can be substituted by any module, such as ResNet [42], NAFNet [51], ConvNeXt [52] block, etc. The autoencoder integrates spectral and spatial information by incorporating identical-scale features from injecting two encoders.
Figure 5. The detailed structure of the three components in CMFNet. The upward and downward arrows indicate upsampling and downsampling, respectively, while the plus sign denotes summation. The term “block” can be substituted by any module, such as ResNet [42], NAFNet [51], ConvNeXt [52] block, etc. The autoencoder integrates spectral and spatial information by incorporating identical-scale features from injecting two encoders.
Remotesensing 16 02899 g005
Figure 6. Visual comparison of GF1 and GF2 images from different pansharpening methods. MAE represents the mean absolute error of each spectral band.
Figure 6. Visual comparison of GF1 and GF2 images from different pansharpening methods. MAE represents the mean absolute error of each spectral band.
Remotesensing 16 02899 g006
Figure 7. Visual comparison of QB and IN images from different pansharpening methods. MAE represents the mean absolute error of each spectral band.
Figure 7. Visual comparison of QB and IN images from different pansharpening methods. MAE represents the mean absolute error of each spectral band.
Remotesensing 16 02899 g007
Figure 8. Visual results generated by different pansharpening methods for water in the scene category (from GF6). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Figure 8. Visual results generated by different pansharpening methods for water in the scene category (from GF6). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Remotesensing 16 02899 g008
Figure 9. Visual results generated by different pansharpening methods for urban in the scene category (from WV4). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Figure 9. Visual results generated by different pansharpening methods for urban in the scene category (from WV4). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Remotesensing 16 02899 g009
Figure 10. Visual results generated by different pansharpening methods for ice/snow in the scene category (from LC7). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Figure 10. Visual results generated by different pansharpening methods for ice/snow in the scene category (from LC7). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Remotesensing 16 02899 g010
Figure 11. Visual results generated by different pansharpening methods for crops in the scene category (from WV3). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Figure 11. Visual results generated by different pansharpening methods for crops in the scene category (from WV3). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Remotesensing 16 02899 g011
Figure 12. Visual results generated by different pansharpening methods for vegetation in the scene category (from WV2). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Figure 12. Visual results generated by different pansharpening methods for vegetation in the scene category (from WV2). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Remotesensing 16 02899 g012
Figure 13. Visual results generated by different pansharpening methods for barren in the scene category (from LC8). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Figure 13. Visual results generated by different pansharpening methods for barren in the scene category (from LC8). The first line of images is displayed in RGB combination; the second line is the MAE difference between the output pansharpening result and the ground truth.
Remotesensing 16 02899 g013
Figure 14. Example results for full-resolution experiments. Displayed in RGB combination.
Figure 14. Example results for full-resolution experiments. Displayed in RGB combination.
Remotesensing 16 02899 g014
Table 1. Comparative analysis of the PanBench dataset against datasets used in other representative literature. (* means multiply by the same value as before, for example 256 * equals 256 × 256.)
Table 1. Comparative analysis of the PanBench dataset against datasets used in other representative literature. (* means multiply by the same value as before, for example 256 * equals 256 × 256.)
MethodPublicationYearGF1GF2GF6LC7LC8WV2WV3WV4QBINPAN
PNN [17]Remote Sens.2016 132 *
PanNet [20]ICCV2017 400 *
MSDCNN [18]J-STARS2018 164 *
TFNet [21]Inform. Fusion2020 512 *
FusionNet [22]TGRS2020 64 *
PSGAN [26]TGRS2020 256 *
GPPNN [19]CVPR2021 128 *
SRPPNN [35]TGRS2021 256 *
MDSSC-GAN [36]TGRS2021 512 *
MIDP [24]CVPR2022 128 *
SFIIN [23]ECCV2022 128 *
PanDiff [29]TGRS2023 64 *
USSCNet [30]Inform. Fusion2023 256 *
PGCU [28]CVPR2023 128 *
CMFNet [Ours]-20241024 *
Table 2. Quantitative metrics for all the comparison methods on PanBench.
Table 2. Quantitative metrics for all the comparison methods on PanBench.
ModelFull ResolutionReduced Resolution
D λ D s QNR ↑PSNR ↑SSIM ↑SAM ↓ERGAS ↓SCC ↑MSE ↓
GSA [61]0.07110.13340.808517.35000.32370.08075.02890.877033.3673
MTF-GLP [62]0.11890.13080.771616.97080.30050.09315.68470.870736.4232
PNN [17]0.05590.12980.822328.90290.78870.07504.49980.899225.5223
PanNet [20]0.06400.11840.831930.14650.84970.07023.80530.923418.5474
MSDCNN [18]0.05420.09810.855729.26750.82370.07614.16770.913921.7871
TFNet [21]0.05650.11050.840432.70180.89310.06002.88160.948611.9169
FusionNet [22]0.09980.16630.750524.43780.71750.08807.50290.791370.0308
GPPNN [19]0.06710.10740.836928.89010.82110.08424.27200.912422.3945
SRPPNN [35]0.05130.09070.864031.31860.86470.06663.35260.934415.5632
PGCU [28]0.11710.09610.799429.96920.82440.07593.91130.916719.5983
CMFNet [Ours]0.05670.10120.851534.49210.91530.05112.39840.96018.8862
Table 3. Quantitative metrics for all the methods on each satellite (non-commercial) of PanBench.
Table 3. Quantitative metrics for all the methods on each satellite (non-commercial) of PanBench.
SatelliteModelReduced ResolutionFull Resolution
PSNR ↑SSIM ↑SAM ↓ERGAS ↓SCC ↑MSE ↓ D λ D s QNR ↑
GF1PNN36.54570.82780.03132.44380.87420.00110.06980.09040.8469
PanNet38.26360.90520.02781.83540.90160.00060.09580.16070.7596
MSDCNN35.57140.84280.03742.49530.87370.00110.07310.20030.7417
TFNet39.45320.93790.02751.51400.94040.00030.08280.17450.7568
FusionNet29.22580.78110.05035.03090.67530.00440.12030.14570.7509
GPPNN33.38890.83600.05703.03940.86470.00120.08770.09510.8263
SRPPNN38.51060.91460.02731.76260.91980.00050.06790.16090.7819
PGCU35.84500.83830.03532.41580.87040.00100.07570.08590.8462
CMFNet42.86530.95910.01961.04580.95800.00010.06860.08660.8517
GF2PNN28.60840.80930.07265.46550.92120.00210.05910.10210.8468
PanNet30.83100.88970.06564.10910.95600.00110.05660.07320.8744
MSDCNN30.01480.86150.07104.66570.94500.00150.05510.07900.8709
TFNet34.12630.93510.05232.88890.97970.00050.06650.07440.8643
FusionNet25.14340.72460.09668.76560.82120.00410.05880.08680.8598
GPPNN30.16460.86660.07254.51820.94720.00140.06330.08220.8605
SRPPNN32.03610.90600.05863.49410.96630.00090.06020.07410.8708
PGCU30.44830.84770.07534.47740.94740.00120.06850.07680.8616
CMFNet36.67370.95540.03942.22470.98700.00030.05500.07710.8725
GF6PNN29.14620.80800.05902.73130.94640.00130.05680.07950.8687
PanNet30.12920.84450.05192.42810.95800.00100.05700.07470.8730
MSDCNN29.67110.83980.06072.59470.95440.00110.05700.07200.8756
TFNet33.08390.90100.04151.75600.97790.00050.05770.08180.8656
FusionNet24.90370.75770.07084.74970.87080.00360.06260.13510.8119
GPPNN29.58260.83830.06042.6056’0.95380.00120.06080.07890.8656
SRPPNN31.34150.86880.04922.11160.96750.00080.05800.07430.8724
PGCU30.41350.83870.05492.37150.96060.00100.05880.06990.8759
CMFNet34.28580.91650.03611.53110.98260.00040.05710.08520.8628
LC7PNN29.41850.84390.01721.96530.97000.00140.03530.07990.8872
PanNet30.92160.89360.01541.65980.97830.00100.03020.07840.8935
MSDCNN30.15640.89010.03081.74920.97990.00110.03000.08990.8824
TFNet35.31500.93100.01450.97340.99000.00030.02880.07810.8950
FusionNet25.62480.80520.01783.07440.92860.00340.03220.12580.8462
GPPNN29.93200.87820.03781.84140.97810.00110.02810.07900.8949
SRPPNN33.86180.91520.01471.15850.98620.00040.03040.07740.8943
PGCU32.28660.89310.01961.39510.98310.00070.02990.08140.8908
CMFNet36.63530.93910.01100.84170.99180.00020.02940.07760.8949
LC8PNN25.82130.77900.10624.77060.90820.00280.09000.13700.7861
PanNet27.50380.85880.09503.78520.94070.00190.06230.10990.8349
MSDCNN26.81830.83810.10154.10620.93260.00220.07810.12830.8040
TFNet29.35480.89270.08082.89300.96290.00120.05390.11620.8363
FusionNet23.43550.73770.11226.73430.83100.00490.08070.07340.8521
GPPNN26.54200.83240.11074.13180.92870.00230.12940.15990.7322
SRPPNN28.06870.86220.09143.34720.95020.00160.05570.11500.8359
PGCU27.37230.84070.09833.67500.94180.00190.10900.15120.7576
CMFNet30.06310.90550.07192.62240.96850.00100.05330.12010.8332
Table 4. Quantitative metrics for all the methods on each satellite (commercial) of PanBench.
Table 4. Quantitative metrics for all the methods on each satellite (commercial) of PanBench.
SatelliteModelReduced ResolutionFull Resolution
PSNR ↑SSIM ↑SAM ↓ERGAS ↓SCC ↑MSE ↓ D λ D s QNR ↑
QBPNN31.37530.89390.07132.96420.96280.00080.05390.10510.8488
PanNet31.62020.90500.06792.91750.96510.00080.05510.10380.8485
MSDCNN30.95980.89350.07253.14800.95920.00090.05000.09890.8572
TFNet35.39110.94570.05921.89940.98510.00030.05500.11340.8397
FusionNet25.41560.79780.08306.16900.86160.00320.04840.11800.8394
GPPNN30.36340.87830.08663.40970.95300.00110.05910.09970.8479
SRPPNN33.92850.93090.06312.21630.97900.00050.05810.11710.8340
PGCU32.31780.89400.07262.79430.96430.00070.06650.12980.8144
CMFNet37.45930.95810.05061.49650.98970.00020.05370.11750.8367
INPNN22.39230.56860.10056.38410.82800.00690.05750.09910.8497
PanNet23.08460.65180.10015.91240.84820.00590.06540.09520.8462
MSDCNN22.54480.60520.10196.25990.83810.00660.07140.11170.8257
TFNet24.16730.69130.09125.28830.86740.00490.06820.10210.8370
FusionNet18.65150.46530.10929.83360.68650.01520.06690.17070.7762
GPPNN22.59940.61580.11066.21290.84000.00650.07140.11570.8216
SRPPNN23.49790.64850.09585.67080.85430.00550.06280.10490.8393
PGCU23.10440.62740.10115.94270.84650.00600.07570.11960.8147
CMFNet25.19650.75490.08364.71920.89600.00390.07230.12190.8152
WV2PNN29.51030.85980.09424.35770.94720.00130.09190.09050.8271
PanNet30.15670.87920.08864.15440.95450.00110.08030.08320.8580
MSDCNN29.08440.85100.09714.61670.94280.00140.09570.08940.8246
TFNet33.37590.92900.06762.83460.97830.00050.08120.08990.8372
FusionNet25.18530.78410.10917.86400.86930.00380.09480.11010.8063
GPPNN28.86120.83770.10064.75380.93630.00150.11040.09470.8068
SRPPNN31.48560.90230.08043.45710.96620.00080.09010.09240.8272
PGCU30.14650.86590.09464.07440.95530.00110.09940.10470.8076
CMFNet34.87660.94240.05772.36610.98390.00040.07850.08740.8420
WV3PNN29.82290.71510.09338.74250.67510.00490.10070.35630.5778
PanNet31.57400.84720.08936.50030.76510.00260.11560.27240.6416
MSDCNN31.13760.82190.08917.02280.75350.00300.09810.27070.6563
TFNet33.00560.90520.08215.16180.8213’0.00160.12130.32060.5934
FusionNet27.91520.69960.103510.10200.61100.00550.12240.19610.7041
GPPNN30.87940.82390.10037.17490.76390.00310.13510.23220.6631
SRPPNN31.88670.85860.09296.29880.77750.00250.11150.29800.6221
PGCU30.50100.79170.10377.27800.73060.00350.11140.23960.7565
CMFNet35.29940.93500.06824.07910.85460.00100.09510.13070.7895
WV4PNN25.77470.78730.12085.36340.95250.00320.06080.08870.8581
PanNet26.20310.81420.11785.14800.95790.00290.05170.07370.8622
MSDCNN25.85970.79380.11235.23530.95570.00310.06400.08970.8538
TFNet28.33490.85390.09923.93410.97430.00180.05470.08070.8712
FusionNet17.10250.58000.139914.13020.73220.02810.05760.23160.7266
GPPNN25.71280.80520.12165.33450.95370.00330.05880.08990.8584
SRPPNN27.54700.83520.11094.34180.96820.00210.05700.08950.8606
PGCU26.40500.81260.11814.91060.96060.00280.05710.10460.8470
CMFNet29.73770.87510.08943.40160.97960.00130.05660.07790.8720
Table 5. Quantitative metrics for all the methods on various scene classification of PanBench.
Table 5. Quantitative metrics for all the methods on various scene classification of PanBench.
SceneModelReduced ResolutionFull Resolution
PSNR ↑SSIM ↑SAM ↓ERGAS ↓SCC ↑MSE ↓ D λ D s QNR ↑
WaterPNN37.38360.89540.04693.71710.82530.00120.09230.11580.8058
PanNet38.04010.91620.04363.17650.83880.00090.11500.14270.7625
MSDCNN36.72250.90350.05053.50420.83320.00110.08760.20300.7297
TFNet38.94140.93390.04632.64330.87900.00060.10520.16520.7494
FusionNet29.90430.82630.06947.97180.63240.01050.13380.23840.6554
GPPNN35.14590.90370.06653.92570.83440.00120.11530.12440.7780
SRPPNN38.59200.92370.04292.96220.85780.00070.09050.15400.7718
PGCU36.87760.90680.05503.36340.82520.00100.10140.15280.7644
CMFNet41.78010.94530.03632.10980.89900.00040.09320.12740.7948
UrbanPNN25.83270.75190.09635.95300.90590.00370.07000.11060.8285
PanNet27.35840.84240.09274.86310.94340.00240.06660.12750.8156
MSDCNN26.45270.79950.09365.36370.92920.00290.06830.12400.8171
TFNet30.48130.90250.07373.47550.97070.00130.06870.14400.7986
FusionNet21.77870.65600.10739.08890.81530.00850.06710.14670.7975
GPPNN26.36700.80100.10265.38580.92800.00300.07960.11930.8118
SRPPNN28.71090.86510.08734.22780.95440.00190.07240.14540.7944
PGCU27.34570.80430.09634.98470.93540.00260.07690.12260.8115
CMFNet32.40340.92990.06322.81820.98070.00090.06530.10290.8393
Ice/snowPNN27.39290.81550.05313.20670.95790.00210.05790.08160.8661
PanNet28.86230.87340.04742.63520.97130.00150.04310.07680.8835
MSDCNN28.36870.86360.05942.82860.96930.00170.04920.08480.8706
TFNet32.94100.91860.03991.71210.98640.00060.04110.07880.8834
FusionNet23.64230.76250.05694.86990.91030.00490.04730.13240.8264
GPPNN28.22100.85530.06632.83150.96820.00170.06240.08830.8570
SRPPNN31.34810.89490.04522.04560.98030.00090.04280.07620.8843
PGCU29.90090.86770.05152.38060.97500.00120.05980.08850.8588
CMFNet34.08100.92900.03421.51320.98890.00050.04140.08210.8800
CropsPNN29.20040.82920.07524.02940.93930.00150.07450.10160.8330
PanNet30.38340.87860.07013.56390.95740.00110.06980.10920.8295
MSDCNN29.11610.83670.07724.07000.94030.00140.07360.09050.8435
TFNet33.79830.92670.05582.40560.98030.00050.07220.11970.8179
FusionNet24.60570.74760.08927.19070.84530.00440.07890.09480.8347
GPPNN28.74920.82480.08634.25150.93430.00160.08680.09070.8314
SRPPNN31.90110.89990.06432.94100.96870.00080.07550.12090.8139
PGCU30.23710.84390.07583.61390.95080.00120.08320.11050.8170
CMFNet35.72720.94490.04641.93830.98660.00030.06920.12030.8199
VegetationPNN26.82140.75460.08665.08170.90210.00310.06170.10430.8415
PanNet28.37240.83200.07994.12490.93440.00210.05750.10550.8435
MSDCNN27.70130.80660.08484.51680.92510.00250.06010.11180.8354
TFNet30.86990.88020.06713.14800.95820.00140.05980.11340.8339
FusionNet23.30240.68340.10027.86400.80620.00670.06260.11650.8283
GPPNN27.63350.80780.09004.48760.92470.00250.07640.11880.8153
SRPPNN29.28230.84510.07583.68860.94340.00190.05870.11040.8381
PGCU28.19400.80300.08564.29160.92900.00230.07350.10890.8272
CMFNet32.42670.90450.05782.65510.96830.00110.05680.11730.8331
BarrenPNN27.87550.76350.06233.44200.92180.00250.06010.08960.8566
PanNet28.92470.81700.05773.07230.93610.00200.05700.08340.8650
MSDCNN28.24460.79880.06703.30010.93050.00230.06000.08990.8562
TFNet31.73990.85950.05102.41640.95320.00140.05690.08700.8616
FusionNet23.76210.70050.06985.56690.83350.00570.05800.14110.8097
GPPNN28.02830.79310.07413.38380.92840.00230.06690.09350.8471
SRPPNN30.42270.83130.05542.71180.94410.00170.05660.08580.8631
PGCU29.26010.80110.06123.05360.93490.00200.06830.09570.8443
CMFNet32.97570.88480.04472.12150.96320.00110.05810.09550.8526
Table 6. Impact of cascaded layers in the multiscale feature fusion.
Table 6. Impact of cascaded layers in the multiscale feature fusion.
Cascading LevelPSNR ↑SSIM ↑SAM ↓ERGAS ↓SCC ↑ MSE ↓
133.03520.89370.05792.78460.949410.9749
234.38520.91390.05162.42660.95959.0428
334.49210.91530.05112.39840.96018.8862
Table 7. Ablation study results on cascaded information injection.
Table 7. Ablation study results on cascaded information injection.
Cascading InjectionPSNR ↑SSIM ↑SAM ↓ERGAS ↓SCC ↑MSE ↓
×34.07570.91020.05282.49260.95689.5769
34.49210.91530.05112.39840.96018.8862
Table 8. The performance of different types of blocks in the model.
Table 8. The performance of different types of blocks in the model.
BlockPSNR ↑SSIM ↑SAM ↓ERGAS ↓SCC ↑MSE ↓
ResNet [42]33.55970.90250.05642.63690.952610.2362
ConvNeXt [52]33.86390.90620.05462.56770.95589.9013
NAFNet [51]34.35110.91470.05222.45900.95808.9712
DiffCR [50]34.49210.91530.05112.39840.96018.8862
Table 9. Ablation study results on the scalability of PanBench dataset. Besides pansharpening, it also supports tasks such as image super-resolution and image colorization.
Table 9. Ablation study results on the scalability of PanBench dataset. Besides pansharpening, it also supports tasks such as image super-resolution and image colorization.
TaskPSNR ↑SSIM ↑SAM ↓ERGAS ↓SCC ↑MSE ↓
Super-resolution (w/o PAN)29.62830.76560.325625.96040.820525.3851
Colorization (w/o MS)25.19880.78110.16867.98950.778250.1688
Pansharpening (MS+PAN)34.38520.91390.05792.78460.94949.0428
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, S.; Zou, X.; Li, K.; Xing, J.; Cao, T.; Tao, P. Towards Robust Pansharpening: A Large-Scale High-Resolution Multi-Scene Dataset and Novel Approach. Remote Sens. 2024, 16, 2899. https://doi.org/10.3390/rs16162899

AMA Style

Wang S, Zou X, Li K, Xing J, Cao T, Tao P. Towards Robust Pansharpening: A Large-Scale High-Resolution Multi-Scene Dataset and Novel Approach. Remote Sensing. 2024; 16(16):2899. https://doi.org/10.3390/rs16162899

Chicago/Turabian Style

Wang, Shiying, Xuechao Zou, Kai Li, Junliang Xing, Tengfei Cao, and Pin Tao. 2024. "Towards Robust Pansharpening: A Large-Scale High-Resolution Multi-Scene Dataset and Novel Approach" Remote Sensing 16, no. 16: 2899. https://doi.org/10.3390/rs16162899

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop